DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective. DeepSpeed delivers extreme-scale model training for everyone, from data scientists training on massive supercomputers to those training on low-end clusters or even on a single GPU. Using current generation of GPU clusters with hundreds of devices, 3D parallelism of DeepSpeed can efficiently train deep learning models with trillions of parameters. With just a single GPU, ZeRO-Offload of DeepSpeed can train models with over 10B parameters, 10x bigger than the state of arts, democratizing multi-billion-parameter model training such that many deep learning scientists can explore bigger and better models. Sparse attention of DeepSpeed powers an order-of-magnitude longer input sequence and obtains up to 6x faster execution comparing with dense transformers.

Features

  • 10x larger models and 10x faster training
  • Minimal code change
  • Extremely memory efficient
  • Extremely long sequence length
  • Extremely communication efficient
  • An initiative to enable next-generation AI capabilities at scale

Project Samples

Project Activity

See All Activity >

License

MIT License

Follow DeepSeed

DeepSeed Web Site

Other Useful Business Software
Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
Try Free
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of DeepSeed!

Additional Project Details

Operating Systems

Windows

Programming Language

Python

Related Categories

Python Libraries, Python Machine Learning Software, Python Deep Learning Frameworks

Registered

2021-09-23