Releases: microsoft/DeepSpeed
Releases · microsoft/DeepSpeed
v0.5.3: Patch release
[zero_to_fp32] adapt to 4-bytes alignment in z2 (#1372) Co-authored-by: Olatunji Ruwase <olruwase@microsoft.com>
v0.5.2: Patch release
Update setup.py (#1361) updated classifiers
v0.5.1: Patch release
Reducing the memory-overhead of creating model for multi-GPU run (#1244) Co-authored-by: Jeff Rasley <jerasley@microsoft.com>
DeepSpeed v0.5.0
- Mixture of Experts (MoE) support
- Curriculum learning
v0.4.5: Patch release
Use correct input size for splits (#1284) * Use correct input size for splits * Use smarter partitioning
v0.4.4: Patch release
[Doc] round_robin_gradients (#1261) * Fix docstring * Make screenshots clickable for easier viewing * Navigation menu in alphabetical order; More clicable screenshots * Rename 1Cycle doc * Tweak naming * Remove no longer used flag * ZeRO3 Offload release * Single GPU results * Rearrange figures * Single GPU text * tweak intro * zero3-offload section * Add asynchronous i/o docs * Fix print_per_steps doc * Document round_robin_gradients * Tweak description * Trigger CI
v0.4.3: Patch release
revert part of #1220 (#1221) https://github.com/microsoft/DeepSpeed/pull/1220 fixed the leak, but lead to another problem. reverting that part so that we could do release and will work on it after the release. @jeffra
v0.4.2: Patch release
clean up logging (#1190) Co-authored-by: Jeff Rasley <jerasley@microsoft.com>
v0.4.1: Patch release
remove torchvision dependency (#1178)
DeepSpeed v0.4.0
DeepSpeed v0.4.0
- [Press release] DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression
- New inference API inference setup
- DeepSpeed Inference: Multi-GPU inference with customized inference kerenls and quantization support
- Mixture-of-Quantization: A novel quantization approach for reducing model size with minimal accuracy impact
- MoQ tutorial for more details.