
DeepSpeed
@deepspeedai
Official account for DeepSpeed, a library that enables unprecedented scale and speed for deep learning training + inference.
日本語 : @DeepSpeedAI_JP
ID: 1262854060320755715
https://www.deepspeed.ai/ 19-05-2020 21:14:20
81 Tweet
3,3K Followers
88 Following



So you've had your fun with Andrej Karpathy 's mingpt. Now its time to scale : introducing min-max-gpt: really small codebase that scales with help of @MSFTDeepSpeed . No huggingface accelerate, transformer. Just deepspeed + torch: maximum hackability github.com/cloneofsimo/mi…


#DeepSpeed joins forces with University of Sydney to unveil an exciting tech #FP6. Just supply your FP16 models, and we deliver: 🚀 1.5x performance boost for #LLMs serving on #GPUs 🚀 Innovative (4+2)-bit system design 🚀 Quality-preserving quantization link: github.com/microsoft/Deep…



Introducing DeepNVMe, a suite of optimizations for fast and efficient I/O operations in DL applications. - POSIX-style APIs - Direct HBM/NVMe xfers via NVIDIA GDS - Cheap Inference scaling via NVMe-Offload Blog: shorturl.at/l7Oue Microsoft Azure NVIDIA Data Center #FMS24 #GPUDirect


💡Check out Comet’s latest integration with DeepSpeed, a deep learning optimization library! 🤝With the @MSFTDeepSpeed + Comet integration automatically start logging training metrics generated by DeepSpeed. Try the quick-start Colab to get started: colab.research.google.com/github/comet-m…


Great to see the amazing DeepSpeed optimizations from Guanhua Wang, Heyang Qin, Masahiro Tanaka, Quentin Anthony, and Sam Ade Jacobs presented by Ammar Ahmad Awan at MUG '24.









