Skip to content

Latest commit

 

History

History
3 lines (2 loc) · 261 Bytes

220114 DeepSpeed-MoE.md

File metadata and controls

3 lines (2 loc) · 261 Bytes

https://arxiv.org/abs/2201.05596

DeepSpeed-MoE: Advancing Mixture-of-Experts Inference and Training to Power Next-Generation AI Scale (Samyam Rajbhandari, Conglong Li, Zhewei Yao, Minjia Zhang, Reza Yazdani Aminabadi, Ammar Ahmad Awan, Jeff Rasley, Yuxiong He)