This repo collects papers, repos, tools for large model system, including training, inference, serving and compression.
Year | Publisher | Title | Framework |
---|---|---|---|
Year | Pub | Title | Framework |
Year | Pub | Title1 | Framework |
Year | Publisher | Title | Framework |
---|---|---|---|
Year | Pub | Title | Framework |
Year | Pub | Title1 | Framework |
Year | Name | Training | Inference | Serving | Comments |
---|---|---|---|---|---|
2023 | EnergonAI | ✗ | ✔ | ✗ | |
2022 | Alpa | ✔ | ✔ | ✔ | Compilation based mixed parallelism |
2021 | Megatron-DeepSpeed | ✔ | ✗ | ✗ | Add MoE model training, Curriculum Learning, 3D Parallelism from DeepSpeed to Megatron |
2021 | TeraPipe | ✔ | ✗ | ✗ | |
2021 | ColossalAI | ✔ | ✔ | ✔ | |
2021 | FasterTransformer | ✗ | ✔ | ✗ | |
2020 | DeepSpeed | ✔ | ✔ | ✗ | General Support of Transformers and MoE with 3d-parallelism |
2019 | Megatron-LM | ✔ | ✗ | ✗ | |
2019 | PipeDream | ✔ | ✗ | ✗ | |
2019 | TorchGipe | ✔ | ✗ | ✗ | The torchgipe has been merged to PyTorch in 2020. |
2019 | PipeDream | ✔ | ✗ | ✗ |