Skip to content

Version v0.3.1 Release Today!

Compare
Choose a tag to compare
@github-actions github-actions released this 01 Aug 07:02
8064771

What's Changed

Release

Chat

Zero

  • [zero] optimize the optimizer step time (#4221) by LuGY
  • [zero] support shard optimizer state dict of zero (#4194) by LuGY
  • [zero] add state dict for low level zero (#4179) by LuGY
  • [zero] allow passing process group to zero12 (#4153) by LuGY
  • [zero]support no_sync method for zero1 plugin (#4138) by LuGY
  • [zero] refactor low level zero for shard evenly (#4030) by LuGY

Nfc

  • [NFC] polish applications/Chat/coati/models/utils.py codestyle (#4277) by yuxuan-lou
  • [NFC] polish applications/Chat/coati/trainer/strategies/base.py code style (#4278) by Zirui Zhu
  • [NFC] polish applications/Chat/coati/models/generation.py code style (#4275) by RichardoLuo
  • [NFC] polish applications/Chat/inference/server.py code style (#4274) by Yuanchen
  • [NFC] fix format of application/Chat/coati/trainer/utils.py (#4273) by アマデウス
  • [NFC] polish applications/Chat/examples/train_reward_model.py code style (#4271) by Xu Kai
  • [NFC] fix: format (#4270) by dayellow
  • [NFC] polish runtime_preparation_pass style (#4266) by Wenhao Chen
  • [NFC] polish unary_elementwise_generator.py code style (#4267) by YeAnbang
  • [NFC] polish applications/Chat/coati/trainer/base.py code style (#4260) by shenggan
  • [NFC] polish applications/Chat/coati/dataset/sft_dataset.py code style (#4259) by Zheng Zangwei (Alex Zheng)
  • [NFC] polish colossalai/booster/plugin/low_level_zero_plugin.py code style (#4256) by 梁爽
  • [NFC] polish colossalai/auto_parallel/offload/amp_optimizer.py code style (#4255) by Yanjia0
  • [NFC] polish colossalai/cli/benchmark/utils.py code style (#4254) by ocd_with_naming
  • [NFC] policy applications/Chat/examples/ray/mmmt_prompt.py code style (#4250) by CZYCW
  • [NFC] polish applications/Chat/coati/models/base/actor.py code style (#4248) by Junming Wu
  • [NFC] polish applications/Chat/inference/requirements.txt code style (#4265) by Camille Zhong
  • [NFC] Fix format for mixed precision (#4253) by Jianghai
  • [nfc]fix ColossalaiOptimizer is not defined (#4122) by digger yu
  • [nfc] fix dim not defined and fix typo (#3991) by digger yu
  • [nfc] fix typo colossalai/zero (#3923) by digger yu
  • [nfc]fix typo colossalai/pipeline tensor nn (#3899) by digger yu
  • [nfc] fix typo colossalai/nn (#3887) by digger yu
  • [nfc] fix typo colossalai/cli fx kernel (#3847) by digger yu

Example

Ci

Checkpointio

Lazy

Kernels

  • [Kernels] added triton-implemented of self attention for colossal-ai (#4241) by Cuiqing Li

Docker

Dtensor

Workflow

Cli

Format

Shardformer

Test

Device

  • [device] support init device mesh from process group (#3990) by Frank Lee

Hotfix

Doc

Examples

Testing

  • [testing] move pytest to be inside the function (#4087) by Frank Lee

Gemini

  • Merge pull request #4056 from Fridge003/hotfix/fix_gemini_chunk_config_searching by Baizhou Zhang
  • [gemini] fix argument naming during chunk configuration searching by Baizhou Zhang
  • [gemini] fixed the gemini checkpoint io (#3934) by Frank Lee
  • [gemini] fixed the gemini checkpoint io (#3934) by Frank Lee

Devops

Sync

  • Merge pull request #4025 from hpcaitech/develop by Frank Lee
  • Merge pull request #3967 from ver217/update-develop by Frank Lee
  • Merge pull request #3942 from hpcaitech/revert-3931-sync/develop-to-shardformer by FoolPlayer
  • Revert "[sync] sync feature/shardformer with develop" by Frank Lee
  • Merge pull request #3931 from FrankLeeeee/sync/develop-to-shardformer by FoolPlayer
  • Merge pull request #3916 from FrankLeeeee/sync/dtensor-with-develop by Frank Lee
  • Merge pull request #3915 from FrankLeeeee/update/develop by Frank Lee

Booster

Evaluate

  • [evaluate] support gpt evaluation with reference (#3972) by Yuanchen

Feature

  • Merge pull request #3926 from hpcaitech/feature/dtensor by Frank Lee

Bf16

Evaluation

Full Changelog: v0.3.1...v0.3.0