275 Commits (6991819a979292b4735b4c3dc30da5265b7841b7)

Author SHA1 Message Date
pre-commit-ci[bot] 7c2f79fa98
[pre-commit.ci] pre-commit autoupdate (#5572) 5 months ago
Haze188 416580b314
[MoE/ZeRO] Moe refactor with zero refactor (#5821) 5 months ago
binmakeswell 4ccaaaab63
[doc] add GPU cloud playground (#5851) 5 months ago
YeAnbang 84eab13078 update sft trainning script 6 months ago
YeAnbang 2abdede1d7 fix readme 6 months ago
YeAnbang 77db21610a replace the customized dataloader setup with the build-in one 6 months ago
YeAnbang 0d7ff10ea5 replace the customized dataloader setup with the build-in one 6 months ago
YeAnbang 790e1362a6 merge 6 months ago
YeAnbang ac1520cb8f remove baichuan from template test due to transformer version conflict 6 months ago
YeAnbang e16ccc272a update ci 6 months ago
YeAnbang 45195ac53d remove local data path 6 months ago
YeAnbang bf57b13dda remove models that require huggingface auth from ci 6 months ago
YeAnbang 0bbac158ed fix datasets version 6 months ago
YeAnbang 62eb28b929 remove duplicated test 6 months ago
YeAnbang b8b5cacf38 fix transformers version 6 months ago
pre-commit-ci[bot] 1b880ce095 [pre-commit.ci] auto fixes from pre-commit.com hooks 6 months ago
YeAnbang 7ae87b3159 fix training script 6 months ago
YeAnbang 0b4a33548c moupdate ci tests, st ci test cases passed, tp failed in generation for ppo, sp is buggy 6 months ago
YeAnbang 7e65b71815 run pre-commit 6 months ago
YeAnbang 929e1e3da4 upgrade ppo dpo rm script 6 months ago
YeAnbang 7a7e86987d upgrade colossal-chat support tp_group>1, add sp for sft 6 months ago
Tong Li 913c920ecc
[Colossal-LLaMA] Fix sft issue for llama2 (#5719) 6 months ago
Hongxin Liu 7f8b16635b
[misc] refactor launch API and tensor constructor (#5666) 7 months ago
linsj20 91fa553775 [Feature] qlora support (#5586) 7 months ago
Tong Li 862fbaaa62
[Feature] Support LLaMA-3 CPT and ST (#5619) 7 months ago
Camille Zhong 89049b0d89
[doc] fix ColossalMoE readme (#5599) 7 months ago
Hongxin Liu 641b1ee71a
[devops] remove post commit ci (#5566) 8 months ago
digger yu a799ca343b
[fix] fix typo s/muiti-node /multi-node etc. (#5448) 8 months ago
Wenhao Chen e614aa34f3
[shardformer, pipeline] add `gradient_checkpointing_ratio` and heterogenous shard policy for llama (#5508) 8 months ago
YeAnbang df5e9c53cf
[ColossalChat] Update RLHF V2 (#5286) 8 months ago
Insu Jang 00525f7772
[shardformer] fix pipeline forward error if custom layer distribution is used (#5189) 8 months ago
Wenhao Chen bb0a668fee
[hotfix] set return_outputs=False in examples and polish code (#5404) 8 months ago
binmakeswell d158fc0e64
[doc] update open-sora demo (#5479) 8 months ago
digger yu 385e85afd4
[hotfix] fix typo s/keywrods/keywords etc. (#5429) 9 months ago
Camille Zhong da885ed540
fix tensor data update for gemini loss caluculation (#5442) 9 months ago
Camille Zhong 743e7fad2f
[colossal-llama2] add stream chat examlple for chat version model (#5428) 9 months ago
hugo-syn c8003d463b
[doc] Fix typo s/infered/inferred/ (#5288) 9 months ago
Dongruixuan Li a7ae2b5b4c
[eval-hotfix] set few_shot_data to None when few shot is disabled (#5422) 9 months ago
binmakeswell 822241a99c
[doc] sora release (#5425) 9 months ago
Camille Zhong 4b8312c08e
fix sft single turn inference example (#5416) 9 months ago
Tong Li a28c971516
update requirements (#5407) 9 months ago
CZYCW b833153fd5
[hotfix] fix variable type for top_p (#5313) 9 months ago
Hongxin Liu 7303801854
[llama] fix training and inference scripts (#5384) 9 months ago
Hongxin Liu 65e5d6baa5 [moe] fix mixtral optim checkpoint (#5344) 10 months ago
Hongxin Liu 956b561b54 [moe] fix mixtral forward default value (#5329) 10 months ago
Hongxin Liu b60be18dcc [moe] fix mixtral checkpoint io (#5314) 10 months ago
Hongxin Liu da39d21b71 [moe] support mixtral (#5309) 10 months ago
Hongxin Liu c904d2ae99 [moe] update capacity computing (#5253) 10 months ago
Xuanlei Zhao 7d8e0338a4 [moe] init mixtral impl 10 months ago
Hongxin Liu 084c91246c
[llama] fix memory issue (#5371) 10 months ago