Commit Graph

272 Commits (7f9ec599be461cef555f4da2f796b46a3631d18f)

Author SHA1 Message Date
YeAnbang 84eab13078 update sft trainning script
6 months ago
YeAnbang 2abdede1d7 fix readme
6 months ago
YeAnbang 77db21610a replace the customized dataloader setup with the build-in one
6 months ago
YeAnbang 0d7ff10ea5 replace the customized dataloader setup with the build-in one
6 months ago
YeAnbang 790e1362a6 merge
6 months ago
YeAnbang ac1520cb8f remove baichuan from template test due to transformer version conflict
6 months ago
YeAnbang e16ccc272a update ci
6 months ago
YeAnbang 45195ac53d remove local data path
6 months ago
YeAnbang bf57b13dda remove models that require huggingface auth from ci
6 months ago
YeAnbang 0bbac158ed fix datasets version
6 months ago
YeAnbang 62eb28b929 remove duplicated test
6 months ago
YeAnbang b8b5cacf38 fix transformers version
6 months ago
pre-commit-ci[bot] 1b880ce095 [pre-commit.ci] auto fixes from pre-commit.com hooks
6 months ago
YeAnbang 7ae87b3159 fix training script
6 months ago
YeAnbang 0b4a33548c moupdate ci tests, st ci test cases passed, tp failed in generation for ppo, sp is buggy
6 months ago
YeAnbang 7e65b71815 run pre-commit
6 months ago
YeAnbang 929e1e3da4 upgrade ppo dpo rm script
6 months ago
YeAnbang 7a7e86987d upgrade colossal-chat support tp_group>1, add sp for sft
6 months ago
Tong Li 913c920ecc
[Colossal-LLaMA] Fix sft issue for llama2 (#5719)
6 months ago
Hongxin Liu 7f8b16635b
[misc] refactor launch API and tensor constructor (#5666)
7 months ago
linsj20 91fa553775 [Feature] qlora support (#5586)
7 months ago
Tong Li 862fbaaa62
[Feature] Support LLaMA-3 CPT and ST (#5619)
7 months ago
Camille Zhong 89049b0d89
[doc] fix ColossalMoE readme (#5599)
7 months ago
Hongxin Liu 641b1ee71a
[devops] remove post commit ci (#5566)
8 months ago
digger yu a799ca343b
[fix] fix typo s/muiti-node /multi-node etc. (#5448)
8 months ago
Wenhao Chen e614aa34f3
[shardformer, pipeline] add `gradient_checkpointing_ratio` and heterogenous shard policy for llama (#5508)
8 months ago
YeAnbang df5e9c53cf
[ColossalChat] Update RLHF V2 (#5286)
8 months ago
Insu Jang 00525f7772
[shardformer] fix pipeline forward error if custom layer distribution is used (#5189)
8 months ago
Wenhao Chen bb0a668fee
[hotfix] set return_outputs=False in examples and polish code (#5404)
8 months ago
binmakeswell d158fc0e64
[doc] update open-sora demo (#5479)
8 months ago
digger yu 385e85afd4
[hotfix] fix typo s/keywrods/keywords etc. (#5429)
9 months ago
Camille Zhong da885ed540
fix tensor data update for gemini loss caluculation (#5442)
9 months ago
Camille Zhong 743e7fad2f
[colossal-llama2] add stream chat examlple for chat version model (#5428)
9 months ago
hugo-syn c8003d463b
[doc] Fix typo s/infered/inferred/ (#5288)
9 months ago
Dongruixuan Li a7ae2b5b4c
[eval-hotfix] set few_shot_data to None when few shot is disabled (#5422)
9 months ago
binmakeswell 822241a99c
[doc] sora release (#5425)
9 months ago
Camille Zhong 4b8312c08e
fix sft single turn inference example (#5416)
9 months ago
Tong Li a28c971516
update requirements (#5407)
9 months ago
CZYCW b833153fd5
[hotfix] fix variable type for top_p (#5313)
9 months ago
Hongxin Liu 7303801854
[llama] fix training and inference scripts (#5384)
9 months ago
Frank Lee efef43b53c
Merge pull request #5372 from hpcaitech/exp/mixtral
10 months ago
Hongxin Liu 65e5d6baa5 [moe] fix mixtral optim checkpoint (#5344)
10 months ago
Hongxin Liu 956b561b54 [moe] fix mixtral forward default value (#5329)
10 months ago
Hongxin Liu b60be18dcc [moe] fix mixtral checkpoint io (#5314)
10 months ago
Hongxin Liu da39d21b71 [moe] support mixtral (#5309)
10 months ago
Hongxin Liu c904d2ae99 [moe] update capacity computing (#5253)
10 months ago
Xuanlei Zhao 7d8e0338a4 [moe] init mixtral impl
10 months ago
Hongxin Liu 084c91246c
[llama] fix memory issue (#5371)
10 months ago
Hongxin Liu eb4f2d90f9
[llama] polish training script and fix optim ckpt (#5368)
10 months ago
Camille Zhong a5756a8720
[eval] update llama npu eval (#5366)
10 months ago