2742 Commits (eedaa3e1ef991d9f9a274d10c046877ba2b10467)
 

Author SHA1 Message Date
Baizhou Zhang 44eab2b27f
[shardformer] support sharded checkpoint IO for models of HybridParallelPlugin (#4506) 1 year ago
flybird11111 de8a65babc
[shardformer] opt fix. (#4514) 1 year ago
LuGY 839847b7d7
[zero]support zero2 with gradient accumulation (#4511) 1 year ago
github-actions[bot] c0efc3ebcb
[format] applied code formatting on changed files in pull request 4479 (#4504) 1 year ago
flybird11111 3353e55c80
[shardformer] vit/llama/t5 ignore the sequence parallelism flag and some fix. (#4498) 1 year ago
Hongxin Liu 27061426f7
[gemini] improve compatibility and add static placement policy (#4479) 1 year ago
Jianghai e04436a82a
[shardformer] tests for 3d parallel (#4493) 1 year ago
flybird11111 59e252ecdb
[shardformer] chatglm support sequence parallel (#4482) 1 year ago
Bin Jia 351351a36e
[shardformer/sequence parallel] not support opt of seq-parallel, add warning and fix a bug in gpt2 pp (#4488) 1 year ago
Jianghai 5545114fd8
rename chatglm to chatglm2 (#4484) 1 year ago
Michelle 285fe7ba71
[chat] update config and prompt (#4139) 1 year ago
Baizhou Zhang 1c7df566e2
[shardformer] support tp+zero for shardformer (#4472) 1 year ago
Jianghai 8739aa7fa0
[shardformer] Pipeline/whisper (#4456) 1 year ago
flybird11111 a27e0bb494
[shardformer] bert support sequence parallel. (#4455) 1 year ago
flybird11111 0ecd71e041
[shardformer] bloom support sequence parallel (#4465) 1 year ago
Bin Jia 7c8be77081
[shardformer/sequence parallel] support gpt2 seq parallel with pp/dp/tp (#4460) 1 year ago
LuGY a78daf6180
[shardformer] support interleaved pipeline (#4448) 1 year ago
Hongxin Liu 26e29d58f0
[devops] add large-scale distributed test marker (#4452) 1 year ago
Baizhou Zhang 6ef33f75aa
[shardformer] support DDP in HybridPlugin/add tp+dp tests (#4446) 1 year ago
Bin Jia 424629fea0
[shardformer/sequence parallel] Cherry pick commit to new branch (#4450) 1 year ago
github-actions[bot] d20dceb9a3
[format] applied code formatting on changed files in pull request 4441 (#4445) 1 year ago
ver217 5d4efdf58f [shardformer] fix import 1 year ago
ver217 73a4144b91 [shardformer] fix embedding 1 year ago
ver217 922302263b [misc] update requirements 1 year ago
Hongxin Liu 172f7fa3cf [misc] resolve code factor issues (#4433) 1 year ago
flybird11111 328a791d10 [shardformer] update bloom/llama/vit/chatglm tests (#4420) 1 year ago
flybird11111 108e54a0b4 [shardformer]update t5 tests for using all optimizations. (#4407) 1 year ago
flybird11111 1edc9b5fb3 [shardformer] update tests for all optimization (#4413) 1 year ago
Baizhou Zhang 7711bd524a [shardformer] rewrite tests for opt/bloom/llama/vit/chatglm (#4395) 1 year ago
flybird11111 21e0a42fd1 [shardformer]fix, test gpt2 for AMP+TP (#4403) 1 year ago
Jianghai 7596e9ae08 [pipeline] rewrite bert tests and fix some bugs (#4409) 1 year ago
flybird1111 d2cd48e0be [shardformer] test all optimizations (#4399) 1 year ago
flybird1111 7a3dfd0c64 [shardformer] update shardformer to use flash attention 2 (#4392) 1 year ago
Baizhou Zhang ed4c448488 [pipeline] rewrite t5 tests & support multi-tensor transmitting in pipeline (#4388) 1 year ago
flybird1111 906426cb44 [Shardformer] Merge flash attention branch to pipeline branch (#4362) 1 year ago
Jianghai a88e92251d [pipeline] add chatglm (#4363) 1 year ago
Baizhou Zhang b1feeced8e [shardformer] add util functions for shardformer tests/fix sync_shared_param (#4366) 1 year ago
Bin Jia 5c6f183192 [test] Hotfix/fix some model test and refactor check util api (#4369) 1 year ago
FoolPlayer c3ca53cf05 [test] skip some not compatible models 1 year ago
FoolPlayer 726541afe2 update some module with new api version 1 year ago
FoolPlayer 879301d0da [shardformer] support Blip2 (#4243) 1 year ago
klhhhhh 8120eca0c0 [shardformer] support ChatGLMForConditionalGeneration & add fusedlayernorm for vit 1 year ago
klhhhhh 4da05052f4 [shardformer] pre-commit check files 1 year ago
klhhhhh 91850fe984 [shardformer] register without auto policy 1 year ago
klhhhhh f155ae89c4 [shardformer] ChatGLM support layernorm sharding 1 year ago
klhhhhh 00f6ef159d [shardformer] delete some file 1 year ago
klhhhhh dad00c42aa [shardformer] support chatglm without layernorm 1 year ago
klhhhhh cbb54d3202 [shardformer] polish code 1 year ago
klhhhhh 1a29e8fc29 [shardformer] polish chatglm code 1 year ago
klhhhhh 8620009dd7 [sharformer] add first version of policy of chatglm 1 year ago