2932 Commits (c7d68b2c2ca3f7fd32056ea952fae4fe239f75ea)
 

Author SHA1 Message Date
littsk 1e0e080837
[bug] Fix the version check bug in colossalai run when generating the cmd. (#4713) 1 year ago
Hongxin Liu 3e05c07bb8
[lazy] support torch 2.0 (#4763) 1 year ago
Wenhao Chen 901ab1eedd
[chat]: add lora merge weights config (#4766) 1 year ago
Baizhou Zhang 493a5efeab
[doc] add shardformer doc to sidebar (#4768) 1 year ago
Hongxin Liu 66f3926019
[doc] clean up outdated docs (#4765) 1 year ago
Baizhou Zhang df66741f77
[bug] fix get_default_parser in examples (#4764) 1 year ago
Baizhou Zhang c0a033700c
[shardformer] fix master param sync for hybrid plugin/rewrite unwrapping logic (#4758) 1 year ago
Wenhao Chen 7b9b86441f
[chat]: update rm, add wandb and fix bugs (#4471) 1 year ago
ppt0011 07c2e3d09c
Merge pull request #4757 from ppt0011/main 1 year ago
Pengtai Xu 4d7537ba25 [doc] put native colossalai plugins first in description section 1 year ago
Pengtai Xu e10d9f087e [doc] add model examples for each plugin 1 year ago
Pengtai Xu a04337bfc3 [doc] put individual plugin explanation in front 1 year ago
Pengtai Xu 10513f203c [doc] explain suitable use case for each plugin 1 year ago
Hongxin Liu 079bf3cb26
[misc] update pre-commit and run all files (#4752) 1 year ago
github-actions[bot] 3c6b831c26
[format] applied code formatting on changed files in pull request 4743 (#4750) 1 year ago
Hongxin Liu b5f9e37c70
[legacy] clean up legacy code (#4743) 1 year ago
Xuanlei Zhao 32e7f99416
[kernel] update triton init #4740 (#4740) 1 year ago
Baizhou Zhang d151dcab74
[doc] explaination of loading large pretrained models (#4741) 1 year ago
flybird11111 4c4482f3ad
[example] llama2 add fine-tune example (#4673) 1 year ago
Xuanlei Zhao ac2797996b
[shardformer] add custom policy in hybrid parallel plugin (#4718) 1 year ago
Baizhou Zhang 451c3465fb
[doc] polish shardformer doc (#4735) 1 year ago
ppt0011 73eb3e8862
Merge pull request #4738 from ppt0011/main 1 year ago
Bin Jia 608cffaed3
[example] add gpt2 HybridParallelPlugin example (#4653) 1 year ago
Bin Jia 6a03c933a0
[shardformer] update seq parallel document (#4730) 1 year ago
Pengtai Xu cd4e61d149 [legacy] remove deterministic data loader test 1 year ago
flybird11111 46162632e5
[shardformer] update pipeline parallel document (#4725) 1 year ago
digger yu e4fc57c3de
Optimized some syntax errors in the documentation and code under applications/ (#4127) 1 year ago
Baizhou Zhang 50e5602c2d
[doc] add shardformer support matrix/update tensor parallel documents (#4728) 1 year ago
github-actions[bot] 8c2dda7410
[format] applied code formatting on changed files in pull request 4726 (#4727) 1 year ago
Baizhou Zhang f911d5b09d
[doc] Add user document for Shardformer (#4702) 1 year ago
binmakeswell ce97790ed7
[doc] fix llama2 code link (#4726) 1 year ago
flybird11111 20190b49a5
[shardformer] to fix whisper test failed due to significant accuracy differences. (#4710) 1 year ago
Yuanheng Zhao e2c0e7f92a
[hotfix] Fix import error: colossal.kernel without triton installed (#4722) 1 year ago
flybird11111 c7d6975d29
[shardformer] fix GPT2DoubleHeadsModel (#4703) 1 year ago
Baizhou Zhang 068372a738
[doc] add potential solution for OOM in llama2 example (#4699) 1 year ago
digger yu 9c2feb2f0b
fix some typo with colossalai/device colossalai/tensor/ etc. (#4171) 1 year ago
Baizhou Zhang d8ceeac14e
[hotfix] fix typo in hybrid parallel io (#4697) 1 year ago
flybird11111 8844691f4b
[shardformer] update shardformer readme (#4689) 1 year ago
Baizhou Zhang 1d454733c4
[doc] Update booster user documents. (#4669) 1 year ago
Cuiqing Li bce0f16702
[Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) 1 year ago
flybird11111 eedaa3e1ef
[shardformer]fix gpt2 double head (#4663) 1 year ago
Hongxin Liu 554aa9592e
[legacy] move communication and nn to legacy and refactor logger (#4671) 1 year ago
Hongxin Liu 536397cc95
[devops] fix concurrency group (#4667) 1 year ago
flybird11111 7486ed7d3a
[shardformer] update llama2/opt finetune example and fix llama2 policy (#4645) 1 year ago
Hongxin Liu a686f9ddc8
[devops] fix concurrency group and compatibility test (#4665) 1 year ago
Baizhou Zhang 295b38fecf
[example] update vit example for hybrid parallel plugin (#4641) 1 year ago
Baizhou Zhang 660eed9124
[pipeline] set optimizer to optional in execute_pipeline (#4630) 1 year ago
eric8607242 c3d5fa3bac
[shardformer] Support customized policy for llamav2 based model with HybridParallelPlugin (#4624) 1 year ago
Hongxin Liu 9709b8f502
[release] update version (#4623) 1 year ago
Hongxin Liu efba0f44b9
Merge pull request #4612 from hpcaitech/feature/shardformer 1 year ago