Commit Graph

2781 Commits (3e05c07bb8921f2a8f9736b6f6673d4e9f1697d0)
 

Author SHA1 Message Date
Hongxin Liu 3e05c07bb8
[lazy] support torch 2.0 (#4763)
1 year ago
Wenhao Chen 901ab1eedd
[chat]: add lora merge weights config (#4766)
1 year ago
Baizhou Zhang 493a5efeab
[doc] add shardformer doc to sidebar (#4768)
1 year ago
Hongxin Liu 66f3926019
[doc] clean up outdated docs (#4765)
1 year ago
Baizhou Zhang df66741f77
[bug] fix get_default_parser in examples (#4764)
1 year ago
Baizhou Zhang c0a033700c
[shardformer] fix master param sync for hybrid plugin/rewrite unwrapping logic (#4758)
1 year ago
Wenhao Chen 7b9b86441f
[chat]: update rm, add wandb and fix bugs (#4471)
1 year ago
ppt0011 07c2e3d09c
Merge pull request #4757 from ppt0011/main
1 year ago
Pengtai Xu 4d7537ba25 [doc] put native colossalai plugins first in description section
1 year ago
Pengtai Xu e10d9f087e [doc] add model examples for each plugin
1 year ago
Pengtai Xu a04337bfc3 [doc] put individual plugin explanation in front
1 year ago
Pengtai Xu 10513f203c [doc] explain suitable use case for each plugin
1 year ago
Hongxin Liu 079bf3cb26
[misc] update pre-commit and run all files (#4752)
1 year ago
github-actions[bot] 3c6b831c26
[format] applied code formatting on changed files in pull request 4743 (#4750)
1 year ago
Hongxin Liu b5f9e37c70
[legacy] clean up legacy code (#4743)
1 year ago
Xuanlei Zhao 32e7f99416
[kernel] update triton init #4740 (#4740)
1 year ago
Baizhou Zhang d151dcab74
[doc] explaination of loading large pretrained models (#4741)
1 year ago
flybird11111 4c4482f3ad
[example] llama2 add fine-tune example (#4673)
1 year ago
Xuanlei Zhao ac2797996b
[shardformer] add custom policy in hybrid parallel plugin (#4718)
1 year ago
Baizhou Zhang 451c3465fb
[doc] polish shardformer doc (#4735)
1 year ago
ppt0011 73eb3e8862
Merge pull request #4738 from ppt0011/main
1 year ago
Bin Jia 608cffaed3
[example] add gpt2 HybridParallelPlugin example (#4653)
1 year ago
Bin Jia 6a03c933a0
[shardformer] update seq parallel document (#4730)
1 year ago
Pengtai Xu cd4e61d149 [legacy] remove deterministic data loader test
1 year ago
flybird11111 46162632e5
[shardformer] update pipeline parallel document (#4725)
1 year ago
digger yu e4fc57c3de
Optimized some syntax errors in the documentation and code under applications/ (#4127)
1 year ago
Baizhou Zhang 50e5602c2d
[doc] add shardformer support matrix/update tensor parallel documents (#4728)
1 year ago
github-actions[bot] 8c2dda7410
[format] applied code formatting on changed files in pull request 4726 (#4727)
1 year ago
Baizhou Zhang f911d5b09d
[doc] Add user document for Shardformer (#4702)
1 year ago
binmakeswell ce97790ed7
[doc] fix llama2 code link (#4726)
1 year ago
flybird11111 20190b49a5
[shardformer] to fix whisper test failed due to significant accuracy differences. (#4710)
1 year ago
Yuanheng Zhao e2c0e7f92a
[hotfix] Fix import error: colossal.kernel without triton installed (#4722)
1 year ago
flybird11111 c7d6975d29
[shardformer] fix GPT2DoubleHeadsModel (#4703)
1 year ago
Baizhou Zhang 068372a738
[doc] add potential solution for OOM in llama2 example (#4699)
1 year ago
digger yu 9c2feb2f0b
fix some typo with colossalai/device colossalai/tensor/ etc. (#4171)
1 year ago
Baizhou Zhang d8ceeac14e
[hotfix] fix typo in hybrid parallel io (#4697)
1 year ago
flybird11111 8844691f4b
[shardformer] update shardformer readme (#4689)
1 year ago
Baizhou Zhang 1d454733c4
[doc] Update booster user documents. (#4669)
1 year ago
Cuiqing Li bce0f16702
[Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577)
1 year ago
flybird11111 eedaa3e1ef
[shardformer]fix gpt2 double head (#4663)
1 year ago
Hongxin Liu 554aa9592e
[legacy] move communication and nn to legacy and refactor logger (#4671)
1 year ago
Hongxin Liu 536397cc95
[devops] fix concurrency group (#4667)
1 year ago
flybird11111 7486ed7d3a
[shardformer] update llama2/opt finetune example and fix llama2 policy (#4645)
1 year ago
Hongxin Liu a686f9ddc8
[devops] fix concurrency group and compatibility test (#4665)
1 year ago
Baizhou Zhang 295b38fecf
[example] update vit example for hybrid parallel plugin (#4641)
1 year ago
Baizhou Zhang 660eed9124
[pipeline] set optimizer to optional in execute_pipeline (#4630)
1 year ago
eric8607242 c3d5fa3bac
[shardformer] Support customized policy for llamav2 based model with HybridParallelPlugin (#4624)
1 year ago
Hongxin Liu 9709b8f502
[release] update version (#4623)
1 year ago
Hongxin Liu efba0f44b9
Merge pull request #4612 from hpcaitech/feature/shardformer
1 year ago
Hongxin Liu fae6c92ead
Merge branch 'main' into feature/shardformer
1 year ago