Making large AI models cheaper, faster and more accessible
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
Hongxin Liu dc2cdaf3e8
[shardformer] optimize seq parallelism (#6086)
1 month ago
..
kit [Feature] Split cross-entropy computation in SP (#5959) 2 months ago
test_analyzer [misc] update pre-commit and run all files (#4752) 1 year ago
test_auto_parallel [test] Fix/fix testcase (#5770) 6 months ago
test_autochunk [misc] update pre-commit and run all files (#4752) 1 year ago
test_booster [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
test_checkpoint_io [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
test_cluster [misc] refactor launch API and tensor constructor (#5666) 7 months ago
test_config [misc] update pre-commit and run all files (#4752) 1 year ago
test_device [misc] refactor launch API and tensor constructor (#5666) 7 months ago
test_fp8 [fp8] Disable all_gather intranode. Disable Redundant all_gather fp8 (#6059) 2 months ago
test_fx [hotfix] fix testcase in test_fx/test_tracer (#5779) 6 months ago
test_infer [release] update version (#6041) 2 months ago
test_lazy [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
test_legacy [FP8] rebase main (#5963) 4 months ago
test_lora [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
test_moe [hotfix] moe hybrid parallelism benchmark & follow-up fix (#6048) 2 months ago
test_optimizer [MoE/ZeRO] Moe refactor with zero refactor (#5821) 5 months ago
test_pipeline [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
test_shardformer [shardformer] optimize seq parallelism (#6086) 1 month ago
test_smoothquant [inference] Add smmoothquant for llama (#4904) 1 year ago
test_tensor [misc] refactor launch API and tensor constructor (#5666) 7 months ago
test_zero [FP8] rebase main (#5963) 4 months ago
__init__.py [zero] Update sharded model v2 using sharded param v2 (#323) 3 years ago