You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai
botbw 4fa6b9509c
[moe] add parallel strategy for shared_expert && fix test for deepseek (#6063)
2 months ago
..
_C
_analyzer
accelerator
amp
auto_parallel
autochunk
booster [doc] FP8 training and communication document (#6050) 2 months ago
checkpoint_io
cli
cluster
context
device
fx
inference
interface
kernel [pre-commit.ci] auto fixes from pre-commit.com hooks 2 months ago
lazy
legacy
logging
moe
nn
pipeline [fp8] hotfix backward hook (#6053) 2 months ago
quantization [fp8] Disable all_gather intranode. Disable Redundant all_gather fp8 (#6059) 2 months ago
shardformer [moe] add parallel strategy for shared_expert && fix test for deepseek (#6063) 2 months ago
tensor
testing
utils
zero [fp8] Disable all_gather intranode. Disable Redundant all_gather fp8 (#6059) 2 months ago
__init__.py
initialize.py [fp8] hotfix backward hook (#6053) 2 months ago