You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/shardformer/layer
wangbluo 3dc08c8a5a
fix
1 month ago
..
__init__.py [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
_operation.py [fp8] Disable all_gather intranode. Disable Redundant all_gather fp8 (#6059) 3 months ago
attn.py fix 1 month ago
dropout.py [misc] update pre-commit and run all files (#4752) 1 year ago
embedding.py [fp8] support hybrid parallel plugin (#5982) 4 months ago
linear.py [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
loss.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
normalization.py [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
parallel_module.py [shardformer] refactor embedding resize (#5603) 7 months ago
qkv_fused_linear.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
utils.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago