You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/shardformer/layer
wangbluo 0b14a5512e
fix
3 months ago
..
__init__.py [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
_operation.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
attn.py fix 3 months ago
dropout.py [misc] update pre-commit and run all files (#4752) 1 year ago
embedding.py [fp8] support hybrid parallel plugin (#5982) 4 months ago
linear.py [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
loss.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
normalization.py [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 3 months ago
parallel_module.py [shardformer] refactor embedding resize (#5603) 8 months ago
qkv_fused_linear.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
utils.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago