You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/shardformer/layer
Hongxin Liu 641b1ee71a
[devops] remove post commit ci (#5566)
8 months ago
..
__init__.py [devops] remove post commit ci (#5566) 8 months ago
_operation.py [shardformer] Sequence Parallelism Optimization (#5533) 8 months ago
attn.py [shardformer] update colo attention to support custom mask (#5510) 8 months ago
dropout.py [misc] update pre-commit and run all files (#4752) 1 year ago
embedding.py [gemini] gemini support tensor parallelism. (#4942) 1 year ago
linear.py [shardformer] Sequence Parallelism Optimization (#5533) 8 months ago
loss.py [shardformer] llama support DistCrossEntropy (#5176) 12 months ago
normalization.py [hotfix] fix typo change enabel to enable under colossalai/shardformer/ (#5317) 9 months ago
parallel_module.py [misc] update pre-commit and run all files (#4752) 1 year ago
qkv_fused_linear.py [shardformer] Sequence Parallelism Optimization (#5533) 8 months ago
utils.py [shardformer] Sequence Parallelism Optimization (#5533) 8 months ago