You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/tests/test_shardformer/test_model
Zhongkai Zhao 5d9a0ae75b
[hotfix] Fix ShardFormer test execution path when using sequence parallelism (#5230)
10 months ago
..
__init__.py
_utils.py [hotfix] Fix ShardFormer test execution path when using sequence parallelism (#5230) 10 months ago
test_shard_bert.py [pipeline]: fix p2p comm, add metadata cache and support llama interleaved pp (#5134) 11 months ago
test_shard_blip2.py
test_shard_bloom.py [hotfix] Add layer norm gradients all-reduce for sequence parallel (#4926) 1 year ago
test_shard_chatglm2.py [hotfix] Add layer norm gradients all-reduce for sequence parallel (#4926) 1 year ago
test_shard_falcon.py [shardformer]: support gpt-j, falcon, Mistral and add interleaved pipeline for bert (#5088) 1 year ago
test_shard_gpt2.py [ci] fix shardformer tests. (#5255) 11 months ago
test_shard_gptj.py [shardformer] llama support DistCrossEntropy (#5176) 12 months ago
test_shard_llama.py [pipeline]: fix p2p comm, add metadata cache and support llama interleaved pp (#5134) 11 months ago
test_shard_mistral.py [shardformer]: support gpt-j, falcon, Mistral and add interleaved pipeline for bert (#5088) 1 year ago
test_shard_opt.py
test_shard_sam.py
test_shard_t5.py [ci] fix shardformer tests. (#5255) 11 months ago
test_shard_vit.py
test_shard_whisper.py [ci] fix shardformer tests. (#5255) 11 months ago