ColossalAI/tests/test_shardformer/test_model
Bin Jia 7c8be77081
[shardformer/sequence parallel] support gpt2 seq parallel with pp/dp/tp (#4460)
* support gpt2 seq parallel with pp/dp/tp

* fix a bug when waiting for stream done

* delete unused gpt2_seq file
2023-08-18 11:21:53 +08:00
..
__init__.py [shardformer] adapted T5 and LLaMa test to use kit (#4049) 2023-07-04 16:05:01 +08:00
_utils.py [shardformer] support DDP in HybridPlugin/add tp+dp tests (#4446) 2023-08-16 16:11:57 +08:00
test_shard_bert.py [shardformer] support DDP in HybridPlugin/add tp+dp tests (#4446) 2023-08-16 16:11:57 +08:00
test_shard_blip2.py [Shardformer] Merge flash attention branch to pipeline branch (#4362) 2023-08-15 23:25:14 +08:00
test_shard_bloom.py [shardformer] support DDP in HybridPlugin/add tp+dp tests (#4446) 2023-08-16 16:11:57 +08:00
test_shard_chatglm.py [shardformer] support DDP in HybridPlugin/add tp+dp tests (#4446) 2023-08-16 16:11:57 +08:00
test_shard_gpt2.py [shardformer/sequence parallel] support gpt2 seq parallel with pp/dp/tp (#4460) 2023-08-18 11:21:53 +08:00
test_shard_llama.py [shardformer] support DDP in HybridPlugin/add tp+dp tests (#4446) 2023-08-16 16:11:57 +08:00
test_shard_opt.py [shardformer] support DDP in HybridPlugin/add tp+dp tests (#4446) 2023-08-16 16:11:57 +08:00
test_shard_sam.py [Shardformer] Merge flash attention branch to pipeline branch (#4362) 2023-08-15 23:25:14 +08:00
test_shard_t5.py [shardformer] support DDP in HybridPlugin/add tp+dp tests (#4446) 2023-08-16 16:11:57 +08:00
test_shard_vit.py [shardformer] support DDP in HybridPlugin/add tp+dp tests (#4446) 2023-08-16 16:11:57 +08:00
test_shard_whisper.py [Shardformer] Merge flash attention branch to pipeline branch (#4362) 2023-08-15 23:25:14 +08:00