.. |
__init__.py
|
[shardformer] adapted T5 and LLaMa test to use kit (#4049)
|
2023-07-04 16:05:01 +08:00 |
_utils.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_pure_pipeline.py
|
[pipeline] support fp32 for HybridPlugin/merge shardformer test and pipeline test into one file (#4354)
|
2023-08-15 23:25:14 +08:00 |
test_shard_bert.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_shard_bert_pipeline.py
|
[pipeline] refactor test pipeline and remove useless utils in pipeline (#4324)
|
2023-08-15 23:25:14 +08:00 |
test_shard_blip2.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_shard_bloom.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_shard_bloom_pipeline.py
|
[pipeline] refactor test pipeline and remove useless utils in pipeline (#4324)
|
2023-08-15 23:25:14 +08:00 |
test_shard_chatglm.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_shard_chatglm_pipeline.py
|
[pipeline] add chatglm (#4363)
|
2023-08-15 23:25:14 +08:00 |
test_shard_gpt2.py
|
[pipeline] rewrite t5 tests & support multi-tensor transmitting in pipeline (#4388)
|
2023-08-15 23:25:14 +08:00 |
test_shard_llama.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_shard_llama_pipeline.py
|
[pipeline] refactor test pipeline and remove useless utils in pipeline (#4324)
|
2023-08-15 23:25:14 +08:00 |
test_shard_opt.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_shard_opt_pipeline.py
|
[pipeline] OPT model pipeline (#4258)
|
2023-08-15 23:25:14 +08:00 |
test_shard_sam.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_shard_t5.py
|
[pipeline] rewrite t5 tests & support multi-tensor transmitting in pipeline (#4388)
|
2023-08-15 23:25:14 +08:00 |
test_shard_vit.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_shard_vit_pipeline.py
|
[shardformer] support pipeline base vit model (#4284)
|
2023-08-15 23:25:14 +08:00 |
test_shard_whisper.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |