.. |
__init__.py
|
[shardformer] adapted T5 and LLaMa test to use kit (#4049)
|
2023-07-04 16:05:01 +08:00 |
_utils.py
|
[shardformer] rewrite tests for opt/bloom/llama/vit/chatglm (#4395)
|
2023-08-15 23:25:14 +08:00 |
test_pure_pipeline.py
|
[pipeline] support fp32 for HybridPlugin/merge shardformer test and pipeline test into one file (#4354)
|
2023-08-15 23:25:14 +08:00 |
test_shard_bert.py
|
[shardformer] update tests for all optimization (#4413)
|
2023-08-15 23:25:14 +08:00 |
test_shard_blip2.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_shard_bloom.py
|
[shardformer] update bloom/llama/vit/chatglm tests (#4420)
|
2023-08-15 23:25:14 +08:00 |
test_shard_chatglm.py
|
[shardformer] update bloom/llama/vit/chatglm tests (#4420)
|
2023-08-15 23:25:14 +08:00 |
test_shard_gpt2.py
|
[shardformer] update bloom/llama/vit/chatglm tests (#4420)
|
2023-08-15 23:25:14 +08:00 |
test_shard_llama.py
|
[shardformer] update bloom/llama/vit/chatglm tests (#4420)
|
2023-08-15 23:25:14 +08:00 |
test_shard_opt.py
|
[shardformer] update bloom/llama/vit/chatglm tests (#4420)
|
2023-08-15 23:25:14 +08:00 |
test_shard_sam.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |
test_shard_t5.py
|
[shardformer]update t5 tests for using all optimizations. (#4407)
|
2023-08-15 23:25:14 +08:00 |
test_shard_vit.py
|
[shardformer] update bloom/llama/vit/chatglm tests (#4420)
|
2023-08-15 23:25:14 +08:00 |
test_shard_whisper.py
|
[Shardformer] Merge flash attention branch to pipeline branch (#4362)
|
2023-08-15 23:25:14 +08:00 |