ColossalAI/colossalai/shardformer/layer
Baizhou Zhang ed4c448488 [pipeline] rewrite t5 tests & support multi-tensor transmitting in pipeline (#4388)
* fix remaining t5 bugs/rewrite t5 tests

* fix multi-tensor communication in pipeline

* rearrange test_config

* fix keyerror in sync_shared_params

* fix get_held_layers & Randomnizer, complete t5 tests

* erase printing

* fix get_held_layers through modifying _release_unheld_layers

* fix _get_recursive_held_layers bug
2023-08-15 23:25:14 +08:00
..
__init__.py [shardformer] support SAM (#4231) 2023-08-15 23:25:14 +08:00
_operation.py [format] applied code formatting on changed files in pull request 4152 (#4157) 2023-07-04 16:07:47 +08:00
dropout.py [shardformer] supported bloom model (#4098) 2023-07-04 16:05:01 +08:00
embedding.py [shardformer] support whisper (#4212) 2023-08-15 23:25:14 +08:00
linear.py [shardformer] support inplace sharding (#4251) 2023-08-15 23:25:14 +08:00
loss.py fix some typo colossalai/shardformer (#4160) 2023-07-04 17:53:39 +08:00
normalization.py [shardformer] support inplace sharding (#4251) 2023-08-15 23:25:14 +08:00
parallel_module.py [shardformer] supported fused qkv checkpoint (#4073) 2023-07-04 16:05:01 +08:00
qkv_fused_linear.py update some module with new api version 2023-08-15 23:25:14 +08:00
utils.py [pipeline] rewrite t5 tests & support multi-tensor transmitting in pipeline (#4388) 2023-08-15 23:25:14 +08:00