ColossalAI/colossalai/shardformer/layer
Baizhou Zhang 44eab2b27f
[shardformer] support sharded checkpoint IO for models of HybridParallelPlugin (#4506)
* add APIs

* implement save_sharded_model

* add test for hybrid checkpointio

* implement naive loading for sharded model

* implement efficient sharded model loading

* open a new file for hybrid checkpoint_io

* small fix

* fix circular importing

* fix docstring

* arrange arguments and apis

* small fix
2023-08-25 22:04:57 +08:00
..
__init__.py [shardformer] fix import 2023-08-15 23:25:14 +08:00
_operation.py [shardformer] bert support sequence parallel. (#4455) 2023-08-18 18:04:55 +08:00
dropout.py
embedding.py [shardformer] fix embedding 2023-08-15 23:25:14 +08:00
linear.py [shardformer] chatglm support sequence parallel (#4482) 2023-08-22 23:59:31 +08:00
loss.py
normalization.py
parallel_module.py [shardformer] support sharded checkpoint IO for models of HybridParallelPlugin (#4506) 2023-08-25 22:04:57 +08:00
qkv_fused_linear.py [shardformer/sequence parallel] Cherry pick commit to new branch (#4450) 2023-08-16 15:41:20 +08:00
utils.py