ColossalAI/colossalai/zero/gemini
Baizhou Zhang 44eab2b27f
[shardformer] support sharded checkpoint IO for models of HybridParallelPlugin (#4506)
* add APIs

* implement save_sharded_model

* add test for hybrid checkpointio

* implement naive loading for sharded model

* implement efficient sharded model loading

* open a new file for hybrid checkpoint_io

* small fix

* fix circular importing

* fix docstring

* arrange arguments and apis

* small fix
2023-08-25 22:04:57 +08:00
..
chunk [gemini] fix argument naming during chunk configuration searching 2023-06-25 13:34:15 +08:00
memory_tracer [nfc] fix typo colossalai/zero (#3923) 2023-06-08 00:01:29 +08:00
__init__.py [zero] reorganize zero/gemini folder structure (#3424) 2023-04-04 13:48:16 +08:00
colo_init_context.py [doc] Fix typo under colossalai and doc(#3618) 2023-04-26 11:38:43 +08:00
gemini_ddp.py [shardformer] support sharded checkpoint IO for models of HybridParallelPlugin (#4506) 2023-08-25 22:04:57 +08:00
gemini_hook.py [zero] reorganize zero/gemini folder structure (#3424) 2023-04-04 13:48:16 +08:00
gemini_mgr.py [zero] reorganize zero/gemini folder structure (#3424) 2023-04-04 13:48:16 +08:00
gemini_optimizer.py [gemini] fix tensor storage cleaning in state dict collection (#4396) 2023-08-10 15:36:46 +08:00
placement_policy.py [zero] reorganize zero/gemini folder structure (#3424) 2023-04-04 13:48:16 +08:00
utils.py fix typo with colossalai/trainer utils zero (#3908) 2023-06-07 16:08:37 +08:00