ColossalAI/colossalai/checkpoint_io
Hongxin Liu 554aa9592e
[legacy] move communication and nn to legacy and refactor logger (#4671)
* [legacy] move communication to legacy (#4640)

* [legacy] refactor logger and clean up legacy codes (#4654)

* [legacy] make logger independent to gpc

* [legacy] make optim independent to registry

* [legacy] move test engine to legacy

* [legacy] move nn to legacy (#4656)

* [legacy] move nn to legacy

* [checkpointio] fix save hf config

* [test] remove useledd rpc pp test

* [legacy] fix nn init

* [example] skip tutorial hybriad parallel example

* [devops] test doc check

* [devops] test doc check
2023-09-11 16:24:28 +08:00
..
__init__.py [shardformer] support sharded checkpoint IO for models of HybridParallelPlugin (#4506) 2023-08-25 22:04:57 +08:00
checkpoint_io_base.py Next commit [checkpointio] Unsharded Optimizer Checkpoint for Gemini Plugin (#4141) 2023-07-07 16:33:06 +08:00
general_checkpoint_io.py Merge branch 'main' into feature/shardformer 2023-09-04 23:43:13 +08:00
hybrid_parallel_checkpoint_io.py [shardformer] support from_pretrained when loading model with HybridParallelPlugin (#4575) 2023-09-01 17:40:01 +08:00
index_file.py [checkpointio] General Checkpointing of Sharded Optimizers (#3984) 2023-06-15 15:21:26 +08:00
utils.py [legacy] move communication and nn to legacy and refactor logger (#4671) 2023-09-11 16:24:28 +08:00