ColossalAI/colossalai/zero/gemini
botbw 2fc85abf43
[gemini] async grad chunk reduce (all-reduce&reduce-scatter) (#5713)
* [gemini] async grad chunk reduce (all-reduce&reduce-scatter)

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* [gemini] add test

* [gemini] rename func

* [gemini] update llama benchmark

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* [gemini] use tensor counter

* [gemini] change default config in GeminiPlugin and GeminiDDP

* [chore] typo

* [gemini] fix sync issue & add test cases

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2024-05-24 10:31:16 +08:00
..
chunk [gemini] async grad chunk reduce (all-reduce&reduce-scatter) (#5713) 2024-05-24 10:31:16 +08:00
memory_tracer [npu] change device to accelerator api (#5239) 2024-01-09 10:20:05 +08:00
__init__.py [shardformer]: support gpt-j, falcon, Mistral and add interleaved pipeline for bert (#5088) 2023-11-28 16:54:42 +08:00
gemini_ddp.py [gemini] async grad chunk reduce (all-reduce&reduce-scatter) (#5713) 2024-05-24 10:31:16 +08:00
gemini_hook.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
gemini_mgr.py [npu] add npu support for gemini and zero (#5067) 2023-11-20 16:12:41 +08:00
gemini_optimizer.py [gemini] async grad chunk reduce (all-reduce&reduce-scatter) (#5713) 2024-05-24 10:31:16 +08:00
placement_policy.py [npu] change device to accelerator api (#5239) 2024-01-09 10:20:05 +08:00
utils.py [npu] change device to accelerator api (#5239) 2024-01-09 10:20:05 +08:00