ColossalAI/colossalai/zero/gemini
Hongxin Liu 8241c0c054
[fp8] support gemini plugin (#5978)
* [fp8] refactor hook

* [fp8] support gemini plugin

* [example] add fp8 option for llama benchmark
2024-08-09 14:09:48 +08:00
..
chunk [Feature]: support FP8 communication in DDP, FSDP, Gemini (#5928) 2024-08-08 15:55:01 +08:00
memory_tracer [npu] change device to accelerator api (#5239) 2024-01-09 10:20:05 +08:00
__init__.py [shardformer]: support gpt-j, falcon, Mistral and add interleaved pipeline for bert (#5088) 2023-11-28 16:54:42 +08:00
gemini_ddp.py [fp8] support gemini plugin (#5978) 2024-08-09 14:09:48 +08:00
gemini_hook.py [gemini] quick fix on possible async operation (#5803) 2024-06-13 10:35:17 +08:00
gemini_mgr.py [chore] remove unnecessary assert since compute list might not be recorded 2024-05-28 05:16:02 +00:00
gemini_optimizer.py [gemini] async grad chunk reduce (all-reduce&reduce-scatter) (#5713) 2024-05-24 10:31:16 +08:00
placement_policy.py [bug] continue fix 2024-05-28 02:41:23 +00:00
utils.py [npu] change device to accelerator api (#5239) 2024-01-09 10:20:05 +08:00