ColossalAI/colossalai/booster
Hongxin Liu 8241c0c054
[fp8] support gemini plugin (#5978)
* [fp8] refactor hook

* [fp8] support gemini plugin

* [example] add fp8 option for llama benchmark
2024-08-09 14:09:48 +08:00
..
mixed_precision [npu] change device to accelerator api (#5239) 2024-01-09 10:20:05 +08:00
plugin [fp8] support gemini plugin (#5978) 2024-08-09 14:09:48 +08:00
__init__.py [booster] implemented the torch ddd + resnet example (#3232) 2023-03-27 10:24:14 +08:00
accelerator.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
booster.py [Feature] qlora support (#5586) 2024-04-28 10:51:27 +08:00