ColossalAI/colossalai/booster/plugin
flybird11111 21aa5de00b
[gemini] hotfix NaN loss while using Gemini + tensor_parallel (#5150)
* fix

aaa

fix

fix

fix

* fix

* fix

* test ci

* fix ci

fix
2023-12-08 11:10:51 +08:00
..
__init__.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
dp_plugin_base.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
gemini_plugin.py [gemini] hotfix NaN loss while using Gemini + tensor_parallel (#5150) 2023-12-08 11:10:51 +08:00
hybrid_parallel_plugin.py [plugin]fix 3d checkpoint load when booster boost without optimizer. (#5135) 2023-11-30 18:37:47 +08:00
low_level_zero_plugin.py [npu] add npu support for gemini and zero (#5067) 2023-11-20 16:12:41 +08:00
moe_hybrid_parallel_plugin.py [moe] support optimizer checkpoint (#5015) 2023-11-08 15:07:03 +00:00
plugin_base.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
pp_plugin_base.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
torch_ddp_plugin.py [doc] polish shardformer doc (#4779) 2023-09-26 10:57:47 +08:00
torch_fsdp_plugin.py [doc] polish shardformer doc (#4779) 2023-09-26 10:57:47 +08:00