ColossalAI/colossalai/amp/naive_amp
Baizhou Zhang 0ceec8f9a9 [pipeline] support fp32 for HybridPlugin/merge shardformer test and pipeline test into one file (#4354)
* add naive optimizer for 3DPlugin/refactor gpt2 shardformer test

* merge tests of PP/DP/TP combinations into one test file

* fix bug when sync grad for dp in HybridPlugin

* update supported precisions for 3DPlugin/fix bug when shifting tp_degree

* improve the passing of lazy_init

* modify lazy_init/use sync_shared_params
2023-08-15 23:25:14 +08:00
..
grad_scaler [zero] fix gradient clipping in hybrid parallelism (#2521) 2023-01-29 15:09:57 +08:00
mixed_precision_mixin [bf16] add bf16 support (#3882) 2023-06-05 15:58:31 +08:00
__init__.py [NFC] polish colossalai/amp/naive_amp/__init__.py code style (#1905) 2022-11-11 17:49:18 +08:00
_fp16_optimizer.py [setup] support pre-build and jit-build of cuda kernels (#2374) 2023-01-06 20:50:26 +08:00
_utils.py [NFC] polish colossalai/amp/naive_amp/_utils.py code style (#1816) 2022-11-09 12:08:47 +08:00
mixed_precision_optimizer.py [pipeline] support fp32 for HybridPlugin/merge shardformer test and pipeline test into one file (#4354) 2023-08-15 23:25:14 +08:00
naive_amp.py [amp] add gradient clipping for unit tests (#2283) 2023-01-04 11:59:56 +08:00