ColossalAI/colossalai/nn/optimizer
ver217 9ec1ce6ab1
[zero] sharded model support the reuse of fp16 shard (#495)
* sharded model supports reuse fp16 shard

* rename variable

* polish code

* polish code

* polish code
2022-03-23 14:59:59 +08:00
..
__init__.py [zero] cpu adam kernel (#288) 2022-03-11 15:50:28 +08:00
colossalai_optimizer.py Develop/experiments (#59) 2021-12-09 15:08:29 +08:00
cpu_adam.py [zero] sharded model support the reuse of fp16 shard (#495) 2022-03-23 14:59:59 +08:00
fused_adam.py [MOE] add unitest for MOE experts layout, gradient handler and kernel (#469) 2022-03-21 13:35:04 +08:00
fused_lamb.py [MOE] add unitest for MOE experts layout, gradient handler and kernel (#469) 2022-03-21 13:35:04 +08:00
fused_sgd.py refactor kernel (#142) 2022-01-13 16:47:17 +08:00
lamb.py Fixed docstring in colossalai (#171) 2022-01-21 10:44:30 +08:00
lars.py Fixed docstring in colossalai (#171) 2022-01-21 10:44:30 +08:00