You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/engine/gradient_handler
Jiarui Fang a445e118cf
[polish] polish singleton and global context (#500)
3 years ago
..
__init__.py
_base_gradient_handler.py
_data_parallel_gradient_handler.py add moe context, moe utilities and refactor gradient handler (#455) 3 years ago
_moe_gradient_handler.py [polish] polish singleton and global context (#500) 3 years ago
_pipeline_parallel_gradient_handler.py [zero] ZeRO supports pipeline parallel (#477) 3 years ago
_sequence_parallel_gradient_handler.py add moe context, moe utilities and refactor gradient handler (#455) 3 years ago
_zero_gradient_handler.py Flake8 code restyle 3 years ago
utils.py add moe context, moe utilities and refactor gradient handler (#455) 3 years ago