You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai
ver217 9ec1ce6ab1
[zero] sharded model support the reuse of fp16 shard (#495)
3 years ago
..
amp [hotfix] fix initialize bug with zero (#442) 3 years ago
builder
communication
context [MOE] remove old MoE legacy (#493) 3 years ago
engine [zero] polish sharded param name (#484) 3 years ago
kernel
logging [log] better logging display with rich (#426) 3 years ago
nn [zero] sharded model support the reuse of fp16 shard (#495) 3 years ago
registry
testing [test] added rerun on exception for testing (#475) 3 years ago
trainer
utils [MOE] remove old MoE legacy (#493) 3 years ago
zero [zero] sharded model support the reuse of fp16 shard (#495) 3 years ago
__init__.py
constants.py
core.py [MOE] polish moe_env (#467) 3 years ago
global_variables.py [MOE] add unitest for MOE experts layout, gradient handler and kernel (#469) 3 years ago
initialize.py [MOE] add unitest for MOE experts layout, gradient handler and kernel (#469) 3 years ago