You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/applications/ColossalMoE/colossal_moe/models
Wenhao Chen e614aa34f3
[shardformer, pipeline] add `gradient_checkpointing_ratio` and heterogenous shard policy for llama (#5508)
8 months ago
..
__init__.py [moe] init mixtral impl 10 months ago
mixtral_checkpoint.py [moe] fix mixtral optim checkpoint (#5344) 10 months ago
mixtral_layer.py [moe] support mixtral (#5309) 10 months ago
mixtral_policy.py [shardformer, pipeline] add `gradient_checkpointing_ratio` and heterogenous shard policy for llama (#5508) 8 months ago