mirror of https://github.com/InternLM/InternLM
* support zero for expert local dp
* fix above codes:
*treat optim.zero_world_size and optim.zero_local_rank as list in model_checkpoint.py and test_model_checkpoint.py
*add overlap and zero check for moe in args_sanity_check(.)
|
||
|---|---|---|
| .. | ||
| apis | ||
| core | ||
| data | ||
| initialize | ||
| model | ||
| moe | ||
| monitor | ||
| solver | ||
| train | ||
| utils | ||
| __init__.py | ||