ColossalAI/colossalai/moe
Hongxin Liu c904d2ae99 [moe] update capacity computing (#5253)
* [moe] top2 allow uneven input

* [moe] update capacity computing

* [moe] remove debug info

* [moe] update capacity computing

* [moe] update capacity computing
2024-02-07 19:21:02 +08:00
..
__init__.py [moe] init mixtral impl 2024-02-07 19:21:02 +08:00
_operation.py [feat] refactored extension module (#5298) 2024-01-25 17:01:48 +08:00
checkpoint.py [moe] init mixtral impl 2024-02-07 19:21:02 +08:00
experts.py [moe] init mixtral impl 2024-02-07 19:21:02 +08:00
layers.py [moe] init mixtral impl 2024-02-07 19:21:02 +08:00
load_balance.py [moe] merge moe into main (#4978) 2023-11-02 02:21:24 +00:00
loss.py [moe] merge moe into main (#4978) 2023-11-02 02:21:24 +00:00
manager.py fix some typo (#5307) 2024-01-25 13:56:27 +08:00
routers.py [moe] update capacity computing (#5253) 2024-02-07 19:21:02 +08:00
utils.py [moe] init mixtral impl 2024-02-07 19:21:02 +08:00