ColossalAI/examples/language
botbw c54c4fcd15
[hotfix] moe hybrid parallelism benchmark & follow-up fix (#6048)
* [example] pass use_fp8_comm flag to all plugins

* [example] add mixtral benchmark

* [moe] refine assertion and check

* [moe] fix mixtral & add more tests

* [moe] consider checking dp * sp group and moe_dp_group

* [mixtral] remove gate tp & add more tests

* [deepseek] fix tp & sp for deepseek

* [mixtral] minor fix

* [deepseek] add deepseek benchmark
2024-09-10 17:30:53 +08:00
..
bert [Feature]: support FP8 communication in DDP, FSDP, Gemini (#5928) 2024-08-08 15:55:01 +08:00
commons [example] make gpt example directory more clear (#2353) 2023-01-06 11:11:26 +08:00
deepseek [hotfix] moe hybrid parallelism benchmark & follow-up fix (#6048) 2024-09-10 17:30:53 +08:00
gpt [Feature] Split cross-entropy computation in SP (#5959) 2024-09-10 12:06:50 +08:00
grok-1 [misc] refactor launch API and tensor constructor (#5666) 2024-04-29 10:40:11 +08:00
llama [hotfix] moe hybrid parallelism benchmark & follow-up fix (#6048) 2024-09-10 17:30:53 +08:00
mixtral [hotfix] moe hybrid parallelism benchmark & follow-up fix (#6048) 2024-09-10 17:30:53 +08:00
opt [fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016) 2024-08-22 09:21:34 +08:00
palm [misc] refactor launch API and tensor constructor (#5666) 2024-04-29 10:40:11 +08:00
__init__.py [example]add gpt2 benchmark example script. (#5295) 2024-03-04 16:18:13 +08:00
data_utils.py [devops] remove post commit ci (#5566) 2024-04-08 15:09:40 +08:00
model_utils.py [example]add gpt2 benchmark example script. (#5295) 2024-03-04 16:18:13 +08:00
performance_evaluator.py [Feature] Split cross-entropy computation in SP (#5959) 2024-09-10 12:06:50 +08:00