## OpenMoE [OpenMoE](https://github.com/XueFuzhao/OpenMoE) is the open-source community's first decoder-only MoE transformer. OpenMoE is implemented in Jax, and [Colossal-AI](https://github.com/hpcaitech/ColossalAI) has pioneered an efficient open-source support for this model in PyTorch, enabling a broader range of users to participate in and use this model. The following example of [Colossal-AI](https://github.com/hpcaitech/ColossalAI) demonstrates finetune and inference methods.
* [2023/11] [Enhanced MoE Parallelism, Open-source MoE Model Training Can Be 9 Times More Efficient](https://www.hpc-ai.tech/blog/enhanced-moe-parallelism-open-source-moe-model-training-can-be-9-times-more-efficient) [[code]](https://github.com/hpcaitech/ColossalAI/tree/main/examples/language/openmoe) [[blog]](https://www.hpc-ai.tech/blog/enhanced-moe-parallelism-open-source-moe-model-training-can-be-9-times-more-efficient) ## Usage ### 1. Installation Please install the latest ColossalAI from source. ```bash BUILD_EXT=1 pip install -U git+https://github.com/hpcaitech/ColossalAI ``` Then install dependencies. ```bash cd ColossalAI/examples/language/openmoe pip install -r requirements.txt ``` Additionally, we recommend you to use torch 1.13.1. We've tested our code on torch 1.13.1 and found it's compatible with our code and flash attention. ### 2. Install kernels (Optional) We have utilized `Triton`, `FlashAttention` and `Apex` kernel for better performance. They are not necessary but we recommend you to install them to fully utilize your hardware. ``` # install triton via pip pip install triton # install flash attention via pip pip install flash-attn==2.0.5 # install apex from source git clone https://github.com/NVIDIA/apex.git cd apex git checkout 741bdf50825a97664db08574981962d66436d16a pip install -v --disable-pip-version-check --no-cache-dir --no-build-isolation --config-settings "--build-option=--cpp_ext" --config-settings "--build-option=--cuda_ext" ./ --global-option="--cuda_ext" ``` ### 3. Train Yon can use colossalai run to launch single-node training: ```bash colossalai run --standalone --nproc_per_node YOUR_GPU_PER_NODE train.py --OTHER_CONFIGURATIONS ``` Yon can also use colossalai run to launch multi-nodes training: ```bash colossalai run --nproc_per_node YOUR_GPU_PER_NODE --hostfile YOUR_HOST_FILE train.py --OTHER_CONFIGURATIONS ``` Here is a sample hostfile: ```text hostname1 hostname2 hostname3 hostname4 ``` The hostname refers to the ip address of your nodes. Make sure master node can access all nodes (including itself) by ssh without password. Here is details about CLI arguments: - Model configuration: `--model_name`. `base` and `8b` are supported for OpenMoE. - Booster plugin: `--plugin`. `ep`, `ep_zero` and `hybrid` are supported. `ep_zero` is recommended for general cases. `ep` can provides least memory consumption and `hybrid` suits large scale training. - Output path: `--output_path`. The path to save your model. The default value is `./outputs`. - Number of epochs: `--num_epochs`. The default value is 1. - Local batch size: `--batch_size`. Batch size per GPU. The default value is 1. - Save interval: `-i`, `--save_interval`. The interval (steps) of saving checkpoints. The default value is 1000. - Mixed precision: `--precision`. The default value is "bf16". "fp16", "bf16" and "fp32" are supported. - Max length: `--max_length`. Max sequence length. Default to 2048. - Dataset: `-d`, `--dataset`. The default dataset is `yizhongw/self_instruct`. It support any dataset from `datasets` with the same data format as it. - Task Name: `--task_name`. Task of corresponding dataset. Default to `super_natural_instructions`. - Learning rate: `--lr`. The default value is 1e-5. - Weight decay: `--weight_decay`. The default value is 0. - Zero stage: `--zero_stage`. Zero stage. Recommend 2 for ep and 1 for ep zero. - Extra dp size: `--extra_dp_size`. Extra moe param dp size for ep_zero plugin. Recommended to be 2 or 4. - Use kernel: `--use_kernel`. Use kernel optim. Need to install flash attention and triton to enable all kernel optimizations. Skip if not installed. - Use layernorm kernel: `--use_layernorm_kernel`. Use layernorm kernel. Need to install apex. Raise error if not installed. - Router aux loss factor: `--router_aux_loss_factor`. Moe router z loss factor. You can refer to STMoE for details. - Router z loss factor: `--router_z_loss_factor`. Moe router aux loss factor. You can refer to STMoE for details. - Label smoothing: `--label_smoothing`. Label smoothing. - Z loss factor: `--z_loss_factor`. The final outputs' classification z loss factor. Load balance: `--load_balance`. Expert load balance. Defaults to False. Recommend enabling. - Load balance interval: `--load_balance_interval`. Expert load balance interval. - Communication overlap: `--comm_overlap`. Use communication overlap for MoE. Recommended to enable for multi-node training. ### 4. Shell Script Examples For your convenience, we provide some shell scripts to train with various configurations. Here we will show an example of how to run training OpenMoE. #### a. Running environment This experiment was performed on a single computing nodes with 8 A800 80GB GPUs in total for OpenMoE-8B. The GPUs are fully connected with NVLink. #### b. Running command We demonstrate how to run three plugins in `train.sh`. You can choose anyone and use your own args. ```bash bash train.sh ``` #### c. Multi-Nodes Training To run on multi-nodes, you can modify the script as: ```bash colossalai run --nproc_per_node YOUR_GPU_PER_NODE --hostfile YOUR_HOST_FILE \ train.py --OTHER_CONFIGURATIONS ``` ## Reference ``` @article{bian2021colossal, title={Colossal-AI: A Unified Deep Learning System For Large-Scale Parallel Training}, author={Bian, Zhengda and Liu, Hongxin and Wang, Boxiang and Huang, Haichen and Li, Yongbin and Wang, Chuanrui and Cui, Fan and You, Yang}, journal={arXiv preprint arXiv:2110.14883}, year={2021} } ``` ```bibtex @misc{openmoe2023, author = {Fuzhao Xue, Zian Zheng, Yao Fu, Jinjie Ni, Zangwei Zheng, Wangchunshu Zhou and Yang You}, title = {OpenMoE: Open Mixture-of-Experts Language Models}, year = {2023}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {\url{https://github.com/XueFuzhao/OpenMoE}}, } ```