# Colossal-AI
- 14x larger batch size, and 5x faster training for Tensor Parallelism = 64 ### GPT-3
- Save 50% GPU resources, and 10.7% acceleration ### GPT-2 - 11x lower GPU memory consumption, and superlinear scaling efficiency with Tensor Parallelism - 24x larger model size on the same hardware - over 3x acceleration ### BERT - 2x faster training, or 50% longer sequence length ### PaLM - [PaLM-colossalai](https://github.com/hpcaitech/PaLM-colossalai): Scalable implementation of Google's Pathways Language Model ([PaLM](https://ai.googleblog.com/2022/04/pathways-language-model-palm-scaling-to.html)). Please visit our [documentation and tutorials](https://www.colossalai.org/) for more details. ## Single GPU Demo ### GPT-2
- 20x larger model size on the same hardware ### PaLM
- 34x larger model size on the same hardware ## Installation ### Download From Official Releases You can visit the [Download](https://www.colossalai.org/download) page to download Colossal-AI with pre-built CUDA extensions. ### Download From Source > The version of Colossal-AI will be in line with the main branch of the repository. Feel free to raise an issue if you encounter any problem. :) ```shell git clone https://github.com/hpcaitech/ColossalAI.git cd ColossalAI # install dependency pip install -r requirements/requirements.txt # install colossalai pip install . ``` If you don't want to install and enable CUDA kernel fusion (compulsory installation when using fused optimizer): ```shell NO_CUDA_EXT=1 pip install . ``` ## Use Docker Run the following command to build a docker image from Dockerfile provided. ```bash cd ColossalAI docker build -t colossalai ./docker ``` Run the following command to start the docker container in interactive mode. ```bash docker run -ti --gpus all --rm --ipc=host colossalai bash ``` ## Community Join the Colossal-AI community on [Forum](https://github.com/hpcaitech/ColossalAI/discussions), [Slack](https://join.slack.com/t/colossalaiworkspace/shared_invite/zt-z7b26eeb-CBp7jouvu~r0~lcFzX832w), and [WeChat](https://raw.githubusercontent.com/hpcaitech/public_assets/main/colossalai/img/WeChat.png "qrcode") to share your suggestions, feedback, and questions with our engineering team. ## Contributing If you wish to contribute to this project, please follow the guideline in [Contributing](./CONTRIBUTING.md). Thanks so much to all of our amazing contributors! *The order of contributor avatars is randomly shuffled.* ## Quick View ### Start Distributed Training in Lines ```python parallel = dict( pipeline=2, tensor=dict(mode='2.5d', depth = 1, size=4) ) ``` ### Start Heterogeneous Training in Lines ```python zero = dict( model_config=dict( tensor_placement_policy='auto', shard_strategy=TensorShardStrategy(), reuse_fp16_shard=True ), optimizer_config=dict(initial_scale=2**5, gpu_margin_mem_ratio=0.2) ) ``` ## Cite Us ``` @article{bian2021colossal, title={Colossal-AI: A Unified Deep Learning System For Large-Scale Parallel Training}, author={Bian, Zhengda and Liu, Hongxin and Wang, Boxiang and Huang, Haichen and Li, Yongbin and Wang, Chuanrui and Cui, Fan and You, Yang}, journal={arXiv preprint arXiv:2110.14883}, year={2021} } ```