mirror of https://github.com/hpcaitech/ColossalAI
[example] initialize tutorial (#1865)
parent
cc27adceb0
commit
fd8f0ca5a8
|
@ -0,0 +1,54 @@
|
||||||
|
# Colossal-AI Tutorial Hands-on
|
||||||
|
|
||||||
|
## Introduction
|
||||||
|
|
||||||
|
Welcome to the [Colossal-AI](https://github.com/hpcaitech/ColossalAI) tutorial, which has been accepted as official tutorials by top conference [SC](https://sc22.supercomputing.org/), [AAAI](https://aaai.org/Conferences/AAAI-23/), [PPoPP](https://ppopp23.sigplan.org/), etc.
|
||||||
|
|
||||||
|
|
||||||
|
[Colossal-AI](https://github.com/hpcaitech/ColossalAI), a unified deep learning system for the big model era, integrates
|
||||||
|
many advanced technologies such as multi-dimensional tensor parallelism, sequence parallelism, heterogeneous memory management,
|
||||||
|
large-scale optimization, adaptive task scheduling, etc. By using Colossal-AI, we could help users to efficiently and
|
||||||
|
quickly deploy large AI model training and inference, reducing large AI model training budgets and scaling down the labor cost of learning and deployment.
|
||||||
|
|
||||||
|
### 🚀 Quick Links
|
||||||
|
|
||||||
|
[**Colossal-AI**](https://github.com/hpcaitech/ColossalAI) |
|
||||||
|
[**Paper**](https://arxiv.org/abs/2110.14883) |
|
||||||
|
[**Documentation**](https://www.colossalai.org/) |
|
||||||
|
[**Forum**](https://github.com/hpcaitech/ColossalAI/discussions) |
|
||||||
|
[**Slack**](https://join.slack.com/t/colossalaiworkspace/shared_invite/zt-z7b26eeb-CBp7jouvu~r0~lcFzX832w)
|
||||||
|
|
||||||
|
|
||||||
|
## Table of Content
|
||||||
|
|
||||||
|
- Multi-dimensional Parallelism
|
||||||
|
- Know the components and sketch of Colossal-AI
|
||||||
|
- Step-by-step from PyTorch to Colossal-AI
|
||||||
|
- Try data/pipeline parallelism and 1D/2D/2.5D/3D tensor parallelism using a unified model
|
||||||
|
- Sequence Parallelism
|
||||||
|
- Try sequence parallelism with BERT
|
||||||
|
- Combination of data/pipeline/sequence parallelism
|
||||||
|
- Faster training and longer sequence length
|
||||||
|
- Auto-Parallelism
|
||||||
|
- Parallelism with normal non-distributed training code
|
||||||
|
- Model tracing + solution solving + runtime communication inserting all in one auto-parallelism system
|
||||||
|
- Try single program, multiple data (SPMD) parallel with auto-parallelism SPMD solver on ResNet50
|
||||||
|
- Large Batch Training Optimization
|
||||||
|
- Comparison of small/large batch size with SGD/LARS optimizer
|
||||||
|
- Acceleration from a larger batch size
|
||||||
|
- Fine-tuning and Serving for OPT from Hugging Face
|
||||||
|
- Try OPT model imported from Hugging Face with Colossal-AI
|
||||||
|
- Fine-tuning OPT with limited hardware using ZeRO, Gemini and parallelism
|
||||||
|
- Deploy the fine-tuned model to inference service
|
||||||
|
- Acceleration of Stable Diffusion
|
||||||
|
- Stable Diffusion with Lightning
|
||||||
|
- Try Lightning Colossal-AI strategy to optimize memory and accelerate speed
|
||||||
|
|
||||||
|
|
||||||
|
## Discussion
|
||||||
|
|
||||||
|
Discussion about the [Colossal-AI](https://github.com/hpcaitech/ColossalAI) project is always welcomed! We would love to exchange ideas with the community to better help this project grow.
|
||||||
|
If you think there is a need to discuss anything, you may jump to our [Slack](https://join.slack.com/t/colossalaiworkspace/shared_invite/zt-z7b26eeb-CBp7jouvu~r0~lcFzX832w).
|
||||||
|
|
||||||
|
If you encounter any problem while running these tutorials, you may want to raise an [issue](https://github.com/hpcaitech/ColossalAI/issues/new/choose) in this repository.
|
||||||
|
|
Loading…
Reference in New Issue