mirror of https://github.com/hpcaitech/ColossalAI
68 lines
2.5 KiB
Markdown
68 lines
2.5 KiB
Markdown
# Auto-Parallelism
|
|
|
|
## Table of contents
|
|
|
|
- [Auto-Parallelism](#auto-parallelism)
|
|
- [Table of contents](#table-of-contents)
|
|
- [📚 Overview](#-overview)
|
|
- [🚀 Quick Start](#-quick-start)
|
|
- [Setup](#setup)
|
|
- [Auto-Parallel Tutorial](#auto-parallel-tutorial)
|
|
- [Auto-Checkpoint Tutorial](#auto-checkpoint-tutorial)
|
|
|
|
|
|
## 📚 Overview
|
|
|
|
This tutorial folder contains a simple demo to run auto-parallelism with ResNet. Meanwhile, this diretory also contains demo scripts to run automatic activation checkpointing, but both features are still experimental for now and no guarantee that they will work for your version of Colossal-AI.
|
|
|
|
## 🚀 Quick Start
|
|
|
|
### Setup
|
|
|
|
1. Create a conda environment
|
|
|
|
```bash
|
|
conda create -n auto python=3.8
|
|
conda activate auto
|
|
```
|
|
|
|
2. Install `requirements` and `coin-or-cbc` for the solver.
|
|
|
|
```bash
|
|
pip install -r requirements.txt
|
|
conda install -c conda-forge coin-or-cbc
|
|
```
|
|
|
|
|
|
### Auto-Parallel Tutorial
|
|
|
|
Run the auto parallel resnet example with 4 GPUs with synthetic dataset.
|
|
|
|
```bash
|
|
colossalai run --nproc_per_node 4 auto_parallel_with_resnet.py
|
|
```
|
|
|
|
You should expect to the log like this. This log shows the edge cost on the computation graph as well as the sharding strategy for an operation. For example, `layer1_0_conv1 S01R = S01R X RR` means that the first dimension (batch) of the input and output is sharded while the weight is not sharded (S means sharded, R means replicated), simply equivalent to data parallel training.
|
|
![](https://raw.githubusercontent.com/hpcaitech/public_assets/main/examples/tutorial/auto-parallel%20demo.png)
|
|
|
|
|
|
### Auto-Checkpoint Tutorial
|
|
|
|
We prepare two bechmarks for you to test the performance of auto checkpoint
|
|
|
|
The first test `auto_ckpt_solver_test.py` will show you the ability of solver to search checkpoint strategy that could fit in the given budget (test on GPT2 Medium and ResNet 50). It will output the benchmark summary and data visualization of peak memory vs. budget memory and relative step time vs. peak memory.
|
|
|
|
The second test `auto_ckpt_batchsize_test.py` will show you the advantage of fitting larger batchsize training into limited GPU memory with the help of our activation checkpoint solver (test on ResNet152). It will output the benchmark summary.
|
|
|
|
The usage of the above two test
|
|
```bash
|
|
# run auto_ckpt_solver_test.py on gpt2 medium
|
|
python auto_ckpt_solver_test.py --model gpt2
|
|
|
|
# run auto_ckpt_solver_test.py on resnet50
|
|
python auto_ckpt_solver_test.py --model resnet50
|
|
|
|
# tun auto_ckpt_batchsize_test.py
|
|
python auto_ckpt_batchsize_test.py
|
|
```
|