mirror of https://github.com/hpcaitech/ColossalAI
aibig-modeldata-parallelismdeep-learningdistributed-computingfoundation-modelsheterogeneous-traininghpcinferencelarge-scalemodel-parallelismpipeline-parallelism
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
flybird11111
26cd6d850c
|
1 year ago | |
---|---|---|
.. | ||
README.md | ||
requirements.txt | ||
test_ci.sh | ||
train.py | 1 year ago |
README.md
Train ViT on CIFAR-10 from scratch
🚀 Quick Start
This example provides a training script, which provides an example of training ViT on CIFAR10 dataset from scratch.
- Training Arguments
-p
,--plugin
: Plugin to use. Choices:torch_ddp
,torch_ddp_fp16
,low_level_zero
. Defaults totorch_ddp
.-r
,--resume
: Resume from checkpoint file path. Defaults to-1
, which means not resuming.-c
,--checkpoint
: The folder to save checkpoints. Defaults to./checkpoint
.-i
,--interval
: Epoch interval to save checkpoints. Defaults to5
. If set to0
, no checkpoint will be saved.--target_acc
: Target accuracy. Raise exception if not reached. Defaults toNone
.
Install requirements
pip install -r requirements.txt
Train
# train with torch DDP with fp32
colossalai run --nproc_per_node 4 train.py -c ./ckpt-fp32
# train with torch DDP with mixed precision training
colossalai run --nproc_per_node 4 train.py -c ./ckpt-fp16 -p torch_ddp_fp16
# train with low level zero
colossalai run --nproc_per_node 4 train.py -c ./ckpt-low_level_zero -p low_level_zero
Expected accuracy performance will be:
Model | Single-GPU Baseline FP32 | Booster DDP with FP32 | Booster DDP with FP16 | Booster Low Level Zero |
---|---|---|---|---|
ViT | 83.00% | 84.03% | 84.00% | 84.43% |