diff --git a/examples/images/diffusion/README.md b/examples/images/diffusion/README.md index 324337426..8583f3be2 100644 --- a/examples/images/diffusion/README.md +++ b/examples/images/diffusion/README.md @@ -52,7 +52,7 @@ export PACKAGE_NAME=pytorch pip install . ``` -### Install [Colossal-AI v0.1.10](https://colossalai.org/download/) From Our Official Website +### Install [Colossal-AI v0.1.12](https://colossalai.org/download/) From Our Official Website ``` pip install colossalai==0.1.12+torch1.12cu11.3 -f https://release.colossalai.org @@ -101,10 +101,10 @@ python main.py --logdir /tmp/ -t -b configs/train_colossalai.yaml You can change the trainging config in the yaml file -- accelerator: acceleratortype, default 'gpu' -- devices: device number used for training, default 4 -- max_epochs: max training epochs -- precision: usefp16 for training or not, default 16, you must use fp16 if you want to apply colossalai +- devices: device number used for training, default 8 +- max_epochs: max training epochs, default 2 +- precision: the precision type used in training, default 16 (fp16), you must use fp16 if you want to apply colossalai +- more information about the configuration of ColossalAIStrategy can be found [here](https://pytorch-lightning.readthedocs.io/en/latest/advanced/model_parallel.html#colossal-ai) ## Finetune Example ### Training on Teyvat Datasets