ColossalAI/examples/language/gpt/README.md

16 lines
492 B
Markdown

## Overview
This example shows how to use Colossal-AI to run huggingface GPT training in distributed manners.
## GPT
We use the GPT2 model from huggingface transformers. The input data is randonly generated.
The `train_gpt_demo.py` provides three distributed plans, i.e. ColossalAI, PyTorch DDP and ZeRO.
The ColossalAI leverages Tensor Parallel and Gemini.
## Quick Start
You can launch training by using the following bash script.
```bash
pip install -r requirements.txt
bash run.sh
```