ColossalAI/examples/language/gpt/README.md

492 B

Overview

This example shows how to use Colossal-AI to run huggingface GPT training in distributed manners.

GPT

We use the GPT2 model from huggingface transformers. The input data is randonly generated. The train_gpt_demo.py provides three distributed plans, i.e. ColossalAI, PyTorch DDP and ZeRO. The ColossalAI leverages Tensor Parallel and Gemini.

Quick Start

You can launch training by using the following bash script.

pip install -r requirements.txt
bash run.sh