ColossalAI/examples/language/gpt/gpt2_configs/gpt2_vanilla.py

27 lines
393 B
Python

from titans.model.gpt import gpt2_small
from torch.optim import Adam
from colossalai.amp import AMP_TYPE
BATCH_SIZE = 1
NUM_EPOCHS = 60
SEQ_LEN = 1024
optimizer = dict(
type=Adam,
lr=0.00015,
weight_decay=1e-2,
)
fp16 = dict(mode=AMP_TYPE.NAIVE)
model = dict(
type=gpt2_small,
checkpoint=True,
)
parallel = dict(
pipeline=1,
tensor=dict(size=1, mode=None),
)