mirror of https://github.com/hpcaitech/ColossalAI
32 lines
500 B
Python
32 lines
500 B
Python
from titans.loss.lm_loss import GPTLMLoss
|
|
from titans.model.gpt import gpt2_small
|
|
from torch.optim import Adam
|
|
|
|
from colossalai.amp import AMP_TYPE
|
|
|
|
BATCH_SIZE = 1
|
|
SEQ_LEN = 1024
|
|
NUM_EPOCHS = 60
|
|
|
|
TENSOR_PARALLEL = 2
|
|
|
|
optimizer = dict(
|
|
type=Adam,
|
|
lr=0.00015,
|
|
weight_decay=1e-2,
|
|
)
|
|
|
|
fp16 = dict(mode=AMP_TYPE.NAIVE)
|
|
|
|
loss = dict(type=GPTLMLoss,)
|
|
|
|
model = dict(
|
|
type=gpt2_small,
|
|
checkpoint=True,
|
|
)
|
|
|
|
parallel = dict(
|
|
pipeline=1,
|
|
tensor=dict(size=TENSOR_PARALLEL, mode='1d'),
|
|
)
|