from colossalai.amp import AMP_TYPE
BATCH_SIZE = 128
NUM_EPOCHS = 200
CONFIG = dict(
fp16=dict(
mode=AMP_TYPE.TORCH
)