You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/shardformer/examples/convergence_benchmark.sh

10 lines
272 B

torchrun --standalone --nproc_per_node=4 convergence_benchmark.py \
--model "bert" \
--pretrain "bert-base-uncased" \
--max_epochs 3 \
--batch_size 2 \
--lr 2.4e-5 \
--fused_layernorm False \
--accumulation_steps 8 \
--warmup_fraction 0.03