mirror of https://github.com/hpcaitech/ColossalAI
f911d5b09d
* create shardformer doc files * add docstring for seq-parallel * update ShardConfig docstring * add links to llama example * add outdated massage * finish introduction & supporting information * finish 'how shardformer works' * finish shardformer.md English doc * fix doctest fail * add Chinese document |
||
---|---|---|
.. | ||
1D_tensor_parallel.md | ||
2D_tensor_parallel.md | ||
2p5D_tensor_parallel.md | ||
3D_tensor_parallel.md | ||
cluster_utils.md | ||
gradient_accumulation.md | ||
gradient_accumulation_with_booster.md | ||
gradient_clipping.md | ||
gradient_clipping_with_booster.md | ||
gradient_handler.md | ||
mixed_precision_training.md | ||
mixed_precision_training_with_booster.md | ||
nvme_offload.md | ||
pipeline_parallel.md | ||
shardformer.md | ||
zero_with_chunk.md |