mirror of https://github.com/hpcaitech/ColossalAI
b2ad0d9e8f
* Use p2p * Cannot bidirectonal send p2p * Refactor tensor creation and serialization in P2P communication * Fix llama forward args in flash attention * Add flop estimate from megatron * Support loading weight not in weight_map when strict=False in hybrid_parallel * Use send_forward_recv_backward, etc in 1f1b * Use dataclass for metdata Remove torch.cuda.synchronize() as suggested * Add comment about the torch.cuda.synchronize for potential error * Typo * Update hybrid_parallel_checkpoint_io.py * Update p2p.py * Update one_f_one_b.py * Update p2p.py --------- Co-authored-by: flybird11111 <1829166702@qq.com> |
||
---|---|---|
.. | ||
__init__.py | ||
checkpoint_io_base.py | ||
general_checkpoint_io.py | ||
hybrid_parallel_checkpoint_io.py | ||
index_file.py | ||
utils.py |