ColossalAI/colossalai
YuliangLiu0306 21d6a48f4d
[autoparallel] add shard option (#2696)
* [autoparallel] add shard option

* polish
2023-02-15 13:48:28 +08:00
..
_C [setup] support pre-build and jit-build of cuda kernels (#2374) 2023-01-06 20:50:26 +08:00
amp [test] fixed the triton version for testing (#2608) 2023-02-07 13:49:38 +08:00
auto_parallel [autoparallel] add shard option (#2696) 2023-02-15 13:48:28 +08:00
autochunk [autochunk] support diffusion for autochunk (#2621) 2023-02-07 16:32:45 +08:00
builder
cli [cli] fixed hostname mismatch error (#2465) 2023-01-12 14:52:09 +08:00
communication [NFC] polish communication/p2p_v2.py code style (#2303) 2023-01-04 15:09:57 +08:00
context Revert "Update parallel_context.py (#2408)" 2023-01-19 12:27:48 +08:00
device [autoparallel] accelerate gpt2 training (#2495) 2023-01-29 11:13:15 +08:00
engine
fx [autoparallel] Patch meta information of `torch.matmul` (#2584) 2023-02-08 11:05:31 +08:00
gemini [gemini] add fake_release_chunk for keep-gathered chunk in the inference mode (#2671) 2023-02-13 14:35:32 +08:00
kernel [kernel] fixed repeated loading of kernels (#2549) 2023-02-03 09:47:13 +08:00
logging [logger] hotfix, missing _FORMAT (#2231) 2022-12-29 22:59:39 +08:00
nn [gemini] add fake_release_chunk for keep-gathered chunk in the inference mode (#2671) 2023-02-13 14:35:32 +08:00
pipeline polish pp middleware (#2476) 2023-01-13 16:56:01 +08:00
registry
tensor [polish] polish ColoTensor and its submodules (#2537) 2023-02-03 11:44:10 +08:00
testing [amp] add gradient clipping for unit tests (#2283) 2023-01-04 11:59:56 +08:00
trainer
utils [gemini] fix colo_init_context (#2683) 2023-02-13 17:53:15 +08:00
zero [zero1&2] only append parameters with gradients (#2681) 2023-02-13 18:00:16 +08:00
__init__.py
constants.py
core.py
global_variables.py
initialize.py Fix False warning in initialize.py (#2456) 2023-01-12 13:49:01 +08:00