.. |
_C
|
[setup] support pre-build and jit-build of cuda kernels (#2374)
|
2023-01-06 20:50:26 +08:00 |
amp
|
[test] fixed the triton version for testing (#2608)
|
2023-02-07 13:49:38 +08:00 |
auto_parallel
|
[autoparallel] add shard option (#2696)
|
2023-02-15 13:48:28 +08:00 |
autochunk
|
[autochunk] support diffusion for autochunk (#2621)
|
2023-02-07 16:32:45 +08:00 |
builder
|
…
|
|
cli
|
[cli] fixed hostname mismatch error (#2465)
|
2023-01-12 14:52:09 +08:00 |
communication
|
[NFC] polish communication/p2p_v2.py code style (#2303)
|
2023-01-04 15:09:57 +08:00 |
context
|
Revert "Update parallel_context.py (#2408)"
|
2023-01-19 12:27:48 +08:00 |
device
|
[autoparallel] accelerate gpt2 training (#2495)
|
2023-01-29 11:13:15 +08:00 |
engine
|
…
|
|
fx
|
[autoparallel] Patch meta information of `torch.matmul` (#2584)
|
2023-02-08 11:05:31 +08:00 |
gemini
|
[gemini] add fake_release_chunk for keep-gathered chunk in the inference mode (#2671)
|
2023-02-13 14:35:32 +08:00 |
kernel
|
[kernel] fixed repeated loading of kernels (#2549)
|
2023-02-03 09:47:13 +08:00 |
logging
|
[logger] hotfix, missing _FORMAT (#2231)
|
2022-12-29 22:59:39 +08:00 |
nn
|
[gemini] add fake_release_chunk for keep-gathered chunk in the inference mode (#2671)
|
2023-02-13 14:35:32 +08:00 |
pipeline
|
polish pp middleware (#2476)
|
2023-01-13 16:56:01 +08:00 |
registry
|
…
|
|
tensor
|
[polish] polish ColoTensor and its submodules (#2537)
|
2023-02-03 11:44:10 +08:00 |
testing
|
[amp] add gradient clipping for unit tests (#2283)
|
2023-01-04 11:59:56 +08:00 |
trainer
|
…
|
|
utils
|
[gemini] fix colo_init_context (#2683)
|
2023-02-13 17:53:15 +08:00 |
zero
|
[zero1&2] only append parameters with gradients (#2681)
|
2023-02-13 18:00:16 +08:00 |
__init__.py
|
…
|
|
constants.py
|
…
|
|
core.py
|
…
|
|
global_variables.py
|
…
|
|
initialize.py
|
Fix False warning in initialize.py (#2456)
|
2023-01-12 13:49:01 +08:00 |