.. |
_C
|
[optimizer] add div_scale for optimizers (#2117)
|
2022-12-12 17:58:57 +08:00 |
amp
|
[kernel] move all symlinks of kernel to `colossalai._C` (#1971)
|
2022-11-17 13:42:33 +08:00 |
auto_parallel
|
[autoparallel] memory estimation for shape consistency (#2144)
|
2022-12-21 10:39:37 +08:00 |
builder
|
[NFC] polish colossalai/builder/__init__.py code style (#1560)
|
2022-09-08 22:11:04 +08:00 |
cli
|
[cli] updated installation cheheck with more inforamtion (#2050)
|
2022-11-30 17:53:55 +08:00 |
communication
|
…
|
|
context
|
updated tp layers
|
2022-11-02 12:19:38 +08:00 |
device
|
[device] update flatten device mesh usage (#2079)
|
2022-12-05 16:16:07 +08:00 |
engine
|
[engin/schedule] use p2p_v2 to recontruct pipeline_schedule (#1408)
|
2022-08-12 11:33:26 +08:00 |
fx
|
register meta func for rnn (#2159)
|
2022-12-21 23:06:18 +08:00 |
gemini
|
[hotfix] fix auto policy of test_sharded_optim_v2 (#2157)
|
2022-12-20 23:03:18 +08:00 |
kernel
|
updated attention kernel (#2133)
|
2022-12-16 10:54:03 +08:00 |
logging
|
fixed logger
|
2022-11-15 16:00:07 +08:00 |
nn
|
[Gemini] GeminiDPP convert to PyTorch Module. (#2151)
|
2022-12-20 10:19:36 +08:00 |
pipeline
|
[PP Middleware] Add bwd and step for PP middleware (#2111)
|
2022-12-12 12:40:03 +08:00 |
registry
|
…
|
|
tensor
|
[autoparallel] memory estimation for shape consistency (#2144)
|
2022-12-21 10:39:37 +08:00 |
testing
|
[zero] test gradient accumulation (#1964)
|
2022-11-29 13:00:30 +08:00 |
trainer
|
[polish] remove useless file _mem_tracer_hook.py (#1963)
|
2022-11-16 15:55:10 +08:00 |
utils
|
[Gemini] Update coloinit_ctx to support meta_tensor (#2147)
|
2022-12-19 22:37:07 +08:00 |
zero
|
[example] add zero1, zero2 example in GPT examples (#2146)
|
2022-12-20 14:30:27 +08:00 |
__init__.py
|
[setup] supported conda-installed torch (#2048)
|
2022-11-30 16:45:15 +08:00 |
constants.py
|
updated tp layers
|
2022-11-02 12:19:38 +08:00 |
core.py
|
…
|
|
global_variables.py
|
updated tp layers
|
2022-11-02 12:19:38 +08:00 |
initialize.py
|
…
|
|