Jiarui Fang
|
bdef9dfdbe
|
[NFC] remove useless graph node code (#2150)
|
2022-12-20 00:33:58 +08:00 |
BlueRum
|
b3f73ce1c8
|
[Gemini] Update coloinit_ctx to support meta_tensor (#2147)
|
2022-12-19 22:37:07 +08:00 |
Jiarui Fang
|
6ad866b684
|
[version] version to v0.1.13 (#2139)
|
2022-12-19 15:38:58 +08:00 |
oahzxl
|
9d516fa68f
|
fix layernorm
|
2022-12-18 20:37:55 +08:00 |
Zihao
|
a128eec9d5
|
register aten._convolution.default (#2137)
|
2022-12-18 19:27:01 +08:00 |
oahzxl
|
e66a18a0bf
|
optimise search
|
2022-12-16 15:06:39 +08:00 |
Jiarui Fang
|
ee287620f0
|
[Gemini] revert ZeROInitCtx related tracer (#2138)
|
2022-12-16 12:37:06 +08:00 |
oahzxl
|
e83e3c6154
|
update memory estimate
|
2022-12-16 11:09:35 +08:00 |
アマデウス
|
077a66dd81
|
updated attention kernel (#2133)
|
2022-12-16 10:54:03 +08:00 |
github-actions[bot]
|
484fe62252
|
Automated submodule synchronization (#2131)
Co-authored-by: github-actions <github-actions@github.com>
|
2022-12-15 09:32:01 +08:00 |
YuliangLiu0306
|
a3c6924deb
|
[autoparallel] process size nodes in runtime pass (#2130)
* [autoparallel] process size nodes in runtime pass
* polish code
|
2022-12-14 16:10:50 +08:00 |
YuliangLiu0306
|
536560ccc0
|
[autoparallel] implement softmax handler (#2132)
|
2022-12-14 16:09:53 +08:00 |
Jiarui Fang
|
c89c66a858
|
[Gemini] update API of the chunkmemstatscollector. (#2129)
|
2022-12-14 00:47:06 +08:00 |
Jiarui Fang
|
2938edf446
|
[Gemini] update the non model data record method in runtime memory tracer (#2128)
|
2022-12-13 17:11:31 +08:00 |
Jiarui Fang
|
deee317b0f
|
[Gemini] test step-tensor mapping using repeated_computed_layers.py (#2127)
|
2022-12-13 16:34:10 +08:00 |
Jiarui Fang
|
8fac837679
|
[Gemini] update non model data calculation method (#2126)
|
2022-12-13 15:44:07 +08:00 |
Fazzie-Maqianli
|
6c4c6a0409
|
Merge pull request #2120 from Fazziekey/example/stablediffusion-v2
[example] support stable diffusion v2
|
2022-12-13 14:38:40 +08:00 |
Fazzie
|
cea4292ae5
|
support stable diffusion v2
|
2022-12-13 14:26:49 +08:00 |
Jiarui Fang
|
5efda69735
|
[Gemini] hotfix the unittest bugs (#2125)
|
2022-12-13 14:14:55 +08:00 |
Jiarui Fang
|
05bb28aacf
|
[Gemini] mapping of preop timestep and param (#2124)
|
2022-12-13 12:50:24 +08:00 |
oahzxl
|
de65e6c3e8
|
support output
|
2022-12-13 11:00:51 +08:00 |
oahzxl
|
cda3e8572a
|
support index dupilictae and update loop
|
2022-12-13 10:02:26 +08:00 |
oahzxl
|
1e0fd11bc1
|
support check_index_duplicate
|
2022-12-13 10:01:30 +08:00 |
github-actions[bot]
|
764bc16f3e
|
Automated submodule synchronization (#2123)
Co-authored-by: github-actions <github-actions@github.com>
|
2022-12-13 09:44:27 +08:00 |
oahzxl
|
8754fa2553
|
change threshold
|
2022-12-12 18:25:47 +08:00 |
oahzxl
|
98f9728e29
|
code style
|
2022-12-12 18:15:47 +08:00 |
YuliangLiu0306
|
cd0af9f7f6
|
[autoparallel] gpt2lp runtimee test (#2113)
|
2022-12-12 18:06:40 +08:00 |
Jiarui Fang
|
9214d1fe28
|
[Gemini] chunk init using runtime visited param order (#2115)
|
2022-12-12 18:06:16 +08:00 |
HELSON
|
e7d3afc9cc
|
[optimizer] add div_scale for optimizers (#2117)
* [optimizer] add div_scale for optimizers
* [zero] use div_scale in zero optimizer
* fix testing error
|
2022-12-12 17:58:57 +08:00 |
oahzxl
|
8511d900a8
|
code style
|
2022-12-12 17:36:17 +08:00 |
oahzxl
|
5cdfcfe1d1
|
code style
|
2022-12-12 17:29:07 +08:00 |
oahzxl
|
b7b67c32ad
|
code style
|
2022-12-12 17:25:38 +08:00 |
oahzxl
|
31a2c5d09f
|
work with outerproductmean and msa
|
2022-12-12 17:24:06 +08:00 |
Jiarui Fang
|
e5aa8333e4
|
[NFC] update chunk manager API (#2119)
|
2022-12-12 16:57:22 +08:00 |
Jiarui Fang
|
e99edfcb51
|
[NFC] polish comments for Chunk class (#2116)
|
2022-12-12 15:39:31 +08:00 |
Ziyue Jiang
|
09d69e1c25
|
[PP Middleware] Add bwd and step for PP middleware (#2111)
* add bwd and step for PP middleware
* pre-commit
Co-authored-by: Ziyue Jiang <ziyue.jiang@gmail.com>
|
2022-12-12 12:40:03 +08:00 |
Jiarui Fang
|
8afc001f4f
|
[Gemini] chunk init use OrderedParamGenerator (#2110)
|
2022-12-11 21:41:13 +08:00 |
oahzxl
|
5de9e46381
|
code format
|
2022-12-10 17:34:48 +08:00 |
oahzxl
|
d31e146687
|
code format
|
2022-12-10 17:34:40 +08:00 |
oahzxl
|
929445116a
|
pass outproduct mean
|
2022-12-10 17:29:51 +08:00 |
HELSON
|
63fbba3c19
|
[zero] add L2 gradient clipping for ZeRO (#2112)
* [zero] add L2 gradient clipping
* [testing] add MlpModel
* [zero] add unit test for grad clipping
* fix atol
|
2022-12-09 18:09:17 +08:00 |
oahzxl
|
979e61db92
|
redesign index tracer, add source and change compute
|
2022-12-09 17:39:02 +08:00 |
Jiarui Fang
|
70a8556946
|
[gemini] get the param visited order during runtime (#2108)
|
2022-12-09 16:13:03 +08:00 |
Jiarui Fang
|
61f31c3cf0
|
[Gemini] NFC, polish search_chunk_configuration (#2107)
|
2022-12-09 15:00:39 +08:00 |
Jiarui Fang
|
8e14344ec9
|
[hotfix] fix a type in ColoInitContext (#2106)
|
2022-12-09 11:44:39 +08:00 |
Jiarui Fang
|
05545bfee9
|
[ColoTensor] throw error when ColoInitContext meets meta parameter. (#2105)
|
2022-12-09 11:39:46 +08:00 |
YuliangLiu0306
|
d87baa85d9
|
[autoparallel] support linear function bias addition (#2104)
|
2022-12-09 10:31:36 +08:00 |
Jiarui Fang
|
6a71d3a0d9
|
[version] 0.1.11rc5 -> 0.1.12 (#2103)
|
2022-12-09 10:12:39 +08:00 |
YuliangLiu0306
|
0fecbb9e20
|
[autoparallel] support addbmm computation (#2102)
|
2022-12-08 21:15:11 +08:00 |
YuliangLiu0306
|
d3d4630495
|
[autoparallel] add sum handler (#2101)
|
2022-12-08 17:02:54 +08:00 |