Shawn-Kong
|
1712da2800
|
[NFC] polish colossalai/gemini/gemini_context.py code style (#2690)
|
2023-02-14 11:55:23 +08:00 |
HELSON
|
8213f89fd2
|
[gemini] add fake_release_chunk for keep-gathered chunk in the inference mode (#2671)
|
2023-02-13 14:35:32 +08:00 |
ver217
|
5b1854309a
|
[hotfix] fix zero ddp warmup check (#2545)
|
2023-02-02 16:42:38 +08:00 |
HELSON
|
707b11d4a0
|
[gemini] update ddp strict mode (#2518)
* [zero] add strict ddp mode for chunk init
* [gemini] update gpt example
|
2023-01-28 14:35:25 +08:00 |
HELSON
|
2bfeb24308
|
[zero] add warning for ignored parameters (#2446)
|
2023-01-11 15:30:09 +08:00 |
HELSON
|
7829aa094e
|
[ddp] add is_ddp_ignored (#2434)
[ddp] rename to is_ddp_ignored
|
2023-01-11 12:22:45 +08:00 |
HELSON
|
bb4e9a311a
|
[zero] add inference mode and its unit test (#2418)
|
2023-01-11 10:07:37 +08:00 |
Jiarui Fang
|
af32022f74
|
[Gemini] fix the convert_to_torch_module bug (#2269)
|
2023-01-03 15:55:35 +08:00 |
Jiarui Fang
|
d5e3e3ec01
|
[example] update gpt example for larger model scale (#2211)
|
2022-12-28 13:54:08 +08:00 |
HELSON
|
2458659919
|
[zero] fix error for BEiT models (#2169)
* [zero] fix error for BEiT models
* [ColoParameter] add unpack operation for tuple arguments
* fix bugs
* fix chunkv2 unit testing
* add assertion for gradient state
|
2022-12-26 15:03:54 +08:00 |
Jiarui Fang
|
b87496a66b
|
[hotfix] fix auto policy of test_sharded_optim_v2 (#2157)
|
2022-12-20 23:03:18 +08:00 |
Jiarui Fang
|
ee287620f0
|
[Gemini] revert ZeROInitCtx related tracer (#2138)
|
2022-12-16 12:37:06 +08:00 |
Jiarui Fang
|
c89c66a858
|
[Gemini] update API of the chunkmemstatscollector. (#2129)
|
2022-12-14 00:47:06 +08:00 |
Jiarui Fang
|
2938edf446
|
[Gemini] update the non model data record method in runtime memory tracer (#2128)
|
2022-12-13 17:11:31 +08:00 |
Jiarui Fang
|
8fac837679
|
[Gemini] update non model data calculation method (#2126)
|
2022-12-13 15:44:07 +08:00 |
Jiarui Fang
|
5efda69735
|
[Gemini] hotfix the unittest bugs (#2125)
|
2022-12-13 14:14:55 +08:00 |
Jiarui Fang
|
05bb28aacf
|
[Gemini] mapping of preop timestep and param (#2124)
|
2022-12-13 12:50:24 +08:00 |
Jiarui Fang
|
9214d1fe28
|
[Gemini] chunk init using runtime visited param order (#2115)
|
2022-12-12 18:06:16 +08:00 |
Jiarui Fang
|
e5aa8333e4
|
[NFC] update chunk manager API (#2119)
|
2022-12-12 16:57:22 +08:00 |
Jiarui Fang
|
e99edfcb51
|
[NFC] polish comments for Chunk class (#2116)
|
2022-12-12 15:39:31 +08:00 |
Jiarui Fang
|
8afc001f4f
|
[Gemini] chunk init use OrderedParamGenerator (#2110)
|
2022-12-11 21:41:13 +08:00 |
HELSON
|
63fbba3c19
|
[zero] add L2 gradient clipping for ZeRO (#2112)
* [zero] add L2 gradient clipping
* [testing] add MlpModel
* [zero] add unit test for grad clipping
* fix atol
|
2022-12-09 18:09:17 +08:00 |
Jiarui Fang
|
70a8556946
|
[gemini] get the param visited order during runtime (#2108)
|
2022-12-09 16:13:03 +08:00 |
Jiarui Fang
|
61f31c3cf0
|
[Gemini] NFC, polish search_chunk_configuration (#2107)
|
2022-12-09 15:00:39 +08:00 |
Jiarui Fang
|
85efb7ac2e
|
[Gemini] gemini use the runtime memory tracer (RMT) (#2099)
|
2022-12-07 23:04:02 +08:00 |
Jiarui Fang
|
4b055351b0
|
[Gemini] make RuntimeMemTracer work correctly (#2096)
|
2022-12-07 16:59:59 +08:00 |
Jiarui Fang
|
1fca5d79ea
|
[Gemini] remove GLOBAL_MODEL_DATA_TRACER (#2091)
|
2022-12-06 22:30:16 +08:00 |
Jiarui Fang
|
28e55c2530
|
[Gemini] remove GLOBAL_CUDA_MEM_INFO (#2090)
|
2022-12-06 22:10:47 +08:00 |
Jiarui Fang
|
25abae6d7f
|
[Gemini] use MemStats in Runtime Memory tracer (#2088)
|
2022-12-06 19:48:20 +08:00 |
Jiarui Fang
|
33f4412102
|
[Gemini] use MemStats to store the tracing data. Seperate it from Collector. (#2084)
|
2022-12-06 16:43:06 +08:00 |
Jiarui Fang
|
1f99205827
|
[Gemini] remove static tracer (#2083)
|
2022-12-06 12:53:58 +08:00 |
Jiarui Fang
|
b3b89865e2
|
[Gemini] ParamOpHook -> ColoParamOpHook (#2080)
|
2022-12-05 17:11:06 +08:00 |
Jiarui Fang
|
a7adad9ccb
|
[Gemini] rename hooks related to runtime mem tracer (#2076)
|
2022-12-05 15:00:03 +08:00 |
Jiarui Fang
|
223332ff7e
|
[Gemini] rename ParamTracerWrapper -> RuntimeMemTracer (#2073)
|
2022-12-05 12:45:11 +08:00 |
Jiarui Fang
|
9f828ef36f
|
[Gemini] remove not used MemtracerWrapper (#2072)
|
2022-12-05 11:57:59 +08:00 |
Zihao
|
38ea4ba1bd
|
[Gemini] fix grad unreleased issue and param recovery issue (#2052)
|
2022-12-02 16:04:19 +08:00 |
Zihao
|
6a9158f1fa
|
[Gemini] free and allocate cuda memory by tensor.storage, add grad hook (#2040)
|
2022-11-30 15:57:45 +08:00 |
Jiarui Fang
|
28aa9a4294
|
[Gemini] more rigorous unit tests for run_fwd_bwd (#2034)
|
2022-11-29 09:26:06 +08:00 |
Zihao
|
95c4532fff
|
[Gemini] paramWrapper paramTracerHook unitest (#2030)
|
2022-11-26 13:30:24 +08:00 |
Jiarui Fang
|
8daf1b4db1
|
[Gemini] patch for supporting orch.add_ function for ColoTensor (#2003)
|
2022-11-25 20:06:35 +08:00 |
Zihao
|
a719b89a41
|
[gemini] param_trace_hook (#2020)
|
2022-11-24 18:08:36 +08:00 |
Jiarui Fang
|
0b0d8f9e17
|
[hotfix] revert bug PRs (#2016)
|
2022-11-24 15:28:58 +08:00 |
Zihao
|
aba3db464d
|
[Gemini] ParamMemHook (#2008)
|
2022-11-24 15:22:51 +08:00 |
Zihao
|
0160a62a3c
|
[Gemini] param_tracer_wrapper and test case (#2009)
|
2022-11-24 14:40:33 +08:00 |
Jiarui Fang
|
3712ac7f90
|
[Gemini] add bert for MemtracerWrapper unintests (#1982)
|
2022-11-18 14:58:28 +08:00 |
Jiarui Fang
|
e481489aa6
|
[Gemini] MemtracerWrapper unittests (#1981)
|
2022-11-18 14:19:40 +08:00 |
Jiarui Fang
|
31922110ad
|
[Gemini] memory trace hook (#1978)
|
2022-11-18 11:52:55 +08:00 |
Jiarui Fang
|
0529fcde06
|
[Gemini] independent runtime tracer (#1974)
|
2022-11-18 10:53:42 +08:00 |
Jiarui Fang
|
7e24b9b9ee
|
[Gemini] clean no used MemTraceOp (#1970)
|
2022-11-17 13:41:54 +08:00 |
Jiarui Fang
|
c4739a725a
|
[Gemini] polish memstats collector (#1962)
|
2022-11-16 15:45:57 +08:00 |