HELSON
|
5521af7877
|
[zero] fix state_dict and load_state_dict for ddp ignored parameters (#2443)
* [ddp] add is_ddp_ignored
[ddp] rename to is_ddp_ignored
* [zero] fix state_dict and load_state_dict
* fix bugs
* [zero] update unit test for ZeroDDP
|
2 years ago |
YuliangLiu0306
|
41429b9b28
|
[autoparallel] add shard option (#2423)
|
2 years ago |
HELSON
|
bb4e9a311a
|
[zero] add inference mode and its unit test (#2418)
|
2 years ago |
oahzxl
|
61fdd3464a
|
update doc
|
2 years ago |
oahzxl
|
36ab2cb783
|
change import
|
2 years ago |
oahzxl
|
7ab2db206f
|
adapt new fx
|
2 years ago |
oahzxl
|
e532679c95
|
Merge branch 'main' of https://github.com/oahzxl/ColossalAI into chunk
|
2 years ago |
oahzxl
|
c1492e5013
|
add test in import
|
2 years ago |
HELSON
|
ea13a201bb
|
[polish] polish code for get_static_torch_model (#2405)
* [gemini] polish code
* [testing] remove code
* [gemini] make more robust
|
2 years ago |
oahzxl
|
212b5b1b5f
|
add comments
|
2 years ago |
oahzxl
|
aafc3516a5
|
add available
|
2 years ago |
oahzxl
|
d5c4f0bf95
|
code style
|
2 years ago |
oahzxl
|
d106b271f8
|
add chunk search test
|
2 years ago |
oahzxl
|
a005965d2d
|
update codegen test
|
2 years ago |
oahzxl
|
3abbaf8bc6
|
update codegen test
|
2 years ago |
oahzxl
|
74b81395a2
|
update codegen test
|
2 years ago |
oahzxl
|
18a51c87fe
|
rename test
|
2 years ago |
oahzxl
|
cb68ee864a
|
set benchmark
|
2 years ago |
Jiarui Fang
|
4e96039649
|
[device] find best logical mesh
|
2 years ago |
Frank Lee
|
40d376c566
|
[setup] support pre-build and jit-build of cuda kernels (#2374)
* [setup] support pre-build and jit-build of cuda kernels
* polish code
* polish code
* polish code
* polish code
* polish code
* polish code
|
2 years ago |
oahzxl
|
a6cdbf9161
|
seperate trace flow
|
2 years ago |
oahzxl
|
da4076846d
|
rename
|
2 years ago |
oahzxl
|
fd87d78a28
|
rename ambiguous variable
|
2 years ago |
oahzxl
|
8a634af2f5
|
close mem and code print
|
2 years ago |
oahzxl
|
1a6d2a740b
|
take apart chunk code gen
|
2 years ago |
HELSON
|
48d33b1b17
|
[gemini] add get static torch model (#2356)
|
2 years ago |
oahzxl
|
d1f0773182
|
rename
|
2 years ago |
oahzxl
|
06a5355d98
|
update test
|
2 years ago |
oahzxl
|
efb1c64c30
|
restruct dir
|
2 years ago |
YuliangLiu0306
|
b5a3a4a65f
|
[device] find best logical mesh
|
2 years ago |
YuliangLiu0306
|
9c9246c0d9
|
[device] alpha beta profiler (#2311)
* [device] alpha beta profiler
* add usage
* fix variable name
|
2 years ago |
Jiarui Fang
|
db6eea3583
|
[builder] reconfig op_builder for pypi install (#2314)
|
2 years ago |
HELSON
|
5d3a2be3af
|
[amp] add gradient clipping for unit tests (#2283)
* [amp] add gradient clipping in unit tests
* fix bugs
|
2 years ago |
zbian
|
e94c79f15b
|
improved allgather & reducescatter for 3d
|
2 years ago |
YuliangLiu0306
|
fb87322773
|
[autoparallel] fix spelling error (#2270)
|
2 years ago |
YuliangLiu0306
|
8897b8f753
|
[autoparallel] autoparallel initialize (#2238)
|
2 years ago |
YuliangLiu0306
|
3b1b91eaf4
|
[autoparallel] record parameter attribute in colotracer (#2217)
* [autoparallel] record parameter attribute in collotracer
* [autoparallel] fix construct_meta_info bug
|
2 years ago |
Boyuan Yao
|
24246f7aa5
|
[autoparallel] Attach input, buffer and output tensor to MetaInfo class (#2162)
* [fx] metainfo class for auto parallel
* [fx] add unit test for linear metainfo
* [fx] fix bwd param for linear
* [fx] modify unit test
* [fx] modify unit test
* [fx] modify import
* [fx] modify import
* [fx] modify import
* [fx] move meta profiler to auto parallel
* [fx] add conv metainfo class
* [fx] restore profiler
* [fx] restore meta profiler
* [autoparallel] modify unit test
* [fx] modify unit test
* [autoparallel] add batchnorm metainfo class
* [autoparallel] fix batchnorm unit test function declaration
* [fx] restore profiler
* [fx] add relu metainfo class
* [fx] restore profiler
* [autoparallel] modify metainfo input
* [autoparallel] add pooling metainfo
* [autoparallel] add F.linear metainfo generator
* [autoparallel] add binary elementwise metainfo
* [fx] recover profiler
* [autoparallel] fix forward memory calculation
* [autoparallel] modify constants.py
* [autoparallel] remove redundant print
* [autoparallel] add F.conv metainfo
* [autoparallel] linear fix
* [autoparallel] memory estimation for communication actions
* [autoparallel] fix docstring
* [autoparallel] fix variables name
* [autoparallel] attach tensor to metainfo class
* [autoparallel] fix dangerous try except
* [autoparallel] attach memory cost to shape consistency node
* [autoparallel] attach shape consistency node's metainfo to the node
* [autoparallel] remove todo in shape consistency memory estimation
* [autoparallel] fix the annotation
|
2 years ago |
YuliangLiu0306
|
78509124d3
|
[autoparallel] update getitem handler (#2207)
|
2 years ago |
YuliangLiu0306
|
4851f2d607
|
[autoparallel] update_getattr_handler (#2193)
|
2 years ago |
YuliangLiu0306
|
f10ce01e31
|
[autoparallel] add gpt2 performance test code (#2194)
|
2 years ago |
HELSON
|
a3100bd50d
|
[testing] add beit model for unit testings (#2196)
* [testing] add beit model
* [beit] fix bugs
* [beit] fix bugs
* [testing] fix bugs
|
2 years ago |
HELSON
|
2458659919
|
[zero] fix error for BEiT models (#2169)
* [zero] fix error for BEiT models
* [ColoParameter] add unpack operation for tuple arguments
* fix bugs
* fix chunkv2 unit testing
* add assertion for gradient state
|
2 years ago |
Jiarui Fang
|
355ffb386e
|
[builder] unified cpu_optim fused_optim inferface (#2190)
|
2 years ago |
Jiarui Fang
|
9587b080ba
|
[builder] use runtime builder for fused_optim (#2189)
|
2 years ago |
Jiarui Fang
|
bc0e271e71
|
[buider] use builder() for cpu adam and fused optim in setup.py (#2187)
|
2 years ago |
Jiarui Fang
|
d42afd30f8
|
[builder] runtime adam and fused_optim builder (#2184)
|
2 years ago |
YuliangLiu0306
|
550f8f8905
|
[autoparallel] integrate_gpt_related_tests (#2134)
* [autoparallel] integrate_gpt_related_tests
* polish code
* polish code
* add GPT2Model into runtime test
|
2 years ago |
Jiarui Fang
|
27327a4c90
|
[example] add palm pytorch version (#2172)
|
2 years ago |
Jiarui Fang
|
b87496a66b
|
[hotfix] fix auto policy of test_sharded_optim_v2 (#2157)
|
2 years ago |