1310 Commits (258b43317c4a5cafb8d3da0ff63c8843443bc448)

Author SHA1 Message Date
YuliangLiu0306 258b43317c
[hotfix] layout converting issue (#3188) 2 years ago
YH 80aed29cd3
[zero] Refactor ZeroContextConfig class using dataclass (#3186) 2 years ago
YH 9d644ff09f
Fix docstr for zero statedict (#3185) 2 years ago
zbian 7bc0afc901 updated flash attention usage 2 years ago
Frank Lee a9b8402d93
[booster] added the accelerator implementation (#3159) 2 years ago
ver217 6ae8ed0407
[lazyinit] add correctness verification (#3147) 2 years ago
Frank Lee ed19290560
[booster] implemented mixed precision class (#3151) 2 years ago
YuliangLiu0306 2eca4cd376
[DTensor] refactor dtensor with new components (#3089) 2 years ago
ver217 ed8f60b93b
[lazyinit] refactor lazy tensor and lazy init ctx (#3131) 2 years ago
Frank Lee 95a36eae63
[kernel] added kernel loader to softmax autograd function (#3093) 2 years ago
Super Daniel fff98f06ed
[analyzer] a minimal implementation of static graph analyzer (#2852) 2 years ago
Xuanlei Zhao 10c61de2f7
[autochunk] support vit (#3084) 2 years ago
YuliangLiu0306 8e4e8601b7
[DTensor] implement layout converter (#3055) 2 years ago
Frank Lee f19b49e164
[booster] init module structure and definition (#3056) 2 years ago
Xuanlei Zhao 2ca9728cbb
[autochunk] refactor chunk memory estimation (#2762) 2 years ago
YuliangLiu0306 29386a54e6
[DTensor] refactor CommSpec (#3034) 2 years ago
YuliangLiu0306 cd2b0eaa8d
[DTensor] refactor sharding spec (#2987) 2 years ago
Ziyue Jiang 400f63012e
[pipeline] Add Simplified Alpa DP Partition (#2507) 2 years ago
Super Daniel b42d3d28ed
[fx] remove depreciated algorithms. (#2312) (#2313) 2 years ago
github-actions[bot] 82503a96f2
[format] applied code formatting on changed files in pull request 2997 (#3008) 2 years ago
binmakeswell 52a5078988
[doc] add ISC tutorial (#2997) 2 years ago
ver217 823f3b9cf4
[doc] add deepspeed citation and copyright (#2996) 2 years ago
YuliangLiu0306 e414e4092b
[DTensor] implementation of dtensor (#2946) 2 years ago
YuliangLiu0306 47fb214b3b
[hotfix] add shard dim to aviod backward communication error (#2954) 2 years ago
ver217 090f14fd6b
[misc] add reference (#2930) 2 years ago
YuliangLiu0306 197d0bf4ed
[autoparallel] apply repeat block to reduce solving time (#2912) 2 years ago
YH a848091141
Fix port exception type (#2925) 2 years ago
zbian 61e687831d fixed using zero with tp cannot access weight correctly 2 years ago
YH 7b13f7db18
[zero] trivial zero optimizer refactoring (#2869) 2 years ago
Jiatong (Julius) Han 8c8a39be95
[hotfix]: Remove math.prod dependency (#2837) 2 years ago
YuliangLiu0306 819e25d8b1
[hotfix] fix autoparallel compatibility test issues (#2754) 2 years ago
YuliangLiu0306 0f392d7403
[autoparallel] find repeat blocks (#2854) 2 years ago
junxu c52edcf0eb
Rename class method of ZeroDDP (#2692) 2 years ago
HELSON 6e4ac08172
[hotfix] fix chunk size can not be divided (#2867) 2 years ago
Boyuan Yao eae77c831d
[autoparallel] Patch meta information for nodes that will not be handled by SPMD solver (#2823) 2 years ago
Boyuan Yao c7764d3f22
[autoparallel] Patch meta information of `torch.where` (#2822) 2 years ago
Boyuan Yao fcc4097efa
[autoparallel] Patch meta information of `torch.tanh()` and `torch.nn.Dropout` (#2773) 2 years ago
Frank Lee 935346430f
[cli] handled version check exceptions (#2848) 2 years ago
Frank Lee 918bc94b6b
[triton] added copyright information for flash attention (#2835) 2 years ago
Boyuan Yao 7ea6bc7f69
[autoparallel] Patch tensor related operations meta information (#2789) 2 years ago
Michelle c008d4ad0c
[NFC] polish colossalai/engine/schedule/_pipeline_schedule.py code style (#2744) 2 years ago
YuliangLiu0306 2059fdd6b0
[hotfix] add copyright for solver and device mesh (#2803) 2 years ago
Boyuan Yao 8593ae1a3f
[autoparallel] rotor solver refactor (#2813) 2 years ago
HELSON 56ddc9ca7a
[hotfix] add correct device for fake_param (#2796) 2 years ago
Boyuan Yao a2b43e393d
[autoparallel] Patch meta information of `torch.nn.Embedding` (#2760) 2 years ago
Boyuan Yao 8e3f66a0d1
[zero] fix wrong import (#2777) 2 years ago
Nikita Shulga 01066152f1
Don't use `torch._six` (#2775) 2 years ago
YuliangLiu0306 1dc003c169
[autoparallel] distinguish different parallel strategies (#2699) 2 years ago
YH ae86a29e23
Refact method of grad store (#2687) 2 years ago
xyupeng 2fd528b9f4
[NFC] polish colossalai/auto_parallel/tensor_shard/deprecated/graph_analysis.py code style (#2737) 2 years ago