327 Commits (457a0de79fd2d3602eba0ac78e606acb6401fc60)

Author SHA1 Message Date
ver217 3092317b80 polish code 3 years ago
ver217 36f9a74ab2 fix sharded param hook and unit test 3 years ago
ver217 001ca624dd impl shard optim v2 and add unit test 3 years ago
Jiarui Fang 74f77e314b [zero] a shard strategy in granularity of tensor (#307) 3 years ago
Jiarui Fang 80364c7686 [zero] sharded tensor (#305) 3 years ago
ver217 b105371ace rename shared adam to sharded optim v2 3 years ago
ver217 70814dc22f fix master params dtype 3 years ago
ver217 795210dd99 add fp32 master params in sharded adam 3 years ago
ver217 a109225bc2 add sharded adam 3 years ago
Jiarui Fang e17e92c54d Polish sharded parameter (#297) 3 years ago
ver217 7aef75ca42 [zero] add sharded grad and refactor grad hooks for ShardedModel (#287) 3 years ago
Frank Lee 9afb5c8b2d fixed typo in ShardParam (#294) 3 years ago
Frank Lee e17e54e32a added buffer sync to naive amp model wrapper (#291) 3 years ago
Jiarui Fang 5a560a060a Feature/zero (#279) 3 years ago
HELSON 0f8c7f9804
Fixed docstring in colossalai (#171) 3 years ago
ver217 9ef05ed1fc
try import deepspeed when using zero (#130) 3 years ago
Frank Lee 91c327cb44
fixed zero level 3 dtype bug (#76) 3 years ago
Frank Lee 35813ed3c4
update examples and sphnix docs for the new api (#63) 3 years ago
ver217 7d3711058f
fix zero3 fp16 and add zero3 model context (#62) 3 years ago
Frank Lee da01c234e1
Develop/experiments (#59) 3 years ago