Baizhou Zhang
|
a2db75546d
|
[doc] polish shardformer doc (#4779)
* fix example format in docstring
* polish shardformer doc
|
2023-09-26 10:57:47 +08:00 |
Baizhou Zhang
|
c0a033700c
|
[shardformer] fix master param sync for hybrid plugin/rewrite unwrapping logic (#4758)
* fix master param sync for hybrid plugin
* rewrite unwrap for ddp/fsdp
* rewrite unwrap for zero/gemini
* rewrite unwrap for hybrid plugin
* fix geemini unwrap
* fix bugs
|
2023-09-20 18:29:37 +08:00 |
Hongxin Liu
|
079bf3cb26
|
[misc] update pre-commit and run all files (#4752)
* [misc] update pre-commit
* [misc] run pre-commit
* [misc] remove useless configuration files
* [misc] ignore cuda for clang-format
|
2023-09-19 14:20:26 +08:00 |
LuGY
|
79cf1b5f33
|
[zero]support no_sync method for zero1 plugin (#4138)
* support no sync for zero1 plugin
* polish
* polish
|
2023-07-31 22:13:29 +08:00 |
Wenhao Chen
|
725af3eeeb
|
[booster] make optimizer argument optional for boost (#3993)
* feat: make optimizer optional in Booster.boost
* test: skip unet test if diffusers version > 0.10.2
|
2023-06-15 17:38:42 +08:00 |
Baizhou Zhang
|
c9cff7e7fa
|
[checkpointio] General Checkpointing of Sharded Optimizers (#3984)
|
2023-06-15 15:21:26 +08:00 |
wukong1992
|
3229f93e30
|
[booster] add warning for torch fsdp plugin doc (#3833)
|
2023-05-25 14:00:02 +08:00 |
wukong1992
|
6b305a99d6
|
[booster] torch fsdp fix ckpt (#3788)
|
2023-05-23 16:58:45 +08:00 |
wukong1992
|
b37797ed3d
|
[booster] support torch fsdp plugin in booster (#3697)
Co-authored-by: 纪少敏 <jishaomin@jishaomindeMBP.lan>
|
2023-05-15 12:14:38 +08:00 |