167 Commits (80c3c8789bdfc095618a0b725a4980cd575c6c6b)

Author SHA1 Message Date
Hongxin Liu 7f8b16635b
[misc] refactor launch API and tensor constructor (#5666) 7 months ago
flybird11111 a0ad587c24
[shardformer] refactor embedding resize (#5603) 7 months ago
Edenzzzz 61da3fbc52 fixed layout converter caching and updated tester 8 months ago
Hongxin Liu 079bf3cb26
[misc] update pre-commit and run all files (#4752) 1 year ago
Hongxin Liu b5f9e37c70
[legacy] clean up legacy code (#4743) 1 year ago
Hongxin Liu 27061426f7
[gemini] improve compatibility and add static placement policy (#4479) 1 year ago
Frank Lee c4b1b65931 [test] fixed tests failed due to dtensor change (#4082) 1 year ago
Frank Lee 8eb09a4c69 [shardformer] support module saving and loading (#4062) 1 year ago
Baizhou Zhang 0bb0b481b4 [gemini] fix argument naming during chunk configuration searching 1 year ago
digger yu e61ffc77c6
fix typo tests/ (#3936) 1 year ago
Frank Lee ddcf58cacf
Revert "[sync] sync feature/shardformer with develop" 1 year ago
Frank Lee eb39154d40
[dtensor] updated api and doc (#3845) 1 year ago
digger-yu 1f73609adb
[CI] fix typo with tests/ etc. (#3727) 2 years ago
Frank Lee 80eba05b0a
[test] refactor tests with spawn (#3452) 2 years ago
ver217 26b7aac0be
[zero] reorganize zero/gemini folder structure (#3424) 2 years ago
YuliangLiu0306 2eca4cd376
[DTensor] refactor dtensor with new components (#3089) 2 years ago
YuliangLiu0306 8e4e8601b7
[DTensor] implement layout converter (#3055) 2 years ago
YuliangLiu0306 29386a54e6
[DTensor] refactor CommSpec (#3034) 2 years ago
YuliangLiu0306 4269196c79
[hotfix] skip auto checkpointing tests (#3029) 2 years ago
YuliangLiu0306 cd2b0eaa8d
[DTensor] refactor sharding spec (#2987) 2 years ago
YuliangLiu0306 e414e4092b
[DTensor] implementation of dtensor (#2946) 2 years ago
HELSON 707b11d4a0
[gemini] update ddp strict mode (#2518) 2 years ago
HELSON 2d1a7dfe5f
[zero] add strict ddp mode (#2508) 2 years ago
HELSON d565a24849
[zero] add unit testings for hybrid parallelism (#2486) 2 years ago
HELSON ea13a201bb
[polish] polish code for get_static_torch_model (#2405) 2 years ago
HELSON a3100bd50d
[testing] add beit model for unit testings (#2196) 2 years ago
Jiarui Fang 1f99205827
[Gemini] remove static tracer (#2083) 2 years ago
Jiarui Fang 2e9cbfca12
[Gemini] add unitests to check gemini correctness (#2015) 2 years ago
Genghan Zhang d655eea515
[autoparallel] mix gather (#1977) 2 years ago
Jiarui Fang f7e276fa71
[Gemini] add GeminiAdamOptimizer (#1960) 2 years ago
Jiarui Fang 52c6ad26e0
[ColoTensor] reconfig ColoInitContext, decouple default_pg and default_dist_spec. (#1953) 2 years ago
Jiarui Fang 9f4fb3f28a
[ColoTensor] ColoInitContext initialize parameters in shard mode. (#1937) 2 years ago
Jiarui Fang 3ce4463fe6
[utils] remove lazy_memory_allocate from ColoInitContext (#1844) 2 years ago
YuliangLiu0306 980ed21723
[autoparallel] shard param and buffer as expected (#1753) 2 years ago
Frank Lee eee84908d4
[autoparallel] handled illegal sharding strategy (#1728) 2 years ago
HELSON f69f9bf223
[zero] add chunk init function for users (#1729) 2 years ago
HELSON b28991dd0a
[feature] A new ZeRO implementation (#1644) 2 years ago
YuliangLiu0306 3f068d1409
[autoparallel] update CommSpec (#1667) 2 years ago
Frank Lee 154d3ef432
[fix] fixed the collective pattern name for consistency (#1649) 2 years ago
Jiarui Fang c5d39215f6
Revert "[feature] new zero implementation (#1623)" (#1643) 2 years ago
HELSON 5be118f405
[feature] new zero implementation (#1623) 2 years ago
YuliangLiu0306 702dbc5288
[tensor] use communication autograd func (#1617) 2 years ago
YuliangLiu0306 4b03c25f85
[tensor]add 1D device mesh (#1492) 2 years ago
YuliangLiu0306 b73fb7a077
[tensor] support runtime ShardingSpec apply (#1453) 2 years ago
YuliangLiu0306 0f3042363c
[tensor] shape consistency generate transform path and communication cost (#1435) 2 years ago
Frank Lee ae1b58cd16
[tensor] added linear implementation for the new sharding spec (#1416) 2 years ago
Jiarui Fang 89c434a0a6
[polish] add test_ops directory (#1431) 2 years ago
Jiarui Fang 10b3df65c8
[FAW] move coloparam setting in test code. (#1429) 2 years ago
Jiarui Fang cb98cf5558
[FAW] parallel FreqAwareEmbedding (#1424) 2 years ago
YuliangLiu0306 33f0744d51
[tensor] add shape consistency feature to support auto spec transform (#1418) 2 years ago