295 Commits (cloud/coati)

Author SHA1 Message Date
ver217 e4f555f29a
[optim] refactor fused sgd (#1134) 2 years ago
ver217 d26902645e
[ddp] add save/load state dict for ColoDDP (#1127) 2 years ago
ver217 f0a954f16d
[ddp] add set_params_to_ignore for ColoDDP (#1122) 2 years ago
ver217 e127b4375b
cast colo ddp v2 inputs/outputs (#1120) 2 years ago
ver217 7d14b473f0
[gemini] gemini mgr supports "cpu" placement policy (#1118) 2 years ago
ver217 895c1c5ee7
[tensor] refactor param op hook (#1097) 2 years ago
Frank Lee cb18922c47
[doc] added documentation to chunk and chunk manager (#1094) 2 years ago
ver217 1f894e033f
[gemini] zero supports gemini (#1093) 2 years ago
Frank Lee 2b2dc1c86b
[pipeline] refactor the pipeline module (#1087) 2 years ago
ver217 be01db37c8
[tensor] refactor chunk mgr and impl MemStatsCollectorV2 (#1077) 2 years ago
Ziyue Jiang 0653c63eaa
[Tensor] 1d row embedding (#1075) 2 years ago
Ziyue Jiang 4fc748f69b
[Tensor] fix optimizer for CPU parallel (#1069) 3 years ago
Jiarui Fang 49832b2344
[refactory] add nn.parallel module (#1068) 3 years ago
Ziyue Jiang 6754f1b77f
fix module utils bug (#1066) 3 years ago
Jiarui Fang a00644079e
reorgnize colotensor directory (#1062) 3 years ago
Ziyue Jiang df9dcbbff6
[Tensor] add hybrid device demo and fix bugs (#1059) 3 years ago
ver217 51b9a49655
[zero] add zero optimizer for ColoTensor (#1046) 3 years ago
ver217 9492a561c3
[tensor] ColoTensor supports ZeRo (#1015) 3 years ago
ver217 cefc29ff06
[tensor] impl ColoDDP for ColoTensor (#1009) 3 years ago
Ziheng Qin 571f12eff3 [NFC] polish colossalai/nn/layer/utils/common.py code style (#983) 3 years ago
shenggan 18542b47fc [NFC] polish colossalai/nn/layer/parallel_2d/layers.py code style (#976) 3 years ago
Zirui Zhu 598cde4a0f [NFC] polish colossalai/nn/layer/parallel_2p5d/layers.py code style (#972) 3 years ago
LuGY fb5bc6cb28 [NFC] polish colossalai/nn/layer/parallel_3d/layers.py code style (#966) 3 years ago
ver217 58580b50fe
Revert "[NFC] Hotfix/format (#984)" (#986) 3 years ago
binmakeswell 0772828fba
[NFC] Hotfix/format (#984) 3 years ago
HELSON e5ea3fdeef
[gemini] add GeminiMemoryManger (#832) 3 years ago
Ziyue Jiang 4b01da24cd
[TP] change the check assert in split batch 2d (#772) 3 years ago
アマデウス b8899e0905
[TP] allow layernorm without bias (#750) 3 years ago
Frank Lee eda30a058e
[compatibility] fixed tensor parallel compatibility with torch 1.9 (#700) 3 years ago
HELSON a9b8300d54
[zero] improve adaptability for not-shard parameters (#708) 3 years ago
アマデウス 3fc8a204dc
[]Corrected 3d vocab parallel embedding (#707) 3 years ago
HELSON b31daed4cf
fix bugs in CPU adam (#633) 3 years ago
Liang Bowen 828e465622
[hotfix] Raise messages for indivisible batch sizes with tensor parallelism (#622) 3 years ago
アマデウス 77ad24bf94
[model checkpoint] updated saving/loading for 3d layers (#597) 3 years ago
アマデウス 93089ed708
[model checkpoint] updated saving/loading for 2.5d layers (#596) 3 years ago
アマデウス c50bfb807b
[model checkpoint] updated saving/loading for 1d layers (#594) 3 years ago
アマデウス 7636d518e1
[model checkpoint] updated saving/loading for 2d layers (#595) 3 years ago
アマデウス cd13b63832
[model checkpoint] reworked unified layers for ease of save/load states (#593) 3 years ago
Ziyue Jiang 1c40ee8749
[TP] add assert for tp1d (#621) 3 years ago
ver217 e619a651fb
polish optimizer docstring (#619) 3 years ago
ver217 8432dc7080
polish moe docsrting (#618) 3 years ago
ver217 104cbbb313
[hotfix] add hybrid adam to __init__ (#584) 3 years ago
HELSON e6d50ec107
[zero] adapt zero for unsharded parameters (#561) 3 years ago
Wesley 46c9ba33da update code format 3 years ago
Wesley 666cfd094a fix parallel_input flag for Linear1D_Col gather_output 3 years ago
Liang Bowen 2c45efc398
html refactor (#555) 3 years ago
LuGY c44d797072
[docs] updatad docs of hybrid adam and cpu adam (#552) 3 years ago
Ziyue Jiang 763dc325f1
[TP] Add gather_out arg to Linear (#541) 3 years ago
HELSON 8c90d4df54
[zero] add zero context manager to change config during initialization (#546) 3 years ago
Liang Bowen ec5086c49c Refactored docstring to google style 3 years ago