Commit Graph

21 Commits (ckpt)

Author SHA1 Message Date
botbw c54c4fcd15
[hotfix] moe hybrid parallelism benchmark & follow-up fix (#6048)
3 months ago
Wang Binluo eea37da6fa
[fp8] Merge feature/fp8_comm to main branch of Colossalai (#6016)
3 months ago
flybird11111 0a51319113
[fp8] zero support fp8 linear. (#6006)
3 months ago
Hongxin Liu 8241c0c054
[fp8] support gemini plugin (#5978)
4 months ago
Hanks b480eec738
[Feature]: support FP8 communication in DDP, FSDP, Gemini (#5928)
4 months ago
flybird11111 0c10afd372
[FP8] rebase main (#5963)
4 months ago
botbw 8e718a1421
[gemini] fixes for benchmarking (#5847)
5 months ago
Edenzzzz 2a25a2aff7
[Feature] optimize PP overlap (#5735)
5 months ago
Edenzzzz 8795bb2e80
Support 4d parallel + flash attention (#5789)
5 months ago
genghaozhe 87665d7922 correct argument help message
6 months ago
genghaozhe b9269d962d add args.prefetch_num for benchmark
6 months ago
hxwang ff507b755e Merge branch 'main' of github.com:hpcaitech/ColossalAI into prefetch
6 months ago
hxwang 63c057cd8e [example] add profile util for llama
6 months ago
botbw 2fc85abf43
[gemini] async grad chunk reduce (all-reduce&reduce-scatter) (#5713)
6 months ago
Hongxin Liu 7f8b16635b
[misc] refactor launch API and tensor constructor (#5666)
7 months ago
Tong Li 68ec99e946
[hotfix] add soft link to support required files (#5661)
7 months ago
Hongxin Liu 1b387ca9fe
[shardformer] refactor pipeline grad ckpt config (#5646)
7 months ago
binmakeswell f4c5aafe29
[example] llama3 (#5631)
7 months ago
Hongxin Liu 0b00def881
[example] add llama2 example (#4527)
1 year ago
binmakeswell ef4b99ebcd add llama example CI
1 year ago
binmakeswell 7ff11b5537
[example] add llama pretraining (#4257)
1 year ago