Commit Graph

2209 Commits (fee2af861078181e18a059800ba09a4edfb311f0)

Author SHA1 Message Date
mickogoin 58abde2857
Update README.md (#2791)
Fixed typo on line 285 from "defualt" to "default"
2023-02-20 10:37:57 +08:00
Marco Rodrigues 89f0017a9c
Typo (#2826) 2023-02-20 10:36:23 +08:00
Jiarui Fang bf0204604f
[exmaple] add bert and albert (#2824) 2023-02-20 10:35:55 +08:00
YuliangLiu0306 cf6409dd40
Hotfix/auto parallel zh doc (#2820)
* [hotfix] fix autoparallel zh docs

* polish

* polish
2023-02-19 15:57:14 +08:00
YuliangLiu0306 2059fdd6b0
[hotfix] add copyright for solver and device mesh (#2803)
* [hotfix] add copyright for solver and device mesh

* add readme

* add alpa license

* polish
2023-02-18 21:14:38 +08:00
LuGY dbd0fd1522
[CI/CD] fix nightly release CD running on forked repo (#2812)
* [CI/CD] fix nightly release CD running on forker repo

* fix misunderstanding of dispatch

* remove some build condition, enable notify even when release failed
2023-02-18 13:27:13 +08:00
Boyuan Yao 8593ae1a3f
[autoparallel] rotor solver refactor (#2813)
* [autoparallel] rotor solver refactor

* [autoparallel] rotor solver refactor
2023-02-18 11:30:15 +08:00
binmakeswell 09f457479d
[doc] update OPT serving (#2804)
* [doc] update OPT serving

* [doc] update OPT serving
2023-02-17 23:21:42 +08:00
HELSON 56ddc9ca7a
[hotfix] add correct device for fake_param (#2796) 2023-02-17 15:29:07 +08:00
ver217 a619a190df
[chatgpt] update readme about checkpoint (#2792)
* [chatgpt] add save/load checkpoint sample code

* [chatgpt] add save/load checkpoint readme

* [chatgpt] refactor save/load checkpoint readme
2023-02-17 12:43:31 +08:00
ver217 4ee311c026
[chatgpt] startegy add prepare method (#2766)
* [chatgpt] startegy add prepare method

* [chatgpt] refactor examples

* [chatgpt] refactor strategy.prepare

* [chatgpt] support save/load checkpoint

* [chatgpt] fix unwrap actor

* [chatgpt] fix unwrap actor
2023-02-17 11:27:27 +08:00
Boyuan Yao a2b43e393d
[autoparallel] Patch meta information of `torch.nn.Embedding` (#2760)
* [autoparallel] embedding metainfo

* [autoparallel] fix function name in test_activation_metainfo

* [autoparallel] undo changes in activation metainfo and related tests
2023-02-17 10:39:48 +08:00
Boyuan Yao 8e3f66a0d1
[zero] fix wrong import (#2777) 2023-02-17 10:26:07 +08:00
Fazzie-Maqianli ba84cd80b2
fix pip install colossal (#2764) 2023-02-17 09:54:21 +08:00
Nikita Shulga 01066152f1
Don't use `torch._six` (#2775)
* Don't use `torch._six`

This is a private API which is gone after https://github.com/pytorch/pytorch/pull/94709

* Update common.py
2023-02-17 09:22:45 +08:00
ver217 a88bc828d5
[chatgpt] disable shard init for colossalai (#2767) 2023-02-16 20:09:34 +08:00
binmakeswell d6d6dec190
[doc] update example and OPT serving link (#2769)
* [doc] update OPT serving link

* [doc] update example and OPT serving link

* [doc] update example and OPT serving link
2023-02-16 20:07:25 +08:00
Frank Lee e376954305
[doc] add opt service doc (#2747) 2023-02-16 15:45:26 +08:00
BlueRum 613efebc5c
[chatgpt] support colossalai strategy to train rm (#2742)
* [chatgpt]fix train_rm bug with lora

* [chatgpt]support colossalai strategy to train rm

* fix pre-commit

* fix pre-commit 2
2023-02-16 11:24:07 +08:00
BlueRum 648183a960
[chatgpt]fix train_rm bug with lora (#2741) 2023-02-16 10:25:17 +08:00
fastalgo b6e3b955c3
Update README.md 2023-02-16 07:39:46 +08:00
binmakeswell 30aee9c45d
[NFC] polish code format
[NFC] polish code format
2023-02-15 23:21:36 +08:00
YuliangLiu0306 1dc003c169
[autoparallel] distinguish different parallel strategies (#2699) 2023-02-15 22:28:28 +08:00
YH ae86a29e23
Refact method of grad store (#2687) 2023-02-15 22:27:58 +08:00
cloudhuang 43dffdaba5
[doc] fixed a typo in GPT readme (#2736) 2023-02-15 22:24:45 +08:00
binmakeswell 93b788b95a Merge branch 'main' into fix/format 2023-02-15 20:23:51 +08:00
xyupeng 2fd528b9f4
[NFC] polish colossalai/auto_parallel/tensor_shard/deprecated/graph_analysis.py code style (#2737) 2023-02-15 22:57:45 +08:00
Zirui Zhu c9e3ee389e
[NFC] polish colossalai/context/process_group_initializer/initializer_2d.py code style (#2726) 2023-02-15 22:27:13 +08:00
Zangwei Zheng 1819373e5c
[NFC] polish colossalai/auto_parallel/tensor_shard/deprecated/op_handler/batch_norm_handler.py code style (#2728) 2023-02-15 22:26:13 +08:00
Wangbo Zhao(黑色枷锁) 8331420520
[NFC] polish colossalai/cli/cli.py code style (#2734) 2023-02-15 22:25:28 +08:00
Frank Lee 5479fdd5b8
[doc] updated documentation version list (#2730) 2023-02-15 17:39:50 +08:00
binmakeswell c5be83afbf
Update version.txt (#2727) 2023-02-15 16:48:08 +08:00
ziyuhuang123 d344313533
[NFC] polish colossalai/auto_parallel/tensor_shard/deprecated/op_handler/embedding_handler.py code style (#2725) 2023-02-15 16:31:40 +08:00
Xue Fuzhao e81caeb4bc
[NFC] polish colossalai/auto_parallel/tensor_shard/deprecated/cost_graph.py code style (#2720)
Co-authored-by: Fuzhao Xue <fuzhao@login2.ls6.tacc.utexas.edu>
2023-02-15 16:12:45 +08:00
yuxuan-lou 51c45c2460
[NFC] polish colossalai/auto_parallel/tensor_shard/deprecated/op_handler/where_handler.py code style (#2723) 2023-02-15 16:12:24 +08:00
CH.Li 7aacfad8af
fix typo (#2721) 2023-02-15 14:54:53 +08:00
ver217 9c0943ecdb
[chatgpt] optimize generation kwargs (#2717)
* [chatgpt] ppo trainer use default generate args

* [chatgpt] example remove generation preparing fn

* [chatgpt] benchmark remove generation preparing fn

* [chatgpt] fix ci
2023-02-15 13:59:58 +08:00
YuliangLiu0306 21d6a48f4d
[autoparallel] add shard option (#2696)
* [autoparallel] add shard option

* polish
2023-02-15 13:48:28 +08:00
YuliangLiu0306 5b24987fa7
[autoparallel] fix parameters sharding bug (#2716) 2023-02-15 12:25:50 +08:00
Frank Lee 2045d45ab7
[doc] updated documentation version list (#2715) 2023-02-15 11:24:18 +08:00
binmakeswell d4d3387f45
[doc] add open-source contribution invitation (#2714)
* [doc] fix typo

* [doc] add invitation
2023-02-15 11:08:35 +08:00
ver217 f6b4ca4e6c
[devops] add chatgpt ci (#2713) 2023-02-15 10:53:54 +08:00
Ziyue Jiang 4603538ddd
[NFC] posh colossalai/context/process_group_initializer/initializer_sequence.py code style (#2712)
Co-authored-by: Ziyue Jiang <ziyue.jiang@gmail.com>
2023-02-15 10:53:38 +08:00
YuliangLiu0306 cb2c6a2415
[autoparallel] refactor runtime pass (#2644)
* [autoparallel] refactor runtime pass

* add unit test

* polish
2023-02-15 10:36:19 +08:00
Frank Lee 89f8975fb8
[workflow] fixed tensor-nvme build caching (#2711) 2023-02-15 10:12:55 +08:00
Zihao b3d10db5f1
[NFC] polish colossalai/cli/launcher/__init__.py code style (#2709) 2023-02-15 09:57:22 +08:00
Fazzie-Maqianli d03f4429c1
add ci (#2641) 2023-02-15 09:55:53 +08:00
YuliangLiu0306 0b2a738393
[autoparallel] remove deprecated codes (#2664) 2023-02-15 09:54:32 +08:00
YuliangLiu0306 7fa6be49d2
[autoparallel] test compatibility for gemini and auto parallel (#2700) 2023-02-15 09:43:29 +08:00
CZYCW 4ac8bfb072
[NFC] polish colossalai/engine/gradient_handler/utils.py code style (#2708) 2023-02-15 09:40:08 +08:00