Tong Li
dacc04ef75
fix bug
2024-11-15 03:55:55 +00:00
Tong Li
375e356a16
update prm
2024-11-15 03:18:41 +00:00
Tong Li
a8b4afb747
add prm
2024-11-14 09:06:59 +00:00
pre-commit-ci[bot]
ab992b89e4
[pre-commit.ci] auto fixes from pre-commit.com hooks
...
for more information, see https://pre-commit.ci
2024-11-14 08:59:32 +00:00
Tong Li
852333423d
update prm
2024-11-14 08:58:43 +00:00
Tong Li
f2f5ff5e24
update init
2024-11-14 08:56:24 +00:00
Tong Li
5719974a1e
Merge branch 'feat/prm' of github.com:TongLi3701/ColossalAI into feat/prm
2024-11-14 08:53:56 +00:00
Tong Li
797a81a8e2
add loss
2024-11-14 08:53:26 +00:00
pre-commit-ci[bot]
38a7f3846d
[pre-commit.ci] auto fixes from pre-commit.com hooks
...
for more information, see https://pre-commit.ci
2024-11-14 08:49:37 +00:00
Tong Li
9995119c28
update init
2024-11-14 08:48:08 +00:00
Tong Li
9ff9dc3d4a
Merge branch 'feat/prm' of github.com:TongLi3701/ColossalAI into feat/prm
2024-11-14 08:31:43 +00:00
Tong Li
c606d1101c
add tokenize func
2024-11-14 08:31:33 +00:00
Tong Li
b6ec337f3d
update tokenize function
2024-11-14 08:28:02 +00:00
pre-commit-ci[bot]
0bfb0d32a8
[pre-commit.ci] auto fixes from pre-commit.com hooks
...
for more information, see https://pre-commit.ci
2024-11-11 11:39:16 +00:00
Tong Li
794e0d4f4a
update conversation
2024-11-11 11:38:14 +00:00
Tong Li
73ebbef3a3
add prm dataset example
2024-11-11 10:17:31 +00:00
Tong Li
1210dbea97
update tokenization function
2024-11-11 07:26:32 +00:00
Tong Li
dcb509c8e3
Merge branch 'feat/prm' of github.com:TongLi3701/ColossalAI into feat/prm
2024-11-08 12:43:07 +00:00
Tong Li
ed817a29f9
update
2024-11-08 12:42:48 +00:00
pre-commit-ci[bot]
308960534f
[pre-commit.ci] auto fixes from pre-commit.com hooks
...
for more information, see https://pre-commit.ci
2024-11-08 12:38:38 +00:00
Tong Li
6c619c9992
update best answer function
2024-11-08 03:30:21 +00:00
Tong Li
30a9443132
[Coati] Refine prompt for better inference ( #6117 )
...
* refine prompt
* update prompt
* [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
---------
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2024-11-08 11:00:37 +08:00
Tong Li
7a60161035
update readme ( #6116 )
2024-11-06 17:24:08 +08:00
Hongxin Liu
a15ab139ad
[plugin] support get_grad_norm ( #6115 )
2024-11-05 18:12:47 +08:00
Hongxin Liu
13ffa08cfa
[release] update version ( #6109 )
2024-11-04 17:26:28 +08:00
pre-commit-ci[bot]
2f583c1549
[pre-commit.ci] pre-commit autoupdate ( #6078 )
...
updates:
- [github.com/psf/black-pre-commit-mirror: 24.8.0 → 24.10.0](https://github.com/psf/black-pre-commit-mirror/compare/24.8.0...24.10.0 )
- [github.com/pre-commit/mirrors-clang-format: v18.1.8 → v19.1.2](https://github.com/pre-commit/mirrors-clang-format/compare/v18.1.8...v19.1.2 )
- [github.com/pre-commit/pre-commit-hooks: v4.6.0 → v5.0.0](https://github.com/pre-commit/pre-commit-hooks/compare/v4.6.0...v5.0.0 )
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2024-10-31 18:18:01 +08:00
Hongxin Liu
c2e8f61592
[checkpointio] fix hybrid plugin model save ( #6106 )
2024-10-31 17:04:53 +08:00
Tong Li
89a9a600bc
[MCTS] Add self-refined MCTS ( #6098 )
...
* add reasoner
* [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
* update code
* delete llama
* update prompts
* update readme
* update readme
---------
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2024-10-24 17:51:19 +08:00
binmakeswell
4294ae83bb
[doc] sora solution news ( #6100 )
...
* [doc] sora solution news
* [doc] sora solution news
2024-10-24 13:24:37 +08:00
Hongxin Liu
80a8ca916a
[extension] hotfix compile check ( #6099 )
2024-10-24 11:11:44 +08:00
Hanks
dee63cc5ef
Merge pull request #6096 from BurkeHulk/hotfix/lora_ckpt
...
[hotfix] fix lora ckpt saving format
2024-10-21 14:13:04 +08:00
BurkeHulk
6d6cafabe2
pre-commit fix
2024-10-21 14:04:32 +08:00
BurkeHulk
b10339df7c
fix lora ckpt save format (ColoTensor to Tensor)
2024-10-21 13:55:43 +08:00
Hongxin Liu
19baab5fd5
[release] update version ( #6094 )
2024-10-21 10:19:08 +08:00
Hongxin Liu
58d8b8a2dd
[misc] fit torch api upgradation and remove legecy import ( #6093 )
...
* [amp] fit torch's new api
* [amp] fix api call
* [amp] fix api call
* [misc] fit torch pytree api upgrade
* [misc] remove legacy import
* [misc] fit torch amp api
* [misc] fit torch amp api
2024-10-18 16:48:52 +08:00
Hongxin Liu
5ddad486ca
[fp8] add fallback and make compile option configurable ( #6092 )
2024-10-18 13:55:31 +08:00
botbw
3b1d7d1ae8
[chore] refactor
2024-10-17 11:04:47 +08:00
botbw
2bcd0b6844
[ckpt] add safetensors util
2024-10-17 11:04:47 +08:00
Hongxin Liu
cd61353bae
[pipeline] hotfix backward for multiple outputs ( #6090 )
...
* [pipeline] hotfix backward for multiple outputs
* [pipeline] hotfix backward for multiple outputs
2024-10-16 17:27:33 +08:00
Wenxuan Tan
62c13e7969
[Ring Attention] Improve comments ( #6085 )
...
* improve comments
* improve comments
---------
Co-authored-by: Edenzzzz <wtan45@wisc.edu>
2024-10-16 11:23:35 +08:00
Wang Binluo
dcd41d0973
Merge pull request #6071 from wangbluo/ring_attention
...
[Ring Attention] fix the 2d ring attn when using multiple machine
2024-10-15 15:17:21 +08:00
wangbluo
83cf2f84fb
fix
2024-10-15 14:50:27 +08:00
wangbluo
bc7eeade33
fix
2024-10-15 13:28:33 +08:00
wangbluo
fd92789af2
fix
2024-10-15 13:26:44 +08:00
wangbluo
6be9862aaf
fix
2024-10-15 11:56:49 +08:00
wangbluo
3dc08c8a5a
fix
2024-10-15 11:01:34 +08:00
wangbluo
8ff7d0c780
fix
2024-10-14 18:16:03 +08:00
wangbluo
fe9208feac
fix
2024-10-14 18:07:56 +08:00
wangbluo
3201377e94
fix
2024-10-14 18:06:24 +08:00
wangbluo
23199e34cc
fix
2024-10-14 18:01:53 +08:00