mirror of https://github.com/THUDM/ChatGLM-6B
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
65 lines
4.3 KiB
65 lines
4.3 KiB
2 years ago
|
## 更新信息
|
||
|
|
||
|
**[2023/0515]** 更新 v1.1 版本 checkpoint,训练数据增加英文数据以平衡中英文数据比例,解决英文回答中夹杂中文词语的现象。
|
||
|
|
||
|
<details><summary><b>以下是更新前后的英文问题对比:</b></summary>
|
||
|
|
||
|
* 问题:Describe a time when you had to make a difficult decision.
|
||
|
- v1.0:
|
||
|
![](resources/english-q1-old.png)
|
||
|
- v1.1:
|
||
|
![](resources/english-q1-new.png)
|
||
|
* 问题:Describe the function of a computer motherboard
|
||
|
- v1.0:
|
||
|
![](resources/english-q2-old.png)
|
||
|
- v1.1:
|
||
|
![](resources/english-q2-new.png)
|
||
|
* 问题:Develop a plan to reduce electricity usage in a home.
|
||
|
- v1.0:
|
||
|
![](resources/english-q3-old.png)
|
||
|
- v1.1:
|
||
|
![](resources/english-q3-new.png)
|
||
|
</details>
|
||
|
|
||
|
**[2023/04/16]** 增加 INT8 量化后的模型 [ChatGLM-6B-INT8](https://huggingface.co/THUDM/chatglm-6b-int8)。增加多卡部署(感谢 [@Cherrysaber](https://github.com/Cherrysaber))。
|
||
|
|
||
|
**[2023/04/06]** 优化web demo的界面(感谢 [@tuteng0915](https://github.com/tuteng0915))。移除embedding中的image token以减小显存占用(需要更新模型文件`pytorch_model-00001-of-00008.bin`和`pytorch_model-00008-of-00008.bin`,感谢 [@silverriver](https://github.com/silverriver) 提出的想法)。去掉了对 `icetk` 的依赖(需要更新模型文件`ice_text.model`)。
|
||
|
|
||
|
**[2023/03/31]** 增加基于 [P-Tuning-v2](https://github.com/THUDM/P-tuning-v2) 的高效参数微调实现,INT4 量化级别下最低只需 7GB 显存即可进行模型微调。详见[高效参数微调方法](ptuning/README.md)。
|
||
|
|
||
|
**[2023/03/23]** 增加 API 部署(感谢 [@LemonQu-GIT](https://github.com/LemonQu-GIT))。~~增加 Embedding 量化模型 [ChatGLM-6B-INT4-QE](https://huggingface.co/THUDM/chatglm-6b-int4-qe)~~ (已停止维护)。增加配备 Apple Silicon 芯片的 Mac 上 GPU 加速的支持。
|
||
|
|
||
|
**[2023/03/19]** 增加流式输出接口 `stream_chat`,已更新到网页版和命令行 Demo。修复输出中的中文标点。增加 INT4 量化后的模型 [ChatGLM-6B-INT4](https://huggingface.co/THUDM/chatglm-6b-int4)
|
||
|
|
||
|
|
||
|
## Update
|
||
|
**[2023/0515]** Update the checkpoint of v1.1 version, add English instruction data for training to balance the proportion of Chinese and English data, which solves the phenomenon of Chinese words mixed in English answers .
|
||
|
|
||
|
<details><summary><b>The following is a comparison of English questions before and after the update</b></summary>
|
||
|
|
||
|
* Question: Describe a time when you had to make a difficult decision.
|
||
|
- v1.0:
|
||
|
![](resources/english-q1-old.png)
|
||
|
- v1.1:
|
||
|
![](resources/english-q1-new.png)
|
||
|
* Question: Describe the function of a computer motherboard
|
||
|
- v1.0:
|
||
|
![](resources/english-q2-old.png)
|
||
|
- v1.1:
|
||
|
![](resources/english-q2-new.png)
|
||
|
* Question: Develop a plan to reduce electricity usage in a home.
|
||
|
- v1.0:
|
||
|
![](resources/english-q3-old.png)
|
||
|
- v1.1:
|
||
|
![](resources/english-q3-new.png)
|
||
|
</details>
|
||
|
|
||
|
**[2023/04/16]** Added INT8 quantized model [ChatGLM-6B-INT8](https://huggingface.co/THUDM/chatglm-6b-int8). Added multi-GPU deployment (thanks to [@Cherrysaber](https://github.com/Cherrysaber)).
|
||
|
|
||
|
**[2023/04/06]** Improve the web demo interface (thanks to [@tuteng0915](https://github.com/tuteng0915)). Remove the image tokens in the embedding layer to reduce the memory usage (need to update the model files `pytorch_model-00001-of-00008.bin` and `pytorch_model-00008-of-00008.bin`, thanks to [@silverriver](https:/ /github.com/silverriver) for proposing the idea). Removed dependency on `icetk` (need to update model file `ice_text.model`).
|
||
|
|
||
|
**[2023/03/31]** Added a parameter-efficient tuning implementation based on [P-Tuning-v2](https://github.com/THUDM/P-tuning-v2). The minimum INT4 quantization level only needs 7GB GPU memory is enough for model tuning. See [Parameter-efficient tuning method](ptuning/README.md) for details.
|
||
|
|
||
|
**[2023/03/23]** Add API deployment, thanks to [@LemonQu-GIT](https://github.com/LemonQu-GIT). Add embedding-quantized model [ChatGLM-6B-INT4-QE](https://huggingface.co/THUDM/chatglm-6b-int4-qe). Add support for GPU inference on Mac with Apple Silicon.
|
||
|
|
||
|
**[2023/03/19]** Add streaming output function `stream_chat`, already applied in web and CLI demo. Fix Chinese punctuations in output. Add quantized model [ChatGLM-6B-INT4](https://huggingface.co/THUDM/chatglm-6b-int4).
|