mirror of https://github.com/THUDM/ChatGLM-6B
Merge 79e210dad3
into 401bf3a8a7
commit
ccf041664d
|
@ -8,7 +8,7 @@
|
|||
## 软件依赖
|
||||
运行微调需要4.27.1版本的`transformers`。除 ChatGLM-6B 的依赖之外,还需要安装以下依赖
|
||||
```
|
||||
pip install rouge_chinese nltk jieba datasets
|
||||
pip install rouge_chinese nltk jieba datasets cpm_kernels sentencepiece accelerate
|
||||
```
|
||||
## 使用方法
|
||||
|
||||
|
|
|
@ -6,7 +6,7 @@ The following uses the [ADGEN](https://aclanthology.org/D19-1321.pdf) (advertisi
|
|||
## Software dependencies
|
||||
Running p-tuning requires version 4.27.1 of `transformers`. In addition to the dependencies of ChatGLM-6B, the following dependencies are required
|
||||
```
|
||||
pip install rouge_chinese nltk jieba datasets
|
||||
pip install rouge_chinese nltk jieba datasets cpm_kernels sentencepiece accelerate
|
||||
```
|
||||
## Instructions
|
||||
|
||||
|
|
|
@ -2,8 +2,9 @@
|
|||
LR=1e-4
|
||||
|
||||
MASTER_PORT=$(shuf -n 1 -i 10000-65535)
|
||||
NUM_GPUS=$(nvidia-smi --list-gpus | wc -l)
|
||||
|
||||
deepspeed --num_gpus=4 --master_port $MASTER_PORT main.py \
|
||||
deepspeed --num_gpus=$NUM_GPUS --master_port $MASTER_PORT main.py \
|
||||
--deepspeed deepspeed.json \
|
||||
--do_train \
|
||||
--train_file AdvertiseGen/train.json \
|
||||
|
|
Loading…
Reference in New Issue