From da626f8b232884b3848eb5a892d58937a8211cfd Mon Sep 17 00:00:00 2001 From: duzx16 Date: Wed, 12 Apr 2023 22:42:36 +0800 Subject: [PATCH] Add instruction for pre_seq_len --- ptuning/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/ptuning/README.md b/ptuning/README.md index a86db16..ab91468 100644 --- a/ptuning/README.md +++ b/ptuning/README.md @@ -155,11 +155,11 @@ for k, v in prefix_state_dict.items(): new_prefix_state_dict[k[len("transformer.prefix_encoder."):]] = v model.transformer.prefix_encoder.load_state_dict(new_prefix_state_dict) ``` +注意你可能需要将 `pre_seq_len` 改成你训练时的实际值。 (2) 如果需要加载的是旧 Checkpoint(包含 ChatGLM-6B 以及 PrefixEncoder 参数),则直接加载整个 Checkpoint: ```python -config = AutoConfig.from_pretrained(CHECKPOINT_PATH, trust_remote_code=True, pre_seq_len=128) model = AutoModel.from_pretrained(CHECKPOINT_PATH, config=config, trust_remote_code=True) ```