From 8101d75ab8cff4de38ad17e4f8d12af473232df1 Mon Sep 17 00:00:00 2001 From: saber <3082548039@qq.com> Date: Sun, 26 Mar 2023 15:29:15 +0800 Subject: [PATCH] fix typo --- README.md | 7 ++++--- README_en.md | 7 ++++--- api.py | 4 ++-- cli_demo.py | 4 ++-- utils.py | 6 +++--- web_demo.py | 4 ++-- web_demo2.py | 4 ++-- 7 files changed, 19 insertions(+), 17 deletions(-) diff --git a/README.md b/README.md index 49b5797..788ed81 100644 --- a/README.md +++ b/README.md @@ -165,10 +165,11 @@ model = AutoModel.from_pretrained("your local path", trust_remote_code=True).hal ```shell pip install accelerate ``` -```python -from utils import load_mode_and_tokenizer -model, tokenizer = load_mode_and_tokenizer("your local path", num_gpus=2) +```python +from utils import load_model_and_tokenizer + +model, tokenizer = load_model_and_tokenizer("your local path", num_gpus=2) ``` 即可将模型部署到多卡上进行推理。 ## ChatGLM-6B 示例 diff --git a/README_en.md b/README_en.md index 9a4b220..9950672 100644 --- a/README_en.md +++ b/README_en.md @@ -154,10 +154,11 @@ model = AutoModel.from_pretrained("THUDM/chatglm-6b-int4", trust_remote_code=Tru ```shell pip install accelerate ``` -```python -from utils import load_mode_and_tokenizer -model, tokenizer = load_mode_and_tokenizer("your local path", num_gpus=2) +```python +from utils import load_model_and_tokenizer + +model, tokenizer = load_model_and_tokenizer("your local path", num_gpus=2) ``` ## ChatGLM-6B Examples diff --git a/api.py b/api.py index ea4d2dc..ab2066e 100644 --- a/api.py +++ b/api.py @@ -4,7 +4,7 @@ import json import uvicorn from fastapi import FastAPI, Request -from utils import load_mode_and_tokenizer +from utils import load_model_and_tokenizer app = FastAPI() @@ -34,4 +34,4 @@ async def create_item(request: Request): if __name__ == '__main__': uvicorn.run('api:app', host='0.0.0.0', port=8000, workers=1) -model, tokenizer = load_mode_and_tokenizer("THUDM/chatglm-6b", num_gpus=1) +model, tokenizer = load_model_and_tokenizer("THUDM/chatglm-6b", num_gpus=1) diff --git a/cli_demo.py b/cli_demo.py index d55fda1..fc48ca4 100644 --- a/cli_demo.py +++ b/cli_demo.py @@ -1,9 +1,9 @@ import os import platform -from utils import load_mode_and_tokenizer +from utils import load_model_and_tokenizer -model, tokenizer = load_mode_and_tokenizer("THUDM/chatglm-6b", num_gpus=1) +model, tokenizer = load_model_and_tokenizer("THUDM/chatglm-6b", num_gpus=1) os_name = platform.system() clear_command = 'cls' if os_name == 'Windows' else 'clear' diff --git a/utils.py b/utils.py index d9cf8d6..054d088 100644 --- a/utils.py +++ b/utils.py @@ -54,9 +54,9 @@ def load_model_on_gpus(checkpoint_path: Union[str, os.PathLike], return model -def load_mode_and_tokenizer(checkpoint_path: Union[str, os.PathLike], - multi_gpu_model_cache_dir: Union[str, os.PathLike] = "./temp_model_dir", - num_gpus: int = 1, **kwargs) -> Tuple[AutoModel, AutoTokenizer]: +def load_model_and_tokenizer(checkpoint_path: Union[str, os.PathLike], + multi_gpu_model_cache_dir: Union[str, os.PathLike] = "./temp_model_dir", + num_gpus: int = 1, **kwargs) -> Tuple[AutoModel, AutoTokenizer]: tokenizer = AutoTokenizer.from_pretrained(checkpoint_path, trust_remote_code=True, **kwargs) if num_gpus < 2: model = AutoModel.from_pretrained(checkpoint_path, trust_remote_code=True, **kwargs).half().cuda() diff --git a/web_demo.py b/web_demo.py index 02af0d1..0311f90 100644 --- a/web_demo.py +++ b/web_demo.py @@ -1,7 +1,7 @@ import gradio as gr -from utils import load_mode_and_tokenizer +from utils import load_model_and_tokenizer -model, tokenizer = load_mode_and_tokenizer("THUDM/chatglm-6b", num_gpus=1) +model, tokenizer = load_model_and_tokenizer("THUDM/chatglm-6b", num_gpus=1) MAX_TURNS = 20 MAX_BOXES = MAX_TURNS * 2 diff --git a/web_demo2.py b/web_demo2.py index 69a663c..82fb033 100644 --- a/web_demo2.py +++ b/web_demo2.py @@ -1,7 +1,7 @@ from transformers import AutoModel, AutoTokenizer import streamlit as st from streamlit_chat import message -from utils import load_mode_and_tokenizer +from utils import load_model_and_tokenizer st.set_page_config( page_title="ChatGLM-6b 演示", @@ -11,7 +11,7 @@ st.set_page_config( @st.cache_resource def get_model(): - model, tokenizer = load_mode_and_tokenizer("THUDM/chatglm-6b", num_gpus=1) + model, tokenizer = load_model_and_tokenizer("THUDM/chatglm-6b", num_gpus=1) return tokenizer, model