mirror of https://github.com/THUDM/ChatGLM2-6B
parent
e84a5f3c14
commit
07d463034b
|
@ -4,7 +4,7 @@ import mdtex2html
|
||||||
from utils import load_model_on_gpus
|
from utils import load_model_on_gpus
|
||||||
|
|
||||||
tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm2-6b", trust_remote_code=True)
|
tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm2-6b", trust_remote_code=True)
|
||||||
model = AutoModel.from_pretrained("THUDM/chatglm2-6b", trust_remote_code=True).cuda()
|
model = AutoModel.from_pretrained("THUDM/chatglm2-6b", trust_remote_code=True, device='cuda')#.cuda()
|
||||||
# 多显卡支持,使用下面两行代替上面一行,将num_gpus改为你实际的显卡数量
|
# 多显卡支持,使用下面两行代替上面一行,将num_gpus改为你实际的显卡数量
|
||||||
# from utils import load_model_on_gpus
|
# from utils import load_model_on_gpus
|
||||||
# model = load_model_on_gpus("THUDM/chatglm2-6b", num_gpus=2)
|
# model = load_model_on_gpus("THUDM/chatglm2-6b", num_gpus=2)
|
||||||
|
|
Loading…
Reference in New Issue