fix according to reviewer comments

pull/745/head
lvhan028 2024-06-25 19:52:50 +08:00
parent f8dbc35e0f
commit 2968b8ab24
2 changed files with 4 additions and 4 deletions

View File

@ -34,7 +34,7 @@ SWIFT supports training, inference, evaluation and deployment of LLMs and MLLMs
LMDeploy is an efficient toolkit for compressing, deploying, and serving LLMs and VLMs. LMDeploy is an efficient toolkit for compressing, deploying, and serving LLMs and VLMs.
With only 4 lines of codes, you can perform `internlm2-chat-7b` inference after `pip install lmdeploy`: With only 4 lines of code, you can perform `internlm2-chat-7b` inference after `pip install lmdeploy`:
```python ```python
from lmdeploy import pipeline from lmdeploy import pipeline
@ -141,7 +141,7 @@ llamafile lets you turn large language model (LLM) weights into executables. It
The best practice of deploying InternLM2 using llamafile is shown as below: The best practice of deploying InternLM2 using llamafile is shown as below:
- Convert the internlm2 model into GGUF model by `llama.cpp`. Suppose we get `internlm2-chat-7b.gguf` in this step - Convert the internlm2 model into GGUF model by `llama.cpp`. Suppose we get `internlm2-chat-7b.gguf` in this step
- create the llamafile - Create the llamafile
```shell ```shell
wget https://github.com/Mozilla-Ocho/llamafile/releases/download/0.8.6/llamafile-0.8.6.zip wget https://github.com/Mozilla-Ocho/llamafile/releases/download/0.8.6/llamafile-0.8.6.zip
@ -157,7 +157,7 @@ internlm2-chat-7b.gguf
999 999
..." > .args ..." > .args
zipalign -j0 \ llamafile-0.8.6/bin/zipalign -j0 \
internlm2.llamafile \ internlm2.llamafile \
internlm2-chat-7b.gguf \ internlm2-chat-7b.gguf \
.args .args

View File

@ -161,7 +161,7 @@ internlm2-chat-7b.gguf
999 999
..." > .args ..." > .args
zipalign -j0 \ llamafile-0.8.6/bin/zipalign -j0 \
internlm2.llamafile \ internlm2.llamafile \
internlm2-chat-7b.gguf \ internlm2-chat-7b.gguf \
.args .args