mirror of https://github.com/THUDM/ChatGLM2-6B
fix bug in chunk.json()
parent
f01259efb4
commit
0ba47403f3
|
@ -3,7 +3,7 @@
|
||||||
# Usage: python openai_api.py
|
# Usage: python openai_api.py
|
||||||
# Visit http://localhost:8000/docs for documents.
|
# Visit http://localhost:8000/docs for documents.
|
||||||
|
|
||||||
|
import json
|
||||||
import time
|
import time
|
||||||
import torch
|
import torch
|
||||||
import uvicorn
|
import uvicorn
|
||||||
|
@ -135,7 +135,7 @@ async def predict(query: str, history: List[List[str]], model_id: str):
|
||||||
finish_reason=None
|
finish_reason=None
|
||||||
)
|
)
|
||||||
chunk = ChatCompletionResponse(model=model_id, choices=[choice_data], object="chat.completion.chunk")
|
chunk = ChatCompletionResponse(model=model_id, choices=[choice_data], object="chat.completion.chunk")
|
||||||
yield "{}".format(chunk.json(exclude_unset=True, ensure_ascii=False))
|
yield json.dumps(chunk.model_dump(exclude_unset=True), ensure_ascii=False)
|
||||||
|
|
||||||
current_length = 0
|
current_length = 0
|
||||||
|
|
||||||
|
@ -152,7 +152,7 @@ async def predict(query: str, history: List[List[str]], model_id: str):
|
||||||
finish_reason=None
|
finish_reason=None
|
||||||
)
|
)
|
||||||
chunk = ChatCompletionResponse(model=model_id, choices=[choice_data], object="chat.completion.chunk")
|
chunk = ChatCompletionResponse(model=model_id, choices=[choice_data], object="chat.completion.chunk")
|
||||||
yield "{}".format(chunk.json(exclude_unset=True, ensure_ascii=False))
|
yield json.dumps(chunk.model_dump(exclude_unset=True), ensure_ascii=False)
|
||||||
|
|
||||||
|
|
||||||
choice_data = ChatCompletionResponseStreamChoice(
|
choice_data = ChatCompletionResponseStreamChoice(
|
||||||
|
@ -161,7 +161,7 @@ async def predict(query: str, history: List[List[str]], model_id: str):
|
||||||
finish_reason="stop"
|
finish_reason="stop"
|
||||||
)
|
)
|
||||||
chunk = ChatCompletionResponse(model=model_id, choices=[choice_data], object="chat.completion.chunk")
|
chunk = ChatCompletionResponse(model=model_id, choices=[choice_data], object="chat.completion.chunk")
|
||||||
yield "{}".format(chunk.json(exclude_unset=True, ensure_ascii=False))
|
yield json.dumps(chunk.model_dump(exclude_unset=True), ensure_ascii=False)
|
||||||
yield '[DONE]'
|
yield '[DONE]'
|
||||||
|
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue