mirror of https://github.com/THUDM/ChatGLM-6B
Merge branch 'main' into dev
commit
1e40acce6a
|
@ -9,7 +9,9 @@ MAX_TURNS = 20
|
||||||
MAX_BOXES = MAX_TURNS * 2
|
MAX_BOXES = MAX_TURNS * 2
|
||||||
|
|
||||||
|
|
||||||
def predict(input, history=[]):
|
def predict(input, history=None):
|
||||||
|
if history is None:
|
||||||
|
history = []
|
||||||
response, history = model.chat(tokenizer, input, history)
|
response, history = model.chat(tokenizer, input, history)
|
||||||
updates = []
|
updates = []
|
||||||
for query, response in history:
|
for query, response in history:
|
||||||
|
|
Loading…
Reference in New Issue