mirror of https://github.com/InternLM/InternLM
fixed the issue that the HF model spontaneously conducted multiple rounds of Q&A and stream_chat method generates garbled characters
Signed-off-by: daijun1 <daijun1@eccom.com.cn>pull/560/head
parent
ac7509389b
commit
97e7d03d09
|
@ -844,6 +844,7 @@ class InternLMForCausalLM(InternLMPreTrainedModel):
|
||||||
self.query = query
|
self.query = query
|
||||||
self.history = history
|
self.history = history
|
||||||
self.response = ""
|
self.response = ""
|
||||||
|
self.chche = []
|
||||||
self.received_inputs = False
|
self.received_inputs = False
|
||||||
self.queue.put((self.response, history + [(self.query, self.response)]))
|
self.queue.put((self.response, history + [(self.query, self.response)]))
|
||||||
|
|
||||||
|
@ -858,11 +859,18 @@ class InternLMForCausalLM(InternLMPreTrainedModel):
|
||||||
self.received_inputs = True
|
self.received_inputs = True
|
||||||
return
|
return
|
||||||
|
|
||||||
token = self.tokenizer.decode([value[-1]], skip_special_tokens=True)
|
self.chche.extend(value.tolist())
|
||||||
|
token = self.tokenizer.decode(self.chche, skip_special_tokens=True)
|
||||||
|
if " " in token and len(token) <= 5:
|
||||||
|
return
|
||||||
|
|
||||||
if token.strip() != "<eoa>":
|
if token.strip() != "<eoa>":
|
||||||
self.response = self.response + token
|
self.response = self.response + token
|
||||||
history = self.history + [(self.query, self.response)]
|
history = self.history + [(self.query, self.response)]
|
||||||
self.queue.put((self.response, history))
|
self.queue.put((self.response, history))
|
||||||
|
self.chche = []
|
||||||
|
else:
|
||||||
|
self.end()
|
||||||
|
|
||||||
def end(self):
|
def end(self):
|
||||||
self.queue.put(None)
|
self.queue.put(None)
|
||||||
|
|
Loading…
Reference in New Issue