From 10e3c9f923caf4fb68ab61e96c244bd5cca9b9da Mon Sep 17 00:00:00 2001 From: yuehuayingxueluo <867460659@qq.com> Date: Tue, 9 Jan 2024 15:53:04 +0800 Subject: [PATCH] rm torch.cuda.synchronize --- colossalai/inference/core/request_handler.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/colossalai/inference/core/request_handler.py b/colossalai/inference/core/request_handler.py index a83e5041d..dd8591e7f 100644 --- a/colossalai/inference/core/request_handler.py +++ b/colossalai/inference/core/request_handler.py @@ -198,8 +198,6 @@ class RequestHandler: if type in config_dict and config_dict[type] is not None: logits = logit_processor(type, logits, config_dict[type]) - torch.cuda.synchronize() - # calculate probs probs = torch.softmax(logits, dim=-1, dtype=torch.float) logprobs = torch.log_softmax(logits, dim=-1, dtype=torch.float)