Browse Source

rm torch.cuda.synchronize

pull/5258/head
yuehuayingxueluo 11 months ago committed by FrankLeeeee
parent
commit
10e3c9f923
  1. 2
      colossalai/inference/core/request_handler.py

2
colossalai/inference/core/request_handler.py

@ -198,8 +198,6 @@ class RequestHandler:
if type in config_dict and config_dict[type] is not None:
logits = logit_processor(type, logits, config_dict[type])
torch.cuda.synchronize()
# calculate probs
probs = torch.softmax(logits, dim=-1, dtype=torch.float)
logprobs = torch.log_softmax(logits, dim=-1, dtype=torch.float)

Loading…
Cancel
Save