From 8f740deb5323bf940724911541456a1ab0329179 Mon Sep 17 00:00:00 2001 From: YH <100389977+yhna940@users.noreply.github.com> Date: Thu, 6 Apr 2023 10:43:31 +0900 Subject: [PATCH] Fix typo (#3448) --- colossalai/tensor/param_op_hook.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/colossalai/tensor/param_op_hook.py b/colossalai/tensor/param_op_hook.py index ed705da0e..9c2e0d4ad 100644 --- a/colossalai/tensor/param_op_hook.py +++ b/colossalai/tensor/param_op_hook.py @@ -168,12 +168,12 @@ def _get_grad_args(*args): # if there is no grad tensor, the backward of PreFwdPostBwd can't be triggered arg_zero = args[0] if not isinstance(arg_zero, tuple): - raise NotImplementedError("Some torch function is incompatible because of its complcated inputs.") + raise NotImplementedError("Some torch function is incompatible because of its complicated inputs.") check_grad_flag = False for obj in arg_zero: check_grad_flag |= _is_grad_tensor(obj) if not check_grad_flag: - raise NotImplementedError("Some torch function is incompatible because of its complcated inputs.") + raise NotImplementedError("Some torch function is incompatible because of its complicated inputs.") return arg_zero, args[1:]