From 204ca2f09afe8717134a8fb56f1946c72d7b18d6 Mon Sep 17 00:00:00 2001 From: dayellow <49357110+dayellow@users.noreply.github.com> Date: Mon, 27 Mar 2023 18:43:43 +0800 Subject: [PATCH] [NFC] polish colossalai/fx/profiler/experimental/profiler_module/embedding.py code style (#3256) Co-authored-by: Minghao Huang <huangminghao@luchentech.com> --- .../fx/profiler/experimental/profiler_module/embedding.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/colossalai/fx/profiler/experimental/profiler_module/embedding.py b/colossalai/fx/profiler/experimental/profiler_module/embedding.py index dca6f9453..a1ade5d3a 100644 --- a/colossalai/fx/profiler/experimental/profiler_module/embedding.py +++ b/colossalai/fx/profiler/experimental/profiler_module/embedding.py @@ -1,5 +1,7 @@ from typing import Tuple + import torch + from ..registry import meta_profiler_module @@ -8,4 +10,4 @@ def torch_nn_embedding(self: torch.nn.Embedding, input: torch.Tensor) -> Tuple[i # nn.Embedding is a dictionary lookup, so technically it has 0 FLOPs. (https://discuss.pytorch.org/t/correct-way-to-calculate-flops-in-model/67198/6) flops = 0 macs = 0 - return flops, macs \ No newline at end of file + return flops, macs