From 204ca2f09afe8717134a8fb56f1946c72d7b18d6 Mon Sep 17 00:00:00 2001
From: dayellow <49357110+dayellow@users.noreply.github.com>
Date: Mon, 27 Mar 2023 18:43:43 +0800
Subject: [PATCH] [NFC] polish
 colossalai/fx/profiler/experimental/profiler_module/embedding.py code style
 (#3256)

Co-authored-by: Minghao Huang <huangminghao@luchentech.com>
---
 .../fx/profiler/experimental/profiler_module/embedding.py     | 4 +++-
 1 file changed, 3 insertions(+), 1 deletion(-)

diff --git a/colossalai/fx/profiler/experimental/profiler_module/embedding.py b/colossalai/fx/profiler/experimental/profiler_module/embedding.py
index dca6f9453..a1ade5d3a 100644
--- a/colossalai/fx/profiler/experimental/profiler_module/embedding.py
+++ b/colossalai/fx/profiler/experimental/profiler_module/embedding.py
@@ -1,5 +1,7 @@
 from typing import Tuple
+
 import torch
+
 from ..registry import meta_profiler_module
 
 
@@ -8,4 +10,4 @@ def torch_nn_embedding(self: torch.nn.Embedding, input: torch.Tensor) -> Tuple[i
     # nn.Embedding is a dictionary lookup, so technically it has 0 FLOPs. (https://discuss.pytorch.org/t/correct-way-to-calculate-flops-in-model/67198/6)
     flops = 0
     macs = 0
-    return flops, macs
\ No newline at end of file
+    return flops, macs