mirror of https://github.com/hpcaitech/ColossalAI
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
22 lines
632 B
22 lines
632 B
from typing import Optional
|
|
|
|
import torch
|
|
|
|
from ..registry import meta_profiler_function
|
|
|
|
|
|
@meta_profiler_function.register(torch.nn.functional.embedding)
|
|
def torch_nn_functional_embedding(
|
|
input: torch.Tensor,
|
|
weight: torch.Tensor,
|
|
padding_idx: Optional[int] = None,
|
|
max_norm: Optional[float] = None,
|
|
norm_type: float = 2.0,
|
|
scale_grad_by_freq: bool = False,
|
|
sparse: bool = False,
|
|
) -> torch.Tensor:
|
|
# F.embedding is a dictionary lookup, so technically it has 0 FLOPs. (https://discuss.pytorch.org/t/correct-way-to-calculate-flops-in-model/67198/6)
|
|
flops = 0
|
|
macs = 0
|
|
return flops, macs
|