mirror of https://github.com/hpcaitech/ColossalAI
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
45 lines
871 B
45 lines
871 B
import torch
|
|
|
|
__all__ = ['ALIAS_ATEN', 'INPLACE_NEW', 'INPLACE_MATH_ATEN', 'CLONE_ATEN', 'RELU_LIKE_OPS', 'RELU_LIKE_MOD']
|
|
|
|
aten = torch.ops.aten
|
|
|
|
ALIAS_ATEN = [
|
|
aten.detach.default,
|
|
aten.t.default,
|
|
aten.transpose.int,
|
|
aten.view.default,
|
|
aten._unsafe_view.default,
|
|
aten._reshape_alias.default,
|
|
]
|
|
|
|
INPLACE_NEW = [
|
|
aten.empty_like.default,
|
|
aten.new_empty_strided.default,
|
|
]
|
|
|
|
INPLACE_MATH_ATEN = [
|
|
aten.add_.Tensor,
|
|
aten.sub_.Tensor,
|
|
aten.div_.Tensor,
|
|
aten.div_.Scalar,
|
|
aten.mul_.Tensor,
|
|
aten.bernoulli_.float,
|
|
]
|
|
|
|
CLONE_ATEN = [
|
|
aten.clone.default,
|
|
]
|
|
|
|
# See illustrations in
|
|
# https://github.com/hpcaitech/ColossalAI/blob/main/colossalai/fx/profiler/constants.py
|
|
OUTPUT_SAVED_OPS = [
|
|
torch.nn.functional.relu,
|
|
torch.nn.functional.softmax,
|
|
]
|
|
|
|
OUTPUT_SAVED_MOD = [
|
|
torch.nn.ReLU,
|
|
torch.nn.Softmax,
|
|
]
|