You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/op_builder/__init__.py

33 lines
1.0 KiB

from .cpu_adam import CPUAdamBuilder
from .fused_optim import FusedOptimBuilder
from .layernorm import LayerNormBuilder
from .moe import MOEBuilder
from .multi_head_attn import MultiHeadAttnBuilder
from .scaled_masked_softmax import ScaledMaskedSoftmaxBuilder
from .scaled_upper_triangle_masked_softmax import ScaledUpperTrainglemaskedSoftmaxBuilder
ALL_OPS = {
"cpu_adam": CPUAdamBuilder,
"fused_optim": FusedOptimBuilder,
"moe": MOEBuilder,
"multi_head_attn": MultiHeadAttnBuilder,
"scaled_masked_softmax": ScaledMaskedSoftmaxBuilder,
"scaled_upper_triangle_masked_softmax": ScaledUpperTrainglemaskedSoftmaxBuilder,
"layernorm": LayerNormBuilder,
}
__all__ = [
"ALL_OPS",
"CPUAdamBuilder",
"FusedOptimBuilder",
"MultiHeadAttnBuilder",
"ScaledMaskedSoftmaxBuilder",
"ScaledUpperTrainglemaskedSoftmaxBuilder",
"MOEBuilder",
"MultiTensorSGDBuilder",
"MultiTensorAdamBuilder",
"MultiTensorLambBuilder",
"MultiTensorScaleBuilder",
"MultiTensorL2NormBuilder",
]