try: import torch_int HAS_TORCH_INT = True except ImportError: HAS_TORCH_INT = False print("Not install torch_int. Please install torch_int from https://github.com/Guangxuan-Xiao/torch-int") if HAS_TORCH_INT: from .llama import LLamaSmoothquantAttention, LlamaSmoothquantMLP