import os import torch import re from .builder import Builder from .utils import append_nvcc_threads, get_cuda_cc_flag class GPTQBuilder(Builder): NAME = "cu_gptq" PREBUILT_IMPORT_PATH = "colossalai._C.cu_gptq" def __init__(self): super().__init__(name=GPTQBuilder.NAME, prebuilt_import_path=GPTQBuilder.PREBUILT_IMPORT_PATH) def include_dirs(self): ret = [self.csrc_abs_path("gptq"), self.get_cuda_home_include()] return ret def sources_files(self): ret = [ self.csrc_abs_path(fname) for fname in [ 'gptq/linear_gptq.cpp', 'gptq/column_remap.cu', 'gptq/cuda_buffers.cu', 'gptq/q4_matmul.cu', 'gptq/q4_matrix.cu' ] ] return ret def cxx_flags(self): return ['-O3'] + self.version_dependent_macros def nvcc_flags(self): extra_cuda_flags = ['-v', '-std=c++14', '-U__CUDA_NO_HALF_OPERATORS__', '-U__CUDA_NO_HALF_CONVERSIONS__', '-U__CUDA_NO_HALF2_OPERATORS__', '-DTHRUST_IGNORE_CUB_VERSION_CHECK', "-lcublas", "-std=c++17" ] for arch in torch.cuda.get_arch_list(): res = re.search(r'sm_(\d+)', arch) if res: arch_cap = res[1] if int(arch_cap) >= 80: extra_cuda_flags.extend(['-gencode', f'arch=compute_{arch_cap},code={arch}']) ret = ['-O3', '--use_fast_math'] + self.version_dependent_macros + extra_cuda_flags return append_nvcc_threads(ret)