mirror of https://github.com/hpcaitech/ColossalAI
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
185 lines
7.1 KiB
185 lines
7.1 KiB
import importlib
|
|
import os
|
|
import time
|
|
from abc import ABC, abstractmethod
|
|
from pathlib import Path
|
|
from typing import List
|
|
|
|
|
|
def print_rank_0(message):
|
|
"""
|
|
Print on only one process to avoid spamming.
|
|
"""
|
|
try:
|
|
import torch.distributed as dist
|
|
if not dist.is_initialized():
|
|
is_main_rank = True
|
|
else:
|
|
is_main_rank = dist.get_rank() == 0
|
|
except ImportError:
|
|
is_main_rank = True
|
|
|
|
if is_main_rank:
|
|
print(message)
|
|
|
|
|
|
class Builder(ABC):
|
|
"""
|
|
Builder is the base class to build extensions for PyTorch.
|
|
|
|
Args:
|
|
name (str): the name of the kernel to be built
|
|
prebuilt_import_path (str): the path where the extension is installed during pip install
|
|
"""
|
|
|
|
def __init__(self, name: str, prebuilt_import_path: str):
|
|
self.name = name
|
|
self.prebuilt_import_path = prebuilt_import_path
|
|
self.version_dependent_macros = ['-DVERSION_GE_1_1', '-DVERSION_GE_1_3', '-DVERSION_GE_1_5']
|
|
|
|
assert prebuilt_import_path.startswith('colossalai._C'), \
|
|
f'The prebuilt_import_path should start with colossalai._C, but got {self.prebuilt_import_path}'
|
|
|
|
def relative_to_abs_path(self, code_path: str) -> str:
|
|
"""
|
|
This function takes in a path relative to the colossalai root directory and return the absolute path.
|
|
"""
|
|
op_builder_module_path = Path(__file__).parent
|
|
|
|
# if we install from source
|
|
# the current file path will be op_builder/builder.py
|
|
# if we install via pip install colossalai
|
|
# the current file path will be colossalai/kernel/op_builder/builder.py
|
|
# this is because that the op_builder inside colossalai is a symlink
|
|
# this symlink will be replaced with actual files if we install via pypi
|
|
# thus we cannot tell the colossalai root directory by checking whether the op_builder
|
|
# is a symlink, we can only tell whether it is inside or outside colossalai
|
|
if str(op_builder_module_path).endswith('colossalai/kernel/op_builder'):
|
|
root_path = op_builder_module_path.parent.parent
|
|
else:
|
|
root_path = op_builder_module_path.parent.joinpath('colossalai')
|
|
|
|
code_abs_path = root_path.joinpath(code_path)
|
|
return str(code_abs_path)
|
|
|
|
def get_cuda_home_include(self):
|
|
"""
|
|
return include path inside the cuda home.
|
|
"""
|
|
from torch.utils.cpp_extension import CUDA_HOME
|
|
if CUDA_HOME is None:
|
|
raise RuntimeError("CUDA_HOME is None, please set CUDA_HOME to compile C++/CUDA kernels in ColossalAI.")
|
|
cuda_include = os.path.join(CUDA_HOME, "include")
|
|
return cuda_include
|
|
|
|
def csrc_abs_path(self, path):
|
|
return os.path.join(self.relative_to_abs_path('kernel/cuda_native/csrc'), path)
|
|
|
|
# functions must be overrided begin
|
|
@abstractmethod
|
|
def sources_files(self) -> List[str]:
|
|
"""
|
|
This function should return a list of source files for extensions.
|
|
"""
|
|
raise NotImplementedError
|
|
|
|
@abstractmethod
|
|
def include_dirs(self) -> List[str]:
|
|
"""
|
|
This function should return a list of inlcude files for extensions.
|
|
"""
|
|
pass
|
|
|
|
@abstractmethod
|
|
def cxx_flags(self) -> List[str]:
|
|
"""
|
|
This function should return a list of cxx compilation flags for extensions.
|
|
"""
|
|
pass
|
|
|
|
@abstractmethod
|
|
def nvcc_flags(self) -> List[str]:
|
|
"""
|
|
This function should return a list of nvcc compilation flags for extensions.
|
|
"""
|
|
pass
|
|
|
|
# functions must be overrided over
|
|
def strip_empty_entries(self, args):
|
|
'''
|
|
Drop any empty strings from the list of compile and link flags
|
|
'''
|
|
return [x for x in args if len(x) > 0]
|
|
|
|
def import_op(self):
|
|
"""
|
|
This function will import the op module by its string name.
|
|
"""
|
|
return importlib.import_module(self.prebuilt_import_path)
|
|
|
|
def load(self, verbose=True):
|
|
"""
|
|
load the kernel during runtime. If the kernel is not built during pip install, it will build the kernel.
|
|
If the kernel is built during runtime, it will be stored in `~/.cache/colossalai/torch_extensions/`. If the
|
|
kernel is built during pip install, it can be accessed through `colossalai._C`.
|
|
|
|
Warning: do not load this kernel repeatedly during model execution as it could slow down the training process.
|
|
|
|
Args:
|
|
verbose (bool, optional): show detailed info. Defaults to True.
|
|
"""
|
|
from torch.utils.cpp_extension import load
|
|
start_build = time.time()
|
|
|
|
try:
|
|
op_module = self.import_op()
|
|
if verbose:
|
|
print_rank_0(f"OP {self.prebuilt_import_path} already exists, skip building.")
|
|
except ImportError:
|
|
# construct the build directory
|
|
import torch
|
|
torch_version_major = torch.__version__.split('.')[0]
|
|
torch_version_minor = torch.__version__.split('.')[1]
|
|
torch_cuda_version = torch.version.cuda
|
|
home_directory = os.path.expanduser('~')
|
|
extension_directory = f".cache/colossalai/torch_extensions/torch{torch_version_major}.{torch_version_minor}_cu{torch_cuda_version}"
|
|
build_directory = os.path.join(home_directory, extension_directory)
|
|
Path(build_directory).mkdir(parents=True, exist_ok=True)
|
|
|
|
if verbose:
|
|
print_rank_0(
|
|
"=========================================================================================")
|
|
print_rank_0(f"No pre-built kernel is found, build and load the {self.name} kernel during runtime now")
|
|
print_rank_0(
|
|
"=========================================================================================")
|
|
|
|
# load the kernel
|
|
op_module = load(name=self.name,
|
|
sources=self.strip_empty_entries(self.sources_files()),
|
|
extra_include_paths=self.strip_empty_entries(self.include_dirs()),
|
|
extra_cflags=self.cxx_flags(),
|
|
extra_cuda_cflags=self.nvcc_flags(),
|
|
extra_ldflags=[],
|
|
build_directory=build_directory,
|
|
verbose=verbose)
|
|
|
|
build_duration = time.time() - start_build
|
|
if verbose:
|
|
print_rank_0(f"Time to load {self.name} op: {build_duration} seconds")
|
|
|
|
return op_module
|
|
|
|
def builder(self) -> 'CUDAExtension':
|
|
"""
|
|
get a CUDAExtension instance used for setup.py
|
|
"""
|
|
from torch.utils.cpp_extension import CUDAExtension
|
|
|
|
return CUDAExtension(name=self.prebuilt_import_path,
|
|
sources=self.strip_empty_entries(self.sources_files()),
|
|
include_dirs=self.strip_empty_entries(self.include_dirs()),
|
|
extra_compile_args={
|
|
'cxx': self.strip_empty_entries(self.cxx_flags()),
|
|
'nvcc': self.strip_empty_entries(self.nvcc_flags())
|
|
})
|