mirror of https://github.com/hpcaitech/ColossalAI
aibig-modeldata-parallelismdeep-learningdistributed-computingfoundation-modelsheterogeneous-traininghpcinferencelarge-scalemodel-parallelismpipeline-parallelism
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
37 lines
1.1 KiB
37 lines
1.1 KiB
from setuptools import find_packages, setup |
|
|
|
|
|
def fetch_requirements(path): |
|
with open(path, "r") as fd: |
|
return [r.strip() for r in fd.readlines()] |
|
|
|
|
|
def fetch_readme(): |
|
with open("README.md", encoding="utf-8") as f: |
|
return f.read() |
|
|
|
|
|
def fetch_version(): |
|
with open("version.txt", "r") as f: |
|
return f.read().strip() |
|
|
|
|
|
setup( |
|
name="colossal_llama", |
|
version=fetch_version(), |
|
packages=find_packages(exclude=("*.egg-info",)), |
|
description="Continual Pre-training and SFT for LLaMA", |
|
long_description=fetch_readme(), |
|
long_description_content_type="text/markdown", |
|
license="Apache Software License 2.0", |
|
url="https://github.com/hpcaitech/ColossalAI/tree/main/applications/Colossal-LLaMA", |
|
install_requires=fetch_requirements("requirements.txt"), |
|
python_requires=">=3.7", |
|
classifiers=[ |
|
"Programming Language :: Python :: 3", |
|
"License :: OSI Approved :: Apache Software License", |
|
"Environment :: GPU :: NVIDIA CUDA", |
|
"Topic :: Scientific/Engineering :: Artificial Intelligence", |
|
"Topic :: System :: Distributed Computing", |
|
], |
|
)
|
|
|