You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/examples/tutorial/auto_parallel/setup.py

14 lines
231 B

from setuptools import find_packages, setup
setup(
name="auto_parallel",
version="0.0.1",
description="",
packages=find_packages(),
install_requires=[
"torch",
"numpy",
"tqdm",
],
)