install torch for tests fix cuda version for autogptq index set torch in requirements so that it installs properly move gptq install around to work with github cicd
47 lines
1.2 KiB
Python
47 lines
1.2 KiB
Python
"""setup.py for axolotl"""
|
|
|
|
from setuptools import find_packages, setup
|
|
|
|
|
|
def parse_requirements():
|
|
_install_requires = []
|
|
_dependency_links = []
|
|
with open("./requirements.txt", encoding="utf-8") as requirements_file:
|
|
lines = [
|
|
r.strip() for r in requirements_file.readlines() if "auto-gptq" not in r
|
|
]
|
|
for line in lines:
|
|
if line.startswith("--extra-index-url"):
|
|
# Handle custom index URLs
|
|
_, url = line.split()
|
|
_dependency_links.append(url)
|
|
elif "flash-attn" not in line and line and line[0] != "#":
|
|
# Handle standard packages
|
|
_install_requires.append(line)
|
|
return _install_requires, _dependency_links
|
|
|
|
|
|
install_requires, dependency_links = parse_requirements()
|
|
|
|
|
|
setup(
|
|
name="axolotl",
|
|
version="0.1",
|
|
description="You know you're going to axolotl questions",
|
|
package_dir={"": "src"},
|
|
packages=find_packages(),
|
|
install_requires=install_requires,
|
|
dependency_links=dependency_links,
|
|
extras_require={
|
|
"gptq": [
|
|
"auto-gptq",
|
|
],
|
|
"flash-attn": [
|
|
"flash-attn==2.0.8",
|
|
],
|
|
"extras": [
|
|
"deepspeed",
|
|
],
|
|
},
|
|
)
|