* auto gptq support * more tweaks and add yml * remove old gptq docker * don't need explicit peft install for tests * fix setup.py to use extra index url install torch for tests fix cuda version for autogptq index set torch in requirements so that it installs properly move gptq install around to work with github cicd * gptq doesn't play well with sample packing * address pr feedback * remove torch install for now * set quantization_config from model config * Fix the implementation for getting quant config from model config
47 lines
1.2 KiB
Python
47 lines
1.2 KiB
Python
"""setup.py for axolotl"""
|
|
|
|
from setuptools import find_packages, setup
|
|
|
|
|
|
def parse_requirements():
|
|
_install_requires = []
|
|
_dependency_links = []
|
|
with open("./requirements.txt", encoding="utf-8") as requirements_file:
|
|
lines = [
|
|
r.strip() for r in requirements_file.readlines() if "auto-gptq" not in r
|
|
]
|
|
for line in lines:
|
|
if line.startswith("--extra-index-url"):
|
|
# Handle custom index URLs
|
|
_, url = line.split()
|
|
_dependency_links.append(url)
|
|
elif "flash-attn" not in line and line and line[0] != "#":
|
|
# Handle standard packages
|
|
_install_requires.append(line)
|
|
return _install_requires, _dependency_links
|
|
|
|
|
|
install_requires, dependency_links = parse_requirements()
|
|
|
|
|
|
setup(
|
|
name="axolotl",
|
|
version="0.1",
|
|
description="You know you're going to axolotl questions",
|
|
package_dir={"": "src"},
|
|
packages=find_packages(),
|
|
install_requires=install_requires,
|
|
dependency_links=dependency_links,
|
|
extras_require={
|
|
"gptq": [
|
|
"auto-gptq",
|
|
],
|
|
"flash-attn": [
|
|
"flash-attn==2.0.8",
|
|
],
|
|
"extras": [
|
|
"deepspeed",
|
|
],
|
|
},
|
|
)
|