auto-gptq -> gptqmodel

This commit is contained in:
Dan Saunders
2025-09-15 17:06:52 -04:00
parent ef150fd973
commit 0437c1a4ba
4 changed files with 17 additions and 94 deletions

View File

@@ -96,7 +96,7 @@ ring-flash-attn = [
"yunchang==0.6.0",
]
mamba-ssm = ["mamba-ssm>=2.2.0", "causal_conv1d>=1.4.0",]
auto-gptq = ["auto-gptq==0.5.1"]
gptqmodel = ["gptqmodel>=4.0.0"]
mlflow = ["mlflow"]
galore = ["galore_torch"]
apollo = ["apollo-torch"]
@@ -119,7 +119,6 @@ dev = [
"codecov",
"codecov-cli",
"tbparse",
"black",
"ruff",
"mypy",
"pre-commit",

View File

@@ -1,33 +0,0 @@
# file generated by setuptools-scm
# don't change, don't track in version control
__all__ = [
"__version__",
"__version_tuple__",
"version",
"version_tuple",
"__commit_id__",
"commit_id",
]
TYPE_CHECKING = False
if TYPE_CHECKING:
from typing import Tuple, Union
VERSION_TUPLE = Tuple[Union[int, str], ...]
COMMIT_ID = Union[str, None]
else:
VERSION_TUPLE = object
COMMIT_ID = object
version: str
__version__: str
__version_tuple__: VERSION_TUPLE
version_tuple: VERSION_TUPLE
commit_id: COMMIT_ID
__commit_id__: COMMIT_ID
__version__ = version = "0.12.1.dev88+gc42ba73d4.d20250915"
__version_tuple__ = version_tuple = (0, 12, 1, "dev88", "gc42ba73d4.d20250915")
__commit_id__ = commit_id = "gc42ba73d4"

View File

@@ -5,7 +5,7 @@ E2E tests for lora llama
import unittest
import pytest
from transformers.utils import is_auto_gptq_available, is_torch_bf16_gpu_available
from transformers.utils import is_gptqmodel_available, is_torch_bf16_gpu_available
from axolotl.common.datasets import load_datasets
from axolotl.train import train
@@ -69,7 +69,7 @@ class TestLoraLlama(unittest.TestCase):
train(cfg=cfg, dataset_meta=dataset_meta)
check_model_output_exists(temp_dir, cfg)
@pytest.mark.skipif(not is_auto_gptq_available(), reason="auto-gptq not available")
@pytest.mark.skipif(not is_gptqmodel_available(), reason="gptqmodel not installed")
@with_temp_dir
def test_lora_gptq_packed(self, temp_dir):
cfg = DictDefault(

71
uv.lock generated
View File

@@ -378,31 +378,6 @@ wheels = [
{ url = "https://files.pythonhosted.org/packages/77/06/bb80f5f86020c4551da315d78b3ab75e8228f89f0162f2c3a819e407941a/attrs-25.3.0-py3-none-any.whl", hash = "sha256:427318ce031701fea540783410126f03899a97ffc6f61596ad581ac2e40e3bc3", size = 63815, upload-time = "2025-03-13T11:10:21.14Z" },
]
[[package]]
name = "auto-gptq"
version = "0.5.1"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "accelerate" },
{ name = "datasets" },
{ name = "gekko" },
{ name = "numpy" },
{ name = "peft" },
{ name = "rouge" },
{ name = "safetensors" },
{ name = "sentencepiece" },
{ name = "torch" },
{ name = "tqdm" },
{ name = "transformers" },
]
sdist = { url = "https://files.pythonhosted.org/packages/db/77/ec5a16c5625b0791dccfe5e42356171332ed3537c1df505d64a162148c8f/auto_gptq-0.5.1.tar.gz", hash = "sha256:a4897a59f8901f3c497c8d0a0422db40a79b23dbaf840fdb5de2fcf4482c8a54", size = 112336, upload-time = "2023-11-09T14:51:41.478Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/ba/b6/0c7fec03da337d645e407a76cdf78386d4749513e22558abc17665cdcaab/auto_gptq-0.5.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3dad2f779e1a2798b9de7fbd75abf953c7eb23bb52108479cb3e09d3fda8b83", size = 4827634, upload-time = "2023-11-09T14:51:20.691Z" },
{ url = "https://files.pythonhosted.org/packages/a6/42/1acacd2d9436502fbfa9b578f8e9aea17142c224f007bb622eb34651c0b0/auto_gptq-0.5.1-cp310-cp310-win_amd64.whl", hash = "sha256:5f84c2b022dbf4cfa379ee2ea64fa559cd9630fa68b60bf312915330248ae9e5", size = 4618662, upload-time = "2023-11-09T14:51:23.742Z" },
{ url = "https://files.pythonhosted.org/packages/80/be/04b6cff1c334b17a6dd22a350eb54758ca4171ff235a55d6901d1d98d018/auto_gptq-0.5.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5b5c3c5f0470f3e969c3cf2fc2d5c83adf434329a1d5826d796410d625b6d7b2", size = 4830567, upload-time = "2023-11-09T14:51:26.447Z" },
{ url = "https://files.pythonhosted.org/packages/c1/7c/54f9fab72bb9c626415bdd3aed1da8d48520cf0e2068e89eafa9f2b6359c/auto_gptq-0.5.1-cp311-cp311-win_amd64.whl", hash = "sha256:7ec9592635fad541a821713cb71e73592c7fc06631369e6528e9ab2a2a9c0f03", size = 4624060, upload-time = "2023-11-09T14:51:29.402Z" },
]
[[package]]
name = "autoawq"
version = "0.2.7.post3"
@@ -489,9 +464,6 @@ dependencies = [
apollo = [
{ name = "apollo-torch" },
]
auto-gptq = [
{ name = "auto-gptq" },
]
dev = [
{ name = "black" },
{ name = "blobfile" },
@@ -518,6 +490,9 @@ flash-attn = [
galore = [
{ name = "galore-torch" },
]
gptqmodel = [
{ name = "gptqmodel" },
]
llmcompressor = [
{ name = "llmcompressor" },
]
@@ -567,7 +542,6 @@ requires-dist = [
{ name = "apollo-torch", marker = "extra == 'apollo'" },
{ name = "apollo-torch", marker = "extra == 'optimizers'" },
{ name = "art" },
{ name = "auto-gptq", marker = "extra == 'auto-gptq'", specifier = "==0.5.1" },
{ name = "autoawq", marker = "sys_platform != 'darwin'", specifier = "==0.2.7.post3" },
{ name = "axolotl-contribs-lgpl", git = "https://github.com/axolotl-ai-cloud/axolotl-contribs-lgpl.git?rev=numpy" },
{ name = "axolotl-contribs-mit", specifier = "==0.0.5" },
@@ -591,6 +565,7 @@ requires-dist = [
{ name = "galore-torch", marker = "extra == 'galore'" },
{ name = "galore-torch", marker = "extra == 'optimizers'" },
{ name = "gcsfs", specifier = ">=2024.5.0" },
{ name = "gptqmodel", marker = "extra == 'gptqmodel'", specifier = ">=4.0.0" },
{ name = "gradio", specifier = "==5.41.1" },
{ name = "hf-transfer" },
{ name = "hf-xet", specifier = "==1.1.5" },
@@ -652,7 +627,7 @@ requires-dist = [
{ name = "yunchang", marker = "extra == 'ring-flash-attn'", specifier = "==0.6.0" },
{ name = "zstandard", specifier = ">=0.23.0" },
]
provides-extras = ["flash-attn", "ring-flash-attn", "mamba-ssm", "auto-gptq", "mlflow", "galore", "apollo", "optimizers", "ray", "vllm", "llmcompressor", "dev"]
provides-extras = ["flash-attn", "ring-flash-attn", "mamba-ssm", "gptqmodel", "mlflow", "galore", "apollo", "optimizers", "ray", "vllm", "llmcompressor", "dev"]
[package.metadata.requires-dev]
dev = [
@@ -2125,18 +2100,6 @@ wheels = [
{ url = "https://files.pythonhosted.org/packages/44/dd/874223310565a336820a70727b61e7dd23f7be6cb91006f2cbb634670142/gcsfs-2025.3.0-py2.py3-none-any.whl", hash = "sha256:afbc2b26a481de66519e9cce7762340ef4781ce01c6663af0d63eda10f6d2c9c", size = 36133, upload-time = "2025-03-08T18:33:53.21Z" },
]
[[package]]
name = "gekko"
version = "1.3.0"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "numpy" },
]
sdist = { url = "https://files.pythonhosted.org/packages/ec/ce/2708c403906692e8816ec430b0a95b4bec57ea8d8f3a86dbf7bd3737878e/gekko-1.3.0.tar.gz", hash = "sha256:4bfb8703550f3d7d79f593a4fef22d495599e1556ed41ca585cf7c2ec940bad2", size = 13079433, upload-time = "2025-04-03T01:41:05.047Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/76/b1/074200e06a7bc772b12ba11363107e7aed9eb31e383b6a610e6528d5be60/gekko-1.3.0-py3-none-any.whl", hash = "sha256:78e6e51e1e3ea2ad5476cd2e70a7f0db0acc5ffafed9459dfc1ba120f13ce3e4", size = 13169302, upload-time = "2025-04-03T01:40:59.773Z" },
]
[[package]]
name = "gguf"
version = "0.17.1"
@@ -2300,6 +2263,12 @@ wheels = [
{ url = "https://files.pythonhosted.org/packages/86/f1/62a193f0227cf15a920390abe675f386dec35f7ae3ffe6da582d3ade42c7/googleapis_common_protos-1.70.0-py3-none-any.whl", hash = "sha256:b8bfcca8c25a2bb253e0e0b0adaf8c00773e5e6af6fd92397576680b807e0fd8", size = 294530, upload-time = "2025-04-14T10:17:01.271Z" },
]
[[package]]
name = "gptqmodel"
version = "4.2.0"
source = { registry = "https://pypi.org/simple" }
sdist = { url = "https://files.pythonhosted.org/packages/5f/b4/87edc315d23ff0ec267c46116e30dd85bc817c6a24086812f8f2175f8010/gptqmodel-4.2.0.tar.gz", hash = "sha256:0dff183893268e15cd683a30e08acaf0b03073d55aa44f7076380ee87094f0ae", size = 329437, upload-time = "2025-09-12T09:35:46.29Z" }
[[package]]
name = "gradio"
version = "5.41.1"
@@ -4649,7 +4618,7 @@ wheels = [
[[package]]
name = "openai"
version = "1.107.2"
version = "1.107.3"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "anyio" },
@@ -4661,9 +4630,9 @@ dependencies = [
{ name = "tqdm" },
{ name = "typing-extensions" },
]
sdist = { url = "https://files.pythonhosted.org/packages/a0/66/61b0c63b68df8a22f8763d7d632ea7255edb4021dca1859f4359a5659b85/openai-1.107.2.tar.gz", hash = "sha256:a11fe8d4318e98e94309308dd3a25108dec4dfc1b606f9b1c5706e8d88bdd3cb", size = 564155, upload-time = "2025-09-12T19:52:21.159Z" }
sdist = { url = "https://files.pythonhosted.org/packages/e3/24/7fb5749bcf66b52209e3ece05cb4eaeae2102e95f8ae77589e8afaf70ba8/openai-1.107.3.tar.gz", hash = "sha256:69bb8032b05c5f00f7660e422f70f9aabc94793b9a30c5f899360ed21e46314f", size = 564194, upload-time = "2025-09-15T20:09:20.159Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/d3/65/e51a77a368eed7b9cc22ce394087ab43f13fa2884724729b716adf2da389/openai-1.107.2-py3-none-any.whl", hash = "sha256:d159d4f3ee3d9c717b248c5d69fe93d7773a80563c8b1ca8e9cad789d3cf0260", size = 946937, upload-time = "2025-09-12T19:52:19.355Z" },
{ url = "https://files.pythonhosted.org/packages/16/1d/58ad0084451f64a9193de48c0afd63047682ffdedb6ae1d494a203e03fd5/openai-1.107.3-py3-none-any.whl", hash = "sha256:4ca54a847235ac04c6320da70fdc06b62d71439de9ec0aa40d5690c3064d4025", size = 947600, upload-time = "2025-09-15T20:09:18.219Z" },
]
[[package]]
@@ -6182,18 +6151,6 @@ wheels = [
{ url = "https://files.pythonhosted.org/packages/c1/02/18ba0727a1c755c528d6a52b363d62c0b7a8e64cf961b3030c046107db4d/ring_flash_attn-0.1.8-py3-none-any.whl", hash = "sha256:296c929516c3b21f7bcdaeca44a99bb541779a7b63979eb0f67837dcb18a2bb9", size = 25437, upload-time = "2025-09-10T11:53:07.565Z" },
]
[[package]]
name = "rouge"
version = "1.0.1"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "six" },
]
sdist = { url = "https://files.pythonhosted.org/packages/db/e4/3420a1ab1e82a280fb6107f7ae99e88eb12383c978fe573c0c64d0327d6b/rouge-1.0.1.tar.gz", hash = "sha256:12b48346ca47d6bcf3c45061f315452b9ccec0620ee895ec85b7efc3d54aae34", size = 14292, upload-time = "2021-07-20T08:45:56.53Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/32/7c/650ae86f92460e9e8ef969cc5008b24798dcf56a9a8947d04c78f550b3f5/rouge-1.0.1-py3-none-any.whl", hash = "sha256:28d118536e8c774dc47d1d15ec266479b4dd0914c4672ce117d4002789bdc644", size = 13725, upload-time = "2021-07-20T08:45:54.605Z" },
]
[[package]]
name = "rouge-score"
version = "0.1.2"