From 0437c1a4ba8c8fdfa58708278fa9b2f2152b2730 Mon Sep 17 00:00:00 2001 From: Dan Saunders Date: Mon, 15 Sep 2025 17:06:52 -0400 Subject: [PATCH] auto-gptq -> gptqmodel --- pyproject.toml | 3 +- src/axolotl/_version.py | 33 --------- .../e2e/patched/test_lora_llama_multipack.py | 4 +- uv.lock | 71 ++++--------------- 4 files changed, 17 insertions(+), 94 deletions(-) delete mode 100644 src/axolotl/_version.py diff --git a/pyproject.toml b/pyproject.toml index 92693764b..7e10eaa6c 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -96,7 +96,7 @@ ring-flash-attn = [ "yunchang==0.6.0", ] mamba-ssm = ["mamba-ssm>=2.2.0", "causal_conv1d>=1.4.0",] -auto-gptq = ["auto-gptq==0.5.1"] +gptqmodel = ["gptqmodel>=4.0.0"] mlflow = ["mlflow"] galore = ["galore_torch"] apollo = ["apollo-torch"] @@ -119,7 +119,6 @@ dev = [ "codecov", "codecov-cli", "tbparse", - "black", "ruff", "mypy", "pre-commit", diff --git a/src/axolotl/_version.py b/src/axolotl/_version.py deleted file mode 100644 index bc67efcfc..000000000 --- a/src/axolotl/_version.py +++ /dev/null @@ -1,33 +0,0 @@ -# file generated by setuptools-scm -# don't change, don't track in version control - -__all__ = [ - "__version__", - "__version_tuple__", - "version", - "version_tuple", - "__commit_id__", - "commit_id", -] - -TYPE_CHECKING = False -if TYPE_CHECKING: - from typing import Tuple, Union - - VERSION_TUPLE = Tuple[Union[int, str], ...] - COMMIT_ID = Union[str, None] -else: - VERSION_TUPLE = object - COMMIT_ID = object - -version: str -__version__: str -__version_tuple__: VERSION_TUPLE -version_tuple: VERSION_TUPLE -commit_id: COMMIT_ID -__commit_id__: COMMIT_ID - -__version__ = version = "0.12.1.dev88+gc42ba73d4.d20250915" -__version_tuple__ = version_tuple = (0, 12, 1, "dev88", "gc42ba73d4.d20250915") - -__commit_id__ = commit_id = "gc42ba73d4" diff --git a/tests/e2e/patched/test_lora_llama_multipack.py b/tests/e2e/patched/test_lora_llama_multipack.py index 1833c750b..d4b7cf099 100644 --- a/tests/e2e/patched/test_lora_llama_multipack.py +++ b/tests/e2e/patched/test_lora_llama_multipack.py @@ -5,7 +5,7 @@ E2E tests for lora llama import unittest import pytest -from transformers.utils import is_auto_gptq_available, is_torch_bf16_gpu_available +from transformers.utils import is_gptqmodel_available, is_torch_bf16_gpu_available from axolotl.common.datasets import load_datasets from axolotl.train import train @@ -69,7 +69,7 @@ class TestLoraLlama(unittest.TestCase): train(cfg=cfg, dataset_meta=dataset_meta) check_model_output_exists(temp_dir, cfg) - @pytest.mark.skipif(not is_auto_gptq_available(), reason="auto-gptq not available") + @pytest.mark.skipif(not is_gptqmodel_available(), reason="gptqmodel not installed") @with_temp_dir def test_lora_gptq_packed(self, temp_dir): cfg = DictDefault( diff --git a/uv.lock b/uv.lock index b002bdeaa..35f085953 100644 --- a/uv.lock +++ b/uv.lock @@ -378,31 +378,6 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/77/06/bb80f5f86020c4551da315d78b3ab75e8228f89f0162f2c3a819e407941a/attrs-25.3.0-py3-none-any.whl", hash = "sha256:427318ce031701fea540783410126f03899a97ffc6f61596ad581ac2e40e3bc3", size = 63815, upload-time = "2025-03-13T11:10:21.14Z" }, ] -[[package]] -name = "auto-gptq" -version = "0.5.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "accelerate" }, - { name = "datasets" }, - { name = "gekko" }, - { name = "numpy" }, - { name = "peft" }, - { name = "rouge" }, - { name = "safetensors" }, - { name = "sentencepiece" }, - { name = "torch" }, - { name = "tqdm" }, - { name = "transformers" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/db/77/ec5a16c5625b0791dccfe5e42356171332ed3537c1df505d64a162148c8f/auto_gptq-0.5.1.tar.gz", hash = "sha256:a4897a59f8901f3c497c8d0a0422db40a79b23dbaf840fdb5de2fcf4482c8a54", size = 112336, upload-time = "2023-11-09T14:51:41.478Z" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/ba/b6/0c7fec03da337d645e407a76cdf78386d4749513e22558abc17665cdcaab/auto_gptq-0.5.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a3dad2f779e1a2798b9de7fbd75abf953c7eb23bb52108479cb3e09d3fda8b83", size = 4827634, upload-time = "2023-11-09T14:51:20.691Z" }, - { url = "https://files.pythonhosted.org/packages/a6/42/1acacd2d9436502fbfa9b578f8e9aea17142c224f007bb622eb34651c0b0/auto_gptq-0.5.1-cp310-cp310-win_amd64.whl", hash = "sha256:5f84c2b022dbf4cfa379ee2ea64fa559cd9630fa68b60bf312915330248ae9e5", size = 4618662, upload-time = "2023-11-09T14:51:23.742Z" }, - { url = "https://files.pythonhosted.org/packages/80/be/04b6cff1c334b17a6dd22a350eb54758ca4171ff235a55d6901d1d98d018/auto_gptq-0.5.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5b5c3c5f0470f3e969c3cf2fc2d5c83adf434329a1d5826d796410d625b6d7b2", size = 4830567, upload-time = "2023-11-09T14:51:26.447Z" }, - { url = "https://files.pythonhosted.org/packages/c1/7c/54f9fab72bb9c626415bdd3aed1da8d48520cf0e2068e89eafa9f2b6359c/auto_gptq-0.5.1-cp311-cp311-win_amd64.whl", hash = "sha256:7ec9592635fad541a821713cb71e73592c7fc06631369e6528e9ab2a2a9c0f03", size = 4624060, upload-time = "2023-11-09T14:51:29.402Z" }, -] - [[package]] name = "autoawq" version = "0.2.7.post3" @@ -489,9 +464,6 @@ dependencies = [ apollo = [ { name = "apollo-torch" }, ] -auto-gptq = [ - { name = "auto-gptq" }, -] dev = [ { name = "black" }, { name = "blobfile" }, @@ -518,6 +490,9 @@ flash-attn = [ galore = [ { name = "galore-torch" }, ] +gptqmodel = [ + { name = "gptqmodel" }, +] llmcompressor = [ { name = "llmcompressor" }, ] @@ -567,7 +542,6 @@ requires-dist = [ { name = "apollo-torch", marker = "extra == 'apollo'" }, { name = "apollo-torch", marker = "extra == 'optimizers'" }, { name = "art" }, - { name = "auto-gptq", marker = "extra == 'auto-gptq'", specifier = "==0.5.1" }, { name = "autoawq", marker = "sys_platform != 'darwin'", specifier = "==0.2.7.post3" }, { name = "axolotl-contribs-lgpl", git = "https://github.com/axolotl-ai-cloud/axolotl-contribs-lgpl.git?rev=numpy" }, { name = "axolotl-contribs-mit", specifier = "==0.0.5" }, @@ -591,6 +565,7 @@ requires-dist = [ { name = "galore-torch", marker = "extra == 'galore'" }, { name = "galore-torch", marker = "extra == 'optimizers'" }, { name = "gcsfs", specifier = ">=2024.5.0" }, + { name = "gptqmodel", marker = "extra == 'gptqmodel'", specifier = ">=4.0.0" }, { name = "gradio", specifier = "==5.41.1" }, { name = "hf-transfer" }, { name = "hf-xet", specifier = "==1.1.5" }, @@ -652,7 +627,7 @@ requires-dist = [ { name = "yunchang", marker = "extra == 'ring-flash-attn'", specifier = "==0.6.0" }, { name = "zstandard", specifier = ">=0.23.0" }, ] -provides-extras = ["flash-attn", "ring-flash-attn", "mamba-ssm", "auto-gptq", "mlflow", "galore", "apollo", "optimizers", "ray", "vllm", "llmcompressor", "dev"] +provides-extras = ["flash-attn", "ring-flash-attn", "mamba-ssm", "gptqmodel", "mlflow", "galore", "apollo", "optimizers", "ray", "vllm", "llmcompressor", "dev"] [package.metadata.requires-dev] dev = [ @@ -2125,18 +2100,6 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/44/dd/874223310565a336820a70727b61e7dd23f7be6cb91006f2cbb634670142/gcsfs-2025.3.0-py2.py3-none-any.whl", hash = "sha256:afbc2b26a481de66519e9cce7762340ef4781ce01c6663af0d63eda10f6d2c9c", size = 36133, upload-time = "2025-03-08T18:33:53.21Z" }, ] -[[package]] -name = "gekko" -version = "1.3.0" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "numpy" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/ec/ce/2708c403906692e8816ec430b0a95b4bec57ea8d8f3a86dbf7bd3737878e/gekko-1.3.0.tar.gz", hash = "sha256:4bfb8703550f3d7d79f593a4fef22d495599e1556ed41ca585cf7c2ec940bad2", size = 13079433, upload-time = "2025-04-03T01:41:05.047Z" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/76/b1/074200e06a7bc772b12ba11363107e7aed9eb31e383b6a610e6528d5be60/gekko-1.3.0-py3-none-any.whl", hash = "sha256:78e6e51e1e3ea2ad5476cd2e70a7f0db0acc5ffafed9459dfc1ba120f13ce3e4", size = 13169302, upload-time = "2025-04-03T01:40:59.773Z" }, -] - [[package]] name = "gguf" version = "0.17.1" @@ -2300,6 +2263,12 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/86/f1/62a193f0227cf15a920390abe675f386dec35f7ae3ffe6da582d3ade42c7/googleapis_common_protos-1.70.0-py3-none-any.whl", hash = "sha256:b8bfcca8c25a2bb253e0e0b0adaf8c00773e5e6af6fd92397576680b807e0fd8", size = 294530, upload-time = "2025-04-14T10:17:01.271Z" }, ] +[[package]] +name = "gptqmodel" +version = "4.2.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/5f/b4/87edc315d23ff0ec267c46116e30dd85bc817c6a24086812f8f2175f8010/gptqmodel-4.2.0.tar.gz", hash = "sha256:0dff183893268e15cd683a30e08acaf0b03073d55aa44f7076380ee87094f0ae", size = 329437, upload-time = "2025-09-12T09:35:46.29Z" } + [[package]] name = "gradio" version = "5.41.1" @@ -4649,7 +4618,7 @@ wheels = [ [[package]] name = "openai" -version = "1.107.2" +version = "1.107.3" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "anyio" }, @@ -4661,9 +4630,9 @@ dependencies = [ { name = "tqdm" }, { name = "typing-extensions" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/a0/66/61b0c63b68df8a22f8763d7d632ea7255edb4021dca1859f4359a5659b85/openai-1.107.2.tar.gz", hash = "sha256:a11fe8d4318e98e94309308dd3a25108dec4dfc1b606f9b1c5706e8d88bdd3cb", size = 564155, upload-time = "2025-09-12T19:52:21.159Z" } +sdist = { url = "https://files.pythonhosted.org/packages/e3/24/7fb5749bcf66b52209e3ece05cb4eaeae2102e95f8ae77589e8afaf70ba8/openai-1.107.3.tar.gz", hash = "sha256:69bb8032b05c5f00f7660e422f70f9aabc94793b9a30c5f899360ed21e46314f", size = 564194, upload-time = "2025-09-15T20:09:20.159Z" } wheels = [ - { url = "https://files.pythonhosted.org/packages/d3/65/e51a77a368eed7b9cc22ce394087ab43f13fa2884724729b716adf2da389/openai-1.107.2-py3-none-any.whl", hash = "sha256:d159d4f3ee3d9c717b248c5d69fe93d7773a80563c8b1ca8e9cad789d3cf0260", size = 946937, upload-time = "2025-09-12T19:52:19.355Z" }, + { url = "https://files.pythonhosted.org/packages/16/1d/58ad0084451f64a9193de48c0afd63047682ffdedb6ae1d494a203e03fd5/openai-1.107.3-py3-none-any.whl", hash = "sha256:4ca54a847235ac04c6320da70fdc06b62d71439de9ec0aa40d5690c3064d4025", size = 947600, upload-time = "2025-09-15T20:09:18.219Z" }, ] [[package]] @@ -6182,18 +6151,6 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/c1/02/18ba0727a1c755c528d6a52b363d62c0b7a8e64cf961b3030c046107db4d/ring_flash_attn-0.1.8-py3-none-any.whl", hash = "sha256:296c929516c3b21f7bcdaeca44a99bb541779a7b63979eb0f67837dcb18a2bb9", size = 25437, upload-time = "2025-09-10T11:53:07.565Z" }, ] -[[package]] -name = "rouge" -version = "1.0.1" -source = { registry = "https://pypi.org/simple" } -dependencies = [ - { name = "six" }, -] -sdist = { url = "https://files.pythonhosted.org/packages/db/e4/3420a1ab1e82a280fb6107f7ae99e88eb12383c978fe573c0c64d0327d6b/rouge-1.0.1.tar.gz", hash = "sha256:12b48346ca47d6bcf3c45061f315452b9ccec0620ee895ec85b7efc3d54aae34", size = 14292, upload-time = "2021-07-20T08:45:56.53Z" } -wheels = [ - { url = "https://files.pythonhosted.org/packages/32/7c/650ae86f92460e9e8ef969cc5008b24798dcf56a9a8947d04c78f550b3f5/rouge-1.0.1-py3-none-any.whl", hash = "sha256:28d118536e8c774dc47d1d15ec266479b4dd0914c4672ce117d4002789bdc644", size = 13725, upload-time = "2021-07-20T08:45:54.605Z" }, -] - [[package]] name = "rouge-score" version = "0.1.2"