use latest hf-xet and don't install vllm for torch 2.7.0 (#2603)

* use latest hf-xet and don't install vllm for torch 2.7.0

* fix runpod hub tests
This commit is contained in:
Wing Lian
2025-04-30 18:27:39 -04:00
committed by GitHub
parent 24ff5f53f8
commit 6ba5c0ed2c
3 changed files with 92 additions and 2 deletions

View File

@@ -30,7 +30,7 @@ jobs:
cuda_version: 12.6.3
python_version: "3.11"
pytorch: 2.7.0
axolotl_extras: vllm
axolotl_extras:
runs-on: axolotl-gpu-runner
steps:
- name: Checkout