move ring flash attn to extras with flash-attn (#2414)

This commit is contained in:
Wing Lian
2025-03-14 09:28:28 -04:00
committed by Dan Saunders
parent e1a02a32b5
commit 7a1a211c99
2 changed files with 2 additions and 10 deletions

View File

@@ -68,4 +68,3 @@ axolotl-contribs-mit==0.0.3
# for sequence parallelism
yunchang==0.6.0
ring-flash-attn>=0.1.4

View File

@@ -17,11 +17,7 @@ def parse_requirements():
lines = [r.strip() for r in requirements_file.readlines()]
for line in lines:
is_extras = (
"flash-attn" in line
or "flash-attention" in line
or "deepspeed" in line
or "mamba-ssm" in line
or "lion-pytorch" in line
"deepspeed" in line or "mamba-ssm" in line or "lion-pytorch" in line
)
if line.startswith("--extra-index-url"):
# Handle custom index URLs
@@ -39,7 +35,6 @@ def parse_requirements():
"bitsandbytes",
"triton",
"mamba-ssm",
"flash-attn",
"xformers",
"autoawq",
"liger-kernel",
@@ -124,9 +119,7 @@ setup(
],
},
extras_require={
"flash-attn": [
"flash-attn==2.7.4.post1",
],
"flash-attn": ["flash-attn==2.7.4.post1", "ring-flash-attn>=0.1.4"],
"deepspeed": [
"deepspeed==0.16.4",
"deepspeed-kernels",