install flash-linear-attention (#3466)

* install flash-linear-attention

* handle prequant weights for fsdp2 and ensure loss is not zero

* fix type for cu_seqlen, uninstall causal_conv1d

* chore: lint

* uv pip uninstall doesn't need confirmation
This commit is contained in:
Wing Lian
2026-03-06 12:40:57 -05:00
committed by GitHub
parent d65e1b960c
commit 876941ffd0
8 changed files with 24 additions and 4 deletions

View File

@@ -27,9 +27,16 @@ def parse_requirements(extras_require_map):
xformers_version = [req for req in _install_requires if "xformers" in req][0]
install_xformers = platform.machine() != "aarch64"
if platform.machine() == "aarch64":
# skip torchao on ARM64
# skip on ARM64
skip_packages = [
"torchao",
"fla-core",
"flash-linear-attention",
]
_install_requires = [
req for req in _install_requires if "torchao" not in req
req
for req in _install_requires
if re.split(r"[>=<]", req)[0].strip() not in skip_packages
]
if "Darwin" in platform.system():
# skip packages not compatible with OSX