install flash-linear-attention (#3466)

* install flash-linear-attention

* handle prequant weights for fsdp2 and ensure loss is not zero

* fix type for cu_seqlen, uninstall causal_conv1d

* chore: lint

* uv pip uninstall doesn't need confirmation
This commit is contained in:
Wing Lian
2026-03-06 12:40:57 -05:00
committed by GitHub
parent d65e1b960c
commit 876941ffd0
8 changed files with 24 additions and 4 deletions

View File

@@ -506,8 +506,11 @@ def patch_initialize_missing_keys_for_fsdp():
def _patched_initialize_missing_keys(self, is_quantized: bool) -> None:
if is_fsdp_enabled() and not is_local_dist_rank_0():
for key in self.state_dict():
param_or_buffer = self.get_parameter_or_buffer(key)
param_or_buffer._is_hf_initialized = True
try:
param_or_buffer = self.get_parameter_or_buffer(key)
param_or_buffer._is_hf_initialized = True
except AttributeError:
pass # may happen when handling pre-quantized weights
self._is_hf_initialized = True
_original_initialize_missing_keys(self, is_quantized)