fix: lora kernels for mistral3 (#3027) [skip ci]

This commit is contained in:
NanoCode012
2025-08-07 20:25:37 +07:00
committed by GitHub
parent 46dfacf255
commit 39fbd3b2b5

View File

@@ -156,6 +156,11 @@ def get_attention_cls_from_config(cfg: DictDefault) -> Type[nn.Module]:
return Llama4TextAttention
if model_type == "mistral3":
from transformers.models.mistral.modeling_mistral import MistralAttention
return MistralAttention
try:
# Dynamically import the module and attention class
module_path = f"transformers.models.{model_type}.modeling_{model_type}"