From 39fbd3b2b558e97c3288b551fd28633d2476898e Mon Sep 17 00:00:00 2001 From: NanoCode012 Date: Thu, 7 Aug 2025 20:25:37 +0700 Subject: [PATCH] fix: lora kernels for mistral3 (#3027) [skip ci] --- src/axolotl/monkeypatch/lora_kernels.py | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/src/axolotl/monkeypatch/lora_kernels.py b/src/axolotl/monkeypatch/lora_kernels.py index b420a965c..be1e1f2ff 100644 --- a/src/axolotl/monkeypatch/lora_kernels.py +++ b/src/axolotl/monkeypatch/lora_kernels.py @@ -156,6 +156,11 @@ def get_attention_cls_from_config(cfg: DictDefault) -> Type[nn.Module]: return Llama4TextAttention + if model_type == "mistral3": + from transformers.models.mistral.modeling_mistral import MistralAttention + + return MistralAttention + try: # Dynamically import the module and attention class module_path = f"transformers.models.{model_type}.modeling_{model_type}"