fix: lora kernels for mistral3 (#3027) [skip ci]
This commit is contained in:
@@ -156,6 +156,11 @@ def get_attention_cls_from_config(cfg: DictDefault) -> Type[nn.Module]:
|
||||
|
||||
return Llama4TextAttention
|
||||
|
||||
if model_type == "mistral3":
|
||||
from transformers.models.mistral.modeling_mistral import MistralAttention
|
||||
|
||||
return MistralAttention
|
||||
|
||||
try:
|
||||
# Dynamically import the module and attention class
|
||||
module_path = f"transformers.models.{model_type}.modeling_{model_type}"
|
||||
|
||||
Reference in New Issue
Block a user