From ba62aa65ee00fe0af731855c5b3686b1b650cfb7 Mon Sep 17 00:00:00 2001 From: Matt Cummins Date: Sun, 15 Jun 2025 13:47:02 -0700 Subject: [PATCH] fixed the lora_target_modules syntax (#2793) --- examples/qwen2-vl/lora-7b.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/qwen2-vl/lora-7b.yaml b/examples/qwen2-vl/lora-7b.yaml index 55773bc3d..e8932b968 100644 --- a/examples/qwen2-vl/lora-7b.yaml +++ b/examples/qwen2-vl/lora-7b.yaml @@ -25,7 +25,7 @@ pad_to_sequence_len: false lora_r: 32 lora_alpha: 16 lora_dropout: 0.05 -lora_target_modules: 'model.layers.[\d]+.(mlp|cross_attn|self_attn).(up|down|gate|q|k|v|o)_proj' +lora_target_modules: 'model.language_model.layers.[\d]+.(mlp|cross_attn|self_attn).(up|down|gate|q|k|v|o)_proj' wandb_project: wandb_entity: