From 89dae7dc6d70f18683aa1257346a49fb9ca4a317 Mon Sep 17 00:00:00 2001 From: bursteratom Date: Fri, 6 Dec 2024 15:41:09 -0500 Subject: [PATCH] lora_target_module --- examples/qwen2-vl/lora-7b.yaml | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/examples/qwen2-vl/lora-7b.yaml b/examples/qwen2-vl/lora-7b.yaml index 765b0caef..e7ab13ddb 100644 --- a/examples/qwen2-vl/lora-7b.yaml +++ b/examples/qwen2-vl/lora-7b.yaml @@ -26,8 +26,7 @@ pad_to_sequence_len: false lora_r: 32 lora_alpha: 16 lora_dropout: 0.05 -# lora_target_linear: true -lora_target_modules: 'torch.nn.linear' +lora_target_modules: 'model.layers.[\d]+.(mlp|cross_attn|self_attn).(up|down|gate|q|k|v|o)_proj' wandb_project: wandb_entity: