docs(config.qmd): add loraplus example (#1577)

* Update qwen2-moe-lora.yaml

* feat(project): update
This commit is contained in:
tpoisonooo
2024-05-06 13:05:28 +08:00
committed by GitHub
parent 70185763f6
commit 1ac899800b

View File

@@ -227,6 +227,12 @@ lora_modules_to_save:
lora_fan_in_fan_out: false
# LoRA+ hyperparameters
# For more details about the following options, see:
# https://arxiv.org/abs/2402.12354 and `src/axolotl/core/train_builder.py`
loraplus_lr_ratio: # loraplus learning rate ratio lr_B / lr_A. Recommended value is 2^4.
loraplus_lr_embedding: # loraplus learning rate for lora embedding layers. Default value is 1e-6.
peft:
# Configuration options for loftq initialization for LoRA
# https://huggingface.co/docs/peft/developer_guides/quantization#loftq-initialization