From 603166d9c5fd6033e3234209882c9d71283ee94c Mon Sep 17 00:00:00 2001 From: NanoCode012 Date: Thu, 7 Aug 2025 13:12:57 +0700 Subject: [PATCH] feat: add example config --- examples/glm4/glm4.5-qlora.yaml | 64 +++++++++++++++++++++++++++++++++ 1 file changed, 64 insertions(+) create mode 100644 examples/glm4/glm4.5-qlora.yaml diff --git a/examples/glm4/glm4.5-qlora.yaml b/examples/glm4/glm4.5-qlora.yaml new file mode 100644 index 000000000..d89a7f42d --- /dev/null +++ b/examples/glm4/glm4.5-qlora.yaml @@ -0,0 +1,64 @@ +base_model: zai-org/GLM-4.5 +# Automatically upload checkpoint and final model to HF +# hub_model_id: username/custom_model_name + +load_in_4bit: true + +datasets: + - path: winglian/pirate-ultrachat-10k + type: chat_template +dataset_prepared_path: last_run_prepared +val_set_size: 0 +output_dir: ./outputs/qlora-out + +adapter: qlora +lora_model_dir: + +sequence_len: 2048 +sample_packing: true +eval_sample_packing: true + + +lora_r: 16 +lora_alpha: 32 +lora_dropout: 0.05 +lora_target_modules: + - gate_proj + - down_proj + - up_proj + - q_proj + - v_proj + - k_proj + - o_proj + +wandb_project: +wandb_entity: +wandb_watch: +wandb_name: +wandb_log_model: + +gradient_accumulation_steps: 2 +micro_batch_size: 2 +num_epochs: 1 +optimizer: adamw_8bit +lr_scheduler: cosine +learning_rate: 0.0002 + +bf16: auto +tf32: false + +gradient_checkpointing: true +resume_from_checkpoint: +logging_steps: 1 +flash_attention: true + +loss_watchdog_threshold: 5.0 +loss_watchdog_patience: 3 + +warmup_ratio: 0.1 +evals_per_epoch: 1 +saves_per_epoch: 1 +weight_decay: 0.0 +special_tokens: + +# save_first_step: true # uncomment this to validate checkpoint saving works with your config