From fe4d6baf9286e0eea18a3e752099a4fa16aef606 Mon Sep 17 00:00:00 2001 From: "Charles O. Goddard" Date: Sat, 26 Aug 2023 18:08:34 -0700 Subject: [PATCH] Add example Llama 2 ReLoRA config (#471) * Add example Llama 2 ReLoRA config * Use adamw_bnb_8bit in example relora config --- examples/llama-2/relora.yml | 73 +++++++++++++++++++++++++++++++++++++ 1 file changed, 73 insertions(+) create mode 100644 examples/llama-2/relora.yml diff --git a/examples/llama-2/relora.yml b/examples/llama-2/relora.yml new file mode 100644 index 000000000..66515dabc --- /dev/null +++ b/examples/llama-2/relora.yml @@ -0,0 +1,73 @@ +base_model: meta-llama/Llama-2-7b-hf +base_model_config: meta-llama/Llama-2-7b-hf +model_type: LlamaForCausalLM +tokenizer_type: LlamaTokenizer +is_llama_derived_model: true + +load_in_8bit: false +load_in_4bit: true +strict: false + +datasets: + - path: teknium/GPT4-LLM-Cleaned + type: alpaca +dataset_prepared_path: last_run_prepared +val_set_size: 0.01 +output_dir: ./relora-out + +adapter: qlora +lora_model_dir: + +sequence_len: 4096 +sample_packing: true + +lora_r: 8 +lora_alpha: 16 +lora_dropout: 0.05 +lora_target_modules: +lora_target_linear: true +lora_fan_in_fan_out: + +relora_steps: 150 +relora_warmup_steps: 10 +relora_cpu_offload: false + +wandb_project: +wandb_entity: +wandb_watch: +wandb_run_id: +wandb_log_model: + +gradient_accumulation_steps: 4 +micro_batch_size: 4 +num_epochs: 3 +optimizer: adamw_bnb_8bit +lr_scheduler: cosine +learning_rate: 0.0002 + +train_on_inputs: false +group_by_length: false +bf16: true +fp16: false +tf32: false + +gradient_checkpointing: true +early_stopping_patience: +resume_from_checkpoint: +local_rank: +logging_steps: 1 +xformers_attention: +flash_attention: true + +warmup_steps: 10 +eval_steps: 20 +save_steps: 50 +debug: +deepspeed: +weight_decay: 0.0 +fsdp: +fsdp_config: +special_tokens: + bos_token: "" + eos_token: "" + unk_token: ""