diff --git a/examples/gptq-lora-7b/README.md b/examples/gptq-lora-7b/README.md deleted file mode 100644 index 0bde51b06..000000000 --- a/examples/gptq-lora-7b/README.md +++ /dev/null @@ -1,8 +0,0 @@ -# LLaMa 7B using LoRA - -This is a good place to start for beginners. This will run on an NVIDIA RTX4090 with no other changes needed. - -```shell -accelerate launch scripts/finetune.py examples/gptq-lora-7b/config.yml - -``` diff --git a/examples/gptq-lora-7b/config.yml b/examples/gptq-lora-7b/config.yml deleted file mode 100644 index d909f7d07..000000000 --- a/examples/gptq-lora-7b/config.yml +++ /dev/null @@ -1,63 +0,0 @@ -base_model: Neko-Institute-of-Science/LLaMA-7B-4bit-128g -base_model_config: Neko-Institute-of-Science/LLaMA-7B-4bit-128g -model_type: LlamaForCausalLM -tokenizer_type: LlamaTokenizer -trust_remote_code: -load_in_8bit: true -gptq: true -datasets: - - path: vicgalle/alpaca-gpt4 - type: alpaca -dataset_prepared_path: last_run_prepared -val_set_size: 0.02 -adapter: -lora_model_dir: -sequence_len: 2048 -max_packed_sequence_len: -lora_r: 8 -lora_alpha: 16 -lora_dropout: 0.05 -lora_target_modules: - - q_proj - - v_proj -lora_fan_in_fan_out: false -wandb_project: llama-7b-lora-int4 -wandb_entity: -wandb_watch: -wandb_run_id: -wandb_log_model: -output_dir: ./llama-7b-lora-int4 -gradient_accumulation_steps: 1 -micro_batch_size: 1 -num_epochs: 3 -optimizer: adamw_bnb_8bit -torchdistx_path: -lr_scheduler: cosine -learning_rate: 0.0000002 -train_on_inputs: false -group_by_length: false -fp16: true -bf16: false -tf32: true -early_stopping_patience: -resume_from_checkpoint: -local_rank: -logging_steps: 5 -xformers_attention: -flash_attention: -gradient_checkpointing: true -gptq_groupsize: 128 -gptq_model_v1: false -warmup_steps: 20 -eval_steps: 110 -save_steps: 660 -debug: -deepspeed: -weight_decay: 0.0001 -fsdp: -fsdp_config: -tokens: - pad_token: "" - bos_token: "" - eos_token: "" - unk_token: "" diff --git a/examples/gptq/lora.yml b/examples/llama-2/gptq-lora.yml similarity index 100% rename from examples/gptq/lora.yml rename to examples/llama-2/gptq-lora.yml diff --git a/src/axolotl/utils/models.py b/src/axolotl/utils/models.py index 52d25f493..7d650c0e7 100644 --- a/src/axolotl/utils/models.py +++ b/src/axolotl/utils/models.py @@ -5,7 +5,6 @@ import logging import math import os from typing import Optional, Tuple # noqa: F401 -from typing import TYPE_CHECKING, Optional, Tuple # noqa: F401 import bitsandbytes as bnb import torch