diff --git a/README.md b/README.md index 8ba63299e..7ebea8678 100644 --- a/README.md +++ b/README.md @@ -39,10 +39,10 @@ pip3 install -U git+https://github.com/huggingface/peft.git accelerate config # finetune lora -accelerate launch scripts/finetune.py examples/lora-openllama-3b/config.yml +accelerate launch scripts/finetune.py examples/openllama-3b/lora.yml # inference -accelerate launch scripts/finetune.py examples/lora-openllama-3b/config.yml \ +accelerate launch scripts/finetune.py examples/openllama-3b/lora.yml \ --inference --lora_model_dir="./lora-out" ``` diff --git a/examples/gptq-lora-7b/README.md b/examples/gptq-lora-7b/README.md index eefe98d3f..0bde51b06 100644 --- a/examples/gptq-lora-7b/README.md +++ b/examples/gptq-lora-7b/README.md @@ -3,6 +3,6 @@ This is a good place to start for beginners. This will run on an NVIDIA RTX4090 with no other changes needed. ```shell -accelerate launch scripts/finetune.py examples/4bit-lora-7b/config.yml +accelerate launch scripts/finetune.py examples/gptq-lora-7b/config.yml ``` diff --git a/examples/openllama-3b/README.md b/examples/openllama-3b/README.md index 9e8f3a9e8..3e9501a54 100644 --- a/examples/openllama-3b/README.md +++ b/examples/openllama-3b/README.md @@ -2,15 +2,15 @@ Basic full tune ```shell -accelerate launch scripts/finetune.py examples/qlora-openllama-3b/config.yml +accelerate launch scripts/finetune.py examples/openllama-3b/config.yml ``` LoRA ```shell -accelerate launch scripts/finetune.py examples/qlora-openllama-3b/lora.yml +accelerate launch scripts/finetune.py examples/openllama-3b/lora.yml ``` QLoRA ```shell -accelerate launch scripts/finetune.py examples/qlora-openllama-3b/qlora.yml +accelerate launch scripts/finetune.py examples/openllama-3b/qlora.yml ```