From 3f6017db9e88dcff1011c38e6aa37888faca4f09 Mon Sep 17 00:00:00 2001 From: Wing Lian Date: Thu, 25 May 2023 22:39:13 -0400 Subject: [PATCH] qlora merge and load requires that base model isn't loaded in 4 or 8 bit --- README.md | 16 +++++++++++----- 1 file changed, 11 insertions(+), 5 deletions(-) diff --git a/README.md b/README.md index f79a49a1f..28cbf21b8 100644 --- a/README.md +++ b/README.md @@ -24,7 +24,7 @@ ## Quickstart ⚡ -**Requirements**: Python 3.9. +**Requirements**: Python 3.9. ```bash git clone https://github.com/OpenAccess-AI-Collective/axolotl @@ -45,7 +45,7 @@ accelerate launch scripts/finetune.py examples/4bit-lora-7b/config.yml \ ### Environment -- Docker +- Docker ```bash docker run --gpus '"all"' --rm -it winglian/axolotl:main ``` @@ -332,7 +332,7 @@ seed: ### Accelerate -Configure accelerate +Configure accelerate ```bash accelerate config @@ -363,12 +363,18 @@ Pass the appropriate flag to the train command: ### Merge LORA to base -Add below flag to train command above +Add below flag to train command above (and using LoRA) ```bash --merge_lora --lora_model_dir="./completed-model" ``` +Add below flag to train command above (and using QLoRA) + +```bash +--merge_lora --lora_model_dir="./completed-model" --load_in_8bit False --load_in_4bit False +``` + ## Common Errors 🧰 > Cuda out of memory @@ -383,7 +389,7 @@ Please reduce any below Try set `fp16: true` ## Need help? 🙋‍♂️ - + Join our [Discord server](https://discord.gg/HhrNrHJPRb) where we can help you ## Contributing 🤝