diff --git a/README.md b/README.md index 097084146..a81b63f1b 100644 --- a/README.md +++ b/README.md @@ -1,10 +1,39 @@ # Axolotl +Axolotl is a tool designed to streamline the fine-tuning of various AI models, offering support for multiple configurations and architectures. + + + + + + +
+ +## Table of Contents +- [Introduction](#axolotl) +- [Supported Features](#axolotl-supports) +- [Quickstart](#quickstart-) +- [Installation](#installation) + - [Docker Installation](#environment) + - [Conda/Pip venv Installation](#condapip-venv) + - [LambdaLabs Installation](#lambdalabs) +- [Dataset](#dataset) + - [How to Add Custom Prompts](#how-to-add-custom-prompts) +- [Config](#config) + - [Train](#train) + - [Inference](#inference) + - [Merge LORA to Base](#merge-lora-to-base) +- [Common Errors](#common-errors-) +- [Need Help?](#need-help-) +- [Badge](#badge-) +- [Community Showcase](#community-showcase) +- [Contributing](#contributing-) + + +
axolotl

- One repo to finetune them all! + Axolotl provides a unified repository for fine-tuning
a variety of AI models with ease

Go ahead and axolotl questions!! @@ -14,6 +43,10 @@

+
+ ## Axolotl supports | | fp16/fp32 | lora | qlora | gptq | gptq w/ lora | gptq w/flash attn | flash attn | xformers attn | @@ -29,6 +62,8 @@ ## Quickstart ⚡ +Get started with Axolotl in just a few steps! This quickstart guide will walk you through setting up and running a basic fine-tuning task. + **Requirements**: Python >=3.9 and Pytorch >=2.0. ```bash @@ -130,6 +165,7 @@ accelerate launch scripts/finetune.py examples/openllama-3b/lora.yml \ ### Dataset +Axolotl supports a variety of dataset formats. Below are some of the formats you can use. Have dataset(s) in one of the following format (JSONL recommended): - `alpaca`: instruction; input(optional) @@ -622,7 +658,7 @@ CUDA_VISIBLE_DEVICES="" python3 scripts/finetune.py ... ## Common Errors 🧰 -> Cuda out of memory +> If you encounter a 'Cuda out of memory' error, it means your GPU ran out of memory during the training process. Here's how to resolve it: Please reduce any below - `micro_batch_size` @@ -630,6 +666,10 @@ Please reduce any below - `gradient_accumulation_steps` - `sequence_len` +> `failed (exitcode: -9)` usually means your system has run out of system memory. +Similarly, you should consider reducing the same settings as when you run out of VRAM. +Additionally, look into upgrading your system RAM which should be simpler than GPU upgrades. + > RuntimeError: expected scalar type Float but found Half Try set `fp16: true` @@ -658,6 +698,8 @@ Building something cool with Axolotl? Consider adding a badge to your model card ## Community Showcase +Check out some of the projects and models that have been built using Axolotl! Have a model you'd like to add to our Community Showcase? Open a PR with your model. + Open Access AI Collective - [Minotaur 13b](https://huggingface.co/openaccess-ai-collective/minotaur-13b) - [Manticore 13b](https://huggingface.co/openaccess-ai-collective/manticore-13b)