diff --git a/README.md b/README.md
index 097084146..a81b63f1b 100644
--- a/README.md
+++ b/README.md
@@ -1,10 +1,39 @@
# Axolotl
+Axolotl is a tool designed to streamline the fine-tuning of various AI models, offering support for multiple configurations and architectures.
+
+
+
+|
+
+## Table of Contents
+- [Introduction](#axolotl)
+- [Supported Features](#axolotl-supports)
+- [Quickstart](#quickstart-)
+- [Installation](#installation)
+ - [Docker Installation](#environment)
+ - [Conda/Pip venv Installation](#condapip-venv)
+ - [LambdaLabs Installation](#lambdalabs)
+- [Dataset](#dataset)
+ - [How to Add Custom Prompts](#how-to-add-custom-prompts)
+- [Config](#config)
+ - [Train](#train)
+ - [Inference](#inference)
+ - [Merge LORA to Base](#merge-lora-to-base)
+- [Common Errors](#common-errors-)
+- [Need Help?](#need-help-)
+- [Badge](#badge-)
+- [Community Showcase](#community-showcase)
+- [Contributing](#contributing-)
+
+ |
+
+
- One repo to finetune them all!
+ Axolotl provides a unified repository for fine-tuning a variety of AI models with ease
Go ahead and axolotl questions!!
@@ -14,6 +43,10 @@
+ |
+
+
+
## Axolotl supports
| | fp16/fp32 | lora | qlora | gptq | gptq w/ lora | gptq w/flash attn | flash attn | xformers attn |
@@ -29,6 +62,8 @@
## Quickstart ⚡
+Get started with Axolotl in just a few steps! This quickstart guide will walk you through setting up and running a basic fine-tuning task.
+
**Requirements**: Python >=3.9 and Pytorch >=2.0.
```bash
@@ -130,6 +165,7 @@ accelerate launch scripts/finetune.py examples/openllama-3b/lora.yml \
### Dataset
+Axolotl supports a variety of dataset formats. Below are some of the formats you can use.
Have dataset(s) in one of the following format (JSONL recommended):
- `alpaca`: instruction; input(optional)
@@ -622,7 +658,7 @@ CUDA_VISIBLE_DEVICES="" python3 scripts/finetune.py ...
## Common Errors 🧰
-> Cuda out of memory
+> If you encounter a 'Cuda out of memory' error, it means your GPU ran out of memory during the training process. Here's how to resolve it:
Please reduce any below
- `micro_batch_size`
@@ -630,6 +666,10 @@ Please reduce any below
- `gradient_accumulation_steps`
- `sequence_len`
+> `failed (exitcode: -9)` usually means your system has run out of system memory.
+Similarly, you should consider reducing the same settings as when you run out of VRAM.
+Additionally, look into upgrading your system RAM which should be simpler than GPU upgrades.
+
> RuntimeError: expected scalar type Float but found Half
Try set `fp16: true`
@@ -658,6 +698,8 @@ Building something cool with Axolotl? Consider adding a badge to your model card
## Community Showcase
+Check out some of the projects and models that have been built using Axolotl! Have a model you'd like to add to our Community Showcase? Open a PR with your model.
+
Open Access AI Collective
- [Minotaur 13b](https://huggingface.co/openaccess-ai-collective/minotaur-13b)
- [Manticore 13b](https://huggingface.co/openaccess-ai-collective/manticore-13b)