Files
axolotl/README.md

17 KiB

Axolotl Logo

Axolotl: Fine-tune LLMs with Unprecedented Ease & Power! 🚀

Your ultimate toolkit for efficient, scalable, and versatile large language model fine-tuning.

Discord Community Official Documentation PyPI Package GitHub Downloads



🎉 Latest Innovations & Updates!

  • 2025/06: Magistral with mistral-common tokenizer support! Dive into examples to train your own Magistral models.
  • 2025/05: Quantization Aware Training (QAT) support! Explore the docs to learn more.
  • 2025/04: Llama 4 support! See examples to train Llama 4 with Axolotl's linearized version!
  • 2025/03: Sequence Parallelism (SP) support! Scale your context length. Read the blog and docs.
  • 2025/03: (Beta) Fine-tuning Multimodal models! Check out the docs.
  • 2025/02: LoRA optimizations! Reduce memory and improve speed. Jump into the docs.
  • 2025/02: GRPO support! Dive into our blog and GRPO example.
  • 2025/01: Reward Modelling / Process Reward Modelling fine-tuning! See docs.

Axolotl Overview: Your LLM Fine-tuning Powerhouse!

Axolotl is a powerful, flexible, and user-friendly tool designed to supercharge your post-training workflows for a wide range of cutting-edge AI models.

🤖 Broad Model Compatibility

  • Train a vast array of models including LLaMA, Mistral, Mixtral, Pythia, and many more.
  • Fully compatible with HuggingFace transformers causal language models, ensuring wide adoption.

🔧 Diverse Training Methodologies

  • Full fine-tuning, LoRA, QLoRA, GPTQ, QAT.
  • Preference Tuning: DPO, IPO, KTO, ORPO.
  • Advanced RL: GRPO.
  • Multimodal and Reward Modelling (RM) / Process Reward Modelling (PRM).

⚙️ Streamlined Configuration

  • Utilize a single, intuitive YAML file across dataset preprocess, training, evaluation, quantization, and inference.

Cutting-Edge Performance Optimizations

📂 Flexible Data Handling

  • Load datasets from local paths, HuggingFace Hub, and major cloud providers (S3, Azure, GCP, OCI).

☁️ Cloud-Ready & Deployable

🚀 Quick Start: Get Fine-tuning in Minutes!

Requirements:

  • ▶ NVIDIA GPU (Ampere or newer for `bf16` and Flash Attention) or AMD GPU
  • ▶ Python 3.11
  • ▶ PyTorch ≥2.5.1

Installation:

pip3 install -U packaging==23.2 setuptools==75.8.0 wheel ninja
pip3 install --no-build-isolation axolotl[flash-attn,deepspeed]

Download example axolotl configs, deepspeed configs

axolotl fetch examples axolotl fetch deepspeed_configs # OPTIONAL

Other installation approaches are described here.

Your First Fine-tune:

# Fetch axolotl examples
axolotl fetch examples

Or, specify a custom path

axolotl fetch examples --dest path/to/folder

Train a model using LoRA

axolotl train examples/llama-3/lora-1b.yml

That's it! Check out our  Getting Started Guide ➜ for a more detailed walkthrough.

📚 Comprehensive Documentation: Unlock Axolotl's Full Potential

Dive deep into Axolotl's capabilities with our extensive documentation:

🤝 Need Help? We're Here for You!

  • Join our vibrant Discord community for real-time support and discussions.
  • Explore our Examples directory for practical use cases.
  • Read our Debugging Guide for troubleshooting tips.
  • Need dedicated support? Please contact wing@axolotl.ai for professional assistance options.

🌟 Contribute to Axolotl!

Contributions are always welcome and highly appreciated! Axolotl thrives on community support. Please see our Contributing Guide for details on how you can help make Axolotl even better.

❤️ Our Esteemed Sponsors

A huge thank you to our visionary sponsors who provide the essential resources to keep Axolotl at the forefront of LLM fine-tuning:

Modal Logo

Modal: Revolutionizing cloud computing for Gen AI. Run jobs, deploy models, and fine-tune LLMs at scale with ease.

Interested in powering the future of Axolotl? Become a sponsor! Contact us at wing@axolotl.ai

📜 License

This project is proudly licensed under the Apache 2.0 License. See the LICENSE file for full details.