diff --git a/README.md b/README.md index b119f6e12..47750bc1a 100644 --- a/README.md +++ b/README.md @@ -1,383 +1,177 @@
- - - - Axolotl - - -

🧠 Streamlined AI Model Post-Training

-

- A powerful, flexible tool designed to streamline post-training for various AI models with enterprise-grade features and optimizations. + + Axolotl Logo + +

Axolotl: Fine-tune LLMs with Unprecedented Ease & Power! 🚀

+

Your ultimate toolkit for efficient, scalable, and versatile large language model fine-tuning.

+ +

+ + Discord Community + + + Official Documentation + + + PyPI Package + + + GitHub Downloads + +

+
+
+ +--- + +
+

🎉 Latest Innovations & Updates!

+ +
+ +

Axolotl Overview: Your LLM Fine-tuning Powerhouse!

+ +
+

Axolotl is a powerful, flexible, and user-friendly tool designed to supercharge your post-training workflows for a wide range of cutting-edge AI models.

+
+ +
+
+

🤖 Broad Model Compatibility

+ +
+ +
+

🔧 Diverse Training Methodologies

+ +
+ +
+

⚙️ Streamlined Configuration

+ +
+ +
+

Cutting-Edge Performance Optimizations

+ +
+ +
+

📂 Flexible Data Handling

+ +
+ +
+

☁️ Cloud-Ready & Deployable

+ +
+
+ +

🚀 Quick Start: Get Fine-tuning in Minutes!

+ +
+

Requirements:

+ + +

Installation:

+
pip3 install -U packaging==23.2 setuptools==75.8.0 wheel ninja
+pip3 install --no-build-isolation axolotl[flash-attn,deepspeed]
+
+# Download example axolotl configs, deepspeed configs
+axolotl fetch examples
+axolotl fetch deepspeed_configs  # OPTIONAL
+

Other installation approaches are described here.

+ +

Your First Fine-tune:

+
# Fetch axolotl examples
+axolotl fetch examples
+
+# Or, specify a custom path
+axolotl fetch examples --dest path/to/folder
+
+# Train a model using LoRA
+axolotl train examples/llama-3/lora-1b.yml
+

+ That's it! Check out our  Getting Started Guide ➜ for a more detailed walkthrough.

-
- - - - - - - - - -
- GitHub License - - GitHub Repo stars -
- tests - - - codecov - -
+

📚 Comprehensive Documentation: Unlock Axolotl's Full Potential

+ +
+

Dive deep into Axolotl's capabilities with our extensive documentation:

+
-
- - contributors - - - Releases +

🤝 Need Help? We're Here for You!

+
+ +

🌟 Contribute to Axolotl!

+

+ Contributions are always welcome and highly appreciated! Axolotl thrives on community support. Please see our Contributing Guide for details on how you can help make Axolotl even better. +

+ +
+

❤️ Our Esteemed Sponsors

+

A huge thank you to our visionary sponsors who provide the essential resources to keep Axolotl at the forefront of LLM fine-tuning:

+ + Modal Logo +

+ Modal: Revolutionizing cloud computing for Gen AI. Run jobs, deploy models, and fine-tune LLMs at scale with ease. +

+

+ Interested in powering the future of Axolotl? Become a sponsor! Contact us at wing@axolotl.ai +

-
- - discord - - - twitter - -
- ---- - -
-

- 🎉 Latest Updates -

- -
-
- - 📅 2025/06: Magistral Support Added - -

- Magistral with mistral-common tokenizer support has been added to Axolotl. - - See examples → - -

-
- -
- - 📅 2025/05: Quantization Aware Training (QAT) - -

- QAT support has been added to Axolotl. - - Explore the docs → - -

-
- -
- - 📅 2025/04: Llama 4 Support - -

- Llama 4 support has been added in Axolotl. - - See examples → - -

-
- -
- - 📅 2025/03: Sequence Parallelism & Multimodal Support - -
-

- • Sequence Parallelism (SP) for scaling context length - - - Blog - | - - Docs - -

-

- • (Beta) Multimodal models fine-tuning - - - Check docs → - -

-
-
- -
- - 📅 2025/02: LoRA Optimizations & GRPO Support - -
-

- • LoRA optimizations for better memory usage and speed - - - Docs → - -

-

- • GRPO support added - - - Blog - | - - Example - -

-
-
- -
- - 📅 2025/01: Reward Modelling Support - -

- Reward Modelling / Process Reward Modelling fine-tuning support added. - - See docs → - -

-
-
-
- ---- - -
-

- ✨ What Makes Axolotl Special -

- -
-
-

🚀 Multiple Model Support

-

- Train LLaMA, Mistral, Mixtral, Pythia, and more. Full compatibility with HuggingFace transformers causal language models. -

-
- -
-

🎯 Advanced Training Methods

-

- Full fine-tuning, LoRA, QLoRA, GPTQ, QAT, Preference Tuning (DPO, IPO, KTO, ORPO), RL (GRPO), Multimodal, and Reward Modelling. -

-
- -
-

⚙️ Easy Configuration

-

- Reuse a single YAML file across dataset preprocessing, training, evaluation, quantization, and inference. -

-
- -
-

⚡ Performance Optimizations

-

- Multipacking, Flash Attention, Xformers, Flex Attention, Liger Kernel, Sequence Parallelism, and Multi-GPU training. -

-
- -
-

📊 Flexible Dataset Handling

-

- Load from local files, HuggingFace datasets, and cloud storage (S3, Azure, GCP, OCI). -

-
- -
-

☁️ Cloud Ready

-

- Pre-built Docker images and PyPI packages for seamless deployment on cloud platforms and local hardware. -

-
-
-
- ---- - -
-

- 🚀 Quick Start -

- -
-

📋 Requirements

-
    -
  • GPU: NVIDIA GPU (Ampere or newer for bf16 and Flash Attention) or AMD GPU
  • -
  • Python: 3.11+
  • -
  • PyTorch: ≥2.5.1
  • -
-
- -
-

💾 Installation

-
-
# Install dependencies
-pip3 install -U packaging==23.2 setuptools==75.8.0 wheel ninja
-
-# Install Axolotl with Flash Attention and DeepSpeed
-pip3 install --no-build-isolation axolotl[flash-attn,deepspeed]
-
-# Download examples and configs
-axolotl fetch examples
-axolotl fetch deepspeed_configs  # OPTIONAL
-
-

- Other installation methods available in our documentation -

-
- -
-

🎯 Your First Fine-tune

-
-
# Fetch examples
-axolotl fetch examples
-
-# Or specify custom path
-axolotl fetch examples --dest path/to/folder
-
-# Start training with LoRA
-axolotl train examples/llama-3/lora-1b.yml
-
-

- That's it! Check our Getting Started Guide for detailed walkthrough -

-
-
- ---- - -
-

- 📚 Documentation Hub -

- -
- -
-

🔧 Installation Options

-

Detailed setup instructions for different environments

-
-
- - -
-

⚙️ Configuration Guide

-

Full configuration options and examples

-
-
- - -
-

📊 Dataset Loading

-

Loading datasets from various sources

-
-
- - -
-

📋 Dataset Guide

-

Supported formats and usage instructions

-
-
- - -
-

🖥️ Multi-GPU Training

-

Scale your training across multiple GPUs

-
-
- - -
-

🌐 Multi-Node Training

-

Distributed training across multiple machines

-
-
- - -
-

📦 Multipacking

-

Efficient batch packing for training

-
-
- - -
-

🔍 API Reference

-

Auto-generated code documentation

-
-
- - -
-

❓ FAQ

-

Frequently asked questions

-
-
-
-
- ---- - -
-

- 🤝 Getting Help -

- -
-
-
💬
-

Community Support

-

Join thousands of developers in our Discord

- Join Discord -
- -
-
📖
-

Examples

-

Browse our comprehensive examples

- View Examples -
- -
-
🔧
-

Debugging

-

Troubleshooting and debugging guide

- Debug Guide -
- -
-
✉️
-

Enterprise Support

-

Need dedicated support? Contact us

- Contact Us -
-
-
- ---- - -
-

- 🌟 Contributing -

- -
-

- We welcome contributions from the community! Whether it's bug fixes, \ No newline at end of file +

📜 License

+

+ This project is proudly licensed under the Apache 2.0 License. See the LICENSE file for full details. +

\ No newline at end of file