From 109cae8122038125fc4e4dd23f091b51d98efb93 Mon Sep 17 00:00:00 2001 From: Quarto GHA Workflow Runner Date: Wed, 19 Feb 2025 14:24:39 +0000 Subject: [PATCH] Built site for gh-pages --- .nojekyll | 2 +- docs/lora_optims.html | 9 ++++- search.json | 2 +- sitemap.xml | 78 +++++++++++++++++++++---------------------- 4 files changed, 49 insertions(+), 42 deletions(-) diff --git a/.nojekyll b/.nojekyll index c29793c81..5d86df9ad 100644 --- a/.nojekyll +++ b/.nojekyll @@ -1 +1 @@ -6e8747fa \ No newline at end of file +2adccae7 \ No newline at end of file diff --git a/docs/lora_optims.html b/docs/lora_optims.html index 4dd6e9b00..27df91675 100644 --- a/docs/lora_optims.html +++ b/docs/lora_optims.html @@ -369,7 +369,14 @@ pre > code.sourceCode > span > a:first-child::before { text-decoration: underlin

Inspired by Unsloth, we’ve implemented two optimizations for LoRA and QLoRA fine-tuning, supporting both single GPU and multi-GPU (in the DDP and DeepSpeed settings) training. These include (1) SwiGLU and GEGLU activation function Triton kernels, and (2) LoRA MLP and attention custom autograd functions. Our goal was to leverage operator fusion and tensor re-use in order to improve speed and reduce memory usage during the forward and backward passes of these calculations.

-

We currently support several common model architectures, including (but not limited to): - llama - mistral - qwen2 - gemma - gemma2

+

We currently support several common model architectures, including (but not limited to):

+

The set of models we support is currently limited by our attention patching strategy, which assumes (and replaces) specific code blocks for query / key / value and output projections:

ORIGINAL_QKV_CODE = """
diff --git a/search.json b/search.json
index f8c595602..a81d3a1fe 100644
--- a/search.json
+++ b/search.json
@@ -1321,7 +1321,7 @@
     "href": "docs/lora_optims.html",
     "title": "LoRA Optimizations",
     "section": "",
-    "text": "Inspired by Unsloth, we’ve implemented two optimizations for LoRA and QLoRA fine-tuning, supporting both single GPU and multi-GPU (in the DDP and DeepSpeed settings) training. These include (1) SwiGLU and GEGLU activation function Triton kernels, and (2) LoRA MLP and attention custom autograd functions. Our goal was to leverage operator fusion and tensor re-use in order to improve speed and reduce memory usage during the forward and backward passes of these calculations.\nWe currently support several common model architectures, including (but not limited to): - llama - mistral - qwen2 - gemma - gemma2"
+    "text": "Inspired by Unsloth, we’ve implemented two optimizations for LoRA and QLoRA fine-tuning, supporting both single GPU and multi-GPU (in the DDP and DeepSpeed settings) training. These include (1) SwiGLU and GEGLU activation function Triton kernels, and (2) LoRA MLP and attention custom autograd functions. Our goal was to leverage operator fusion and tensor re-use in order to improve speed and reduce memory usage during the forward and backward passes of these calculations.\nWe currently support several common model architectures, including (but not limited to):"
   },
   {
     "objectID": "docs/lora_optims.html#usage",
diff --git a/sitemap.xml b/sitemap.xml
index 36b28fe37..e5618dffe 100644
--- a/sitemap.xml
+++ b/sitemap.xml
@@ -2,158 +2,158 @@
 
   
     https://axolotl-ai-cloud.github.io/axolotl/index.html
-    2025-02-19T05:02:45.640Z
+    2025-02-19T14:23:44.385Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/rlhf.html
-    2025-02-19T05:02:45.629Z
+    2025-02-19T14:23:44.374Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/amd_hpc.html
-    2025-02-19T05:02:45.624Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/faq.html
-    2025-02-19T05:02:45.625Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/inference.html
-    2025-02-19T05:02:45.628Z
+    2025-02-19T14:23:44.373Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/unsloth.html
-    2025-02-19T05:02:45.629Z
+    2025-02-19T14:23:44.374Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/multimodal.html
-    2025-02-19T05:02:45.628Z
+    2025-02-19T14:23:44.373Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/input_output.html
-    2025-02-19T05:02:45.628Z
+    2025-02-19T14:23:44.373Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/index.html
-    2025-02-19T05:02:45.624Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/template_free.html
-    2025-02-19T05:02:45.625Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/inst_tune.html
-    2025-02-19T05:02:45.625Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/stepwise_supervised.html
-    2025-02-19T05:02:45.625Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/reward_modelling.html
-    2025-02-19T05:02:45.629Z
+    2025-02-19T14:23:44.374Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/fsdp_qlora.html
-    2025-02-19T05:02:45.625Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/lr_groups.html
-    2025-02-19T05:02:45.628Z
+    2025-02-19T14:23:44.373Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/mac.html
-    2025-02-19T05:02:45.628Z
+    2025-02-19T14:23:44.373Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset_preprocessing.html
-    2025-02-19T05:02:45.625Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html
-    2025-02-19T05:02:45.644Z
+    2025-02-19T14:23:44.388Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/examples/colab-notebooks/colab-axolotl-example.html
-    2025-02-19T05:02:45.629Z
+    2025-02-19T14:23:44.374Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/FAQS.html
-    2025-02-19T05:02:45.622Z
+    2025-02-19T14:23:44.369Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/TODO.html
-    2025-02-19T05:02:45.623Z
+    2025-02-19T14:23:44.369Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/src/axolotl/integrations/kd/topk_logprob/LICENSE.html
-    2025-02-19T05:02:45.644Z
+    2025-02-19T14:23:44.389Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/src/axolotl/integrations/LICENSE.html
-    2025-02-19T05:02:45.643Z
+    2025-02-19T14:23:44.388Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/installation.html
-    2025-02-19T05:02:45.628Z
+    2025-02-19T14:23:44.373Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/torchao.html
-    2025-02-19T05:02:45.629Z
+    2025-02-19T14:23:44.374Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/ray-integration.html
-    2025-02-19T05:02:45.629Z
+    2025-02-19T14:23:44.374Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/cli.html
-    2025-02-19T05:02:45.624Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/debugging.html
-    2025-02-19T05:02:45.625Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/tokenized.html
-    2025-02-19T05:02:45.625Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/pretraining.html
-    2025-02-19T05:02:45.625Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/conversation.html
-    2025-02-19T05:02:45.624Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/multi-gpu.html
-    2025-02-19T05:02:45.628Z
+    2025-02-19T14:23:44.373Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/batch_vs_grad.html
-    2025-02-19T05:02:45.624Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/getting-started.html
-    2025-02-19T05:02:45.625Z
+    2025-02-19T14:23:44.371Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/config.html
-    2025-02-19T05:02:45.624Z
+    2025-02-19T14:23:44.370Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/multipack.html
-    2025-02-19T05:02:45.629Z
+    2025-02-19T14:23:44.374Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/multi-node.html
-    2025-02-19T05:02:45.628Z
+    2025-02-19T14:23:44.373Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/nccl.html
-    2025-02-19T05:02:45.629Z
+    2025-02-19T14:23:44.374Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/lora_optims.html
-    2025-02-19T05:02:45.628Z
+    2025-02-19T14:23:44.373Z