diff --git a/.nojekyll b/.nojekyll index c29793c81..5d86df9ad 100644 --- a/.nojekyll +++ b/.nojekyll @@ -1 +1 @@ -6e8747fa \ No newline at end of file +2adccae7 \ No newline at end of file diff --git a/docs/lora_optims.html b/docs/lora_optims.html index 4dd6e9b00..27df91675 100644 --- a/docs/lora_optims.html +++ b/docs/lora_optims.html @@ -369,7 +369,14 @@ pre > code.sourceCode > span > a:first-child::before { text-decoration: underlin
Inspired by Unsloth, we’ve implemented two optimizations for LoRA and QLoRA fine-tuning, supporting both single GPU and multi-GPU (in the DDP and DeepSpeed settings) training. These include (1) SwiGLU and GEGLU activation function Triton kernels, and (2) LoRA MLP and attention custom autograd functions. Our goal was to leverage operator fusion and tensor re-use in order to improve speed and reduce memory usage during the forward and backward passes of these calculations.
-We currently support several common model architectures, including (but not limited to): - llama - mistral - qwen2 - gemma - gemma2
We currently support several common model architectures, including (but not limited to):
+llamamistralqwen2gemmagemma2The set of models we support is currently limited by our attention patching strategy, which assumes (and replaces) specific code blocks for query / key / value and output projections:
ORIGINAL_QKV_CODE = """
diff --git a/search.json b/search.json
index f8c595602..a81d3a1fe 100644
--- a/search.json
+++ b/search.json
@@ -1321,7 +1321,7 @@
"href": "docs/lora_optims.html",
"title": "LoRA Optimizations",
"section": "",
- "text": "Inspired by Unsloth, we’ve implemented two optimizations for LoRA and QLoRA fine-tuning, supporting both single GPU and multi-GPU (in the DDP and DeepSpeed settings) training. These include (1) SwiGLU and GEGLU activation function Triton kernels, and (2) LoRA MLP and attention custom autograd functions. Our goal was to leverage operator fusion and tensor re-use in order to improve speed and reduce memory usage during the forward and backward passes of these calculations.\nWe currently support several common model architectures, including (but not limited to): - llama - mistral - qwen2 - gemma - gemma2"
+ "text": "Inspired by Unsloth, we’ve implemented two optimizations for LoRA and QLoRA fine-tuning, supporting both single GPU and multi-GPU (in the DDP and DeepSpeed settings) training. These include (1) SwiGLU and GEGLU activation function Triton kernels, and (2) LoRA MLP and attention custom autograd functions. Our goal was to leverage operator fusion and tensor re-use in order to improve speed and reduce memory usage during the forward and backward passes of these calculations.\nWe currently support several common model architectures, including (but not limited to):"
},
{
"objectID": "docs/lora_optims.html#usage",
diff --git a/sitemap.xml b/sitemap.xml
index 36b28fe37..e5618dffe 100644
--- a/sitemap.xml
+++ b/sitemap.xml
@@ -2,158 +2,158 @@
https://axolotl-ai-cloud.github.io/axolotl/index.html
- 2025-02-19T05:02:45.640Z
+ 2025-02-19T14:23:44.385Z
https://axolotl-ai-cloud.github.io/axolotl/docs/rlhf.html
- 2025-02-19T05:02:45.629Z
+ 2025-02-19T14:23:44.374Z
https://axolotl-ai-cloud.github.io/axolotl/docs/amd_hpc.html
- 2025-02-19T05:02:45.624Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/faq.html
- 2025-02-19T05:02:45.625Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/inference.html
- 2025-02-19T05:02:45.628Z
+ 2025-02-19T14:23:44.373Z
https://axolotl-ai-cloud.github.io/axolotl/docs/unsloth.html
- 2025-02-19T05:02:45.629Z
+ 2025-02-19T14:23:44.374Z
https://axolotl-ai-cloud.github.io/axolotl/docs/multimodal.html
- 2025-02-19T05:02:45.628Z
+ 2025-02-19T14:23:44.373Z
https://axolotl-ai-cloud.github.io/axolotl/docs/input_output.html
- 2025-02-19T05:02:45.628Z
+ 2025-02-19T14:23:44.373Z
https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/index.html
- 2025-02-19T05:02:45.624Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/template_free.html
- 2025-02-19T05:02:45.625Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/inst_tune.html
- 2025-02-19T05:02:45.625Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/stepwise_supervised.html
- 2025-02-19T05:02:45.625Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/reward_modelling.html
- 2025-02-19T05:02:45.629Z
+ 2025-02-19T14:23:44.374Z
https://axolotl-ai-cloud.github.io/axolotl/docs/fsdp_qlora.html
- 2025-02-19T05:02:45.625Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/lr_groups.html
- 2025-02-19T05:02:45.628Z
+ 2025-02-19T14:23:44.373Z
https://axolotl-ai-cloud.github.io/axolotl/docs/mac.html
- 2025-02-19T05:02:45.628Z
+ 2025-02-19T14:23:44.373Z
https://axolotl-ai-cloud.github.io/axolotl/docs/dataset_preprocessing.html
- 2025-02-19T05:02:45.625Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html
- 2025-02-19T05:02:45.644Z
+ 2025-02-19T14:23:44.388Z
https://axolotl-ai-cloud.github.io/axolotl/examples/colab-notebooks/colab-axolotl-example.html
- 2025-02-19T05:02:45.629Z
+ 2025-02-19T14:23:44.374Z
https://axolotl-ai-cloud.github.io/axolotl/FAQS.html
- 2025-02-19T05:02:45.622Z
+ 2025-02-19T14:23:44.369Z
https://axolotl-ai-cloud.github.io/axolotl/TODO.html
- 2025-02-19T05:02:45.623Z
+ 2025-02-19T14:23:44.369Z
https://axolotl-ai-cloud.github.io/axolotl/src/axolotl/integrations/kd/topk_logprob/LICENSE.html
- 2025-02-19T05:02:45.644Z
+ 2025-02-19T14:23:44.389Z
https://axolotl-ai-cloud.github.io/axolotl/src/axolotl/integrations/LICENSE.html
- 2025-02-19T05:02:45.643Z
+ 2025-02-19T14:23:44.388Z
https://axolotl-ai-cloud.github.io/axolotl/docs/installation.html
- 2025-02-19T05:02:45.628Z
+ 2025-02-19T14:23:44.373Z
https://axolotl-ai-cloud.github.io/axolotl/docs/torchao.html
- 2025-02-19T05:02:45.629Z
+ 2025-02-19T14:23:44.374Z
https://axolotl-ai-cloud.github.io/axolotl/docs/ray-integration.html
- 2025-02-19T05:02:45.629Z
+ 2025-02-19T14:23:44.374Z
https://axolotl-ai-cloud.github.io/axolotl/docs/cli.html
- 2025-02-19T05:02:45.624Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/debugging.html
- 2025-02-19T05:02:45.625Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/tokenized.html
- 2025-02-19T05:02:45.625Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/pretraining.html
- 2025-02-19T05:02:45.625Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/conversation.html
- 2025-02-19T05:02:45.624Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/multi-gpu.html
- 2025-02-19T05:02:45.628Z
+ 2025-02-19T14:23:44.373Z
https://axolotl-ai-cloud.github.io/axolotl/docs/batch_vs_grad.html
- 2025-02-19T05:02:45.624Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/getting-started.html
- 2025-02-19T05:02:45.625Z
+ 2025-02-19T14:23:44.371Z
https://axolotl-ai-cloud.github.io/axolotl/docs/config.html
- 2025-02-19T05:02:45.624Z
+ 2025-02-19T14:23:44.370Z
https://axolotl-ai-cloud.github.io/axolotl/docs/multipack.html
- 2025-02-19T05:02:45.629Z
+ 2025-02-19T14:23:44.374Z
https://axolotl-ai-cloud.github.io/axolotl/docs/multi-node.html
- 2025-02-19T05:02:45.628Z
+ 2025-02-19T14:23:44.373Z
https://axolotl-ai-cloud.github.io/axolotl/docs/nccl.html
- 2025-02-19T05:02:45.629Z
+ 2025-02-19T14:23:44.374Z
https://axolotl-ai-cloud.github.io/axolotl/docs/lora_optims.html
- 2025-02-19T05:02:45.628Z
+ 2025-02-19T14:23:44.373Z