From d7f9f4e61f6525a95adb7161239cf1fbffcccdb4 Mon Sep 17 00:00:00 2001
From: Quarto GHA Workflow Runner
Date: Mon, 7 Jul 2025 21:10:36 +0000
Subject: [PATCH] Built site for gh-pages
---
.nojekyll | 2 +-
docs/custom_integrations.html | 409 +++++++++++++++++-----------------
search.json | 35 ++-
sitemap.xml | 378 +++++++++++++++----------------
4 files changed, 422 insertions(+), 402 deletions(-)
diff --git a/.nojekyll b/.nojekyll
index adb0eac46..ce3b02b59 100644
--- a/.nojekyll
+++ b/.nojekyll
@@ -1 +1 @@
-6fea3a42
\ No newline at end of file
+2aa75668
\ No newline at end of file
diff --git a/docs/custom_integrations.html b/docs/custom_integrations.html
index 11dfc0712..44c037289 100644
--- a/docs/custom_integrations.html
+++ b/docs/custom_integrations.html
@@ -475,6 +475,7 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
Supported Models
Citation
+ DenseMixer
Grokfast
Usage
@@ -484,31 +485,31 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
- Liger Kernels
+ LLMCompressor
Language Model Evaluation Harness (LM Eval)
+ Liger Kernels
+
Spectrum
- LLMCompressor
-
Adding a new integration
@@ -609,25 +610,33 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
Please see reference here
+
+DenseMixer
+See DenseMixer
+Simply add the following to your axolotl YAML config:
+plugins :
+ - axolotl.integrations.densemixer.DenseMixerPlugin
+Please see reference here
+
Grokfast
See https://github.com/ironjr/grokfast
Usage
-plugins :
- - axolotl.integrations.grokfast.GrokfastPlugin
-
-grokfast_alpha : 2.0
-grokfast_lamb : 0.98
+plugins :
+ - axolotl.integrations.grokfast.GrokfastPlugin
+
+grokfast_alpha : 2.0
+grokfast_lamb : 0.98
Citation
-@article {lee2024grokfast ,
- title ={{Grokfast}: Accelerated Grokking by Amplifying Slow Gradients},
- author ={Lee, Jaerin and Kang, Bong Gyun and Kim, Kihoon and Lee, Kyoung Mu},
- journal ={arXiv preprint arXiv:2405.20233},
- year ={2024}
- }
+@article {lee2024grokfast ,
+ title ={{Grokfast}: Accelerated Grokking by Amplifying Slow Gradients},
+ author ={Lee, Jaerin and Kang, Bong Gyun and Kim, Kihoon and Lee, Kyoung Mu},
+ journal ={arXiv preprint arXiv:2405.20233},
+ year ={2024}
+ }
Please see reference here
@@ -635,145 +644,25 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
Knowledge Distillation (KD)
Usage
-plugins :
- - "axolotl.integrations.kd.KDPlugin"
-
-kd_trainer : True
-kd_ce_alpha : 0.1
-kd_alpha : 0.9
-kd_temperature : 1.0
-
-torch_compile : True # torch>=2.5.1, recommended to reduce vram
-
-datasets :
- - path : ...
- type : "axolotl.integrations.kd.chat_template"
- field_messages : "messages_combined"
- logprobs_field : "llm_text_generation_vllm_logprobs" # for kd only, field of logprobs
+plugins :
+ - "axolotl.integrations.kd.KDPlugin"
+
+kd_trainer : True
+kd_ce_alpha : 0.1
+kd_alpha : 0.9
+kd_temperature : 1.0
+
+torch_compile : True # torch>=2.5.1, recommended to reduce vram
+
+datasets :
+ - path : ...
+ type : "axolotl.integrations.kd.chat_template"
+ field_messages : "messages_combined"
+ logprobs_field : "llm_text_generation_vllm_logprobs" # for kd only, field of logprobs
An example dataset can be found at axolotl-ai-co/evolkit-logprobs-pipeline-75k-v2-sample
Please see reference here
-
-Liger Kernels
-Liger Kernel provides efficient Triton kernels for LLM training, offering:
-
-20% increase in multi-GPU training throughput
-60% reduction in memory usage
-Compatibility with both FSDP and DeepSpeed
-
-See https://github.com/linkedin/Liger-Kernel
-
-Usage
-plugins :
- - axolotl.integrations.liger.LigerPlugin
-liger_rope : true
-liger_rms_norm : true
-liger_glu_activation : true
-liger_layer_norm : true
-liger_fused_linear_cross_entropy : true
-
-
-Supported Models
-
-deepseek_v2
-gemma
-gemma2
-gemma3
-granite
-jamba
-llama
-mistral
-mixtral
-mllama
-mllama_text_model
-olmo2
-paligemma
-phi3
-qwen2
-qwen2_5_vl
-qwen2_vl
-
-
-
-Citation
-@article {hsu2024ligerkernelefficienttriton ,
- title ={Liger Kernel: Efficient Triton Kernels for LLM Training},
- author ={Pin-Lun Hsu and Yun Dai and Vignesh Kothapalli and Qingquan Song and Shao Tang and Siyu Zhu and Steven Shimizu and Shivam Sahni and Haowen Ning and Yanning Chen},
- year ={2024},
- eprint ={2410.10989},
- archivePrefix ={arXiv},
- primaryClass ={cs.LG},
- url ={https://arxiv.org/abs/2410.10989},
- journal ={arXiv preprint arXiv:2410.10989},
- }
-Please see reference here
-
-
-
-Language Model Evaluation Harness (LM Eval)
-Run evaluation on model using the popular lm-evaluation-harness library.
-See https://github.com/EleutherAI/lm-evaluation-harness
-
-Usage
-plugins :
- - axolotl.integrations.lm_eval.LMEvalPlugin
-
-lm_eval_tasks :
- - gsm8k
- - hellaswag
- - arc_easy
-
-lm_eval_batch_size : # Batch size for evaluation
-output_dir : # Directory to save evaluation results
-
-
-Citation
-@misc {eval -harness ,
- author = {Gao, Leo and Tow, Jonathan and Abbasi, Baber and Biderman, Stella and Black, Sid and DiPofi, Anthony and Foster, Charles and Golding, Laurence and Hsu, Jeffrey and Le Noac'h, Alain and Li, Haonan and McDonell, Kyle and Muennighoff, Niklas and Ociepa, Chris and Phang, Jason and Reynolds, Laria and Schoelkopf, Hailey and Skowron, Aviya and Sutawika, Lintang and Tang, Eric and Thite, Anish and Wang, Ben and Wang, Kevin and Zou, Andy},
- title = {A framework for few-shot language model evaluation},
- month = 07,
- year = 2024,
- publisher = {Zenodo},
- version = {v0.4.3},
- doi = {10.5281/zenodo.12608602},
- url = {https://zenodo.org/records/12608602}
- }
-Please see reference here
-
-
-
-Spectrum
-by Eric Hartford, Lucas Atkins, Fernando Fernandes, David Golchinfar
-This plugin contains code to freeze the bottom fraction of modules in a model, based on the Signal-to-Noise Ratio (SNR).
-See https://github.com/cognitivecomputations/spectrum
-
-Overview
-Spectrum is a tool for scanning and evaluating the Signal-to-Noise Ratio (SNR) of layers in large language models.
-By identifying the top n% of layers with the highest SNR, you can optimize training efficiency.
-
-
-Usage
-plugins :
- - axolotl.integrations.spectrum.SpectrumPlugin
-
-spectrum_top_fraction : 0.5
-spectrum_model_name : meta-llama/Meta-Llama-3.1-8B
-
-
-Citation
-@misc {hartford2024spectrumtargetedtrainingsignal ,
- title ={Spectrum: Targeted Training on Signal to Noise Ratio},
- author ={Eric Hartford and Lucas Atkins and Fernando Fernandes Neto and David Golchinfar},
- year ={2024},
- eprint ={2406.06623},
- archivePrefix ={arXiv},
- primaryClass ={cs.LG},
- url ={https://arxiv.org/abs/2406.06623},
- }
-Please see reference here
-
-
LLMCompressor
Fine-tune sparsified models in Axolotl using Neural Magic’s LLMCompressor .
@@ -784,34 +673,34 @@ By identifying the top n% of layers with the highest SNR, you can optimize train
Requirements
Axolotl with llmcompressor extras:
-pip install "axolotl[llmcompressor]"
+pip install "axolotl[llmcompressor]"
Requires llmcompressor >= 0.5.1
This will install all necessary dependencies to fine-tune sparsified models using the integration.
-
-Usage
+
+Usage
To enable sparse fine-tuning with this integration, include the plugin in your Axolotl config:
-plugins :
- - axolotl.integrations.llm_compressor.LLMCompressorPlugin
-
-llmcompressor :
- recipe :
- finetuning_stage :
- finetuning_modifiers :
- ConstantPruningModifier :
- targets : [
- 're:.*q_proj.weight' ,
- 're:.*k_proj.weight' ,
- 're:.*v_proj.weight' ,
- 're:.*o_proj.weight' ,
- 're:.*gate_proj.weight' ,
- 're:.*up_proj.weight' ,
- 're:.*down_proj.weight' ,
- ]
- start : 0
- save_compressed : true
+plugins :
+ - axolotl.integrations.llm_compressor.LLMCompressorPlugin
+
+llmcompressor :
+ recipe :
+ finetuning_stage :
+ finetuning_modifiers :
+ ConstantPruningModifier :
+ targets : [
+ 're:.*q_proj.weight' ,
+ 're:.*k_proj.weight' ,
+ 're:.*v_proj.weight' ,
+ 're:.*o_proj.weight' ,
+ 're:.*gate_proj.weight' ,
+ 're:.*up_proj.weight' ,
+ 're:.*down_proj.weight' ,
+ ]
+ start : 0
+ save_compressed : true
This plugin does not apply pruning or sparsification itself — it is intended for fine-tuning models that have already been sparsified .
Pre-sparsified checkpoints can be:
- Generated using LLMCompressor
@@ -838,22 +727,22 @@ By identifying the top n% of layers with the highest SNR, you can optimize train
After fine-tuning your sparse model, you can leverage vLLM for efficient inference.
You can also use LLMCompressor to apply additional quantization to your fine-tuned
sparse model before inference for even greater performance benefits.:
-from vllm import LLM, SamplingParams
-
- prompts = [
- "Hello, my name is" ,
- "The president of the United States is" ,
- "The capital of France is" ,
- "The future of AI is" ,
- ]
- sampling_params = SamplingParams(temperature= 0.8 , top_p= 0.95 )
- llm = LLM("path/to/your/sparse/model" )
- outputs = llm.generate(prompts, sampling_params)
-
-for output in outputs:
- prompt = output.prompt
- generated_text = output.outputs[0 ].text
- print (f"Prompt: { prompt!r} , Generated text: { generated_text!r} " )
+from vllm import LLM, SamplingParams
+
+ prompts = [
+ "Hello, my name is" ,
+ "The president of the United States is" ,
+ "The capital of France is" ,
+ "The future of AI is" ,
+ ]
+ sampling_params = SamplingParams(temperature= 0.8 , top_p= 0.95 )
+ llm = LLM("path/to/your/sparse/model" )
+ outputs = llm.generate(prompts, sampling_params)
+
+for output in outputs:
+ prompt = output.prompt
+ generated_text = output.outputs[0 ].text
+ print (f"Prompt: { prompt!r} , Generated text: { generated_text!r} " )
For more details on vLLM’s capabilities and advanced configuration options, see the official vLLM documentation .
@@ -863,6 +752,126 @@ sparse model before inference for even greater performance benefits.:
Please see reference here
+
+Language Model Evaluation Harness (LM Eval)
+Run evaluation on model using the popular lm-evaluation-harness library.
+See https://github.com/EleutherAI/lm-evaluation-harness
+
+Usage
+plugins :
+ - axolotl.integrations.lm_eval.LMEvalPlugin
+
+lm_eval_tasks :
+ - gsm8k
+ - hellaswag
+ - arc_easy
+
+lm_eval_batch_size : # Batch size for evaluation
+output_dir : # Directory to save evaluation results
+
+
+Citation
+@misc {eval -harness ,
+ author = {Gao, Leo and Tow, Jonathan and Abbasi, Baber and Biderman, Stella and Black, Sid and DiPofi, Anthony and Foster, Charles and Golding, Laurence and Hsu, Jeffrey and Le Noac'h, Alain and Li, Haonan and McDonell, Kyle and Muennighoff, Niklas and Ociepa, Chris and Phang, Jason and Reynolds, Laria and Schoelkopf, Hailey and Skowron, Aviya and Sutawika, Lintang and Tang, Eric and Thite, Anish and Wang, Ben and Wang, Kevin and Zou, Andy},
+ title = {A framework for few-shot language model evaluation},
+ month = 07,
+ year = 2024,
+ publisher = {Zenodo},
+ version = {v0.4.3},
+ doi = {10.5281/zenodo.12608602},
+ url = {https://zenodo.org/records/12608602}
+ }
+Please see reference here
+
+
+
+Liger Kernels
+Liger Kernel provides efficient Triton kernels for LLM training, offering:
+
+20% increase in multi-GPU training throughput
+60% reduction in memory usage
+Compatibility with both FSDP and DeepSpeed
+
+See https://github.com/linkedin/Liger-Kernel
+
+Usage
+plugins :
+ - axolotl.integrations.liger.LigerPlugin
+liger_rope : true
+liger_rms_norm : true
+liger_glu_activation : true
+liger_layer_norm : true
+liger_fused_linear_cross_entropy : true
+
+
+Supported Models
+
+deepseek_v2
+gemma
+gemma2
+gemma3
+granite
+jamba
+llama
+mistral
+mixtral
+mllama
+mllama_text_model
+olmo2
+paligemma
+phi3
+qwen2
+qwen2_5_vl
+qwen2_vl
+
+
+
+Citation
+@article {hsu2024ligerkernelefficienttriton ,
+ title ={Liger Kernel: Efficient Triton Kernels for LLM Training},
+ author ={Pin-Lun Hsu and Yun Dai and Vignesh Kothapalli and Qingquan Song and Shao Tang and Siyu Zhu and Steven Shimizu and Shivam Sahni and Haowen Ning and Yanning Chen},
+ year ={2024},
+ eprint ={2410.10989},
+ archivePrefix ={arXiv},
+ primaryClass ={cs.LG},
+ url ={https://arxiv.org/abs/2410.10989},
+ journal ={arXiv preprint arXiv:2410.10989},
+ }
+Please see reference here
+
+
+
+Spectrum
+by Eric Hartford, Lucas Atkins, Fernando Fernandes, David Golchinfar
+This plugin contains code to freeze the bottom fraction of modules in a model, based on the Signal-to-Noise Ratio (SNR).
+See https://github.com/cognitivecomputations/spectrum
+
+Overview
+Spectrum is a tool for scanning and evaluating the Signal-to-Noise Ratio (SNR) of layers in large language models.
+By identifying the top n% of layers with the highest SNR, you can optimize training efficiency.
+
+
+Usage
+plugins :
+ - axolotl.integrations.spectrum.SpectrumPlugin
+
+spectrum_top_fraction : 0.5
+spectrum_model_name : meta-llama/Meta-Llama-3.1-8B
+
+
+Citation
+@misc {hartford2024spectrumtargetedtrainingsignal ,
+ title ={Spectrum: Targeted Training on Signal to Noise Ratio},
+ author ={Eric Hartford and Lucas Atkins and Fernando Fernandes Neto and David Golchinfar},
+ year ={2024},
+ eprint ={2406.06623},
+ archivePrefix ={arXiv},
+ primaryClass ={cs.LG},
+ url ={https://arxiv.org/abs/2406.06623},
+ }
+Please see reference here
+
+
Adding a new integration
Plugins can be used to customize the behavior of the training pipeline through hooks . See axolotl.integrations.BasePlugin for the possible hooks.
@@ -903,10 +912,10 @@ Warning
If you could not load your integration, please ensure you are pip installing in editable mode.
-
+
and correctly spelled the integration name in the config file.
-
plugins :
- - axolotl.integrations.your_integration_name.YourIntegrationPlugin
+
plugins :
+ - axolotl.integrations.your_integration_name.YourIntegrationPlugin
diff --git a/search.json b/search.json
index 67fbd84a4..4335f4077 100644
--- a/search.json
+++ b/search.json
@@ -3070,6 +3070,17 @@
"Custom Integrations"
]
},
+ {
+ "objectID": "docs/custom_integrations.html#densemixer",
+ "href": "docs/custom_integrations.html#densemixer",
+ "title": "Custom Integrations",
+ "section": "DenseMixer",
+ "text": "DenseMixer\nSee DenseMixer\nSimply add the following to your axolotl YAML config:\nplugins:\n - axolotl.integrations.densemixer.DenseMixerPlugin\nPlease see reference here",
+ "crumbs": [
+ "Advanced Features",
+ "Custom Integrations"
+ ]
+ },
{
"objectID": "docs/custom_integrations.html#grokfast",
"href": "docs/custom_integrations.html#grokfast",
@@ -3093,11 +3104,11 @@
]
},
{
- "objectID": "docs/custom_integrations.html#liger-kernels",
- "href": "docs/custom_integrations.html#liger-kernels",
+ "objectID": "docs/custom_integrations.html#llmcompressor",
+ "href": "docs/custom_integrations.html#llmcompressor",
"title": "Custom Integrations",
- "section": "Liger Kernels",
- "text": "Liger Kernels\nLiger Kernel provides efficient Triton kernels for LLM training, offering:\n\n20% increase in multi-GPU training throughput\n60% reduction in memory usage\nCompatibility with both FSDP and DeepSpeed\n\nSee https://github.com/linkedin/Liger-Kernel\n\nUsage\nplugins:\n - axolotl.integrations.liger.LigerPlugin\nliger_rope: true\nliger_rms_norm: true\nliger_glu_activation: true\nliger_layer_norm: true\nliger_fused_linear_cross_entropy: true\n\n\nSupported Models\n\ndeepseek_v2\ngemma\ngemma2\ngemma3\ngranite\njamba\nllama\nmistral\nmixtral\nmllama\nmllama_text_model\nolmo2\npaligemma\nphi3\nqwen2\nqwen2_5_vl\nqwen2_vl\n\n\n\nCitation\n@article{hsu2024ligerkernelefficienttriton,\n title={Liger Kernel: Efficient Triton Kernels for LLM Training},\n author={Pin-Lun Hsu and Yun Dai and Vignesh Kothapalli and Qingquan Song and Shao Tang and Siyu Zhu and Steven Shimizu and Shivam Sahni and Haowen Ning and Yanning Chen},\n year={2024},\n eprint={2410.10989},\n archivePrefix={arXiv},\n primaryClass={cs.LG},\n url={https://arxiv.org/abs/2410.10989},\n journal={arXiv preprint arXiv:2410.10989},\n}\nPlease see reference here",
+ "section": "LLMCompressor",
+ "text": "LLMCompressor\nFine-tune sparsified models in Axolotl using Neural Magic’s LLMCompressor.\nThis integration enables fine-tuning of models sparsified using LLMCompressor within the Axolotl training framework. By combining LLMCompressor’s model compression capabilities with Axolotl’s distributed training pipelines, users can efficiently fine-tune sparse models at scale.\nIt uses Axolotl’s plugin system to hook into the fine-tuning flows while maintaining sparsity throughout training.\n\n\nRequirements\n\nAxolotl with llmcompressor extras:\npip install \"axolotl[llmcompressor]\"\nRequires llmcompressor >= 0.5.1\n\nThis will install all necessary dependencies to fine-tune sparsified models using the integration.\n\n\n\nUsage\nTo enable sparse fine-tuning with this integration, include the plugin in your Axolotl config:\nplugins:\n - axolotl.integrations.llm_compressor.LLMCompressorPlugin\n\nllmcompressor:\n recipe:\n finetuning_stage:\n finetuning_modifiers:\n ConstantPruningModifier:\n targets: [\n 're:.*q_proj.weight',\n 're:.*k_proj.weight',\n 're:.*v_proj.weight',\n 're:.*o_proj.weight',\n 're:.*gate_proj.weight',\n 're:.*up_proj.weight',\n 're:.*down_proj.weight',\n ]\n start: 0\n save_compressed: true\nThis plugin does not apply pruning or sparsification itself — it is intended for fine-tuning models that have already been sparsified.\nPre-sparsified checkpoints can be:\n- Generated using LLMCompressor\n- Downloaded from Neural Magic’s Hugging Face page\n- Any custom LLM with compatible sparsity patterns that you’ve created yourself\nTo learn more about writing and customizing LLMCompressor recipes, refer to the official documentation:\nhttps://github.com/vllm-project/llm-compressor/blob/main/README.md\n\n\nStorage Optimization with save_compressed\nSetting save_compressed: true in your configuration enables saving models in a compressed format, which:\n- Reduces disk space usage by approximately 40%\n- Maintains compatibility with vLLM for accelerated inference\n- Maintains compatibility with llmcompressor for further optimization (example: quantization)\nThis option is highly recommended when working with sparse models to maximize the benefits of model compression.\n\n\nExample Config\nSee examples/llama-3/sparse-finetuning.yaml for a complete example.\n\n\n\nInference with vLLM\nAfter fine-tuning your sparse model, you can leverage vLLM for efficient inference.\nYou can also use LLMCompressor to apply additional quantization to your fine-tuned\nsparse model before inference for even greater performance benefits.:\nfrom vllm import LLM, SamplingParams\n\nprompts = [\n \"Hello, my name is\",\n \"The president of the United States is\",\n \"The capital of France is\",\n \"The future of AI is\",\n]\nsampling_params = SamplingParams(temperature=0.8, top_p=0.95)\nllm = LLM(\"path/to/your/sparse/model\")\noutputs = llm.generate(prompts, sampling_params)\n\nfor output in outputs:\n prompt = output.prompt\n generated_text = output.outputs[0].text\n print(f\"Prompt: {prompt!r}, Generated text: {generated_text!r}\")\nFor more details on vLLM’s capabilities and advanced configuration options, see the official vLLM documentation.\n\n\nLearn More\nFor details on available sparsity and quantization schemes, fine-tuning recipes, and usage examples, visit the official LLMCompressor repository:\nhttps://github.com/vllm-project/llm-compressor\nPlease see reference here",
"crumbs": [
"Advanced Features",
"Custom Integrations"
@@ -3115,22 +3126,22 @@
]
},
{
- "objectID": "docs/custom_integrations.html#spectrum",
- "href": "docs/custom_integrations.html#spectrum",
+ "objectID": "docs/custom_integrations.html#liger-kernels",
+ "href": "docs/custom_integrations.html#liger-kernels",
"title": "Custom Integrations",
- "section": "Spectrum",
- "text": "Spectrum\nby Eric Hartford, Lucas Atkins, Fernando Fernandes, David Golchinfar\nThis plugin contains code to freeze the bottom fraction of modules in a model, based on the Signal-to-Noise Ratio (SNR).\nSee https://github.com/cognitivecomputations/spectrum\n\nOverview\nSpectrum is a tool for scanning and evaluating the Signal-to-Noise Ratio (SNR) of layers in large language models.\nBy identifying the top n% of layers with the highest SNR, you can optimize training efficiency.\n\n\nUsage\nplugins:\n - axolotl.integrations.spectrum.SpectrumPlugin\n\nspectrum_top_fraction: 0.5\nspectrum_model_name: meta-llama/Meta-Llama-3.1-8B\n\n\nCitation\n@misc{hartford2024spectrumtargetedtrainingsignal,\n title={Spectrum: Targeted Training on Signal to Noise Ratio},\n author={Eric Hartford and Lucas Atkins and Fernando Fernandes Neto and David Golchinfar},\n year={2024},\n eprint={2406.06623},\n archivePrefix={arXiv},\n primaryClass={cs.LG},\n url={https://arxiv.org/abs/2406.06623},\n}\nPlease see reference here",
+ "section": "Liger Kernels",
+ "text": "Liger Kernels\nLiger Kernel provides efficient Triton kernels for LLM training, offering:\n\n20% increase in multi-GPU training throughput\n60% reduction in memory usage\nCompatibility with both FSDP and DeepSpeed\n\nSee https://github.com/linkedin/Liger-Kernel\n\nUsage\nplugins:\n - axolotl.integrations.liger.LigerPlugin\nliger_rope: true\nliger_rms_norm: true\nliger_glu_activation: true\nliger_layer_norm: true\nliger_fused_linear_cross_entropy: true\n\n\nSupported Models\n\ndeepseek_v2\ngemma\ngemma2\ngemma3\ngranite\njamba\nllama\nmistral\nmixtral\nmllama\nmllama_text_model\nolmo2\npaligemma\nphi3\nqwen2\nqwen2_5_vl\nqwen2_vl\n\n\n\nCitation\n@article{hsu2024ligerkernelefficienttriton,\n title={Liger Kernel: Efficient Triton Kernels for LLM Training},\n author={Pin-Lun Hsu and Yun Dai and Vignesh Kothapalli and Qingquan Song and Shao Tang and Siyu Zhu and Steven Shimizu and Shivam Sahni and Haowen Ning and Yanning Chen},\n year={2024},\n eprint={2410.10989},\n archivePrefix={arXiv},\n primaryClass={cs.LG},\n url={https://arxiv.org/abs/2410.10989},\n journal={arXiv preprint arXiv:2410.10989},\n}\nPlease see reference here",
"crumbs": [
"Advanced Features",
"Custom Integrations"
]
},
{
- "objectID": "docs/custom_integrations.html#llmcompressor",
- "href": "docs/custom_integrations.html#llmcompressor",
+ "objectID": "docs/custom_integrations.html#spectrum",
+ "href": "docs/custom_integrations.html#spectrum",
"title": "Custom Integrations",
- "section": "LLMCompressor",
- "text": "LLMCompressor\nFine-tune sparsified models in Axolotl using Neural Magic’s LLMCompressor.\nThis integration enables fine-tuning of models sparsified using LLMCompressor within the Axolotl training framework. By combining LLMCompressor’s model compression capabilities with Axolotl’s distributed training pipelines, users can efficiently fine-tune sparse models at scale.\nIt uses Axolotl’s plugin system to hook into the fine-tuning flows while maintaining sparsity throughout training.\n\n\nRequirements\n\nAxolotl with llmcompressor extras:\npip install \"axolotl[llmcompressor]\"\nRequires llmcompressor >= 0.5.1\n\nThis will install all necessary dependencies to fine-tune sparsified models using the integration.\n\n\n\nUsage\nTo enable sparse fine-tuning with this integration, include the plugin in your Axolotl config:\nplugins:\n - axolotl.integrations.llm_compressor.LLMCompressorPlugin\n\nllmcompressor:\n recipe:\n finetuning_stage:\n finetuning_modifiers:\n ConstantPruningModifier:\n targets: [\n 're:.*q_proj.weight',\n 're:.*k_proj.weight',\n 're:.*v_proj.weight',\n 're:.*o_proj.weight',\n 're:.*gate_proj.weight',\n 're:.*up_proj.weight',\n 're:.*down_proj.weight',\n ]\n start: 0\n save_compressed: true\nThis plugin does not apply pruning or sparsification itself — it is intended for fine-tuning models that have already been sparsified.\nPre-sparsified checkpoints can be:\n- Generated using LLMCompressor\n- Downloaded from Neural Magic’s Hugging Face page\n- Any custom LLM with compatible sparsity patterns that you’ve created yourself\nTo learn more about writing and customizing LLMCompressor recipes, refer to the official documentation:\nhttps://github.com/vllm-project/llm-compressor/blob/main/README.md\n\n\nStorage Optimization with save_compressed\nSetting save_compressed: true in your configuration enables saving models in a compressed format, which:\n- Reduces disk space usage by approximately 40%\n- Maintains compatibility with vLLM for accelerated inference\n- Maintains compatibility with llmcompressor for further optimization (example: quantization)\nThis option is highly recommended when working with sparse models to maximize the benefits of model compression.\n\n\nExample Config\nSee examples/llama-3/sparse-finetuning.yaml for a complete example.\n\n\n\nInference with vLLM\nAfter fine-tuning your sparse model, you can leverage vLLM for efficient inference.\nYou can also use LLMCompressor to apply additional quantization to your fine-tuned\nsparse model before inference for even greater performance benefits.:\nfrom vllm import LLM, SamplingParams\n\nprompts = [\n \"Hello, my name is\",\n \"The president of the United States is\",\n \"The capital of France is\",\n \"The future of AI is\",\n]\nsampling_params = SamplingParams(temperature=0.8, top_p=0.95)\nllm = LLM(\"path/to/your/sparse/model\")\noutputs = llm.generate(prompts, sampling_params)\n\nfor output in outputs:\n prompt = output.prompt\n generated_text = output.outputs[0].text\n print(f\"Prompt: {prompt!r}, Generated text: {generated_text!r}\")\nFor more details on vLLM’s capabilities and advanced configuration options, see the official vLLM documentation.\n\n\nLearn More\nFor details on available sparsity and quantization schemes, fine-tuning recipes, and usage examples, visit the official LLMCompressor repository:\nhttps://github.com/vllm-project/llm-compressor\nPlease see reference here",
+ "section": "Spectrum",
+ "text": "Spectrum\nby Eric Hartford, Lucas Atkins, Fernando Fernandes, David Golchinfar\nThis plugin contains code to freeze the bottom fraction of modules in a model, based on the Signal-to-Noise Ratio (SNR).\nSee https://github.com/cognitivecomputations/spectrum\n\nOverview\nSpectrum is a tool for scanning and evaluating the Signal-to-Noise Ratio (SNR) of layers in large language models.\nBy identifying the top n% of layers with the highest SNR, you can optimize training efficiency.\n\n\nUsage\nplugins:\n - axolotl.integrations.spectrum.SpectrumPlugin\n\nspectrum_top_fraction: 0.5\nspectrum_model_name: meta-llama/Meta-Llama-3.1-8B\n\n\nCitation\n@misc{hartford2024spectrumtargetedtrainingsignal,\n title={Spectrum: Targeted Training on Signal to Noise Ratio},\n author={Eric Hartford and Lucas Atkins and Fernando Fernandes Neto and David Golchinfar},\n year={2024},\n eprint={2406.06623},\n archivePrefix={arXiv},\n primaryClass={cs.LG},\n url={https://arxiv.org/abs/2406.06623},\n}\nPlease see reference here",
"crumbs": [
"Advanced Features",
"Custom Integrations"
diff --git a/sitemap.xml b/sitemap.xml
index ac719c6f1..e3a39aca3 100644
--- a/sitemap.xml
+++ b/sitemap.xml
@@ -2,758 +2,758 @@
https://docs.axolotl.ai/docs/unsloth.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.508Z
https://docs.axolotl.ai/docs/dataset-formats/conversation.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.502Z
https://docs.axolotl.ai/docs/dataset-formats/stepwise_supervised.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.502Z
https://docs.axolotl.ai/docs/dataset-formats/tokenized.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.503Z
https://docs.axolotl.ai/docs/mac.html
- 2025-07-07T19:24:01.498Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/nccl.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/multi-node.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/docker.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.503Z
https://docs.axolotl.ai/docs/lr_groups.html
- 2025-07-07T19:24:01.498Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/inference.html
- 2025-07-07T19:24:01.498Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/cli.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.502Z
https://docs.axolotl.ai/docs/config-reference.html
- 2025-07-07T19:27:44.889Z
+ 2025-07-07T21:08:49.663Z
https://docs.axolotl.ai/docs/multi-gpu.html
- 2025-07-07T19:24:01.498Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/debugging.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.503Z
https://docs.axolotl.ai/docs/multimodal.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/api/cli.sweeps.html
- 2025-07-07T19:27:30.521Z
+ 2025-07-07T21:08:36.131Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.llama3.html
- 2025-07-07T19:27:30.850Z
+ 2025-07-07T21:08:36.457Z
https://docs.axolotl.ai/docs/api/utils.schedulers.html
- 2025-07-07T19:27:31.238Z
+ 2025-07-07T21:08:36.840Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_xformers.html
- 2025-07-07T19:27:31.052Z
+ 2025-07-07T21:08:36.655Z
https://docs.axolotl.ai/docs/api/cli.cloud.modal_.html
- 2025-07-07T19:27:30.570Z
+ 2025-07-07T21:08:36.180Z
https://docs.axolotl.ai/docs/api/kernels.geglu.html
- 2025-07-07T19:27:31.004Z
+ 2025-07-07T21:08:36.609Z
https://docs.axolotl.ai/docs/api/core.trainers.utils.html
- 2025-07-07T19:27:30.643Z
+ 2025-07-07T21:08:36.252Z
https://docs.axolotl.ai/docs/api/core.datasets.chat.html
- 2025-07-07T19:27:30.369Z
+ 2025-07-07T21:08:35.983Z
https://docs.axolotl.ai/docs/api/utils.schemas.peft.html
- 2025-07-07T19:27:31.353Z
+ 2025-07-07T21:08:36.954Z
https://docs.axolotl.ai/docs/api/monkeypatch.btlm_attn_hijack_flash.html
- 2025-07-07T19:27:31.115Z
+ 2025-07-07T21:08:36.718Z
https://docs.axolotl.ai/docs/api/prompt_strategies.chat_template.html
- 2025-07-07T19:27:30.752Z
+ 2025-07-07T21:08:36.360Z
https://docs.axolotl.ai/docs/api/prompt_strategies.kto.user_defined.html
- 2025-07-07T19:27:30.883Z
+ 2025-07-07T21:08:36.490Z
https://docs.axolotl.ai/docs/api/cli.cloud.base.html
- 2025-07-07T19:27:30.563Z
+ 2025-07-07T21:08:36.173Z
https://docs.axolotl.ai/docs/api/kernels.swiglu.html
- 2025-07-07T19:27:31.014Z
+ 2025-07-07T21:08:36.619Z
https://docs.axolotl.ai/docs/api/prompt_strategies.stepwise_supervised.html
- 2025-07-07T19:27:30.817Z
+ 2025-07-07T21:08:36.424Z
https://docs.axolotl.ai/docs/api/prompt_strategies.bradley_terry.llama3.html
- 2025-07-07T19:27:30.908Z
+ 2025-07-07T21:08:36.514Z
https://docs.axolotl.ai/docs/api/prompt_strategies.completion.html
- 2025-07-07T19:27:30.806Z
+ 2025-07-07T21:08:36.414Z
https://docs.axolotl.ai/docs/api/kernels.utils.html
- 2025-07-07T19:27:31.024Z
+ 2025-07-07T21:08:36.628Z
https://docs.axolotl.ai/docs/api/common.datasets.html
- 2025-07-07T19:27:31.568Z
+ 2025-07-07T21:08:37.165Z
https://docs.axolotl.ai/docs/api/utils.schemas.datasets.html
- 2025-07-07T19:27:31.344Z
+ 2025-07-07T21:08:36.945Z
https://docs.axolotl.ai/docs/api/core.builders.rl.html
- 2025-07-07T19:27:30.324Z
+ 2025-07-07T21:08:35.938Z
https://docs.axolotl.ai/docs/api/evaluate.html
- 2025-07-07T19:27:30.228Z
+ 2025-07-07T21:08:35.843Z
https://docs.axolotl.ai/docs/api/kernels.quantize.html
- 2025-07-07T19:27:31.022Z
+ 2025-07-07T21:08:36.626Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_flash.html
- 2025-07-07T19:27:31.050Z
+ 2025-07-07T21:08:36.654Z
https://docs.axolotl.ai/docs/api/core.trainers.mixins.rng_state_loader.html
- 2025-07-07T19:27:30.687Z
+ 2025-07-07T21:08:36.295Z
https://docs.axolotl.ai/docs/api/integrations.base.html
- 2025-07-07T19:27:31.528Z
+ 2025-07-07T21:08:37.125Z
https://docs.axolotl.ai/docs/api/cli.merge_lora.html
- 2025-07-07T19:27:30.494Z
+ 2025-07-07T21:08:36.105Z
https://docs.axolotl.ai/docs/api/cli.merge_sharded_fsdp_weights.html
- 2025-07-07T19:27:30.506Z
+ 2025-07-07T21:08:36.117Z
https://docs.axolotl.ai/docs/api/monkeypatch.transformers_fa_utils.html
- 2025-07-07T19:27:31.132Z
+ 2025-07-07T21:08:36.734Z
https://docs.axolotl.ai/docs/api/prompt_strategies.llama2_chat.html
- 2025-07-07T19:27:30.800Z
+ 2025-07-07T21:08:36.408Z
https://docs.axolotl.ai/docs/api/utils.collators.mm_chat.html
- 2025-07-07T19:27:31.598Z
+ 2025-07-07T21:08:37.195Z
https://docs.axolotl.ai/docs/api/utils.data.sft.html
- 2025-07-07T19:27:31.279Z
+ 2025-07-07T21:08:36.880Z
https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_instruct.html
- 2025-07-07T19:27:30.768Z
+ 2025-07-07T21:08:36.375Z
https://docs.axolotl.ai/docs/api/integrations.liger.args.html
- 2025-07-07T19:27:31.543Z
+ 2025-07-07T21:08:37.141Z
https://docs.axolotl.ai/docs/api/monkeypatch.mistral_attn_hijack_flash.html
- 2025-07-07T19:27:31.067Z
+ 2025-07-07T21:08:36.669Z
https://docs.axolotl.ai/docs/api/cli.vllm_serve.html
- 2025-07-07T19:27:30.560Z
+ 2025-07-07T21:08:36.170Z
https://docs.axolotl.ai/docs/api/monkeypatch.utils.html
- 2025-07-07T19:27:31.113Z
+ 2025-07-07T21:08:36.716Z
https://docs.axolotl.ai/docs/api/loaders.patch_manager.html
- 2025-07-07T19:27:30.677Z
+ 2025-07-07T21:08:36.285Z
https://docs.axolotl.ai/docs/api/utils.schemas.integrations.html
- 2025-07-07T19:27:31.374Z
+ 2025-07-07T21:08:36.975Z
https://docs.axolotl.ai/docs/api/utils.callbacks.perplexity.html
- 2025-07-07T19:27:31.646Z
+ 2025-07-07T21:08:37.242Z
https://docs.axolotl.ai/docs/api/cli.utils.html
- 2025-07-07T19:27:30.553Z
+ 2025-07-07T21:08:36.163Z
https://docs.axolotl.ai/docs/api/utils.schemas.config.html
- 2025-07-07T19:27:31.314Z
+ 2025-07-07T21:08:36.915Z
https://docs.axolotl.ai/docs/api/prompt_strategies.input_output.html
- 2025-07-07T19:27:30.812Z
+ 2025-07-07T21:08:36.420Z
https://docs.axolotl.ai/docs/api/utils.distributed.html
- 2025-07-07T19:27:31.258Z
+ 2025-07-07T21:08:36.861Z
https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_disk.html
- 2025-07-07T19:27:31.165Z
+ 2025-07-07T21:08:36.768Z
https://docs.axolotl.ai/docs/api/monkeypatch.trainer_fsdp_optim.html
- 2025-07-07T19:27:31.125Z
+ 2025-07-07T21:08:36.728Z
https://docs.axolotl.ai/docs/api/core.builders.base.html
- 2025-07-07T19:27:30.311Z
+ 2025-07-07T21:08:35.925Z
https://docs.axolotl.ai/docs/api/core.trainers.trl.html
- 2025-07-07T19:27:30.602Z
+ 2025-07-07T21:08:36.211Z
https://docs.axolotl.ai/docs/api/cli.evaluate.html
- 2025-07-07T19:27:30.425Z
+ 2025-07-07T21:08:36.038Z
https://docs.axolotl.ai/docs/api/utils.optimizers.adopt.html
- 2025-07-07T19:27:31.270Z
+ 2025-07-07T21:08:36.872Z
https://docs.axolotl.ai/docs/api/utils.callbacks.qat.html
- 2025-07-07T19:27:31.665Z
+ 2025-07-07T21:08:37.261Z
https://docs.axolotl.ai/docs/api/core.trainers.dpo.trainer.html
- 2025-07-07T19:27:30.618Z
+ 2025-07-07T21:08:36.228Z
https://docs.axolotl.ai/docs/api/core.chat.format.shared.html
- 2025-07-07T19:27:30.364Z
+ 2025-07-07T21:08:35.978Z
https://docs.axolotl.ai/docs/api/monkeypatch.relora.html
- 2025-07-07T19:27:31.075Z
+ 2025-07-07T21:08:36.678Z
https://docs.axolotl.ai/docs/api/cli.config.html
- 2025-07-07T19:27:30.471Z
+ 2025-07-07T21:08:36.082Z
https://docs.axolotl.ai/docs/api/cli.preprocess.html
- 2025-07-07T19:27:30.515Z
+ 2025-07-07T21:08:36.125Z
https://docs.axolotl.ai/docs/api/core.trainers.base.html
- 2025-07-07T19:27:30.585Z
+ 2025-07-07T21:08:36.195Z
https://docs.axolotl.ai/docs/api/convert.html
- 2025-07-07T19:27:30.253Z
+ 2025-07-07T21:08:35.867Z
https://docs.axolotl.ai/docs/api/prompt_strategies.pygmalion.html
- 2025-07-07T19:27:30.834Z
+ 2025-07-07T21:08:36.441Z
https://docs.axolotl.ai/docs/api/utils.schemas.trl.html
- 2025-07-07T19:27:31.357Z
+ 2025-07-07T21:08:36.957Z
https://docs.axolotl.ai/docs/api/cli.args.html
- 2025-07-07T19:27:30.445Z
+ 2025-07-07T21:08:36.058Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chat_template.html
- 2025-07-07T19:27:30.840Z
+ 2025-07-07T21:08:36.447Z
https://docs.axolotl.ai/docs/api/loaders.constants.html
- 2025-07-07T19:27:30.678Z
+ 2025-07-07T21:08:36.287Z
https://docs.axolotl.ai/docs/api/logging_config.html
- 2025-07-07T19:27:30.305Z
+ 2025-07-07T21:08:35.919Z
https://docs.axolotl.ai/docs/api/cli.inference.html
- 2025-07-07T19:27:30.485Z
+ 2025-07-07T21:08:36.096Z
https://docs.axolotl.ai/docs/api/utils.ctx_managers.sequence_parallel.html
- 2025-07-07T19:27:30.718Z
+ 2025-07-07T21:08:36.326Z
https://docs.axolotl.ai/docs/api/integrations.spectrum.args.html
- 2025-07-07T19:27:31.550Z
+ 2025-07-07T21:08:37.147Z
https://docs.axolotl.ai/docs/api/utils.schemas.training.html
- 2025-07-07T19:27:31.326Z
+ 2025-07-07T21:08:36.927Z
https://docs.axolotl.ai/docs/api/prompt_strategies.orcamini.html
- 2025-07-07T19:27:30.828Z
+ 2025-07-07T21:08:36.435Z
https://docs.axolotl.ai/docs/api/utils.freeze.html
- 2025-07-07T19:27:31.196Z
+ 2025-07-07T21:08:36.798Z
https://docs.axolotl.ai/docs/api/loaders.tokenizer.html
- 2025-07-07T19:27:30.662Z
+ 2025-07-07T21:08:36.270Z
https://docs.axolotl.ai/docs/api/utils.bench.html
- 2025-07-07T19:27:31.188Z
+ 2025-07-07T21:08:36.790Z
https://docs.axolotl.ai/docs/api/utils.quantization.html
- 2025-07-07T19:27:31.300Z
+ 2025-07-07T21:08:36.901Z
https://docs.axolotl.ai/docs/batch_vs_grad.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.502Z
https://docs.axolotl.ai/docs/input_output.html
- 2025-07-07T19:24:01.498Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/sequence_parallelism.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.508Z
https://docs.axolotl.ai/docs/reward_modelling.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.508Z
https://docs.axolotl.ai/index.html
- 2025-07-07T19:24:01.514Z
+ 2025-07-07T21:05:27.525Z
https://docs.axolotl.ai/src/axolotl/integrations/LICENSE.html
- 2025-07-07T19:24:01.518Z
+ 2025-07-07T21:05:27.529Z
https://docs.axolotl.ai/FAQS.html
- 2025-07-07T19:24:01.493Z
+ 2025-07-07T21:05:27.500Z
https://docs.axolotl.ai/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html
- 2025-07-07T19:24:01.518Z
+ 2025-07-07T21:05:27.529Z
https://docs.axolotl.ai/TODO.html
- 2025-07-07T19:24:01.493Z
+ 2025-07-07T21:05:27.500Z
https://docs.axolotl.ai/examples/colab-notebooks/colab-axolotl-example.html
- 2025-07-07T19:24:01.501Z
+ 2025-07-07T21:05:27.510Z
https://docs.axolotl.ai/docs/torchao.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.508Z
https://docs.axolotl.ai/docs/ray-integration.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.508Z
https://docs.axolotl.ai/docs/quantize.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/qat.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/api/utils.lora.html
- 2025-07-07T19:27:31.179Z
+ 2025-07-07T21:08:36.781Z
https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_w_system.html
- 2025-07-07T19:27:30.780Z
+ 2025-07-07T21:08:36.387Z
https://docs.axolotl.ai/docs/api/monkeypatch.stablelm_attn_hijack_flash.html
- 2025-07-07T19:27:31.122Z
+ 2025-07-07T21:08:36.725Z
https://docs.axolotl.ai/docs/api/utils.collators.core.html
- 2025-07-07T19:27:31.570Z
+ 2025-07-07T21:08:37.168Z
https://docs.axolotl.ai/docs/api/prompt_strategies.metharme.html
- 2025-07-07T19:27:30.824Z
+ 2025-07-07T21:08:36.431Z
https://docs.axolotl.ai/docs/api/utils.callbacks.profiler.html
- 2025-07-07T19:27:31.650Z
+ 2025-07-07T21:08:37.245Z
https://docs.axolotl.ai/docs/api/utils.data.pretraining.html
- 2025-07-07T19:27:31.271Z
+ 2025-07-07T21:08:36.873Z
https://docs.axolotl.ai/docs/api/utils.callbacks.lisa.html
- 2025-07-07T19:27:31.651Z
+ 2025-07-07T21:08:37.247Z
https://docs.axolotl.ai/docs/api/utils.trainer.html
- 2025-07-07T19:27:31.213Z
+ 2025-07-07T21:08:36.816Z
https://docs.axolotl.ai/docs/api/integrations.cut_cross_entropy.args.html
- 2025-07-07T19:27:31.531Z
+ 2025-07-07T21:08:37.129Z
https://docs.axolotl.ai/docs/api/utils.schemas.model.html
- 2025-07-07T19:27:31.321Z
+ 2025-07-07T21:08:36.922Z
https://docs.axolotl.ai/docs/api/monkeypatch.data.batch_dataset_fetcher.html
- 2025-07-07T19:27:31.135Z
+ 2025-07-07T21:08:36.737Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.zephyr.html
- 2025-07-07T19:27:30.862Z
+ 2025-07-07T21:08:36.469Z
https://docs.axolotl.ai/docs/api/datasets.html
- 2025-07-07T19:27:30.239Z
+ 2025-07-07T21:08:35.854Z
https://docs.axolotl.ai/docs/api/utils.schemas.enums.html
- 2025-07-07T19:27:31.385Z
+ 2025-07-07T21:08:36.985Z
https://docs.axolotl.ai/docs/api/integrations.kd.trainer.html
- 2025-07-07T19:27:31.540Z
+ 2025-07-07T21:08:37.137Z
https://docs.axolotl.ai/docs/api/monkeypatch.lora_kernels.html
- 2025-07-07T19:27:31.105Z
+ 2025-07-07T21:08:36.708Z
https://docs.axolotl.ai/docs/api/utils.collators.batching.html
- 2025-07-07T19:27:31.590Z
+ 2025-07-07T21:08:37.187Z
https://docs.axolotl.ai/docs/api/core.trainers.grpo.sampler.html
- 2025-07-07T19:27:30.642Z
+ 2025-07-07T21:08:36.251Z
https://docs.axolotl.ai/docs/api/prompt_strategies.base.html
- 2025-07-07T19:27:30.720Z
+ 2025-07-07T21:08:36.327Z
https://docs.axolotl.ai/docs/api/monkeypatch.multipack.html
- 2025-07-07T19:27:31.068Z
+ 2025-07-07T21:08:36.671Z
https://docs.axolotl.ai/docs/api/prompt_strategies.orpo.chat_template.html
- 2025-07-07T19:27:30.904Z
+ 2025-07-07T21:08:36.510Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.passthrough.html
- 2025-07-07T19:27:30.865Z
+ 2025-07-07T21:08:36.472Z
https://docs.axolotl.ai/docs/api/core.chat.format.chatml.html
- 2025-07-07T19:27:30.361Z
+ 2025-07-07T21:08:35.975Z
https://docs.axolotl.ai/docs/api/core.trainers.mixins.scheduler.html
- 2025-07-07T19:27:30.694Z
+ 2025-07-07T21:08:36.302Z
https://docs.axolotl.ai/docs/api/utils.model_shard_quant.html
- 2025-07-07T19:27:31.185Z
+ 2025-07-07T21:08:36.787Z
https://docs.axolotl.ai/docs/api/prompt_strategies.kto.chatml.html
- 2025-07-07T19:27:30.882Z
+ 2025-07-07T21:08:36.488Z
https://docs.axolotl.ai/docs/api/utils.tokenization.html
- 2025-07-07T19:27:31.172Z
+ 2025-07-07T21:08:36.775Z
https://docs.axolotl.ai/docs/api/loaders.model.html
- 2025-07-07T19:27:30.654Z
+ 2025-07-07T21:08:36.262Z
https://docs.axolotl.ai/docs/api/utils.callbacks.mlflow_.html
- 2025-07-07T19:27:31.655Z
+ 2025-07-07T21:08:37.250Z
https://docs.axolotl.ai/docs/api/core.trainers.grpo.trainer.html
- 2025-07-07T19:27:30.629Z
+ 2025-07-07T21:08:36.239Z
https://docs.axolotl.ai/docs/api/cli.main.html
- 2025-07-07T19:27:30.408Z
+ 2025-07-07T21:08:36.022Z
https://docs.axolotl.ai/docs/api/utils.callbacks.comet_.html
- 2025-07-07T19:27:31.658Z
+ 2025-07-07T21:08:37.254Z
https://docs.axolotl.ai/docs/api/utils.chat_templates.html
- 2025-07-07T19:27:31.174Z
+ 2025-07-07T21:08:36.776Z
https://docs.axolotl.ai/docs/api/utils.schemas.utils.html
- 2025-07-07T19:27:31.391Z
+ 2025-07-07T21:08:36.991Z
https://docs.axolotl.ai/docs/api/common.architectures.html
- 2025-07-07T19:27:31.551Z
+ 2025-07-07T21:08:37.149Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_expand_mask.html
- 2025-07-07T19:27:31.076Z
+ 2025-07-07T21:08:36.679Z
https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_chat.html
- 2025-07-07T19:27:30.766Z
+ 2025-07-07T21:08:36.374Z
https://docs.axolotl.ai/docs/api/utils.samplers.multipack.html
- 2025-07-07T19:27:31.640Z
+ 2025-07-07T21:08:37.236Z
https://docs.axolotl.ai/docs/api/integrations.grokfast.optimizer.html
- 2025-07-07T19:27:31.532Z
+ 2025-07-07T21:08:37.130Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chatml.html
- 2025-07-07T19:27:30.861Z
+ 2025-07-07T21:08:36.467Z
https://docs.axolotl.ai/docs/api/monkeypatch.mixtral.html
- 2025-07-07T19:27:31.136Z
+ 2025-07-07T21:08:36.739Z
https://docs.axolotl.ai/docs/api/train.html
- 2025-07-07T19:27:30.218Z
+ 2025-07-07T21:08:35.833Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_patch_multipack.html
- 2025-07-07T19:27:31.116Z
+ 2025-07-07T21:08:36.719Z
https://docs.axolotl.ai/docs/api/index.html
- 2025-07-07T19:27:30.154Z
+ 2025-07-07T21:08:35.771Z
https://docs.axolotl.ai/docs/api/loaders.adapter.html
- 2025-07-07T19:27:30.669Z
+ 2025-07-07T21:08:36.277Z
https://docs.axolotl.ai/docs/api/utils.schemas.multimodal.html
- 2025-07-07T19:27:31.362Z
+ 2025-07-07T21:08:36.963Z
https://docs.axolotl.ai/docs/api/kernels.lora.html
- 2025-07-07T19:27:30.993Z
+ 2025-07-07T21:08:36.598Z
https://docs.axolotl.ai/docs/api/prompt_strategies.kto.llama3.html
- 2025-07-07T19:27:30.873Z
+ 2025-07-07T21:08:36.480Z
https://docs.axolotl.ai/docs/api/cli.checks.html
- 2025-07-07T19:27:30.452Z
+ 2025-07-07T21:08:36.065Z
https://docs.axolotl.ai/docs/api/cli.quantize.html
- 2025-07-07T19:27:30.575Z
+ 2025-07-07T21:08:36.185Z
https://docs.axolotl.ai/docs/api/integrations.lm_eval.args.html
- 2025-07-07T19:27:31.546Z
+ 2025-07-07T21:08:37.144Z
https://docs.axolotl.ai/docs/api/core.chat.messages.html
- 2025-07-07T19:27:30.360Z
+ 2025-07-07T21:08:35.974Z
https://docs.axolotl.ai/docs/api/core.builders.causal.html
- 2025-07-07T19:27:30.316Z
+ 2025-07-07T21:08:35.930Z
https://docs.axolotl.ai/docs/api/core.trainers.relora.html
- 2025-07-07T19:27:30.612Z
+ 2025-07-07T21:08:36.221Z
https://docs.axolotl.ai/docs/api/models.mamba.modeling_mamba.html
- 2025-07-07T19:27:31.569Z
+ 2025-07-07T21:08:37.166Z
https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_cpu.html
- 2025-07-07T19:27:31.139Z
+ 2025-07-07T21:08:36.742Z
https://docs.axolotl.ai/docs/api/core.trainers.mamba.html
- 2025-07-07T19:27:30.607Z
+ 2025-07-07T21:08:36.217Z
https://docs.axolotl.ai/docs/api/core.datasets.transforms.chat_builder.html
- 2025-07-07T19:27:30.377Z
+ 2025-07-07T21:08:35.991Z
https://docs.axolotl.ai/docs/api/loaders.processor.html
- 2025-07-07T19:27:30.663Z
+ 2025-07-07T21:08:36.272Z
https://docs.axolotl.ai/docs/api/core.chat.format.llama3x.html
- 2025-07-07T19:27:30.363Z
+ 2025-07-07T21:08:35.976Z
https://docs.axolotl.ai/docs/api/prompt_strategies.messages.chat.html
- 2025-07-07T19:27:30.839Z
+ 2025-07-07T21:08:36.445Z
https://docs.axolotl.ai/docs/api/cli.train.html
- 2025-07-07T19:27:30.417Z
+ 2025-07-07T21:08:36.030Z
https://docs.axolotl.ai/docs/api/core.trainers.mixins.optimizer.html
- 2025-07-07T19:27:30.684Z
+ 2025-07-07T21:08:36.292Z
https://docs.axolotl.ai/docs/api/utils.collators.mamba.html
- 2025-07-07T19:27:31.593Z
+ 2025-07-07T21:08:37.190Z
https://docs.axolotl.ai/docs/api/monkeypatch.unsloth_.html
- 2025-07-07T19:27:31.133Z
+ 2025-07-07T21:08:36.736Z
https://docs.axolotl.ai/docs/api/utils.dict.html
- 2025-07-07T19:27:31.262Z
+ 2025-07-07T21:08:36.864Z
https://docs.axolotl.ai/docs/api/prompt_strategies.user_defined.html
- 2025-07-07T19:27:30.788Z
+ 2025-07-07T21:08:36.395Z
https://docs.axolotl.ai/docs/api/core.training_args.html
- 2025-07-07T19:27:30.337Z
+ 2025-07-07T21:08:35.951Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.user_defined.html
- 2025-07-07T19:27:30.864Z
+ 2025-07-07T21:08:36.471Z
https://docs.axolotl.ai/docs/api/prompt_tokenizers.html
- 2025-07-07T19:27:30.295Z
+ 2025-07-07T21:08:35.910Z
https://docs.axolotl.ai/docs/api/common.const.html
- 2025-07-07T19:27:31.553Z
+ 2025-07-07T21:08:37.150Z
https://docs.axolotl.ai/docs/fsdp_qlora.html
- 2025-07-07T19:24:01.496Z
+ 2025-07-07T21:05:27.503Z
https://docs.axolotl.ai/docs/custom_integrations.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.502Z
https://docs.axolotl.ai/docs/getting-started.html
- 2025-07-07T19:24:01.496Z
+ 2025-07-07T21:05:27.503Z
https://docs.axolotl.ai/docs/faq.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.503Z
https://docs.axolotl.ai/docs/lora_optims.html
- 2025-07-07T19:24:01.498Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/rlhf.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.508Z
https://docs.axolotl.ai/docs/amd_hpc.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.502Z
https://docs.axolotl.ai/docs/installation.html
- 2025-07-07T19:24:01.498Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/multipack.html
- 2025-07-07T19:24:01.499Z
+ 2025-07-07T21:05:27.507Z
https://docs.axolotl.ai/docs/dataset_preprocessing.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.503Z
https://docs.axolotl.ai/docs/dataset_loading.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.503Z
https://docs.axolotl.ai/docs/dataset-formats/inst_tune.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.502Z
https://docs.axolotl.ai/docs/dataset-formats/template_free.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.503Z
https://docs.axolotl.ai/docs/dataset-formats/index.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.502Z
https://docs.axolotl.ai/docs/dataset-formats/pretraining.html
- 2025-07-07T19:24:01.495Z
+ 2025-07-07T21:05:27.502Z