From ba0dd2cdec8d608acf9f3971b24e10b57a21940e Mon Sep 17 00:00:00 2001 From: Quarto GHA Workflow Runner Date: Wed, 17 Sep 2025 17:32:44 +0000 Subject: [PATCH] Built site for gh-pages --- .nojekyll | 2 +- docs/api/index.html | 2 +- docs/api/logging_config.html | 14 +- .../colab-axolotl-example.html | 6 +- search.json | 8 +- sitemap.xml | 396 +++++++++--------- 6 files changed, 214 insertions(+), 214 deletions(-) diff --git a/.nojekyll b/.nojekyll index a55cc4c5f..7dc9a4cc9 100644 --- a/.nojekyll +++ b/.nojekyll @@ -1 +1 @@ -e807e5c1 \ No newline at end of file +51905dfe \ No newline at end of file diff --git a/docs/api/index.html b/docs/api/index.html index f7564e31e..cbae8080d 100644 --- a/docs/api/index.html +++ b/docs/api/index.html @@ -525,7 +525,7 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true}); logging_config -Common logging module for axolotl +Common logging module for axolotl. core.builders.base diff --git a/docs/api/logging_config.html b/docs/api/logging_config.html index fa6836b0d..49cc5679c 100644 --- a/docs/api/logging_config.html +++ b/docs/api/logging_config.html @@ -528,7 +528,7 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});

logging_config

logging_config

-

Common logging module for axolotl

+

Common logging module for axolotl.

Classes

@@ -541,11 +541,11 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true}); - + - + @@ -556,14 +556,14 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});

AxolotlLogger

logging_config.AxolotlLogger(name, level=logging.NOTSET)
-

A Logger that automatically rejects non-axolotl INFOs.

+

Logger that applies filtering to non-axolotl loggers.

AxolotlOrWarnErrorFilter

logging_config.AxolotlOrWarnErrorFilter(**kwargs)
-

Allows ANY WARNING or higher (unless overridden by LOG_LEVEL) -Allows axolotl.* at INFO or higher (unless overridden by AXOLOTL_LOG_LEVEL) -Drops all other records (i.e. non-axolotl.INFO, DEBUG, etc. by default)

+

Allows ANY WARNING or higher (unless overridden by LOG_LEVEL). Allows axolotl.* at +INFO or higher (unless overridden by AXOLOTL_LOG_LEVEL). Drops all other records +(i.e. non-axolotl.INFO, DEBUG, etc. by default).

ColorfulFormatter

diff --git a/examples/colab-notebooks/colab-axolotl-example.html b/examples/colab-notebooks/colab-axolotl-example.html index 30626c7d4..4fa189615 100644 --- a/examples/colab-notebooks/colab-axolotl-example.html +++ b/examples/colab-notebooks/colab-axolotl-example.html @@ -691,10 +691,10 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
-
from axolotl.utils import patch_optimized_env
+
from axolotl.utils import set_pytorch_cuda_alloc_conf
 
-# speedup downloads from HF 🤗 and set "PYTORCH_CUDA_ALLOC_CONF" env to save memory
-patch_optimized_env()
+# Set "PYTORCH_CUDA_ALLOC_CONF" env to save memory +set_pytorch_cuda_alloc_conf()
diff --git a/search.json b/search.json index 4f9292145..5e249f1a9 100644 --- a/search.json +++ b/search.json @@ -977,14 +977,14 @@ "href": "docs/api/index.html", "title": "API Reference", "section": "", - "text": "Core functionality for training\n\n\n\ntrain\nPrepare and train a model on a dataset. Can also infer from a model or merge lora\n\n\nevaluate\nModule for evaluating models.\n\n\ndatasets\nModule containing dataset functionality.\n\n\nconvert\nModule containing File Reader, File Writer, Json Parser, and Jsonl Serializer classes\n\n\nprompt_tokenizers\nModule containing PromptTokenizingStrategy and Prompter classes\n\n\nlogging_config\nCommon logging module for axolotl\n\n\ncore.builders.base\nBase class for trainer builder\n\n\ncore.builders.causal\nBuilder for causal trainers\n\n\ncore.builders.rl\nBuilder for RLHF trainers\n\n\ncore.training_args\nextra axolotl specific training args\n\n\ncore.chat.messages\ninternal message representations of chat messages\n\n\ncore.chat.format.chatml\nChatML transformation functions for MessageContents\n\n\ncore.chat.format.llama3x\nLlama 3.x chat formatting functions for MessageContents\n\n\ncore.chat.format.shared\nshared functions for format transforms\n\n\ncore.datasets.chat\nchat dataset module\n\n\ncore.datasets.transforms.chat_builder\nThis module contains a function that builds a transform that takes a row from the\n\n\n\n\n\n\nCommand-line interface\n\n\n\ncli.main\nClick CLI definitions for various axolotl commands.\n\n\ncli.train\nCLI to run training on a model.\n\n\ncli.evaluate\nCLI to run evaluation on a model.\n\n\ncli.args\nModule for axolotl CLI command arguments.\n\n\ncli.art\nAxolotl ASCII logo utils.\n\n\ncli.checks\nVarious checks for Axolotl CLI.\n\n\ncli.config\nConfiguration loading and processing.\n\n\ncli.delinearize_llama4\nCLI tool to delinearize quantized/Linearized Llama-4 models.\n\n\ncli.inference\nCLI to run inference on a trained model.\n\n\ncli.merge_lora\nCLI to merge a trained LoRA into a base model.\n\n\ncli.merge_sharded_fsdp_weights\nCLI to merge sharded FSDP model checkpoints into a single combined checkpoint.\n\n\ncli.preprocess\nCLI to run preprocessing of a dataset.\n\n\ncli.quantize\nCLI to post-training quantize a model using torchao\n\n\ncli.vllm_serve\nCLI to start the vllm server for online RL\n\n\ncli.cloud.base\nbase class for cloud platforms from cli\n\n\ncli.cloud.modal_\nModal Cloud support from CLI\n\n\ncli.utils\nInit for axolotl.cli.utils module.\n\n\ncli.utils.args\nUtilities for axolotl CLI args.\n\n\ncli.utils.fetch\nUtilities for axolotl fetch CLI command.\n\n\ncli.utils.load\nUtilities for model, tokenizer, etc. loading.\n\n\ncli.utils.sweeps\nUtilities for handling sweeps over configs for axolotl train CLI command\n\n\ncli.utils.train\nUtilities for axolotl train CLI command.\n\n\n\n\n\n\nTraining implementations\n\n\n\ncore.trainers.base\nModule for customized trainers\n\n\ncore.trainers.trl\nModule for TRL RL trainers\n\n\ncore.trainers.mamba\nModule for mamba trainer\n\n\ncore.trainers.dpo.trainer\nDPO trainer for axolotl\n\n\ncore.trainers.grpo.trainer\nAxolotl GRPO trainers (with and without sequence parallelism handling)\n\n\ncore.trainers.grpo.sampler\nRepeat random sampler (similar to the one implemented in\n\n\ncore.trainers.utils\nUtils for Axolotl trainers\n\n\n\n\n\n\nFunctionality for loading and patching models, tokenizers, etc.\n\n\n\nloaders.model\nModel loader class implementation for loading, configuring, and patching various models.\n\n\nloaders.tokenizer\nTokenizer loading functionality and associated utils\n\n\nloaders.processor\nProcessor loading functionality for multi-modal models\n\n\nloaders.adapter\nAdapter loading functionality, including LoRA / QLoRA and associated utils\n\n\nloaders.patch_manager\nPatch manager class implementation to complement axolotl.loaders.ModelLoader.\n\n\nloaders.constants\nShared constants for axolotl.loaders module\n\n\n\n\n\n\nMixin classes for augmenting trainers\n\n\n\ncore.trainers.mixins.optimizer\nModule for Axolotl trainer optimizer mixin\n\n\ncore.trainers.mixins.rng_state_loader\nTemporary fix/override for bug in resume from checkpoint\n\n\ncore.trainers.mixins.scheduler\nModule for Axolotl trainer scheduler mixin\n\n\n\n\n\n\nContext managers for altering trainer behaviors\n\n\n\nutils.ctx_managers.sequence_parallel\nModule for Axolotl trainer sequence parallelism manager and utilities\n\n\n\n\n\n\nPrompt formatting strategies\n\n\n\nprompt_strategies.base\nmodule for base dataset transform strategies\n\n\nprompt_strategies.chat_template\nHF Chat Templates prompt strategy\n\n\nprompt_strategies.alpaca_chat\nModule for Alpaca prompt strategy classes\n\n\nprompt_strategies.alpaca_instruct\nModule loading the AlpacaInstructPromptTokenizingStrategy class\n\n\nprompt_strategies.alpaca_w_system\nPrompt strategies loader for alpaca instruction datasets with system prompts\n\n\nprompt_strategies.user_defined\nUser Defined prompts with configuration from the YML config\n\n\nprompt_strategies.llama2_chat\nPrompt Strategy for finetuning Llama2 chat models\n\n\nprompt_strategies.completion\nBasic completion text\n\n\nprompt_strategies.input_output\nModule for plain input/output prompt pairs\n\n\nprompt_strategies.stepwise_supervised\nModule for stepwise datasets, typically including a prompt and reasoning traces,\n\n\nprompt_strategies.metharme\nModule containing the MetharmenPromptTokenizingStrategy and MetharmePrompter class\n\n\nprompt_strategies.orcamini\nPrompt Strategy for finetuning Orca Mini (v2) models\n\n\nprompt_strategies.pygmalion\nModule containing the PygmalionPromptTokenizingStrategy and PygmalionPrompter class\n\n\nprompt_strategies.messages.chat\nChat dataset wrapping strategy for new internal messages representations\n\n\nprompt_strategies.dpo.chat_template\nDPO prompt strategies for using tokenizer chat templates.\n\n\nprompt_strategies.dpo.llama3\nDPO strategies for llama-3 chat template\n\n\nprompt_strategies.dpo.chatml\nDPO strategies for chatml\n\n\nprompt_strategies.dpo.zephyr\nDPO strategies for zephyr\n\n\nprompt_strategies.dpo.user_defined\nUser-defined DPO strategies\n\n\nprompt_strategies.dpo.passthrough\nDPO prompt strategies passthrough/zero-processing strategy\n\n\nprompt_strategies.kto.llama3\nKTO strategies for llama-3 chat template\n\n\nprompt_strategies.kto.chatml\nKTO strategies for chatml\n\n\nprompt_strategies.kto.user_defined\nUser-defined KTO strategies\n\n\nprompt_strategies.orpo.chat_template\nchatml prompt tokenization strategy for ORPO\n\n\nprompt_strategies.bradley_terry.llama3\nchatml transforms for datasets with system, input, chosen, rejected to match llama3 chat template\n\n\n\n\n\n\nLow-level performance optimizations\n\n\n\nkernels.lora\nModule for definition of Low-Rank Adaptation (LoRA) Triton kernels.\n\n\nkernels.geglu\nModule for definition of GEGLU Triton kernels.\n\n\nkernels.swiglu\nModule for definition of SwiGLU Triton kernels.\n\n\nkernels.quantize\nDequantization utilities for bitsandbytes integration.\n\n\nkernels.utils\nUtilities for axolotl.kernels submodules.\n\n\n\n\n\n\nRuntime patches for model optimizations\n\n\n\nmonkeypatch.llama_attn_hijack_flash\nFlash attention monkey patch for llama model\n\n\nmonkeypatch.llama_attn_hijack_xformers\nDirectly copied the code from https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/modules/llama_attn_hijack.py and made some adjustments\n\n\nmonkeypatch.mistral_attn_hijack_flash\nFlash attention monkey patch for mistral model\n\n\nmonkeypatch.multipack\nmultipack patching for v2 of sample packing\n\n\nmonkeypatch.relora\nImplements the ReLoRA training procedure from https://arxiv.org/abs/2307.05695, minus the initial full fine-tune.\n\n\nmonkeypatch.llama_expand_mask\nexpands the binary attention mask per 3.2.2 of https://arxiv.org/pdf/2107.02027.pdf\n\n\nmonkeypatch.lora_kernels\nModule for patching custom LoRA Triton kernels and torch.autograd functions.\n\n\nmonkeypatch.utils\nShared utils for the monkeypatches\n\n\nmonkeypatch.btlm_attn_hijack_flash\nFlash attention monkey patch for cerebras btlm model\n\n\nmonkeypatch.llama_patch_multipack\nPatched LlamaAttention to use torch.nn.functional.scaled_dot_product_attention\n\n\nmonkeypatch.stablelm_attn_hijack_flash\nPyTorch StableLM Epoch model.\n\n\nmonkeypatch.trainer_fsdp_optim\nfix for FSDP optimizer save in trainer w 4.47.0\n\n\nmonkeypatch.transformers_fa_utils\nsee https://github.com/huggingface/transformers/pull/35834\n\n\nmonkeypatch.unsloth_\nmodule for patching with unsloth optimizations\n\n\nmonkeypatch.data.batch_dataset_fetcher\nMonkey patches for the dataset fetcher to handle batches of packed indexes.\n\n\nmonkeypatch.mixtral\nPatches to support multipack for mixtral\n\n\nmonkeypatch.gradient_checkpointing.offload_cpu\nCPU offloaded checkpointing\n\n\nmonkeypatch.gradient_checkpointing.offload_disk\nDISCO - DIsk-based Storage and Checkpointing with Optimized prefetching\n\n\n\n\n\n\nUtility functions\n\n\n\nutils.tokenization\nModule for tokenization utilities\n\n\nutils.chat_templates\nThis module provides functionality for selecting chat templates based on user choices.\n\n\nutils.lora\nmodule to get the state dict of a merged lora model\n\n\nutils.model_shard_quant\nmodule to handle loading model on cpu/meta device for FSDP\n\n\nutils.bench\nBenchmarking and measurement utilities\n\n\nutils.freeze\nmodule to freeze/unfreeze parameters by name\n\n\nutils.trainer\nModule containing the Trainer class and related functions\n\n\nutils.schedulers\nModule for custom LRScheduler class\n\n\nutils.distributed\nUtilities for distributed functionality.\n\n\nutils.dict\nModule containing the DictDefault class\n\n\nutils.optimizers.adopt\nCopied from https://github.com/iShohei220/adopt\n\n\nutils.data.streaming\nData handling specific to streaming datasets.\n\n\nutils.data.sft\nData handling specific to SFT.\n\n\nutils.quantization\nUtilities for quantization including QAT and PTQ using torchao.\n\n\n\n\n\n\nPydantic data models for Axolotl config\n\n\n\nutils.schemas.config\nModule with Pydantic models for configuration.\n\n\nutils.schemas.model\nPydantic models for model input / output, etc. configuration\n\n\nutils.schemas.training\nPydantic models for training hyperparameters\n\n\nutils.schemas.datasets\nPydantic models for datasets-related configuration\n\n\nutils.schemas.peft\nPydantic models for PEFT-related configuration\n\n\nutils.schemas.trl\nPydantic models for TRL trainer configuration\n\n\nutils.schemas.multimodal\nPydantic models for multimodal-related configuration\n\n\nutils.schemas.integrations\nPydantic models for Axolotl integrations\n\n\nutils.schemas.enums\nEnums for Axolotl input config\n\n\nutils.schemas.utils\nUtilities for Axolotl Pydantic models\n\n\n\n\n\n\nThird-party integrations and extensions\n\n\n\nintegrations.base\nBase class for all plugins.\n\n\nintegrations.cut_cross_entropy.args\nModule for handling Cut Cross Entropy input arguments.\n\n\nintegrations.grokfast.optimizer\n\n\n\nintegrations.kd.trainer\nKD trainer\n\n\nintegrations.liger.args\nModule for handling LIGER input arguments.\n\n\nintegrations.lm_eval.args\nModule for handling lm eval harness input arguments.\n\n\nintegrations.spectrum.args\nModule for handling Spectrum input arguments.\n\n\n\n\n\n\nCommon utilities and shared functionality\n\n\n\ncommon.architectures\nCommon architecture specific constants\n\n\ncommon.const\nVarious shared constants\n\n\ncommon.datasets\nDataset loading utilities.\n\n\n\n\n\n\nCustom model implementations\n\n\n\nmodels.mamba.modeling_mamba\n\n\n\n\n\n\n\nData processing utilities\n\n\n\nutils.collators.core\nbasic shared collator constants\n\n\nutils.collators.batching\nData collators for axolotl to pad labels and position_ids for packed sequences\n\n\nutils.collators.mamba\ncollators for Mamba\n\n\nutils.collators.mm_chat\nCollators for multi-modal chat messages and packing\n\n\nutils.samplers.multipack\nMultipack Batch Sampler - An efficient batch sampler for packing variable-length sequences\n\n\n\n\n\n\nTraining callbacks\n\n\n\nutils.callbacks.perplexity\ncallback to calculate perplexity as an evaluation metric.\n\n\nutils.callbacks.profiler\nHF Trainer callback for creating pytorch profiling snapshots\n\n\nutils.callbacks.lisa\nmodule for LISA\n\n\nutils.callbacks.mlflow_\nMLFlow module for trainer callbacks\n\n\nutils.callbacks.comet_\nComet module for trainer callbacks\n\n\nutils.callbacks.qat\nQAT Callback for HF Causal Trainer" + "text": "Core functionality for training\n\n\n\ntrain\nPrepare and train a model on a dataset. Can also infer from a model or merge lora\n\n\nevaluate\nModule for evaluating models.\n\n\ndatasets\nModule containing dataset functionality.\n\n\nconvert\nModule containing File Reader, File Writer, Json Parser, and Jsonl Serializer classes\n\n\nprompt_tokenizers\nModule containing PromptTokenizingStrategy and Prompter classes\n\n\nlogging_config\nCommon logging module for axolotl.\n\n\ncore.builders.base\nBase class for trainer builder\n\n\ncore.builders.causal\nBuilder for causal trainers\n\n\ncore.builders.rl\nBuilder for RLHF trainers\n\n\ncore.training_args\nextra axolotl specific training args\n\n\ncore.chat.messages\ninternal message representations of chat messages\n\n\ncore.chat.format.chatml\nChatML transformation functions for MessageContents\n\n\ncore.chat.format.llama3x\nLlama 3.x chat formatting functions for MessageContents\n\n\ncore.chat.format.shared\nshared functions for format transforms\n\n\ncore.datasets.chat\nchat dataset module\n\n\ncore.datasets.transforms.chat_builder\nThis module contains a function that builds a transform that takes a row from the\n\n\n\n\n\n\nCommand-line interface\n\n\n\ncli.main\nClick CLI definitions for various axolotl commands.\n\n\ncli.train\nCLI to run training on a model.\n\n\ncli.evaluate\nCLI to run evaluation on a model.\n\n\ncli.args\nModule for axolotl CLI command arguments.\n\n\ncli.art\nAxolotl ASCII logo utils.\n\n\ncli.checks\nVarious checks for Axolotl CLI.\n\n\ncli.config\nConfiguration loading and processing.\n\n\ncli.delinearize_llama4\nCLI tool to delinearize quantized/Linearized Llama-4 models.\n\n\ncli.inference\nCLI to run inference on a trained model.\n\n\ncli.merge_lora\nCLI to merge a trained LoRA into a base model.\n\n\ncli.merge_sharded_fsdp_weights\nCLI to merge sharded FSDP model checkpoints into a single combined checkpoint.\n\n\ncli.preprocess\nCLI to run preprocessing of a dataset.\n\n\ncli.quantize\nCLI to post-training quantize a model using torchao\n\n\ncli.vllm_serve\nCLI to start the vllm server for online RL\n\n\ncli.cloud.base\nbase class for cloud platforms from cli\n\n\ncli.cloud.modal_\nModal Cloud support from CLI\n\n\ncli.utils\nInit for axolotl.cli.utils module.\n\n\ncli.utils.args\nUtilities for axolotl CLI args.\n\n\ncli.utils.fetch\nUtilities for axolotl fetch CLI command.\n\n\ncli.utils.load\nUtilities for model, tokenizer, etc. loading.\n\n\ncli.utils.sweeps\nUtilities for handling sweeps over configs for axolotl train CLI command\n\n\ncli.utils.train\nUtilities for axolotl train CLI command.\n\n\n\n\n\n\nTraining implementations\n\n\n\ncore.trainers.base\nModule for customized trainers\n\n\ncore.trainers.trl\nModule for TRL RL trainers\n\n\ncore.trainers.mamba\nModule for mamba trainer\n\n\ncore.trainers.dpo.trainer\nDPO trainer for axolotl\n\n\ncore.trainers.grpo.trainer\nAxolotl GRPO trainers (with and without sequence parallelism handling)\n\n\ncore.trainers.grpo.sampler\nRepeat random sampler (similar to the one implemented in\n\n\ncore.trainers.utils\nUtils for Axolotl trainers\n\n\n\n\n\n\nFunctionality for loading and patching models, tokenizers, etc.\n\n\n\nloaders.model\nModel loader class implementation for loading, configuring, and patching various models.\n\n\nloaders.tokenizer\nTokenizer loading functionality and associated utils\n\n\nloaders.processor\nProcessor loading functionality for multi-modal models\n\n\nloaders.adapter\nAdapter loading functionality, including LoRA / QLoRA and associated utils\n\n\nloaders.patch_manager\nPatch manager class implementation to complement axolotl.loaders.ModelLoader.\n\n\nloaders.constants\nShared constants for axolotl.loaders module\n\n\n\n\n\n\nMixin classes for augmenting trainers\n\n\n\ncore.trainers.mixins.optimizer\nModule for Axolotl trainer optimizer mixin\n\n\ncore.trainers.mixins.rng_state_loader\nTemporary fix/override for bug in resume from checkpoint\n\n\ncore.trainers.mixins.scheduler\nModule for Axolotl trainer scheduler mixin\n\n\n\n\n\n\nContext managers for altering trainer behaviors\n\n\n\nutils.ctx_managers.sequence_parallel\nModule for Axolotl trainer sequence parallelism manager and utilities\n\n\n\n\n\n\nPrompt formatting strategies\n\n\n\nprompt_strategies.base\nmodule for base dataset transform strategies\n\n\nprompt_strategies.chat_template\nHF Chat Templates prompt strategy\n\n\nprompt_strategies.alpaca_chat\nModule for Alpaca prompt strategy classes\n\n\nprompt_strategies.alpaca_instruct\nModule loading the AlpacaInstructPromptTokenizingStrategy class\n\n\nprompt_strategies.alpaca_w_system\nPrompt strategies loader for alpaca instruction datasets with system prompts\n\n\nprompt_strategies.user_defined\nUser Defined prompts with configuration from the YML config\n\n\nprompt_strategies.llama2_chat\nPrompt Strategy for finetuning Llama2 chat models\n\n\nprompt_strategies.completion\nBasic completion text\n\n\nprompt_strategies.input_output\nModule for plain input/output prompt pairs\n\n\nprompt_strategies.stepwise_supervised\nModule for stepwise datasets, typically including a prompt and reasoning traces,\n\n\nprompt_strategies.metharme\nModule containing the MetharmenPromptTokenizingStrategy and MetharmePrompter class\n\n\nprompt_strategies.orcamini\nPrompt Strategy for finetuning Orca Mini (v2) models\n\n\nprompt_strategies.pygmalion\nModule containing the PygmalionPromptTokenizingStrategy and PygmalionPrompter class\n\n\nprompt_strategies.messages.chat\nChat dataset wrapping strategy for new internal messages representations\n\n\nprompt_strategies.dpo.chat_template\nDPO prompt strategies for using tokenizer chat templates.\n\n\nprompt_strategies.dpo.llama3\nDPO strategies for llama-3 chat template\n\n\nprompt_strategies.dpo.chatml\nDPO strategies for chatml\n\n\nprompt_strategies.dpo.zephyr\nDPO strategies for zephyr\n\n\nprompt_strategies.dpo.user_defined\nUser-defined DPO strategies\n\n\nprompt_strategies.dpo.passthrough\nDPO prompt strategies passthrough/zero-processing strategy\n\n\nprompt_strategies.kto.llama3\nKTO strategies for llama-3 chat template\n\n\nprompt_strategies.kto.chatml\nKTO strategies for chatml\n\n\nprompt_strategies.kto.user_defined\nUser-defined KTO strategies\n\n\nprompt_strategies.orpo.chat_template\nchatml prompt tokenization strategy for ORPO\n\n\nprompt_strategies.bradley_terry.llama3\nchatml transforms for datasets with system, input, chosen, rejected to match llama3 chat template\n\n\n\n\n\n\nLow-level performance optimizations\n\n\n\nkernels.lora\nModule for definition of Low-Rank Adaptation (LoRA) Triton kernels.\n\n\nkernels.geglu\nModule for definition of GEGLU Triton kernels.\n\n\nkernels.swiglu\nModule for definition of SwiGLU Triton kernels.\n\n\nkernels.quantize\nDequantization utilities for bitsandbytes integration.\n\n\nkernels.utils\nUtilities for axolotl.kernels submodules.\n\n\n\n\n\n\nRuntime patches for model optimizations\n\n\n\nmonkeypatch.llama_attn_hijack_flash\nFlash attention monkey patch for llama model\n\n\nmonkeypatch.llama_attn_hijack_xformers\nDirectly copied the code from https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/modules/llama_attn_hijack.py and made some adjustments\n\n\nmonkeypatch.mistral_attn_hijack_flash\nFlash attention monkey patch for mistral model\n\n\nmonkeypatch.multipack\nmultipack patching for v2 of sample packing\n\n\nmonkeypatch.relora\nImplements the ReLoRA training procedure from https://arxiv.org/abs/2307.05695, minus the initial full fine-tune.\n\n\nmonkeypatch.llama_expand_mask\nexpands the binary attention mask per 3.2.2 of https://arxiv.org/pdf/2107.02027.pdf\n\n\nmonkeypatch.lora_kernels\nModule for patching custom LoRA Triton kernels and torch.autograd functions.\n\n\nmonkeypatch.utils\nShared utils for the monkeypatches\n\n\nmonkeypatch.btlm_attn_hijack_flash\nFlash attention monkey patch for cerebras btlm model\n\n\nmonkeypatch.llama_patch_multipack\nPatched LlamaAttention to use torch.nn.functional.scaled_dot_product_attention\n\n\nmonkeypatch.stablelm_attn_hijack_flash\nPyTorch StableLM Epoch model.\n\n\nmonkeypatch.trainer_fsdp_optim\nfix for FSDP optimizer save in trainer w 4.47.0\n\n\nmonkeypatch.transformers_fa_utils\nsee https://github.com/huggingface/transformers/pull/35834\n\n\nmonkeypatch.unsloth_\nmodule for patching with unsloth optimizations\n\n\nmonkeypatch.data.batch_dataset_fetcher\nMonkey patches for the dataset fetcher to handle batches of packed indexes.\n\n\nmonkeypatch.mixtral\nPatches to support multipack for mixtral\n\n\nmonkeypatch.gradient_checkpointing.offload_cpu\nCPU offloaded checkpointing\n\n\nmonkeypatch.gradient_checkpointing.offload_disk\nDISCO - DIsk-based Storage and Checkpointing with Optimized prefetching\n\n\n\n\n\n\nUtility functions\n\n\n\nutils.tokenization\nModule for tokenization utilities\n\n\nutils.chat_templates\nThis module provides functionality for selecting chat templates based on user choices.\n\n\nutils.lora\nmodule to get the state dict of a merged lora model\n\n\nutils.model_shard_quant\nmodule to handle loading model on cpu/meta device for FSDP\n\n\nutils.bench\nBenchmarking and measurement utilities\n\n\nutils.freeze\nmodule to freeze/unfreeze parameters by name\n\n\nutils.trainer\nModule containing the Trainer class and related functions\n\n\nutils.schedulers\nModule for custom LRScheduler class\n\n\nutils.distributed\nUtilities for distributed functionality.\n\n\nutils.dict\nModule containing the DictDefault class\n\n\nutils.optimizers.adopt\nCopied from https://github.com/iShohei220/adopt\n\n\nutils.data.streaming\nData handling specific to streaming datasets.\n\n\nutils.data.sft\nData handling specific to SFT.\n\n\nutils.quantization\nUtilities for quantization including QAT and PTQ using torchao.\n\n\n\n\n\n\nPydantic data models for Axolotl config\n\n\n\nutils.schemas.config\nModule with Pydantic models for configuration.\n\n\nutils.schemas.model\nPydantic models for model input / output, etc. configuration\n\n\nutils.schemas.training\nPydantic models for training hyperparameters\n\n\nutils.schemas.datasets\nPydantic models for datasets-related configuration\n\n\nutils.schemas.peft\nPydantic models for PEFT-related configuration\n\n\nutils.schemas.trl\nPydantic models for TRL trainer configuration\n\n\nutils.schemas.multimodal\nPydantic models for multimodal-related configuration\n\n\nutils.schemas.integrations\nPydantic models for Axolotl integrations\n\n\nutils.schemas.enums\nEnums for Axolotl input config\n\n\nutils.schemas.utils\nUtilities for Axolotl Pydantic models\n\n\n\n\n\n\nThird-party integrations and extensions\n\n\n\nintegrations.base\nBase class for all plugins.\n\n\nintegrations.cut_cross_entropy.args\nModule for handling Cut Cross Entropy input arguments.\n\n\nintegrations.grokfast.optimizer\n\n\n\nintegrations.kd.trainer\nKD trainer\n\n\nintegrations.liger.args\nModule for handling LIGER input arguments.\n\n\nintegrations.lm_eval.args\nModule for handling lm eval harness input arguments.\n\n\nintegrations.spectrum.args\nModule for handling Spectrum input arguments.\n\n\n\n\n\n\nCommon utilities and shared functionality\n\n\n\ncommon.architectures\nCommon architecture specific constants\n\n\ncommon.const\nVarious shared constants\n\n\ncommon.datasets\nDataset loading utilities.\n\n\n\n\n\n\nCustom model implementations\n\n\n\nmodels.mamba.modeling_mamba\n\n\n\n\n\n\n\nData processing utilities\n\n\n\nutils.collators.core\nbasic shared collator constants\n\n\nutils.collators.batching\nData collators for axolotl to pad labels and position_ids for packed sequences\n\n\nutils.collators.mamba\ncollators for Mamba\n\n\nutils.collators.mm_chat\nCollators for multi-modal chat messages and packing\n\n\nutils.samplers.multipack\nMultipack Batch Sampler - An efficient batch sampler for packing variable-length sequences\n\n\n\n\n\n\nTraining callbacks\n\n\n\nutils.callbacks.perplexity\ncallback to calculate perplexity as an evaluation metric.\n\n\nutils.callbacks.profiler\nHF Trainer callback for creating pytorch profiling snapshots\n\n\nutils.callbacks.lisa\nmodule for LISA\n\n\nutils.callbacks.mlflow_\nMLFlow module for trainer callbacks\n\n\nutils.callbacks.comet_\nComet module for trainer callbacks\n\n\nutils.callbacks.qat\nQAT Callback for HF Causal Trainer" }, { "objectID": "docs/api/index.html#core", "href": "docs/api/index.html#core", "title": "API Reference", "section": "", - "text": "Core functionality for training\n\n\n\ntrain\nPrepare and train a model on a dataset. Can also infer from a model or merge lora\n\n\nevaluate\nModule for evaluating models.\n\n\ndatasets\nModule containing dataset functionality.\n\n\nconvert\nModule containing File Reader, File Writer, Json Parser, and Jsonl Serializer classes\n\n\nprompt_tokenizers\nModule containing PromptTokenizingStrategy and Prompter classes\n\n\nlogging_config\nCommon logging module for axolotl\n\n\ncore.builders.base\nBase class for trainer builder\n\n\ncore.builders.causal\nBuilder for causal trainers\n\n\ncore.builders.rl\nBuilder for RLHF trainers\n\n\ncore.training_args\nextra axolotl specific training args\n\n\ncore.chat.messages\ninternal message representations of chat messages\n\n\ncore.chat.format.chatml\nChatML transformation functions for MessageContents\n\n\ncore.chat.format.llama3x\nLlama 3.x chat formatting functions for MessageContents\n\n\ncore.chat.format.shared\nshared functions for format transforms\n\n\ncore.datasets.chat\nchat dataset module\n\n\ncore.datasets.transforms.chat_builder\nThis module contains a function that builds a transform that takes a row from the" + "text": "Core functionality for training\n\n\n\ntrain\nPrepare and train a model on a dataset. Can also infer from a model or merge lora\n\n\nevaluate\nModule for evaluating models.\n\n\ndatasets\nModule containing dataset functionality.\n\n\nconvert\nModule containing File Reader, File Writer, Json Parser, and Jsonl Serializer classes\n\n\nprompt_tokenizers\nModule containing PromptTokenizingStrategy and Prompter classes\n\n\nlogging_config\nCommon logging module for axolotl.\n\n\ncore.builders.base\nBase class for trainer builder\n\n\ncore.builders.causal\nBuilder for causal trainers\n\n\ncore.builders.rl\nBuilder for RLHF trainers\n\n\ncore.training_args\nextra axolotl specific training args\n\n\ncore.chat.messages\ninternal message representations of chat messages\n\n\ncore.chat.format.chatml\nChatML transformation functions for MessageContents\n\n\ncore.chat.format.llama3x\nLlama 3.x chat formatting functions for MessageContents\n\n\ncore.chat.format.shared\nshared functions for format transforms\n\n\ncore.datasets.chat\nchat dataset module\n\n\ncore.datasets.transforms.chat_builder\nThis module contains a function that builds a transform that takes a row from the" }, { "objectID": "docs/api/index.html#cli", @@ -1292,14 +1292,14 @@ "href": "docs/api/logging_config.html", "title": "logging_config", "section": "", - "text": "logging_config\nCommon logging module for axolotl\n\n\n\n\n\nName\nDescription\n\n\n\n\nAxolotlLogger\nA Logger that automatically rejects non-axolotl INFOs.\n\n\nAxolotlOrWarnErrorFilter\nAllows ANY WARNING or higher (unless overridden by LOG_LEVEL)\n\n\nColorfulFormatter\nFormatter to add coloring to log messages by log type\n\n\n\n\n\nlogging_config.AxolotlLogger(name, level=logging.NOTSET)\nA Logger that automatically rejects non-axolotl INFOs.\n\n\n\nlogging_config.AxolotlOrWarnErrorFilter(**kwargs)\nAllows ANY WARNING or higher (unless overridden by LOG_LEVEL)\nAllows axolotl.* at INFO or higher (unless overridden by AXOLOTL_LOG_LEVEL)\nDrops all other records (i.e. non-axolotl.INFO, DEBUG, etc. by default)\n\n\n\nlogging_config.ColorfulFormatter()\nFormatter to add coloring to log messages by log type\n\n\n\n\n\n\n\nName\nDescription\n\n\n\n\nconfigure_logging\nConfigure with default logging\n\n\n\n\n\nlogging_config.configure_logging()\nConfigure with default logging" + "text": "logging_config\nCommon logging module for axolotl.\n\n\n\n\n\nName\nDescription\n\n\n\n\nAxolotlLogger\nLogger that applies filtering to non-axolotl loggers.\n\n\nAxolotlOrWarnErrorFilter\nAllows ANY WARNING or higher (unless overridden by LOG_LEVEL). Allows axolotl.* at\n\n\nColorfulFormatter\nFormatter to add coloring to log messages by log type\n\n\n\n\n\nlogging_config.AxolotlLogger(name, level=logging.NOTSET)\nLogger that applies filtering to non-axolotl loggers.\n\n\n\nlogging_config.AxolotlOrWarnErrorFilter(**kwargs)\nAllows ANY WARNING or higher (unless overridden by LOG_LEVEL). Allows axolotl.* at\nINFO or higher (unless overridden by AXOLOTL_LOG_LEVEL). Drops all other records\n(i.e. non-axolotl.INFO, DEBUG, etc. by default).\n\n\n\nlogging_config.ColorfulFormatter()\nFormatter to add coloring to log messages by log type\n\n\n\n\n\n\n\nName\nDescription\n\n\n\n\nconfigure_logging\nConfigure with default logging\n\n\n\n\n\nlogging_config.configure_logging()\nConfigure with default logging" }, { "objectID": "docs/api/logging_config.html#classes", "href": "docs/api/logging_config.html#classes", "title": "logging_config", "section": "", - "text": "Name\nDescription\n\n\n\n\nAxolotlLogger\nA Logger that automatically rejects non-axolotl INFOs.\n\n\nAxolotlOrWarnErrorFilter\nAllows ANY WARNING or higher (unless overridden by LOG_LEVEL)\n\n\nColorfulFormatter\nFormatter to add coloring to log messages by log type\n\n\n\n\n\nlogging_config.AxolotlLogger(name, level=logging.NOTSET)\nA Logger that automatically rejects non-axolotl INFOs.\n\n\n\nlogging_config.AxolotlOrWarnErrorFilter(**kwargs)\nAllows ANY WARNING or higher (unless overridden by LOG_LEVEL)\nAllows axolotl.* at INFO or higher (unless overridden by AXOLOTL_LOG_LEVEL)\nDrops all other records (i.e. non-axolotl.INFO, DEBUG, etc. by default)\n\n\n\nlogging_config.ColorfulFormatter()\nFormatter to add coloring to log messages by log type" + "text": "Name\nDescription\n\n\n\n\nAxolotlLogger\nLogger that applies filtering to non-axolotl loggers.\n\n\nAxolotlOrWarnErrorFilter\nAllows ANY WARNING or higher (unless overridden by LOG_LEVEL). Allows axolotl.* at\n\n\nColorfulFormatter\nFormatter to add coloring to log messages by log type\n\n\n\n\n\nlogging_config.AxolotlLogger(name, level=logging.NOTSET)\nLogger that applies filtering to non-axolotl loggers.\n\n\n\nlogging_config.AxolotlOrWarnErrorFilter(**kwargs)\nAllows ANY WARNING or higher (unless overridden by LOG_LEVEL). Allows axolotl.* at\nINFO or higher (unless overridden by AXOLOTL_LOG_LEVEL). Drops all other records\n(i.e. non-axolotl.INFO, DEBUG, etc. by default).\n\n\n\nlogging_config.ColorfulFormatter()\nFormatter to add coloring to log messages by log type" }, { "objectID": "docs/api/logging_config.html#functions", diff --git a/sitemap.xml b/sitemap.xml index 13801b961..4d2e51f0b 100644 --- a/sitemap.xml +++ b/sitemap.xml @@ -2,794 +2,794 @@ https://docs.axolotl.ai/examples/colab-notebooks/colab-axolotl-example.html - 2025-09-17T09:38:24.776Z + 2025-09-17T17:27:12.024Z https://docs.axolotl.ai/src/axolotl/integrations/LICENSE.html - 2025-09-17T09:38:24.792Z + 2025-09-17T17:27:12.040Z https://docs.axolotl.ai/FAQS.html - 2025-09-17T09:38:24.766Z + 2025-09-17T17:27:12.014Z https://docs.axolotl.ai/docs/unsloth.html - 2025-09-17T09:38:24.772Z + 2025-09-17T17:27:12.020Z https://docs.axolotl.ai/docs/reward_modelling.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/docker.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.016Z https://docs.axolotl.ai/docs/batch_vs_grad.html - 2025-09-17T09:38:24.767Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/streaming.html - 2025-09-17T09:38:24.772Z + 2025-09-17T17:27:12.020Z https://docs.axolotl.ai/docs/nccl.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/quantize.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/nd_parallelism.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/custom_integrations.html - 2025-09-17T09:38:24.767Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/debugging.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.016Z https://docs.axolotl.ai/docs/gradient_checkpointing.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.016Z https://docs.axolotl.ai/docs/multimodal.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/dataset_loading.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.016Z https://docs.axolotl.ai/docs/api/integrations.spectrum.args.html - 2025-09-17T09:41:55.684Z + 2025-09-17T17:30:32.480Z https://docs.axolotl.ai/docs/api/loaders.adapter.html - 2025-09-17T09:41:54.825Z + 2025-09-17T17:30:31.622Z https://docs.axolotl.ai/docs/api/core.builders.base.html - 2025-09-17T09:41:54.462Z + 2025-09-17T17:30:31.256Z https://docs.axolotl.ai/docs/api/cli.merge_sharded_fsdp_weights.html - 2025-09-17T09:41:54.662Z + 2025-09-17T17:30:31.459Z https://docs.axolotl.ai/docs/api/cli.cloud.modal_.html - 2025-09-17T09:41:54.692Z + 2025-09-17T17:30:31.489Z https://docs.axolotl.ai/docs/api/monkeypatch.stablelm_attn_hijack_flash.html - 2025-09-17T09:41:55.244Z + 2025-09-17T17:30:32.042Z https://docs.axolotl.ai/docs/api/integrations.kd.trainer.html - 2025-09-17T09:41:55.674Z + 2025-09-17T17:30:32.470Z https://docs.axolotl.ai/docs/api/cli.delinearize_llama4.html - 2025-09-17T09:41:54.627Z + 2025-09-17T17:30:31.423Z https://docs.axolotl.ai/docs/api/cli.evaluate.html - 2025-09-17T09:41:54.575Z + 2025-09-17T17:30:31.371Z https://docs.axolotl.ai/docs/api/utils.schemas.model.html - 2025-09-17T09:41:55.454Z + 2025-09-17T17:30:32.251Z https://docs.axolotl.ai/docs/api/cli.art.html - 2025-09-17T09:41:54.598Z + 2025-09-17T17:30:31.394Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.llama3.html - 2025-09-17T09:41:55.009Z + 2025-09-17T17:30:31.805Z https://docs.axolotl.ai/docs/api/core.chat.format.llama3x.html - 2025-09-17T09:41:54.510Z + 2025-09-17T17:30:31.304Z https://docs.axolotl.ai/docs/api/prompt_strategies.kto.chatml.html - 2025-09-17T09:41:55.040Z + 2025-09-17T17:30:31.836Z https://docs.axolotl.ai/docs/api/cli.args.html - 2025-09-17T09:41:54.594Z + 2025-09-17T17:30:31.391Z https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_cpu.html - 2025-09-17T09:41:55.269Z + 2025-09-17T17:30:32.067Z https://docs.axolotl.ai/docs/api/evaluate.html - 2025-09-17T09:41:54.384Z + 2025-09-17T17:30:31.179Z https://docs.axolotl.ai/docs/api/prompt_strategies.input_output.html - 2025-09-17T09:41:54.971Z + 2025-09-17T17:30:31.767Z https://docs.axolotl.ai/docs/api/cli.vllm_serve.html - 2025-09-17T09:41:54.682Z + 2025-09-17T17:30:31.479Z https://docs.axolotl.ai/docs/api/cli.quantize.html - 2025-09-17T09:41:54.675Z + 2025-09-17T17:30:31.472Z https://docs.axolotl.ai/docs/api/cli.utils.sweeps.html - 2025-09-17T09:41:54.722Z + 2025-09-17T17:30:31.518Z https://docs.axolotl.ai/docs/api/core.builders.rl.html - 2025-09-17T09:41:54.472Z + 2025-09-17T17:30:31.266Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chatml.html - 2025-09-17T09:41:55.019Z + 2025-09-17T17:30:31.815Z https://docs.axolotl.ai/docs/api/prompt_strategies.metharme.html - 2025-09-17T09:41:54.982Z + 2025-09-17T17:30:31.779Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.passthrough.html - 2025-09-17T09:41:55.023Z + 2025-09-17T17:30:31.820Z https://docs.axolotl.ai/docs/api/core.datasets.chat.html - 2025-09-17T09:41:54.517Z + 2025-09-17T17:30:31.311Z https://docs.axolotl.ai/docs/api/utils.distributed.html - 2025-09-17T09:41:55.390Z + 2025-09-17T17:30:32.189Z https://docs.axolotl.ai/docs/api/integrations.base.html - 2025-09-17T09:41:55.662Z + 2025-09-17T17:30:32.458Z https://docs.axolotl.ai/docs/api/index.html - 2025-09-17T09:41:54.315Z + 2025-09-17T17:30:31.110Z https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_disk.html - 2025-09-17T09:41:55.295Z + 2025-09-17T17:30:32.093Z https://docs.axolotl.ai/docs/api/utils.quantization.html - 2025-09-17T09:41:55.431Z + 2025-09-17T17:30:32.230Z https://docs.axolotl.ai/docs/api/utils.dict.html - 2025-09-17T09:41:55.396Z + 2025-09-17T17:30:32.195Z https://docs.axolotl.ai/docs/api/core.trainers.dpo.trainer.html - 2025-09-17T09:41:54.776Z + 2025-09-17T17:30:31.572Z https://docs.axolotl.ai/docs/api/utils.schemas.config.html - 2025-09-17T09:41:55.447Z + 2025-09-17T17:30:32.244Z https://docs.axolotl.ai/docs/api/prompt_strategies.orcamini.html - 2025-09-17T09:41:54.986Z + 2025-09-17T17:30:31.782Z https://docs.axolotl.ai/docs/api/utils.collators.mamba.html - 2025-09-17T09:41:55.728Z + 2025-09-17T17:30:32.523Z https://docs.axolotl.ai/docs/api/datasets.html - 2025-09-17T09:41:54.391Z + 2025-09-17T17:30:31.185Z https://docs.axolotl.ai/docs/api/convert.html - 2025-09-17T09:41:54.404Z + 2025-09-17T17:30:31.199Z https://docs.axolotl.ai/docs/api/prompt_strategies.kto.llama3.html - 2025-09-17T09:41:55.032Z + 2025-09-17T17:30:31.828Z https://docs.axolotl.ai/docs/api/utils.tokenization.html - 2025-09-17T09:41:55.302Z + 2025-09-17T17:30:32.100Z https://docs.axolotl.ai/docs/api/core.trainers.base.html - 2025-09-17T09:41:54.748Z + 2025-09-17T17:30:31.545Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_expand_mask.html - 2025-09-17T09:41:55.198Z + 2025-09-17T17:30:31.996Z https://docs.axolotl.ai/docs/api/loaders.model.html - 2025-09-17T09:41:54.810Z + 2025-09-17T17:30:31.607Z https://docs.axolotl.ai/docs/api/logging_config.html - 2025-09-17T09:41:54.456Z + 2025-09-17T17:30:31.250Z https://docs.axolotl.ai/docs/api/core.chat.format.shared.html - 2025-09-17T09:41:54.512Z + 2025-09-17T17:30:31.306Z https://docs.axolotl.ai/docs/api/models.mamba.modeling_mamba.html - 2025-09-17T09:41:55.704Z + 2025-09-17T17:30:32.499Z https://docs.axolotl.ai/docs/api/utils.schemas.enums.html - 2025-09-17T09:41:55.518Z + 2025-09-17T17:30:32.315Z https://docs.axolotl.ai/docs/api/utils.callbacks.profiler.html - 2025-09-17T09:41:55.783Z + 2025-09-17T17:30:32.579Z https://docs.axolotl.ai/docs/api/utils.bench.html - 2025-09-17T09:41:55.317Z + 2025-09-17T17:30:32.116Z https://docs.axolotl.ai/docs/api/utils.samplers.multipack.html - 2025-09-17T09:41:55.773Z + 2025-09-17T17:30:32.569Z https://docs.axolotl.ai/docs/api/core.trainers.mamba.html - 2025-09-17T09:41:54.769Z + 2025-09-17T17:30:31.565Z https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_chat.html - 2025-09-17T09:41:54.925Z + 2025-09-17T17:30:31.722Z https://docs.axolotl.ai/docs/api/monkeypatch.btlm_attn_hijack_flash.html - 2025-09-17T09:41:55.236Z + 2025-09-17T17:30:32.034Z https://docs.axolotl.ai/docs/api/core.chat.messages.html - 2025-09-17T09:41:54.507Z + 2025-09-17T17:30:31.301Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_patch_multipack.html - 2025-09-17T09:41:55.238Z + 2025-09-17T17:30:32.036Z https://docs.axolotl.ai/docs/api/monkeypatch.mixtral.html - 2025-09-17T09:41:55.265Z + 2025-09-17T17:30:32.064Z https://docs.axolotl.ai/docs/api/core.trainers.mixins.scheduler.html - 2025-09-17T09:41:54.852Z + 2025-09-17T17:30:31.649Z https://docs.axolotl.ai/docs/api/loaders.patch_manager.html - 2025-09-17T09:41:54.835Z + 2025-09-17T17:30:31.632Z https://docs.axolotl.ai/docs/api/utils.optimizers.adopt.html - 2025-09-17T09:41:55.403Z + 2025-09-17T17:30:32.202Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_xformers.html - 2025-09-17T09:41:55.190Z + 2025-09-17T17:30:31.988Z https://docs.axolotl.ai/docs/api/utils.callbacks.perplexity.html - 2025-09-17T09:41:55.780Z + 2025-09-17T17:30:32.575Z https://docs.axolotl.ai/docs/api/loaders.tokenizer.html - 2025-09-17T09:41:54.819Z + 2025-09-17T17:30:31.615Z https://docs.axolotl.ai/docs/api/prompt_strategies.pygmalion.html - 2025-09-17T09:41:54.993Z + 2025-09-17T17:30:31.789Z https://docs.axolotl.ai/docs/api/utils.lora.html - 2025-09-17T09:41:55.308Z + 2025-09-17T17:30:32.107Z https://docs.axolotl.ai/docs/api/prompt_strategies.user_defined.html - 2025-09-17T09:41:54.946Z + 2025-09-17T17:30:31.743Z https://docs.axolotl.ai/docs/api/core.trainers.mixins.rng_state_loader.html - 2025-09-17T09:41:54.846Z + 2025-09-17T17:30:31.642Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.user_defined.html - 2025-09-17T09:41:55.022Z + 2025-09-17T17:30:31.818Z https://docs.axolotl.ai/docs/api/kernels.swiglu.html - 2025-09-17T09:41:55.174Z + 2025-09-17T17:30:31.972Z https://docs.axolotl.ai/docs/api/utils.schemas.multimodal.html - 2025-09-17T09:41:55.497Z + 2025-09-17T17:30:32.294Z https://docs.axolotl.ai/docs/api/cli.main.html - 2025-09-17T09:41:54.559Z + 2025-09-17T17:30:31.353Z https://docs.axolotl.ai/docs/api/utils.ctx_managers.sequence_parallel.html - 2025-09-17T09:41:54.876Z + 2025-09-17T17:30:31.673Z https://docs.axolotl.ai/docs/api/monkeypatch.multipack.html - 2025-09-17T09:41:55.193Z + 2025-09-17T17:30:31.991Z https://docs.axolotl.ai/docs/api/core.chat.format.chatml.html - 2025-09-17T09:41:54.509Z + 2025-09-17T17:30:31.303Z https://docs.axolotl.ai/docs/api/cli.cloud.base.html - 2025-09-17T09:41:54.686Z + 2025-09-17T17:30:31.482Z https://docs.axolotl.ai/docs/api/core.trainers.trl.html - 2025-09-17T09:41:54.763Z + 2025-09-17T17:30:31.560Z https://docs.axolotl.ai/docs/mixed_precision.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/installation.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/mac.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/sequence_parallelism.html - 2025-09-17T09:38:24.772Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/faq.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.016Z https://docs.axolotl.ai/docs/dataset-formats/tokenized.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/dataset-formats/stepwise_supervised.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/dataset-formats/pretraining.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/dataset-formats/template_free.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/dataset-formats/index.html - 2025-09-17T09:38:24.767Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/dataset-formats/conversation.html - 2025-09-17T09:38:24.767Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/dataset-formats/inst_tune.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/lr_groups.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/inference.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.018Z https://docs.axolotl.ai/docs/lora_optims.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/multipack.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/amd_hpc.html - 2025-09-17T09:38:24.767Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/api/monkeypatch.trainer_fsdp_optim.html - 2025-09-17T09:41:55.247Z + 2025-09-17T17:30:32.045Z https://docs.axolotl.ai/docs/api/cli.inference.html - 2025-09-17T09:41:54.642Z + 2025-09-17T17:30:31.437Z https://docs.axolotl.ai/docs/api/utils.collators.mm_chat.html - 2025-09-17T09:41:55.732Z + 2025-09-17T17:30:32.528Z https://docs.axolotl.ai/docs/api/utils.model_shard_quant.html - 2025-09-17T09:41:55.314Z + 2025-09-17T17:30:32.112Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_flash.html - 2025-09-17T09:41:55.189Z + 2025-09-17T17:30:31.986Z https://docs.axolotl.ai/docs/api/integrations.grokfast.optimizer.html - 2025-09-17T09:41:55.666Z + 2025-09-17T17:30:32.462Z https://docs.axolotl.ai/docs/api/utils.schemas.training.html - 2025-09-17T09:41:55.461Z + 2025-09-17T17:30:32.258Z https://docs.axolotl.ai/docs/api/utils.schemas.integrations.html - 2025-09-17T09:41:55.509Z + 2025-09-17T17:30:32.306Z https://docs.axolotl.ai/docs/api/utils.schemas.utils.html - 2025-09-17T09:41:55.523Z + 2025-09-17T17:30:32.321Z https://docs.axolotl.ai/docs/api/cli.checks.html - 2025-09-17T09:41:54.604Z + 2025-09-17T17:30:31.400Z https://docs.axolotl.ai/docs/api/utils.callbacks.qat.html - 2025-09-17T09:41:55.799Z + 2025-09-17T17:30:32.594Z https://docs.axolotl.ai/docs/api/monkeypatch.mistral_attn_hijack_flash.html - 2025-09-17T09:41:55.192Z + 2025-09-17T17:30:31.989Z https://docs.axolotl.ai/docs/api/utils.data.sft.html - 2025-09-17T09:41:55.411Z + 2025-09-17T17:30:32.210Z https://docs.axolotl.ai/docs/api/utils.chat_templates.html - 2025-09-17T09:41:55.303Z + 2025-09-17T17:30:32.102Z https://docs.axolotl.ai/docs/api/monkeypatch.utils.html - 2025-09-17T09:41:55.235Z + 2025-09-17T17:30:32.033Z https://docs.axolotl.ai/docs/api/cli.utils.fetch.html - 2025-09-17T09:41:54.710Z + 2025-09-17T17:30:31.507Z https://docs.axolotl.ai/docs/api/core.builders.causal.html - 2025-09-17T09:41:54.467Z + 2025-09-17T17:30:31.261Z https://docs.axolotl.ai/docs/api/utils.schemas.trl.html - 2025-09-17T09:41:55.491Z + 2025-09-17T17:30:32.289Z https://docs.axolotl.ai/docs/api/utils.data.streaming.html - 2025-09-17T09:41:55.405Z + 2025-09-17T17:30:32.204Z https://docs.axolotl.ai/docs/api/monkeypatch.unsloth_.html - 2025-09-17T09:41:55.255Z + 2025-09-17T17:30:32.053Z https://docs.axolotl.ai/docs/api/prompt_strategies.base.html - 2025-09-17T09:41:54.878Z + 2025-09-17T17:30:31.674Z https://docs.axolotl.ai/docs/api/cli.merge_lora.html - 2025-09-17T09:41:54.650Z + 2025-09-17T17:30:31.446Z https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_instruct.html - 2025-09-17T09:41:54.926Z + 2025-09-17T17:30:31.723Z https://docs.axolotl.ai/docs/api/loaders.constants.html - 2025-09-17T09:41:54.837Z + 2025-09-17T17:30:31.633Z https://docs.axolotl.ai/docs/api/monkeypatch.transformers_fa_utils.html - 2025-09-17T09:41:55.253Z + 2025-09-17T17:30:32.052Z https://docs.axolotl.ai/docs/api/monkeypatch.relora.html - 2025-09-17T09:41:55.197Z + 2025-09-17T17:30:31.994Z https://docs.axolotl.ai/docs/api/utils.schedulers.html - 2025-09-17T09:41:55.370Z + 2025-09-17T17:30:32.169Z https://docs.axolotl.ai/docs/api/kernels.lora.html - 2025-09-17T09:41:55.153Z + 2025-09-17T17:30:31.950Z https://docs.axolotl.ai/docs/api/utils.collators.batching.html - 2025-09-17T09:41:55.724Z + 2025-09-17T17:30:32.520Z https://docs.axolotl.ai/docs/api/prompt_strategies.completion.html - 2025-09-17T09:41:54.965Z + 2025-09-17T17:30:31.761Z https://docs.axolotl.ai/docs/api/core.trainers.grpo.trainer.html - 2025-09-17T09:41:54.787Z + 2025-09-17T17:30:31.583Z https://docs.axolotl.ai/docs/api/common.architectures.html - 2025-09-17T09:41:55.686Z + 2025-09-17T17:30:32.481Z https://docs.axolotl.ai/docs/api/cli.utils.html - 2025-09-17T09:41:54.693Z + 2025-09-17T17:30:31.490Z https://docs.axolotl.ai/docs/api/prompt_strategies.orpo.chat_template.html - 2025-09-17T09:41:55.061Z + 2025-09-17T17:30:31.858Z https://docs.axolotl.ai/docs/api/core.training_args.html - 2025-09-17T09:41:54.484Z + 2025-09-17T17:30:31.278Z https://docs.axolotl.ai/docs/api/utils.schemas.datasets.html - 2025-09-17T09:41:55.479Z + 2025-09-17T17:30:32.277Z https://docs.axolotl.ai/docs/api/prompt_strategies.chat_template.html - 2025-09-17T09:41:54.911Z + 2025-09-17T17:30:31.708Z https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_w_system.html - 2025-09-17T09:41:54.938Z + 2025-09-17T17:30:31.735Z https://docs.axolotl.ai/docs/api/integrations.cut_cross_entropy.args.html - 2025-09-17T09:41:55.665Z + 2025-09-17T17:30:32.461Z https://docs.axolotl.ai/docs/api/common.const.html - 2025-09-17T09:41:55.687Z + 2025-09-17T17:30:32.483Z https://docs.axolotl.ai/docs/api/monkeypatch.lora_kernels.html - 2025-09-17T09:41:55.227Z + 2025-09-17T17:30:32.025Z https://docs.axolotl.ai/docs/api/kernels.quantize.html - 2025-09-17T09:41:55.182Z + 2025-09-17T17:30:31.979Z https://docs.axolotl.ai/docs/api/cli.config.html - 2025-09-17T09:41:54.622Z + 2025-09-17T17:30:31.418Z https://docs.axolotl.ai/docs/api/prompt_strategies.llama2_chat.html - 2025-09-17T09:41:54.959Z + 2025-09-17T17:30:31.755Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.zephyr.html - 2025-09-17T09:41:55.020Z + 2025-09-17T17:30:31.817Z https://docs.axolotl.ai/docs/api/kernels.geglu.html - 2025-09-17T09:41:55.164Z + 2025-09-17T17:30:31.961Z https://docs.axolotl.ai/docs/api/utils.freeze.html - 2025-09-17T09:41:55.325Z + 2025-09-17T17:30:32.123Z https://docs.axolotl.ai/docs/api/utils.callbacks.mlflow_.html - 2025-09-17T09:41:55.788Z + 2025-09-17T17:30:32.584Z https://docs.axolotl.ai/docs/api/core.trainers.grpo.sampler.html - 2025-09-17T09:41:54.799Z + 2025-09-17T17:30:31.595Z https://docs.axolotl.ai/docs/api/cli.utils.train.html - 2025-09-17T09:41:54.734Z + 2025-09-17T17:30:31.530Z https://docs.axolotl.ai/docs/api/cli.utils.load.html - 2025-09-17T09:41:54.716Z + 2025-09-17T17:30:31.513Z https://docs.axolotl.ai/docs/api/utils.collators.core.html - 2025-09-17T09:41:55.705Z + 2025-09-17T17:30:32.500Z https://docs.axolotl.ai/docs/api/integrations.liger.args.html - 2025-09-17T09:41:55.677Z + 2025-09-17T17:30:32.473Z https://docs.axolotl.ai/docs/api/cli.preprocess.html - 2025-09-17T09:41:54.671Z + 2025-09-17T17:30:31.467Z https://docs.axolotl.ai/docs/api/kernels.utils.html - 2025-09-17T09:41:55.183Z + 2025-09-17T17:30:31.981Z https://docs.axolotl.ai/docs/api/utils.callbacks.lisa.html - 2025-09-17T09:41:55.785Z + 2025-09-17T17:30:32.580Z https://docs.axolotl.ai/docs/api/monkeypatch.data.batch_dataset_fetcher.html - 2025-09-17T09:41:55.264Z + 2025-09-17T17:30:32.062Z https://docs.axolotl.ai/docs/api/core.trainers.mixins.optimizer.html - 2025-09-17T09:41:54.842Z + 2025-09-17T17:30:31.639Z https://docs.axolotl.ai/docs/api/loaders.processor.html - 2025-09-17T09:41:54.820Z + 2025-09-17T17:30:31.616Z https://docs.axolotl.ai/docs/api/core.trainers.utils.html - 2025-09-17T09:41:54.801Z + 2025-09-17T17:30:31.597Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chat_template.html - 2025-09-17T09:41:54.998Z + 2025-09-17T17:30:31.795Z https://docs.axolotl.ai/docs/api/prompt_tokenizers.html - 2025-09-17T09:41:54.446Z + 2025-09-17T17:30:31.240Z https://docs.axolotl.ai/docs/api/utils.callbacks.comet_.html - 2025-09-17T09:41:55.792Z + 2025-09-17T17:30:32.587Z https://docs.axolotl.ai/docs/api/core.datasets.transforms.chat_builder.html - 2025-09-17T09:41:54.525Z + 2025-09-17T17:30:31.319Z https://docs.axolotl.ai/docs/api/utils.trainer.html - 2025-09-17T09:41:55.343Z + 2025-09-17T17:30:32.141Z https://docs.axolotl.ai/docs/api/cli.train.html - 2025-09-17T09:41:54.567Z + 2025-09-17T17:30:31.362Z https://docs.axolotl.ai/docs/api/integrations.lm_eval.args.html - 2025-09-17T09:41:55.681Z + 2025-09-17T17:30:32.477Z https://docs.axolotl.ai/docs/api/prompt_strategies.messages.chat.html - 2025-09-17T09:41:54.997Z + 2025-09-17T17:30:31.793Z https://docs.axolotl.ai/docs/api/utils.schemas.peft.html - 2025-09-17T09:41:55.488Z + 2025-09-17T17:30:32.285Z https://docs.axolotl.ai/docs/api/prompt_strategies.bradley_terry.llama3.html - 2025-09-17T09:41:55.065Z + 2025-09-17T17:30:31.862Z https://docs.axolotl.ai/docs/api/cli.utils.args.html - 2025-09-17T09:41:54.705Z + 2025-09-17T17:30:31.502Z https://docs.axolotl.ai/docs/api/prompt_strategies.stepwise_supervised.html - 2025-09-17T09:41:54.975Z + 2025-09-17T17:30:31.772Z https://docs.axolotl.ai/docs/api/common.datasets.html - 2025-09-17T09:41:55.703Z + 2025-09-17T17:30:32.498Z https://docs.axolotl.ai/docs/api/prompt_strategies.kto.user_defined.html - 2025-09-17T09:41:55.041Z + 2025-09-17T17:30:31.837Z https://docs.axolotl.ai/docs/api/train.html - 2025-09-17T09:41:54.374Z + 2025-09-17T17:30:31.168Z https://docs.axolotl.ai/docs/multi-node.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/input_output.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.018Z https://docs.axolotl.ai/docs/ray-integration.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/getting-started.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.016Z https://docs.axolotl.ai/docs/optimizers.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/multi-gpu.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/dataset_preprocessing.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.016Z https://docs.axolotl.ai/docs/torchao.html - 2025-09-17T09:38:24.772Z + 2025-09-17T17:27:12.020Z https://docs.axolotl.ai/docs/config-reference.html - 2025-09-17T09:42:11.046Z + 2025-09-17T17:30:47.286Z https://docs.axolotl.ai/docs/rlhf.html - 2025-09-17T09:38:24.772Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/docs/cli.html - 2025-09-17T09:38:24.767Z + 2025-09-17T17:27:12.015Z https://docs.axolotl.ai/docs/fsdp_qlora.html - 2025-09-17T09:38:24.768Z + 2025-09-17T17:27:12.016Z https://docs.axolotl.ai/docs/qat.html - 2025-09-17T09:38:24.771Z + 2025-09-17T17:27:12.019Z https://docs.axolotl.ai/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html - 2025-09-17T09:38:24.792Z + 2025-09-17T17:27:12.040Z https://docs.axolotl.ai/index.html - 2025-09-17T09:38:24.788Z + 2025-09-17T17:27:12.036Z
AxolotlLoggerA Logger that automatically rejects non-axolotl INFOs.Logger that applies filtering to non-axolotl loggers.
AxolotlOrWarnErrorFilterAllows ANY WARNING or higher (unless overridden by LOG_LEVEL)Allows ANY WARNING or higher (unless overridden by LOG_LEVEL). Allows axolotl.* at
ColorfulFormatter