diff --git a/.nojekyll b/.nojekyll
index 50b586081..20334997f 100644
--- a/.nojekyll
+++ b/.nojekyll
@@ -1 +1 @@
-999b5f89
\ No newline at end of file
+b4822350
\ No newline at end of file
diff --git a/docs/faq.html b/docs/faq.html
index 003eadc7a..fffc5f4a7 100644
--- a/docs/faq.html
+++ b/docs/faq.html
@@ -506,11 +506,11 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
A: Usually an issue with the GPUs communicating with each other. See the NCCL doc
-Q: Exitcode -9
+Q: exitcode: -9
A: This usually happens when you run out of system RAM.
-Q: Exitcode -7 while using deepspeed
+Q: exitcode: -7 while using deepspeed
A: Try upgrading deepspeed w: pip install -U deepspeed
diff --git a/search.json b/search.json
index 39c2ded53..a315ff45e 100644
--- a/search.json
+++ b/search.json
@@ -3260,7 +3260,7 @@
"href": "docs/faq.html",
"title": "FAQ",
"section": "",
- "text": "General\nQ: The trainer stopped and hasn’t progressed in several minutes.\n\nA: Usually an issue with the GPUs communicating with each other. See the NCCL doc\n\nQ: Exitcode -9\n\nA: This usually happens when you run out of system RAM.\n\nQ: Exitcode -7 while using deepspeed\n\nA: Try upgrading deepspeed w: pip install -U deepspeed\n\nQ: AttributeError: ‘DummyOptim’ object has no attribute ‘step’\nQ: ModuleNotFoundError: No module named ‘mpi4py’ using single GPU with deepspeed\n\nA: You may be using deepspeed with single gpu. Please remove the deepspeed: section in the yaml file or --deepspeed CLI flag.\n\nQ: The codes is stuck on saving preprocessed datasets.\n\nA: This is usually an issue with the GPU. This can be resolved through setting the os environment variable CUDA_VISIBLE_DEVICES=0. If you are on runpod, this is usually a pod issue. Starting a new pod should take care of it.\n\nQ: Received mismatch error on merge adapters / loading adapters between torch.Size of checkpoint and model.\n\nA: This is likely due to vocab size mismatch. By default, Axolotl expands the model’s embeddings if the tokenizer has more tokens than the model. Please use the axolotl merge-lora command to merge the adapters instead of using your own scripts.\n\n\nOn the other hand, if the model has more tokens than the tokenizer, Axolotl does not shrink the model’s embeddings unless shrink_embeddings: true is set in the config.\n\nQ: How to call Axolotl via custom python scripts?\n\nA: Since Axolotl is just Python, please see src/axolotl/cli/main.py on how each command is called.\n\nQ: How to know the value to use for fsdp_transformer_layer_cls_to_wrap?\n\nA: This is the class name of the transformer layer to wrap with FSDP. For example, for LlamaForCausalLM, the value is LlamaDecoderLayer. To find this for a specific model, check the model’s PreTrainedModel definition and look for _no_split_modules variable in the modeling_<model_name>.py file within transformers library.\n\nQ: ValueError: Asking to pad but the tokenizer does not have a padding token. Please select a token to use as pad_token\n\nA: This is because the tokenizer does not have a padding token. Please add a padding token to the tokenizer via:\n\n\nspecial_tokens:\n # str. If you're not sure, set to same as `eos_token`.\n pad_token: \"...\"\n\n\n\nChat templates\nQ: jinja2.exceptions.UndefinedError: 'dict object' has no attribute 'content' / 'role' / ____\n\nA: This means that the property mapping for the stated attribute does not exist when building chat_template prompt. For example, if no attribute 'content', please check you have added the correct mapping for content under message_property_mappings.\n\nQ: Empty template generated for turn ___\n\nA: The content is empty for that turn.\n\nQ: Could not find content start/end boundary for turn __\n\nA: The specific turn’s start/end could not be detected. Please ensure you have set the eos_token following your chat_template. Otherwise, this could be a chat_template which doesn’t use proper boundaries for each turn (like system). On the rare occurrence, make sure your content is not [[dummy_message]]. Please let us know about this.\n\nQ: Content end boundary is before start boundary for turn ___\n\nA: This is an edge case which should not occur. Please create an Issue if this happens.\n\nQ: Content end boundary is the same as start boundary for turn ___. This is likely an empty turn.\n\nA: This is likely an empty turn.\n\nQ: The EOS token is incorrectly being masked or not being masked / EOS token __ not found in chat template.\n\nA: There can be two reasons:\n\n\n\nThis is because of the mismatch between tokenizer.eos_token and EOS token in template. Please make sure to set eos_token: under special_tokens: to the same EOS token as in template.\n\n\n\n\nThe EOS token is not in the template. Please check if your template is correct. As an example, phi_35 template does not use its dedicated EOS token <|endoftext|> at the end.\n\n\nQ: “chat_template choice is tokenizer_default but tokenizer’s chat_template is null. Please add a chat_template in tokenizer config”\n\nA: This is because the tokenizer does not have a chat template. Please add a chat template in the tokenizer config. See chat_template for more details.\n\nQ: The EOT token(s) are incorrectly being masked or not being masked / EOT token __ not found in chat template.\n\nA: There can be two reasons:\n\n\n\nThe EOT token is different from the EOS token and was not specified under eot_tokens:. Please set eot_tokens: to the same EOT token(s) as in template.\n\n\n\n\nThere is more than one EOT token per turn in the template. Please raise an issue with examples as we recognize this as an edge case.\n\n\nQ: EOT token encoding failed. Please check if the token is valid and can be encoded.\n\nA: There could be some issue with the tokenizer or unicode encoding. Please raise an issue with examples with the EOT token & tokenizer causing the issue.\n\nQ: EOT token __ is encoded as multiple tokens.\n\nA: This is because the EOT token is encoded as multiple tokens which can cause unexpected behavior. Please add it under tokens: or (recommended) override unused added_tokens via added_tokens_overrides:.\n\nQ: Conflict between train_on_eos and train_on_eot. eos_token is in eot_tokens and train_on_eos != train_on_eot\n\nA: This is because the EOS token is in the eot_tokens: while mismatch between train_on_eos: and train_on_eot:. This will cause one to override the other. Please ensure that train_on_eos: and train_on_eot: are the same or remove the EOS token from eot_tokens:.\n\nQ: If eot_tokens: is not provided, what happens?\n\nA: If eot_tokens: is not provided, the default behavior is the same as before. EOS tokens used to delimit turns are masked/unmasked depending on whether the turn is trainable.\n\n\nInternally, eot_tokens: tokenizer.eos_token and train_on_eot: train_on_eos (which defaults to turn). This transition helps clarify the naming and behavior of EOT/EOS tokens.\n\nQ: Data processing error: CAS service error\n\nA: Try disabling XET with export HF_HUB_DISABLE_XET=1\n\nQ: torch._inductor.exc.LoweringException: NoValidChoicesError: No choices to select, please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice.\n\nA: Depending on the version of torch, you may need to include this in your YAML:\n\n\nflex_attn_compile_kwargs:\n dynamic: false\n mode: max-autotune-no-cudagraphs",
+ "text": "General\nQ: The trainer stopped and hasn’t progressed in several minutes.\n\nA: Usually an issue with the GPUs communicating with each other. See the NCCL doc\n\nQ: exitcode: -9\n\nA: This usually happens when you run out of system RAM.\n\nQ: exitcode: -7 while using deepspeed\n\nA: Try upgrading deepspeed w: pip install -U deepspeed\n\nQ: AttributeError: ‘DummyOptim’ object has no attribute ‘step’\nQ: ModuleNotFoundError: No module named ‘mpi4py’ using single GPU with deepspeed\n\nA: You may be using deepspeed with single gpu. Please remove the deepspeed: section in the yaml file or --deepspeed CLI flag.\n\nQ: The codes is stuck on saving preprocessed datasets.\n\nA: This is usually an issue with the GPU. This can be resolved through setting the os environment variable CUDA_VISIBLE_DEVICES=0. If you are on runpod, this is usually a pod issue. Starting a new pod should take care of it.\n\nQ: Received mismatch error on merge adapters / loading adapters between torch.Size of checkpoint and model.\n\nA: This is likely due to vocab size mismatch. By default, Axolotl expands the model’s embeddings if the tokenizer has more tokens than the model. Please use the axolotl merge-lora command to merge the adapters instead of using your own scripts.\n\n\nOn the other hand, if the model has more tokens than the tokenizer, Axolotl does not shrink the model’s embeddings unless shrink_embeddings: true is set in the config.\n\nQ: How to call Axolotl via custom python scripts?\n\nA: Since Axolotl is just Python, please see src/axolotl/cli/main.py on how each command is called.\n\nQ: How to know the value to use for fsdp_transformer_layer_cls_to_wrap?\n\nA: This is the class name of the transformer layer to wrap with FSDP. For example, for LlamaForCausalLM, the value is LlamaDecoderLayer. To find this for a specific model, check the model’s PreTrainedModel definition and look for _no_split_modules variable in the modeling_<model_name>.py file within transformers library.\n\nQ: ValueError: Asking to pad but the tokenizer does not have a padding token. Please select a token to use as pad_token\n\nA: This is because the tokenizer does not have a padding token. Please add a padding token to the tokenizer via:\n\n\nspecial_tokens:\n # str. If you're not sure, set to same as `eos_token`.\n pad_token: \"...\"\n\n\n\nChat templates\nQ: jinja2.exceptions.UndefinedError: 'dict object' has no attribute 'content' / 'role' / ____\n\nA: This means that the property mapping for the stated attribute does not exist when building chat_template prompt. For example, if no attribute 'content', please check you have added the correct mapping for content under message_property_mappings.\n\nQ: Empty template generated for turn ___\n\nA: The content is empty for that turn.\n\nQ: Could not find content start/end boundary for turn __\n\nA: The specific turn’s start/end could not be detected. Please ensure you have set the eos_token following your chat_template. Otherwise, this could be a chat_template which doesn’t use proper boundaries for each turn (like system). On the rare occurrence, make sure your content is not [[dummy_message]]. Please let us know about this.\n\nQ: Content end boundary is before start boundary for turn ___\n\nA: This is an edge case which should not occur. Please create an Issue if this happens.\n\nQ: Content end boundary is the same as start boundary for turn ___. This is likely an empty turn.\n\nA: This is likely an empty turn.\n\nQ: The EOS token is incorrectly being masked or not being masked / EOS token __ not found in chat template.\n\nA: There can be two reasons:\n\n\n\nThis is because of the mismatch between tokenizer.eos_token and EOS token in template. Please make sure to set eos_token: under special_tokens: to the same EOS token as in template.\n\n\n\n\nThe EOS token is not in the template. Please check if your template is correct. As an example, phi_35 template does not use its dedicated EOS token <|endoftext|> at the end.\n\n\nQ: “chat_template choice is tokenizer_default but tokenizer’s chat_template is null. Please add a chat_template in tokenizer config”\n\nA: This is because the tokenizer does not have a chat template. Please add a chat template in the tokenizer config. See chat_template for more details.\n\nQ: The EOT token(s) are incorrectly being masked or not being masked / EOT token __ not found in chat template.\n\nA: There can be two reasons:\n\n\n\nThe EOT token is different from the EOS token and was not specified under eot_tokens:. Please set eot_tokens: to the same EOT token(s) as in template.\n\n\n\n\nThere is more than one EOT token per turn in the template. Please raise an issue with examples as we recognize this as an edge case.\n\n\nQ: EOT token encoding failed. Please check if the token is valid and can be encoded.\n\nA: There could be some issue with the tokenizer or unicode encoding. Please raise an issue with examples with the EOT token & tokenizer causing the issue.\n\nQ: EOT token __ is encoded as multiple tokens.\n\nA: This is because the EOT token is encoded as multiple tokens which can cause unexpected behavior. Please add it under tokens: or (recommended) override unused added_tokens via added_tokens_overrides:.\n\nQ: Conflict between train_on_eos and train_on_eot. eos_token is in eot_tokens and train_on_eos != train_on_eot\n\nA: This is because the EOS token is in the eot_tokens: while mismatch between train_on_eos: and train_on_eot:. This will cause one to override the other. Please ensure that train_on_eos: and train_on_eot: are the same or remove the EOS token from eot_tokens:.\n\nQ: If eot_tokens: is not provided, what happens?\n\nA: If eot_tokens: is not provided, the default behavior is the same as before. EOS tokens used to delimit turns are masked/unmasked depending on whether the turn is trainable.\n\n\nInternally, eot_tokens: tokenizer.eos_token and train_on_eot: train_on_eos (which defaults to turn). This transition helps clarify the naming and behavior of EOT/EOS tokens.\n\nQ: Data processing error: CAS service error\n\nA: Try disabling XET with export HF_HUB_DISABLE_XET=1\n\nQ: torch._inductor.exc.LoweringException: NoValidChoicesError: No choices to select, please consider adding ATEN into max_autotune_gemm_backends config (defined in torch/_inductor/config.py) to allow at least one choice.\n\nA: Depending on the version of torch, you may need to include this in your YAML:\n\n\nflex_attn_compile_kwargs:\n dynamic: false\n mode: max-autotune-no-cudagraphs",
"crumbs": [
"Troubleshooting",
"FAQ"
diff --git a/sitemap.xml b/sitemap.xml
index 932cdaedd..9059b8e61 100644
--- a/sitemap.xml
+++ b/sitemap.xml
@@ -2,758 +2,758 @@
https://docs.axolotl.ai/docs/unsloth.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/dataset-formats/conversation.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/dataset-formats/stepwise_supervised.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/dataset-formats/tokenized.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/mac.html
- 2025-06-19T15:17:03.062Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/nccl.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/multi-node.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/docker.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.768Z
https://docs.axolotl.ai/docs/lr_groups.html
- 2025-06-19T15:17:03.062Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/inference.html
- 2025-06-19T15:17:03.062Z
+ 2025-06-19T15:28:07.770Z
https://docs.axolotl.ai/docs/cli.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/config-reference.html
- 2025-06-19T15:20:20.894Z
+ 2025-06-19T15:31:37.242Z
https://docs.axolotl.ai/docs/multi-gpu.html
- 2025-06-19T15:17:03.062Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/debugging.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.768Z
https://docs.axolotl.ai/docs/multimodal.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/api/cli.sweeps.html
- 2025-06-19T15:20:07.487Z
+ 2025-06-19T15:31:24.676Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.llama3.html
- 2025-06-19T15:20:07.810Z
+ 2025-06-19T15:31:25.001Z
https://docs.axolotl.ai/docs/api/utils.schedulers.html
- 2025-06-19T15:20:08.198Z
+ 2025-06-19T15:31:25.394Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_xformers.html
- 2025-06-19T15:20:08.006Z
+ 2025-06-19T15:31:25.199Z
https://docs.axolotl.ai/docs/api/cli.cloud.modal_.html
- 2025-06-19T15:20:07.535Z
+ 2025-06-19T15:31:24.724Z
https://docs.axolotl.ai/docs/api/kernels.geglu.html
- 2025-06-19T15:20:07.960Z
+ 2025-06-19T15:31:25.153Z
https://docs.axolotl.ai/docs/api/core.trainers.utils.html
- 2025-06-19T15:20:07.607Z
+ 2025-06-19T15:31:24.796Z
https://docs.axolotl.ai/docs/api/core.datasets.chat.html
- 2025-06-19T15:20:07.339Z
+ 2025-06-19T15:31:24.528Z
https://docs.axolotl.ai/docs/api/utils.schemas.peft.html
- 2025-06-19T15:20:08.309Z
+ 2025-06-19T15:31:25.515Z
https://docs.axolotl.ai/docs/api/monkeypatch.btlm_attn_hijack_flash.html
- 2025-06-19T15:20:08.068Z
+ 2025-06-19T15:31:25.261Z
https://docs.axolotl.ai/docs/api/prompt_strategies.chat_template.html
- 2025-06-19T15:20:07.714Z
+ 2025-06-19T15:31:24.903Z
https://docs.axolotl.ai/docs/api/prompt_strategies.kto.user_defined.html
- 2025-06-19T15:20:07.842Z
+ 2025-06-19T15:31:25.033Z
https://docs.axolotl.ai/docs/api/cli.cloud.base.html
- 2025-06-19T15:20:07.528Z
+ 2025-06-19T15:31:24.718Z
https://docs.axolotl.ai/docs/api/kernels.swiglu.html
- 2025-06-19T15:20:07.970Z
+ 2025-06-19T15:31:25.163Z
https://docs.axolotl.ai/docs/api/prompt_strategies.stepwise_supervised.html
- 2025-06-19T15:20:07.777Z
+ 2025-06-19T15:31:24.968Z
https://docs.axolotl.ai/docs/api/prompt_strategies.bradley_terry.llama3.html
- 2025-06-19T15:20:07.866Z
+ 2025-06-19T15:31:25.057Z
https://docs.axolotl.ai/docs/api/prompt_strategies.completion.html
- 2025-06-19T15:20:07.767Z
+ 2025-06-19T15:31:24.957Z
https://docs.axolotl.ai/docs/api/kernels.utils.html
- 2025-06-19T15:20:07.979Z
+ 2025-06-19T15:31:25.172Z
https://docs.axolotl.ai/docs/api/common.datasets.html
- 2025-06-19T15:20:08.518Z
+ 2025-06-19T15:31:25.726Z
https://docs.axolotl.ai/docs/api/utils.schemas.datasets.html
- 2025-06-19T15:20:08.300Z
+ 2025-06-19T15:31:25.503Z
https://docs.axolotl.ai/docs/api/core.builders.rl.html
- 2025-06-19T15:20:07.294Z
+ 2025-06-19T15:31:24.483Z
https://docs.axolotl.ai/docs/api/evaluate.html
- 2025-06-19T15:20:07.201Z
+ 2025-06-19T15:31:24.387Z
https://docs.axolotl.ai/docs/api/kernels.quantize.html
- 2025-06-19T15:20:07.978Z
+ 2025-06-19T15:31:25.170Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_flash.html
- 2025-06-19T15:20:08.005Z
+ 2025-06-19T15:31:25.197Z
https://docs.axolotl.ai/docs/api/core.trainers.mixins.rng_state_loader.html
- 2025-06-19T15:20:07.650Z
+ 2025-06-19T15:31:24.839Z
https://docs.axolotl.ai/docs/api/integrations.base.html
- 2025-06-19T15:20:08.479Z
+ 2025-06-19T15:31:25.687Z
https://docs.axolotl.ai/docs/api/cli.merge_lora.html
- 2025-06-19T15:20:07.460Z
+ 2025-06-19T15:31:24.650Z
https://docs.axolotl.ai/docs/api/cli.merge_sharded_fsdp_weights.html
- 2025-06-19T15:20:07.472Z
+ 2025-06-19T15:31:24.662Z
https://docs.axolotl.ai/docs/api/monkeypatch.transformers_fa_utils.html
- 2025-06-19T15:20:08.085Z
+ 2025-06-19T15:31:25.278Z
https://docs.axolotl.ai/docs/api/prompt_strategies.llama2_chat.html
- 2025-06-19T15:20:07.761Z
+ 2025-06-19T15:31:24.951Z
https://docs.axolotl.ai/docs/api/utils.collators.mm_chat.html
- 2025-06-19T15:20:08.548Z
+ 2025-06-19T15:31:25.756Z
https://docs.axolotl.ai/docs/api/utils.data.sft.html
- 2025-06-19T15:20:08.237Z
+ 2025-06-19T15:31:25.434Z
https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_instruct.html
- 2025-06-19T15:20:07.728Z
+ 2025-06-19T15:31:24.918Z
https://docs.axolotl.ai/docs/api/integrations.liger.args.html
- 2025-06-19T15:20:08.494Z
+ 2025-06-19T15:31:25.702Z
https://docs.axolotl.ai/docs/api/monkeypatch.mistral_attn_hijack_flash.html
- 2025-06-19T15:20:08.021Z
+ 2025-06-19T15:31:25.214Z
https://docs.axolotl.ai/docs/api/cli.vllm_serve.html
- 2025-06-19T15:20:07.525Z
+ 2025-06-19T15:31:24.715Z
https://docs.axolotl.ai/docs/api/monkeypatch.utils.html
- 2025-06-19T15:20:08.067Z
+ 2025-06-19T15:31:25.260Z
https://docs.axolotl.ai/docs/api/loaders.patch_manager.html
- 2025-06-19T15:20:07.640Z
+ 2025-06-19T15:31:24.829Z
https://docs.axolotl.ai/docs/api/utils.schemas.integrations.html
- 2025-06-19T15:20:08.329Z
+ 2025-06-19T15:31:25.535Z
https://docs.axolotl.ai/docs/api/utils.callbacks.perplexity.html
- 2025-06-19T15:20:08.594Z
+ 2025-06-19T15:31:25.802Z
https://docs.axolotl.ai/docs/api/cli.utils.html
- 2025-06-19T15:20:07.518Z
+ 2025-06-19T15:31:24.708Z
https://docs.axolotl.ai/docs/api/utils.schemas.config.html
- 2025-06-19T15:20:08.271Z
+ 2025-06-19T15:31:25.469Z
https://docs.axolotl.ai/docs/api/prompt_strategies.input_output.html
- 2025-06-19T15:20:07.772Z
+ 2025-06-19T15:31:24.963Z
https://docs.axolotl.ai/docs/api/utils.distributed.html
- 2025-06-19T15:20:08.218Z
+ 2025-06-19T15:31:25.415Z
https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_disk.html
- 2025-06-19T15:20:08.118Z
+ 2025-06-19T15:31:25.312Z
https://docs.axolotl.ai/docs/api/monkeypatch.trainer_fsdp_optim.html
- 2025-06-19T15:20:08.079Z
+ 2025-06-19T15:31:25.272Z
https://docs.axolotl.ai/docs/api/core.builders.base.html
- 2025-06-19T15:20:07.282Z
+ 2025-06-19T15:31:24.470Z
https://docs.axolotl.ai/docs/api/core.trainers.trl.html
- 2025-06-19T15:20:07.566Z
+ 2025-06-19T15:31:24.756Z
https://docs.axolotl.ai/docs/api/cli.evaluate.html
- 2025-06-19T15:20:07.394Z
+ 2025-06-19T15:31:24.584Z
https://docs.axolotl.ai/docs/api/utils.optimizers.adopt.html
- 2025-06-19T15:20:08.229Z
+ 2025-06-19T15:31:25.426Z
https://docs.axolotl.ai/docs/api/utils.callbacks.qat.html
- 2025-06-19T15:20:08.613Z
+ 2025-06-19T15:31:25.821Z
https://docs.axolotl.ai/docs/api/core.trainers.dpo.trainer.html
- 2025-06-19T15:20:07.583Z
+ 2025-06-19T15:31:24.772Z
https://docs.axolotl.ai/docs/api/core.chat.format.shared.html
- 2025-06-19T15:20:07.334Z
+ 2025-06-19T15:31:24.523Z
https://docs.axolotl.ai/docs/api/monkeypatch.relora.html
- 2025-06-19T15:20:08.029Z
+ 2025-06-19T15:31:25.222Z
https://docs.axolotl.ai/docs/api/cli.config.html
- 2025-06-19T15:20:07.438Z
+ 2025-06-19T15:31:24.627Z
https://docs.axolotl.ai/docs/api/cli.preprocess.html
- 2025-06-19T15:20:07.481Z
+ 2025-06-19T15:31:24.670Z
https://docs.axolotl.ai/docs/api/core.trainers.base.html
- 2025-06-19T15:20:07.550Z
+ 2025-06-19T15:31:24.740Z
https://docs.axolotl.ai/docs/api/convert.html
- 2025-06-19T15:20:07.225Z
+ 2025-06-19T15:31:24.412Z
https://docs.axolotl.ai/docs/api/prompt_strategies.pygmalion.html
- 2025-06-19T15:20:07.794Z
+ 2025-06-19T15:31:24.985Z
https://docs.axolotl.ai/docs/api/utils.schemas.trl.html
- 2025-06-19T15:20:08.312Z
+ 2025-06-19T15:31:25.518Z
https://docs.axolotl.ai/docs/api/cli.args.html
- 2025-06-19T15:20:07.413Z
+ 2025-06-19T15:31:24.603Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chat_template.html
- 2025-06-19T15:20:07.800Z
+ 2025-06-19T15:31:24.990Z
https://docs.axolotl.ai/docs/api/loaders.constants.html
- 2025-06-19T15:20:07.642Z
+ 2025-06-19T15:31:24.830Z
https://docs.axolotl.ai/docs/api/logging_config.html
- 2025-06-19T15:20:07.275Z
+ 2025-06-19T15:31:24.464Z
https://docs.axolotl.ai/docs/api/cli.inference.html
- 2025-06-19T15:20:07.452Z
+ 2025-06-19T15:31:24.642Z
https://docs.axolotl.ai/docs/api/utils.ctx_managers.sequence_parallel.html
- 2025-06-19T15:20:07.680Z
+ 2025-06-19T15:31:24.869Z
https://docs.axolotl.ai/docs/api/integrations.spectrum.args.html
- 2025-06-19T15:20:08.501Z
+ 2025-06-19T15:31:25.708Z
https://docs.axolotl.ai/docs/api/utils.schemas.training.html
- 2025-06-19T15:20:08.283Z
+ 2025-06-19T15:31:25.481Z
https://docs.axolotl.ai/docs/api/prompt_strategies.orcamini.html
- 2025-06-19T15:20:07.787Z
+ 2025-06-19T15:31:24.978Z
https://docs.axolotl.ai/docs/api/utils.freeze.html
- 2025-06-19T15:20:08.157Z
+ 2025-06-19T15:31:25.350Z
https://docs.axolotl.ai/docs/api/loaders.tokenizer.html
- 2025-06-19T15:20:07.625Z
+ 2025-06-19T15:31:24.814Z
https://docs.axolotl.ai/docs/api/utils.bench.html
- 2025-06-19T15:20:08.149Z
+ 2025-06-19T15:31:25.343Z
https://docs.axolotl.ai/docs/api/utils.quantization.html
- 2025-06-19T15:20:08.258Z
+ 2025-06-19T15:31:25.455Z
https://docs.axolotl.ai/docs/batch_vs_grad.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/input_output.html
- 2025-06-19T15:17:03.062Z
+ 2025-06-19T15:28:07.770Z
https://docs.axolotl.ai/docs/sequence_parallelism.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/reward_modelling.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/index.html
- 2025-06-19T15:17:03.076Z
+ 2025-06-19T15:28:07.784Z
https://docs.axolotl.ai/src/axolotl/integrations/LICENSE.html
- 2025-06-19T15:17:03.080Z
+ 2025-06-19T15:28:07.788Z
https://docs.axolotl.ai/FAQS.html
- 2025-06-19T15:17:03.057Z
+ 2025-06-19T15:28:07.765Z
https://docs.axolotl.ai/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html
- 2025-06-19T15:17:03.080Z
+ 2025-06-19T15:28:07.788Z
https://docs.axolotl.ai/TODO.html
- 2025-06-19T15:17:03.057Z
+ 2025-06-19T15:28:07.766Z
https://docs.axolotl.ai/examples/colab-notebooks/colab-axolotl-example.html
- 2025-06-19T15:17:03.064Z
+ 2025-06-19T15:28:07.772Z
https://docs.axolotl.ai/docs/torchao.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/ray-integration.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/quantize.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/qat.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/api/utils.lora.html
- 2025-06-19T15:20:08.140Z
+ 2025-06-19T15:31:25.334Z
https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_w_system.html
- 2025-06-19T15:20:07.740Z
+ 2025-06-19T15:31:24.931Z
https://docs.axolotl.ai/docs/api/monkeypatch.stablelm_attn_hijack_flash.html
- 2025-06-19T15:20:08.075Z
+ 2025-06-19T15:31:25.269Z
https://docs.axolotl.ai/docs/api/utils.collators.core.html
- 2025-06-19T15:20:08.521Z
+ 2025-06-19T15:31:25.729Z
https://docs.axolotl.ai/docs/api/prompt_strategies.metharme.html
- 2025-06-19T15:20:07.784Z
+ 2025-06-19T15:31:24.974Z
https://docs.axolotl.ai/docs/api/utils.callbacks.profiler.html
- 2025-06-19T15:20:08.597Z
+ 2025-06-19T15:31:25.806Z
https://docs.axolotl.ai/docs/api/utils.data.pretraining.html
- 2025-06-19T15:20:08.230Z
+ 2025-06-19T15:31:25.427Z
https://docs.axolotl.ai/docs/api/utils.callbacks.lisa.html
- 2025-06-19T15:20:08.599Z
+ 2025-06-19T15:31:25.807Z
https://docs.axolotl.ai/docs/api/utils.trainer.html
- 2025-06-19T15:20:08.174Z
+ 2025-06-19T15:31:25.368Z
https://docs.axolotl.ai/docs/api/integrations.cut_cross_entropy.args.html
- 2025-06-19T15:20:08.482Z
+ 2025-06-19T15:31:25.690Z
https://docs.axolotl.ai/docs/api/utils.schemas.model.html
- 2025-06-19T15:20:08.278Z
+ 2025-06-19T15:31:25.476Z
https://docs.axolotl.ai/docs/api/monkeypatch.data.batch_dataset_fetcher.html
- 2025-06-19T15:20:08.088Z
+ 2025-06-19T15:31:25.281Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.zephyr.html
- 2025-06-19T15:20:07.821Z
+ 2025-06-19T15:31:25.012Z
https://docs.axolotl.ai/docs/api/datasets.html
- 2025-06-19T15:20:07.211Z
+ 2025-06-19T15:31:24.398Z
https://docs.axolotl.ai/docs/api/utils.schemas.enums.html
- 2025-06-19T15:20:08.340Z
+ 2025-06-19T15:31:25.546Z
https://docs.axolotl.ai/docs/api/integrations.kd.trainer.html
- 2025-06-19T15:20:08.491Z
+ 2025-06-19T15:31:25.698Z
https://docs.axolotl.ai/docs/api/monkeypatch.lora_kernels.html
- 2025-06-19T15:20:08.059Z
+ 2025-06-19T15:31:25.252Z
https://docs.axolotl.ai/docs/api/utils.collators.batching.html
- 2025-06-19T15:20:08.540Z
+ 2025-06-19T15:31:25.747Z
https://docs.axolotl.ai/docs/api/core.trainers.grpo.sampler.html
- 2025-06-19T15:20:07.606Z
+ 2025-06-19T15:31:24.795Z
https://docs.axolotl.ai/docs/api/prompt_strategies.base.html
- 2025-06-19T15:20:07.681Z
+ 2025-06-19T15:31:24.870Z
https://docs.axolotl.ai/docs/api/monkeypatch.multipack.html
- 2025-06-19T15:20:08.022Z
+ 2025-06-19T15:31:25.215Z
https://docs.axolotl.ai/docs/api/prompt_strategies.orpo.chat_template.html
- 2025-06-19T15:20:07.862Z
+ 2025-06-19T15:31:25.053Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.passthrough.html
- 2025-06-19T15:20:07.824Z
+ 2025-06-19T15:31:25.015Z
https://docs.axolotl.ai/docs/api/core.chat.format.chatml.html
- 2025-06-19T15:20:07.331Z
+ 2025-06-19T15:31:24.520Z
https://docs.axolotl.ai/docs/api/core.trainers.mixins.scheduler.html
- 2025-06-19T15:20:07.657Z
+ 2025-06-19T15:31:24.846Z
https://docs.axolotl.ai/docs/api/utils.model_shard_quant.html
- 2025-06-19T15:20:08.145Z
+ 2025-06-19T15:31:25.339Z
https://docs.axolotl.ai/docs/api/prompt_strategies.kto.chatml.html
- 2025-06-19T15:20:07.840Z
+ 2025-06-19T15:31:25.031Z
https://docs.axolotl.ai/docs/api/utils.tokenization.html
- 2025-06-19T15:20:08.125Z
+ 2025-06-19T15:31:25.319Z
https://docs.axolotl.ai/docs/api/loaders.model.html
- 2025-06-19T15:20:07.617Z
+ 2025-06-19T15:31:24.806Z
https://docs.axolotl.ai/docs/api/utils.callbacks.mlflow_.html
- 2025-06-19T15:20:08.602Z
+ 2025-06-19T15:31:25.811Z
https://docs.axolotl.ai/docs/api/core.trainers.grpo.trainer.html
- 2025-06-19T15:20:07.594Z
+ 2025-06-19T15:31:24.782Z
https://docs.axolotl.ai/docs/api/cli.main.html
- 2025-06-19T15:20:07.377Z
+ 2025-06-19T15:31:24.567Z
https://docs.axolotl.ai/docs/api/utils.callbacks.comet_.html
- 2025-06-19T15:20:08.606Z
+ 2025-06-19T15:31:25.814Z
https://docs.axolotl.ai/docs/api/utils.chat_templates.html
- 2025-06-19T15:20:08.135Z
+ 2025-06-19T15:31:25.329Z
https://docs.axolotl.ai/docs/api/utils.schemas.utils.html
- 2025-06-19T15:20:08.345Z
+ 2025-06-19T15:31:25.552Z
https://docs.axolotl.ai/docs/api/common.architectures.html
- 2025-06-19T15:20:08.502Z
+ 2025-06-19T15:31:25.710Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_expand_mask.html
- 2025-06-19T15:20:08.030Z
+ 2025-06-19T15:31:25.223Z
https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_chat.html
- 2025-06-19T15:20:07.727Z
+ 2025-06-19T15:31:24.917Z
https://docs.axolotl.ai/docs/api/utils.samplers.multipack.html
- 2025-06-19T15:20:08.588Z
+ 2025-06-19T15:31:25.796Z
https://docs.axolotl.ai/docs/api/integrations.grokfast.optimizer.html
- 2025-06-19T15:20:08.483Z
+ 2025-06-19T15:31:25.691Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chatml.html
- 2025-06-19T15:20:07.820Z
+ 2025-06-19T15:31:25.011Z
https://docs.axolotl.ai/docs/api/monkeypatch.mixtral.html
- 2025-06-19T15:20:08.089Z
+ 2025-06-19T15:31:25.283Z
https://docs.axolotl.ai/docs/api/train.html
- 2025-06-19T15:20:07.190Z
+ 2025-06-19T15:31:24.377Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_patch_multipack.html
- 2025-06-19T15:20:08.070Z
+ 2025-06-19T15:31:25.263Z
https://docs.axolotl.ai/docs/api/index.html
- 2025-06-19T15:20:07.129Z
+ 2025-06-19T15:31:24.314Z
https://docs.axolotl.ai/docs/api/loaders.adapter.html
- 2025-06-19T15:20:07.632Z
+ 2025-06-19T15:31:24.821Z
https://docs.axolotl.ai/docs/api/utils.schemas.multimodal.html
- 2025-06-19T15:20:08.317Z
+ 2025-06-19T15:31:25.523Z
https://docs.axolotl.ai/docs/api/kernels.lora.html
- 2025-06-19T15:20:07.950Z
+ 2025-06-19T15:31:25.142Z
https://docs.axolotl.ai/docs/api/prompt_strategies.kto.llama3.html
- 2025-06-19T15:20:07.832Z
+ 2025-06-19T15:31:25.023Z
https://docs.axolotl.ai/docs/api/cli.checks.html
- 2025-06-19T15:20:07.420Z
+ 2025-06-19T15:31:24.609Z
https://docs.axolotl.ai/docs/api/cli.quantize.html
- 2025-06-19T15:20:07.540Z
+ 2025-06-19T15:31:24.729Z
https://docs.axolotl.ai/docs/api/integrations.lm_eval.args.html
- 2025-06-19T15:20:08.497Z
+ 2025-06-19T15:31:25.705Z
https://docs.axolotl.ai/docs/api/core.chat.messages.html
- 2025-06-19T15:20:07.329Z
+ 2025-06-19T15:31:24.518Z
https://docs.axolotl.ai/docs/api/core.builders.causal.html
- 2025-06-19T15:20:07.286Z
+ 2025-06-19T15:31:24.475Z
https://docs.axolotl.ai/docs/api/core.trainers.relora.html
- 2025-06-19T15:20:07.577Z
+ 2025-06-19T15:31:24.765Z
https://docs.axolotl.ai/docs/api/models.mamba.modeling_mamba.html
- 2025-06-19T15:20:08.519Z
+ 2025-06-19T15:31:25.727Z
https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_cpu.html
- 2025-06-19T15:20:08.093Z
+ 2025-06-19T15:31:25.286Z
https://docs.axolotl.ai/docs/api/core.trainers.mamba.html
- 2025-06-19T15:20:07.572Z
+ 2025-06-19T15:31:24.761Z
https://docs.axolotl.ai/docs/api/core.datasets.transforms.chat_builder.html
- 2025-06-19T15:20:07.346Z
+ 2025-06-19T15:31:24.535Z
https://docs.axolotl.ai/docs/api/loaders.processor.html
- 2025-06-19T15:20:07.627Z
+ 2025-06-19T15:31:24.816Z
https://docs.axolotl.ai/docs/api/core.chat.format.llama3x.html
- 2025-06-19T15:20:07.332Z
+ 2025-06-19T15:31:24.521Z
https://docs.axolotl.ai/docs/api/prompt_strategies.messages.chat.html
- 2025-06-19T15:20:07.798Z
+ 2025-06-19T15:31:24.989Z
https://docs.axolotl.ai/docs/api/cli.train.html
- 2025-06-19T15:20:07.385Z
+ 2025-06-19T15:31:24.575Z
https://docs.axolotl.ai/docs/api/core.trainers.mixins.optimizer.html
- 2025-06-19T15:20:07.647Z
+ 2025-06-19T15:31:24.836Z
https://docs.axolotl.ai/docs/api/utils.collators.mamba.html
- 2025-06-19T15:20:08.543Z
+ 2025-06-19T15:31:25.751Z
https://docs.axolotl.ai/docs/api/monkeypatch.unsloth_.html
- 2025-06-19T15:20:08.087Z
+ 2025-06-19T15:31:25.280Z
https://docs.axolotl.ai/docs/api/utils.dict.html
- 2025-06-19T15:20:08.221Z
+ 2025-06-19T15:31:25.418Z
https://docs.axolotl.ai/docs/api/prompt_strategies.user_defined.html
- 2025-06-19T15:20:07.748Z
+ 2025-06-19T15:31:24.939Z
https://docs.axolotl.ai/docs/api/core.training_args.html
- 2025-06-19T15:20:07.306Z
+ 2025-06-19T15:31:24.495Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.user_defined.html
- 2025-06-19T15:20:07.823Z
+ 2025-06-19T15:31:25.014Z
https://docs.axolotl.ai/docs/api/prompt_tokenizers.html
- 2025-06-19T15:20:07.266Z
+ 2025-06-19T15:31:24.454Z
https://docs.axolotl.ai/docs/api/common.const.html
- 2025-06-19T15:20:08.503Z
+ 2025-06-19T15:31:25.711Z
https://docs.axolotl.ai/docs/fsdp_qlora.html
- 2025-06-19T15:17:03.060Z
+ 2025-06-19T15:28:07.768Z
https://docs.axolotl.ai/docs/custom_integrations.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/getting-started.html
- 2025-06-19T15:17:03.060Z
+ 2025-06-19T15:28:07.768Z
https://docs.axolotl.ai/docs/faq.html
- 2025-06-19T15:17:03.060Z
+ 2025-06-19T15:28:07.768Z
https://docs.axolotl.ai/docs/lora_optims.html
- 2025-06-19T15:17:03.062Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/rlhf.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/amd_hpc.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/installation.html
- 2025-06-19T15:17:03.062Z
+ 2025-06-19T15:28:07.770Z
https://docs.axolotl.ai/docs/multipack.html
- 2025-06-19T15:17:03.063Z
+ 2025-06-19T15:28:07.771Z
https://docs.axolotl.ai/docs/dataset_preprocessing.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.768Z
https://docs.axolotl.ai/docs/dataset_loading.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/dataset-formats/inst_tune.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/dataset-formats/template_free.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/dataset-formats/index.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z
https://docs.axolotl.ai/docs/dataset-formats/pretraining.html
- 2025-06-19T15:17:03.059Z
+ 2025-06-19T15:28:07.767Z