diff --git a/search.json b/search.json
index 10001c63b..2ee30a320 100644
--- a/search.json
+++ b/search.json
@@ -2989,14 +2989,14 @@
"href": "docs/api/train.html",
"title": "train",
"section": "",
- "text": "train\nPrepare and train a model on a dataset. Can also infer from a model or merge lora\n\n\n\n\n\nName\nDescription\n\n\n\n\ncreate_model_card\nCreate a model card for the trained model if needed.\n\n\ndetermine_resume_checkpoint\nDetermine the checkpoint to resume from based on configuration.\n\n\nexecute_training\nExecute the training process with appropriate SDP kernel configurations.\n\n\nhandle_untrained_tokens_fix\nApply fixes for untrained tokens if configured.\n\n\nsave_initial_configs\nSave initial configurations before training.\n\n\nsave_trained_model\nSave the trained model according to configuration and training setup.\n\n\nsetup_model_and_tokenizer\nLoad the tokenizer, processor (for multimodal models), and model based on\n\n\nsetup_model_and_trainer\nLoad model, tokenizer, trainer, etc. Helper function to encapsulate the full\n\n\nsetup_model_card\nSet up the Axolotl badge and add the Axolotl config to the model card if available.\n\n\nsetup_reference_model\nSet up the reference model for RL training if needed.\n\n\nsetup_signal_handler\nSet up signal handler for graceful termination.\n\n\ntrain\nTrain a model on the given dataset.\n\n\n\n\n\ntrain.create_model_card(cfg, trainer)\nCreate a model card for the trained model if needed.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nTrainer\nThe trainer object with model card creation capabilities.\nrequired\n\n\n\n\n\n\n\ntrain.determine_resume_checkpoint(cfg)\nDetermine the checkpoint to resume from based on configuration.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nstr | None\nPath to the checkpoint to resume from, or None if not resuming.\n\n\n\n\n\n\n\ntrain.execute_training(cfg, trainer, resume_from_checkpoint)\nExecute the training process with appropriate SDP kernel configurations.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nAny\nThe configured trainer object.\nrequired\n\n\nresume_from_checkpoint\nstr | None\nPath to checkpoint to resume from, if applicable.\nrequired\n\n\n\n\n\n\n\ntrain.handle_untrained_tokens_fix(\n cfg,\n model,\n tokenizer,\n train_dataset,\n safe_serialization,\n)\nApply fixes for untrained tokens if configured.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to apply fixes to.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer for token identification.\nrequired\n\n\ntrain_dataset\nDataset\nThe training dataset to use.\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization when saving.\nrequired\n\n\n\n\n\n\n\ntrain.save_initial_configs(cfg, tokenizer, model, peft_config, processor)\nSave initial configurations before training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer to save.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to save configuration for.\nrequired\n\n\npeft_config\nPeftConfig | None\nThe PEFT configuration to save if applicable.\nrequired\n\n\n\n\n\n\n\ntrain.save_trained_model(cfg, trainer, model, safe_serialization)\nSave the trained model according to configuration and training setup.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nAny\nThe trainer object.\nrequired\n\n\nmodel\nPreTrainedModel\nThe trained model to save.\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization.\nrequired\n\n\n\n\n\n\n\ntrain.setup_model_and_tokenizer(cfg)\nLoad the tokenizer, processor (for multimodal models), and model based on\nconfiguration.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple[PreTrainedModel, PreTrainedTokenizer, PeftConfig | None, ProcessorMixin | None]\nTuple containing model, tokenizer, peft_config (if LoRA / QLoRA, else None), and processor (if multimodal, else None).\n\n\n\n\n\n\n\ntrain.setup_model_and_trainer(cfg, dataset_meta)\nLoad model, tokenizer, trainer, etc. Helper function to encapsulate the full\ntrainer setup.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nThe configuration dictionary with training parameters.\nrequired\n\n\ndataset_meta\nTrainDatasetMeta\nObject with training, validation datasets and metadata.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple['HFRLTrainerBuilder' | 'HFCausalTrainerBuilder', PeftModel | PreTrainedModel, PreTrainedTokenizer, PeftConfig | None, ProcessorMixin | None]\nTuple of: - Trainer (Causal or RLHF) - Model - Tokenizer - PEFT config - Processor\n\n\n\n\n\n\n\ntrain.setup_model_card(cfg)\nSet up the Axolotl badge and add the Axolotl config to the model card if available.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\n\n\n\n\n\ntrain.setup_reference_model(cfg, tokenizer)\nSet up the reference model for RL training if needed.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer to use for the reference model.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nPreTrainedModel | None\nReference model if needed for RL training, None otherwise.\n\n\n\n\n\n\n\ntrain.setup_signal_handler(cfg, model, safe_serialization)\nSet up signal handler for graceful termination.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to save on termination\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization when saving\nrequired\n\n\n\n\n\n\n\ntrain.train(cfg, dataset_meta)\nTrain a model on the given dataset.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nThe configuration dictionary with training parameters\nrequired\n\n\ndataset_meta\nTrainDatasetMeta\nObject with training, validation datasets and metadata\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple[PeftModel | PreTrainedModel, PreTrainedTokenizer, Trainer]\nTuple of (model, tokenizer) after training"
+ "text": "train\nPrepare and train a model on a dataset. Can also infer from a model or merge lora\n\n\n\n\n\nName\nDescription\n\n\n\n\ncreate_model_card\nCreate a model card for the trained model if needed.\n\n\nexecute_training\nExecute the training process with appropriate SDP kernel configurations.\n\n\nhandle_untrained_tokens_fix\nApply fixes for untrained tokens if configured.\n\n\nsave_initial_configs\nSave initial configurations before training.\n\n\nsave_trained_model\nSave the trained model according to configuration and training setup.\n\n\nsetup_model_and_tokenizer\nLoad the tokenizer, processor (for multimodal models), and model based on\n\n\nsetup_model_and_trainer\nLoad model, tokenizer, trainer, etc. Helper function to encapsulate the full\n\n\nsetup_model_card\nSet up the Axolotl badge and add the Axolotl config to the model card if available.\n\n\nsetup_reference_model\nSet up the reference model for RL training if needed.\n\n\nsetup_signal_handler\nSet up signal handler for graceful termination.\n\n\ntrain\nTrain a model on the given dataset.\n\n\n\n\n\ntrain.create_model_card(cfg, trainer)\nCreate a model card for the trained model if needed.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nTrainer\nThe trainer object with model card creation capabilities.\nrequired\n\n\n\n\n\n\n\ntrain.execute_training(cfg, trainer, resume_from_checkpoint)\nExecute the training process with appropriate SDP kernel configurations.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nAny\nThe configured trainer object.\nrequired\n\n\nresume_from_checkpoint\nstr | None\nPath to checkpoint to resume from, if applicable.\nrequired\n\n\n\n\n\n\n\ntrain.handle_untrained_tokens_fix(\n cfg,\n model,\n tokenizer,\n train_dataset,\n safe_serialization,\n)\nApply fixes for untrained tokens if configured.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to apply fixes to.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer for token identification.\nrequired\n\n\ntrain_dataset\nDataset\nThe training dataset to use.\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization when saving.\nrequired\n\n\n\n\n\n\n\ntrain.save_initial_configs(cfg, tokenizer, model, peft_config, processor)\nSave initial configurations before training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer to save.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to save configuration for.\nrequired\n\n\npeft_config\nPeftConfig | None\nThe PEFT configuration to save if applicable.\nrequired\n\n\n\n\n\n\n\ntrain.save_trained_model(cfg, trainer, model, safe_serialization)\nSave the trained model according to configuration and training setup.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nAny\nThe trainer object.\nrequired\n\n\nmodel\nPreTrainedModel\nThe trained model to save.\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization.\nrequired\n\n\n\n\n\n\n\ntrain.setup_model_and_tokenizer(cfg)\nLoad the tokenizer, processor (for multimodal models), and model based on\nconfiguration.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple[PreTrainedModel, PreTrainedTokenizer, PeftConfig | None, ProcessorMixin | None]\nTuple containing model, tokenizer, peft_config (if LoRA / QLoRA, else None), and processor (if multimodal, else None).\n\n\n\n\n\n\n\ntrain.setup_model_and_trainer(cfg, dataset_meta)\nLoad model, tokenizer, trainer, etc. Helper function to encapsulate the full\ntrainer setup.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nThe configuration dictionary with training parameters.\nrequired\n\n\ndataset_meta\nTrainDatasetMeta\nObject with training, validation datasets and metadata.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple['HFRLTrainerBuilder' | 'HFCausalTrainerBuilder', PeftModel | PreTrainedModel, PreTrainedTokenizer, PeftConfig | None, ProcessorMixin | None]\nTuple of: - Trainer (Causal or RLHF) - Model - Tokenizer - PEFT config - Processor\n\n\n\n\n\n\n\ntrain.setup_model_card(cfg)\nSet up the Axolotl badge and add the Axolotl config to the model card if available.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\n\n\n\n\n\ntrain.setup_reference_model(cfg, tokenizer)\nSet up the reference model for RL training if needed.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer to use for the reference model.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nPreTrainedModel | None\nReference model if needed for RL training, None otherwise.\n\n\n\n\n\n\n\ntrain.setup_signal_handler(cfg, model, safe_serialization)\nSet up signal handler for graceful termination.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to save on termination\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization when saving\nrequired\n\n\n\n\n\n\n\ntrain.train(cfg, dataset_meta)\nTrain a model on the given dataset.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nThe configuration dictionary with training parameters\nrequired\n\n\ndataset_meta\nTrainDatasetMeta\nObject with training, validation datasets and metadata\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple[PeftModel | PreTrainedModel, PreTrainedTokenizer, Trainer]\nTuple of (model, tokenizer) after training"
},
{
"objectID": "docs/api/train.html#functions",
"href": "docs/api/train.html#functions",
"title": "train",
"section": "",
- "text": "Name\nDescription\n\n\n\n\ncreate_model_card\nCreate a model card for the trained model if needed.\n\n\ndetermine_resume_checkpoint\nDetermine the checkpoint to resume from based on configuration.\n\n\nexecute_training\nExecute the training process with appropriate SDP kernel configurations.\n\n\nhandle_untrained_tokens_fix\nApply fixes for untrained tokens if configured.\n\n\nsave_initial_configs\nSave initial configurations before training.\n\n\nsave_trained_model\nSave the trained model according to configuration and training setup.\n\n\nsetup_model_and_tokenizer\nLoad the tokenizer, processor (for multimodal models), and model based on\n\n\nsetup_model_and_trainer\nLoad model, tokenizer, trainer, etc. Helper function to encapsulate the full\n\n\nsetup_model_card\nSet up the Axolotl badge and add the Axolotl config to the model card if available.\n\n\nsetup_reference_model\nSet up the reference model for RL training if needed.\n\n\nsetup_signal_handler\nSet up signal handler for graceful termination.\n\n\ntrain\nTrain a model on the given dataset.\n\n\n\n\n\ntrain.create_model_card(cfg, trainer)\nCreate a model card for the trained model if needed.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nTrainer\nThe trainer object with model card creation capabilities.\nrequired\n\n\n\n\n\n\n\ntrain.determine_resume_checkpoint(cfg)\nDetermine the checkpoint to resume from based on configuration.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nstr | None\nPath to the checkpoint to resume from, or None if not resuming.\n\n\n\n\n\n\n\ntrain.execute_training(cfg, trainer, resume_from_checkpoint)\nExecute the training process with appropriate SDP kernel configurations.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nAny\nThe configured trainer object.\nrequired\n\n\nresume_from_checkpoint\nstr | None\nPath to checkpoint to resume from, if applicable.\nrequired\n\n\n\n\n\n\n\ntrain.handle_untrained_tokens_fix(\n cfg,\n model,\n tokenizer,\n train_dataset,\n safe_serialization,\n)\nApply fixes for untrained tokens if configured.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to apply fixes to.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer for token identification.\nrequired\n\n\ntrain_dataset\nDataset\nThe training dataset to use.\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization when saving.\nrequired\n\n\n\n\n\n\n\ntrain.save_initial_configs(cfg, tokenizer, model, peft_config, processor)\nSave initial configurations before training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer to save.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to save configuration for.\nrequired\n\n\npeft_config\nPeftConfig | None\nThe PEFT configuration to save if applicable.\nrequired\n\n\n\n\n\n\n\ntrain.save_trained_model(cfg, trainer, model, safe_serialization)\nSave the trained model according to configuration and training setup.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nAny\nThe trainer object.\nrequired\n\n\nmodel\nPreTrainedModel\nThe trained model to save.\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization.\nrequired\n\n\n\n\n\n\n\ntrain.setup_model_and_tokenizer(cfg)\nLoad the tokenizer, processor (for multimodal models), and model based on\nconfiguration.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple[PreTrainedModel, PreTrainedTokenizer, PeftConfig | None, ProcessorMixin | None]\nTuple containing model, tokenizer, peft_config (if LoRA / QLoRA, else None), and processor (if multimodal, else None).\n\n\n\n\n\n\n\ntrain.setup_model_and_trainer(cfg, dataset_meta)\nLoad model, tokenizer, trainer, etc. Helper function to encapsulate the full\ntrainer setup.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nThe configuration dictionary with training parameters.\nrequired\n\n\ndataset_meta\nTrainDatasetMeta\nObject with training, validation datasets and metadata.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple['HFRLTrainerBuilder' | 'HFCausalTrainerBuilder', PeftModel | PreTrainedModel, PreTrainedTokenizer, PeftConfig | None, ProcessorMixin | None]\nTuple of: - Trainer (Causal or RLHF) - Model - Tokenizer - PEFT config - Processor\n\n\n\n\n\n\n\ntrain.setup_model_card(cfg)\nSet up the Axolotl badge and add the Axolotl config to the model card if available.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\n\n\n\n\n\ntrain.setup_reference_model(cfg, tokenizer)\nSet up the reference model for RL training if needed.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer to use for the reference model.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nPreTrainedModel | None\nReference model if needed for RL training, None otherwise.\n\n\n\n\n\n\n\ntrain.setup_signal_handler(cfg, model, safe_serialization)\nSet up signal handler for graceful termination.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to save on termination\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization when saving\nrequired\n\n\n\n\n\n\n\ntrain.train(cfg, dataset_meta)\nTrain a model on the given dataset.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nThe configuration dictionary with training parameters\nrequired\n\n\ndataset_meta\nTrainDatasetMeta\nObject with training, validation datasets and metadata\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple[PeftModel | PreTrainedModel, PreTrainedTokenizer, Trainer]\nTuple of (model, tokenizer) after training"
+ "text": "Name\nDescription\n\n\n\n\ncreate_model_card\nCreate a model card for the trained model if needed.\n\n\nexecute_training\nExecute the training process with appropriate SDP kernel configurations.\n\n\nhandle_untrained_tokens_fix\nApply fixes for untrained tokens if configured.\n\n\nsave_initial_configs\nSave initial configurations before training.\n\n\nsave_trained_model\nSave the trained model according to configuration and training setup.\n\n\nsetup_model_and_tokenizer\nLoad the tokenizer, processor (for multimodal models), and model based on\n\n\nsetup_model_and_trainer\nLoad model, tokenizer, trainer, etc. Helper function to encapsulate the full\n\n\nsetup_model_card\nSet up the Axolotl badge and add the Axolotl config to the model card if available.\n\n\nsetup_reference_model\nSet up the reference model for RL training if needed.\n\n\nsetup_signal_handler\nSet up signal handler for graceful termination.\n\n\ntrain\nTrain a model on the given dataset.\n\n\n\n\n\ntrain.create_model_card(cfg, trainer)\nCreate a model card for the trained model if needed.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nTrainer\nThe trainer object with model card creation capabilities.\nrequired\n\n\n\n\n\n\n\ntrain.execute_training(cfg, trainer, resume_from_checkpoint)\nExecute the training process with appropriate SDP kernel configurations.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nAny\nThe configured trainer object.\nrequired\n\n\nresume_from_checkpoint\nstr | None\nPath to checkpoint to resume from, if applicable.\nrequired\n\n\n\n\n\n\n\ntrain.handle_untrained_tokens_fix(\n cfg,\n model,\n tokenizer,\n train_dataset,\n safe_serialization,\n)\nApply fixes for untrained tokens if configured.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to apply fixes to.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer for token identification.\nrequired\n\n\ntrain_dataset\nDataset\nThe training dataset to use.\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization when saving.\nrequired\n\n\n\n\n\n\n\ntrain.save_initial_configs(cfg, tokenizer, model, peft_config, processor)\nSave initial configurations before training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer to save.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to save configuration for.\nrequired\n\n\npeft_config\nPeftConfig | None\nThe PEFT configuration to save if applicable.\nrequired\n\n\n\n\n\n\n\ntrain.save_trained_model(cfg, trainer, model, safe_serialization)\nSave the trained model according to configuration and training setup.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntrainer\nAny\nThe trainer object.\nrequired\n\n\nmodel\nPreTrainedModel\nThe trained model to save.\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization.\nrequired\n\n\n\n\n\n\n\ntrain.setup_model_and_tokenizer(cfg)\nLoad the tokenizer, processor (for multimodal models), and model based on\nconfiguration.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple[PreTrainedModel, PreTrainedTokenizer, PeftConfig | None, ProcessorMixin | None]\nTuple containing model, tokenizer, peft_config (if LoRA / QLoRA, else None), and processor (if multimodal, else None).\n\n\n\n\n\n\n\ntrain.setup_model_and_trainer(cfg, dataset_meta)\nLoad model, tokenizer, trainer, etc. Helper function to encapsulate the full\ntrainer setup.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nThe configuration dictionary with training parameters.\nrequired\n\n\ndataset_meta\nTrainDatasetMeta\nObject with training, validation datasets and metadata.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple['HFRLTrainerBuilder' | 'HFCausalTrainerBuilder', PeftModel | PreTrainedModel, PreTrainedTokenizer, PeftConfig | None, ProcessorMixin | None]\nTuple of: - Trainer (Causal or RLHF) - Model - Tokenizer - PEFT config - Processor\n\n\n\n\n\n\n\ntrain.setup_model_card(cfg)\nSet up the Axolotl badge and add the Axolotl config to the model card if available.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\n\n\n\n\n\ntrain.setup_reference_model(cfg, tokenizer)\nSet up the reference model for RL training if needed.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\ntokenizer\nPreTrainedTokenizer\nThe tokenizer to use for the reference model.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nPreTrainedModel | None\nReference model if needed for RL training, None otherwise.\n\n\n\n\n\n\n\ntrain.setup_signal_handler(cfg, model, safe_serialization)\nSet up signal handler for graceful termination.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nDictionary mapping axolotl config keys to values.\nrequired\n\n\nmodel\nPreTrainedModel\nThe model to save on termination\nrequired\n\n\nsafe_serialization\nbool\nWhether to use safe serialization when saving\nrequired\n\n\n\n\n\n\n\ntrain.train(cfg, dataset_meta)\nTrain a model on the given dataset.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\nDictDefault\nThe configuration dictionary with training parameters\nrequired\n\n\ndataset_meta\nTrainDatasetMeta\nObject with training, validation datasets and metadata\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\ntuple[PeftModel | PreTrainedModel, PreTrainedTokenizer, Trainer]\nTuple of (model, tokenizer) after training"
},
{
"objectID": "docs/api/integrations.grokfast.optimizer.html",
diff --git a/sitemap.xml b/sitemap.xml
index f47fbc153..d92f4214c 100644
--- a/sitemap.xml
+++ b/sitemap.xml
@@ -2,790 +2,790 @@
https://docs.axolotl.ai/index.html
- 2025-08-15T14:53:04.773Z
+ 2025-08-16T01:25:11.151Zhttps://docs.axolotl.ai/src/axolotl/integrations/LICENSE.html
- 2025-08-15T14:53:04.777Z
+ 2025-08-16T01:25:11.155Zhttps://docs.axolotl.ai/docs/gradient_checkpointing.html
- 2025-08-15T14:53:04.754Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/mixed_precision.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/sequence_parallelism.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.135Zhttps://docs.axolotl.ai/docs/docker.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/torchao.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.135Zhttps://docs.axolotl.ai/docs/multi-gpu.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/dataset_preprocessing.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/debugging.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/rlhf.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.135Zhttps://docs.axolotl.ai/docs/lr_groups.html
- 2025-08-15T14:53:04.756Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/multimodal.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/ray-integration.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.135Zhttps://docs.axolotl.ai/docs/input_output.html
- 2025-08-15T14:53:04.756Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/inference.html
- 2025-08-15T14:53:04.756Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/fsdp_qlora.html
- 2025-08-15T14:53:04.754Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/multipack.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.input_output.html
- 2025-08-15T14:56:16.581Z
+ 2025-08-16T01:28:45.040Zhttps://docs.axolotl.ai/docs/api/monkeypatch.llama_patch_multipack.html
- 2025-08-15T14:56:16.848Z
+ 2025-08-16T01:28:45.302Zhttps://docs.axolotl.ai/docs/api/cli.art.html
- 2025-08-15T14:56:16.212Z
+ 2025-08-16T01:28:44.677Zhttps://docs.axolotl.ai/docs/api/cli.quantize.html
- 2025-08-15T14:56:16.288Z
+ 2025-08-16T01:28:44.752Zhttps://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_flash.html
- 2025-08-15T14:56:16.800Z
+ 2025-08-16T01:28:45.254Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.profiler.html
- 2025-08-15T14:56:17.379Z
+ 2025-08-16T01:28:45.835Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.stepwise_supervised.html
- 2025-08-15T14:56:16.585Z
+ 2025-08-16T01:28:45.044Zhttps://docs.axolotl.ai/docs/api/integrations.cut_cross_entropy.args.html
- 2025-08-15T14:56:17.263Z
+ 2025-08-16T01:28:45.718Zhttps://docs.axolotl.ai/docs/api/utils.data.sft.html
- 2025-08-15T14:56:17.012Z
+ 2025-08-16T01:28:45.465Zhttps://docs.axolotl.ai/docs/api/monkeypatch.unsloth_.html
- 2025-08-15T14:56:16.865Z
+ 2025-08-16T01:28:45.319Zhttps://docs.axolotl.ai/docs/api/kernels.geglu.html
- 2025-08-15T14:56:16.775Z
+ 2025-08-16T01:28:45.230Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.orpo.chat_template.html
- 2025-08-15T14:56:16.670Z
+ 2025-08-16T01:28:45.129Zhttps://docs.axolotl.ai/docs/api/cli.utils.sweeps.html
- 2025-08-15T14:56:16.335Z
+ 2025-08-16T01:28:44.798Zhttps://docs.axolotl.ai/docs/api/cli.delinearize_llama4.html
- 2025-08-15T14:56:16.241Z
+ 2025-08-16T01:28:44.705Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.pygmalion.html
- 2025-08-15T14:56:16.602Z
+ 2025-08-16T01:28:45.061Zhttps://docs.axolotl.ai/docs/api/evaluate.html
- 2025-08-15T14:56:15.997Z
+ 2025-08-16T01:28:44.458Zhttps://docs.axolotl.ai/docs/api/utils.data.pretraining.html
- 2025-08-15T14:56:17.005Z
+ 2025-08-16T01:28:45.458Zhttps://docs.axolotl.ai/docs/api/index.html
- 2025-08-15T14:56:15.924Z
+ 2025-08-16T01:28:44.389Zhttps://docs.axolotl.ai/docs/api/monkeypatch.stablelm_attn_hijack_flash.html
- 2025-08-15T14:56:16.854Z
+ 2025-08-16T01:28:45.308Zhttps://docs.axolotl.ai/docs/api/monkeypatch.utils.html
- 2025-08-15T14:56:16.845Z
+ 2025-08-16T01:28:45.300Zhttps://docs.axolotl.ai/docs/api/cli.checks.html
- 2025-08-15T14:56:16.218Z
+ 2025-08-16T01:28:44.683Zhttps://docs.axolotl.ai/docs/api/utils.chat_templates.html
- 2025-08-15T14:56:16.905Z
+ 2025-08-16T01:28:45.359Zhttps://docs.axolotl.ai/docs/api/core.builders.rl.html
- 2025-08-15T14:56:16.087Z
+ 2025-08-16T01:28:44.552Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.messages.chat.html
- 2025-08-15T14:56:16.606Z
+ 2025-08-16T01:28:45.065Zhttps://docs.axolotl.ai/docs/api/core.trainers.mixins.optimizer.html
- 2025-08-15T14:56:16.450Z
+ 2025-08-16T01:28:44.913Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.orcamini.html
- 2025-08-15T14:56:16.596Z
+ 2025-08-16T01:28:45.054Zhttps://docs.axolotl.ai/docs/api/core.trainers.mixins.scheduler.html
- 2025-08-15T14:56:16.460Z
+ 2025-08-16T01:28:44.923Zhttps://docs.axolotl.ai/docs/api/cli.utils.fetch.html
- 2025-08-15T14:56:16.323Z
+ 2025-08-16T01:28:44.787Zhttps://docs.axolotl.ai/docs/api/utils.schemas.datasets.html
- 2025-08-15T14:56:17.078Z
+ 2025-08-16T01:28:45.532Zhttps://docs.axolotl.ai/docs/api/cli.cloud.base.html
- 2025-08-15T14:56:16.299Z
+ 2025-08-16T01:28:44.762Zhttps://docs.axolotl.ai/docs/api/cli.utils.args.html
- 2025-08-15T14:56:16.318Z
+ 2025-08-16T01:28:44.781Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.comet_.html
- 2025-08-15T14:56:17.387Z
+ 2025-08-16T01:28:45.843Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.mlflow_.html
- 2025-08-15T14:56:17.384Z
+ 2025-08-16T01:28:45.840Zhttps://docs.axolotl.ai/docs/api/core.builders.causal.html
- 2025-08-15T14:56:16.082Z
+ 2025-08-16T01:28:44.547Zhttps://docs.axolotl.ai/docs/api/cli.train.html
- 2025-08-15T14:56:16.182Z
+ 2025-08-16T01:28:44.646Zhttps://docs.axolotl.ai/docs/api/utils.schemas.integrations.html
- 2025-08-15T14:56:17.107Z
+ 2025-08-16T01:28:45.562Zhttps://docs.axolotl.ai/docs/api/integrations.lm_eval.args.html
- 2025-08-15T14:56:17.278Z
+ 2025-08-16T01:28:45.733Zhttps://docs.axolotl.ai/docs/api/cli.evaluate.html
- 2025-08-15T14:56:16.190Z
+ 2025-08-16T01:28:44.654Zhttps://docs.axolotl.ai/docs/api/utils.trainer.html
- 2025-08-15T14:56:16.943Z
+ 2025-08-16T01:28:45.398Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.kto.llama3.html
- 2025-08-15T14:56:16.641Z
+ 2025-08-16T01:28:45.100Zhttps://docs.axolotl.ai/docs/api/convert.html
- 2025-08-15T14:56:16.021Z
+ 2025-08-16T01:28:44.482Zhttps://docs.axolotl.ai/docs/api/utils.schemas.multimodal.html
- 2025-08-15T14:56:17.095Z
+ 2025-08-16T01:28:45.550Zhttps://docs.axolotl.ai/docs/api/loaders.patch_manager.html
- 2025-08-15T14:56:16.443Z
+ 2025-08-16T01:28:44.906Zhttps://docs.axolotl.ai/docs/api/utils.schemas.training.html
- 2025-08-15T14:56:17.060Z
+ 2025-08-16T01:28:45.514Zhttps://docs.axolotl.ai/docs/api/utils.schemas.config.html
- 2025-08-15T14:56:17.047Z
+ 2025-08-16T01:28:45.500Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.kto.user_defined.html
- 2025-08-15T14:56:16.650Z
+ 2025-08-16T01:28:45.109Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.bradley_terry.llama3.html
- 2025-08-15T14:56:16.674Z
+ 2025-08-16T01:28:45.133Zhttps://docs.axolotl.ai/docs/api/cli.vllm_serve.html
- 2025-08-15T14:56:16.295Z
+ 2025-08-16T01:28:44.759Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_w_system.html
- 2025-08-15T14:56:16.549Z
+ 2025-08-16T01:28:45.007Zhttps://docs.axolotl.ai/docs/api/cli.merge_lora.html
- 2025-08-15T14:56:16.263Z
+ 2025-08-16T01:28:44.728Zhttps://docs.axolotl.ai/docs/api/utils.ctx_managers.sequence_parallel.html
- 2025-08-15T14:56:16.487Z
+ 2025-08-16T01:28:44.946Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_instruct.html
- 2025-08-15T14:56:16.537Z
+ 2025-08-16T01:28:44.995Zhttps://docs.axolotl.ai/docs/api/utils.bench.html
- 2025-08-15T14:56:16.919Z
+ 2025-08-16T01:28:45.373Zhttps://docs.axolotl.ai/docs/api/common.datasets.html
- 2025-08-15T14:56:17.299Z
+ 2025-08-16T01:28:45.754Zhttps://docs.axolotl.ai/docs/api/cli.utils.train.html
- 2025-08-15T14:56:16.347Z
+ 2025-08-16T01:28:44.810Zhttps://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_xformers.html
- 2025-08-15T14:56:16.801Z
+ 2025-08-16T01:28:45.255Zhttps://docs.axolotl.ai/docs/api/core.chat.messages.html
- 2025-08-15T14:56:16.123Z
+ 2025-08-16T01:28:44.587Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chat_template.html
- 2025-08-15T14:56:16.608Z
+ 2025-08-16T01:28:45.067Zhttps://docs.axolotl.ai/docs/api/core.trainers.trl.html
- 2025-08-15T14:56:16.372Z
+ 2025-08-16T01:28:44.835Zhttps://docs.axolotl.ai/docs/api/cli.preprocess.html
- 2025-08-15T14:56:16.283Z
+ 2025-08-16T01:28:44.747Zhttps://docs.axolotl.ai/docs/api/kernels.swiglu.html
- 2025-08-15T14:56:16.785Z
+ 2025-08-16T01:28:45.240Zhttps://docs.axolotl.ai/docs/api/kernels.quantize.html
- 2025-08-15T14:56:16.793Z
+ 2025-08-16T01:28:45.247Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.chat_template.html
- 2025-08-15T14:56:16.521Z
+ 2025-08-16T01:28:44.980Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.kto.chatml.html
- 2025-08-15T14:56:16.649Z
+ 2025-08-16T01:28:45.108Zhttps://docs.axolotl.ai/docs/api/core.trainers.grpo.trainer.html
- 2025-08-15T14:56:16.395Z
+ 2025-08-16T01:28:44.858Zhttps://docs.axolotl.ai/docs/api/monkeypatch.mistral_attn_hijack_flash.html
- 2025-08-15T14:56:16.803Z
+ 2025-08-16T01:28:45.257Zhttps://docs.axolotl.ai/docs/api/core.datasets.chat.html
- 2025-08-15T14:56:16.132Z
+ 2025-08-16T01:28:44.596Zhttps://docs.axolotl.ai/docs/api/cli.args.html
- 2025-08-15T14:56:16.209Z
+ 2025-08-16T01:28:44.673Zhttps://docs.axolotl.ai/docs/api/cli.main.html
- 2025-08-15T14:56:16.173Z
+ 2025-08-16T01:28:44.637Zhttps://docs.axolotl.ai/docs/api/core.trainers.dpo.trainer.html
- 2025-08-15T14:56:16.385Z
+ 2025-08-16T01:28:44.847Zhttps://docs.axolotl.ai/docs/api/utils.schemas.trl.html
- 2025-08-15T14:56:17.090Z
+ 2025-08-16T01:28:45.544Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.passthrough.html
- 2025-08-15T14:56:16.633Z
+ 2025-08-16T01:28:45.092Zhttps://docs.axolotl.ai/docs/api/prompt_tokenizers.html
- 2025-08-15T14:56:16.062Z
+ 2025-08-16T01:28:44.527Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_chat.html
- 2025-08-15T14:56:16.535Z
+ 2025-08-16T01:28:44.994Zhttps://docs.axolotl.ai/docs/api/logging_config.html
- 2025-08-15T14:56:16.072Z
+ 2025-08-16T01:28:44.536Zhttps://docs.axolotl.ai/docs/dataset-formats/tokenized.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/dataset-formats/index.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/dataset-formats/pretraining.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/dataset-formats/inst_tune.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/qat.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.135Zhttps://docs.axolotl.ai/examples/colab-notebooks/colab-axolotl-example.html
- 2025-08-15T14:53:04.762Z
+ 2025-08-16T01:25:11.139Zhttps://docs.axolotl.ai/FAQS.html
- 2025-08-15T14:53:04.751Z
+ 2025-08-16T01:25:11.129Zhttps://docs.axolotl.ai/docs/installation.html
- 2025-08-15T14:53:04.756Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/dataset-formats/stepwise_supervised.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/dataset-formats/template_free.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/dataset-formats/conversation.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/api/utils.dict.html
- 2025-08-15T14:56:16.996Z
+ 2025-08-16T01:28:45.449Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.completion.html
- 2025-08-15T14:56:16.575Z
+ 2025-08-16T01:28:45.034Zhttps://docs.axolotl.ai/docs/api/utils.collators.core.html
- 2025-08-15T14:56:17.302Z
+ 2025-08-16T01:28:45.757Zhttps://docs.axolotl.ai/docs/api/cli.inference.html
- 2025-08-15T14:56:16.255Z
+ 2025-08-16T01:28:44.719Zhttps://docs.axolotl.ai/docs/api/utils.freeze.html
- 2025-08-15T14:56:16.926Z
+ 2025-08-16T01:28:45.381Zhttps://docs.axolotl.ai/docs/api/core.trainers.grpo.sampler.html
- 2025-08-15T14:56:16.407Z
+ 2025-08-16T01:28:44.870Zhttps://docs.axolotl.ai/docs/api/core.trainers.mixins.rng_state_loader.html
- 2025-08-15T14:56:16.454Z
+ 2025-08-16T01:28:44.916Zhttps://docs.axolotl.ai/docs/api/cli.utils.html
- 2025-08-15T14:56:16.307Z
+ 2025-08-16T01:28:44.770Zhttps://docs.axolotl.ai/docs/api/core.chat.format.shared.html
- 2025-08-15T14:56:16.127Z
+ 2025-08-16T01:28:44.591Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.lisa.html
- 2025-08-15T14:56:17.381Z
+ 2025-08-16T01:28:45.836Zhttps://docs.axolotl.ai/docs/api/utils.collators.mm_chat.html
- 2025-08-15T14:56:17.329Z
+ 2025-08-16T01:28:45.784Zhttps://docs.axolotl.ai/docs/api/core.trainers.utils.html
- 2025-08-15T14:56:16.409Z
+ 2025-08-16T01:28:44.872Zhttps://docs.axolotl.ai/docs/api/utils.optimizers.adopt.html
- 2025-08-15T14:56:17.004Z
+ 2025-08-16T01:28:45.457Zhttps://docs.axolotl.ai/docs/api/integrations.base.html
- 2025-08-15T14:56:17.260Z
+ 2025-08-16T01:28:45.714Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.user_defined.html
- 2025-08-15T14:56:16.631Z
+ 2025-08-16T01:28:45.090Zhttps://docs.axolotl.ai/docs/api/monkeypatch.btlm_attn_hijack_flash.html
- 2025-08-15T14:56:16.847Z
+ 2025-08-16T01:28:45.301Zhttps://docs.axolotl.ai/docs/api/utils.quantization.html
- 2025-08-15T14:56:17.032Z
+ 2025-08-16T01:28:45.486Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.qat.html
- 2025-08-15T14:56:17.394Z
+ 2025-08-16T01:28:45.850Zhttps://docs.axolotl.ai/docs/api/core.builders.base.html
- 2025-08-15T14:56:16.078Z
+ 2025-08-16T01:28:44.543Zhttps://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_cpu.html
- 2025-08-15T14:56:16.871Z
+ 2025-08-16T01:28:45.325Zhttps://docs.axolotl.ai/docs/api/integrations.kd.trainer.html
- 2025-08-15T14:56:17.272Z
+ 2025-08-16T01:28:45.726Zhttps://docs.axolotl.ai/docs/api/integrations.liger.args.html
- 2025-08-15T14:56:17.275Z
+ 2025-08-16T01:28:45.729Zhttps://docs.axolotl.ai/docs/api/utils.collators.mamba.html
- 2025-08-15T14:56:17.324Z
+ 2025-08-16T01:28:45.779Zhttps://docs.axolotl.ai/docs/api/loaders.model.html
- 2025-08-15T14:56:16.419Z
+ 2025-08-16T01:28:44.881Zhttps://docs.axolotl.ai/docs/api/utils.schedulers.html
- 2025-08-15T14:56:16.971Z
+ 2025-08-16T01:28:45.424Zhttps://docs.axolotl.ai/docs/api/kernels.lora.html
- 2025-08-15T14:56:16.765Z
+ 2025-08-16T01:28:45.219Zhttps://docs.axolotl.ai/docs/api/utils.model_shard_quant.html
- 2025-08-15T14:56:16.915Z
+ 2025-08-16T01:28:45.369Zhttps://docs.axolotl.ai/docs/api/core.chat.format.llama3x.html
- 2025-08-15T14:56:16.125Z
+ 2025-08-16T01:28:44.590Zhttps://docs.axolotl.ai/docs/api/core.trainers.mamba.html
- 2025-08-15T14:56:16.378Z
+ 2025-08-16T01:28:44.841Zhttps://docs.axolotl.ai/docs/api/utils.schemas.enums.html
- 2025-08-15T14:56:17.117Z
+ 2025-08-16T01:28:45.572Zhttps://docs.axolotl.ai/docs/api/monkeypatch.mixtral.html
- 2025-08-15T14:56:16.868Z
+ 2025-08-16T01:28:45.322Zhttps://docs.axolotl.ai/docs/api/kernels.utils.html
- 2025-08-15T14:56:16.794Z
+ 2025-08-16T01:28:45.249Zhttps://docs.axolotl.ai/docs/api/core.training_args.html
- 2025-08-15T14:56:16.100Z
+ 2025-08-16T01:28:44.564Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.perplexity.html
- 2025-08-15T14:56:17.375Z
+ 2025-08-16T01:28:45.831Zhttps://docs.axolotl.ai/docs/api/cli.cloud.modal_.html
- 2025-08-15T14:56:16.305Z
+ 2025-08-16T01:28:44.768Zhttps://docs.axolotl.ai/docs/api/cli.utils.load.html
- 2025-08-15T14:56:16.329Z
+ 2025-08-16T01:28:44.792Zhttps://docs.axolotl.ai/docs/api/train.html
- 2025-08-15T14:56:15.986Z
+ 2025-08-16T01:28:44.447Zhttps://docs.axolotl.ai/docs/api/integrations.grokfast.optimizer.html
- 2025-08-15T14:56:17.264Z
+ 2025-08-16T01:28:45.719Zhttps://docs.axolotl.ai/docs/api/utils.samplers.multipack.html
- 2025-08-15T14:56:17.369Z
+ 2025-08-16T01:28:45.825Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.metharme.html
- 2025-08-15T14:56:16.592Z
+ 2025-08-16T01:28:45.051Zhttps://docs.axolotl.ai/docs/api/monkeypatch.llama_expand_mask.html
- 2025-08-15T14:56:16.809Z
+ 2025-08-16T01:28:45.263Zhttps://docs.axolotl.ai/docs/api/monkeypatch.data.batch_dataset_fetcher.html
- 2025-08-15T14:56:16.866Z
+ 2025-08-16T01:28:45.321Zhttps://docs.axolotl.ai/docs/api/utils.lora.html
- 2025-08-15T14:56:16.910Z
+ 2025-08-16T01:28:45.364Zhttps://docs.axolotl.ai/docs/api/loaders.tokenizer.html
- 2025-08-15T14:56:16.427Z
+ 2025-08-16T01:28:44.889Zhttps://docs.axolotl.ai/docs/api/core.chat.format.chatml.html
- 2025-08-15T14:56:16.124Z
+ 2025-08-16T01:28:44.588Zhttps://docs.axolotl.ai/docs/api/utils.collators.batching.html
- 2025-08-15T14:56:17.321Z
+ 2025-08-16T01:28:45.776Zhttps://docs.axolotl.ai/docs/api/cli.merge_sharded_fsdp_weights.html
- 2025-08-15T14:56:16.275Z
+ 2025-08-16T01:28:44.739Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.llama2_chat.html
- 2025-08-15T14:56:16.569Z
+ 2025-08-16T01:28:45.028Zhttps://docs.axolotl.ai/docs/api/utils.tokenization.html
- 2025-08-15T14:56:16.903Z
+ 2025-08-16T01:28:45.357Zhttps://docs.axolotl.ai/docs/api/common.architectures.html
- 2025-08-15T14:56:17.283Z
+ 2025-08-16T01:28:45.738Zhttps://docs.axolotl.ai/docs/api/core.datasets.transforms.chat_builder.html
- 2025-08-15T14:56:16.140Z
+ 2025-08-16T01:28:44.604Zhttps://docs.axolotl.ai/docs/api/core.trainers.base.html
- 2025-08-15T14:56:16.357Z
+ 2025-08-16T01:28:44.820Zhttps://docs.axolotl.ai/docs/api/monkeypatch.lora_kernels.html
- 2025-08-15T14:56:16.837Z
+ 2025-08-16T01:28:45.292Zhttps://docs.axolotl.ai/docs/api/utils.schemas.utils.html
- 2025-08-15T14:56:17.123Z
+ 2025-08-16T01:28:45.578Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.llama3.html
- 2025-08-15T14:56:16.618Z
+ 2025-08-16T01:28:45.077Zhttps://docs.axolotl.ai/docs/api/cli.config.html
- 2025-08-15T14:56:16.236Z
+ 2025-08-16T01:28:44.701Zhttps://docs.axolotl.ai/docs/api/utils.schemas.peft.html
- 2025-08-15T14:56:17.086Z
+ 2025-08-16T01:28:45.541Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.user_defined.html
- 2025-08-15T14:56:16.556Z
+ 2025-08-16T01:28:45.015Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.zephyr.html
- 2025-08-15T14:56:16.630Z
+ 2025-08-16T01:28:45.089Zhttps://docs.axolotl.ai/docs/api/monkeypatch.multipack.html
- 2025-08-15T14:56:16.804Z
+ 2025-08-16T01:28:45.258Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.base.html
- 2025-08-15T14:56:16.488Z
+ 2025-08-16T01:28:44.948Zhttps://docs.axolotl.ai/docs/api/models.mamba.modeling_mamba.html
- 2025-08-15T14:56:17.300Z
+ 2025-08-16T01:28:45.755Zhttps://docs.axolotl.ai/docs/api/monkeypatch.relora.html
- 2025-08-15T14:56:16.808Z
+ 2025-08-16T01:28:45.262Zhttps://docs.axolotl.ai/docs/api/common.const.html
- 2025-08-15T14:56:17.284Z
+ 2025-08-16T01:28:45.739Zhttps://docs.axolotl.ai/docs/api/monkeypatch.trainer_fsdp_optim.html
- 2025-08-15T14:56:16.857Z
+ 2025-08-16T01:28:45.311Zhttps://docs.axolotl.ai/docs/api/utils.distributed.html
- 2025-08-15T14:56:16.991Z
+ 2025-08-16T01:28:45.444Zhttps://docs.axolotl.ai/docs/api/loaders.constants.html
- 2025-08-15T14:56:16.445Z
+ 2025-08-16T01:28:44.907Zhttps://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_disk.html
- 2025-08-15T14:56:16.897Z
+ 2025-08-16T01:28:45.351Zhttps://docs.axolotl.ai/docs/api/datasets.html
- 2025-08-15T14:56:16.008Z
+ 2025-08-16T01:28:44.469Zhttps://docs.axolotl.ai/docs/api/monkeypatch.transformers_fa_utils.html
- 2025-08-15T14:56:16.863Z
+ 2025-08-16T01:28:45.318Zhttps://docs.axolotl.ai/docs/api/loaders.processor.html
- 2025-08-15T14:56:16.428Z
+ 2025-08-16T01:28:44.891Zhttps://docs.axolotl.ai/docs/api/integrations.spectrum.args.html
- 2025-08-15T14:56:17.281Z
+ 2025-08-16T01:28:45.736Zhttps://docs.axolotl.ai/docs/api/loaders.adapter.html
- 2025-08-15T14:56:16.434Z
+ 2025-08-16T01:28:44.896Zhttps://docs.axolotl.ai/docs/api/utils.schemas.model.html
- 2025-08-15T14:56:17.054Z
+ 2025-08-16T01:28:45.507Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chatml.html
- 2025-08-15T14:56:16.628Z
+ 2025-08-16T01:28:45.087Zhttps://docs.axolotl.ai/docs/batch_vs_grad.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.130Zhttps://docs.axolotl.ai/docs/mac.html
- 2025-08-15T14:53:04.756Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/nd_parallelism.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/dataset_loading.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/lora_optims.html
- 2025-08-15T14:53:04.756Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/unsloth.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.135Zhttps://docs.axolotl.ai/docs/config-reference.html
- 2025-08-15T14:56:30.358Z
+ 2025-08-16T01:28:59.918Zhttps://docs.axolotl.ai/docs/custom_integrations.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.130Zhttps://docs.axolotl.ai/docs/faq.html
- 2025-08-15T14:53:04.754Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/amd_hpc.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.130Zhttps://docs.axolotl.ai/docs/multi-node.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/cli.html
- 2025-08-15T14:53:04.753Z
+ 2025-08-16T01:25:11.130Zhttps://docs.axolotl.ai/docs/nccl.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/optimizers.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.134Zhttps://docs.axolotl.ai/docs/getting-started.html
- 2025-08-15T14:53:04.754Z
+ 2025-08-16T01:25:11.131Zhttps://docs.axolotl.ai/docs/quantize.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.135Zhttps://docs.axolotl.ai/docs/reward_modelling.html
- 2025-08-15T14:53:04.757Z
+ 2025-08-16T01:25:11.135Zhttps://docs.axolotl.ai/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html
- 2025-08-15T14:53:04.777Z
+ 2025-08-16T01:25:11.155Z