From 158ce4eed104af8d167845308fec25373b06bc4f Mon Sep 17 00:00:00 2001 From: Quarto GHA Workflow Runner Date: Tue, 29 Apr 2025 16:08:03 +0000 Subject: [PATCH] Built site for gh-pages --- .nojekyll | 2 +- docs/api/integrations.base.html | 6 +- search.json | 4 +- sitemap.xml | 340 ++++++++++++++++---------------- 4 files changed, 175 insertions(+), 177 deletions(-) diff --git a/.nojekyll b/.nojekyll index 0ebb97932..333d4ae6f 100644 --- a/.nojekyll +++ b/.nojekyll @@ -1 +1 @@ -165a63b7 \ No newline at end of file +d53f4ad0 \ No newline at end of file diff --git a/docs/api/integrations.base.html b/docs/api/integrations.base.html index 413b16f92..c3069280f 100644 --- a/docs/api/integrations.base.html +++ b/docs/api/integrations.base.html @@ -845,10 +845,9 @@ List[callable]: A list of callback functions to be added to the TrainingArgs.

create_lr_scheduler
-
integrations.base.PluginManager.create_lr_scheduler(cfg, trainer, optimizer)
+
integrations.base.PluginManager.create_lr_scheduler(trainer, optimizer)

Calls the create_lr_scheduler method of all registered plugins and returns the first non-None scheduler.

Parameters: -cfg (dict): The configuration for the plugins. trainer (object): The trainer object for training. optimizer (object): The optimizer for training.

Returns: @@ -856,10 +855,9 @@ object: The created learning rate scheduler, or None if none was found.

create_optimizer
-
integrations.base.PluginManager.create_optimizer(cfg, trainer)
+
integrations.base.PluginManager.create_optimizer(trainer)

Calls the create_optimizer method of all registered plugins and returns the first non-None optimizer.

Parameters: -cfg (dict): The configuration for the plugins. trainer (object): The trainer object for training.

Returns: object: The created optimizer, or None if none was found.

diff --git a/search.json b/search.json index 3cb1dca79..08576f148 100644 --- a/search.json +++ b/search.json @@ -2528,14 +2528,14 @@ "href": "docs/api/integrations.base.html", "title": "integrations.base", "section": "", - "text": "integrations.base\nBase class for all plugins.\nA plugin is a reusable, modular, and self-contained piece of code that extends the functionality of Axolotl.\nPlugins can be used to integrate third-party models, modify the training process, or add new features.\nTo create a new plugin, you need to inherit from the BasePlugin class and implement the required methods.\n\n\n\n\n\nName\nDescription\n\n\n\n\nBaseOptimizerFactory\nBase class for factories to create custom optimizers\n\n\nBasePlugin\nBase class for all plugins. Defines the interface for plugin methods.\n\n\nPluginManager\nThe PluginManager class is responsible for loading and managing plugins.\n\n\n\n\n\nintegrations.base.BaseOptimizerFactory()\nBase class for factories to create custom optimizers\n\n\n\nintegrations.base.BasePlugin(self)\nBase class for all plugins. Defines the interface for plugin methods.\nAttributes:\nNone\nMethods:\nregister(cfg): Registers the plugin with the given configuration.\npre_model_load(cfg): Performs actions before the model is loaded.\npost_model_build(cfg, model): Performs actions after the model is loaded, but before LoRA adapters are applied.\npre_lora_load(cfg, model): Performs actions before LoRA weights are loaded.\npost_lora_load(cfg, model): Performs actions after LoRA weights are loaded.\npost_model_load(cfg, model): Performs actions after the model is loaded, inclusive of any adapters.\ncreate_optimizer(cfg, trainer): Creates and returns an optimizer for training.\ncreate_lr_scheduler(cfg, trainer, optimizer): Creates and returns a learning rate scheduler.\nadd_callbacks_pre_trainer(cfg, model): Adds callbacks to the trainer before training.\nadd_callbacks_post_trainer(cfg, trainer): Adds callbacks to the trainer after training.\n\n\n\n\n\nName\nDescription\n\n\n\n\nadd_callbacks_post_trainer\nAdds callbacks to the trainer after creating the trainer.\n\n\nadd_callbacks_pre_trainer\nsetup callbacks before creating the trainer.\n\n\ncreate_lr_scheduler\nCreates and returns a learning rate scheduler.\n\n\ncreate_optimizer\nCreates and returns an optimizer for training.\n\n\nget_input_args\nReturns a pydantic model for the plugin’s input arguments.\n\n\nget_trainer_cls\nReturns a custom class for the trainer.\n\n\npost_lora_load\nPerforms actions after LoRA weights are loaded.\n\n\npost_model_build\nPerforms actions after the model is built/loaded, but before any adapters are applied.\n\n\npost_model_load\nPerforms actions after the model is loaded.\n\n\npost_train\nPerforms actions after training is complete.\n\n\npost_train_unload\nPerforms actions after training is complete and the model is unloaded.\n\n\npre_lora_load\nPerforms actions before LoRA weights are loaded.\n\n\npre_model_load\nPerforms actions before the model is loaded.\n\n\nregister\nRegisters the plugin with the given configuration.\n\n\n\n\n\nintegrations.base.BasePlugin.add_callbacks_post_trainer(cfg, trainer)\nAdds callbacks to the trainer after creating the trainer.\nThis is useful for callbacks that require access to the model or trainer.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\nReturns:\nList[callable]: A list of callback functions to be added\n\n\n\nintegrations.base.BasePlugin.add_callbacks_pre_trainer(cfg, model)\nsetup callbacks before creating the trainer.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs\n\n\n\nintegrations.base.BasePlugin.create_lr_scheduler(cfg, trainer, optimizer)\nCreates and returns a learning rate scheduler.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\noptimizer (object): The optimizer for training.\nReturns:\nobject: The created learning rate scheduler.\n\n\n\nintegrations.base.BasePlugin.create_optimizer(cfg, trainer)\nCreates and returns an optimizer for training.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\nReturns:\nobject: The created optimizer.\n\n\n\nintegrations.base.BasePlugin.get_input_args()\nReturns a pydantic model for the plugin’s input arguments.\n\n\n\nintegrations.base.BasePlugin.get_trainer_cls(cfg)\nReturns a custom class for the trainer.\nParameters:\ncfg (dict): The global axolotl configuration.\nReturns:\nclass: The class for the trainer.\n\n\n\nintegrations.base.BasePlugin.post_lora_load(cfg, model)\nPerforms actions after LoRA weights are loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_model_build(cfg, model)\nPerforms actions after the model is built/loaded, but before any adapters are applied.\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\ndict\nThe configuration for the plugin.\nrequired\n\n\n\n\n\n\n\nintegrations.base.BasePlugin.post_model_load(cfg, model)\nPerforms actions after the model is loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_train(cfg, model)\nPerforms actions after training is complete.\nParameters:\ncfg (dict): The axolotl configuration\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_train_unload(cfg)\nPerforms actions after training is complete and the model is unloaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.pre_lora_load(cfg, model)\nPerforms actions before LoRA weights are loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.pre_model_load(cfg)\nPerforms actions before the model is loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.register(cfg)\nRegisters the plugin with the given configuration.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\n\n\nintegrations.base.PluginManager()\nThe PluginManager class is responsible for loading and managing plugins.\nIt should be a singleton so it can be accessed from anywhere in the codebase.\nAttributes:\nplugins (ListBasePlugin): A list of loaded plugins.\nMethods:\nget_instance(): Static method to get the singleton instance of PluginManager.\nregister(plugin_name: str): Registers a new plugin by its name.\npre_model_load(cfg): Calls the pre_model_load method of all registered plugins.\n\n\n\n\n\nName\nDescription\n\n\n\n\nadd_callbacks_post_trainer\nCalls the add_callbacks_post_trainer method of all registered plugins.\n\n\nadd_callbacks_pre_trainer\nCalls the add_callbacks_pre_trainer method of all registered plugins.\n\n\ncreate_lr_scheduler\nCalls the create_lr_scheduler method of all registered plugins and returns the first non-None scheduler.\n\n\ncreate_optimizer\nCalls the create_optimizer method of all registered plugins and returns the first non-None optimizer.\n\n\nget_input_args\nReturns a list of Pydantic classes for all registered plugins’ input arguments.’\n\n\nget_instance\nReturns the singleton instance of PluginManager.\n\n\nget_trainer_cls\nCalls the get_trainer_cls method of all registered plugins and returns the first non-None trainer class.\n\n\npost_lora_load\nCalls the post_lora_load method of all registered plugins.\n\n\npost_model_build\nCalls the post_model_build method of all registered plugins after the model has been built/loaded,\n\n\npost_model_load\nCalls the post_model_load method of all registered plugins after the model has been loaded\n\n\npost_train\nCalls the post_train method of all registered plugins.\n\n\npost_train_unload\nCalls the post_train_unload method of all registered plugins.\n\n\npre_lora_load\nCalls the pre_lora_load method of all registered plugins.\n\n\npre_model_load\nCalls the pre_model_load method of all registered plugins.\n\n\nregister\nRegisters a new plugin by its name.\n\n\n\n\n\nintegrations.base.PluginManager.add_callbacks_post_trainer(cfg, trainer)\nCalls the add_callbacks_post_trainer method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\ntrainer (object): The trainer object for training.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs.\n\n\n\nintegrations.base.PluginManager.add_callbacks_pre_trainer(cfg, model)\nCalls the add_callbacks_pre_trainer method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs.\n\n\n\nintegrations.base.PluginManager.create_lr_scheduler(cfg, trainer, optimizer)\nCalls the create_lr_scheduler method of all registered plugins and returns the first non-None scheduler.\nParameters:\ncfg (dict): The configuration for the plugins.\ntrainer (object): The trainer object for training.\noptimizer (object): The optimizer for training.\nReturns:\nobject: The created learning rate scheduler, or None if none was found.\n\n\n\nintegrations.base.PluginManager.create_optimizer(cfg, trainer)\nCalls the create_optimizer method of all registered plugins and returns the first non-None optimizer.\nParameters:\ncfg (dict): The configuration for the plugins.\ntrainer (object): The trainer object for training.\nReturns:\nobject: The created optimizer, or None if none was found.\n\n\n\nintegrations.base.PluginManager.get_input_args()\nReturns a list of Pydantic classes for all registered plugins’ input arguments.’\nReturns:\nlist[str]: A list of Pydantic classes for all registered plugins’ input arguments.’\n\n\n\nintegrations.base.PluginManager.get_instance()\nReturns the singleton instance of PluginManager.\nIf the instance doesn’t exist, it creates a new one.\n\n\n\nintegrations.base.PluginManager.get_trainer_cls(cfg)\nCalls the get_trainer_cls method of all registered plugins and returns the first non-None trainer class.\nParameters:\ncfg (dict): The configuration for the plugins.\nReturns:\nobject: The trainer class, or None if none was found.\n\n\n\nintegrations.base.PluginManager.post_lora_load(cfg, model)\nCalls the post_lora_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_model_build(cfg, model)\nCalls the post_model_build method of all registered plugins after the model has been built/loaded,\nbut before any adapters have been applied.\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\ndict\nThe configuration for the plugins.\nrequired\n\n\nmodel\nobject\nThe loaded model.\nrequired\n\n\n\n\n\n\n\nintegrations.base.PluginManager.post_model_load(cfg, model)\nCalls the post_model_load method of all registered plugins after the model has been loaded\ninclusive of any adapters\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_train(cfg, model)\nCalls the post_train method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_train_unload(cfg)\nCalls the post_train_unload method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.pre_lora_load(cfg, model)\nCalls the pre_lora_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.pre_model_load(cfg)\nCalls the pre_model_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.register(plugin_name)\nRegisters a new plugin by its name.\nParameters:\nplugin_name (str): The name of the plugin to be registered.\nReturns:\nNone\nRaises:\nImportError: If the plugin module cannot be imported.\n\n\n\n\n\n\n\n\n\nName\nDescription\n\n\n\n\nload_plugin\nLoads a plugin based on the given plugin name.\n\n\n\n\n\nintegrations.base.load_plugin(plugin_name)\nLoads a plugin based on the given plugin name.\nThe plugin name should be in the format “module_name.class_name”.\nThis function splits the plugin name into module and class, imports the module,\nretrieves the class from the module, and creates an instance of the class.\nParameters:\nplugin_name (str): The name of the plugin to be loaded. The name should be in the format “module_name.class_name”.\nReturns:\nBasePlugin: An instance of the loaded plugin.\nRaises:\nImportError: If the plugin module cannot be imported." + "text": "integrations.base\nBase class for all plugins.\nA plugin is a reusable, modular, and self-contained piece of code that extends the functionality of Axolotl.\nPlugins can be used to integrate third-party models, modify the training process, or add new features.\nTo create a new plugin, you need to inherit from the BasePlugin class and implement the required methods.\n\n\n\n\n\nName\nDescription\n\n\n\n\nBaseOptimizerFactory\nBase class for factories to create custom optimizers\n\n\nBasePlugin\nBase class for all plugins. Defines the interface for plugin methods.\n\n\nPluginManager\nThe PluginManager class is responsible for loading and managing plugins.\n\n\n\n\n\nintegrations.base.BaseOptimizerFactory()\nBase class for factories to create custom optimizers\n\n\n\nintegrations.base.BasePlugin(self)\nBase class for all plugins. Defines the interface for plugin methods.\nAttributes:\nNone\nMethods:\nregister(cfg): Registers the plugin with the given configuration.\npre_model_load(cfg): Performs actions before the model is loaded.\npost_model_build(cfg, model): Performs actions after the model is loaded, but before LoRA adapters are applied.\npre_lora_load(cfg, model): Performs actions before LoRA weights are loaded.\npost_lora_load(cfg, model): Performs actions after LoRA weights are loaded.\npost_model_load(cfg, model): Performs actions after the model is loaded, inclusive of any adapters.\ncreate_optimizer(cfg, trainer): Creates and returns an optimizer for training.\ncreate_lr_scheduler(cfg, trainer, optimizer): Creates and returns a learning rate scheduler.\nadd_callbacks_pre_trainer(cfg, model): Adds callbacks to the trainer before training.\nadd_callbacks_post_trainer(cfg, trainer): Adds callbacks to the trainer after training.\n\n\n\n\n\nName\nDescription\n\n\n\n\nadd_callbacks_post_trainer\nAdds callbacks to the trainer after creating the trainer.\n\n\nadd_callbacks_pre_trainer\nsetup callbacks before creating the trainer.\n\n\ncreate_lr_scheduler\nCreates and returns a learning rate scheduler.\n\n\ncreate_optimizer\nCreates and returns an optimizer for training.\n\n\nget_input_args\nReturns a pydantic model for the plugin’s input arguments.\n\n\nget_trainer_cls\nReturns a custom class for the trainer.\n\n\npost_lora_load\nPerforms actions after LoRA weights are loaded.\n\n\npost_model_build\nPerforms actions after the model is built/loaded, but before any adapters are applied.\n\n\npost_model_load\nPerforms actions after the model is loaded.\n\n\npost_train\nPerforms actions after training is complete.\n\n\npost_train_unload\nPerforms actions after training is complete and the model is unloaded.\n\n\npre_lora_load\nPerforms actions before LoRA weights are loaded.\n\n\npre_model_load\nPerforms actions before the model is loaded.\n\n\nregister\nRegisters the plugin with the given configuration.\n\n\n\n\n\nintegrations.base.BasePlugin.add_callbacks_post_trainer(cfg, trainer)\nAdds callbacks to the trainer after creating the trainer.\nThis is useful for callbacks that require access to the model or trainer.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\nReturns:\nList[callable]: A list of callback functions to be added\n\n\n\nintegrations.base.BasePlugin.add_callbacks_pre_trainer(cfg, model)\nsetup callbacks before creating the trainer.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs\n\n\n\nintegrations.base.BasePlugin.create_lr_scheduler(cfg, trainer, optimizer)\nCreates and returns a learning rate scheduler.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\noptimizer (object): The optimizer for training.\nReturns:\nobject: The created learning rate scheduler.\n\n\n\nintegrations.base.BasePlugin.create_optimizer(cfg, trainer)\nCreates and returns an optimizer for training.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\nReturns:\nobject: The created optimizer.\n\n\n\nintegrations.base.BasePlugin.get_input_args()\nReturns a pydantic model for the plugin’s input arguments.\n\n\n\nintegrations.base.BasePlugin.get_trainer_cls(cfg)\nReturns a custom class for the trainer.\nParameters:\ncfg (dict): The global axolotl configuration.\nReturns:\nclass: The class for the trainer.\n\n\n\nintegrations.base.BasePlugin.post_lora_load(cfg, model)\nPerforms actions after LoRA weights are loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_model_build(cfg, model)\nPerforms actions after the model is built/loaded, but before any adapters are applied.\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\ndict\nThe configuration for the plugin.\nrequired\n\n\n\n\n\n\n\nintegrations.base.BasePlugin.post_model_load(cfg, model)\nPerforms actions after the model is loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_train(cfg, model)\nPerforms actions after training is complete.\nParameters:\ncfg (dict): The axolotl configuration\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_train_unload(cfg)\nPerforms actions after training is complete and the model is unloaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.pre_lora_load(cfg, model)\nPerforms actions before LoRA weights are loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.pre_model_load(cfg)\nPerforms actions before the model is loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.register(cfg)\nRegisters the plugin with the given configuration.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\n\n\nintegrations.base.PluginManager()\nThe PluginManager class is responsible for loading and managing plugins.\nIt should be a singleton so it can be accessed from anywhere in the codebase.\nAttributes:\nplugins (ListBasePlugin): A list of loaded plugins.\nMethods:\nget_instance(): Static method to get the singleton instance of PluginManager.\nregister(plugin_name: str): Registers a new plugin by its name.\npre_model_load(cfg): Calls the pre_model_load method of all registered plugins.\n\n\n\n\n\nName\nDescription\n\n\n\n\nadd_callbacks_post_trainer\nCalls the add_callbacks_post_trainer method of all registered plugins.\n\n\nadd_callbacks_pre_trainer\nCalls the add_callbacks_pre_trainer method of all registered plugins.\n\n\ncreate_lr_scheduler\nCalls the create_lr_scheduler method of all registered plugins and returns the first non-None scheduler.\n\n\ncreate_optimizer\nCalls the create_optimizer method of all registered plugins and returns the first non-None optimizer.\n\n\nget_input_args\nReturns a list of Pydantic classes for all registered plugins’ input arguments.’\n\n\nget_instance\nReturns the singleton instance of PluginManager.\n\n\nget_trainer_cls\nCalls the get_trainer_cls method of all registered plugins and returns the first non-None trainer class.\n\n\npost_lora_load\nCalls the post_lora_load method of all registered plugins.\n\n\npost_model_build\nCalls the post_model_build method of all registered plugins after the model has been built/loaded,\n\n\npost_model_load\nCalls the post_model_load method of all registered plugins after the model has been loaded\n\n\npost_train\nCalls the post_train method of all registered plugins.\n\n\npost_train_unload\nCalls the post_train_unload method of all registered plugins.\n\n\npre_lora_load\nCalls the pre_lora_load method of all registered plugins.\n\n\npre_model_load\nCalls the pre_model_load method of all registered plugins.\n\n\nregister\nRegisters a new plugin by its name.\n\n\n\n\n\nintegrations.base.PluginManager.add_callbacks_post_trainer(cfg, trainer)\nCalls the add_callbacks_post_trainer method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\ntrainer (object): The trainer object for training.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs.\n\n\n\nintegrations.base.PluginManager.add_callbacks_pre_trainer(cfg, model)\nCalls the add_callbacks_pre_trainer method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs.\n\n\n\nintegrations.base.PluginManager.create_lr_scheduler(trainer, optimizer)\nCalls the create_lr_scheduler method of all registered plugins and returns the first non-None scheduler.\nParameters:\ntrainer (object): The trainer object for training.\noptimizer (object): The optimizer for training.\nReturns:\nobject: The created learning rate scheduler, or None if none was found.\n\n\n\nintegrations.base.PluginManager.create_optimizer(trainer)\nCalls the create_optimizer method of all registered plugins and returns the first non-None optimizer.\nParameters:\ntrainer (object): The trainer object for training.\nReturns:\nobject: The created optimizer, or None if none was found.\n\n\n\nintegrations.base.PluginManager.get_input_args()\nReturns a list of Pydantic classes for all registered plugins’ input arguments.’\nReturns:\nlist[str]: A list of Pydantic classes for all registered plugins’ input arguments.’\n\n\n\nintegrations.base.PluginManager.get_instance()\nReturns the singleton instance of PluginManager.\nIf the instance doesn’t exist, it creates a new one.\n\n\n\nintegrations.base.PluginManager.get_trainer_cls(cfg)\nCalls the get_trainer_cls method of all registered plugins and returns the first non-None trainer class.\nParameters:\ncfg (dict): The configuration for the plugins.\nReturns:\nobject: The trainer class, or None if none was found.\n\n\n\nintegrations.base.PluginManager.post_lora_load(cfg, model)\nCalls the post_lora_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_model_build(cfg, model)\nCalls the post_model_build method of all registered plugins after the model has been built/loaded,\nbut before any adapters have been applied.\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\ndict\nThe configuration for the plugins.\nrequired\n\n\nmodel\nobject\nThe loaded model.\nrequired\n\n\n\n\n\n\n\nintegrations.base.PluginManager.post_model_load(cfg, model)\nCalls the post_model_load method of all registered plugins after the model has been loaded\ninclusive of any adapters\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_train(cfg, model)\nCalls the post_train method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_train_unload(cfg)\nCalls the post_train_unload method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.pre_lora_load(cfg, model)\nCalls the pre_lora_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.pre_model_load(cfg)\nCalls the pre_model_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.register(plugin_name)\nRegisters a new plugin by its name.\nParameters:\nplugin_name (str): The name of the plugin to be registered.\nReturns:\nNone\nRaises:\nImportError: If the plugin module cannot be imported.\n\n\n\n\n\n\n\n\n\nName\nDescription\n\n\n\n\nload_plugin\nLoads a plugin based on the given plugin name.\n\n\n\n\n\nintegrations.base.load_plugin(plugin_name)\nLoads a plugin based on the given plugin name.\nThe plugin name should be in the format “module_name.class_name”.\nThis function splits the plugin name into module and class, imports the module,\nretrieves the class from the module, and creates an instance of the class.\nParameters:\nplugin_name (str): The name of the plugin to be loaded. The name should be in the format “module_name.class_name”.\nReturns:\nBasePlugin: An instance of the loaded plugin.\nRaises:\nImportError: If the plugin module cannot be imported." }, { "objectID": "docs/api/integrations.base.html#classes", "href": "docs/api/integrations.base.html#classes", "title": "integrations.base", "section": "", - "text": "Name\nDescription\n\n\n\n\nBaseOptimizerFactory\nBase class for factories to create custom optimizers\n\n\nBasePlugin\nBase class for all plugins. Defines the interface for plugin methods.\n\n\nPluginManager\nThe PluginManager class is responsible for loading and managing plugins.\n\n\n\n\n\nintegrations.base.BaseOptimizerFactory()\nBase class for factories to create custom optimizers\n\n\n\nintegrations.base.BasePlugin(self)\nBase class for all plugins. Defines the interface for plugin methods.\nAttributes:\nNone\nMethods:\nregister(cfg): Registers the plugin with the given configuration.\npre_model_load(cfg): Performs actions before the model is loaded.\npost_model_build(cfg, model): Performs actions after the model is loaded, but before LoRA adapters are applied.\npre_lora_load(cfg, model): Performs actions before LoRA weights are loaded.\npost_lora_load(cfg, model): Performs actions after LoRA weights are loaded.\npost_model_load(cfg, model): Performs actions after the model is loaded, inclusive of any adapters.\ncreate_optimizer(cfg, trainer): Creates and returns an optimizer for training.\ncreate_lr_scheduler(cfg, trainer, optimizer): Creates and returns a learning rate scheduler.\nadd_callbacks_pre_trainer(cfg, model): Adds callbacks to the trainer before training.\nadd_callbacks_post_trainer(cfg, trainer): Adds callbacks to the trainer after training.\n\n\n\n\n\nName\nDescription\n\n\n\n\nadd_callbacks_post_trainer\nAdds callbacks to the trainer after creating the trainer.\n\n\nadd_callbacks_pre_trainer\nsetup callbacks before creating the trainer.\n\n\ncreate_lr_scheduler\nCreates and returns a learning rate scheduler.\n\n\ncreate_optimizer\nCreates and returns an optimizer for training.\n\n\nget_input_args\nReturns a pydantic model for the plugin’s input arguments.\n\n\nget_trainer_cls\nReturns a custom class for the trainer.\n\n\npost_lora_load\nPerforms actions after LoRA weights are loaded.\n\n\npost_model_build\nPerforms actions after the model is built/loaded, but before any adapters are applied.\n\n\npost_model_load\nPerforms actions after the model is loaded.\n\n\npost_train\nPerforms actions after training is complete.\n\n\npost_train_unload\nPerforms actions after training is complete and the model is unloaded.\n\n\npre_lora_load\nPerforms actions before LoRA weights are loaded.\n\n\npre_model_load\nPerforms actions before the model is loaded.\n\n\nregister\nRegisters the plugin with the given configuration.\n\n\n\n\n\nintegrations.base.BasePlugin.add_callbacks_post_trainer(cfg, trainer)\nAdds callbacks to the trainer after creating the trainer.\nThis is useful for callbacks that require access to the model or trainer.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\nReturns:\nList[callable]: A list of callback functions to be added\n\n\n\nintegrations.base.BasePlugin.add_callbacks_pre_trainer(cfg, model)\nsetup callbacks before creating the trainer.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs\n\n\n\nintegrations.base.BasePlugin.create_lr_scheduler(cfg, trainer, optimizer)\nCreates and returns a learning rate scheduler.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\noptimizer (object): The optimizer for training.\nReturns:\nobject: The created learning rate scheduler.\n\n\n\nintegrations.base.BasePlugin.create_optimizer(cfg, trainer)\nCreates and returns an optimizer for training.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\nReturns:\nobject: The created optimizer.\n\n\n\nintegrations.base.BasePlugin.get_input_args()\nReturns a pydantic model for the plugin’s input arguments.\n\n\n\nintegrations.base.BasePlugin.get_trainer_cls(cfg)\nReturns a custom class for the trainer.\nParameters:\ncfg (dict): The global axolotl configuration.\nReturns:\nclass: The class for the trainer.\n\n\n\nintegrations.base.BasePlugin.post_lora_load(cfg, model)\nPerforms actions after LoRA weights are loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_model_build(cfg, model)\nPerforms actions after the model is built/loaded, but before any adapters are applied.\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\ndict\nThe configuration for the plugin.\nrequired\n\n\n\n\n\n\n\nintegrations.base.BasePlugin.post_model_load(cfg, model)\nPerforms actions after the model is loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_train(cfg, model)\nPerforms actions after training is complete.\nParameters:\ncfg (dict): The axolotl configuration\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_train_unload(cfg)\nPerforms actions after training is complete and the model is unloaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.pre_lora_load(cfg, model)\nPerforms actions before LoRA weights are loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.pre_model_load(cfg)\nPerforms actions before the model is loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.register(cfg)\nRegisters the plugin with the given configuration.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\n\n\nintegrations.base.PluginManager()\nThe PluginManager class is responsible for loading and managing plugins.\nIt should be a singleton so it can be accessed from anywhere in the codebase.\nAttributes:\nplugins (ListBasePlugin): A list of loaded plugins.\nMethods:\nget_instance(): Static method to get the singleton instance of PluginManager.\nregister(plugin_name: str): Registers a new plugin by its name.\npre_model_load(cfg): Calls the pre_model_load method of all registered plugins.\n\n\n\n\n\nName\nDescription\n\n\n\n\nadd_callbacks_post_trainer\nCalls the add_callbacks_post_trainer method of all registered plugins.\n\n\nadd_callbacks_pre_trainer\nCalls the add_callbacks_pre_trainer method of all registered plugins.\n\n\ncreate_lr_scheduler\nCalls the create_lr_scheduler method of all registered plugins and returns the first non-None scheduler.\n\n\ncreate_optimizer\nCalls the create_optimizer method of all registered plugins and returns the first non-None optimizer.\n\n\nget_input_args\nReturns a list of Pydantic classes for all registered plugins’ input arguments.’\n\n\nget_instance\nReturns the singleton instance of PluginManager.\n\n\nget_trainer_cls\nCalls the get_trainer_cls method of all registered plugins and returns the first non-None trainer class.\n\n\npost_lora_load\nCalls the post_lora_load method of all registered plugins.\n\n\npost_model_build\nCalls the post_model_build method of all registered plugins after the model has been built/loaded,\n\n\npost_model_load\nCalls the post_model_load method of all registered plugins after the model has been loaded\n\n\npost_train\nCalls the post_train method of all registered plugins.\n\n\npost_train_unload\nCalls the post_train_unload method of all registered plugins.\n\n\npre_lora_load\nCalls the pre_lora_load method of all registered plugins.\n\n\npre_model_load\nCalls the pre_model_load method of all registered plugins.\n\n\nregister\nRegisters a new plugin by its name.\n\n\n\n\n\nintegrations.base.PluginManager.add_callbacks_post_trainer(cfg, trainer)\nCalls the add_callbacks_post_trainer method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\ntrainer (object): The trainer object for training.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs.\n\n\n\nintegrations.base.PluginManager.add_callbacks_pre_trainer(cfg, model)\nCalls the add_callbacks_pre_trainer method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs.\n\n\n\nintegrations.base.PluginManager.create_lr_scheduler(cfg, trainer, optimizer)\nCalls the create_lr_scheduler method of all registered plugins and returns the first non-None scheduler.\nParameters:\ncfg (dict): The configuration for the plugins.\ntrainer (object): The trainer object for training.\noptimizer (object): The optimizer for training.\nReturns:\nobject: The created learning rate scheduler, or None if none was found.\n\n\n\nintegrations.base.PluginManager.create_optimizer(cfg, trainer)\nCalls the create_optimizer method of all registered plugins and returns the first non-None optimizer.\nParameters:\ncfg (dict): The configuration for the plugins.\ntrainer (object): The trainer object for training.\nReturns:\nobject: The created optimizer, or None if none was found.\n\n\n\nintegrations.base.PluginManager.get_input_args()\nReturns a list of Pydantic classes for all registered plugins’ input arguments.’\nReturns:\nlist[str]: A list of Pydantic classes for all registered plugins’ input arguments.’\n\n\n\nintegrations.base.PluginManager.get_instance()\nReturns the singleton instance of PluginManager.\nIf the instance doesn’t exist, it creates a new one.\n\n\n\nintegrations.base.PluginManager.get_trainer_cls(cfg)\nCalls the get_trainer_cls method of all registered plugins and returns the first non-None trainer class.\nParameters:\ncfg (dict): The configuration for the plugins.\nReturns:\nobject: The trainer class, or None if none was found.\n\n\n\nintegrations.base.PluginManager.post_lora_load(cfg, model)\nCalls the post_lora_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_model_build(cfg, model)\nCalls the post_model_build method of all registered plugins after the model has been built/loaded,\nbut before any adapters have been applied.\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\ndict\nThe configuration for the plugins.\nrequired\n\n\nmodel\nobject\nThe loaded model.\nrequired\n\n\n\n\n\n\n\nintegrations.base.PluginManager.post_model_load(cfg, model)\nCalls the post_model_load method of all registered plugins after the model has been loaded\ninclusive of any adapters\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_train(cfg, model)\nCalls the post_train method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_train_unload(cfg)\nCalls the post_train_unload method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.pre_lora_load(cfg, model)\nCalls the pre_lora_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.pre_model_load(cfg)\nCalls the pre_model_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.register(plugin_name)\nRegisters a new plugin by its name.\nParameters:\nplugin_name (str): The name of the plugin to be registered.\nReturns:\nNone\nRaises:\nImportError: If the plugin module cannot be imported." + "text": "Name\nDescription\n\n\n\n\nBaseOptimizerFactory\nBase class for factories to create custom optimizers\n\n\nBasePlugin\nBase class for all plugins. Defines the interface for plugin methods.\n\n\nPluginManager\nThe PluginManager class is responsible for loading and managing plugins.\n\n\n\n\n\nintegrations.base.BaseOptimizerFactory()\nBase class for factories to create custom optimizers\n\n\n\nintegrations.base.BasePlugin(self)\nBase class for all plugins. Defines the interface for plugin methods.\nAttributes:\nNone\nMethods:\nregister(cfg): Registers the plugin with the given configuration.\npre_model_load(cfg): Performs actions before the model is loaded.\npost_model_build(cfg, model): Performs actions after the model is loaded, but before LoRA adapters are applied.\npre_lora_load(cfg, model): Performs actions before LoRA weights are loaded.\npost_lora_load(cfg, model): Performs actions after LoRA weights are loaded.\npost_model_load(cfg, model): Performs actions after the model is loaded, inclusive of any adapters.\ncreate_optimizer(cfg, trainer): Creates and returns an optimizer for training.\ncreate_lr_scheduler(cfg, trainer, optimizer): Creates and returns a learning rate scheduler.\nadd_callbacks_pre_trainer(cfg, model): Adds callbacks to the trainer before training.\nadd_callbacks_post_trainer(cfg, trainer): Adds callbacks to the trainer after training.\n\n\n\n\n\nName\nDescription\n\n\n\n\nadd_callbacks_post_trainer\nAdds callbacks to the trainer after creating the trainer.\n\n\nadd_callbacks_pre_trainer\nsetup callbacks before creating the trainer.\n\n\ncreate_lr_scheduler\nCreates and returns a learning rate scheduler.\n\n\ncreate_optimizer\nCreates and returns an optimizer for training.\n\n\nget_input_args\nReturns a pydantic model for the plugin’s input arguments.\n\n\nget_trainer_cls\nReturns a custom class for the trainer.\n\n\npost_lora_load\nPerforms actions after LoRA weights are loaded.\n\n\npost_model_build\nPerforms actions after the model is built/loaded, but before any adapters are applied.\n\n\npost_model_load\nPerforms actions after the model is loaded.\n\n\npost_train\nPerforms actions after training is complete.\n\n\npost_train_unload\nPerforms actions after training is complete and the model is unloaded.\n\n\npre_lora_load\nPerforms actions before LoRA weights are loaded.\n\n\npre_model_load\nPerforms actions before the model is loaded.\n\n\nregister\nRegisters the plugin with the given configuration.\n\n\n\n\n\nintegrations.base.BasePlugin.add_callbacks_post_trainer(cfg, trainer)\nAdds callbacks to the trainer after creating the trainer.\nThis is useful for callbacks that require access to the model or trainer.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\nReturns:\nList[callable]: A list of callback functions to be added\n\n\n\nintegrations.base.BasePlugin.add_callbacks_pre_trainer(cfg, model)\nsetup callbacks before creating the trainer.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs\n\n\n\nintegrations.base.BasePlugin.create_lr_scheduler(cfg, trainer, optimizer)\nCreates and returns a learning rate scheduler.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\noptimizer (object): The optimizer for training.\nReturns:\nobject: The created learning rate scheduler.\n\n\n\nintegrations.base.BasePlugin.create_optimizer(cfg, trainer)\nCreates and returns an optimizer for training.\nParameters:\ncfg (dict): The configuration for the plugin.\ntrainer (object): The trainer object for training.\nReturns:\nobject: The created optimizer.\n\n\n\nintegrations.base.BasePlugin.get_input_args()\nReturns a pydantic model for the plugin’s input arguments.\n\n\n\nintegrations.base.BasePlugin.get_trainer_cls(cfg)\nReturns a custom class for the trainer.\nParameters:\ncfg (dict): The global axolotl configuration.\nReturns:\nclass: The class for the trainer.\n\n\n\nintegrations.base.BasePlugin.post_lora_load(cfg, model)\nPerforms actions after LoRA weights are loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_model_build(cfg, model)\nPerforms actions after the model is built/loaded, but before any adapters are applied.\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\ndict\nThe configuration for the plugin.\nrequired\n\n\n\n\n\n\n\nintegrations.base.BasePlugin.post_model_load(cfg, model)\nPerforms actions after the model is loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_train(cfg, model)\nPerforms actions after training is complete.\nParameters:\ncfg (dict): The axolotl configuration\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.post_train_unload(cfg)\nPerforms actions after training is complete and the model is unloaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.pre_lora_load(cfg, model)\nPerforms actions before LoRA weights are loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.pre_model_load(cfg)\nPerforms actions before the model is loaded.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\nintegrations.base.BasePlugin.register(cfg)\nRegisters the plugin with the given configuration.\nParameters:\ncfg (dict): The configuration for the plugin.\nReturns:\nNone\n\n\n\n\n\nintegrations.base.PluginManager()\nThe PluginManager class is responsible for loading and managing plugins.\nIt should be a singleton so it can be accessed from anywhere in the codebase.\nAttributes:\nplugins (ListBasePlugin): A list of loaded plugins.\nMethods:\nget_instance(): Static method to get the singleton instance of PluginManager.\nregister(plugin_name: str): Registers a new plugin by its name.\npre_model_load(cfg): Calls the pre_model_load method of all registered plugins.\n\n\n\n\n\nName\nDescription\n\n\n\n\nadd_callbacks_post_trainer\nCalls the add_callbacks_post_trainer method of all registered plugins.\n\n\nadd_callbacks_pre_trainer\nCalls the add_callbacks_pre_trainer method of all registered plugins.\n\n\ncreate_lr_scheduler\nCalls the create_lr_scheduler method of all registered plugins and returns the first non-None scheduler.\n\n\ncreate_optimizer\nCalls the create_optimizer method of all registered plugins and returns the first non-None optimizer.\n\n\nget_input_args\nReturns a list of Pydantic classes for all registered plugins’ input arguments.’\n\n\nget_instance\nReturns the singleton instance of PluginManager.\n\n\nget_trainer_cls\nCalls the get_trainer_cls method of all registered plugins and returns the first non-None trainer class.\n\n\npost_lora_load\nCalls the post_lora_load method of all registered plugins.\n\n\npost_model_build\nCalls the post_model_build method of all registered plugins after the model has been built/loaded,\n\n\npost_model_load\nCalls the post_model_load method of all registered plugins after the model has been loaded\n\n\npost_train\nCalls the post_train method of all registered plugins.\n\n\npost_train_unload\nCalls the post_train_unload method of all registered plugins.\n\n\npre_lora_load\nCalls the pre_lora_load method of all registered plugins.\n\n\npre_model_load\nCalls the pre_model_load method of all registered plugins.\n\n\nregister\nRegisters a new plugin by its name.\n\n\n\n\n\nintegrations.base.PluginManager.add_callbacks_post_trainer(cfg, trainer)\nCalls the add_callbacks_post_trainer method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\ntrainer (object): The trainer object for training.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs.\n\n\n\nintegrations.base.PluginManager.add_callbacks_pre_trainer(cfg, model)\nCalls the add_callbacks_pre_trainer method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nList[callable]: A list of callback functions to be added to the TrainingArgs.\n\n\n\nintegrations.base.PluginManager.create_lr_scheduler(trainer, optimizer)\nCalls the create_lr_scheduler method of all registered plugins and returns the first non-None scheduler.\nParameters:\ntrainer (object): The trainer object for training.\noptimizer (object): The optimizer for training.\nReturns:\nobject: The created learning rate scheduler, or None if none was found.\n\n\n\nintegrations.base.PluginManager.create_optimizer(trainer)\nCalls the create_optimizer method of all registered plugins and returns the first non-None optimizer.\nParameters:\ntrainer (object): The trainer object for training.\nReturns:\nobject: The created optimizer, or None if none was found.\n\n\n\nintegrations.base.PluginManager.get_input_args()\nReturns a list of Pydantic classes for all registered plugins’ input arguments.’\nReturns:\nlist[str]: A list of Pydantic classes for all registered plugins’ input arguments.’\n\n\n\nintegrations.base.PluginManager.get_instance()\nReturns the singleton instance of PluginManager.\nIf the instance doesn’t exist, it creates a new one.\n\n\n\nintegrations.base.PluginManager.get_trainer_cls(cfg)\nCalls the get_trainer_cls method of all registered plugins and returns the first non-None trainer class.\nParameters:\ncfg (dict): The configuration for the plugins.\nReturns:\nobject: The trainer class, or None if none was found.\n\n\n\nintegrations.base.PluginManager.post_lora_load(cfg, model)\nCalls the post_lora_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_model_build(cfg, model)\nCalls the post_model_build method of all registered plugins after the model has been built/loaded,\nbut before any adapters have been applied.\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\ncfg\ndict\nThe configuration for the plugins.\nrequired\n\n\nmodel\nobject\nThe loaded model.\nrequired\n\n\n\n\n\n\n\nintegrations.base.PluginManager.post_model_load(cfg, model)\nCalls the post_model_load method of all registered plugins after the model has been loaded\ninclusive of any adapters\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_train(cfg, model)\nCalls the post_train method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.post_train_unload(cfg)\nCalls the post_train_unload method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.pre_lora_load(cfg, model)\nCalls the pre_lora_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nmodel (object): The loaded model.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.pre_model_load(cfg)\nCalls the pre_model_load method of all registered plugins.\nParameters:\ncfg (dict): The configuration for the plugins.\nReturns:\nNone\n\n\n\nintegrations.base.PluginManager.register(plugin_name)\nRegisters a new plugin by its name.\nParameters:\nplugin_name (str): The name of the plugin to be registered.\nReturns:\nNone\nRaises:\nImportError: If the plugin module cannot be imported." }, { "objectID": "docs/api/integrations.base.html#functions", diff --git a/sitemap.xml b/sitemap.xml index 45057e0e5..bdabe5d96 100644 --- a/sitemap.xml +++ b/sitemap.xml @@ -2,682 +2,682 @@ https://docs.axolotl.ai/examples/colab-notebooks/colab-axolotl-example.html - 2025-04-29T14:01:58.464Z + 2025-04-29T16:05:56.146Z https://docs.axolotl.ai/index.html - 2025-04-29T14:01:58.476Z + 2025-04-29T16:05:56.158Z https://docs.axolotl.ai/docs/rlhf.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/unsloth.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/dataset_preprocessing.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/input_output.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/dataset_loading.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/api/utils.collators.mamba.html - 2025-04-29T14:02:29.185Z + 2025-04-29T16:06:33.720Z https://docs.axolotl.ai/docs/api/utils.optimizers.adopt.html - 2025-04-29T14:02:28.961Z + 2025-04-29T16:06:33.498Z https://docs.axolotl.ai/docs/api/prompt_strategies.user_defined.html - 2025-04-29T14:02:28.472Z + 2025-04-29T16:06:33.013Z https://docs.axolotl.ai/docs/api/utils.chat_templates.html - 2025-04-29T14:02:28.864Z + 2025-04-29T16:06:33.400Z https://docs.axolotl.ai/docs/api/cli.merge_lora.html - 2025-04-29T14:02:28.301Z + 2025-04-29T16:06:32.845Z https://docs.axolotl.ai/docs/api/monkeypatch.multipack.html - 2025-04-29T14:02:28.745Z + 2025-04-29T16:06:33.284Z https://docs.axolotl.ai/docs/api/core.chat.format.shared.html - 2025-04-29T14:02:28.177Z + 2025-04-29T16:06:32.721Z https://docs.axolotl.ai/docs/api/utils.schemas.integrations.html - 2025-04-29T14:02:29.034Z + 2025-04-29T16:06:33.571Z https://docs.axolotl.ai/docs/api/utils.freeze.html - 2025-04-29T14:02:28.889Z + 2025-04-29T16:06:33.425Z https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_w_system.html - 2025-04-29T14:02:28.464Z + 2025-04-29T16:06:33.005Z https://docs.axolotl.ai/docs/api/monkeypatch.attention.mllama.html - 2025-04-29T14:02:28.815Z + 2025-04-29T16:06:33.352Z https://docs.axolotl.ai/docs/api/utils.schemas.model.html - 2025-04-29T14:02:28.982Z + 2025-04-29T16:06:33.519Z https://docs.axolotl.ai/docs/api/core.datasets.transforms.chat_builder.html - 2025-04-29T14:02:28.190Z + 2025-04-29T16:06:32.734Z https://docs.axolotl.ai/docs/api/monkeypatch.btlm_attn_hijack_flash.html - 2025-04-29T14:02:28.788Z + 2025-04-29T16:06:33.327Z https://docs.axolotl.ai/docs/api/models.mamba.modeling_mamba.html - 2025-04-29T14:02:29.161Z + 2025-04-29T16:06:33.696Z https://docs.axolotl.ai/docs/api/core.datasets.chat.html - 2025-04-29T14:02:28.182Z + 2025-04-29T16:06:32.726Z https://docs.axolotl.ai/docs/api/utils.model_shard_quant.html - 2025-04-29T14:02:28.878Z + 2025-04-29T16:06:33.414Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_patch_multipack.html - 2025-04-29T14:02:28.789Z + 2025-04-29T16:06:33.329Z https://docs.axolotl.ai/docs/api/utils.lora_embeddings.html - 2025-04-29T14:02:28.872Z + 2025-04-29T16:06:33.409Z https://docs.axolotl.ai/docs/api/utils.schemas.peft.html - 2025-04-29T14:02:29.013Z + 2025-04-29T16:06:33.550Z https://docs.axolotl.ai/docs/api/core.chat.format.llama3x.html - 2025-04-29T14:02:28.175Z + 2025-04-29T16:06:32.720Z https://docs.axolotl.ai/docs/api/monkeypatch.mistral_attn_hijack_flash.html - 2025-04-29T14:02:28.743Z + 2025-04-29T16:06:33.283Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chat_template.html - 2025-04-29T14:02:28.524Z + 2025-04-29T16:06:33.065Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.user_defined.html - 2025-04-29T14:02:28.548Z + 2025-04-29T16:06:33.089Z https://docs.axolotl.ai/docs/api/utils.distributed.html - 2025-04-29T14:02:28.950Z + 2025-04-29T16:06:33.487Z https://docs.axolotl.ai/docs/api/prompt_strategies.pygmalion.html - 2025-04-29T14:02:28.519Z + 2025-04-29T16:06:33.060Z https://docs.axolotl.ai/docs/api/utils.dict.html - 2025-04-29T14:02:28.954Z + 2025-04-29T16:06:33.491Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_flash.html - 2025-04-29T14:02:28.728Z + 2025-04-29T16:06:33.267Z https://docs.axolotl.ai/docs/api/prompt_strategies.base.html - 2025-04-29T14:02:28.418Z + 2025-04-29T16:06:32.960Z https://docs.axolotl.ai/docs/api/monkeypatch.data.batch_dataset_fetcher.html - 2025-04-29T14:02:28.817Z + 2025-04-29T16:06:33.354Z https://docs.axolotl.ai/docs/api/kernels.swiglu.html - 2025-04-29T14:02:28.693Z + 2025-04-29T16:06:33.233Z https://docs.axolotl.ai/docs/api/core.chat.messages.html - 2025-04-29T14:02:28.172Z + 2025-04-29T16:06:32.717Z https://docs.axolotl.ai/docs/api/index.html - 2025-04-29T14:02:27.907Z + 2025-04-29T16:06:32.453Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.zephyr.html - 2025-04-29T14:02:28.546Z + 2025-04-29T16:06:33.087Z https://docs.axolotl.ai/docs/api/convert.html - 2025-04-29T14:02:28.000Z + 2025-04-29T16:06:32.545Z https://docs.axolotl.ai/docs/api/utils.schemas.multimodal.html - 2025-04-29T14:02:29.022Z + 2025-04-29T16:06:33.559Z https://docs.axolotl.ai/docs/api/core.trainers.base.html - 2025-04-29T14:02:28.387Z + 2025-04-29T16:06:32.929Z https://docs.axolotl.ai/docs/api/evaluate.html - 2025-04-29T14:02:27.979Z + 2025-04-29T16:06:32.524Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_xformers.html - 2025-04-29T14:02:28.729Z + 2025-04-29T16:06:33.268Z https://docs.axolotl.ai/docs/api/kernels.quantize.html - 2025-04-29T14:02:28.701Z + 2025-04-29T16:06:33.240Z https://docs.axolotl.ai/docs/api/utils.callbacks.mlflow_.html - 2025-04-29T14:02:29.213Z + 2025-04-29T16:06:33.748Z https://docs.axolotl.ai/docs/api/utils.callbacks.profiler.html - 2025-04-29T14:02:29.208Z + 2025-04-29T16:06:33.743Z https://docs.axolotl.ai/docs/api/core.trainers.dpo.trainer.html - 2025-04-29T14:02:28.413Z + 2025-04-29T16:06:32.956Z https://docs.axolotl.ai/docs/api/cli.vllm_serve.html - 2025-04-29T14:02:28.363Z + 2025-04-29T16:06:32.906Z https://docs.axolotl.ai/docs/api/train.html - 2025-04-29T14:02:27.969Z + 2025-04-29T16:06:32.513Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chatml.html - 2025-04-29T14:02:28.545Z + 2025-04-29T16:06:33.086Z https://docs.axolotl.ai/docs/api/utils.schemas.trl.html - 2025-04-29T14:02:29.016Z + 2025-04-29T16:06:33.553Z https://docs.axolotl.ai/docs/api/kernels.geglu.html - 2025-04-29T14:02:28.683Z + 2025-04-29T16:06:33.223Z https://docs.axolotl.ai/docs/api/utils.bench.html - 2025-04-29T14:02:28.881Z + 2025-04-29T16:06:33.418Z https://docs.axolotl.ai/docs/api/monkeypatch.transformers_fa_utils.html - 2025-04-29T14:02:28.807Z + 2025-04-29T16:06:33.344Z https://docs.axolotl.ai/docs/api/integrations.liger.args.html - 2025-04-29T14:02:29.134Z + 2025-04-29T16:06:33.669Z https://docs.axolotl.ai/docs/api/core.trainer_builder.html - 2025-04-29T14:02:28.062Z + 2025-04-29T16:06:32.607Z https://docs.axolotl.ai/docs/api/utils.schemas.utils.html - 2025-04-29T14:02:29.046Z + 2025-04-29T16:06:33.583Z https://docs.axolotl.ai/docs/api/kernels.lora.html - 2025-04-29T14:02:28.673Z + 2025-04-29T16:06:33.213Z https://docs.axolotl.ai/docs/api/prompt_strategies.bradley_terry.llama3.html - 2025-04-29T14:02:28.591Z + 2025-04-29T16:06:33.131Z https://docs.axolotl.ai/docs/api/core.trainers.grpo.trainer.html - 2025-04-29T14:02:28.417Z + 2025-04-29T16:06:32.959Z https://docs.axolotl.ai/docs/api/prompt_strategies.messages.chat.html - 2025-04-29T14:02:28.523Z + 2025-04-29T16:06:33.064Z https://docs.axolotl.ai/docs/api/utils.tokenization.html - 2025-04-29T14:02:28.854Z + 2025-04-29T16:06:33.391Z https://docs.axolotl.ai/docs/api/prompt_strategies.chat_template.html - 2025-04-29T14:02:28.436Z + 2025-04-29T16:06:32.978Z https://docs.axolotl.ai/docs/api/prompt_strategies.stepwise_supervised.html - 2025-04-29T14:02:28.501Z + 2025-04-29T16:06:33.042Z https://docs.axolotl.ai/docs/api/utils.samplers.multipack.html - 2025-04-29T14:02:29.198Z + 2025-04-29T16:06:33.733Z https://docs.axolotl.ai/docs/api/cli.args.html - 2025-04-29T14:02:28.254Z + 2025-04-29T16:06:32.798Z https://docs.axolotl.ai/docs/api/utils.callbacks.perplexity.html - 2025-04-29T14:02:29.204Z + 2025-04-29T16:06:33.739Z https://docs.axolotl.ai/docs/api/utils.gradient_checkpointing.unsloth.html - 2025-04-29T14:02:28.967Z + 2025-04-29T16:06:33.504Z https://docs.axolotl.ai/docs/mac.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/config.html - 2025-04-29T14:01:58.459Z + 2025-04-29T16:05:56.141Z https://docs.axolotl.ai/docs/multimodal.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/lr_groups.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/dataset-formats/index.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/dataset-formats/stepwise_supervised.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/dataset-formats/pretraining.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/multi-node.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/sequence_parallelism.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/batch_vs_grad.html - 2025-04-29T14:01:58.459Z + 2025-04-29T16:05:56.141Z https://docs.axolotl.ai/docs/amd_hpc.html - 2025-04-29T14:01:58.459Z + 2025-04-29T16:05:56.141Z https://docs.axolotl.ai/docs/faq.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/custom_integrations.html - 2025-04-29T14:01:58.459Z + 2025-04-29T16:05:56.141Z https://docs.axolotl.ai/src/axolotl/integrations/LICENSE.html - 2025-04-29T14:01:58.480Z + 2025-04-29T16:05:56.161Z https://docs.axolotl.ai/TODO.html - 2025-04-29T14:01:58.458Z + 2025-04-29T16:05:56.140Z https://docs.axolotl.ai/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html - 2025-04-29T14:01:58.480Z + 2025-04-29T16:05:56.162Z https://docs.axolotl.ai/docs/getting-started.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/multipack.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/multi-gpu.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/installation.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/cli.html - 2025-04-29T14:01:58.459Z + 2025-04-29T16:05:56.141Z https://docs.axolotl.ai/docs/dataset-formats/inst_tune.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/dataset-formats/tokenized.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/dataset-formats/conversation.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/dataset-formats/template_free.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/reward_modelling.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/lora_optims.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/nccl.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/api/logging_config.html - 2025-04-29T14:02:28.047Z + 2025-04-29T16:06:32.592Z https://docs.axolotl.ai/docs/api/utils.trainer.html - 2025-04-29T14:02:28.906Z + 2025-04-29T16:06:33.442Z https://docs.axolotl.ai/docs/api/monkeypatch.unsloth_.html - 2025-04-29T14:02:28.808Z + 2025-04-29T16:06:33.346Z https://docs.axolotl.ai/docs/api/cli.evaluate.html - 2025-04-29T14:02:28.237Z + 2025-04-29T16:06:32.781Z https://docs.axolotl.ai/docs/api/kernels.utils.html - 2025-04-29T14:02:28.702Z + 2025-04-29T16:06:33.242Z https://docs.axolotl.ai/docs/api/datasets.html - 2025-04-29T14:02:27.987Z + 2025-04-29T16:06:32.531Z https://docs.axolotl.ai/docs/api/utils.models.html - 2025-04-29T14:02:28.847Z + 2025-04-29T16:06:33.384Z https://docs.axolotl.ai/docs/api/prompt_strategies.kto.llama3.html - 2025-04-29T14:02:28.557Z + 2025-04-29T16:06:33.098Z https://docs.axolotl.ai/docs/api/cli.preprocess.html - 2025-04-29T14:02:28.321Z + 2025-04-29T16:06:32.864Z https://docs.axolotl.ai/docs/api/cli.merge_sharded_fsdp_weights.html - 2025-04-29T14:02:28.313Z + 2025-04-29T16:06:32.856Z https://docs.axolotl.ai/docs/api/integrations.base.html - 2025-04-29T14:02:29.118Z + 2025-04-29T16:06:33.654Z https://docs.axolotl.ai/docs/api/prompt_strategies.orpo.chat_template.html - 2025-04-29T14:02:28.587Z + 2025-04-29T16:06:33.128Z https://docs.axolotl.ai/docs/api/utils.schemas.enums.html - 2025-04-29T14:02:29.041Z + 2025-04-29T16:06:33.577Z https://docs.axolotl.ai/docs/api/utils.callbacks.comet_.html - 2025-04-29T14:02:29.216Z + 2025-04-29T16:06:33.751Z https://docs.axolotl.ai/docs/api/prompt_strategies.input_output.html - 2025-04-29T14:02:28.497Z + 2025-04-29T16:06:33.038Z https://docs.axolotl.ai/docs/api/utils.schedulers.html - 2025-04-29T14:02:28.930Z + 2025-04-29T16:06:33.467Z https://docs.axolotl.ai/docs/api/cli.cloud.base.html - 2025-04-29T14:02:28.366Z + 2025-04-29T16:06:32.909Z https://docs.axolotl.ai/docs/api/cli.utils.html - 2025-04-29T14:02:28.358Z + 2025-04-29T16:06:32.901Z https://docs.axolotl.ai/docs/api/monkeypatch.lora_kernels.html - 2025-04-29T14:02:28.778Z + 2025-04-29T16:06:33.318Z https://docs.axolotl.ai/docs/api/prompt_strategies.completion.html - 2025-04-29T14:02:28.491Z + 2025-04-29T16:06:33.032Z https://docs.axolotl.ai/docs/api/utils.schemas.config.html - 2025-04-29T14:02:28.975Z + 2025-04-29T16:06:33.512Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_expand_mask.html - 2025-04-29T14:02:28.753Z + 2025-04-29T16:06:33.293Z https://docs.axolotl.ai/docs/api/cli.config.html - 2025-04-29T14:02:28.279Z + 2025-04-29T16:06:32.822Z https://docs.axolotl.ai/docs/api/utils.collators.core.html - 2025-04-29T14:02:29.162Z + 2025-04-29T16:06:33.698Z https://docs.axolotl.ai/docs/api/utils.data.sft.html - 2025-04-29T14:02:28.964Z + 2025-04-29T16:06:33.501Z https://docs.axolotl.ai/docs/api/integrations.spectrum.args.html - 2025-04-29T14:02:29.140Z + 2025-04-29T16:06:33.676Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.llama3.html - 2025-04-29T14:02:28.535Z + 2025-04-29T16:06:33.075Z https://docs.axolotl.ai/docs/api/cli.inference.html - 2025-04-29T14:02:28.293Z + 2025-04-29T16:06:32.836Z https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_instruct.html - 2025-04-29T14:02:28.451Z + 2025-04-29T16:06:32.993Z https://docs.axolotl.ai/docs/api/utils.collators.batching.html - 2025-04-29T14:02:29.181Z + 2025-04-29T16:06:33.716Z https://docs.axolotl.ai/docs/api/utils.schemas.datasets.html - 2025-04-29T14:02:29.005Z + 2025-04-29T16:06:33.541Z https://docs.axolotl.ai/docs/api/utils.data.pretraining.html - 2025-04-29T14:02:28.963Z + 2025-04-29T16:06:33.500Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.passthrough.html - 2025-04-29T14:02:28.549Z + 2025-04-29T16:06:33.090Z https://docs.axolotl.ai/docs/api/monkeypatch.utils.html - 2025-04-29T14:02:28.786Z + 2025-04-29T16:06:33.326Z https://docs.axolotl.ai/docs/api/utils.callbacks.lisa.html - 2025-04-29T14:02:29.209Z + 2025-04-29T16:06:33.744Z https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_chat.html - 2025-04-29T14:02:28.450Z + 2025-04-29T16:06:32.992Z https://docs.axolotl.ai/docs/api/prompt_strategies.kto.user_defined.html - 2025-04-29T14:02:28.567Z + 2025-04-29T16:06:33.107Z https://docs.axolotl.ai/docs/api/core.training_args.html - 2025-04-29T14:02:28.150Z + 2025-04-29T16:06:32.694Z https://docs.axolotl.ai/docs/api/utils.collators.mm_chat.html - 2025-04-29T14:02:29.190Z + 2025-04-29T16:06:33.725Z https://docs.axolotl.ai/docs/api/common.architectures.html - 2025-04-29T14:02:29.142Z + 2025-04-29T16:06:33.678Z https://docs.axolotl.ai/docs/api/integrations.kd.trainer.html - 2025-04-29T14:02:29.130Z + 2025-04-29T16:06:33.666Z https://docs.axolotl.ai/docs/api/prompt_strategies.llama2_chat.html - 2025-04-29T14:02:28.485Z + 2025-04-29T16:06:33.026Z https://docs.axolotl.ai/docs/api/core.trainers.trl.html - 2025-04-29T14:02:28.404Z + 2025-04-29T16:06:32.946Z https://docs.axolotl.ai/docs/api/prompt_strategies.orcamini.html - 2025-04-29T14:02:28.512Z + 2025-04-29T16:06:33.053Z https://docs.axolotl.ai/docs/api/integrations.cut_cross_entropy.args.html - 2025-04-29T14:02:29.122Z + 2025-04-29T16:06:33.658Z https://docs.axolotl.ai/docs/api/utils.lora.html - 2025-04-29T14:02:28.869Z + 2025-04-29T16:06:33.405Z https://docs.axolotl.ai/docs/api/cli.sweeps.html - 2025-04-29T14:02:28.327Z + 2025-04-29T16:06:32.870Z https://docs.axolotl.ai/docs/api/core.chat.format.chatml.html - 2025-04-29T14:02:28.174Z + 2025-04-29T16:06:32.718Z https://docs.axolotl.ai/docs/api/common.const.html - 2025-04-29T14:02:29.143Z + 2025-04-29T16:06:33.679Z https://docs.axolotl.ai/docs/api/prompt_strategies.metharme.html - 2025-04-29T14:02:28.508Z + 2025-04-29T16:06:33.049Z https://docs.axolotl.ai/docs/api/monkeypatch.stablelm_attn_hijack_flash.html - 2025-04-29T14:02:28.795Z + 2025-04-29T16:06:33.334Z https://docs.axolotl.ai/docs/api/integrations.grokfast.optimizer.html - 2025-04-29T14:02:29.123Z + 2025-04-29T16:06:33.659Z https://docs.axolotl.ai/docs/api/cli.checks.html - 2025-04-29T14:02:28.261Z + 2025-04-29T16:06:32.805Z https://docs.axolotl.ai/docs/api/common.datasets.html - 2025-04-29T14:02:29.160Z + 2025-04-29T16:06:33.695Z https://docs.axolotl.ai/docs/api/integrations.lm_eval.args.html - 2025-04-29T14:02:29.137Z + 2025-04-29T16:06:33.673Z https://docs.axolotl.ai/docs/api/monkeypatch.mixtral.html - 2025-04-29T14:02:28.818Z + 2025-04-29T16:06:33.355Z https://docs.axolotl.ai/docs/api/prompt_strategies.kto.chatml.html - 2025-04-29T14:02:28.565Z + 2025-04-29T16:06:33.106Z https://docs.axolotl.ai/docs/api/cli.train.html - 2025-04-29T14:02:28.229Z + 2025-04-29T16:06:32.773Z https://docs.axolotl.ai/docs/api/prompt_tokenizers.html - 2025-04-29T14:02:28.042Z + 2025-04-29T16:06:32.587Z https://docs.axolotl.ai/docs/api/utils.schemas.training.html - 2025-04-29T14:02:28.987Z + 2025-04-29T16:06:33.524Z https://docs.axolotl.ai/docs/api/monkeypatch.relora.html - 2025-04-29T14:02:28.752Z + 2025-04-29T16:06:33.292Z https://docs.axolotl.ai/docs/api/cli.cloud.modal_.html - 2025-04-29T14:02:28.373Z + 2025-04-29T16:06:32.915Z https://docs.axolotl.ai/docs/api/cli.main.html - 2025-04-29T14:02:28.221Z + 2025-04-29T16:06:32.765Z https://docs.axolotl.ai/docs/api/monkeypatch.trainer_fsdp_optim.html - 2025-04-29T14:02:28.798Z + 2025-04-29T16:06:33.338Z https://docs.axolotl.ai/docs/fsdp_qlora.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/debugging.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/ray-integration.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/docker.html - 2025-04-29T14:01:58.460Z + 2025-04-29T16:05:56.142Z https://docs.axolotl.ai/docs/inference.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/docs/torchao.html - 2025-04-29T14:01:58.463Z + 2025-04-29T16:05:56.145Z https://docs.axolotl.ai/FAQS.html - 2025-04-29T14:01:58.458Z + 2025-04-29T16:05:56.140Z