diff --git a/.nojekyll b/.nojekyll
index 94539e1f7..4c2dd3dae 100644
--- a/.nojekyll
+++ b/.nojekyll
@@ -1 +1 @@
-3cf13af5
\ No newline at end of file
+5b444df2
\ No newline at end of file
diff --git a/docs/api/utils.samplers.multipack.html b/docs/api/utils.samplers.multipack.html
index 7e78081b2..96ef76381 100644
--- a/docs/api/utils.samplers.multipack.html
+++ b/docs/api/utils.samplers.multipack.html
@@ -562,12 +562,12 @@ into fixed-capacity batches to optimize memory usage and training throughput.
batch_size,
batch_max_len,
lengths,
- packing_efficiency_estimate=1.0,
- drop_last=True,
- num_count_samples=4,
- sequential=False,
- group_size=100000,
- bin_size=200,
+ bin_size,
+ packing_efficiency_estimate=1.0,
+ drop_last=True,
+ num_count_samples=4,
+ sequential=False,
+ group_size=100000,
num_processes=None,
safe_mode=True,
mp_start_method='fork',
diff --git a/search.json b/search.json
index ea42006f2..6eb44508f 100644
--- a/search.json
+++ b/search.json
@@ -2658,14 +2658,14 @@
"href": "docs/api/utils.samplers.multipack.html",
"title": "utils.samplers.multipack",
"section": "",
- "text": "utils.samplers.multipack\nMultipack Batch Sampler - An efficient batch sampler for packing variable-length sequences\ninto fixed-capacity batches to optimize memory usage and training throughput.\n\n\n\n\n\nName\nDescription\n\n\n\n\nMultipackBatchSampler\nBatch sampler class for efficient packing of variable-length sequences\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler(\n sampler,\n batch_size,\n batch_max_len,\n lengths,\n packing_efficiency_estimate=1.0,\n drop_last=True,\n num_count_samples=4,\n sequential=False,\n group_size=100000,\n bin_size=200,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n **kwargs,\n)\nBatch sampler class for efficient packing of variable-length sequences\nThis sampler packs sequences into fixed-capacity bins (batches) to maximize\nGPU memory utilization and training throughput by reducing padding.\nIt supports both parallel packing (using FFD algorithm) and\nsequential packing (preserving original sequence order).\n\n\n\n\n\nName\nDescription\n\n\n\n\nefficiency\nCalculate the packing efficiency (ratio of tokens used to total token slots).\n\n\ngather_efficiency\nGather and synchronize packing efficiency estimates across all distributed\n\n\ngather_len_batches\nGather and synchronize batch counts across all distributed ranks. Returns\n\n\ngenerate_batches\nGenerate packed batches for training.\n\n\nset_epoch\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.efficiency()\nCalculate the packing efficiency (ratio of tokens used to total token slots).\nHigher is better - 1.0 would mean perfect packing with no wasted space.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_efficiency()\nGather and synchronize packing efficiency estimates across all distributed\nranks.\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nfloat\nA conservative efficiency estimate based on the measurements.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)\nGather and synchronize batch counts across all distributed ranks. Returns\nthe minimum number of batches available on any rank.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.generate_batches(set_stats=False)\nGenerate packed batches for training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nset_stats\nbool\nWhether to update efficiency statistics.\nFalse\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[list[int]]]\nList of batches, where each batch contains multiple bins, and each bin contains multiple sequence indices.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.set_epoch(epoch)\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\n\n\n\n\nName\nDescription\n\n\n\n\nallocate_sequentially\nSequential allocator that preserves example order.\n\n\nffd_check\nFirst-fit-decreasing bin packing algorithm check.\n\n\npack_group\nPack a group of sequences into bins using First-Fit Decreasing algorithm.\n\n\npack_parallel\nPack sequences into bins using parallel processing.\n\n\n\n\n\nutils.samplers.multipack.allocate_sequentially(\n sequence_lengths,\n rank,\n bin_capacity,\n num_ranks,\n)\nSequential allocator that preserves example order.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nThe lengths of all examples.\nrequired\n\n\nrank\nint\nThe current rank (for distributed training).\nrequired\n\n\nbin_capacity\nint\nThe capacity of each bin (maximum sequence length).\nrequired\n\n\nnum_ranks\nint\nNumber of ranks (processes / GPUs).\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\nrank_batches\nlist[list[int]]\nList of batches for the current rank.\n\n\ntotal_tokens_used\nint\nNumber of actual example tokens.\n\n\ntotal_token_slots\nint\nMaximum theoretical number of example tokens (number of bins * bin capacity).\n\n\n\n\n\n\n\nutils.samplers.multipack.ffd_check(sequence_lengths, bin_capacity, num_bins)\nFirst-fit-decreasing bin packing algorithm check.\nChecks if sequences with the given lengths could fit in the specified number of\nbins.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin.\nrequired\n\n\nnum_bins\nint\nNumber of bins available.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nbool\nTrue if all sequences can be packed, False otherwise.\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_group(\n sequence_lengths,\n group_offset,\n bin_capacity,\n max_bins,\n bin_size,\n safe_mode=True,\n)\nPack a group of sequences into bins using First-Fit Decreasing algorithm.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\ngroup_offset\nint\nOffset to apply to indices when returning results.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin.\nrequired\n\n\nmax_bins\nint\nMaximum number of bins to use.\nrequired\n\n\nbin_size\nint\nMaximum number of sequences per bin.\nrequired\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach.\nTrue\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[int]]\nList of bins, where each bin contains indices of sequences assigned to it.\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_parallel(\n sequence_lengths,\n bin_capacity,\n group_size,\n bin_size,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n)\nPack sequences into bins using parallel processing.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin as total number of tokens.\nrequired\n\n\ngroup_size\nint\nNumber of sequences to process in each group.\nrequired\n\n\nbin_size\nint\nMaximum number of bins to use.\nrequired\n\n\nnum_processes\nint | None\nNumber of parallel processes to use.\nNone\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach.\nTrue\n\n\nmp_start_method\nstr | None\nMultiprocessing start method (‘fork’, ‘spawn’, ‘forkserver’). ‘spawn’ is often safer with Numba/PyTorch. Set to None to use system default.\n'fork'\n\n\n\nReturns:\nList of bins, where each bin contains indices of sequences assigned to it."
+ "text": "utils.samplers.multipack\nMultipack Batch Sampler - An efficient batch sampler for packing variable-length sequences\ninto fixed-capacity batches to optimize memory usage and training throughput.\n\n\n\n\n\nName\nDescription\n\n\n\n\nMultipackBatchSampler\nBatch sampler class for efficient packing of variable-length sequences\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler(\n sampler,\n batch_size,\n batch_max_len,\n lengths,\n bin_size,\n packing_efficiency_estimate=1.0,\n drop_last=True,\n num_count_samples=4,\n sequential=False,\n group_size=100000,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n **kwargs,\n)\nBatch sampler class for efficient packing of variable-length sequences\nThis sampler packs sequences into fixed-capacity bins (batches) to maximize\nGPU memory utilization and training throughput by reducing padding.\nIt supports both parallel packing (using FFD algorithm) and\nsequential packing (preserving original sequence order).\n\n\n\n\n\nName\nDescription\n\n\n\n\nefficiency\nCalculate the packing efficiency (ratio of tokens used to total token slots).\n\n\ngather_efficiency\nGather and synchronize packing efficiency estimates across all distributed\n\n\ngather_len_batches\nGather and synchronize batch counts across all distributed ranks. Returns\n\n\ngenerate_batches\nGenerate packed batches for training.\n\n\nset_epoch\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.efficiency()\nCalculate the packing efficiency (ratio of tokens used to total token slots).\nHigher is better - 1.0 would mean perfect packing with no wasted space.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_efficiency()\nGather and synchronize packing efficiency estimates across all distributed\nranks.\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nfloat\nA conservative efficiency estimate based on the measurements.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)\nGather and synchronize batch counts across all distributed ranks. Returns\nthe minimum number of batches available on any rank.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.generate_batches(set_stats=False)\nGenerate packed batches for training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nset_stats\nbool\nWhether to update efficiency statistics.\nFalse\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[list[int]]]\nList of batches, where each batch contains multiple bins, and each bin contains multiple sequence indices.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.set_epoch(epoch)\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\n\n\n\n\nName\nDescription\n\n\n\n\nallocate_sequentially\nSequential allocator that preserves example order.\n\n\nffd_check\nFirst-fit-decreasing bin packing algorithm check.\n\n\npack_group\nPack a group of sequences into bins using First-Fit Decreasing algorithm.\n\n\npack_parallel\nPack sequences into bins using parallel processing.\n\n\n\n\n\nutils.samplers.multipack.allocate_sequentially(\n sequence_lengths,\n rank,\n bin_capacity,\n num_ranks,\n)\nSequential allocator that preserves example order.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nThe lengths of all examples.\nrequired\n\n\nrank\nint\nThe current rank (for distributed training).\nrequired\n\n\nbin_capacity\nint\nThe capacity of each bin (maximum sequence length).\nrequired\n\n\nnum_ranks\nint\nNumber of ranks (processes / GPUs).\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\nrank_batches\nlist[list[int]]\nList of batches for the current rank.\n\n\ntotal_tokens_used\nint\nNumber of actual example tokens.\n\n\ntotal_token_slots\nint\nMaximum theoretical number of example tokens (number of bins * bin capacity).\n\n\n\n\n\n\n\nutils.samplers.multipack.ffd_check(sequence_lengths, bin_capacity, num_bins)\nFirst-fit-decreasing bin packing algorithm check.\nChecks if sequences with the given lengths could fit in the specified number of\nbins.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin.\nrequired\n\n\nnum_bins\nint\nNumber of bins available.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nbool\nTrue if all sequences can be packed, False otherwise.\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_group(\n sequence_lengths,\n group_offset,\n bin_capacity,\n max_bins,\n bin_size,\n safe_mode=True,\n)\nPack a group of sequences into bins using First-Fit Decreasing algorithm.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\ngroup_offset\nint\nOffset to apply to indices when returning results.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin.\nrequired\n\n\nmax_bins\nint\nMaximum number of bins to use.\nrequired\n\n\nbin_size\nint\nMaximum number of sequences per bin.\nrequired\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach.\nTrue\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[int]]\nList of bins, where each bin contains indices of sequences assigned to it.\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_parallel(\n sequence_lengths,\n bin_capacity,\n group_size,\n bin_size,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n)\nPack sequences into bins using parallel processing.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin as total number of tokens.\nrequired\n\n\ngroup_size\nint\nNumber of sequences to process in each group.\nrequired\n\n\nbin_size\nint\nMaximum number of bins to use.\nrequired\n\n\nnum_processes\nint | None\nNumber of parallel processes to use.\nNone\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach.\nTrue\n\n\nmp_start_method\nstr | None\nMultiprocessing start method (‘fork’, ‘spawn’, ‘forkserver’). ‘spawn’ is often safer with Numba/PyTorch. Set to None to use system default.\n'fork'\n\n\n\nReturns:\nList of bins, where each bin contains indices of sequences assigned to it."
},
{
"objectID": "docs/api/utils.samplers.multipack.html#classes",
"href": "docs/api/utils.samplers.multipack.html#classes",
"title": "utils.samplers.multipack",
"section": "",
- "text": "Name\nDescription\n\n\n\n\nMultipackBatchSampler\nBatch sampler class for efficient packing of variable-length sequences\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler(\n sampler,\n batch_size,\n batch_max_len,\n lengths,\n packing_efficiency_estimate=1.0,\n drop_last=True,\n num_count_samples=4,\n sequential=False,\n group_size=100000,\n bin_size=200,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n **kwargs,\n)\nBatch sampler class for efficient packing of variable-length sequences\nThis sampler packs sequences into fixed-capacity bins (batches) to maximize\nGPU memory utilization and training throughput by reducing padding.\nIt supports both parallel packing (using FFD algorithm) and\nsequential packing (preserving original sequence order).\n\n\n\n\n\nName\nDescription\n\n\n\n\nefficiency\nCalculate the packing efficiency (ratio of tokens used to total token slots).\n\n\ngather_efficiency\nGather and synchronize packing efficiency estimates across all distributed\n\n\ngather_len_batches\nGather and synchronize batch counts across all distributed ranks. Returns\n\n\ngenerate_batches\nGenerate packed batches for training.\n\n\nset_epoch\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.efficiency()\nCalculate the packing efficiency (ratio of tokens used to total token slots).\nHigher is better - 1.0 would mean perfect packing with no wasted space.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_efficiency()\nGather and synchronize packing efficiency estimates across all distributed\nranks.\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nfloat\nA conservative efficiency estimate based on the measurements.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)\nGather and synchronize batch counts across all distributed ranks. Returns\nthe minimum number of batches available on any rank.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.generate_batches(set_stats=False)\nGenerate packed batches for training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nset_stats\nbool\nWhether to update efficiency statistics.\nFalse\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[list[int]]]\nList of batches, where each batch contains multiple bins, and each bin contains multiple sequence indices.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.set_epoch(epoch)\nSet the epoch number, used for reproducible shuffling across epochs"
+ "text": "Name\nDescription\n\n\n\n\nMultipackBatchSampler\nBatch sampler class for efficient packing of variable-length sequences\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler(\n sampler,\n batch_size,\n batch_max_len,\n lengths,\n bin_size,\n packing_efficiency_estimate=1.0,\n drop_last=True,\n num_count_samples=4,\n sequential=False,\n group_size=100000,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n **kwargs,\n)\nBatch sampler class for efficient packing of variable-length sequences\nThis sampler packs sequences into fixed-capacity bins (batches) to maximize\nGPU memory utilization and training throughput by reducing padding.\nIt supports both parallel packing (using FFD algorithm) and\nsequential packing (preserving original sequence order).\n\n\n\n\n\nName\nDescription\n\n\n\n\nefficiency\nCalculate the packing efficiency (ratio of tokens used to total token slots).\n\n\ngather_efficiency\nGather and synchronize packing efficiency estimates across all distributed\n\n\ngather_len_batches\nGather and synchronize batch counts across all distributed ranks. Returns\n\n\ngenerate_batches\nGenerate packed batches for training.\n\n\nset_epoch\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.efficiency()\nCalculate the packing efficiency (ratio of tokens used to total token slots).\nHigher is better - 1.0 would mean perfect packing with no wasted space.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_efficiency()\nGather and synchronize packing efficiency estimates across all distributed\nranks.\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nfloat\nA conservative efficiency estimate based on the measurements.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)\nGather and synchronize batch counts across all distributed ranks. Returns\nthe minimum number of batches available on any rank.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.generate_batches(set_stats=False)\nGenerate packed batches for training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nset_stats\nbool\nWhether to update efficiency statistics.\nFalse\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[list[int]]]\nList of batches, where each batch contains multiple bins, and each bin contains multiple sequence indices.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.set_epoch(epoch)\nSet the epoch number, used for reproducible shuffling across epochs"
},
{
"objectID": "docs/api/utils.samplers.multipack.html#functions",
diff --git a/sitemap.xml b/sitemap.xml
index 62b3954c0..1a389c490 100644
--- a/sitemap.xml
+++ b/sitemap.xml
@@ -2,802 +2,802 @@
https://docs.axolotl.ai/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html
- 2025-12-04T14:44:55.031Z
+ 2025-12-08T14:16:36.173Z
https://docs.axolotl.ai/docs/mac.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/cli.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.144Z
https://docs.axolotl.ai/docs/mixed_precision.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/installation.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/dataset_loading.html
- 2025-12-04T14:44:55.005Z
+ 2025-12-08T14:16:36.144Z
https://docs.axolotl.ai/docs/sequence_parallelism.html
- 2025-12-04T14:44:55.009Z
+ 2025-12-08T14:16:36.150Z
https://docs.axolotl.ai/docs/optimizations.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/gradient_checkpointing.html
- 2025-12-04T14:44:55.005Z
+ 2025-12-08T14:16:36.145Z
https://docs.axolotl.ai/docs/docker.html
- 2025-12-04T14:44:55.005Z
+ 2025-12-08T14:16:36.145Z
https://docs.axolotl.ai/docs/input_output.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.148Z
https://docs.axolotl.ai/docs/multi-gpu.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/dataset-formats/template_free.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.144Z
https://docs.axolotl.ai/docs/dataset-formats/tokenized.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.144Z
https://docs.axolotl.ai/docs/dataset-formats/pretraining.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.144Z
https://docs.axolotl.ai/docs/dataset-formats/stepwise_supervised.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.144Z
https://docs.axolotl.ai/docs/rlhf.html
- 2025-12-04T14:44:55.009Z
+ 2025-12-08T14:16:36.150Z
https://docs.axolotl.ai/docs/multi-node.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/batch_vs_grad.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.143Z
https://docs.axolotl.ai/docs/nd_parallelism.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/fsdp_qlora.html
- 2025-12-04T14:44:55.005Z
+ 2025-12-08T14:16:36.145Z
https://docs.axolotl.ai/docs/quantize.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.150Z
https://docs.axolotl.ai/docs/reward_modelling.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.150Z
https://docs.axolotl.ai/docs/api/core.chat.format.llama3x.html
- 2025-12-04T15:01:38.011Z
+ 2025-12-08T14:20:16.823Z
https://docs.axolotl.ai/docs/api/monkeypatch.unsloth_.html
- 2025-12-04T15:01:38.917Z
+ 2025-12-08T14:20:17.719Z
https://docs.axolotl.ai/docs/api/integrations.kd.trainer.html
- 2025-12-04T15:01:39.422Z
+ 2025-12-08T14:20:18.218Z
https://docs.axolotl.ai/docs/api/integrations.grokfast.optimizer.html
- 2025-12-04T15:01:39.413Z
+ 2025-12-08T14:20:18.209Z
https://docs.axolotl.ai/docs/api/core.chat.format.chatml.html
- 2025-12-04T15:01:38.010Z
+ 2025-12-08T14:20:16.821Z
https://docs.axolotl.ai/docs/api/utils.callbacks.perplexity.html
- 2025-12-04T15:01:39.548Z
+ 2025-12-08T14:20:18.344Z
https://docs.axolotl.ai/docs/api/utils.callbacks.profiler.html
- 2025-12-04T15:01:39.553Z
+ 2025-12-08T14:20:18.349Z
https://docs.axolotl.ai/docs/api/cli.preprocess.html
- 2025-12-04T15:01:38.204Z
+ 2025-12-08T14:20:17.013Z
https://docs.axolotl.ai/docs/api/cli.utils.load.html
- 2025-12-04T15:01:38.260Z
+ 2025-12-08T14:20:17.068Z
https://docs.axolotl.ai/docs/api/cli.inference.html
- 2025-12-04T15:01:38.170Z
+ 2025-12-08T14:20:16.979Z
https://docs.axolotl.ai/docs/api/cli.args.html
- 2025-12-04T15:01:38.113Z
+ 2025-12-08T14:20:16.924Z
https://docs.axolotl.ai/docs/api/prompt_tokenizers.html
- 2025-12-04T15:01:37.933Z
+ 2025-12-08T14:20:16.745Z
https://docs.axolotl.ai/docs/api/cli.vllm_serve.html
- 2025-12-04T15:01:38.219Z
+ 2025-12-08T14:20:17.027Z
https://docs.axolotl.ai/docs/api/monkeypatch.data.batch_dataset_fetcher.html
- 2025-12-04T15:01:38.929Z
+ 2025-12-08T14:20:17.730Z
https://docs.axolotl.ai/docs/api/loaders.constants.html
- 2025-12-04T15:01:38.406Z
+ 2025-12-08T14:20:17.213Z
https://docs.axolotl.ai/docs/api/prompt_strategies.input_output.html
- 2025-12-04T15:01:38.568Z
+ 2025-12-08T14:20:17.373Z
https://docs.axolotl.ai/docs/api/core.trainers.trl.html
- 2025-12-04T15:01:38.317Z
+ 2025-12-08T14:20:17.124Z
https://docs.axolotl.ai/docs/api/models.mamba.modeling_mamba.html
- 2025-12-04T15:01:39.457Z
+ 2025-12-08T14:20:18.253Z
https://docs.axolotl.ai/docs/api/monkeypatch.stablelm_attn_hijack_flash.html
- 2025-12-04T15:01:38.904Z
+ 2025-12-08T14:20:17.705Z
https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_chat.html
- 2025-12-04T15:01:38.511Z
+ 2025-12-08T14:20:17.317Z
https://docs.axolotl.ai/docs/api/prompt_strategies.metharme.html
- 2025-12-04T15:01:38.581Z
+ 2025-12-08T14:20:17.386Z
https://docs.axolotl.ai/docs/api/utils.freeze.html
- 2025-12-04T15:01:39.004Z
+ 2025-12-08T14:20:17.805Z
https://docs.axolotl.ai/docs/api/logging_config.html
- 2025-12-04T15:01:37.944Z
+ 2025-12-08T14:20:16.757Z
https://docs.axolotl.ai/docs/api/monkeypatch.trainer_fsdp_optim.html
- 2025-12-04T15:01:38.908Z
+ 2025-12-08T14:20:17.710Z
https://docs.axolotl.ai/docs/api/cli.art.html
- 2025-12-04T15:01:38.117Z
+ 2025-12-08T14:20:16.928Z
https://docs.axolotl.ai/docs/api/utils.collators.mamba.html
- 2025-12-04T15:01:39.486Z
+ 2025-12-08T14:20:18.282Z
https://docs.axolotl.ai/docs/api/cli.utils.train.html
- 2025-12-04T15:01:38.281Z
+ 2025-12-08T14:20:17.089Z
https://docs.axolotl.ai/docs/api/integrations.base.html
- 2025-12-04T15:01:39.407Z
+ 2025-12-08T14:20:18.204Z
https://docs.axolotl.ai/docs/api/core.trainers.dpo.trainer.html
- 2025-12-04T15:01:38.332Z
+ 2025-12-08T14:20:17.139Z
https://docs.axolotl.ai/docs/api/cli.checks.html
- 2025-12-04T15:01:38.125Z
+ 2025-12-08T14:20:16.935Z
https://docs.axolotl.ai/docs/api/common.architectures.html
- 2025-12-04T15:01:39.436Z
+ 2025-12-08T14:20:18.232Z
https://docs.axolotl.ai/docs/api/utils.bench.html
- 2025-12-04T15:01:38.994Z
+ 2025-12-08T14:20:17.795Z
https://docs.axolotl.ai/docs/api/utils.data.streaming.html
- 2025-12-04T15:01:39.101Z
+ 2025-12-08T14:20:17.900Z
https://docs.axolotl.ai/docs/api/utils.chat_templates.html
- 2025-12-04T15:01:38.977Z
+ 2025-12-08T14:20:17.778Z
https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_w_system.html
- 2025-12-04T15:01:38.528Z
+ 2025-12-08T14:20:17.333Z
https://docs.axolotl.ai/docs/api/core.trainers.utils.html
- 2025-12-04T15:01:38.361Z
+ 2025-12-08T14:20:17.168Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.zephyr.html
- 2025-12-04T15:01:38.634Z
+ 2025-12-08T14:20:17.439Z
https://docs.axolotl.ai/docs/api/cli.evaluate.html
- 2025-12-04T15:01:38.089Z
+ 2025-12-08T14:20:16.900Z
https://docs.axolotl.ai/docs/api/monkeypatch.relora.html
- 2025-12-04T15:01:38.847Z
+ 2025-12-08T14:20:17.649Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chatml.html
- 2025-12-04T15:01:38.633Z
+ 2025-12-08T14:20:17.437Z
https://docs.axolotl.ai/docs/api/cli.utils.fetch.html
- 2025-12-04T15:01:38.253Z
+ 2025-12-08T14:20:17.061Z
https://docs.axolotl.ai/docs/api/utils.schemas.datasets.html
- 2025-12-04T15:01:39.186Z
+ 2025-12-08T14:20:17.984Z
https://docs.axolotl.ai/docs/api/common.datasets.html
- 2025-12-04T15:01:39.456Z
+ 2025-12-08T14:20:18.252Z
https://docs.axolotl.ai/docs/api/utils.callbacks.mlflow_.html
- 2025-12-04T15:01:39.559Z
+ 2025-12-08T14:20:18.355Z
https://docs.axolotl.ai/docs/api/prompt_strategies.kto.chatml.html
- 2025-12-04T15:01:38.658Z
+ 2025-12-08T14:20:17.463Z
https://docs.axolotl.ai/docs/api/utils.schemas.model.html
- 2025-12-04T15:01:39.156Z
+ 2025-12-08T14:20:17.954Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_xformers.html
- 2025-12-04T15:01:38.839Z
+ 2025-12-08T14:20:17.641Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_patch_multipack.html
- 2025-12-04T15:01:38.897Z
+ 2025-12-08T14:20:17.698Z
https://docs.axolotl.ai/docs/api/core.trainers.base.html
- 2025-12-04T15:01:38.299Z
+ 2025-12-08T14:20:17.106Z
https://docs.axolotl.ai/docs/api/cli.utils.args.html
- 2025-12-04T15:01:38.247Z
+ 2025-12-08T14:20:17.055Z
https://docs.axolotl.ai/docs/api/utils.schemas.trl.html
- 2025-12-04T15:01:39.201Z
+ 2025-12-08T14:20:17.999Z
https://docs.axolotl.ai/docs/api/core.builders.base.html
- 2025-12-04T15:01:37.952Z
+ 2025-12-08T14:20:16.764Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_flash.html
- 2025-12-04T15:01:38.837Z
+ 2025-12-08T14:20:17.639Z
https://docs.axolotl.ai/docs/api/cli.cloud.modal_.html
- 2025-12-04T15:01:38.231Z
+ 2025-12-08T14:20:17.039Z
https://docs.axolotl.ai/docs/api/utils.optimizers.adopt.html
- 2025-12-04T15:01:39.099Z
+ 2025-12-08T14:20:17.898Z
https://docs.axolotl.ai/docs/api/integrations.spectrum.args.html
- 2025-12-04T15:01:39.434Z
+ 2025-12-08T14:20:18.230Z
https://docs.axolotl.ai/docs/api/cli.main.html
- 2025-12-04T15:01:38.069Z
+ 2025-12-08T14:20:16.880Z
https://docs.axolotl.ai/docs/api/kernels.lora.html
- 2025-12-04T15:01:38.794Z
+ 2025-12-08T14:20:17.597Z
https://docs.axolotl.ai/docs/api/utils.model_shard_quant.html
- 2025-12-04T15:01:38.990Z
+ 2025-12-08T14:20:17.791Z
https://docs.axolotl.ai/docs/api/utils.distributed.html
- 2025-12-04T15:01:39.082Z
+ 2025-12-08T14:20:17.882Z
https://docs.axolotl.ai/docs/api/loaders.model.html
- 2025-12-04T15:01:38.373Z
+ 2025-12-08T14:20:17.180Z
https://docs.axolotl.ai/docs/api/monkeypatch.mistral_attn_hijack_flash.html
- 2025-12-04T15:01:38.840Z
+ 2025-12-08T14:20:17.643Z
https://docs.axolotl.ai/docs/api/evaluate.html
- 2025-12-04T15:01:37.857Z
+ 2025-12-08T14:20:16.670Z
https://docs.axolotl.ai/docs/api/cli.delinearize_llama4.html
- 2025-12-04T15:01:38.152Z
+ 2025-12-08T14:20:16.962Z
https://docs.axolotl.ai/docs/api/utils.trainer.html
- 2025-12-04T15:01:39.024Z
+ 2025-12-08T14:20:17.825Z
https://docs.axolotl.ai/docs/api/cli.quantize.html
- 2025-12-04T15:01:38.210Z
+ 2025-12-08T14:20:17.019Z
https://docs.axolotl.ai/docs/api/common.const.html
- 2025-12-04T15:01:39.438Z
+ 2025-12-08T14:20:18.234Z
https://docs.axolotl.ai/docs/api/prompt_strategies.llama2_chat.html
- 2025-12-04T15:01:38.553Z
+ 2025-12-08T14:20:17.358Z
https://docs.axolotl.ai/docs/api/utils.collators.mm_chat.html
- 2025-12-04T15:01:39.492Z
+ 2025-12-08T14:20:18.288Z
https://docs.axolotl.ai/docs/api/core.datasets.chat.html
- 2025-12-04T15:01:38.019Z
+ 2025-12-08T14:20:16.831Z
https://docs.axolotl.ai/docs/api/prompt_strategies.bradley_terry.llama3.html
- 2025-12-04T15:01:38.690Z
+ 2025-12-08T14:20:17.494Z
https://docs.axolotl.ai/docs/api/core.trainers.mamba.html
- 2025-12-04T15:01:38.324Z
+ 2025-12-08T14:20:17.131Z
https://docs.axolotl.ai/docs/api/core.builders.causal.html
- 2025-12-04T15:01:37.958Z
+ 2025-12-08T14:20:16.770Z
https://docs.axolotl.ai/docs/api/kernels.geglu.html
- 2025-12-04T15:01:38.807Z
+ 2025-12-08T14:20:17.610Z
https://docs.axolotl.ai/docs/api/utils.schemas.utils.html
- 2025-12-04T15:01:39.242Z
+ 2025-12-08T14:20:18.040Z
https://docs.axolotl.ai/docs/custom_integrations.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.144Z
https://docs.axolotl.ai/index.html
- 2025-12-04T14:44:55.026Z
+ 2025-12-08T14:16:36.168Z
https://docs.axolotl.ai/examples/colab-notebooks/colab-axolotl-example.html
- 2025-12-04T14:44:55.013Z
+ 2025-12-08T14:16:36.155Z
https://docs.axolotl.ai/FAQS.html
- 2025-12-04T14:44:55.002Z
+ 2025-12-08T14:16:36.141Z
https://docs.axolotl.ai/docs/inference.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.148Z
https://docs.axolotl.ai/docs/api/core.datasets.transforms.chat_builder.html
- 2025-12-04T15:01:38.029Z
+ 2025-12-08T14:20:16.840Z
https://docs.axolotl.ai/docs/api/train.html
- 2025-12-04T15:01:37.844Z
+ 2025-12-08T14:20:16.658Z
https://docs.axolotl.ai/docs/api/utils.schemas.training.html
- 2025-12-04T15:01:39.164Z
+ 2025-12-08T14:20:17.963Z
https://docs.axolotl.ai/docs/api/kernels.quantize.html
- 2025-12-04T15:01:38.828Z
+ 2025-12-08T14:20:17.631Z
https://docs.axolotl.ai/docs/api/utils.dict.html
- 2025-12-04T15:01:39.089Z
+ 2025-12-08T14:20:17.889Z
https://docs.axolotl.ai/docs/api/utils.quantization.html
- 2025-12-04T15:01:39.132Z
+ 2025-12-08T14:20:17.930Z
https://docs.axolotl.ai/docs/api/utils.collators.batching.html
- 2025-12-04T15:01:39.482Z
+ 2025-12-08T14:20:18.278Z
https://docs.axolotl.ai/docs/api/monkeypatch.multipack.html
- 2025-12-04T15:01:38.842Z
+ 2025-12-08T14:20:17.645Z
https://docs.axolotl.ai/docs/api/cli.config.html
- 2025-12-04T15:01:38.146Z
+ 2025-12-08T14:20:16.956Z
https://docs.axolotl.ai/docs/api/core.trainers.mixins.optimizer.html
- 2025-12-04T15:01:38.412Z
+ 2025-12-08T14:20:17.219Z
https://docs.axolotl.ai/docs/api/utils.lora.html
- 2025-12-04T15:01:38.983Z
+ 2025-12-08T14:20:17.784Z
https://docs.axolotl.ai/docs/api/core.trainers.grpo.sampler.html
- 2025-12-04T15:01:38.360Z
+ 2025-12-08T14:20:17.166Z
https://docs.axolotl.ai/docs/api/prompt_strategies.orcamini.html
- 2025-12-04T15:01:38.586Z
+ 2025-12-08T14:20:17.391Z
https://docs.axolotl.ai/docs/api/monkeypatch.mixtral.html
- 2025-12-04T15:01:38.930Z
+ 2025-12-08T14:20:17.732Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.llama3.html
- 2025-12-04T15:01:38.620Z
+ 2025-12-08T14:20:17.425Z
https://docs.axolotl.ai/docs/api/index.html
- 2025-12-04T15:01:37.765Z
+ 2025-12-08T14:20:16.580Z
https://docs.axolotl.ai/docs/api/utils.ctx_managers.sequence_parallel.html
- 2025-12-04T15:01:38.453Z
+ 2025-12-08T14:20:17.260Z
https://docs.axolotl.ai/docs/api/cli.utils.html
- 2025-12-04T15:01:38.233Z
+ 2025-12-08T14:20:17.041Z
https://docs.axolotl.ai/docs/api/cli.merge_lora.html
- 2025-12-04T15:01:38.180Z
+ 2025-12-08T14:20:16.989Z
https://docs.axolotl.ai/docs/api/kernels.utils.html
- 2025-12-04T15:01:38.830Z
+ 2025-12-08T14:20:17.633Z
https://docs.axolotl.ai/docs/api/prompt_strategies.base.html
- 2025-12-04T15:01:38.455Z
+ 2025-12-08T14:20:17.261Z
https://docs.axolotl.ai/docs/api/utils.callbacks.comet_.html
- 2025-12-04T15:01:39.563Z
+ 2025-12-08T14:20:18.359Z
https://docs.axolotl.ai/docs/api/utils.schemas.multimodal.html
- 2025-12-04T15:01:39.207Z
+ 2025-12-08T14:20:18.005Z
https://docs.axolotl.ai/docs/api/prompt_strategies.chat_template.html
- 2025-12-04T15:01:38.495Z
+ 2025-12-08T14:20:17.301Z
https://docs.axolotl.ai/docs/api/utils.callbacks.qat.html
- 2025-12-04T15:01:39.572Z
+ 2025-12-08T14:20:18.367Z
https://docs.axolotl.ai/docs/api/utils.samplers.multipack.html
- 2025-12-04T15:01:39.541Z
+ 2025-12-08T14:20:18.337Z
https://docs.axolotl.ai/docs/api/prompt_strategies.orpo.chat_template.html
- 2025-12-04T15:01:38.685Z
+ 2025-12-08T14:20:17.489Z
https://docs.axolotl.ai/docs/api/core.chat.format.shared.html
- 2025-12-04T15:01:38.013Z
+ 2025-12-08T14:20:16.825Z
https://docs.axolotl.ai/docs/api/utils.collators.core.html
- 2025-12-04T15:01:39.459Z
+ 2025-12-08T14:20:18.255Z
https://docs.axolotl.ai/docs/api/integrations.lm_eval.args.html
- 2025-12-04T15:01:39.430Z
+ 2025-12-08T14:20:18.226Z
https://docs.axolotl.ai/docs/api/core.trainers.grpo.trainer.html
- 2025-12-04T15:01:38.345Z
+ 2025-12-08T14:20:17.152Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chat_template.html
- 2025-12-04T15:01:38.607Z
+ 2025-12-08T14:20:17.412Z
https://docs.axolotl.ai/docs/api/monkeypatch.btlm_attn_hijack_flash.html
- 2025-12-04T15:01:38.895Z
+ 2025-12-08T14:20:17.697Z
https://docs.axolotl.ai/docs/api/utils.schemas.enums.html
- 2025-12-04T15:01:39.235Z
+ 2025-12-08T14:20:18.033Z
https://docs.axolotl.ai/docs/api/utils.schemas.config.html
- 2025-12-04T15:01:39.147Z
+ 2025-12-08T14:20:17.946Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.passthrough.html
- 2025-12-04T15:01:38.638Z
+ 2025-12-08T14:20:17.443Z
https://docs.axolotl.ai/docs/api/convert.html
- 2025-12-04T15:01:37.881Z
+ 2025-12-08T14:20:16.694Z
https://docs.axolotl.ai/docs/api/utils.schemas.integrations.html
- 2025-12-04T15:01:39.224Z
+ 2025-12-08T14:20:18.022Z
https://docs.axolotl.ai/docs/api/prompt_strategies.kto.llama3.html
- 2025-12-04T15:01:38.648Z
+ 2025-12-08T14:20:17.453Z
https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.user_defined.html
- 2025-12-04T15:01:38.636Z
+ 2025-12-08T14:20:17.441Z
https://docs.axolotl.ai/docs/api/core.trainers.mixins.scheduler.html
- 2025-12-04T15:01:38.424Z
+ 2025-12-08T14:20:17.231Z
https://docs.axolotl.ai/docs/api/core.chat.messages.html
- 2025-12-04T15:01:38.008Z
+ 2025-12-08T14:20:16.819Z
https://docs.axolotl.ai/docs/api/prompt_strategies.user_defined.html
- 2025-12-04T15:01:38.538Z
+ 2025-12-08T14:20:17.343Z
https://docs.axolotl.ai/docs/api/prompt_strategies.messages.chat.html
- 2025-12-04T15:01:38.599Z
+ 2025-12-08T14:20:17.404Z
https://docs.axolotl.ai/docs/api/monkeypatch.lora_kernels.html
- 2025-12-04T15:01:38.883Z
+ 2025-12-08T14:20:17.685Z
https://docs.axolotl.ai/docs/api/prompt_strategies.stepwise_supervised.html
- 2025-12-04T15:01:38.573Z
+ 2025-12-08T14:20:17.378Z
https://docs.axolotl.ai/docs/api/prompt_strategies.completion.html
- 2025-12-04T15:01:38.560Z
+ 2025-12-08T14:20:17.366Z
https://docs.axolotl.ai/docs/api/core.trainers.mixins.rng_state_loader.html
- 2025-12-04T15:01:38.416Z
+ 2025-12-08T14:20:17.223Z
https://docs.axolotl.ai/docs/api/cli.train.html
- 2025-12-04T15:01:38.079Z
+ 2025-12-08T14:20:16.890Z
https://docs.axolotl.ai/docs/api/cli.merge_sharded_fsdp_weights.html
- 2025-12-04T15:01:38.194Z
+ 2025-12-08T14:20:17.003Z
https://docs.axolotl.ai/docs/api/loaders.adapter.html
- 2025-12-04T15:01:38.392Z
+ 2025-12-08T14:20:17.198Z
https://docs.axolotl.ai/docs/api/core.training_args.html
- 2025-12-04T15:01:37.979Z
+ 2025-12-08T14:20:16.791Z
https://docs.axolotl.ai/docs/api/utils.callbacks.lisa.html
- 2025-12-04T15:01:39.555Z
+ 2025-12-08T14:20:18.351Z
https://docs.axolotl.ai/docs/api/loaders.processor.html
- 2025-12-04T15:01:38.385Z
+ 2025-12-08T14:20:17.191Z
https://docs.axolotl.ai/docs/api/core.builders.rl.html
- 2025-12-04T15:01:37.964Z
+ 2025-12-08T14:20:16.775Z
https://docs.axolotl.ai/docs/api/monkeypatch.utils.html
- 2025-12-04T15:01:38.893Z
+ 2025-12-08T14:20:17.695Z
https://docs.axolotl.ai/docs/api/prompt_strategies.kto.user_defined.html
- 2025-12-04T15:01:38.660Z
+ 2025-12-08T14:20:17.464Z
https://docs.axolotl.ai/docs/api/integrations.cut_cross_entropy.args.html
- 2025-12-04T15:01:39.412Z
+ 2025-12-08T14:20:18.208Z
https://docs.axolotl.ai/docs/api/kernels.swiglu.html
- 2025-12-04T15:01:38.819Z
+ 2025-12-08T14:20:17.622Z
https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_disk.html
- 2025-12-04T15:01:38.967Z
+ 2025-12-08T14:20:17.768Z
https://docs.axolotl.ai/docs/api/cli.cloud.base.html
- 2025-12-04T15:01:38.223Z
+ 2025-12-08T14:20:17.031Z
https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_instruct.html
- 2025-12-04T15:01:38.513Z
+ 2025-12-08T14:20:17.319Z
https://docs.axolotl.ai/docs/api/prompt_strategies.pygmalion.html
- 2025-12-04T15:01:38.594Z
+ 2025-12-08T14:20:17.399Z
https://docs.axolotl.ai/docs/api/utils.schemas.peft.html
- 2025-12-04T15:01:39.197Z
+ 2025-12-08T14:20:17.995Z
https://docs.axolotl.ai/docs/api/integrations.liger.args.html
- 2025-12-04T15:01:39.426Z
+ 2025-12-08T14:20:18.222Z
https://docs.axolotl.ai/docs/api/loaders.patch_manager.html
- 2025-12-04T15:01:38.404Z
+ 2025-12-08T14:20:17.211Z
https://docs.axolotl.ai/docs/api/monkeypatch.transformers_fa_utils.html
- 2025-12-04T15:01:38.916Z
+ 2025-12-08T14:20:17.717Z
https://docs.axolotl.ai/docs/api/utils.data.sft.html
- 2025-12-04T15:01:39.108Z
+ 2025-12-08T14:20:17.907Z
https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_cpu.html
- 2025-12-04T15:01:38.934Z
+ 2025-12-08T14:20:17.736Z
https://docs.axolotl.ai/docs/api/monkeypatch.llama_expand_mask.html
- 2025-12-04T15:01:38.848Z
+ 2025-12-08T14:20:17.651Z
https://docs.axolotl.ai/docs/api/loaders.tokenizer.html
- 2025-12-04T15:01:38.383Z
+ 2025-12-08T14:20:17.190Z
https://docs.axolotl.ai/docs/api/utils.tokenization.html
- 2025-12-04T15:01:38.975Z
+ 2025-12-08T14:20:17.776Z
https://docs.axolotl.ai/docs/api/datasets.html
- 2025-12-04T15:01:37.865Z
+ 2025-12-08T14:20:16.678Z
https://docs.axolotl.ai/docs/api/cli.utils.sweeps.html
- 2025-12-04T15:01:38.267Z
+ 2025-12-08T14:20:17.075Z
https://docs.axolotl.ai/docs/api/utils.schedulers.html
- 2025-12-04T15:01:39.057Z
+ 2025-12-08T14:20:17.858Z
https://docs.axolotl.ai/docs/optimizers.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.150Z
https://docs.axolotl.ai/docs/torchao.html
- 2025-12-04T14:44:55.009Z
+ 2025-12-08T14:16:36.150Z
https://docs.axolotl.ai/docs/dataset_preprocessing.html
- 2025-12-04T14:44:55.005Z
+ 2025-12-08T14:16:36.145Z
https://docs.axolotl.ai/docs/faq.html
- 2025-12-04T14:44:55.005Z
+ 2025-12-08T14:16:36.145Z
https://docs.axolotl.ai/docs/ray-integration.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.150Z
https://docs.axolotl.ai/docs/multimodal.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/config-reference.html
- 2025-12-04T15:01:53.952Z
+ 2025-12-08T14:20:32.647Z
https://docs.axolotl.ai/docs/dataset-formats/index.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.144Z
https://docs.axolotl.ai/docs/dataset-formats/inst_tune.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.144Z
https://docs.axolotl.ai/docs/dataset-formats/conversation.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.144Z
https://docs.axolotl.ai/docs/debugging.html
- 2025-12-04T14:44:55.005Z
+ 2025-12-08T14:16:36.145Z
https://docs.axolotl.ai/docs/amd_hpc.html
- 2025-12-04T14:44:55.004Z
+ 2025-12-08T14:16:36.143Z
https://docs.axolotl.ai/docs/lora_optims.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/streaming.html
- 2025-12-04T14:44:55.009Z
+ 2025-12-08T14:16:36.150Z
https://docs.axolotl.ai/docs/multipack.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/qat.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.150Z
https://docs.axolotl.ai/docs/lr_groups.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/getting-started.html
- 2025-12-04T14:44:55.005Z
+ 2025-12-08T14:16:36.145Z
https://docs.axolotl.ai/docs/nccl.html
- 2025-12-04T14:44:55.008Z
+ 2025-12-08T14:16:36.149Z
https://docs.axolotl.ai/docs/telemetry.html
- 2025-12-04T14:44:55.009Z
+ 2025-12-08T14:16:36.150Z
https://docs.axolotl.ai/docs/unsloth.html
- 2025-12-04T14:44:55.009Z
+ 2025-12-08T14:16:36.151Z
https://docs.axolotl.ai/src/axolotl/integrations/LICENSE.html
- 2025-12-04T14:44:55.031Z
+ 2025-12-08T14:16:36.173Z