diff --git a/.nojekyll b/.nojekyll
index 7009b12ee..303544c0a 100644
--- a/.nojekyll
+++ b/.nojekyll
@@ -1 +1 @@
-604c0b1b
\ No newline at end of file
+3b828c6a
\ No newline at end of file
diff --git a/docs/api/utils.samplers.multipack.html b/docs/api/utils.samplers.multipack.html
index b33d68e06..680542123 100644
--- a/docs/api/utils.samplers.multipack.html
+++ b/docs/api/utils.samplers.multipack.html
@@ -920,16 +920,17 @@ Returns the minimum number of batches available on any rank
Pack sequences into bins using parallel processing
Parameters
-
-
-
-
+
+
+
+
@@ -976,32 +977,16 @@ Returns the minimum number of batches available on any rank
If True, use a more conservative packing approach
True
-
-
-
-
-
Returns
-
-
-
-
-
-
-
-
-
Name
-
Type
-
Description
-
-
-
-
-
-
List of bins, where each bin contains indices of sequences assigned to it
+
mp_start_method
+
str | None
+
Multiprocessing start method (‘fork’, ‘spawn’, ‘forkserver’). ‘spawn’ is often safer with Numba/PyTorch. Set to None to use system default.
+
'spawn'
+
Returns:
+List of bins, where each bin contains indices of sequences assigned to it
diff --git a/search.json b/search.json
index e70865c97..395d015a5 100644
--- a/search.json
+++ b/search.json
@@ -644,7 +644,7 @@
"href": "docs/api/utils.samplers.multipack.html",
"title": "utils.samplers.multipack",
"section": "",
- "text": "utils.samplers.multipack\nMultipack Batch Sampler - An efficient batch sampler for packing variable-length sequences\ninto fixed-capacity batches to optimize memory usage and training throughput.\n\n\n\n\n\nName\nDescription\n\n\n\n\nMultipackBatchSampler\nBatch sampler class for efficient packing of variable-length sequences\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler(\n self,\n sampler,\n batch_size,\n batch_max_len,\n lengths,\n packing_efficiency_estimate=1.0,\n drop_last=False,\n num_count_samples=16,\n sequential=False,\n group_size=100000,\n bin_size=200,\n num_processes=None,\n safe_mode=True,\n **kwargs,\n)\nBatch sampler class for efficient packing of variable-length sequences\nThis sampler packs sequences into fixed-capacity bins (batches) to maximize\nGPU memory utilization and training throughput by reducing padding.\nIt supports both parallel packing (using FFD algorithm) and\nsequential packing (preserving original sequence order).\n\n\n\n\n\nName\nDescription\n\n\n\n\nefficiency\nCalculate the packing efficiency (ratio of tokens used to total token slots)\n\n\ngather_efficiency\nGather and synchronize packing efficiency estimates across all distributed ranks\n\n\ngather_len_batches\nGather and synchronize batch counts across all distributed ranks\n\n\ngenerate_batches\nGenerate packed batches for training\n\n\nset_epoch\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.efficiency()\nCalculate the packing efficiency (ratio of tokens used to total token slots)\nHigher is better - 1.0 would mean perfect packing with no wasted space\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_efficiency()\nGather and synchronize packing efficiency estimates across all distributed ranks\nReturns a conservative efficiency estimate based on the measurements\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)\nGather and synchronize batch counts across all distributed ranks\nReturns the minimum number of batches available on any rank\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.generate_batches(set_stats=False)\nGenerate packed batches for training\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nset_stats\n\nWhether to update efficiency statistics\nFalse\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nList of batches, where each batch contains multiple bins,\n\n\n\n\nand each bin contains multiple sequence indices\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.set_epoch(epoch)\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\n\n\n\n\nName\nDescription\n\n\n\n\nallocate_sequentially\nSequential allocator that preserves example order\n\n\nffd_check\nFirst-fit-decreasing bin packing algorithm check\n\n\npack_group\nPack a group of sequences into bins using First-Fit Decreasing algorithm\n\n\npack_parallel\nPack sequences into bins using parallel processing\n\n\n\n\n\nutils.samplers.multipack.allocate_sequentially(\n sequence_lengths,\n rank,\n bin_capacity,\n num_ranks,\n)\nSequential allocator that preserves example order\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nThe lengths of all examples\nrequired\n\n\nrank\nint\nThe current rank (for distributed training)\nrequired\n\n\nbin_capacity\nint\nThe capacity of each bin (maximum sequence length)\nrequired\n\n\nnum_ranks\nint\nNumber of ranks (processes/GPUs)\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\nrank_batches\n\nList of batches for the current rank\n\n\ntotal_tokens_used\n\nNumber of actual example tokens\n\n\ntotal_token_slots\n\nMaximum theoretical number of example tokens (number of bins * bin capacity)\n\n\n\n\n\n\n\nutils.samplers.multipack.ffd_check(sequence_lengths, bin_capacity, num_bins)\nFirst-fit-decreasing bin packing algorithm check\nChecks if sequences with the given lengths could fit in the specified number of bins\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin\nrequired\n\n\nnum_bins\nint\nNumber of bins available\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nTrue if all sequences can be packed, False otherwise\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_group(\n sequence_lengths,\n group_offset,\n bin_capacity,\n max_bins,\n bin_size,\n safe_mode=True,\n)\nPack a group of sequences into bins using First-Fit Decreasing algorithm\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\ngroup_offset\nint\nOffset to apply to indices when returning results\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin\nrequired\n\n\nmax_bins\nint\nMaximum number of bins to use\nrequired\n\n\nbin_size\nint\nMaximum number of sequences per bin\nrequired\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach\nTrue\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nList of bins, where each bin contains indices of sequences assigned to it\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_parallel(\n sequence_lengths,\n bin_capacity,\n group_size,\n bin_size,\n num_processes=None,\n safe_mode=True,\n)\nPack sequences into bins using parallel processing\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin as total number of tokens\nrequired\n\n\ngroup_size\nint\nNumber of sequences to process in each group\nrequired\n\n\nbin_size\nint\nMaximum number of bins to use\nrequired\n\n\nnum_processes\nint | None\nNumber of parallel processes to use\nNone\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach\nTrue\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nList of bins, where each bin contains indices of sequences assigned to it"
+ "text": "utils.samplers.multipack\nMultipack Batch Sampler - An efficient batch sampler for packing variable-length sequences\ninto fixed-capacity batches to optimize memory usage and training throughput.\n\n\n\n\n\nName\nDescription\n\n\n\n\nMultipackBatchSampler\nBatch sampler class for efficient packing of variable-length sequences\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler(\n self,\n sampler,\n batch_size,\n batch_max_len,\n lengths,\n packing_efficiency_estimate=1.0,\n drop_last=False,\n num_count_samples=16,\n sequential=False,\n group_size=100000,\n bin_size=200,\n num_processes=None,\n safe_mode=True,\n **kwargs,\n)\nBatch sampler class for efficient packing of variable-length sequences\nThis sampler packs sequences into fixed-capacity bins (batches) to maximize\nGPU memory utilization and training throughput by reducing padding.\nIt supports both parallel packing (using FFD algorithm) and\nsequential packing (preserving original sequence order).\n\n\n\n\n\nName\nDescription\n\n\n\n\nefficiency\nCalculate the packing efficiency (ratio of tokens used to total token slots)\n\n\ngather_efficiency\nGather and synchronize packing efficiency estimates across all distributed ranks\n\n\ngather_len_batches\nGather and synchronize batch counts across all distributed ranks\n\n\ngenerate_batches\nGenerate packed batches for training\n\n\nset_epoch\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.efficiency()\nCalculate the packing efficiency (ratio of tokens used to total token slots)\nHigher is better - 1.0 would mean perfect packing with no wasted space\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_efficiency()\nGather and synchronize packing efficiency estimates across all distributed ranks\nReturns a conservative efficiency estimate based on the measurements\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)\nGather and synchronize batch counts across all distributed ranks\nReturns the minimum number of batches available on any rank\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.generate_batches(set_stats=False)\nGenerate packed batches for training\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nset_stats\n\nWhether to update efficiency statistics\nFalse\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nList of batches, where each batch contains multiple bins,\n\n\n\n\nand each bin contains multiple sequence indices\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.set_epoch(epoch)\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\n\n\n\n\nName\nDescription\n\n\n\n\nallocate_sequentially\nSequential allocator that preserves example order\n\n\nffd_check\nFirst-fit-decreasing bin packing algorithm check\n\n\npack_group\nPack a group of sequences into bins using First-Fit Decreasing algorithm\n\n\npack_parallel\nPack sequences into bins using parallel processing\n\n\n\n\n\nutils.samplers.multipack.allocate_sequentially(\n sequence_lengths,\n rank,\n bin_capacity,\n num_ranks,\n)\nSequential allocator that preserves example order\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nThe lengths of all examples\nrequired\n\n\nrank\nint\nThe current rank (for distributed training)\nrequired\n\n\nbin_capacity\nint\nThe capacity of each bin (maximum sequence length)\nrequired\n\n\nnum_ranks\nint\nNumber of ranks (processes/GPUs)\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\nrank_batches\n\nList of batches for the current rank\n\n\ntotal_tokens_used\n\nNumber of actual example tokens\n\n\ntotal_token_slots\n\nMaximum theoretical number of example tokens (number of bins * bin capacity)\n\n\n\n\n\n\n\nutils.samplers.multipack.ffd_check(sequence_lengths, bin_capacity, num_bins)\nFirst-fit-decreasing bin packing algorithm check\nChecks if sequences with the given lengths could fit in the specified number of bins\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin\nrequired\n\n\nnum_bins\nint\nNumber of bins available\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nTrue if all sequences can be packed, False otherwise\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_group(\n sequence_lengths,\n group_offset,\n bin_capacity,\n max_bins,\n bin_size,\n safe_mode=True,\n)\nPack a group of sequences into bins using First-Fit Decreasing algorithm\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\ngroup_offset\nint\nOffset to apply to indices when returning results\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin\nrequired\n\n\nmax_bins\nint\nMaximum number of bins to use\nrequired\n\n\nbin_size\nint\nMaximum number of sequences per bin\nrequired\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach\nTrue\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nList of bins, where each bin contains indices of sequences assigned to it\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_parallel(\n sequence_lengths,\n bin_capacity,\n group_size,\n bin_size,\n num_processes=None,\n safe_mode=True,\n mp_start_method='spawn',\n)\nPack sequences into bins using parallel processing\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin as total number of tokens\nrequired\n\n\ngroup_size\nint\nNumber of sequences to process in each group\nrequired\n\n\nbin_size\nint\nMaximum number of bins to use\nrequired\n\n\nnum_processes\nint | None\nNumber of parallel processes to use\nNone\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach\nTrue\n\n\nmp_start_method\nstr | None\nMultiprocessing start method (‘fork’, ‘spawn’, ‘forkserver’). ‘spawn’ is often safer with Numba/PyTorch. Set to None to use system default.\n'spawn'\n\n\n\nReturns:\nList of bins, where each bin contains indices of sequences assigned to it"
},
{
"objectID": "docs/api/utils.samplers.multipack.html#classes",
@@ -658,7 +658,7 @@
"href": "docs/api/utils.samplers.multipack.html#functions",
"title": "utils.samplers.multipack",
"section": "",
- "text": "Name\nDescription\n\n\n\n\nallocate_sequentially\nSequential allocator that preserves example order\n\n\nffd_check\nFirst-fit-decreasing bin packing algorithm check\n\n\npack_group\nPack a group of sequences into bins using First-Fit Decreasing algorithm\n\n\npack_parallel\nPack sequences into bins using parallel processing\n\n\n\n\n\nutils.samplers.multipack.allocate_sequentially(\n sequence_lengths,\n rank,\n bin_capacity,\n num_ranks,\n)\nSequential allocator that preserves example order\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nThe lengths of all examples\nrequired\n\n\nrank\nint\nThe current rank (for distributed training)\nrequired\n\n\nbin_capacity\nint\nThe capacity of each bin (maximum sequence length)\nrequired\n\n\nnum_ranks\nint\nNumber of ranks (processes/GPUs)\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\nrank_batches\n\nList of batches for the current rank\n\n\ntotal_tokens_used\n\nNumber of actual example tokens\n\n\ntotal_token_slots\n\nMaximum theoretical number of example tokens (number of bins * bin capacity)\n\n\n\n\n\n\n\nutils.samplers.multipack.ffd_check(sequence_lengths, bin_capacity, num_bins)\nFirst-fit-decreasing bin packing algorithm check\nChecks if sequences with the given lengths could fit in the specified number of bins\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin\nrequired\n\n\nnum_bins\nint\nNumber of bins available\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nTrue if all sequences can be packed, False otherwise\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_group(\n sequence_lengths,\n group_offset,\n bin_capacity,\n max_bins,\n bin_size,\n safe_mode=True,\n)\nPack a group of sequences into bins using First-Fit Decreasing algorithm\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\ngroup_offset\nint\nOffset to apply to indices when returning results\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin\nrequired\n\n\nmax_bins\nint\nMaximum number of bins to use\nrequired\n\n\nbin_size\nint\nMaximum number of sequences per bin\nrequired\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach\nTrue\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nList of bins, where each bin contains indices of sequences assigned to it\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_parallel(\n sequence_lengths,\n bin_capacity,\n group_size,\n bin_size,\n num_processes=None,\n safe_mode=True,\n)\nPack sequences into bins using parallel processing\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin as total number of tokens\nrequired\n\n\ngroup_size\nint\nNumber of sequences to process in each group\nrequired\n\n\nbin_size\nint\nMaximum number of bins to use\nrequired\n\n\nnum_processes\nint | None\nNumber of parallel processes to use\nNone\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach\nTrue\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nList of bins, where each bin contains indices of sequences assigned to it"
+ "text": "Name\nDescription\n\n\n\n\nallocate_sequentially\nSequential allocator that preserves example order\n\n\nffd_check\nFirst-fit-decreasing bin packing algorithm check\n\n\npack_group\nPack a group of sequences into bins using First-Fit Decreasing algorithm\n\n\npack_parallel\nPack sequences into bins using parallel processing\n\n\n\n\n\nutils.samplers.multipack.allocate_sequentially(\n sequence_lengths,\n rank,\n bin_capacity,\n num_ranks,\n)\nSequential allocator that preserves example order\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nThe lengths of all examples\nrequired\n\n\nrank\nint\nThe current rank (for distributed training)\nrequired\n\n\nbin_capacity\nint\nThe capacity of each bin (maximum sequence length)\nrequired\n\n\nnum_ranks\nint\nNumber of ranks (processes/GPUs)\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\nrank_batches\n\nList of batches for the current rank\n\n\ntotal_tokens_used\n\nNumber of actual example tokens\n\n\ntotal_token_slots\n\nMaximum theoretical number of example tokens (number of bins * bin capacity)\n\n\n\n\n\n\n\nutils.samplers.multipack.ffd_check(sequence_lengths, bin_capacity, num_bins)\nFirst-fit-decreasing bin packing algorithm check\nChecks if sequences with the given lengths could fit in the specified number of bins\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin\nrequired\n\n\nnum_bins\nint\nNumber of bins available\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nTrue if all sequences can be packed, False otherwise\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_group(\n sequence_lengths,\n group_offset,\n bin_capacity,\n max_bins,\n bin_size,\n safe_mode=True,\n)\nPack a group of sequences into bins using First-Fit Decreasing algorithm\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\ngroup_offset\nint\nOffset to apply to indices when returning results\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin\nrequired\n\n\nmax_bins\nint\nMaximum number of bins to use\nrequired\n\n\nbin_size\nint\nMaximum number of sequences per bin\nrequired\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach\nTrue\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\n\nList of bins, where each bin contains indices of sequences assigned to it\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_parallel(\n sequence_lengths,\n bin_capacity,\n group_size,\n bin_size,\n num_processes=None,\n safe_mode=True,\n mp_start_method='spawn',\n)\nPack sequences into bins using parallel processing\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin as total number of tokens\nrequired\n\n\ngroup_size\nint\nNumber of sequences to process in each group\nrequired\n\n\nbin_size\nint\nMaximum number of bins to use\nrequired\n\n\nnum_processes\nint | None\nNumber of parallel processes to use\nNone\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach\nTrue\n\n\nmp_start_method\nstr | None\nMultiprocessing start method (‘fork’, ‘spawn’, ‘forkserver’). ‘spawn’ is often safer with Numba/PyTorch. Set to None to use system default.\n'spawn'\n\n\n\nReturns:\nList of bins, where each bin contains indices of sequences assigned to it"
},
{
"objectID": "docs/api/utils.callbacks.comet_.html",
diff --git a/sitemap.xml b/sitemap.xml
index a6b2031f1..d4bf416cc 100644
--- a/sitemap.xml
+++ b/sitemap.xml
@@ -2,682 +2,682 @@
https://docs.axolotl.ai/FAQS.html
- 2025-05-10T00:29:11.331Z
+ 2025-05-12T14:51:27.715Zhttps://docs.axolotl.ai/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html
- 2025-05-10T00:29:11.353Z
+ 2025-05-12T14:51:27.737Zhttps://docs.axolotl.ai/TODO.html
- 2025-05-10T00:29:11.331Z
+ 2025-05-12T14:51:27.715Zhttps://docs.axolotl.ai/docs/multimodal.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/multi-node.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/lora_optims.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.stepwise_supervised.html
- 2025-05-10T00:29:40.213Z
+ 2025-05-12T14:51:55.897Zhttps://docs.axolotl.ai/docs/api/core.datasets.transforms.chat_builder.html
- 2025-05-10T00:29:39.897Z
+ 2025-05-12T14:51:55.586Zhttps://docs.axolotl.ai/docs/api/common.architectures.html
- 2025-05-10T00:29:40.898Z
+ 2025-05-12T14:51:56.575Zhttps://docs.axolotl.ai/docs/api/utils.schemas.trl.html
- 2025-05-10T00:29:40.736Z
+ 2025-05-12T14:51:56.417Zhttps://docs.axolotl.ai/docs/api/utils.lora_embeddings.html
- 2025-05-10T00:29:40.590Z
+ 2025-05-12T14:51:56.272Zhttps://docs.axolotl.ai/docs/api/core.chat.messages.html
- 2025-05-10T00:29:39.880Z
+ 2025-05-12T14:51:55.568Zhttps://docs.axolotl.ai/docs/api/monkeypatch.mistral_attn_hijack_flash.html
- 2025-05-10T00:29:40.461Z
+ 2025-05-12T14:51:56.144Zhttps://docs.axolotl.ai/docs/api/integrations.lm_eval.args.html
- 2025-05-10T00:29:40.893Z
+ 2025-05-12T14:51:56.571Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.chat_template.html
- 2025-05-10T00:29:40.148Z
+ 2025-05-12T14:51:55.832Zhttps://docs.axolotl.ai/docs/api/integrations.kd.trainer.html
- 2025-05-10T00:29:40.886Z
+ 2025-05-12T14:51:56.564Zhttps://docs.axolotl.ai/docs/api/monkeypatch.transformers_fa_utils.html
- 2025-05-10T00:29:40.524Z
+ 2025-05-12T14:51:56.205Zhttps://docs.axolotl.ai/docs/api/index.html
- 2025-05-10T00:29:39.612Z
+ 2025-05-12T14:51:55.303Zhttps://docs.axolotl.ai/docs/api/core.trainers.dpo.trainer.html
- 2025-05-10T00:29:40.125Z
+ 2025-05-12T14:51:55.809Zhttps://docs.axolotl.ai/docs/api/evaluate.html
- 2025-05-10T00:29:39.683Z
+ 2025-05-12T14:51:55.374Zhttps://docs.axolotl.ai/docs/api/utils.schemas.utils.html
- 2025-05-10T00:29:40.766Z
+ 2025-05-12T14:51:56.447Zhttps://docs.axolotl.ai/docs/api/core.trainers.base.html
- 2025-05-10T00:29:40.098Z
+ 2025-05-12T14:51:55.782Zhttps://docs.axolotl.ai/docs/api/kernels.lora.html
- 2025-05-10T00:29:40.387Z
+ 2025-05-12T14:51:56.070Zhttps://docs.axolotl.ai/docs/api/common.const.html
- 2025-05-10T00:29:40.899Z
+ 2025-05-12T14:51:56.577Zhttps://docs.axolotl.ai/docs/api/utils.schemas.training.html
- 2025-05-10T00:29:40.706Z
+ 2025-05-12T14:51:56.388Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_chat.html
- 2025-05-10T00:29:40.162Z
+ 2025-05-12T14:51:55.846Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.lisa.html
- 2025-05-10T00:29:41.000Z
+ 2025-05-12T14:51:56.676Zhttps://docs.axolotl.ai/docs/api/monkeypatch.llama_patch_multipack.html
- 2025-05-10T00:29:40.508Z
+ 2025-05-12T14:51:56.190Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.bradley_terry.llama3.html
- 2025-05-10T00:29:40.304Z
+ 2025-05-12T14:51:55.987Zhttps://docs.axolotl.ai/docs/api/integrations.spectrum.args.html
- 2025-05-10T00:29:40.896Z
+ 2025-05-12T14:51:56.574Zhttps://docs.axolotl.ai/docs/api/utils.trainer.html
- 2025-05-10T00:29:40.624Z
+ 2025-05-12T14:51:56.306Zhttps://docs.axolotl.ai/docs/api/monkeypatch.mixtral.html
- 2025-05-10T00:29:40.535Z
+ 2025-05-12T14:51:56.217Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.orcamini.html
- 2025-05-10T00:29:40.224Z
+ 2025-05-12T14:51:55.908Zhttps://docs.axolotl.ai/docs/api/train.html
- 2025-05-10T00:29:39.673Z
+ 2025-05-12T14:51:55.364Zhttps://docs.axolotl.ai/docs/api/convert.html
- 2025-05-10T00:29:39.704Z
+ 2025-05-12T14:51:55.395Zhttps://docs.axolotl.ai/docs/api/utils.samplers.multipack.html
- 2025-05-10T00:29:40.988Z
+ 2025-05-12T14:51:56.665Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.comet_.html
- 2025-05-10T00:29:41.008Z
+ 2025-05-12T14:51:56.684Zhttps://docs.axolotl.ai/docs/api/utils.lora.html
- 2025-05-10T00:29:40.586Z
+ 2025-05-12T14:51:56.268Zhttps://docs.axolotl.ai/docs/api/utils.schemas.peft.html
- 2025-05-10T00:29:40.733Z
+ 2025-05-12T14:51:56.414Zhttps://docs.axolotl.ai/docs/api/utils.schemas.multimodal.html
- 2025-05-10T00:29:40.741Z
+ 2025-05-12T14:51:56.422Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.user_defined.html
- 2025-05-10T00:29:40.261Z
+ 2025-05-12T14:51:55.944Zhttps://docs.axolotl.ai/docs/api/monkeypatch.multipack.html
- 2025-05-10T00:29:40.463Z
+ 2025-05-12T14:51:56.145Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.llama3.html
- 2025-05-10T00:29:40.247Z
+ 2025-05-12T14:51:55.931Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.perplexity.html
- 2025-05-10T00:29:40.995Z
+ 2025-05-12T14:51:56.671Zhttps://docs.axolotl.ai/docs/api/logging_config.html
- 2025-05-10T00:29:39.753Z
+ 2025-05-12T14:51:55.443Zhttps://docs.axolotl.ai/docs/api/integrations.cut_cross_entropy.args.html
- 2025-05-10T00:29:40.877Z
+ 2025-05-12T14:51:56.555Zhttps://docs.axolotl.ai/docs/api/utils.schemas.integrations.html
- 2025-05-10T00:29:40.754Z
+ 2025-05-12T14:51:56.434Zhttps://docs.axolotl.ai/docs/api/monkeypatch.data.batch_dataset_fetcher.html
- 2025-05-10T00:29:40.533Z
+ 2025-05-12T14:51:56.215Zhttps://docs.axolotl.ai/docs/api/cli.merge_sharded_fsdp_weights.html
- 2025-05-10T00:29:40.023Z
+ 2025-05-12T14:51:55.708Zhttps://docs.axolotl.ai/docs/api/cli.preprocess.html
- 2025-05-10T00:29:40.031Z
+ 2025-05-12T14:51:55.717Zhttps://docs.axolotl.ai/docs/api/prompt_tokenizers.html
- 2025-05-10T00:29:39.747Z
+ 2025-05-12T14:51:55.438Zhttps://docs.axolotl.ai/docs/api/utils.collators.core.html
- 2025-05-10T00:29:40.919Z
+ 2025-05-12T14:51:56.597Zhttps://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_xformers.html
- 2025-05-10T00:29:40.446Z
+ 2025-05-12T14:51:56.129Zhttps://docs.axolotl.ai/docs/api/cli.merge_lora.html
- 2025-05-10T00:29:40.011Z
+ 2025-05-12T14:51:55.697Zhttps://docs.axolotl.ai/docs/api/core.chat.format.chatml.html
- 2025-05-10T00:29:39.881Z
+ 2025-05-12T14:51:55.570Zhttps://docs.axolotl.ai/docs/api/core.trainer_builder.html
- 2025-05-10T00:29:39.768Z
+ 2025-05-12T14:51:55.458Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.kto.chatml.html
- 2025-05-10T00:29:40.278Z
+ 2025-05-12T14:51:55.962Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.completion.html
- 2025-05-10T00:29:40.203Z
+ 2025-05-12T14:51:55.887Zhttps://docs.axolotl.ai/docs/api/integrations.liger.args.html
- 2025-05-10T00:29:40.890Z
+ 2025-05-12T14:51:56.567Zhttps://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_flash.html
- 2025-05-10T00:29:40.444Z
+ 2025-05-12T14:51:56.128Zhttps://docs.axolotl.ai/docs/api/integrations.grokfast.optimizer.html
- 2025-05-10T00:29:40.878Z
+ 2025-05-12T14:51:56.556Zhttps://docs.axolotl.ai/docs/api/utils.chat_templates.html
- 2025-05-10T00:29:40.581Z
+ 2025-05-12T14:51:56.263Zhttps://docs.axolotl.ai/docs/api/utils.model_shard_quant.html
- 2025-05-10T00:29:40.595Z
+ 2025-05-12T14:51:56.277Zhttps://docs.axolotl.ai/docs/api/monkeypatch.trainer_fsdp_optim.html
- 2025-05-10T00:29:40.517Z
+ 2025-05-12T14:51:56.199Zhttps://docs.axolotl.ai/docs/api/monkeypatch.unsloth_.html
- 2025-05-10T00:29:40.525Z
+ 2025-05-12T14:51:56.207Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.zephyr.html
- 2025-05-10T00:29:40.259Z
+ 2025-05-12T14:51:55.943Zhttps://docs.axolotl.ai/docs/api/kernels.quantize.html
- 2025-05-10T00:29:40.415Z
+ 2025-05-12T14:51:56.100Zhttps://docs.axolotl.ai/docs/api/cli.cloud.base.html
- 2025-05-10T00:29:40.077Z
+ 2025-05-12T14:51:55.762Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_instruct.html
- 2025-05-10T00:29:40.163Z
+ 2025-05-12T14:51:55.847Zhttps://docs.axolotl.ai/docs/api/core.chat.format.llama3x.html
- 2025-05-10T00:29:39.883Z
+ 2025-05-12T14:51:55.571Zhttps://docs.axolotl.ai/docs/reward_modelling.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/config.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.716Zhttps://docs.axolotl.ai/docs/dataset_loading.html
- 2025-05-10T00:29:11.333Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/multi-gpu.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/installation.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/docker.html
- 2025-05-10T00:29:11.333Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/multipack.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/debugging.html
- 2025-05-10T00:29:11.333Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/inference.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/dataset-formats/conversation.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/dataset-formats/stepwise_supervised.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/dataset-formats/inst_tune.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/custom_integrations.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.716Zhttps://docs.axolotl.ai/docs/mac.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/rlhf.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/examples/colab-notebooks/colab-axolotl-example.html
- 2025-05-10T00:29:11.337Z
+ 2025-05-12T14:51:27.721Zhttps://docs.axolotl.ai/docs/getting-started.html
- 2025-05-10T00:29:11.333Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/nccl.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/dataset-formats/tokenized.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/dataset-formats/index.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/dataset-formats/template_free.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/dataset-formats/pretraining.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/batch_vs_grad.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.716Zhttps://docs.axolotl.ai/docs/lr_groups.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/fsdp_qlora.html
- 2025-05-10T00:29:11.333Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/sequence_parallelism.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/amd_hpc.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.716Zhttps://docs.axolotl.ai/docs/ray-integration.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/dataset_preprocessing.html
- 2025-05-10T00:29:11.333Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/faq.html
- 2025-05-10T00:29:11.333Z
+ 2025-05-12T14:51:27.717Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chatml.html
- 2025-05-10T00:29:40.258Z
+ 2025-05-12T14:51:55.941Zhttps://docs.axolotl.ai/docs/api/kernels.geglu.html
- 2025-05-10T00:29:40.398Z
+ 2025-05-12T14:51:56.080Zhttps://docs.axolotl.ai/docs/api/core.chat.format.shared.html
- 2025-05-10T00:29:39.884Z
+ 2025-05-12T14:51:55.573Zhttps://docs.axolotl.ai/docs/api/cli.vllm_serve.html
- 2025-05-10T00:29:40.074Z
+ 2025-05-12T14:51:55.758Zhttps://docs.axolotl.ai/docs/api/core.training_args.html
- 2025-05-10T00:29:39.857Z
+ 2025-05-12T14:51:55.546Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.passthrough.html
- 2025-05-10T00:29:40.262Z
+ 2025-05-12T14:51:55.946Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.pygmalion.html
- 2025-05-10T00:29:40.231Z
+ 2025-05-12T14:51:55.915Zhttps://docs.axolotl.ai/docs/api/cli.checks.html
- 2025-05-10T00:29:39.970Z
+ 2025-05-12T14:51:55.656Zhttps://docs.axolotl.ai/docs/api/monkeypatch.attention.mllama.html
- 2025-05-10T00:29:40.532Z
+ 2025-05-12T14:51:56.214Zhttps://docs.axolotl.ai/docs/api/cli.main.html
- 2025-05-10T00:29:39.929Z
+ 2025-05-12T14:51:55.617Zhttps://docs.axolotl.ai/docs/api/cli.utils.html
- 2025-05-10T00:29:40.069Z
+ 2025-05-12T14:51:55.754Zhttps://docs.axolotl.ai/docs/api/utils.tokenization.html
- 2025-05-10T00:29:40.571Z
+ 2025-05-12T14:51:56.254Zhttps://docs.axolotl.ai/docs/api/monkeypatch.btlm_attn_hijack_flash.html
- 2025-05-10T00:29:40.507Z
+ 2025-05-12T14:51:56.189Zhttps://docs.axolotl.ai/docs/api/utils.models.html
- 2025-05-10T00:29:40.564Z
+ 2025-05-12T14:51:56.246Zhttps://docs.axolotl.ai/docs/api/core.datasets.chat.html
- 2025-05-10T00:29:39.889Z
+ 2025-05-12T14:51:55.578Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.llama2_chat.html
- 2025-05-10T00:29:40.197Z
+ 2025-05-12T14:51:55.881Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.messages.chat.html
- 2025-05-10T00:29:40.235Z
+ 2025-05-12T14:51:55.919Zhttps://docs.axolotl.ai/docs/api/utils.dict.html
- 2025-05-10T00:29:40.672Z
+ 2025-05-12T14:51:56.353Zhttps://docs.axolotl.ai/docs/api/cli.config.html
- 2025-05-10T00:29:39.988Z
+ 2025-05-12T14:51:55.675Zhttps://docs.axolotl.ai/docs/api/core.trainers.trl.html
- 2025-05-10T00:29:40.115Z
+ 2025-05-12T14:51:55.799Zhttps://docs.axolotl.ai/docs/api/utils.data.pretraining.html
- 2025-05-10T00:29:40.681Z
+ 2025-05-12T14:51:56.362Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.kto.user_defined.html
- 2025-05-10T00:29:40.280Z
+ 2025-05-12T14:51:55.963Zhttps://docs.axolotl.ai/docs/api/utils.schedulers.html
- 2025-05-10T00:29:40.649Z
+ 2025-05-12T14:51:56.330Zhttps://docs.axolotl.ai/docs/api/utils.freeze.html
- 2025-05-10T00:29:40.607Z
+ 2025-05-12T14:51:56.289Zhttps://docs.axolotl.ai/docs/api/kernels.swiglu.html
- 2025-05-10T00:29:40.408Z
+ 2025-05-12T14:51:56.091Zhttps://docs.axolotl.ai/docs/api/integrations.base.html
- 2025-05-10T00:29:40.874Z
+ 2025-05-12T14:51:56.552Zhttps://docs.axolotl.ai/docs/api/cli.cloud.modal_.html
- 2025-05-10T00:29:40.084Z
+ 2025-05-12T14:51:55.768Zhttps://docs.axolotl.ai/docs/api/core.trainers.grpo.trainer.html
- 2025-05-10T00:29:40.128Z
+ 2025-05-12T14:51:55.812Zhttps://docs.axolotl.ai/docs/api/cli.sweeps.html
- 2025-05-10T00:29:40.037Z
+ 2025-05-12T14:51:55.722Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.input_output.html
- 2025-05-10T00:29:40.209Z
+ 2025-05-12T14:51:55.893Zhttps://docs.axolotl.ai/docs/api/utils.gradient_checkpointing.unsloth.html
- 2025-05-10T00:29:40.686Z
+ 2025-05-12T14:51:56.367Zhttps://docs.axolotl.ai/docs/api/cli.args.html
- 2025-05-10T00:29:39.963Z
+ 2025-05-12T14:51:55.650Zhttps://docs.axolotl.ai/docs/api/cli.inference.html
- 2025-05-10T00:29:40.002Z
+ 2025-05-12T14:51:55.688Zhttps://docs.axolotl.ai/docs/api/monkeypatch.lora_kernels.html
- 2025-05-10T00:29:40.497Z
+ 2025-05-12T14:51:56.179Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.metharme.html
- 2025-05-10T00:29:40.220Z
+ 2025-05-12T14:51:55.904Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chat_template.html
- 2025-05-10T00:29:40.237Z
+ 2025-05-12T14:51:55.921Zhttps://docs.axolotl.ai/docs/api/utils.schemas.config.html
- 2025-05-10T00:29:40.694Z
+ 2025-05-12T14:51:56.376Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.orpo.chat_template.html
- 2025-05-10T00:29:40.300Z
+ 2025-05-12T14:51:55.984Zhttps://docs.axolotl.ai/docs/api/monkeypatch.utils.html
- 2025-05-10T00:29:40.505Z
+ 2025-05-12T14:51:56.187Zhttps://docs.axolotl.ai/docs/api/monkeypatch.relora.html
- 2025-05-10T00:29:40.470Z
+ 2025-05-12T14:51:56.152Zhttps://docs.axolotl.ai/docs/api/utils.optimizers.adopt.html
- 2025-05-10T00:29:40.679Z
+ 2025-05-12T14:51:56.361Zhttps://docs.axolotl.ai/docs/api/cli.train.html
- 2025-05-10T00:29:39.938Z
+ 2025-05-12T14:51:55.625Zhttps://docs.axolotl.ai/docs/api/utils.collators.mamba.html
- 2025-05-10T00:29:40.942Z
+ 2025-05-12T14:51:56.619Zhttps://docs.axolotl.ai/docs/api/monkeypatch.stablelm_attn_hijack_flash.html
- 2025-05-10T00:29:40.514Z
+ 2025-05-12T14:51:56.196Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.base.html
- 2025-05-10T00:29:40.130Z
+ 2025-05-12T14:51:55.814Zhttps://docs.axolotl.ai/docs/api/datasets.html
- 2025-05-10T00:29:39.691Z
+ 2025-05-12T14:51:55.382Zhttps://docs.axolotl.ai/docs/api/utils.schemas.enums.html
- 2025-05-10T00:29:40.761Z
+ 2025-05-12T14:51:56.441Zhttps://docs.axolotl.ai/docs/api/models.mamba.modeling_mamba.html
- 2025-05-10T00:29:40.917Z
+ 2025-05-12T14:51:56.595Zhttps://docs.axolotl.ai/docs/api/monkeypatch.llama_expand_mask.html
- 2025-05-10T00:29:40.472Z
+ 2025-05-12T14:51:56.154Zhttps://docs.axolotl.ai/docs/api/cli.evaluate.html
- 2025-05-10T00:29:39.946Z
+ 2025-05-12T14:51:55.633Zhttps://docs.axolotl.ai/docs/api/utils.data.sft.html
- 2025-05-10T00:29:40.682Z
+ 2025-05-12T14:51:56.364Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.user_defined.html
- 2025-05-10T00:29:40.184Z
+ 2025-05-12T14:51:55.868Zhttps://docs.axolotl.ai/docs/api/utils.schemas.datasets.html
- 2025-05-10T00:29:40.724Z
+ 2025-05-12T14:51:56.405Zhttps://docs.axolotl.ai/docs/api/common.datasets.html
- 2025-05-10T00:29:40.916Z
+ 2025-05-12T14:51:56.594Zhttps://docs.axolotl.ai/docs/api/kernels.utils.html
- 2025-05-10T00:29:40.417Z
+ 2025-05-12T14:51:56.102Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.kto.llama3.html
- 2025-05-10T00:29:40.270Z
+ 2025-05-12T14:51:55.954Zhttps://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_w_system.html
- 2025-05-10T00:29:40.175Z
+ 2025-05-12T14:51:55.860Zhttps://docs.axolotl.ai/docs/api/utils.collators.mm_chat.html
- 2025-05-10T00:29:40.947Z
+ 2025-05-12T14:51:56.624Zhttps://docs.axolotl.ai/docs/api/utils.bench.html
- 2025-05-10T00:29:40.599Z
+ 2025-05-12T14:51:56.281Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.mlflow_.html
- 2025-05-10T00:29:41.004Z
+ 2025-05-12T14:51:56.680Zhttps://docs.axolotl.ai/docs/api/utils.callbacks.profiler.html
- 2025-05-10T00:29:40.999Z
+ 2025-05-12T14:51:56.675Zhttps://docs.axolotl.ai/docs/api/utils.collators.batching.html
- 2025-05-10T00:29:40.938Z
+ 2025-05-12T14:51:56.616Zhttps://docs.axolotl.ai/docs/api/utils.distributed.html
- 2025-05-10T00:29:40.668Z
+ 2025-05-12T14:51:56.350Zhttps://docs.axolotl.ai/docs/api/utils.schemas.model.html
- 2025-05-10T00:29:40.701Z
+ 2025-05-12T14:51:56.382Zhttps://docs.axolotl.ai/docs/torchao.html
- 2025-05-10T00:29:11.337Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/input_output.html
- 2025-05-10T00:29:11.336Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/docs/cli.html
- 2025-05-10T00:29:11.332Z
+ 2025-05-12T14:51:27.716Zhttps://docs.axolotl.ai/docs/unsloth.html
- 2025-05-10T00:29:11.337Z
+ 2025-05-12T14:51:27.720Zhttps://docs.axolotl.ai/src/axolotl/integrations/LICENSE.html
- 2025-05-10T00:29:11.353Z
+ 2025-05-12T14:51:27.737Zhttps://docs.axolotl.ai/index.html
- 2025-05-10T00:29:11.349Z
+ 2025-05-12T14:51:27.733Z