diff --git a/.nojekyll b/.nojekyll index 94539e1f7..4c2dd3dae 100644 --- a/.nojekyll +++ b/.nojekyll @@ -1 +1 @@ -3cf13af5 \ No newline at end of file +5b444df2 \ No newline at end of file diff --git a/docs/api/utils.samplers.multipack.html b/docs/api/utils.samplers.multipack.html index 7e78081b2..96ef76381 100644 --- a/docs/api/utils.samplers.multipack.html +++ b/docs/api/utils.samplers.multipack.html @@ -562,12 +562,12 @@ into fixed-capacity batches to optimize memory usage and training throughput.

batch_size, batch_max_len, lengths, - packing_efficiency_estimate=1.0, - drop_last=True, - num_count_samples=4, - sequential=False, - group_size=100000, - bin_size=200, + bin_size, + packing_efficiency_estimate=1.0, + drop_last=True, + num_count_samples=4, + sequential=False, + group_size=100000, num_processes=None, safe_mode=True, mp_start_method='fork', diff --git a/search.json b/search.json index ea42006f2..6eb44508f 100644 --- a/search.json +++ b/search.json @@ -2658,14 +2658,14 @@ "href": "docs/api/utils.samplers.multipack.html", "title": "utils.samplers.multipack", "section": "", - "text": "utils.samplers.multipack\nMultipack Batch Sampler - An efficient batch sampler for packing variable-length sequences\ninto fixed-capacity batches to optimize memory usage and training throughput.\n\n\n\n\n\nName\nDescription\n\n\n\n\nMultipackBatchSampler\nBatch sampler class for efficient packing of variable-length sequences\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler(\n sampler,\n batch_size,\n batch_max_len,\n lengths,\n packing_efficiency_estimate=1.0,\n drop_last=True,\n num_count_samples=4,\n sequential=False,\n group_size=100000,\n bin_size=200,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n **kwargs,\n)\nBatch sampler class for efficient packing of variable-length sequences\nThis sampler packs sequences into fixed-capacity bins (batches) to maximize\nGPU memory utilization and training throughput by reducing padding.\nIt supports both parallel packing (using FFD algorithm) and\nsequential packing (preserving original sequence order).\n\n\n\n\n\nName\nDescription\n\n\n\n\nefficiency\nCalculate the packing efficiency (ratio of tokens used to total token slots).\n\n\ngather_efficiency\nGather and synchronize packing efficiency estimates across all distributed\n\n\ngather_len_batches\nGather and synchronize batch counts across all distributed ranks. Returns\n\n\ngenerate_batches\nGenerate packed batches for training.\n\n\nset_epoch\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.efficiency()\nCalculate the packing efficiency (ratio of tokens used to total token slots).\nHigher is better - 1.0 would mean perfect packing with no wasted space.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_efficiency()\nGather and synchronize packing efficiency estimates across all distributed\nranks.\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nfloat\nA conservative efficiency estimate based on the measurements.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)\nGather and synchronize batch counts across all distributed ranks. Returns\nthe minimum number of batches available on any rank.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.generate_batches(set_stats=False)\nGenerate packed batches for training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nset_stats\nbool\nWhether to update efficiency statistics.\nFalse\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[list[int]]]\nList of batches, where each batch contains multiple bins, and each bin contains multiple sequence indices.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.set_epoch(epoch)\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\n\n\n\n\nName\nDescription\n\n\n\n\nallocate_sequentially\nSequential allocator that preserves example order.\n\n\nffd_check\nFirst-fit-decreasing bin packing algorithm check.\n\n\npack_group\nPack a group of sequences into bins using First-Fit Decreasing algorithm.\n\n\npack_parallel\nPack sequences into bins using parallel processing.\n\n\n\n\n\nutils.samplers.multipack.allocate_sequentially(\n sequence_lengths,\n rank,\n bin_capacity,\n num_ranks,\n)\nSequential allocator that preserves example order.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nThe lengths of all examples.\nrequired\n\n\nrank\nint\nThe current rank (for distributed training).\nrequired\n\n\nbin_capacity\nint\nThe capacity of each bin (maximum sequence length).\nrequired\n\n\nnum_ranks\nint\nNumber of ranks (processes / GPUs).\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\nrank_batches\nlist[list[int]]\nList of batches for the current rank.\n\n\ntotal_tokens_used\nint\nNumber of actual example tokens.\n\n\ntotal_token_slots\nint\nMaximum theoretical number of example tokens (number of bins * bin capacity).\n\n\n\n\n\n\n\nutils.samplers.multipack.ffd_check(sequence_lengths, bin_capacity, num_bins)\nFirst-fit-decreasing bin packing algorithm check.\nChecks if sequences with the given lengths could fit in the specified number of\nbins.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin.\nrequired\n\n\nnum_bins\nint\nNumber of bins available.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nbool\nTrue if all sequences can be packed, False otherwise.\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_group(\n sequence_lengths,\n group_offset,\n bin_capacity,\n max_bins,\n bin_size,\n safe_mode=True,\n)\nPack a group of sequences into bins using First-Fit Decreasing algorithm.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\ngroup_offset\nint\nOffset to apply to indices when returning results.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin.\nrequired\n\n\nmax_bins\nint\nMaximum number of bins to use.\nrequired\n\n\nbin_size\nint\nMaximum number of sequences per bin.\nrequired\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach.\nTrue\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[int]]\nList of bins, where each bin contains indices of sequences assigned to it.\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_parallel(\n sequence_lengths,\n bin_capacity,\n group_size,\n bin_size,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n)\nPack sequences into bins using parallel processing.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin as total number of tokens.\nrequired\n\n\ngroup_size\nint\nNumber of sequences to process in each group.\nrequired\n\n\nbin_size\nint\nMaximum number of bins to use.\nrequired\n\n\nnum_processes\nint | None\nNumber of parallel processes to use.\nNone\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach.\nTrue\n\n\nmp_start_method\nstr | None\nMultiprocessing start method (‘fork’, ‘spawn’, ‘forkserver’). ‘spawn’ is often safer with Numba/PyTorch. Set to None to use system default.\n'fork'\n\n\n\nReturns:\nList of bins, where each bin contains indices of sequences assigned to it." + "text": "utils.samplers.multipack\nMultipack Batch Sampler - An efficient batch sampler for packing variable-length sequences\ninto fixed-capacity batches to optimize memory usage and training throughput.\n\n\n\n\n\nName\nDescription\n\n\n\n\nMultipackBatchSampler\nBatch sampler class for efficient packing of variable-length sequences\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler(\n sampler,\n batch_size,\n batch_max_len,\n lengths,\n bin_size,\n packing_efficiency_estimate=1.0,\n drop_last=True,\n num_count_samples=4,\n sequential=False,\n group_size=100000,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n **kwargs,\n)\nBatch sampler class for efficient packing of variable-length sequences\nThis sampler packs sequences into fixed-capacity bins (batches) to maximize\nGPU memory utilization and training throughput by reducing padding.\nIt supports both parallel packing (using FFD algorithm) and\nsequential packing (preserving original sequence order).\n\n\n\n\n\nName\nDescription\n\n\n\n\nefficiency\nCalculate the packing efficiency (ratio of tokens used to total token slots).\n\n\ngather_efficiency\nGather and synchronize packing efficiency estimates across all distributed\n\n\ngather_len_batches\nGather and synchronize batch counts across all distributed ranks. Returns\n\n\ngenerate_batches\nGenerate packed batches for training.\n\n\nset_epoch\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.efficiency()\nCalculate the packing efficiency (ratio of tokens used to total token slots).\nHigher is better - 1.0 would mean perfect packing with no wasted space.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_efficiency()\nGather and synchronize packing efficiency estimates across all distributed\nranks.\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nfloat\nA conservative efficiency estimate based on the measurements.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)\nGather and synchronize batch counts across all distributed ranks. Returns\nthe minimum number of batches available on any rank.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.generate_batches(set_stats=False)\nGenerate packed batches for training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nset_stats\nbool\nWhether to update efficiency statistics.\nFalse\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[list[int]]]\nList of batches, where each batch contains multiple bins, and each bin contains multiple sequence indices.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.set_epoch(epoch)\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\n\n\n\n\nName\nDescription\n\n\n\n\nallocate_sequentially\nSequential allocator that preserves example order.\n\n\nffd_check\nFirst-fit-decreasing bin packing algorithm check.\n\n\npack_group\nPack a group of sequences into bins using First-Fit Decreasing algorithm.\n\n\npack_parallel\nPack sequences into bins using parallel processing.\n\n\n\n\n\nutils.samplers.multipack.allocate_sequentially(\n sequence_lengths,\n rank,\n bin_capacity,\n num_ranks,\n)\nSequential allocator that preserves example order.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nThe lengths of all examples.\nrequired\n\n\nrank\nint\nThe current rank (for distributed training).\nrequired\n\n\nbin_capacity\nint\nThe capacity of each bin (maximum sequence length).\nrequired\n\n\nnum_ranks\nint\nNumber of ranks (processes / GPUs).\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\nrank_batches\nlist[list[int]]\nList of batches for the current rank.\n\n\ntotal_tokens_used\nint\nNumber of actual example tokens.\n\n\ntotal_token_slots\nint\nMaximum theoretical number of example tokens (number of bins * bin capacity).\n\n\n\n\n\n\n\nutils.samplers.multipack.ffd_check(sequence_lengths, bin_capacity, num_bins)\nFirst-fit-decreasing bin packing algorithm check.\nChecks if sequences with the given lengths could fit in the specified number of\nbins.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin.\nrequired\n\n\nnum_bins\nint\nNumber of bins available.\nrequired\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nbool\nTrue if all sequences can be packed, False otherwise.\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_group(\n sequence_lengths,\n group_offset,\n bin_capacity,\n max_bins,\n bin_size,\n safe_mode=True,\n)\nPack a group of sequences into bins using First-Fit Decreasing algorithm.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\ngroup_offset\nint\nOffset to apply to indices when returning results.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin.\nrequired\n\n\nmax_bins\nint\nMaximum number of bins to use.\nrequired\n\n\nbin_size\nint\nMaximum number of sequences per bin.\nrequired\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach.\nTrue\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[int]]\nList of bins, where each bin contains indices of sequences assigned to it.\n\n\n\n\n\n\n\nutils.samplers.multipack.pack_parallel(\n sequence_lengths,\n bin_capacity,\n group_size,\n bin_size,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n)\nPack sequences into bins using parallel processing.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nsequence_lengths\nnp.ndarray\nArray of sequence lengths.\nrequired\n\n\nbin_capacity\nint\nMaximum capacity of each bin as total number of tokens.\nrequired\n\n\ngroup_size\nint\nNumber of sequences to process in each group.\nrequired\n\n\nbin_size\nint\nMaximum number of bins to use.\nrequired\n\n\nnum_processes\nint | None\nNumber of parallel processes to use.\nNone\n\n\nsafe_mode\nbool\nIf True, use a more conservative packing approach.\nTrue\n\n\nmp_start_method\nstr | None\nMultiprocessing start method (‘fork’, ‘spawn’, ‘forkserver’). ‘spawn’ is often safer with Numba/PyTorch. Set to None to use system default.\n'fork'\n\n\n\nReturns:\nList of bins, where each bin contains indices of sequences assigned to it." }, { "objectID": "docs/api/utils.samplers.multipack.html#classes", "href": "docs/api/utils.samplers.multipack.html#classes", "title": "utils.samplers.multipack", "section": "", - "text": "Name\nDescription\n\n\n\n\nMultipackBatchSampler\nBatch sampler class for efficient packing of variable-length sequences\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler(\n sampler,\n batch_size,\n batch_max_len,\n lengths,\n packing_efficiency_estimate=1.0,\n drop_last=True,\n num_count_samples=4,\n sequential=False,\n group_size=100000,\n bin_size=200,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n **kwargs,\n)\nBatch sampler class for efficient packing of variable-length sequences\nThis sampler packs sequences into fixed-capacity bins (batches) to maximize\nGPU memory utilization and training throughput by reducing padding.\nIt supports both parallel packing (using FFD algorithm) and\nsequential packing (preserving original sequence order).\n\n\n\n\n\nName\nDescription\n\n\n\n\nefficiency\nCalculate the packing efficiency (ratio of tokens used to total token slots).\n\n\ngather_efficiency\nGather and synchronize packing efficiency estimates across all distributed\n\n\ngather_len_batches\nGather and synchronize batch counts across all distributed ranks. Returns\n\n\ngenerate_batches\nGenerate packed batches for training.\n\n\nset_epoch\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.efficiency()\nCalculate the packing efficiency (ratio of tokens used to total token slots).\nHigher is better - 1.0 would mean perfect packing with no wasted space.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_efficiency()\nGather and synchronize packing efficiency estimates across all distributed\nranks.\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nfloat\nA conservative efficiency estimate based on the measurements.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)\nGather and synchronize batch counts across all distributed ranks. Returns\nthe minimum number of batches available on any rank.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.generate_batches(set_stats=False)\nGenerate packed batches for training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nset_stats\nbool\nWhether to update efficiency statistics.\nFalse\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[list[int]]]\nList of batches, where each batch contains multiple bins, and each bin contains multiple sequence indices.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.set_epoch(epoch)\nSet the epoch number, used for reproducible shuffling across epochs" + "text": "Name\nDescription\n\n\n\n\nMultipackBatchSampler\nBatch sampler class for efficient packing of variable-length sequences\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler(\n sampler,\n batch_size,\n batch_max_len,\n lengths,\n bin_size,\n packing_efficiency_estimate=1.0,\n drop_last=True,\n num_count_samples=4,\n sequential=False,\n group_size=100000,\n num_processes=None,\n safe_mode=True,\n mp_start_method='fork',\n **kwargs,\n)\nBatch sampler class for efficient packing of variable-length sequences\nThis sampler packs sequences into fixed-capacity bins (batches) to maximize\nGPU memory utilization and training throughput by reducing padding.\nIt supports both parallel packing (using FFD algorithm) and\nsequential packing (preserving original sequence order).\n\n\n\n\n\nName\nDescription\n\n\n\n\nefficiency\nCalculate the packing efficiency (ratio of tokens used to total token slots).\n\n\ngather_efficiency\nGather and synchronize packing efficiency estimates across all distributed\n\n\ngather_len_batches\nGather and synchronize batch counts across all distributed ranks. Returns\n\n\ngenerate_batches\nGenerate packed batches for training.\n\n\nset_epoch\nSet the epoch number, used for reproducible shuffling across epochs\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.efficiency()\nCalculate the packing efficiency (ratio of tokens used to total token slots).\nHigher is better - 1.0 would mean perfect packing with no wasted space.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_efficiency()\nGather and synchronize packing efficiency estimates across all distributed\nranks.\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nfloat\nA conservative efficiency estimate based on the measurements.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)\nGather and synchronize batch counts across all distributed ranks. Returns\nthe minimum number of batches available on any rank.\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.generate_batches(set_stats=False)\nGenerate packed batches for training.\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\nDefault\n\n\n\n\nset_stats\nbool\nWhether to update efficiency statistics.\nFalse\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nName\nType\nDescription\n\n\n\n\n\nlist[list[list[int]]]\nList of batches, where each batch contains multiple bins, and each bin contains multiple sequence indices.\n\n\n\n\n\n\n\nutils.samplers.multipack.MultipackBatchSampler.set_epoch(epoch)\nSet the epoch number, used for reproducible shuffling across epochs" }, { "objectID": "docs/api/utils.samplers.multipack.html#functions", diff --git a/sitemap.xml b/sitemap.xml index 62b3954c0..1a389c490 100644 --- a/sitemap.xml +++ b/sitemap.xml @@ -2,802 +2,802 @@ https://docs.axolotl.ai/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html - 2025-12-04T14:44:55.031Z + 2025-12-08T14:16:36.173Z https://docs.axolotl.ai/docs/mac.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/cli.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.144Z https://docs.axolotl.ai/docs/mixed_precision.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/installation.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/dataset_loading.html - 2025-12-04T14:44:55.005Z + 2025-12-08T14:16:36.144Z https://docs.axolotl.ai/docs/sequence_parallelism.html - 2025-12-04T14:44:55.009Z + 2025-12-08T14:16:36.150Z https://docs.axolotl.ai/docs/optimizations.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/gradient_checkpointing.html - 2025-12-04T14:44:55.005Z + 2025-12-08T14:16:36.145Z https://docs.axolotl.ai/docs/docker.html - 2025-12-04T14:44:55.005Z + 2025-12-08T14:16:36.145Z https://docs.axolotl.ai/docs/input_output.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.148Z https://docs.axolotl.ai/docs/multi-gpu.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/dataset-formats/template_free.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.144Z https://docs.axolotl.ai/docs/dataset-formats/tokenized.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.144Z https://docs.axolotl.ai/docs/dataset-formats/pretraining.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.144Z https://docs.axolotl.ai/docs/dataset-formats/stepwise_supervised.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.144Z https://docs.axolotl.ai/docs/rlhf.html - 2025-12-04T14:44:55.009Z + 2025-12-08T14:16:36.150Z https://docs.axolotl.ai/docs/multi-node.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/batch_vs_grad.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.143Z https://docs.axolotl.ai/docs/nd_parallelism.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/fsdp_qlora.html - 2025-12-04T14:44:55.005Z + 2025-12-08T14:16:36.145Z https://docs.axolotl.ai/docs/quantize.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.150Z https://docs.axolotl.ai/docs/reward_modelling.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.150Z https://docs.axolotl.ai/docs/api/core.chat.format.llama3x.html - 2025-12-04T15:01:38.011Z + 2025-12-08T14:20:16.823Z https://docs.axolotl.ai/docs/api/monkeypatch.unsloth_.html - 2025-12-04T15:01:38.917Z + 2025-12-08T14:20:17.719Z https://docs.axolotl.ai/docs/api/integrations.kd.trainer.html - 2025-12-04T15:01:39.422Z + 2025-12-08T14:20:18.218Z https://docs.axolotl.ai/docs/api/integrations.grokfast.optimizer.html - 2025-12-04T15:01:39.413Z + 2025-12-08T14:20:18.209Z https://docs.axolotl.ai/docs/api/core.chat.format.chatml.html - 2025-12-04T15:01:38.010Z + 2025-12-08T14:20:16.821Z https://docs.axolotl.ai/docs/api/utils.callbacks.perplexity.html - 2025-12-04T15:01:39.548Z + 2025-12-08T14:20:18.344Z https://docs.axolotl.ai/docs/api/utils.callbacks.profiler.html - 2025-12-04T15:01:39.553Z + 2025-12-08T14:20:18.349Z https://docs.axolotl.ai/docs/api/cli.preprocess.html - 2025-12-04T15:01:38.204Z + 2025-12-08T14:20:17.013Z https://docs.axolotl.ai/docs/api/cli.utils.load.html - 2025-12-04T15:01:38.260Z + 2025-12-08T14:20:17.068Z https://docs.axolotl.ai/docs/api/cli.inference.html - 2025-12-04T15:01:38.170Z + 2025-12-08T14:20:16.979Z https://docs.axolotl.ai/docs/api/cli.args.html - 2025-12-04T15:01:38.113Z + 2025-12-08T14:20:16.924Z https://docs.axolotl.ai/docs/api/prompt_tokenizers.html - 2025-12-04T15:01:37.933Z + 2025-12-08T14:20:16.745Z https://docs.axolotl.ai/docs/api/cli.vllm_serve.html - 2025-12-04T15:01:38.219Z + 2025-12-08T14:20:17.027Z https://docs.axolotl.ai/docs/api/monkeypatch.data.batch_dataset_fetcher.html - 2025-12-04T15:01:38.929Z + 2025-12-08T14:20:17.730Z https://docs.axolotl.ai/docs/api/loaders.constants.html - 2025-12-04T15:01:38.406Z + 2025-12-08T14:20:17.213Z https://docs.axolotl.ai/docs/api/prompt_strategies.input_output.html - 2025-12-04T15:01:38.568Z + 2025-12-08T14:20:17.373Z https://docs.axolotl.ai/docs/api/core.trainers.trl.html - 2025-12-04T15:01:38.317Z + 2025-12-08T14:20:17.124Z https://docs.axolotl.ai/docs/api/models.mamba.modeling_mamba.html - 2025-12-04T15:01:39.457Z + 2025-12-08T14:20:18.253Z https://docs.axolotl.ai/docs/api/monkeypatch.stablelm_attn_hijack_flash.html - 2025-12-04T15:01:38.904Z + 2025-12-08T14:20:17.705Z https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_chat.html - 2025-12-04T15:01:38.511Z + 2025-12-08T14:20:17.317Z https://docs.axolotl.ai/docs/api/prompt_strategies.metharme.html - 2025-12-04T15:01:38.581Z + 2025-12-08T14:20:17.386Z https://docs.axolotl.ai/docs/api/utils.freeze.html - 2025-12-04T15:01:39.004Z + 2025-12-08T14:20:17.805Z https://docs.axolotl.ai/docs/api/logging_config.html - 2025-12-04T15:01:37.944Z + 2025-12-08T14:20:16.757Z https://docs.axolotl.ai/docs/api/monkeypatch.trainer_fsdp_optim.html - 2025-12-04T15:01:38.908Z + 2025-12-08T14:20:17.710Z https://docs.axolotl.ai/docs/api/cli.art.html - 2025-12-04T15:01:38.117Z + 2025-12-08T14:20:16.928Z https://docs.axolotl.ai/docs/api/utils.collators.mamba.html - 2025-12-04T15:01:39.486Z + 2025-12-08T14:20:18.282Z https://docs.axolotl.ai/docs/api/cli.utils.train.html - 2025-12-04T15:01:38.281Z + 2025-12-08T14:20:17.089Z https://docs.axolotl.ai/docs/api/integrations.base.html - 2025-12-04T15:01:39.407Z + 2025-12-08T14:20:18.204Z https://docs.axolotl.ai/docs/api/core.trainers.dpo.trainer.html - 2025-12-04T15:01:38.332Z + 2025-12-08T14:20:17.139Z https://docs.axolotl.ai/docs/api/cli.checks.html - 2025-12-04T15:01:38.125Z + 2025-12-08T14:20:16.935Z https://docs.axolotl.ai/docs/api/common.architectures.html - 2025-12-04T15:01:39.436Z + 2025-12-08T14:20:18.232Z https://docs.axolotl.ai/docs/api/utils.bench.html - 2025-12-04T15:01:38.994Z + 2025-12-08T14:20:17.795Z https://docs.axolotl.ai/docs/api/utils.data.streaming.html - 2025-12-04T15:01:39.101Z + 2025-12-08T14:20:17.900Z https://docs.axolotl.ai/docs/api/utils.chat_templates.html - 2025-12-04T15:01:38.977Z + 2025-12-08T14:20:17.778Z https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_w_system.html - 2025-12-04T15:01:38.528Z + 2025-12-08T14:20:17.333Z https://docs.axolotl.ai/docs/api/core.trainers.utils.html - 2025-12-04T15:01:38.361Z + 2025-12-08T14:20:17.168Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.zephyr.html - 2025-12-04T15:01:38.634Z + 2025-12-08T14:20:17.439Z https://docs.axolotl.ai/docs/api/cli.evaluate.html - 2025-12-04T15:01:38.089Z + 2025-12-08T14:20:16.900Z https://docs.axolotl.ai/docs/api/monkeypatch.relora.html - 2025-12-04T15:01:38.847Z + 2025-12-08T14:20:17.649Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chatml.html - 2025-12-04T15:01:38.633Z + 2025-12-08T14:20:17.437Z https://docs.axolotl.ai/docs/api/cli.utils.fetch.html - 2025-12-04T15:01:38.253Z + 2025-12-08T14:20:17.061Z https://docs.axolotl.ai/docs/api/utils.schemas.datasets.html - 2025-12-04T15:01:39.186Z + 2025-12-08T14:20:17.984Z https://docs.axolotl.ai/docs/api/common.datasets.html - 2025-12-04T15:01:39.456Z + 2025-12-08T14:20:18.252Z https://docs.axolotl.ai/docs/api/utils.callbacks.mlflow_.html - 2025-12-04T15:01:39.559Z + 2025-12-08T14:20:18.355Z https://docs.axolotl.ai/docs/api/prompt_strategies.kto.chatml.html - 2025-12-04T15:01:38.658Z + 2025-12-08T14:20:17.463Z https://docs.axolotl.ai/docs/api/utils.schemas.model.html - 2025-12-04T15:01:39.156Z + 2025-12-08T14:20:17.954Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_xformers.html - 2025-12-04T15:01:38.839Z + 2025-12-08T14:20:17.641Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_patch_multipack.html - 2025-12-04T15:01:38.897Z + 2025-12-08T14:20:17.698Z https://docs.axolotl.ai/docs/api/core.trainers.base.html - 2025-12-04T15:01:38.299Z + 2025-12-08T14:20:17.106Z https://docs.axolotl.ai/docs/api/cli.utils.args.html - 2025-12-04T15:01:38.247Z + 2025-12-08T14:20:17.055Z https://docs.axolotl.ai/docs/api/utils.schemas.trl.html - 2025-12-04T15:01:39.201Z + 2025-12-08T14:20:17.999Z https://docs.axolotl.ai/docs/api/core.builders.base.html - 2025-12-04T15:01:37.952Z + 2025-12-08T14:20:16.764Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_flash.html - 2025-12-04T15:01:38.837Z + 2025-12-08T14:20:17.639Z https://docs.axolotl.ai/docs/api/cli.cloud.modal_.html - 2025-12-04T15:01:38.231Z + 2025-12-08T14:20:17.039Z https://docs.axolotl.ai/docs/api/utils.optimizers.adopt.html - 2025-12-04T15:01:39.099Z + 2025-12-08T14:20:17.898Z https://docs.axolotl.ai/docs/api/integrations.spectrum.args.html - 2025-12-04T15:01:39.434Z + 2025-12-08T14:20:18.230Z https://docs.axolotl.ai/docs/api/cli.main.html - 2025-12-04T15:01:38.069Z + 2025-12-08T14:20:16.880Z https://docs.axolotl.ai/docs/api/kernels.lora.html - 2025-12-04T15:01:38.794Z + 2025-12-08T14:20:17.597Z https://docs.axolotl.ai/docs/api/utils.model_shard_quant.html - 2025-12-04T15:01:38.990Z + 2025-12-08T14:20:17.791Z https://docs.axolotl.ai/docs/api/utils.distributed.html - 2025-12-04T15:01:39.082Z + 2025-12-08T14:20:17.882Z https://docs.axolotl.ai/docs/api/loaders.model.html - 2025-12-04T15:01:38.373Z + 2025-12-08T14:20:17.180Z https://docs.axolotl.ai/docs/api/monkeypatch.mistral_attn_hijack_flash.html - 2025-12-04T15:01:38.840Z + 2025-12-08T14:20:17.643Z https://docs.axolotl.ai/docs/api/evaluate.html - 2025-12-04T15:01:37.857Z + 2025-12-08T14:20:16.670Z https://docs.axolotl.ai/docs/api/cli.delinearize_llama4.html - 2025-12-04T15:01:38.152Z + 2025-12-08T14:20:16.962Z https://docs.axolotl.ai/docs/api/utils.trainer.html - 2025-12-04T15:01:39.024Z + 2025-12-08T14:20:17.825Z https://docs.axolotl.ai/docs/api/cli.quantize.html - 2025-12-04T15:01:38.210Z + 2025-12-08T14:20:17.019Z https://docs.axolotl.ai/docs/api/common.const.html - 2025-12-04T15:01:39.438Z + 2025-12-08T14:20:18.234Z https://docs.axolotl.ai/docs/api/prompt_strategies.llama2_chat.html - 2025-12-04T15:01:38.553Z + 2025-12-08T14:20:17.358Z https://docs.axolotl.ai/docs/api/utils.collators.mm_chat.html - 2025-12-04T15:01:39.492Z + 2025-12-08T14:20:18.288Z https://docs.axolotl.ai/docs/api/core.datasets.chat.html - 2025-12-04T15:01:38.019Z + 2025-12-08T14:20:16.831Z https://docs.axolotl.ai/docs/api/prompt_strategies.bradley_terry.llama3.html - 2025-12-04T15:01:38.690Z + 2025-12-08T14:20:17.494Z https://docs.axolotl.ai/docs/api/core.trainers.mamba.html - 2025-12-04T15:01:38.324Z + 2025-12-08T14:20:17.131Z https://docs.axolotl.ai/docs/api/core.builders.causal.html - 2025-12-04T15:01:37.958Z + 2025-12-08T14:20:16.770Z https://docs.axolotl.ai/docs/api/kernels.geglu.html - 2025-12-04T15:01:38.807Z + 2025-12-08T14:20:17.610Z https://docs.axolotl.ai/docs/api/utils.schemas.utils.html - 2025-12-04T15:01:39.242Z + 2025-12-08T14:20:18.040Z https://docs.axolotl.ai/docs/custom_integrations.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.144Z https://docs.axolotl.ai/index.html - 2025-12-04T14:44:55.026Z + 2025-12-08T14:16:36.168Z https://docs.axolotl.ai/examples/colab-notebooks/colab-axolotl-example.html - 2025-12-04T14:44:55.013Z + 2025-12-08T14:16:36.155Z https://docs.axolotl.ai/FAQS.html - 2025-12-04T14:44:55.002Z + 2025-12-08T14:16:36.141Z https://docs.axolotl.ai/docs/inference.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.148Z https://docs.axolotl.ai/docs/api/core.datasets.transforms.chat_builder.html - 2025-12-04T15:01:38.029Z + 2025-12-08T14:20:16.840Z https://docs.axolotl.ai/docs/api/train.html - 2025-12-04T15:01:37.844Z + 2025-12-08T14:20:16.658Z https://docs.axolotl.ai/docs/api/utils.schemas.training.html - 2025-12-04T15:01:39.164Z + 2025-12-08T14:20:17.963Z https://docs.axolotl.ai/docs/api/kernels.quantize.html - 2025-12-04T15:01:38.828Z + 2025-12-08T14:20:17.631Z https://docs.axolotl.ai/docs/api/utils.dict.html - 2025-12-04T15:01:39.089Z + 2025-12-08T14:20:17.889Z https://docs.axolotl.ai/docs/api/utils.quantization.html - 2025-12-04T15:01:39.132Z + 2025-12-08T14:20:17.930Z https://docs.axolotl.ai/docs/api/utils.collators.batching.html - 2025-12-04T15:01:39.482Z + 2025-12-08T14:20:18.278Z https://docs.axolotl.ai/docs/api/monkeypatch.multipack.html - 2025-12-04T15:01:38.842Z + 2025-12-08T14:20:17.645Z https://docs.axolotl.ai/docs/api/cli.config.html - 2025-12-04T15:01:38.146Z + 2025-12-08T14:20:16.956Z https://docs.axolotl.ai/docs/api/core.trainers.mixins.optimizer.html - 2025-12-04T15:01:38.412Z + 2025-12-08T14:20:17.219Z https://docs.axolotl.ai/docs/api/utils.lora.html - 2025-12-04T15:01:38.983Z + 2025-12-08T14:20:17.784Z https://docs.axolotl.ai/docs/api/core.trainers.grpo.sampler.html - 2025-12-04T15:01:38.360Z + 2025-12-08T14:20:17.166Z https://docs.axolotl.ai/docs/api/prompt_strategies.orcamini.html - 2025-12-04T15:01:38.586Z + 2025-12-08T14:20:17.391Z https://docs.axolotl.ai/docs/api/monkeypatch.mixtral.html - 2025-12-04T15:01:38.930Z + 2025-12-08T14:20:17.732Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.llama3.html - 2025-12-04T15:01:38.620Z + 2025-12-08T14:20:17.425Z https://docs.axolotl.ai/docs/api/index.html - 2025-12-04T15:01:37.765Z + 2025-12-08T14:20:16.580Z https://docs.axolotl.ai/docs/api/utils.ctx_managers.sequence_parallel.html - 2025-12-04T15:01:38.453Z + 2025-12-08T14:20:17.260Z https://docs.axolotl.ai/docs/api/cli.utils.html - 2025-12-04T15:01:38.233Z + 2025-12-08T14:20:17.041Z https://docs.axolotl.ai/docs/api/cli.merge_lora.html - 2025-12-04T15:01:38.180Z + 2025-12-08T14:20:16.989Z https://docs.axolotl.ai/docs/api/kernels.utils.html - 2025-12-04T15:01:38.830Z + 2025-12-08T14:20:17.633Z https://docs.axolotl.ai/docs/api/prompt_strategies.base.html - 2025-12-04T15:01:38.455Z + 2025-12-08T14:20:17.261Z https://docs.axolotl.ai/docs/api/utils.callbacks.comet_.html - 2025-12-04T15:01:39.563Z + 2025-12-08T14:20:18.359Z https://docs.axolotl.ai/docs/api/utils.schemas.multimodal.html - 2025-12-04T15:01:39.207Z + 2025-12-08T14:20:18.005Z https://docs.axolotl.ai/docs/api/prompt_strategies.chat_template.html - 2025-12-04T15:01:38.495Z + 2025-12-08T14:20:17.301Z https://docs.axolotl.ai/docs/api/utils.callbacks.qat.html - 2025-12-04T15:01:39.572Z + 2025-12-08T14:20:18.367Z https://docs.axolotl.ai/docs/api/utils.samplers.multipack.html - 2025-12-04T15:01:39.541Z + 2025-12-08T14:20:18.337Z https://docs.axolotl.ai/docs/api/prompt_strategies.orpo.chat_template.html - 2025-12-04T15:01:38.685Z + 2025-12-08T14:20:17.489Z https://docs.axolotl.ai/docs/api/core.chat.format.shared.html - 2025-12-04T15:01:38.013Z + 2025-12-08T14:20:16.825Z https://docs.axolotl.ai/docs/api/utils.collators.core.html - 2025-12-04T15:01:39.459Z + 2025-12-08T14:20:18.255Z https://docs.axolotl.ai/docs/api/integrations.lm_eval.args.html - 2025-12-04T15:01:39.430Z + 2025-12-08T14:20:18.226Z https://docs.axolotl.ai/docs/api/core.trainers.grpo.trainer.html - 2025-12-04T15:01:38.345Z + 2025-12-08T14:20:17.152Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chat_template.html - 2025-12-04T15:01:38.607Z + 2025-12-08T14:20:17.412Z https://docs.axolotl.ai/docs/api/monkeypatch.btlm_attn_hijack_flash.html - 2025-12-04T15:01:38.895Z + 2025-12-08T14:20:17.697Z https://docs.axolotl.ai/docs/api/utils.schemas.enums.html - 2025-12-04T15:01:39.235Z + 2025-12-08T14:20:18.033Z https://docs.axolotl.ai/docs/api/utils.schemas.config.html - 2025-12-04T15:01:39.147Z + 2025-12-08T14:20:17.946Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.passthrough.html - 2025-12-04T15:01:38.638Z + 2025-12-08T14:20:17.443Z https://docs.axolotl.ai/docs/api/convert.html - 2025-12-04T15:01:37.881Z + 2025-12-08T14:20:16.694Z https://docs.axolotl.ai/docs/api/utils.schemas.integrations.html - 2025-12-04T15:01:39.224Z + 2025-12-08T14:20:18.022Z https://docs.axolotl.ai/docs/api/prompt_strategies.kto.llama3.html - 2025-12-04T15:01:38.648Z + 2025-12-08T14:20:17.453Z https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.user_defined.html - 2025-12-04T15:01:38.636Z + 2025-12-08T14:20:17.441Z https://docs.axolotl.ai/docs/api/core.trainers.mixins.scheduler.html - 2025-12-04T15:01:38.424Z + 2025-12-08T14:20:17.231Z https://docs.axolotl.ai/docs/api/core.chat.messages.html - 2025-12-04T15:01:38.008Z + 2025-12-08T14:20:16.819Z https://docs.axolotl.ai/docs/api/prompt_strategies.user_defined.html - 2025-12-04T15:01:38.538Z + 2025-12-08T14:20:17.343Z https://docs.axolotl.ai/docs/api/prompt_strategies.messages.chat.html - 2025-12-04T15:01:38.599Z + 2025-12-08T14:20:17.404Z https://docs.axolotl.ai/docs/api/monkeypatch.lora_kernels.html - 2025-12-04T15:01:38.883Z + 2025-12-08T14:20:17.685Z https://docs.axolotl.ai/docs/api/prompt_strategies.stepwise_supervised.html - 2025-12-04T15:01:38.573Z + 2025-12-08T14:20:17.378Z https://docs.axolotl.ai/docs/api/prompt_strategies.completion.html - 2025-12-04T15:01:38.560Z + 2025-12-08T14:20:17.366Z https://docs.axolotl.ai/docs/api/core.trainers.mixins.rng_state_loader.html - 2025-12-04T15:01:38.416Z + 2025-12-08T14:20:17.223Z https://docs.axolotl.ai/docs/api/cli.train.html - 2025-12-04T15:01:38.079Z + 2025-12-08T14:20:16.890Z https://docs.axolotl.ai/docs/api/cli.merge_sharded_fsdp_weights.html - 2025-12-04T15:01:38.194Z + 2025-12-08T14:20:17.003Z https://docs.axolotl.ai/docs/api/loaders.adapter.html - 2025-12-04T15:01:38.392Z + 2025-12-08T14:20:17.198Z https://docs.axolotl.ai/docs/api/core.training_args.html - 2025-12-04T15:01:37.979Z + 2025-12-08T14:20:16.791Z https://docs.axolotl.ai/docs/api/utils.callbacks.lisa.html - 2025-12-04T15:01:39.555Z + 2025-12-08T14:20:18.351Z https://docs.axolotl.ai/docs/api/loaders.processor.html - 2025-12-04T15:01:38.385Z + 2025-12-08T14:20:17.191Z https://docs.axolotl.ai/docs/api/core.builders.rl.html - 2025-12-04T15:01:37.964Z + 2025-12-08T14:20:16.775Z https://docs.axolotl.ai/docs/api/monkeypatch.utils.html - 2025-12-04T15:01:38.893Z + 2025-12-08T14:20:17.695Z https://docs.axolotl.ai/docs/api/prompt_strategies.kto.user_defined.html - 2025-12-04T15:01:38.660Z + 2025-12-08T14:20:17.464Z https://docs.axolotl.ai/docs/api/integrations.cut_cross_entropy.args.html - 2025-12-04T15:01:39.412Z + 2025-12-08T14:20:18.208Z https://docs.axolotl.ai/docs/api/kernels.swiglu.html - 2025-12-04T15:01:38.819Z + 2025-12-08T14:20:17.622Z https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_disk.html - 2025-12-04T15:01:38.967Z + 2025-12-08T14:20:17.768Z https://docs.axolotl.ai/docs/api/cli.cloud.base.html - 2025-12-04T15:01:38.223Z + 2025-12-08T14:20:17.031Z https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_instruct.html - 2025-12-04T15:01:38.513Z + 2025-12-08T14:20:17.319Z https://docs.axolotl.ai/docs/api/prompt_strategies.pygmalion.html - 2025-12-04T15:01:38.594Z + 2025-12-08T14:20:17.399Z https://docs.axolotl.ai/docs/api/utils.schemas.peft.html - 2025-12-04T15:01:39.197Z + 2025-12-08T14:20:17.995Z https://docs.axolotl.ai/docs/api/integrations.liger.args.html - 2025-12-04T15:01:39.426Z + 2025-12-08T14:20:18.222Z https://docs.axolotl.ai/docs/api/loaders.patch_manager.html - 2025-12-04T15:01:38.404Z + 2025-12-08T14:20:17.211Z https://docs.axolotl.ai/docs/api/monkeypatch.transformers_fa_utils.html - 2025-12-04T15:01:38.916Z + 2025-12-08T14:20:17.717Z https://docs.axolotl.ai/docs/api/utils.data.sft.html - 2025-12-04T15:01:39.108Z + 2025-12-08T14:20:17.907Z https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_cpu.html - 2025-12-04T15:01:38.934Z + 2025-12-08T14:20:17.736Z https://docs.axolotl.ai/docs/api/monkeypatch.llama_expand_mask.html - 2025-12-04T15:01:38.848Z + 2025-12-08T14:20:17.651Z https://docs.axolotl.ai/docs/api/loaders.tokenizer.html - 2025-12-04T15:01:38.383Z + 2025-12-08T14:20:17.190Z https://docs.axolotl.ai/docs/api/utils.tokenization.html - 2025-12-04T15:01:38.975Z + 2025-12-08T14:20:17.776Z https://docs.axolotl.ai/docs/api/datasets.html - 2025-12-04T15:01:37.865Z + 2025-12-08T14:20:16.678Z https://docs.axolotl.ai/docs/api/cli.utils.sweeps.html - 2025-12-04T15:01:38.267Z + 2025-12-08T14:20:17.075Z https://docs.axolotl.ai/docs/api/utils.schedulers.html - 2025-12-04T15:01:39.057Z + 2025-12-08T14:20:17.858Z https://docs.axolotl.ai/docs/optimizers.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.150Z https://docs.axolotl.ai/docs/torchao.html - 2025-12-04T14:44:55.009Z + 2025-12-08T14:16:36.150Z https://docs.axolotl.ai/docs/dataset_preprocessing.html - 2025-12-04T14:44:55.005Z + 2025-12-08T14:16:36.145Z https://docs.axolotl.ai/docs/faq.html - 2025-12-04T14:44:55.005Z + 2025-12-08T14:16:36.145Z https://docs.axolotl.ai/docs/ray-integration.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.150Z https://docs.axolotl.ai/docs/multimodal.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/config-reference.html - 2025-12-04T15:01:53.952Z + 2025-12-08T14:20:32.647Z https://docs.axolotl.ai/docs/dataset-formats/index.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.144Z https://docs.axolotl.ai/docs/dataset-formats/inst_tune.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.144Z https://docs.axolotl.ai/docs/dataset-formats/conversation.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.144Z https://docs.axolotl.ai/docs/debugging.html - 2025-12-04T14:44:55.005Z + 2025-12-08T14:16:36.145Z https://docs.axolotl.ai/docs/amd_hpc.html - 2025-12-04T14:44:55.004Z + 2025-12-08T14:16:36.143Z https://docs.axolotl.ai/docs/lora_optims.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/streaming.html - 2025-12-04T14:44:55.009Z + 2025-12-08T14:16:36.150Z https://docs.axolotl.ai/docs/multipack.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/qat.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.150Z https://docs.axolotl.ai/docs/lr_groups.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/getting-started.html - 2025-12-04T14:44:55.005Z + 2025-12-08T14:16:36.145Z https://docs.axolotl.ai/docs/nccl.html - 2025-12-04T14:44:55.008Z + 2025-12-08T14:16:36.149Z https://docs.axolotl.ai/docs/telemetry.html - 2025-12-04T14:44:55.009Z + 2025-12-08T14:16:36.150Z https://docs.axolotl.ai/docs/unsloth.html - 2025-12-04T14:44:55.009Z + 2025-12-08T14:16:36.151Z https://docs.axolotl.ai/src/axolotl/integrations/LICENSE.html - 2025-12-04T14:44:55.031Z + 2025-12-08T14:16:36.173Z