remove monkeypatch

This commit is contained in:
Salman Mohammadi
2026-01-21 10:06:58 +00:00
parent 7a4f33802d
commit 0a0115493d
2 changed files with 0 additions and 8 deletions

View File

@@ -21,9 +21,7 @@ val_set_size: 0.05
output_dir: ./outputs/gemma-3-1b-fft-dft
sequence_len: 2048
sample_packing: true
# Enable Dynamic Fine-Tuning loss
use_dynamic_finetuning: true
wandb_project:

View File

@@ -96,7 +96,6 @@ class PatchManager:
# self._apply_flex_attention_patches()
self._apply_flash_attention_patches()
self._apply_chunked_cross_entropy_patch()
self._apply_dft_loss_patch()
self._apply_fsdp_patches()
self._apply_adapter_patches()
self._apply_model_specific_patches()
@@ -158,11 +157,6 @@ class PatchManager:
else:
patch_chunked_ce_loss_fn()
def _apply_dft_loss_patch(self):
# DFT loss is now applied via compute_loss_func in the trainer builder
# See: src/axolotl/core/builders/causal.py
pass
def _apply_fsdp_patches(self):
"""Apply patches for FSDP configurations."""
if self.cfg.context_parallel_size > 1 or (