remove monkeypatch
This commit is contained in:
@@ -21,9 +21,7 @@ val_set_size: 0.05
|
||||
output_dir: ./outputs/gemma-3-1b-fft-dft
|
||||
|
||||
sequence_len: 2048
|
||||
sample_packing: true
|
||||
|
||||
# Enable Dynamic Fine-Tuning loss
|
||||
use_dynamic_finetuning: true
|
||||
|
||||
wandb_project:
|
||||
|
||||
@@ -96,7 +96,6 @@ class PatchManager:
|
||||
# self._apply_flex_attention_patches()
|
||||
self._apply_flash_attention_patches()
|
||||
self._apply_chunked_cross_entropy_patch()
|
||||
self._apply_dft_loss_patch()
|
||||
self._apply_fsdp_patches()
|
||||
self._apply_adapter_patches()
|
||||
self._apply_model_specific_patches()
|
||||
@@ -158,11 +157,6 @@ class PatchManager:
|
||||
else:
|
||||
patch_chunked_ce_loss_fn()
|
||||
|
||||
def _apply_dft_loss_patch(self):
|
||||
# DFT loss is now applied via compute_loss_func in the trainer builder
|
||||
# See: src/axolotl/core/builders/causal.py
|
||||
pass
|
||||
|
||||
def _apply_fsdp_patches(self):
|
||||
"""Apply patches for FSDP configurations."""
|
||||
if self.cfg.context_parallel_size > 1 or (
|
||||
|
||||
Reference in New Issue
Block a user