chore: consolidate eval_strat, loraplus, lr sched, max_length

This commit is contained in:
NanoCode012
2025-01-28 15:04:18 +07:00
parent fd271b2547
commit 053e5fd7d1
2 changed files with 42 additions and 65 deletions

View File

@@ -529,7 +529,7 @@ profiler_steps: # enable the pytorch profiler to capture the first N steps of tr
loss_watchdog_threshold: # High loss value, indicating the learning has broken down (a good estimate is ~2 times the loss at the start of training)
loss_watchdog_patience: # Number of high-loss steps in a row before the trainer aborts (default: 3)
# Save model as safetensors (require safetensors package)
# Save model as safetensors (require safetensors package). Transformers default True
save_safetensors:
# Whether to mask out or include the human's prompt from the training labels