Update tinyllama lora.yml to fix eval packing issue (#1362)

This commit is contained in:
Sebastian Raschka
2024-03-05 13:36:29 -06:00
committed by GitHub
parent 2598c9f045
commit 8984bf1722

View File

@@ -15,6 +15,7 @@ output_dir: ./lora-out
sequence_len: 4096
sample_packing: true
eval_sample_packing: false
pad_to_sequence_len: true
adapter: lora