Merge pull request #12 from NanoCode012/feat/eval_config

Add eval_batch_size for evaluation
This commit is contained in:
Wing Lian
2023-05-07 01:46:53 -04:00
committed by GitHub
2 changed files with 2 additions and 0 deletions

View File

@@ -85,6 +85,7 @@ output_dir: ./completed-model
# training hyperparameters
batch_size: 8
micro_batch_size: 2
eval_batch_size: 2
num_epochs: 3
warmup_steps: 100
learning_rate: 0.00003

View File

@@ -47,6 +47,7 @@ def setup_trainer(cfg, train_dataset, eval_dataset, model, tokenizer):
training_args = transformers.TrainingArguments(
per_device_train_batch_size=cfg.micro_batch_size,
per_device_eval_batch_size=cfg.eval_batch_size,
gradient_accumulation_steps=cfg.gradient_accumulation_steps,
num_train_epochs=cfg.num_epochs,
learning_rate=cfg.learning_rate,