From 33d40179ba08ab7db53d1a30745889704f41e9a1 Mon Sep 17 00:00:00 2001 From: NanoCode012 Date: Wed, 31 May 2023 14:04:49 +0900 Subject: [PATCH] Increase max_new_tokens Co-authored-by: Wing Lian --- scripts/finetune.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/scripts/finetune.py b/scripts/finetune.py index 1b84859a4..731e98bff 100644 --- a/scripts/finetune.py +++ b/scripts/finetune.py @@ -83,7 +83,7 @@ def do_inference(cfg, model, tokenizer, prompter="AlpacaPrompter"): with torch.no_grad(): generation_config = GenerationConfig( repetition_penalty=1.1, - max_new_tokens=100, + max_new_tokens=1024, temperature=0.9, top_p=0.95, top_k=40,