From 937f44f0211d1722a2b7f8450e22559c03bf8df4 Mon Sep 17 00:00:00 2001 From: Wing Lian Date: Sat, 15 Apr 2023 00:03:43 -0400 Subject: [PATCH] helpful info output --- configs/llama_65B_alpaca.yml | 2 +- requirements.txt | 3 +-- scripts/finetune.py | 3 +++ 3 files changed, 5 insertions(+), 3 deletions(-) diff --git a/configs/llama_65B_alpaca.yml b/configs/llama_65B_alpaca.yml index 54fe0786f..d3f98e3ec 100644 --- a/configs/llama_65B_alpaca.yml +++ b/configs/llama_65B_alpaca.yml @@ -1,4 +1,4 @@ -base_model: huggyllama/llama-7b +base_model: huggyllama/llama-65b model_type: LlamaForCausalLM tokenizer_type: LlamaTokenizer load_in_8bit: true diff --git a/requirements.txt b/requirements.txt index 8e12e8b7b..230b86b26 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,5 +1,5 @@ -git+https://github.com/huggingface/transformers.git git+https://github.com/huggingface/peft.git +git+https://github.com/huggingface/transformers.git attrdict fire PyYAML==6.0 @@ -12,4 +12,3 @@ wandb flash-attn deepspeed einops - diff --git a/scripts/finetune.py b/scripts/finetune.py index 090f8099e..bbba3dc93 100644 --- a/scripts/finetune.py +++ b/scripts/finetune.py @@ -258,7 +258,9 @@ def train( datasets = [] if not isinstance(cfg.datasets, list) and isinstance(cfg.datasets, str): # assumption that we are loading a previously saved/cached dataset + print("Loading prepared dataset from disk...") dataset = load_from_disk(cfg.datasets) + print("Prepared dataset loaded from disk...") else: for d in cfg.datasets: ds: IterableDataset = load_dataset( @@ -289,6 +291,7 @@ def train( dataset = Dataset.from_list( [_ for _ in constant_len_dataset] ).train_test_split(test_size=cfg.val_set_size, shuffle=True, seed=42) + print("Saving prepared dataset to disk...") dataset.save_to_disk("data/last_run") train_dataset = dataset["train"]