helpful info output

This commit is contained in:
Wing Lian
2023-04-15 00:03:43 -04:00
parent 902dd0ab47
commit 937f44f021
3 changed files with 5 additions and 3 deletions

View File

@@ -1,4 +1,4 @@
base_model: huggyllama/llama-7b
base_model: huggyllama/llama-65b
model_type: LlamaForCausalLM
tokenizer_type: LlamaTokenizer
load_in_8bit: true

View File

@@ -1,5 +1,5 @@
git+https://github.com/huggingface/transformers.git
git+https://github.com/huggingface/peft.git
git+https://github.com/huggingface/transformers.git
attrdict
fire
PyYAML==6.0
@@ -12,4 +12,3 @@ wandb
flash-attn
deepspeed
einops

View File

@@ -258,7 +258,9 @@ def train(
datasets = []
if not isinstance(cfg.datasets, list) and isinstance(cfg.datasets, str):
# assumption that we are loading a previously saved/cached dataset
print("Loading prepared dataset from disk...")
dataset = load_from_disk(cfg.datasets)
print("Prepared dataset loaded from disk...")
else:
for d in cfg.datasets:
ds: IterableDataset = load_dataset(
@@ -289,6 +291,7 @@ def train(
dataset = Dataset.from_list(
[_ for _ in constant_len_dataset]
).train_test_split(test_size=cfg.val_set_size, shuffle=True, seed=42)
print("Saving prepared dataset to disk...")
dataset.save_to_disk("data/last_run")
train_dataset = dataset["train"]