helpful info output
This commit is contained in:
@@ -1,4 +1,4 @@
|
||||
base_model: huggyllama/llama-7b
|
||||
base_model: huggyllama/llama-65b
|
||||
model_type: LlamaForCausalLM
|
||||
tokenizer_type: LlamaTokenizer
|
||||
load_in_8bit: true
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
git+https://github.com/huggingface/transformers.git
|
||||
git+https://github.com/huggingface/peft.git
|
||||
git+https://github.com/huggingface/transformers.git
|
||||
attrdict
|
||||
fire
|
||||
PyYAML==6.0
|
||||
@@ -12,4 +12,3 @@ wandb
|
||||
flash-attn
|
||||
deepspeed
|
||||
einops
|
||||
|
||||
|
||||
@@ -258,7 +258,9 @@ def train(
|
||||
datasets = []
|
||||
if not isinstance(cfg.datasets, list) and isinstance(cfg.datasets, str):
|
||||
# assumption that we are loading a previously saved/cached dataset
|
||||
print("Loading prepared dataset from disk...")
|
||||
dataset = load_from_disk(cfg.datasets)
|
||||
print("Prepared dataset loaded from disk...")
|
||||
else:
|
||||
for d in cfg.datasets:
|
||||
ds: IterableDataset = load_dataset(
|
||||
@@ -289,6 +291,7 @@ def train(
|
||||
dataset = Dataset.from_list(
|
||||
[_ for _ in constant_len_dataset]
|
||||
).train_test_split(test_size=cfg.val_set_size, shuffle=True, seed=42)
|
||||
print("Saving prepared dataset to disk...")
|
||||
dataset.save_to_disk("data/last_run")
|
||||
|
||||
train_dataset = dataset["train"]
|
||||
|
||||
Reference in New Issue
Block a user