diff --git a/README.md b/README.md index 0f9a5c632..759e6932c 100644 --- a/README.md +++ b/README.md @@ -364,6 +364,9 @@ dataset_prepared_path: data/last_run_prepared push_dataset_to_hub: # repo path # push checkpoints to hub hub_model_id: # repo path to push finetuned model +# how to push checkpoints to hub +# https://huggingface.co/docs/transformers/v4.31.0/en/main_classes/trainer#transformers.TrainingArguments.hub_strategy +hub_strategy: # whether to use hf `use_auth_token` for loading datasets. Useful for fetching private datasets # required to be true when used in combination with `push_dataset_to_hub` hf_use_auth_token: # boolean diff --git a/src/axolotl/utils/trainer.py b/src/axolotl/utils/trainer.py index b143cb01f..a446b2625 100644 --- a/src/axolotl/utils/trainer.py +++ b/src/axolotl/utils/trainer.py @@ -440,6 +440,9 @@ def setup_trainer(cfg, train_dataset, eval_dataset, model, tokenizer, total_num_ training_arguments_kwargs["push_to_hub"] = True training_arguments_kwargs["hub_private_repo"] = True + if cfg.hub_strategy: + training_arguments_kwargs["hub_strategy"] = cfg.hub_strategy + if cfg.save_safetensors: training_arguments_kwargs["save_safetensors"] = cfg.save_safetensors