From 8608d8003ea060e0fed81b578ddb3f70e20a58b9 Mon Sep 17 00:00:00 2001 From: xhedit <71542719+xhedit@users.noreply.github.com> Date: Wed, 31 Jan 2024 06:46:55 -0500 Subject: [PATCH] Fix typo (#1231) [skip ci] --- src/axolotl/utils/models.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/axolotl/utils/models.py b/src/axolotl/utils/models.py index e2401b7fe..c12518b2d 100644 --- a/src/axolotl/utils/models.py +++ b/src/axolotl/utils/models.py @@ -756,7 +756,7 @@ def load_llama_adapter(model, cfg): ) if cfg.lora_model_dir: - LOG.debug("Loading pretained PEFT - llama_adapter") + LOG.debug("Loading pretrained PEFT - llama_adapter") model = PeftModel.from_pretrained( model, cfg.lora_model_dir, @@ -825,7 +825,7 @@ def load_lora(model, cfg, inference=False, config_only=False): return None, lora_config if cfg.lora_model_dir: - LOG.debug("Loading pretained PEFT - LoRA") + LOG.debug("Loading pretrained PEFT - LoRA") model_kwargs: Any = {} if cfg.lora_on_cpu: model_kwargs["max_memory"] = {"cpu": "256GiB"}