Fix typo (#1231) [skip ci]

This commit is contained in:
xhedit
2024-01-31 06:46:55 -05:00
committed by GitHub
parent 4cb7900a56
commit 8608d8003e

View File

@@ -756,7 +756,7 @@ def load_llama_adapter(model, cfg):
)
if cfg.lora_model_dir:
LOG.debug("Loading pretained PEFT - llama_adapter")
LOG.debug("Loading pretrained PEFT - llama_adapter")
model = PeftModel.from_pretrained(
model,
cfg.lora_model_dir,
@@ -825,7 +825,7 @@ def load_lora(model, cfg, inference=False, config_only=False):
return None, lora_config
if cfg.lora_model_dir:
LOG.debug("Loading pretained PEFT - LoRA")
LOG.debug("Loading pretrained PEFT - LoRA")
model_kwargs: Any = {}
if cfg.lora_on_cpu:
model_kwargs["max_memory"] = {"cpu": "256GiB"}