repalce linear layers for qlora as well as add peft

This commit is contained in:
Wing Lian
2023-11-01 22:31:02 -04:00
parent 026172eaa8
commit 87e8f13056

View File

@@ -502,7 +502,7 @@ def load_adapter(model, cfg, adapter, inference=False):
except NotImplementedError:
LOG.warning("enable_input_require_grads not implemented on model")
if adapter == "qlora" and cfg.tensor_parallel:
return load_tp_qlora(model)
model, _ = load_tp_qlora(model)
if adapter in ["lora", "qlora"]:
return load_lora(model, cfg, inference=inference)
if adapter == "llama-adapter":