From 01790217801313158e36acb4c76f6fdd93d1cc90 Mon Sep 17 00:00:00 2001 From: Sunny Liu Date: Mon, 21 Apr 2025 22:29:24 -0400 Subject: [PATCH] fix attribute error --- src/axolotl/utils/schemas/peft.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/src/axolotl/utils/schemas/peft.py b/src/axolotl/utils/schemas/peft.py index 0fc50b783..b7ff4335f 100644 --- a/src/axolotl/utils/schemas/peft.py +++ b/src/axolotl/utils/schemas/peft.py @@ -87,24 +87,24 @@ class LoraConfig(BaseModel): def validate_qlora(self): if self.adapter == "qlora": if self.merge_lora: - if self.quantization.bits == 8 or self.cfg.load_in_8bit: + if self.quantization.bits == 8 or self.load_in_8bit: raise ValueError("Can't merge qlora if loaded in 8bit") if self.quantization.backend == "gptq": raise ValueError("Can't merge qlora if using gptq") - if self.quantization.bits == 4 or self.cfg.load_in_4bit: + if self.quantization.bits == 4 or self.load_in_4bit: raise ValueError("Can't merge qlora if loaded in 4bit") else: if self.quantization: - if self.quantization.bits == 8 or self.cfg.load_in_8bit: + if self.quantization.bits == 8 or self.load_in_8bit: raise ValueError("Can't load qlora in 8bit") if self.quantization.backend == "gptq": raise ValueError("Can't load qlora if using gptq") - if not self.quantization.bits == 4 or self.cfg.load_in_4bit: + if not self.quantization.bits == 4 or self.load_in_4bit: raise ValueError("Require quantization.bits <= 4 for qlora") return self