fix condition and add logging

This commit is contained in:
Aman Karmani
2023-08-27 20:09:26 +00:00
parent 1f613e5aa7
commit 3a011ea1ef

View File

@@ -355,7 +355,7 @@ def load_model(
if hasattr(module, "weight"):
module.to(torch.float32)
needs_fa2_dtype = not cfg.adapter
needs_fa2_dtype = cfg.adapter is not None
if not cfg.gptq and (
(cfg.adapter == "lora" and load_in_8bit)
or (cfg.adapter == "qlora" and cfg.load_in_4bit)
@@ -369,6 +369,7 @@ def load_model(
# LlamaRMSNorm layers are in fp32 after kbit_training or full finetune, so we need to
# convert them back to fp16/bf16 for flash-attn compatibility.
if needs_fa2_dtype and (cfg.flash_attention and cfg.is_llama_derived_model):
LOG.info("converting modules to %s for flash attention", cfg.torch_dtype)
for name, module in model.named_modules():
if "norm" in name:
module.to(cfg.torch_dtype)