From d03887fad5044a90b1984baaad36387079ecd4f6 Mon Sep 17 00:00:00 2001 From: Maxime <672982+maximegmd@users.noreply.github.com> Date: Sat, 26 Aug 2023 22:45:45 +0200 Subject: [PATCH] ignore: address pr review --- src/axolotl/utils/models.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/axolotl/utils/models.py b/src/axolotl/utils/models.py index fc2cf04b3..71e27a2bc 100644 --- a/src/axolotl/utils/models.py +++ b/src/axolotl/utils/models.py @@ -368,7 +368,7 @@ def load_model( # LlamaRMSNorm layers are in fp32 after kbit_training or full finetune, so we need to # convert them back to fp16/bf16 for flash-attn compatibility. - if (fix_dtype or cfg.adapter == "" or cfg.adapter is None) and ( + if (fix_dtype or not cfg.adapter) and ( cfg.flash_attention and cfg.is_llama_derived_model ): for name, module in model.named_modules():