Address PR suggestion

Co-authored-by: Wing Lian <wing.lian@gmail.com>
This commit is contained in:
NanoCode012
2023-06-11 10:52:12 +09:00
committed by GitHub
parent 919727b4d7
commit e285e24f7f

View File

@@ -202,7 +202,7 @@ def load_model(
else True,
)
load_in_8bit = False
elif cfg.is_llama_derived_model and "LlamaForCausalLM" in globals():
elif cfg.is_llama_derived_model:
try:
from transformers import LlamaForCausalLM
except ImportError: