diff --git a/src/axolotl/monkeypatch/llama_attn_hijack_xformers.py b/src/axolotl/monkeypatch/llama_attn_hijack_xformers.py index bb5728ef1..d4b7165aa 100644 --- a/src/axolotl/monkeypatch/llama_attn_hijack_xformers.py +++ b/src/axolotl/monkeypatch/llama_attn_hijack_xformers.py @@ -25,7 +25,6 @@ def hijack_llama_sdp_attention(): transformers.models.llama.modeling_llama.LlamaAttention.forward = ( sdp_attention_forward ) - logging.info("Replaced attention with sdp_attention") def xformers_forward(