From 32eeeb5b646460200660bbd5bcbe1e541af386e2 Mon Sep 17 00:00:00 2001 From: Casper Date: Sun, 22 Oct 2023 19:22:32 +0200 Subject: [PATCH] Hotfix for not saving correctly (#762) --- src/axolotl/monkeypatch/fused_modules.py | 0 src/axolotl/monkeypatch/llama_attn_hijack_flash.py | 1 + 2 files changed, 1 insertion(+) delete mode 100644 src/axolotl/monkeypatch/fused_modules.py diff --git a/src/axolotl/monkeypatch/fused_modules.py b/src/axolotl/monkeypatch/fused_modules.py deleted file mode 100644 index e69de29bb..000000000 diff --git a/src/axolotl/monkeypatch/llama_attn_hijack_flash.py b/src/axolotl/monkeypatch/llama_attn_hijack_flash.py index 386f4bfac..f0fa807fa 100644 --- a/src/axolotl/monkeypatch/llama_attn_hijack_flash.py +++ b/src/axolotl/monkeypatch/llama_attn_hijack_flash.py @@ -152,6 +152,7 @@ class FusedAttention(LlamaAttention): new_attn.q_proj.weight.data = q_proj new_attn.k_proj.weight.data = k_proj new_attn.v_proj.weight.data = v_proj + new_attn.o_proj.weight.data = self.o_proj.weight.data set_module_name(model, name, new_attn)