inline comment change

This commit is contained in:
Dan Saunders
2025-01-14 16:59:43 +00:00
parent fd8ad6fcbf
commit 28694219a5

View File

@@ -384,7 +384,7 @@ class LlamaDifferentialAttentionBase(nn.Module):
"""
attn = self.subln(attn)
# NOTE: this may need to be added back in, but doesn't interact well with
# `diff_attn_mix`.
# `diff_attn_mix`, and doesn't allow us to preserve the original model output.
# attn = attn * self.diff_attn_mix * (1 - self.lambda_init)
attn = attn.transpose(1, 2).reshape(bsz, q_len, self.config.hidden_size)