From 74e7cfd28fca8dd6c80c501094298538384ad580 Mon Sep 17 00:00:00 2001 From: Dan Saunders Date: Fri, 11 Apr 2025 04:06:37 +0000 Subject: [PATCH] update --- .../monkeypatch/attention/ring_attn/adapters/batch_ring.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/src/axolotl/monkeypatch/attention/ring_attn/adapters/batch_ring.py b/src/axolotl/monkeypatch/attention/ring_attn/adapters/batch_ring.py index ea4f9c57a..91f1d0c67 100644 --- a/src/axolotl/monkeypatch/attention/ring_attn/adapters/batch_ring.py +++ b/src/axolotl/monkeypatch/attention/ring_attn/adapters/batch_ring.py @@ -34,8 +34,7 @@ def create_ring_flash_attention_forward(process_group: dist.ProcessGroup) -> Cal Create a ring flash attention forward function compatible with HuggingFace's interface. Args: - process_group: A PyTorch distributed process group that defines the communication - topology for the ring attention pattern. + process_group: A PyTorch distributed process group. Returns: A function that implements the ring flash attention forward pass with the