more test

This commit is contained in:
bursteratom
2025-02-02 00:43:41 -05:00
parent fa7355404c
commit 10de67e8ea
2 changed files with 5 additions and 3 deletions

View File

@@ -50,8 +50,10 @@ def create_block_causal_mask(
residue_len = max_seq_len - torch.sum(seq_lens[sample_idx])
block_attn_masks.append(
torch.zeros(
residue_len, residue_len, dtype=torch.bool, device=seq_lens[0][0].device
torch.tril(
torch.zeros(
residue_len, residue_len, dtype=torch.bool, device=seq_lens[sample_idx].device
)
)
)

View File

@@ -182,7 +182,7 @@ class FlexBatchSamplerDataCollatorForSeq2Seq(DataCollatorForSeq2Seq):
collated_seq_lens, max_seq_len = get_seqlens_from_pos_ids(out["position_ids"])
# out["attention_mask"] = packed_block_causal_mask(collated_seq_lens)
out["attention_mask"] = create_block_causal_mask(collated_seq_lens, max_seq_len)
raise ValueError(f"{out['attention_mask'].shape}")
# raise ValueError(f"{out['attention_mask'].shape}")
return out