Compare commits
76 Commits
colab-misc
...
flx_attn_s
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
328bb0466b | ||
|
|
e792b54bab | ||
|
|
82d04ea060 | ||
|
|
0ef1f011fe | ||
|
|
c0a1d205c7 | ||
|
|
d0e739da24 | ||
|
|
3f6be519d5 | ||
|
|
adcbc7459b | ||
|
|
470ba65c44 | ||
|
|
8e1adc154d | ||
|
|
e5b36900e4 | ||
|
|
9f6c89b12b | ||
|
|
b0871c8d3b | ||
|
|
d3ea379a23 | ||
|
|
0ebab63309 | ||
|
|
e98581f6f5 | ||
|
|
b832b11c8f | ||
|
|
b692d394b1 | ||
|
|
2319e5276d | ||
|
|
9a43a0925d | ||
|
|
10de67e8ea | ||
|
|
fa7355404c | ||
|
|
907424a2e8 | ||
|
|
3f4fd3c1eb | ||
|
|
48c3c47071 | ||
|
|
3ed9c117fb | ||
|
|
84960003ed | ||
|
|
93a268e43d | ||
|
|
065f6d477e | ||
|
|
96ad741cd5 | ||
|
|
ba88bc7840 | ||
|
|
b31796a681 | ||
|
|
5ca57cb55a | ||
|
|
0149de7fb0 | ||
|
|
8c34c65181 | ||
|
|
555aa5772a | ||
|
|
e8b2789086 | ||
|
|
85752cdfc9 | ||
|
|
f2f23c8041 | ||
|
|
8b3eec7f6e | ||
|
|
bb9bea3110 | ||
|
|
0dd18a3681 | ||
|
|
152e988d3c | ||
|
|
27532825a9 | ||
|
|
06f83a54a5 | ||
|
|
d7b133dc1f | ||
|
|
f3bec17917 | ||
|
|
b7deb5241c | ||
|
|
cee310dcfa | ||
|
|
d1be6e228d | ||
|
|
5f9f77f384 | ||
|
|
b2a34380b3 | ||
|
|
80bfc50d1f | ||
|
|
a5360c172c | ||
|
|
013a9b73fc | ||
|
|
aad62428e0 | ||
|
|
a6f2c5d583 | ||
|
|
dbcd11e533 | ||
|
|
c06a6be915 | ||
|
|
d3a0cb5edb | ||
|
|
8b47e456b0 | ||
|
|
2319ac729c | ||
|
|
f99cae0e7b | ||
|
|
888cd9407f | ||
|
|
bd62d6e10a | ||
|
|
5eae134110 | ||
|
|
b7d27bdfa4 | ||
|
|
da97a21bdc | ||
|
|
e0d4b88598 | ||
|
|
fac059a209 | ||
|
|
9c9ac1cf0b | ||
|
|
2346f21b2b | ||
|
|
0b47281f51 | ||
|
|
543daaf46f | ||
|
|
bcd9ad44e0 | ||
|
|
61ad375bf4 |
@@ -831,7 +831,9 @@ class HFCausalTrainerBuilder(TrainerBuilderBase):
|
||||
if "max_length" in kwargs:
|
||||
kwargs.pop("max_length")
|
||||
elif use_batch_sampler_collator:
|
||||
if self.cfg.model_config_type in SUPPORTED_MULTIPACK_MODEL_TYPES:
|
||||
if self.cfg.flex_attention is True:
|
||||
collator = V2BatchSamplerDataCollatorForSeq2Seq
|
||||
elif self.cfg.model_config_type in SUPPORTED_MULTIPACK_MODEL_TYPES:
|
||||
collator = V2BatchSamplerDataCollatorForSeq2Seq
|
||||
elif (
|
||||
self.cfg.model_config_type in ["llama"]
|
||||
|
||||
@@ -95,6 +95,103 @@ def get_cu_seqlens(attn_mask):
|
||||
return torch.stack(results).to(dtype=torch.int32), torch.stack(max_seq_lens)
|
||||
|
||||
|
||||
def get_packed_mask_from_pos_ids(position_ids):
|
||||
if len(position_ids.shape) == 1:
|
||||
position_ids = position_ids.unsqueeze(0)
|
||||
|
||||
device = position_ids.device
|
||||
results = []
|
||||
|
||||
for i, row in enumerate(position_ids):
|
||||
# Count the number of consecutive zeros from the right side
|
||||
padding_length = (row == 0).int().flip(dims=[0]).cumprod(dim=0).sum().item()
|
||||
|
||||
# Adjust the row to exclude padding
|
||||
adjusted_row = row[:-padding_length] if padding_length else row.clone()
|
||||
|
||||
# Find where the position resets to 0 (indicating a new sequence)
|
||||
seq_starts = torch.cat(
|
||||
[
|
||||
torch.tensor([True], dtype=torch.bool, device=device),
|
||||
adjusted_row[1:] == 0,
|
||||
]
|
||||
)
|
||||
# Get the indices where the sequence starts
|
||||
start_indices = torch.cat(
|
||||
[
|
||||
torch.nonzero(seq_starts).unbind(dim=1)[0],
|
||||
torch.tensor([len(adjusted_row)], dtype=torch.int32, device=device),
|
||||
]
|
||||
)
|
||||
# Calculate the sequence lengths
|
||||
seq_lengths = start_indices[1:] - start_indices[:-1]
|
||||
# Append the padding length to the sequence lengths
|
||||
doc_mask = torch.ones(len(row), dtype=torch.int32, device=device)
|
||||
for i, seq_len in enumerate(seq_lengths):
|
||||
start_id = start_indices[i]
|
||||
doc_mask[start_id : start_id + seq_len] = (
|
||||
(i+1) * doc_mask[start_id : start_id + seq_len]
|
||||
)
|
||||
if padding_length:
|
||||
doc_mask[len(adjusted_row) :] = 0 * doc_mask[len(adjusted_row) :]
|
||||
|
||||
results.append(doc_mask)
|
||||
|
||||
return torch.stack(results)
|
||||
|
||||
|
||||
def get_seqlens_from_pos_ids(position_ids):
|
||||
"""generate a sequence length set using pos ids for doc mask creation in flex attention"""
|
||||
if len(position_ids.shape) == 1:
|
||||
position_ids = position_ids.unsqueeze(0)
|
||||
max_seq_len = position_ids.shape[1]
|
||||
|
||||
device = position_ids.device
|
||||
results = []
|
||||
totalseqlens = []
|
||||
|
||||
for row in position_ids:
|
||||
# Count the number of consecutive zeros from the right side
|
||||
padding_length = (row == 0).int().flip(dims=[0]).cumprod(dim=0).sum().item()
|
||||
|
||||
# Adjust the row to exclude padding
|
||||
adjusted_row = row[:-padding_length] if padding_length else row.clone()
|
||||
|
||||
# Find where the position resets to 0 (indicating a new sequence)
|
||||
seq_starts = torch.cat(
|
||||
[
|
||||
torch.tensor([True], dtype=torch.bool, device=device),
|
||||
adjusted_row[1:] == 0,
|
||||
]
|
||||
)
|
||||
# Get the indices where the sequence starts
|
||||
start_indices = torch.cat(
|
||||
[
|
||||
torch.nonzero(seq_starts).unbind(dim=1)[0],
|
||||
torch.tensor([len(adjusted_row)], dtype=torch.int32, device=device),
|
||||
]
|
||||
)
|
||||
# Calculate the sequence lengths
|
||||
seq_lengths = start_indices[1:] - start_indices[:-1]
|
||||
# Append the padding length to the sequence lengths
|
||||
if padding_length:
|
||||
seq_lengths = torch.cat(
|
||||
[
|
||||
seq_lengths,
|
||||
torch.tensor(
|
||||
[len(row) - torch.sum(seq_lengths)],
|
||||
dtype=torch.int32,
|
||||
device=device,
|
||||
),
|
||||
]
|
||||
)
|
||||
|
||||
results.append(seq_lengths)
|
||||
totalseqlens.append(len(adjusted_row))
|
||||
|
||||
return results, torch.tensor(totalseqlens, dtype=torch.int32, device=device)
|
||||
|
||||
|
||||
def get_cu_seqlens_from_pos_ids(position_ids):
|
||||
"""generate a cumulative sequence length mask for flash attention using pos ids"""
|
||||
if len(position_ids.shape) == 1:
|
||||
@@ -176,7 +273,10 @@ def mask_2d_to_4d(
|
||||
when they attend to each other within that sequence.
|
||||
This expansion transforms the mask to lower triangular form to prevent future peeking.
|
||||
"""
|
||||
bsz, src_len = mask.size()
|
||||
|
||||
if len(mask.size()) == 4:
|
||||
return mask
|
||||
bsz, src_len = int(mask.size()[0]), int(mask.size()[1])
|
||||
tgt_len = tgt_len if tgt_len is not None else src_len
|
||||
|
||||
mask = mask.unsqueeze(1).unsqueeze(2)
|
||||
|
||||
@@ -823,6 +823,7 @@ class AxolotlInputConfig(
|
||||
xformers_attention: Optional[bool] = None
|
||||
sdp_attention: Optional[bool] = None
|
||||
s2_attention: Optional[bool] = None
|
||||
flex_attention: Optional[bool] = None
|
||||
flash_attention: Optional[bool] = None
|
||||
flash_attn_cross_entropy: Optional[bool] = None
|
||||
flash_attn_rms_norm: Optional[bool] = None
|
||||
@@ -1789,6 +1790,26 @@ class AxolotlConfigWCapabilities(AxolotlInputConfig):
|
||||
)
|
||||
return data
|
||||
|
||||
@model_validator(mode="before")
|
||||
@classmethod
|
||||
def check_flex_torch_version(cls, data):
|
||||
if (data.get("flex_attention") is not None) and (
|
||||
data.get("flex_attention") is True
|
||||
):
|
||||
env_capabilities = data.get("env_capabilities", {})
|
||||
torch_version = env_capabilities.get("torch_version")
|
||||
|
||||
if torch_version is None:
|
||||
import torch
|
||||
|
||||
torch_version = str(torch.__version__).split("+", maxsplit=1)[0]
|
||||
|
||||
if version.parse(torch_version) < version.parse("2.5.1"):
|
||||
raise ValueError(
|
||||
"Flex attention is not supported on torch version < 2.5.1"
|
||||
)
|
||||
return data
|
||||
|
||||
@model_validator(mode="before")
|
||||
@classmethod
|
||||
def check_torch_compile_auto(cls, data):
|
||||
|
||||
@@ -403,7 +403,7 @@ class ModelLoader:
|
||||
|
||||
if (
|
||||
self.cfg.model_config_type in SUPPORTED_MULTIPACK_MODEL_TYPES
|
||||
and self.cfg.flash_attention
|
||||
and (self.cfg.flash_attention or self.cfg.flex_attention)
|
||||
and self.cfg.sample_packing
|
||||
):
|
||||
if "auto_map" in self.model_config:
|
||||
@@ -707,7 +707,13 @@ class ModelLoader:
|
||||
"""
|
||||
sample packing uses custom FA2 patch
|
||||
"""
|
||||
if self.cfg.flash_attention:
|
||||
|
||||
if self.cfg.flex_attention:
|
||||
self.model_kwargs["attn_implementation"] = "flex_attention"
|
||||
self.model_config._attn_implementation = ( # pylint: disable=protected-access
|
||||
"flex_attention"
|
||||
)
|
||||
elif self.cfg.flash_attention:
|
||||
if not self.cfg.sample_packing and self.cfg.s2_attention:
|
||||
pass
|
||||
self.model_kwargs["attn_implementation"] = "flash_attention_2"
|
||||
@@ -1113,7 +1119,7 @@ class ModelLoader:
|
||||
should_convert = (
|
||||
# LlamaRMSNorm layers are in fp32 after kbit_training or full finetune, so we need to
|
||||
# convert them back to fp16/bf16 for flash-attn compatibility.
|
||||
((needs_fa2_dtype or self.cfg.flash_attention) and not qlora_fsdp)
|
||||
((needs_fa2_dtype or self.cfg.flash_attention or self.cfg.flex_attention) and not qlora_fsdp)
|
||||
or self.cfg.cut_cross_entropy # Cut cross entropy requires embedding layers to be in fp16/bf16 for backward pass
|
||||
)
|
||||
|
||||
|
||||
Reference in New Issue
Block a user