Compare commits
1 Commits
attn-imple
...
weight-sca
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
f608d263a6 |
@@ -137,6 +137,50 @@ This means the policy has diverged significantly from the weights used by vLLM f
|
|||||||
- Increase `gradient_accumulation_steps` to smooth out noisy batches.
|
- Increase `gradient_accumulation_steps` to smooth out noisy batches.
|
||||||
- Check for NaN issues (see next section).
|
- Check for NaN issues (see next section).
|
||||||
|
|
||||||
|
## MoE Weight Scale Drift
|
||||||
|
|
||||||
|
**Symptom**: Model works on short prompts but loses coherence on long conversations — repeating itself, "philosophizing", or generating broken code. Particularly affects MoE models with recurrent/SSM components (e.g. DeltaNet linear attention).
|
||||||
|
|
||||||
|
**Root cause**: In MoE models trained with AdamW, rarely-activated experts accumulate smaller second-moment estimates. This gives them a disproportionately large effective learning rate, causing their weights to drift to higher variance than the group norm. In recurrent components like `conv1d` in DeltaNet layers, this amplifies short-range context and washes out long-range state.
|
||||||
|
|
||||||
|
**Detection**: Use `normalize_weight_scales` with `dry_run: true` to scan for anomalies without modifying weights:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
normalize_weight_scales:
|
||||||
|
- name_pattern: 'linear_attn\.conv1d\.weight'
|
||||||
|
threshold: 1.3
|
||||||
|
dry_run: true
|
||||||
|
```
|
||||||
|
|
||||||
|
This logs any tensors matching the pattern whose standard deviation exceeds 1.3x the group median. Example output:
|
||||||
|
|
||||||
|
```
|
||||||
|
normalize_weight_scales [DRY RUN]: pattern 'linear_attn\.conv1d\.weight' —
|
||||||
|
3/30 tensors outside 1.3x threshold (median std=0.062733):
|
||||||
|
layers.36.linear_attn.conv1d.weight: std=0.101870 (1.62x median)
|
||||||
|
layers.37.linear_attn.conv1d.weight: std=0.102362 (1.63x median)
|
||||||
|
layers.38.linear_attn.conv1d.weight: std=0.089227 (1.42x median)
|
||||||
|
```
|
||||||
|
|
||||||
|
Each rule accepts:
|
||||||
|
|
||||||
|
- `name_pattern`: regex matched against parameter names. All matching tensors form a group.
|
||||||
|
- `threshold`: flag tensors whose std deviates from the group median by more than this factor (default: 1.5).
|
||||||
|
- `dry_run`: when `true`, log anomalies without modifying weights (default: `false`).
|
||||||
|
|
||||||
|
Multiple rules can target different tensor patterns:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
normalize_weight_scales:
|
||||||
|
- name_pattern: 'linear_attn\.conv1d\.weight'
|
||||||
|
threshold: 1.3
|
||||||
|
- name_pattern: 'experts\.gate_up_proj'
|
||||||
|
threshold: 1.5
|
||||||
|
dry_run: true # just check these, don't fix
|
||||||
|
```
|
||||||
|
|
||||||
|
The transform runs after model loading but before adapter injection, so it modifies the base model weights directly.
|
||||||
|
|
||||||
## NaN and Inf Handling
|
## NaN and Inf Handling
|
||||||
|
|
||||||
### Common Causes
|
### Common Causes
|
||||||
|
|||||||
@@ -38,6 +38,7 @@ from axolotl.utils.dict import DictDefault
|
|||||||
from axolotl.utils.distributed import cleanup_distributed
|
from axolotl.utils.distributed import cleanup_distributed
|
||||||
from axolotl.utils.freeze import freeze_layers_except
|
from axolotl.utils.freeze import freeze_layers_except
|
||||||
from axolotl.utils.logging import get_logger
|
from axolotl.utils.logging import get_logger
|
||||||
|
from axolotl.utils.normalize_weights import normalize_weight_scales
|
||||||
from axolotl.utils.schemas.enums import RLType
|
from axolotl.utils.schemas.enums import RLType
|
||||||
from axolotl.utils.train import determine_last_checkpoint
|
from axolotl.utils.train import determine_last_checkpoint
|
||||||
from axolotl.utils.trainer import setup_trainer
|
from axolotl.utils.trainer import setup_trainer
|
||||||
@@ -105,6 +106,10 @@ def setup_model_and_tokenizer(
|
|||||||
event_type="peft-config-load", properties=peft_config.to_dict()
|
event_type="peft-config-load", properties=peft_config.to_dict()
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# Normalize weight scales for MoE/hybrid models with drifted expert weights
|
||||||
|
if cfg.normalize_weight_scales:
|
||||||
|
normalize_weight_scales(model, cfg.normalize_weight_scales)
|
||||||
|
|
||||||
# Apply freezing if specified
|
# Apply freezing if specified
|
||||||
if cfg.unfrozen_parameters:
|
if cfg.unfrozen_parameters:
|
||||||
freeze_layers_except(model, cfg.unfrozen_parameters)
|
freeze_layers_except(model, cfg.unfrozen_parameters)
|
||||||
|
|||||||
115
src/axolotl/utils/normalize_weights.py
Normal file
115
src/axolotl/utils/normalize_weights.py
Normal file
@@ -0,0 +1,115 @@
|
|||||||
|
"""
|
||||||
|
Detect and fix weight scale anomalies in MoE/hybrid models.
|
||||||
|
|
||||||
|
In MoE models trained with AdamW, rarely-activated experts accumulate smaller
|
||||||
|
second-moment estimates, giving them a disproportionately large effective
|
||||||
|
learning rate. Over time this causes their weights to drift to higher
|
||||||
|
variance than the median for the same tensor across layers.
|
||||||
|
|
||||||
|
For recurrent / SSM / DeltaNet components (e.g. ``conv1d.weight`` in linear
|
||||||
|
attention layers), this drift corrupts the hidden state and degrades long-
|
||||||
|
context performance — the model "forgets" after a few tokens.
|
||||||
|
|
||||||
|
This module provides a configurable transform that detects outlier weight
|
||||||
|
scales per-tensor-pattern and rescales them to the group median.
|
||||||
|
"""
|
||||||
|
|
||||||
|
import re
|
||||||
|
from collections import defaultdict
|
||||||
|
|
||||||
|
import torch
|
||||||
|
|
||||||
|
from axolotl.utils.distributed import is_main_process
|
||||||
|
from axolotl.utils.logging import get_logger
|
||||||
|
|
||||||
|
LOG = get_logger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def normalize_weight_scales(model, rules):
|
||||||
|
"""Normalize weight scales for tensor groups that have outlier variance.
|
||||||
|
|
||||||
|
Parameters
|
||||||
|
----------
|
||||||
|
model : torch.nn.Module
|
||||||
|
The loaded model (before adapter injection).
|
||||||
|
rules : list[dict]
|
||||||
|
Each rule is a dict with keys:
|
||||||
|
|
||||||
|
- ``name_pattern`` (str): regex matched against each named parameter.
|
||||||
|
Parameters that match are grouped together, and outliers within the
|
||||||
|
group are rescaled.
|
||||||
|
- ``threshold`` (float, default 1.5): a parameter is flagged when its
|
||||||
|
std deviates from the group median by more than this factor
|
||||||
|
(ratio > threshold or ratio < 1/threshold).
|
||||||
|
- ``dry_run`` (bool, default False): when True, log anomalies but do
|
||||||
|
not modify weights.
|
||||||
|
|
||||||
|
Returns
|
||||||
|
-------
|
||||||
|
int
|
||||||
|
Number of tensors that were rescaled (0 in dry-run mode).
|
||||||
|
"""
|
||||||
|
total_fixed = 0
|
||||||
|
|
||||||
|
for rule in rules:
|
||||||
|
pattern = rule.get("name_pattern")
|
||||||
|
if not pattern:
|
||||||
|
LOG.warning("normalize_weight_scales: rule missing 'name_pattern', skipping")
|
||||||
|
continue
|
||||||
|
|
||||||
|
threshold = float(rule.get("threshold", 1.5))
|
||||||
|
dry_run = bool(rule.get("dry_run", False))
|
||||||
|
regex = re.compile(pattern)
|
||||||
|
|
||||||
|
# Collect matching tensors
|
||||||
|
matches = []
|
||||||
|
for name, param in model.named_parameters():
|
||||||
|
if regex.search(name):
|
||||||
|
with torch.no_grad():
|
||||||
|
std = param.data.float().std().item()
|
||||||
|
matches.append((name, param, std))
|
||||||
|
|
||||||
|
if len(matches) < 3:
|
||||||
|
if is_main_process():
|
||||||
|
LOG.info(
|
||||||
|
f"normalize_weight_scales: pattern '{pattern}' matched "
|
||||||
|
f"{len(matches)} tensors (need >=3 to detect outliers), skipping"
|
||||||
|
)
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Compute group median std
|
||||||
|
stds = [s for _, _, s in matches]
|
||||||
|
median_std = float(sorted(stds)[len(stds) // 2])
|
||||||
|
|
||||||
|
if median_std < 1e-10:
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Detect and fix outliers
|
||||||
|
outliers = []
|
||||||
|
for name, param, std in matches:
|
||||||
|
ratio = std / median_std
|
||||||
|
if ratio > threshold or ratio < (1.0 / threshold):
|
||||||
|
outliers.append((name, param, std, ratio))
|
||||||
|
if not dry_run:
|
||||||
|
scale_factor = median_std / std
|
||||||
|
param.data.mul_(scale_factor)
|
||||||
|
total_fixed += 1
|
||||||
|
|
||||||
|
# Report
|
||||||
|
if is_main_process() and outliers:
|
||||||
|
mode = "DRY RUN" if dry_run else "FIXED"
|
||||||
|
LOG.warning(
|
||||||
|
f"normalize_weight_scales [{mode}]: pattern '{pattern}' — "
|
||||||
|
f"{len(outliers)}/{len(matches)} tensors outside "
|
||||||
|
f"{threshold:.1f}x threshold (median std={median_std:.6f}):"
|
||||||
|
)
|
||||||
|
for name, _, std, ratio in outliers:
|
||||||
|
LOG.warning(f" {name}: std={std:.6f} ({ratio:.2f}x median)")
|
||||||
|
elif is_main_process():
|
||||||
|
LOG.info(
|
||||||
|
f"normalize_weight_scales: pattern '{pattern}' — "
|
||||||
|
f"{len(matches)} tensors, all within {threshold:.1f}x threshold "
|
||||||
|
f"(median std={median_std:.6f})"
|
||||||
|
)
|
||||||
|
|
||||||
|
return total_fixed
|
||||||
@@ -578,6 +578,19 @@ class AxolotlInputConfig(
|
|||||||
},
|
},
|
||||||
)
|
)
|
||||||
|
|
||||||
|
normalize_weight_scales: list[dict] | None = Field(
|
||||||
|
default=None,
|
||||||
|
json_schema_extra={
|
||||||
|
"description": "Detect and rescale outlier weight tensors caused by AdamW + rare "
|
||||||
|
"MoE expert drift. Each entry is a rule with: "
|
||||||
|
"'name_pattern' (regex matching parameter names to group), "
|
||||||
|
"'threshold' (float, default 1.5 — flag tensors whose std deviates from the "
|
||||||
|
"group median by more than this factor), "
|
||||||
|
"'dry_run' (bool, default false — log anomalies without modifying weights). "
|
||||||
|
"Example: [{name_pattern: 'linear_attn\\.conv1d\\.weight', threshold: 1.3}]"
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
unfrozen_parameters: list[str] | None = Field(
|
unfrozen_parameters: list[str] | None = Field(
|
||||||
default=None,
|
default=None,
|
||||||
json_schema_extra={
|
json_schema_extra={
|
||||||
|
|||||||
Reference in New Issue
Block a user