Compare commits

...

1 Commits

Author SHA1 Message Date
Wing Lian
f608d263a6 configurable weight scale normalization for MoE expert drift 2026-04-09 15:37:16 +00:00
4 changed files with 177 additions and 0 deletions

View File

@@ -137,6 +137,50 @@ This means the policy has diverged significantly from the weights used by vLLM f
- Increase `gradient_accumulation_steps` to smooth out noisy batches.
- Check for NaN issues (see next section).
## MoE Weight Scale Drift
**Symptom**: Model works on short prompts but loses coherence on long conversations — repeating itself, "philosophizing", or generating broken code. Particularly affects MoE models with recurrent/SSM components (e.g. DeltaNet linear attention).
**Root cause**: In MoE models trained with AdamW, rarely-activated experts accumulate smaller second-moment estimates. This gives them a disproportionately large effective learning rate, causing their weights to drift to higher variance than the group norm. In recurrent components like `conv1d` in DeltaNet layers, this amplifies short-range context and washes out long-range state.
**Detection**: Use `normalize_weight_scales` with `dry_run: true` to scan for anomalies without modifying weights:
```yaml
normalize_weight_scales:
- name_pattern: 'linear_attn\.conv1d\.weight'
threshold: 1.3
dry_run: true
```
This logs any tensors matching the pattern whose standard deviation exceeds 1.3x the group median. Example output:
```
normalize_weight_scales [DRY RUN]: pattern 'linear_attn\.conv1d\.weight' —
3/30 tensors outside 1.3x threshold (median std=0.062733):
layers.36.linear_attn.conv1d.weight: std=0.101870 (1.62x median)
layers.37.linear_attn.conv1d.weight: std=0.102362 (1.63x median)
layers.38.linear_attn.conv1d.weight: std=0.089227 (1.42x median)
```
Each rule accepts:
- `name_pattern`: regex matched against parameter names. All matching tensors form a group.
- `threshold`: flag tensors whose std deviates from the group median by more than this factor (default: 1.5).
- `dry_run`: when `true`, log anomalies without modifying weights (default: `false`).
Multiple rules can target different tensor patterns:
```yaml
normalize_weight_scales:
- name_pattern: 'linear_attn\.conv1d\.weight'
threshold: 1.3
- name_pattern: 'experts\.gate_up_proj'
threshold: 1.5
dry_run: true # just check these, don't fix
```
The transform runs after model loading but before adapter injection, so it modifies the base model weights directly.
## NaN and Inf Handling
### Common Causes

View File

@@ -38,6 +38,7 @@ from axolotl.utils.dict import DictDefault
from axolotl.utils.distributed import cleanup_distributed
from axolotl.utils.freeze import freeze_layers_except
from axolotl.utils.logging import get_logger
from axolotl.utils.normalize_weights import normalize_weight_scales
from axolotl.utils.schemas.enums import RLType
from axolotl.utils.train import determine_last_checkpoint
from axolotl.utils.trainer import setup_trainer
@@ -105,6 +106,10 @@ def setup_model_and_tokenizer(
event_type="peft-config-load", properties=peft_config.to_dict()
)
# Normalize weight scales for MoE/hybrid models with drifted expert weights
if cfg.normalize_weight_scales:
normalize_weight_scales(model, cfg.normalize_weight_scales)
# Apply freezing if specified
if cfg.unfrozen_parameters:
freeze_layers_except(model, cfg.unfrozen_parameters)

View File

@@ -0,0 +1,115 @@
"""
Detect and fix weight scale anomalies in MoE/hybrid models.
In MoE models trained with AdamW, rarely-activated experts accumulate smaller
second-moment estimates, giving them a disproportionately large effective
learning rate. Over time this causes their weights to drift to higher
variance than the median for the same tensor across layers.
For recurrent / SSM / DeltaNet components (e.g. ``conv1d.weight`` in linear
attention layers), this drift corrupts the hidden state and degrades long-
context performance — the model "forgets" after a few tokens.
This module provides a configurable transform that detects outlier weight
scales per-tensor-pattern and rescales them to the group median.
"""
import re
from collections import defaultdict
import torch
from axolotl.utils.distributed import is_main_process
from axolotl.utils.logging import get_logger
LOG = get_logger(__name__)
def normalize_weight_scales(model, rules):
"""Normalize weight scales for tensor groups that have outlier variance.
Parameters
----------
model : torch.nn.Module
The loaded model (before adapter injection).
rules : list[dict]
Each rule is a dict with keys:
- ``name_pattern`` (str): regex matched against each named parameter.
Parameters that match are grouped together, and outliers within the
group are rescaled.
- ``threshold`` (float, default 1.5): a parameter is flagged when its
std deviates from the group median by more than this factor
(ratio > threshold or ratio < 1/threshold).
- ``dry_run`` (bool, default False): when True, log anomalies but do
not modify weights.
Returns
-------
int
Number of tensors that were rescaled (0 in dry-run mode).
"""
total_fixed = 0
for rule in rules:
pattern = rule.get("name_pattern")
if not pattern:
LOG.warning("normalize_weight_scales: rule missing 'name_pattern', skipping")
continue
threshold = float(rule.get("threshold", 1.5))
dry_run = bool(rule.get("dry_run", False))
regex = re.compile(pattern)
# Collect matching tensors
matches = []
for name, param in model.named_parameters():
if regex.search(name):
with torch.no_grad():
std = param.data.float().std().item()
matches.append((name, param, std))
if len(matches) < 3:
if is_main_process():
LOG.info(
f"normalize_weight_scales: pattern '{pattern}' matched "
f"{len(matches)} tensors (need >=3 to detect outliers), skipping"
)
continue
# Compute group median std
stds = [s for _, _, s in matches]
median_std = float(sorted(stds)[len(stds) // 2])
if median_std < 1e-10:
continue
# Detect and fix outliers
outliers = []
for name, param, std in matches:
ratio = std / median_std
if ratio > threshold or ratio < (1.0 / threshold):
outliers.append((name, param, std, ratio))
if not dry_run:
scale_factor = median_std / std
param.data.mul_(scale_factor)
total_fixed += 1
# Report
if is_main_process() and outliers:
mode = "DRY RUN" if dry_run else "FIXED"
LOG.warning(
f"normalize_weight_scales [{mode}]: pattern '{pattern}'"
f"{len(outliers)}/{len(matches)} tensors outside "
f"{threshold:.1f}x threshold (median std={median_std:.6f}):"
)
for name, _, std, ratio in outliers:
LOG.warning(f" {name}: std={std:.6f} ({ratio:.2f}x median)")
elif is_main_process():
LOG.info(
f"normalize_weight_scales: pattern '{pattern}'"
f"{len(matches)} tensors, all within {threshold:.1f}x threshold "
f"(median std={median_std:.6f})"
)
return total_fixed

View File

@@ -578,6 +578,19 @@ class AxolotlInputConfig(
},
)
normalize_weight_scales: list[dict] | None = Field(
default=None,
json_schema_extra={
"description": "Detect and rescale outlier weight tensors caused by AdamW + rare "
"MoE expert drift. Each entry is a rule with: "
"'name_pattern' (regex matching parameter names to group), "
"'threshold' (float, default 1.5 — flag tensors whose std deviates from the "
"group median by more than this factor), "
"'dry_run' (bool, default false — log anomalies without modifying weights). "
"Example: [{name_pattern: 'linear_attn\\.conv1d\\.weight', threshold: 1.3}]"
},
)
unfrozen_parameters: list[str] | None = Field(
default=None,
json_schema_extra={