Files
axolotl/tests/utils/data/test_hash.py
Edward Zion Saji 55a7950e3d fix: DPO tool role KeyError (#3217), dataset hash output_dir (#3303), config validators (#3538) [skip ci]
* fix: DPO tool role KeyError, dataset hash output_dir, config validators [skip-e2e]

- Add 'tool' to default role_map_inv in dpo/chat_template.py default() and
  argilla_chat() so datasets with tool-call messages no longer raise
  KeyError: 'tool' (closes #3217)

- Fix generate_dataset_hash_from_config to use canonical tokenizer config +
  overrides content instead of tokenizer.name_or_path when added_tokens_overrides
  is set, preventing cache busting when only output_dir changes (closes #3303)

- Add three Pydantic config validators to AxolotlConfigWCapabilities:
  * save_strategy: 'best' requires metric_for_best_model
  * streaming=True is incompatible with val_set_size > 0
  * lora_target_modules list entries must be valid Python regex patterns

- Tests for all three changes

* review: condense comment in shared.py, swap Mistral model for SmolLM2-135M in test_hash

* chore: lint

* move the validators out of the w/ capabilities schema

---------

Co-authored-by: Wing Lian <wing@axolotl.ai>
2026-04-01 19:57:07 -04:00

136 lines
4.8 KiB
Python

"""
Tests for generate_dataset_hash_from_config.
Regression test for https://github.com/axolotl-ai-cloud/axolotl/issues/3303:
changing output_dir should not bust the dataset cache when added_tokens_overrides
is set.
"""
from axolotl.utils.data.shared import generate_dataset_hash_from_config
from axolotl.utils.dict import DictDefault
def _base_cfg(**kwargs):
return DictDefault(
{
"sequence_len": 2048,
"sample_packing": False,
"eval_sample_packing": False,
"group_by_length": False,
"kd_temperature": None,
"dataset_exact_deduplication": False,
"tokenizer_config": "NousResearch/Llama-3.2-1B",
**kwargs,
}
)
def _datasets():
return [
DictDefault(
{
"path": "mhenrichsen/alpaca_2k_test",
"type": "alpaca",
"shards": None,
"conversation": None,
"split": "train",
"temperature": None,
}
)
]
class TestGenerateDatasetHashFromConfig:
def test_same_config_same_hash(self):
"""Identical configs produce identical hashes."""
cfg = _base_cfg()
h1 = generate_dataset_hash_from_config(
cfg, _datasets(), "NousResearch/Llama-3.2-1B"
)
h2 = generate_dataset_hash_from_config(
cfg, _datasets(), "NousResearch/Llama-3.2-1B"
)
assert h1 == h2
def test_different_tokenizer_different_hash(self):
"""A different tokenizer path produces a different hash."""
cfg = _base_cfg()
h1 = generate_dataset_hash_from_config(
cfg, _datasets(), "NousResearch/Llama-3.2-1B"
)
h2 = generate_dataset_hash_from_config(
cfg, _datasets(), "HuggingFaceTB/SmolLM2-135M"
)
assert h1 != h2
def test_different_sequence_len_different_hash(self):
cfg_a = _base_cfg(sequence_len=2048)
cfg_b = _base_cfg(sequence_len=4096)
h1 = generate_dataset_hash_from_config(cfg_a, _datasets(), "tok")
h2 = generate_dataset_hash_from_config(cfg_b, _datasets(), "tok")
assert h1 != h2
# --- Regression: added_tokens_overrides + output_dir ---
def test_added_tokens_overrides_hash_stable_across_output_dir(self):
"""Hash must not change when only output_dir changes (issue #3303).
When added_tokens_overrides is set the tokenizer is saved into output_dir,
making tokenizer.name_or_path an absolute path that includes output_dir.
The hash should be derived from the canonical tokenizer config + overrides,
not from the output-dir-dependent path.
"""
cfg_run1 = _base_cfg(
output_dir="/tmp/run_1",
added_tokens_overrides={32000: "<PAD>", 32001: "<MASK>"},
)
cfg_run2 = _base_cfg(
output_dir="/tmp/run_2_different_name",
added_tokens_overrides={32000: "<PAD>", 32001: "<MASK>"},
)
# Simulate what happens in practice: tokenizer.name_or_path becomes the
# output_dir-based path after modify_tokenizer_files() saves the tokenizer.
tokenizer_name_run1 = "/tmp/run_1/modified_tokenizer"
tokenizer_name_run2 = "/tmp/run_2_different_name/modified_tokenizer"
h1 = generate_dataset_hash_from_config(
cfg_run1, _datasets(), tokenizer_name_run1
)
h2 = generate_dataset_hash_from_config(
cfg_run2, _datasets(), tokenizer_name_run2
)
assert h1 == h2, (
"Dataset cache hash must not change when only output_dir changes "
"while added_tokens_overrides stays the same (issue #3303)."
)
def test_added_tokens_overrides_different_overrides_different_hash(self):
"""Different added_tokens_overrides produce different hashes."""
cfg_a = _base_cfg(
output_dir="/tmp/run_a",
added_tokens_overrides={32000: "<PAD>"},
)
cfg_b = _base_cfg(
output_dir="/tmp/run_a", # same output_dir
added_tokens_overrides={32000: "<OTHER>"},
)
tokenizer_path = "/tmp/run_a/modified_tokenizer"
h1 = generate_dataset_hash_from_config(cfg_a, _datasets(), tokenizer_path)
h2 = generate_dataset_hash_from_config(cfg_b, _datasets(), tokenizer_path)
assert h1 != h2
def test_no_added_tokens_overrides_uses_tokenizer_name_as_before(self):
"""Without added_tokens_overrides the old behaviour is preserved."""
cfg = _base_cfg() # no added_tokens_overrides
tokenizer_name = "NousResearch/Llama-3.2-1B"
h1 = generate_dataset_hash_from_config(cfg, _datasets(), tokenizer_name)
# Changing tokenizer_name still changes the hash
h2 = generate_dataset_hash_from_config(cfg, _datasets(), "some/other-model")
assert h1 != h2