more fixes 20240228 (#1342) [skip ci]

* add missing evals_per_epoch setting

* more pydantic fixes

* more fixes

* move test from normalization to validation

* increase eval size for sample packing tests
This commit is contained in:
Wing Lian
2024-02-28 12:57:45 -05:00
committed by GitHub
parent c1a7b3dd69
commit 0f985e12fe
7 changed files with 28 additions and 20 deletions

View File

@@ -43,7 +43,7 @@ class TestLoraLlama(unittest.TestCase):
"lora_alpha": 64,
"lora_dropout": 0.05,
"lora_target_linear": True,
"val_set_size": 0.1,
"val_set_size": 0.2,
"special_tokens": {
"unk_token": "<unk>",
"bos_token": "<s>",

View File

@@ -25,20 +25,6 @@ class NormalizeConfigTestCase(unittest.TestCase):
}
)
def test_lr_as_float(self):
cfg = (
self._get_base_cfg()
| DictDefault( # pylint: disable=unsupported-binary-operation
{
"learning_rate": "5e-5",
}
)
)
normalize_config(cfg)
assert cfg.learning_rate == 0.00005
def test_base_model_config_set_when_empty(self):
cfg = self._get_base_cfg()
del cfg.base_model_config

View File

@@ -176,6 +176,20 @@ class TestValidation(BaseValidation):
with pytest.raises(ValueError, match=r".*At least two of*"):
validate_config(cfg)
def test_lr_as_float(self, minimal_cfg):
cfg = (
DictDefault( # pylint: disable=unsupported-binary-operation
{
"learning_rate": "5e-5",
}
)
| minimal_cfg
)
new_cfg = validate_config(cfg)
assert new_cfg.learning_rate == 0.00005
def test_qlora(self, minimal_cfg):
base_cfg = (
DictDefault(