diff --git a/tests/e2e/multigpu/test_llama.py b/tests/e2e/multigpu/test_llama.py index f00c54bad..758491d63 100644 --- a/tests/e2e/multigpu/test_llama.py +++ b/tests/e2e/multigpu/test_llama.py @@ -114,7 +114,7 @@ class TestMultiGPULlama: "lora_alpha": 16, "lora_dropout": 0.05, "lora_target_linear": True, - "val_set_size": 0.01, + "val_set_size": 0.05, "special_tokens": { "pad_token": "<|endoftext|>", }, @@ -122,6 +122,7 @@ class TestMultiGPULlama: { "path": "tatsu-lab/alpaca", "type": "alpaca", + "split": "train[:20%]", }, ], "num_epochs": 1, diff --git a/tests/e2e/multigpu/test_ray.py b/tests/e2e/multigpu/test_ray.py index 14b1c0a86..9be7c6f50 100644 --- a/tests/e2e/multigpu/test_ray.py +++ b/tests/e2e/multigpu/test_ray.py @@ -31,7 +31,7 @@ class TestMultiGPURay: cfg = DictDefault( { "base_model": "HuggingFaceTB/SmolLM2-135M", - "sequence_len": 2048, + "sequence_len": 1024, "adapter": "lora", "lora_r": 8, "lora_alpha": 16, @@ -94,8 +94,8 @@ class TestMultiGPURay: "base_model": "HuggingFaceTB/SmolLM2-135M", "sample_packing": True, "pad_to_sequence_len": True, - "sequence_len": 2048, - "val_set_size": 0.05, + "sequence_len": 1024, + "val_set_size": 0.01, "special_tokens": { "pad_token": "<|endoftext|>", },