diff --git a/tests/e2e/test_fused_llama.py b/tests/e2e/test_fused_llama.py index 73393de20..513df69f9 100644 --- a/tests/e2e/test_fused_llama.py +++ b/tests/e2e/test_fused_llama.py @@ -27,7 +27,7 @@ class TestFusedLlama(unittest.TestCase): """ @with_temp_dir - def test_fft_packing(self, output_dir): + def test_fft_packing(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault( { @@ -52,7 +52,7 @@ class TestFusedLlama(unittest.TestCase): "num_epochs": 2, "micro_batch_size": 2, "gradient_accumulation_steps": 1, - "output_dir": output_dir, + "output_dir": temp_dir, "learning_rate": 0.00001, "optimizer": "adamw_torch", "lr_scheduler": "cosine", @@ -70,4 +70,4 @@ class TestFusedLlama(unittest.TestCase): dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta) - assert (Path(output_dir) / "pytorch_model.bin").exists() + assert (Path(temp_dir) / "pytorch_model.bin").exists() diff --git a/tests/e2e/test_lora_llama.py b/tests/e2e/test_lora_llama.py index 730020a79..c13243dd8 100644 --- a/tests/e2e/test_lora_llama.py +++ b/tests/e2e/test_lora_llama.py @@ -25,7 +25,7 @@ class TestLoraLlama(unittest.TestCase): """ @with_temp_dir - def test_lora(self, output_dir): + def test_lora(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault( { @@ -53,7 +53,7 @@ class TestLoraLlama(unittest.TestCase): "num_epochs": 2, "micro_batch_size": 8, "gradient_accumulation_steps": 1, - "output_dir": output_dir, + "output_dir": temp_dir, "learning_rate": 0.00001, "optimizer": "adamw_torch", "lr_scheduler": "cosine", @@ -64,10 +64,10 @@ class TestLoraLlama(unittest.TestCase): dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta) - assert (Path(output_dir) / "adapter_model.bin").exists() + assert (Path(temp_dir) / "adapter_model.bin").exists() @with_temp_dir - def test_lora_packing(self, output_dir): + def test_lora_packing(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault( { @@ -97,7 +97,7 @@ class TestLoraLlama(unittest.TestCase): "num_epochs": 2, "micro_batch_size": 8, "gradient_accumulation_steps": 1, - "output_dir": output_dir, + "output_dir": temp_dir, "learning_rate": 0.00001, "optimizer": "adamw_torch", "lr_scheduler": "cosine", @@ -108,10 +108,10 @@ class TestLoraLlama(unittest.TestCase): dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta) - assert (Path(output_dir) / "adapter_model.bin").exists() + assert (Path(temp_dir) / "adapter_model.bin").exists() @with_temp_dir - def test_lora_gptq(self, output_dir): + def test_lora_gptq(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault( { @@ -145,7 +145,7 @@ class TestLoraLlama(unittest.TestCase): "save_steps": 0.5, "micro_batch_size": 8, "gradient_accumulation_steps": 1, - "output_dir": output_dir, + "output_dir": temp_dir, "learning_rate": 0.00001, "optimizer": "adamw_torch", "lr_scheduler": "cosine", @@ -156,4 +156,4 @@ class TestLoraLlama(unittest.TestCase): dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta) - assert (Path(output_dir) / "adapter_model.bin").exists() + assert (Path(temp_dir) / "adapter_model.bin").exists() diff --git a/tests/e2e/test_mistral.py b/tests/e2e/test_mistral.py index bab714750..57d85e51e 100644 --- a/tests/e2e/test_mistral.py +++ b/tests/e2e/test_mistral.py @@ -27,7 +27,7 @@ class TestMistral(unittest.TestCase): """ @with_temp_dir - def test_lora(self, output_dir): + def test_lora(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault( { @@ -55,7 +55,7 @@ class TestMistral(unittest.TestCase): "num_epochs": 2, "micro_batch_size": 2, "gradient_accumulation_steps": 1, - "output_dir": output_dir, + "output_dir": temp_dir, "learning_rate": 0.00001, "optimizer": "adamw_torch", "lr_scheduler": "cosine", @@ -69,10 +69,10 @@ class TestMistral(unittest.TestCase): dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta) - assert (Path(output_dir) / "adapter_model.bin").exists() + assert (Path(temp_dir) / "adapter_model.bin").exists() @with_temp_dir - def test_ft(self, output_dir): + def test_ft(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault( { @@ -94,7 +94,7 @@ class TestMistral(unittest.TestCase): "num_epochs": 2, "micro_batch_size": 2, "gradient_accumulation_steps": 1, - "output_dir": output_dir, + "output_dir": temp_dir, "learning_rate": 0.00001, "optimizer": "adamw_torch", "lr_scheduler": "cosine", @@ -112,4 +112,4 @@ class TestMistral(unittest.TestCase): dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta) - assert (Path(output_dir) / "pytorch_model.bin").exists() + assert (Path(temp_dir) / "pytorch_model.bin").exists() diff --git a/tests/e2e/test_mistral_samplepack.py b/tests/e2e/test_mistral_samplepack.py index eb05a84b8..cefbd7dc0 100644 --- a/tests/e2e/test_mistral_samplepack.py +++ b/tests/e2e/test_mistral_samplepack.py @@ -27,7 +27,7 @@ class TestMistral(unittest.TestCase): """ @with_temp_dir - def test_lora_packing(self, output_dir): + def test_lora_packing(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault( { @@ -56,7 +56,7 @@ class TestMistral(unittest.TestCase): "num_epochs": 2, "micro_batch_size": 2, "gradient_accumulation_steps": 1, - "output_dir": output_dir, + "output_dir": temp_dir, "learning_rate": 0.00001, "optimizer": "adamw_torch", "lr_scheduler": "cosine", @@ -70,10 +70,10 @@ class TestMistral(unittest.TestCase): dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta) - assert (Path(output_dir) / "adapter_model.bin").exists() + assert (Path(temp_dir) / "adapter_model.bin").exists() @with_temp_dir - def test_ft_packing(self, output_dir): + def test_ft_packing(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault( { @@ -96,7 +96,7 @@ class TestMistral(unittest.TestCase): "num_epochs": 2, "micro_batch_size": 2, "gradient_accumulation_steps": 1, - "output_dir": output_dir, + "output_dir": temp_dir, "learning_rate": 0.00001, "optimizer": "adamw_torch", "lr_scheduler": "cosine", @@ -114,4 +114,4 @@ class TestMistral(unittest.TestCase): dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta) - assert (Path(output_dir) / "pytorch_model.bin").exists() + assert (Path(temp_dir) / "pytorch_model.bin").exists() diff --git a/tests/e2e/test_phi.py b/tests/e2e/test_phi.py index 78abe272e..b3e2ec95d 100644 --- a/tests/e2e/test_phi.py +++ b/tests/e2e/test_phi.py @@ -25,7 +25,7 @@ class TestPhi(unittest.TestCase): """ @with_temp_dir - def test_ft(self, output_dir): + def test_ft(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault( { @@ -55,7 +55,7 @@ class TestPhi(unittest.TestCase): "num_epochs": 1, "micro_batch_size": 1, "gradient_accumulation_steps": 1, - "output_dir": output_dir, + "output_dir": temp_dir, "learning_rate": 0.00001, "optimizer": "adamw_bnb_8bit", "lr_scheduler": "cosine", @@ -67,10 +67,10 @@ class TestPhi(unittest.TestCase): dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta) - assert (Path(output_dir) / "pytorch_model.bin").exists() + assert (Path(temp_dir) / "pytorch_model.bin").exists() @with_temp_dir - def test_ft_packed(self, output_dir): + def test_ft_packed(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault( { @@ -100,7 +100,7 @@ class TestPhi(unittest.TestCase): "num_epochs": 1, "micro_batch_size": 1, "gradient_accumulation_steps": 1, - "output_dir": output_dir, + "output_dir": temp_dir, "learning_rate": 0.00001, "optimizer": "adamw_bnb_8bit", "lr_scheduler": "cosine", @@ -112,4 +112,4 @@ class TestPhi(unittest.TestCase): dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args) train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta) - assert (Path(output_dir) / "pytorch_model.bin").exists() + assert (Path(temp_dir) / "pytorch_model.bin").exists() diff --git a/tests/e2e/utils.py b/tests/e2e/utils.py index 1c044778f..8b6c566d1 100644 --- a/tests/e2e/utils.py +++ b/tests/e2e/utils.py @@ -14,7 +14,7 @@ def with_temp_dir(test_func): temp_dir = tempfile.mkdtemp() try: # Pass the temporary directory to the test function - test_func(temp_dir, *args, **kwargs) + test_func(*args, temp_dir=temp_dir, **kwargs) finally: # Clean up the directory after the test shutil.rmtree(temp_dir)