From 0136f510f2dcbf039a20e1fdd6c5c256016f6390 Mon Sep 17 00:00:00 2001 From: Wing Lian Date: Wed, 31 May 2023 12:05:43 -0400 Subject: [PATCH] don't worry about duplicate code here --- tests/test_packed_dataset.py | 1 + tests/test_prompt_tokenizers.py | 1 + 2 files changed, 2 insertions(+) diff --git a/tests/test_packed_dataset.py b/tests/test_packed_dataset.py index ced0360fe..1f19d0ecc 100644 --- a/tests/test_packed_dataset.py +++ b/tests/test_packed_dataset.py @@ -17,6 +17,7 @@ class TestPacking(unittest.TestCase): """ def setUp(self) -> None: + # pylint: disable=duplicate-code self.tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b") self.tokenizer.add_special_tokens( { diff --git a/tests/test_prompt_tokenizers.py b/tests/test_prompt_tokenizers.py index fa85fe5f6..89209e84f 100644 --- a/tests/test_prompt_tokenizers.py +++ b/tests/test_prompt_tokenizers.py @@ -18,6 +18,7 @@ class TestPromptTokenizationStrategies(unittest.TestCase): """ def setUp(self) -> None: + # pylint: disable=duplicate-code self.tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b") self.tokenizer.add_special_tokens( {