diff --git a/tests/test_packed_dataset.py b/tests/test_packed_dataset.py index ced0360fe..1f19d0ecc 100644 --- a/tests/test_packed_dataset.py +++ b/tests/test_packed_dataset.py @@ -17,6 +17,7 @@ class TestPacking(unittest.TestCase): """ def setUp(self) -> None: + # pylint: disable=duplicate-code self.tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b") self.tokenizer.add_special_tokens( { diff --git a/tests/test_prompt_tokenizers.py b/tests/test_prompt_tokenizers.py index fa85fe5f6..89209e84f 100644 --- a/tests/test_prompt_tokenizers.py +++ b/tests/test_prompt_tokenizers.py @@ -18,6 +18,7 @@ class TestPromptTokenizationStrategies(unittest.TestCase): """ def setUp(self) -> None: + # pylint: disable=duplicate-code self.tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b") self.tokenizer.add_special_tokens( {