From be22551435ed644b6b746058069afc2edbbbf334 Mon Sep 17 00:00:00 2001 From: NanoCode012 Date: Mon, 29 May 2023 15:33:40 +0900 Subject: [PATCH] Fix unsupported operand type(s) for | --- src/axolotl/prompt_tokenizers.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/axolotl/prompt_tokenizers.py b/src/axolotl/prompt_tokenizers.py index ceb65e2ab..761441a7e 100644 --- a/src/axolotl/prompt_tokenizers.py +++ b/src/axolotl/prompt_tokenizers.py @@ -4,7 +4,7 @@ import abc import copy import functools import logging -from typing import Dict, List, Tuple +from typing import Dict, List, Tuple, Union from transformers import PreTrainedTokenizer @@ -433,7 +433,7 @@ def parse_tokenized_to_result( current_len: int, res: Dict[str, List[int]], labels: list[int], - pad_token_id: int | None = None, + pad_token_id: Union[int, None] = None, ) -> Tuple[Dict[str, List[int]], int]: """ Parses the tokenized prompt and append the tokenized input_ids, attention_mask and labels to the result