diff --git a/src/axolotl/datasets.py b/src/axolotl/datasets.py index cd4db4e72..5b627e258 100644 --- a/src/axolotl/datasets.py +++ b/src/axolotl/datasets.py @@ -64,7 +64,9 @@ class TokenizedPromptDataset(Dataset): desc="Strategy Filtering Rows", ) - import ipdb; ipdb.set_trace() + import ipdb + + ipdb.set_trace() return dataset.map( self.prompt_tokenizer.tokenize_prompt, diff --git a/src/axolotl/prompt_tokenizers.py b/src/axolotl/prompt_tokenizers.py index dd882710f..e16581ba7 100644 --- a/src/axolotl/prompt_tokenizers.py +++ b/src/axolotl/prompt_tokenizers.py @@ -67,7 +67,9 @@ class PromptTokenizingStrategy(abc.ABC): LOG.warning("Empty text requested for tokenization.") return empty - import ipdb; ipdb.set_trace() + import ipdb + + ipdb.set_trace() result = self.tokenizer( prompt, diff --git a/src/axolotl/utils/data/sft.py b/src/axolotl/utils/data/sft.py index d09c42704..ba6070a6e 100644 --- a/src/axolotl/utils/data/sft.py +++ b/src/axolotl/utils/data/sft.py @@ -486,7 +486,9 @@ def get_dataset_wrapper( f"Loading dataset: {config_dataset['path']} with base_type: {d_base_type} and prompt_style: {d_prompt_style}" ) - import ipdb; ipdb.set_trace() + import ipdb + + ipdb.set_trace() if ( isinstance(dataset, Dataset)