Compare commits
13 Commits
fix/issue-
...
preprocess
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
f8e92407ff | ||
|
|
c12906134d | ||
|
|
8154d26614 | ||
|
|
fefcbc300d | ||
|
|
7d479348ee | ||
|
|
ce0259db13 | ||
|
|
2798817cf9 | ||
|
|
0e1b081e49 | ||
|
|
8df37ad91f | ||
|
|
9b74298328 | ||
|
|
ae8738aa87 | ||
|
|
ec52561a0c | ||
|
|
eadb16c709 |
@@ -129,10 +129,12 @@ def load_preference_datasets(
|
||||
total_num_steps = None
|
||||
|
||||
if cli_args.debug or cfg.debug:
|
||||
if not cfg.rl == "grpo":
|
||||
LOG.info("check_dataset_labels...")
|
||||
|
||||
tokenizer = load_tokenizer(cfg)
|
||||
train_samples = sample_dataset(train_dataset, cli_args.debug_num_examples)
|
||||
|
||||
check_dataset_labels(
|
||||
train_samples,
|
||||
tokenizer,
|
||||
|
||||
@@ -4,30 +4,73 @@ module for base dataset transform strategies
|
||||
|
||||
import importlib
|
||||
import logging
|
||||
import sys
|
||||
|
||||
LOG = logging.getLogger("axolotl")
|
||||
|
||||
|
||||
def import_from_path(module_name: str, file_path: str):
|
||||
"""
|
||||
Import a module from a file path.
|
||||
|
||||
Args:
|
||||
module_name: Name of the module.
|
||||
file_path: Path to the file.
|
||||
|
||||
Returns:
|
||||
module: The imported module.
|
||||
"""
|
||||
spec = importlib.util.spec_from_file_location(module_name, file_path)
|
||||
if spec is None:
|
||||
raise ImportError(f"Could not create module spec for: {file_path}")
|
||||
module = importlib.util.module_from_spec(spec)
|
||||
|
||||
sys.modules[module_name] = module
|
||||
loader = importlib.machinery.SourceFileLoader(module_name, file_path)
|
||||
spec.loader = loader
|
||||
loader.exec_module(module)
|
||||
return module
|
||||
|
||||
|
||||
def load(strategy, cfg, module_base=None, **kwargs):
|
||||
try:
|
||||
if len(strategy.split(".")) == 1:
|
||||
strategy = strategy + ".default"
|
||||
load_fn = strategy.split(".")[-1]
|
||||
func = None
|
||||
if len(strategy.split(".")) > 1:
|
||||
try:
|
||||
importlib.import_module(
|
||||
mod = importlib.import_module(
|
||||
strategy.split(".")[-2],
|
||||
".".join(strategy.split(".")[:-2]),
|
||||
)
|
||||
module_base = ".".join(strategy.split(".")[:-2])
|
||||
strategy = strategy.split(".")[-2]
|
||||
func = getattr(mod, load_fn)
|
||||
return func(cfg, **kwargs)
|
||||
except ModuleNotFoundError:
|
||||
strategy = "." + ".".join(strategy.split(".")[:-1])
|
||||
pass
|
||||
|
||||
try:
|
||||
mod = importlib.import_module(
|
||||
"." + ".".join(strategy.split(".")[:-1]), module_base
|
||||
)
|
||||
func = getattr(mod, load_fn)
|
||||
return func(cfg, **kwargs)
|
||||
except ModuleNotFoundError:
|
||||
pass
|
||||
|
||||
try:
|
||||
file_path = "/".join(strategy.split(".")[:-1]) + ".py"
|
||||
module_name = strategy.split(".")[-2]
|
||||
mod = import_from_path(module_name, file_path)
|
||||
func = getattr(mod, load_fn)
|
||||
if func is not None:
|
||||
return func(cfg, **kwargs)
|
||||
except FileNotFoundError:
|
||||
pass
|
||||
else:
|
||||
strategy = "." + ".".join(strategy.split(".")[:-1])
|
||||
mod = importlib.import_module(strategy, module_base)
|
||||
func = getattr(mod, load_fn)
|
||||
return func(cfg, **kwargs)
|
||||
except Exception: # pylint: disable=broad-exception-caught
|
||||
|
||||
LOG.warning(f"unable to load strategy {strategy}")
|
||||
return None
|
||||
return func
|
||||
|
||||
85
tests/e2e/solo/test_preprocess.py
Normal file
85
tests/e2e/solo/test_preprocess.py
Normal file
@@ -0,0 +1,85 @@
|
||||
"""
|
||||
E2E tests for preprocessing
|
||||
"""
|
||||
|
||||
import logging
|
||||
import os
|
||||
import unittest
|
||||
|
||||
import transformers
|
||||
|
||||
from axolotl.cli.args import PreprocessCliArgs
|
||||
from axolotl.common.datasets import load_preference_datasets
|
||||
from axolotl.utils.config import normalize_config, validate_config
|
||||
from axolotl.utils.dict import DictDefault
|
||||
|
||||
from ..utils import with_temp_dir
|
||||
|
||||
LOG = logging.getLogger("axolotl.tests.e2e")
|
||||
os.environ["WANDB_DISABLED"] = "true"
|
||||
|
||||
|
||||
class TestCustomRewardFunctionLoading(unittest.TestCase):
|
||||
"""
|
||||
Test case for GRPO training using single GPU
|
||||
"""
|
||||
|
||||
def _utils_write_rewards(self):
|
||||
# write cfg to yaml file
|
||||
with open("rewards.py", "w", encoding="utf-8") as fout:
|
||||
fout.write(
|
||||
"""import random
|
||||
def rand_reward_func(completions, **kwargs) -> list[float]:
|
||||
return [random.uniform(0, 1) for _ in completions]
|
||||
|
||||
def oai_gsm8k_transform(cfg, *args, **kwargs):
|
||||
def transform_fn(example, tokenizer=None):
|
||||
label = example["answer"].split("####")[-1].strip().replace(",", "")
|
||||
return {
|
||||
"prompt": [{"role": "user", "content": example["question"]},],
|
||||
"answer": label,
|
||||
}
|
||||
return transform_fn, {"remove_columns": ["question"]}
|
||||
"""
|
||||
)
|
||||
|
||||
@with_temp_dir
|
||||
def test_custom_rewards_fn_preprocess(self, temp_dir):
|
||||
# pylint: disable=duplicate-code
|
||||
cfg = DictDefault(
|
||||
{
|
||||
"base_model": "HuggingFaceTB/SmolLM2-135M",
|
||||
"strict": False,
|
||||
"rl": "grpo",
|
||||
"trl": {
|
||||
"beta": 0.001,
|
||||
"max_completion_length": 256,
|
||||
"use_vllm": True,
|
||||
"num_generations": 4,
|
||||
"reward_funcs": [
|
||||
"rewards.rand_reward_func"
|
||||
], # format: '{file_name}.{fn_name}'
|
||||
"reward_weights": [1.0],
|
||||
},
|
||||
"datasets": [
|
||||
{
|
||||
"path": "openai/gsm8k",
|
||||
"name": "main",
|
||||
"type": "rewards.oai_gsm8k_transform",
|
||||
},
|
||||
],
|
||||
"dataset_prepared_path": temp_dir,
|
||||
"gradient_accumulation_steps": 1,
|
||||
"micro_batch_size": 1,
|
||||
"learning_rate": 0.000005,
|
||||
}
|
||||
)
|
||||
|
||||
self._utils_write_rewards()
|
||||
|
||||
cfg = validate_config(cfg)
|
||||
normalize_config(cfg)
|
||||
parser = transformers.HfArgumentParser(PreprocessCliArgs)
|
||||
cli_args, _ = parser.parse_args_into_dataclasses(return_remaining_strings=True)
|
||||
|
||||
load_preference_datasets(cfg=cfg, cli_args=cli_args)
|
||||
Reference in New Issue
Block a user