| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """ |
| Preprocess the MATH-lighteval dataset to parquet format |
| """ |
|
|
| import argparse |
| import json |
| import os |
|
|
| import datasets |
|
|
| from verl.utils.hdfs_io import copy, makedirs |
| from verl.utils.reward_score.math_reward import last_boxed_only_string, remove_boxed |
|
|
|
|
| def extract_solution(solution_str): |
| return remove_boxed(last_boxed_only_string(solution_str)) |
|
|
|
|
| if __name__ == "__main__": |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--local_dir", default=None) |
| parser.add_argument("--hdfs_dir", default=None) |
| parser.add_argument("--local_dataset_path", default=None, help="The local path to the raw dataset, if it exists.") |
| parser.add_argument( |
| "--local_save_dir", default="~/data/math", help="The save directory for the preprocessed dataset." |
| ) |
|
|
| args = parser.parse_args() |
| local_dataset_path = args.local_dataset_path |
|
|
| |
| |
| data_source = "DigitalLearningGmbH/MATH-lighteval" |
| print(f"Loading the {data_source} dataset from huggingface...", flush=True) |
| if local_dataset_path is not None: |
| dataset = datasets.load_dataset( |
| local_dataset_path, |
| ) |
| else: |
| dataset = datasets.load_dataset( |
| data_source, |
| ) |
|
|
| train_dataset = dataset["train"] |
| test_dataset = dataset["test"] |
|
|
| instruction_following = "Let's think step by step and output the final answer within \\boxed{}." |
|
|
| |
| def make_map_fn(split): |
| def process_fn(example, idx): |
| question = example.pop("problem") |
|
|
| question = question + " " + instruction_following |
|
|
| answer = example.pop("solution") |
| solution = extract_solution(answer) |
| data = { |
| "data_source": data_source, |
| "prompt": [{"role": "user", "content": question}], |
| "ability": "math", |
| "reward_model": {"style": "rule", "ground_truth": solution}, |
| "extra_info": {"split": split, "index": idx}, |
| } |
| return data |
|
|
| return process_fn |
|
|
| train_dataset = train_dataset.map(function=make_map_fn("train"), with_indices=True) |
| test_dataset = test_dataset.map(function=make_map_fn("test"), with_indices=True) |
|
|
| local_save_dir = args.local_dir |
| if local_save_dir is not None: |
| print("Warning: Argument 'local_dir' is deprecated. Please use 'local_save_dir' instead.") |
| else: |
| local_save_dir = args.local_save_dir |
|
|
| local_dir = os.path.expanduser(local_save_dir) |
| hdfs_dir = args.hdfs_dir |
|
|
| train_dataset.to_parquet(os.path.join(local_dir, "train.parquet")) |
| test_dataset.to_parquet(os.path.join(local_dir, "test.parquet")) |
| |
| example = train_dataset[0] |
| with open(os.path.join(local_dir, "train_example.json"), "w") as f: |
| json.dump(example, f, indent=2) |
| example = test_dataset[0] |
| with open(os.path.join(local_dir, "test_example.json"), "w") as f: |
| json.dump(example, f, indent=2) |
| if hdfs_dir is not None: |
| makedirs(hdfs_dir) |
|
|
| copy(src=local_dir, dst=hdfs_dir) |
|
|