| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """ |
| Preprocess Hellaswag dataset. |
| |
| """ |
|
|
| import argparse |
| import os |
| import re |
|
|
| import datasets |
|
|
| from verl.utils.hdfs_io import copy, makedirs |
|
|
|
|
| def preprocess(text): |
| text = text.strip() |
| |
| text = text.replace(" [title]", ". ") |
| text = re.sub("\\[.*?\\]", "", text) |
| text = text.replace(" ", " ") |
| return text |
|
|
|
|
| if __name__ == "__main__": |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--local_dir", default=None, help="The save directory for the preprocessed dataset.") |
| parser.add_argument("--hdfs_dir", default=None) |
| parser.add_argument("--local_dataset_path", default=None, help="The local path to the raw dataset, if it exists.") |
| parser.add_argument( |
| "--local_save_dir", default="~/data/hellaswag", help="The save directory for the preprocessed dataset." |
| ) |
|
|
| args = parser.parse_args() |
| local_dataset_path = args.local_dataset_path |
|
|
| data_source = "Rowan/hellaswag" |
|
|
| if local_dataset_path is not None: |
| dataset = datasets.load_dataset(local_dataset_path) |
| else: |
| dataset = datasets.load_dataset(data_source, trust_remote_code=True) |
|
|
| train_dataset = dataset["train"] |
| val_dataset = dataset["validation"] |
| test_dataset = dataset["test"] |
|
|
| instruction = "Please complete the following sentence.\n" |
|
|
| def make_map_fn(split): |
| def process_fn(doc, idx): |
| ctx = doc["ctx_a"] + " " + doc["ctx_b"].capitalize() |
| query = preprocess(doc["activity_label"] + ": " + ctx) |
| choices = [preprocess(ending) for ending in doc["endings"]] |
| gold = int(doc["label"]) |
|
|
| data = { |
| "data_source": data_source, |
| "prompt": [{"role": "user", "content": query}], |
| "ability": "nlp", |
| "reward_model": { |
| "style": "model", |
| "eval": "multiple_choice", |
| "ground_truth": gold, |
| "choices": choices, |
| }, |
| "extra_info": {"split": split, "index": idx}, |
| } |
| return data |
|
|
| return process_fn |
|
|
| |
| train_dataset = train_dataset.filter(lambda x: len(x["label"]) > 0) |
| val_dataset = val_dataset.filter(lambda x: len(x["label"]) > 0) |
| test_dataset = test_dataset.filter(lambda x: len(x["label"]) > 0) |
|
|
| train_dataset = train_dataset.map(function=make_map_fn("train"), with_indices=True) |
| val_dataset = val_dataset.map(function=make_map_fn("validation"), with_indices=True) |
| test_dataset = test_dataset.map(function=make_map_fn("test"), with_indices=True) |
|
|
| hdfs_dir = args.hdfs_dir |
| local_save_dir = args.local_dir |
| if local_save_dir is not None: |
| print("Warning: Argument 'local_dir' is deprecated. Please use 'local_save_dir' instead.") |
| else: |
| local_save_dir = args.local_save_dir |
|
|
| train_dataset.to_parquet(os.path.join(local_save_dir, "train.parquet")) |
| val_dataset.to_parquet(os.path.join(local_save_dir, "validation.parquet")) |
| test_dataset.to_parquet(os.path.join(local_save_dir, "test.parquet")) |
|
|
| if hdfs_dir is not None: |
| makedirs(hdfs_dir) |
|
|
| copy(src=local_save_dir, dst=hdfs_dir) |
|
|