| from __future__ import annotations |
|
|
| from pathlib import Path |
| import numpy as np |
|
|
| import datasets |
|
|
|
|
| _HF_AFFIX = { |
| "ara": "arabic", |
| "cmn": "mandarin", |
| "eng": "", |
| "deu": "german", |
| "fra": "french", |
| "hin": "hindi", |
| "ita": "italian", |
| "nld": "dutch", |
| "pol": "polish", |
| "por": "portuguese", |
| "spa": "spanish", |
| } |
|
|
| _HF_AFFIX_REV = {v:k for k,v in _HF_AFFIX.items()} |
|
|
| _REVISION_DICT = { |
| "ara": "65eb7455a05cb77b3ae0c69d444569a8eee54628", |
| "cmn": "617d3e9fccd186277297cc305f6588af7384b008", |
| "eng": "9d2ac89df04254e5c427bcc8d61b6d6c83a1f59b", |
| "deu": "5229a5cc475f36c08d03ca52f0ccb005705e60d2", |
| "fra": "5d3085f2129139abc10d2b58becd4d4f2978e5d5", |
| "hin": "e9e68e1a4db04726b9278192377049d0f9693012", |
| "ita": "21e3d5c827cb60619a89988b24979850a7af85a5", |
| "nld": "d622427417d37a8d74e110e6289bc29af4ba4056", |
| "pol": "28d7098e2e5a211c4810d0a4d8deccc5889e55b6", |
| "por": "323bdf67e0fbd3d7f8086fad0971b5bd5a62524b", |
| "spa": "a7ea759535bb9fad6361cca151cf94a46e88edf3", |
| } |
|
|
| def _transform(dataset): |
| target_cols = ["test_case", "label_gold"] |
| new_cols = ['text', 'is_hateful'] |
| rename_dict = dict(zip(target_cols, ["text", "is_hateful"])) |
| dataset = dataset.rename_columns(rename_dict) |
| keep_cols = new_cols + ["functionality"] |
| remove_cols = [col for col in dataset["test"].column_names if col not in keep_cols] |
| dataset = dataset.remove_columns(remove_cols) |
| return dataset |
|
|
|
|
| def make_dataset(): |
| """ |
| Load dataset from HuggingFace hub |
| """ |
| ds = {} |
| for lang in _HF_AFFIX.values(): |
| lcode = _HF_AFFIX_REV[lang] |
| path = f'Paul/hatecheck-{lang}'.rstrip('-') |
| dataset = datasets.load_dataset( |
| path=path, revision=_REVISION_DICT[lcode] |
| ) |
| dataset = _transform(dataset) |
| out_path = Path('..') / lcode / 'test.jsonl' |
| dataset['test'].to_json(out_path) |
| ds[lcode] = dataset |
| return ds |
| |
|
|
| if __name__ == '__main__': |
| dataset = make_dataset() |
| AVG_CHAR = 0 |
| for lang in _HF_AFFIX: |
| AVG_CHAR += np.mean([len(x['text']) for x in dataset[lang]['test']]) |
| print(f'avg char: {AVG_CHAR / len(_HF_AFFIX)}') |
|
|
|
|