| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """Taskmaster-3: A goal oriented conversations dataset for movie ticketing domain """ |
|
|
|
|
| import json |
|
|
| import datasets |
|
|
|
|
| _CITATION = """\ |
| @inproceedings{48484, |
| title = {Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset}, |
| author = {Bill Byrne and Karthik Krishnamoorthi and Chinnadhurai Sankar and Arvind Neelakantan and Daniel Duckworth and Semih Yavuz and Ben Goodrich and Amit Dubey and Kyu-Young Kim and Andy Cedilnik}, |
| year = {2019} |
| } |
| """ |
|
|
| _DESCRIPTION = """\ |
| Taskmaster is dataset for goal oriented conversations. The Taskmaster-3 dataset consists of 23,757 movie ticketing dialogs. \ |
| By "movie ticketing" we mean conversations where the customer's goal is to purchase tickets after deciding \ |
| on theater, time, movie name, number of tickets, and date, or opt out of the transaction. This collection \ |
| was created using the "self-dialog" method. This means a single, crowd-sourced worker is \ |
| paid to create a conversation writing turns for both speakers, i.e. the customer and the ticketing agent. |
| """ |
|
|
| _HOMEPAGE = "https://github.com/google-research-datasets/Taskmaster/tree/master/TM-3-2020" |
|
|
| _BASE_URL = "https://raw.githubusercontent.com/google-research-datasets/Taskmaster/master/TM-3-2020/data" |
|
|
|
|
| class Taskmaster3(datasets.GeneratorBasedBuilder): |
| """Taskmaster-3: A goal oriented conversations dataset for movie ticketing domain""" |
|
|
| VERSION = datasets.Version("1.0.0") |
|
|
| def _info(self): |
| features = { |
| "conversation_id": datasets.Value("string"), |
| "vertical": datasets.Value("string"), |
| "instructions": datasets.Value("string"), |
| "scenario": datasets.Value("string"), |
| "utterances": [ |
| { |
| "index": datasets.Value("int32"), |
| "speaker": datasets.Value("string"), |
| "text": datasets.Value("string"), |
| "apis": [ |
| { |
| "name": datasets.Value("string"), |
| "index": datasets.Value("int32"), |
| "args": [ |
| { |
| "arg_name": datasets.Value("string"), |
| "arg_value": datasets.Value("string"), |
| } |
| ], |
| "response": [ |
| { |
| "response_name": datasets.Value("string"), |
| "response_value": datasets.Value("string"), |
| } |
| ], |
| } |
| ], |
| "segments": [ |
| { |
| "start_index": datasets.Value("int32"), |
| "end_index": datasets.Value("int32"), |
| "text": datasets.Value("string"), |
| "annotations": [{"name": datasets.Value("string")}], |
| } |
| ], |
| } |
| ], |
| } |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features(features), |
| supervised_keys=None, |
| homepage=_HOMEPAGE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| urls = [f"{_BASE_URL}/data_{i:02}.json" for i in range(20)] |
| dialog_files = dl_manager.download(urls) |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={"dialog_files": dialog_files}, |
| ), |
| ] |
|
|
| def _generate_examples(self, dialog_files): |
| for filepath in dialog_files: |
| with open(filepath, encoding="utf-8") as f: |
| dialogs = json.load(f) |
| for dialog in dialogs: |
| example = self._prepare_example(dialog) |
| yield example["conversation_id"], example |
|
|
| def _prepare_example(self, dialog): |
| utterances = dialog["utterances"] |
| for utterance in utterances: |
| if "segments" not in utterance: |
| utterance["segments"] = [] |
|
|
| if "apis" in utterance: |
| utterance["apis"] = self._transform_apis(utterance["apis"]) |
| else: |
| utterance["apis"] = [] |
| return dialog |
|
|
| def _transform_apis(self, apis): |
| for api in apis: |
| if "args" in api: |
| api["args"] = [{"arg_name": k, "arg_value": v} for k, v in api["args"].items()] |
| else: |
| api["args"] = [] |
|
|
| if "response" in api: |
| api["response"] = [{"response_name": k, "response_value": v} for k, v in api["response"].items()] |
| else: |
| api["response"] = [] |
|
|
| return apis |
|
|