| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """Microsoft Research Sequential Question Answering (SQA) Dataset""" |
|
|
|
|
| import ast |
| import csv |
| import os |
|
|
| import pandas as pd |
|
|
| import datasets |
|
|
|
|
| |
| _CITATION = """\ |
| @inproceedings{iyyer2017search, |
| title={Search-based neural structured learning for sequential question answering}, |
| author={Iyyer, Mohit and Yih, Wen-tau and Chang, Ming-Wei}, |
| booktitle={Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, |
| pages={1821--1831}, |
| year={2017} |
| } |
| """ |
|
|
| _DESCRIPTION = """\ |
| Recent work in semantic parsing for question answering has focused on long and complicated questions, \ |
| many of which would seem unnatural if asked in a normal conversation between two humans. \ |
| In an effort to explore a conversational QA setting, we present a more realistic task: \ |
| answering sequences of simple but inter-related questions. We created SQA by asking crowdsourced workers \ |
| to decompose 2,022 questions from WikiTableQuestions (WTQ), which contains highly-compositional questions about \ |
| tables from Wikipedia. We had three workers decompose each WTQ question, resulting in a dataset of 6,066 sequences \ |
| that contain 17,553 questions in total. Each question is also associated with answers in the form of cell \ |
| locations in the tables. |
| """ |
|
|
| _HOMEPAGE = "https://msropendata.com/datasets/b25190ed-0f59-47b1-9211-5962858142c2" |
|
|
| _LICENSE = "Microsoft Research Data License Agreement" |
|
|
| _URL = "https://download.microsoft.com/download/1/D/C/1DC270D2-1B53-4A61-A2E3-88AB3E4E6E1F/SQA%20Release%201.0.zip" |
|
|
|
|
| def _load_table_data(table_file): |
| """Load additional data from a csv table file. |
| |
| Args: |
| table_file: Path to the csv file. |
| |
| Returns: |
| header: a list of headers in the table. |
| rows: 2d array of data in the table. |
| """ |
| rows = [] |
| table_data = pd.read_csv(table_file) |
| |
| header = list(table_data.columns) |
| for row_data in table_data.values: |
| rows.append([str(_) for _ in list(row_data)]) |
|
|
| return header, rows |
|
|
|
|
| def _parse_answer_coordinates(answer_coordinate_str): |
| """Parsing answer_coordinates field to a list of answer coordinates. |
| The original code is from https://github.com/google-research/tapas. |
| |
| Args: |
| answer_coordinate_str: A string representation of a Python list of tuple |
| strings. |
| For example: "['(1, 4)','(1, 3)', ...]" |
| |
| Returns: |
| answer_coordinates: A list of answer cordinates. |
| """ |
| try: |
| answer_coordinates = [] |
| coords = ast.literal_eval(answer_coordinate_str) |
| for row_index, column_index in sorted(ast.literal_eval(coord) for coord in coords): |
| answer_coordinates.append({"row_index": row_index, "column_index": column_index}) |
| return answer_coordinates |
| except SyntaxError: |
| raise ValueError("Unable to evaluate %s" % answer_coordinate_str) |
|
|
|
|
| def _parse_answer_text(answer_text_str): |
| """Parsing `answer_text` field to list of answers. |
| The original code is from https://github.com/google-research/tapas. |
| Args: |
| answer_text_str: A string representation of a Python list of strings. |
| For example: "[u'test', u'hello', ...]" |
| |
| Returns: |
| answer_texts: A list of answers. |
| """ |
| try: |
| answer_texts = [] |
| for value in ast.literal_eval(answer_text_str): |
| answer_texts.append(value) |
| return answer_texts |
| except SyntaxError: |
| raise ValueError("Unable to evaluate %s" % answer_text_str) |
|
|
|
|
| class MsrSQA(datasets.GeneratorBasedBuilder): |
| """Microsoft Research Sequential Question Answering (SQA) Dataset""" |
|
|
| VERSION = datasets.Version("1.0.0") |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| "id": datasets.Value("string"), |
| "annotator": datasets.Value("int32"), |
| "position": datasets.Value("int32"), |
| "question": datasets.Value("string"), |
| "question_and_history": datasets.Sequence(datasets.Value("string")), |
| "table_file": datasets.Value("string"), |
| "table_header": datasets.features.Sequence(datasets.Value("string")), |
| "table_data": datasets.features.Sequence(datasets.features.Sequence(datasets.Value("string"))), |
| "answer_coordinates": datasets.features.Sequence( |
| {"row_index": datasets.Value("int32"), "column_index": datasets.Value("int32")} |
| ), |
| "answer_text": datasets.features.Sequence(datasets.Value("string")), |
| } |
| ), |
| supervised_keys=None, |
| homepage=_HOMEPAGE, |
| license=_LICENSE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| """Returns SplitGenerators.""" |
| data_dir = os.path.join(dl_manager.download_and_extract(_URL), "SQA Release 1.0") |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={"filepath": os.path.join(data_dir, "random-split-1-train.tsv"), "data_dir": data_dir}, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.VALIDATION, |
| gen_kwargs={"filepath": os.path.join(data_dir, "random-split-1-dev.tsv"), "data_dir": data_dir}, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.TEST, |
| gen_kwargs={"filepath": os.path.join(data_dir, "test.tsv"), "data_dir": data_dir}, |
| ), |
| ] |
|
|
| def _generate_examples(self, filepath, data_dir): |
| """Yields examples.""" |
| with open(filepath, encoding="utf-8") as f: |
| reader = csv.DictReader(f, delimiter="\t") |
| question_and_history = [] |
| for idx, item in enumerate(reader): |
| item["answer_text"] = _parse_answer_text(item["answer_text"]) |
| item["answer_coordinates"] = _parse_answer_coordinates(item["answer_coordinates"]) |
| header, table_data = _load_table_data(os.path.join(data_dir, item["table_file"])) |
| item["table_header"] = header |
| item["table_data"] = table_data |
| if item["position"] == "0": |
| question_and_history = [] |
| question_and_history.append(item["question"]) |
| item["question_and_history"] = question_and_history |
| yield idx, item |
|
|