| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| """The General Language Understanding Evaluation (GLUE) benchmark.""" |
|
|
|
|
| import json |
| import textwrap |
|
|
| import datasets |
|
|
|
|
| _XGLUE_CITATION = """\ |
| @article{Liang2020XGLUEAN, |
| title={XGLUE: A New Benchmark Dataset for Cross-lingual Pre-training, Understanding and Generation}, |
| author={Yaobo Liang and Nan Duan and Yeyun Gong and Ning Wu and Fenfei Guo and Weizhen Qi |
| and Ming Gong and Linjun Shou and Daxin Jiang and Guihong Cao and Xiaodong Fan and Ruofei |
| Zhang and Rahul Agrawal and Edward Cui and Sining Wei and Taroon Bharti and Ying Qiao |
| and Jiun-Hung Chen and Winnie Wu and Shuguang Liu and Fan Yang and Daniel Campos |
| and Rangan Majumder and Ming Zhou}, |
| journal={arXiv}, |
| year={2020}, |
| volume={abs/2004.01401} |
| } |
| """ |
|
|
| _XGLUE_DESCRIPTION = """\ |
| XGLUE is a new benchmark dataset to evaluate the performance of cross-lingual pre-trained |
| models with respect to cross-lingual natural language understanding and generation. |
| The benchmark is composed of the following 11 tasks: |
| - NER |
| - POS Tagging (POS) |
| - News Classification (NC) |
| - MLQA |
| - XNLI |
| - PAWS-X |
| - Query-Ad Matching (QADSM) |
| - Web Page Ranking (WPR) |
| - QA Matching (QAM) |
| - Question Generation (QG) |
| - News Title Generation (NTG) |
| |
| For more information, please take a look at https://microsoft.github.io/XGLUE/. |
| """ |
|
|
| _XGLUE_ALL_DATA = "https://xglue.blob.core.windows.net/xglue/xglue_full_dataset.tar.gz" |
|
|
| _LANGUAGES = { |
| "ner": ["en", "de", "es", "nl"], |
| "pos": ["en", "de", "es", "nl", "bg", "el", "fr", "pl", "tr", "vi", "zh", "ur", "hi", "it", "ar", "ru", "th"], |
| "mlqa": ["en", "de", "ar", "es", "hi", "vi", "zh"], |
| "nc": ["en", "de", "es", "fr", "ru"], |
| "xnli": ["en", "ar", "bg", "de", "el", "es", "fr", "hi", "ru", "sw", "th", "tr", "ur", "vi", "zh"], |
| "paws-x": ["en", "de", "es", "fr"], |
| "qadsm": ["en", "de", "fr"], |
| "wpr": ["en", "de", "es", "fr", "it", "pt", "zh"], |
| "qam": ["en", "de", "fr"], |
| "qg": ["en", "de", "es", "fr", "it", "pt"], |
| "ntg": ["en", "de", "es", "fr", "ru"], |
| } |
|
|
| _PATHS = { |
| "mlqa": { |
| "train": "squad1.1/train-v1.1.json", |
| "dev": "MLQA_V1/dev/dev-context-{0}-question-{0}.json", |
| "test": "MLQA_V1/test/test-context-{0}-question-{0}.json", |
| }, |
| "xnli": {"train": "multinli.train.en.tsv", "dev": "{}.dev", "test": "{}.test"}, |
| "paws-x": { |
| "train": "en/train.tsv", |
| "dev": "{}/dev_2k.tsv", |
| "test": "{}/test_2k.tsv", |
| }, |
| } |
| for name in ["ner", "pos"]: |
| _PATHS[name] = {"train": "en.train", "dev": "{}.dev", "test": "{}.test"} |
| for name in ["nc", "qadsm", "wpr", "qam"]: |
| _PATHS[name] = { |
| "train": "xglue." + name + ".en.train", |
| "dev": "xglue." + name + ".{}.dev", |
| "test": "xglue." + name + ".{}.test", |
| } |
| for name in ["qg", "ntg"]: |
| _PATHS[name] = {"train": "xglue." + name + ".en", "dev": "xglue." + name + ".{}", "test": "xglue." + name + ".{}"} |
|
|
|
|
| class XGlueConfig(datasets.BuilderConfig): |
| """BuilderConfig for XGLUE.""" |
|
|
| def __init__( |
| self, |
| data_dir, |
| citation, |
| url, |
| **kwargs, |
| ): |
| """BuilderConfig for XGLUE. |
| |
| Args: |
| data_dir: `string`, the path to the folder containing the files in the |
| downloaded .tar |
| citation: `string`, citation for the data set |
| url: `string`, url for information about the data set |
| **kwargs: keyword arguments forwarded to super. |
| """ |
| super(XGlueConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs) |
| self.data_dir = data_dir |
| self.citation = citation |
| self.url = url |
|
|
|
|
| class XGlue(datasets.GeneratorBasedBuilder): |
| """The Cross-lingual Pre-training, Understanding and Generation (XGlue) Benchmark.""" |
|
|
| BUILDER_CONFIGS = [ |
| XGlueConfig( |
| name="ner", |
| description=textwrap.dedent( |
| """\ |
| The shared task of CoNLL-2003 concerns language-independent named entity recognition. |
| We will concentrate on four types of named entities: |
| persons, locations, organizations and names of miscellaneous entities |
| that do not belong to the previous three groups. |
| """ |
| ), |
| data_dir="NER", |
| citation=textwrap.dedent( |
| """\ |
| @article{Sang2003IntroductionTT, |
| title={Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition}, |
| author={Erik F. Tjong Kim Sang and Fien De Meulder}, |
| journal={ArXiv}, |
| year={2003}, |
| volume={cs.CL/0306050} |
| }, |
| @article{Sang2002IntroductionTT, |
| title={Introduction to the CoNLL-2002 Shared Task: Language-Independent Named Entity Recognition}, |
| author={Erik F. Tjong Kim Sang}, |
| journal={ArXiv}, |
| year={2002}, |
| volume={cs.CL/0209010} |
| }""" |
| ), |
| url="https://www.clips.uantwerpen.be/conll2003/ner/", |
| ), |
| XGlueConfig( |
| name="pos", |
| description=textwrap.dedent( |
| """\ |
| Universal Dependencies (UD) is a project that is developing cross-linguistically consistent treebank |
| annotation for many languages, with the goal of facilitating multilingual parser development, cross-lingual |
| learning, and parsing research from a language typology perspective. The annotation scheme is based on an |
| evolution of (universal) Stanford dependencies (de Marneffe et al., 2006, 2008, 2014), Google universal |
| part-of-speech tags (Petrov et al., 2012), and the Interset interlingua for morphosyntactic tagsets |
| (Zeman, 2008). The general philosophy is to provide a universal inventory of categories and guidelines |
| to facilitate consistent annotation of similar constructions across languages, while |
| allowing language-specific extensions when necessary. |
| """ |
| ), |
| data_dir="POS", |
| citation=textwrap.dedent( |
| """\ |
| @misc{11234/1-3105, |
| title={Universal Dependencies 2.5}, |
| author={Zeman, Daniel and Nivre, Joakim and Abrams, Mitchell and Aepli, et al.}, |
| url={http://hdl.handle.net/11234/1-3105}, |
| note={{LINDAT}/{CLARIAH}-{CZ} digital library at the Institute of Formal and Applied Linguistics ({{\'U}FAL}), Faculty of Mathematics and Physics, Charles University}, |
| copyright={Licence Universal Dependencies v2.5}, |
| year={2019} |
| }""" |
| ), |
| url="https://universaldependencies.org/", |
| ), |
| XGlueConfig( |
| name="mlqa", |
| description=textwrap.dedent( |
| """\ |
| MLQA (MultiLingual Question Answering) is a benchmark dataset for evaluating cross-lingual question answering |
| performance. MLQA consists of over 5K extractive QA instances (12K in English) in SQuAD format in seven languages |
| - English, Arabic, German, Spanish, Hindi, Vietnamese and Simplified Chinese. |
| MLQA is highly parallel, with QA instances parallel between 4 different languages on average. |
| """ |
| ), |
| data_dir="MLQA", |
| citation=textwrap.dedent( |
| """\ |
| @article{Lewis2019MLQAEC, |
| title={MLQA: Evaluating Cross-lingual Extractive Question Answering}, |
| author={Patrick Lewis and Barlas Oguz and Ruty Rinott and Sebastian Riedel and Holger Schwenk}, |
| journal={ArXiv}, |
| year={2019}, |
| volume={abs/1910.07475} |
| }""" |
| ), |
| url="https://github.com/facebookresearch/MLQA", |
| ), |
| XGlueConfig( |
| name="nc", |
| description=textwrap.dedent( |
| """\ |
| This task aims to predict the category given a news article. It covers |
| 5 languages, including English, Spanish, French, |
| German and Russian. Each labeled instance is a |
| 3-tuple: <news title, news body, category>. The |
| category number is 10. We crawl this dataset from |
| a commercial news website. Accuracy (ACC) of |
| the multi-class classification is used as the metric. |
| """ |
| ), |
| data_dir="NC", |
| citation="", |
| url="", |
| ), |
| XGlueConfig( |
| name="xnli", |
| description=textwrap.dedent( |
| """\ |
| XNLI is a subset of a few thousand examples from MNLI which has been translated |
| into a 14 different languages (some low-ish resource). As with MNLI, the goal is |
| to predict textual entailment (does sentence A imply/contradict/neither sentence |
| B) and is a classification task (given two sentences, predict one of three |
| labels). |
| """ |
| ), |
| data_dir="XNLI", |
| citation=textwrap.dedent( |
| """\ |
| @inproceedings{Conneau2018XNLIEC, |
| title={XNLI: Evaluating Cross-lingual Sentence Representations}, |
| author={Alexis Conneau and Guillaume Lample and Ruty Rinott and Adina Williams and Samuel R. Bowman and Holger Schwenk and Veselin Stoyanov}, |
| booktitle={EMNLP}, |
| year={2018} |
| }""" |
| ), |
| url="https://github.com/facebookresearch/XNLI", |
| ), |
| XGlueConfig( |
| name="paws-x", |
| description=textwrap.dedent( |
| """\ |
| PAWS-X contains 23,659 human translated PAWS (Paraphrase Adversaries from Word Scrambling) evaluation pairs and 296,406 machine translated training pairs in six typologically distinct languages: French, Spanish, German, Chinese, Japanese, and Korean. All translated pairs are sourced from examples in PAWS-Wiki. |
| """ |
| ), |
| data_dir="PAWSX", |
| citation=textwrap.dedent( |
| """\ |
| @article{Yang2019PAWSXAC, |
| title={PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase Identification}, |
| author={Yinfei Yang and Yuan Zhang and Chris Tar and Jason Baldridge}, |
| journal={ArXiv}, |
| year={2019}, |
| volume={abs/1908.11828} |
| }""" |
| ), |
| url="https://github.com/google-research-datasets/paws/tree/master/pawsx", |
| ), |
| XGlueConfig( |
| name="qadsm", |
| description=textwrap.dedent( |
| """\ |
| Query-Ad Matching (QADSM) task aims |
| to predict whether an advertisement (ad) is relevant to an input query. It covers 3 languages, including English, French and German. Each labeled instance is a 4-tuple: <query, ad title, ad description, label>. The label indicates whether the |
| ad is relevant to the query (Good), or not (Bad). |
| This dataset was constructed based on a commercial search engine. Accuracy (ACC) of the binary classification should be used as the metric. |
| """ |
| ), |
| data_dir="QADSM", |
| citation="", |
| url="", |
| ), |
| XGlueConfig( |
| name="wpr", |
| description=textwrap.dedent( |
| """\ |
| Tthe Web Page Ranking (WPR) task aims to |
| predict whether a web page is relevant to an input query. It covers 7 languages, including English, German, French, Spanish, Italian, Portuguese and Chinese. Each labeled instance is a |
| 4-tuple: <query, web page title, web page snippet, label>. The relevance label contains 5 ratings: Perfect (4), Excellent (3), Good (2), Fair (1) |
| and Bad (0). The dataset is constructed based on a |
| commercial search engine. Normalize Discounted |
| Cumulative Gain (nDCG) should be used as the metric. |
| """ |
| ), |
| data_dir="WPR", |
| citation="", |
| url="", |
| ), |
| XGlueConfig( |
| name="qam", |
| description=textwrap.dedent( |
| """\ |
| The QA Matching (QAM) task aims to predict whether a <question, passage> pair is a QA pair. |
| It covers 3 languages, including English, French |
| and German. Each labeled instance is a 3-tuple: |
| <question, passage, label>. The label indicates |
| whether the passage is the answer of the question |
| (1), or not (0). This dataset is constructed based on |
| a commercial search engine. Accuracy (ACC) of |
| the binary classification should be used as the metric. |
| """ |
| ), |
| data_dir="QAM", |
| citation="", |
| url="", |
| ), |
| XGlueConfig( |
| name="qg", |
| description=textwrap.dedent( |
| """\ |
| The Question Generation (QG) task aims to |
| generate a question for a given passage. <passage, question> pairs were collected from a commercial search engine. It covers 6 languages, including English, French, German, Spanish, Italian and |
| Portuguese. BLEU-4 score should be used as the metric. |
| """ |
| ), |
| data_dir="QG", |
| citation="", |
| url="", |
| ), |
| XGlueConfig( |
| name="ntg", |
| description=textwrap.dedent( |
| """\ |
| News Title Generation (NTG) task aims |
| to generate a proper title for a given news body. |
| We collect <news body, news title> pairs from a |
| commercial news website. It covers 5 languages, |
| including German, English, French, Spanish and |
| Russian. BLEU-4 score should be used as the metric. |
| """ |
| ), |
| data_dir="NTG", |
| citation="", |
| url="", |
| ), |
| ] |
|
|
| def _info(self): |
| if self.config.name == "ner": |
| features = { |
| "words": datasets.Sequence(datasets.Value("string")), |
| "ner": datasets.Sequence( |
| datasets.features.ClassLabel( |
| names=[ |
| "O", |
| "B-PER", |
| "I-PER", |
| "B-ORG", |
| "I-ORG", |
| "B-LOC", |
| "I-LOC", |
| "B-MISC", |
| "I-MISC", |
| ] |
| ) |
| ), |
| } |
| elif self.config.name == "pos": |
| features = { |
| "words": datasets.Sequence(datasets.Value("string")), |
| "pos": datasets.Sequence( |
| datasets.features.ClassLabel( |
| names=[ |
| "ADJ", |
| "ADP", |
| "ADV", |
| "AUX", |
| "CCONJ", |
| "DET", |
| "INTJ", |
| "NOUN", |
| "NUM", |
| "PART", |
| "PRON", |
| "PROPN", |
| "PUNCT", |
| "SCONJ", |
| "SYM", |
| "VERB", |
| "X", |
| ] |
| ) |
| ), |
| } |
| elif self.config.name == "mlqa": |
| features = { |
| "context": datasets.Value("string"), |
| "question": datasets.Value("string"), |
| "answers": datasets.features.Sequence( |
| {"answer_start": datasets.Value("int32"), "text": datasets.Value("string")} |
| ), |
| |
| } |
| elif self.config.name == "nc": |
| features = { |
| "news_title": datasets.Value("string"), |
| "news_body": datasets.Value("string"), |
| "news_category": datasets.ClassLabel( |
| names=[ |
| "foodanddrink", |
| "sports", |
| "travel", |
| "finance", |
| "lifestyle", |
| "news", |
| "entertainment", |
| "health", |
| "video", |
| "autos", |
| ] |
| ), |
| } |
| elif self.config.name == "xnli": |
| features = { |
| "premise": datasets.Value("string"), |
| "hypothesis": datasets.Value("string"), |
| "label": datasets.features.ClassLabel(names=["entailment", "neutral", "contradiction"]), |
| } |
| elif self.config.name == "paws-x": |
| features = { |
| "sentence1": datasets.Value("string"), |
| "sentence2": datasets.Value("string"), |
| "label": datasets.features.ClassLabel(names=["different", "same"]), |
| } |
| elif self.config.name == "qadsm": |
| features = { |
| "query": datasets.Value("string"), |
| "ad_title": datasets.Value("string"), |
| "ad_description": datasets.Value("string"), |
| "relevance_label": datasets.features.ClassLabel(names=["Bad", "Good"]), |
| } |
| elif self.config.name == "wpr": |
| features = { |
| "query": datasets.Value("string"), |
| "web_page_title": datasets.Value("string"), |
| "web_page_snippet": datasets.Value("string"), |
| "relavance_label": datasets.features.ClassLabel(names=["Bad", "Fair", "Good", "Excellent", "Perfect"]), |
| } |
| elif self.config.name == "qam": |
| features = { |
| "question": datasets.Value("string"), |
| "answer": datasets.Value("string"), |
| "label": datasets.features.ClassLabel(names=["False", "True"]), |
| } |
| elif self.config.name == "qg": |
| features = { |
| "answer_passage": datasets.Value("string"), |
| "question": datasets.Value("string"), |
| } |
| elif self.config.name == "ntg": |
| features = { |
| "news_body": datasets.Value("string"), |
| "news_title": datasets.Value("string"), |
| } |
|
|
| return datasets.DatasetInfo( |
| description=_XGLUE_DESCRIPTION, |
| features=datasets.Features(features), |
| homepage=self.config.url, |
| citation=self.config.citation + "\n" + _XGLUE_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| archive = dl_manager.download(_XGLUE_ALL_DATA) |
| data_folder = f"xglue_full_dataset/{self.config.data_dir}" |
| name = self.config.name |
|
|
| languages = _LANGUAGES[name] |
| return ( |
| [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={ |
| "archive": dl_manager.iter_archive(archive), |
| "data_path": f"{data_folder}/{_PATHS[name]['train']}", |
| "split": "train", |
| }, |
| ), |
| ] |
| + [ |
| datasets.SplitGenerator( |
| name=datasets.Split(f"validation.{lang}"), |
| gen_kwargs={ |
| "archive": dl_manager.iter_archive(archive), |
| "data_path": f"{data_folder}/{_PATHS[name]['dev'].format(lang)}", |
| "split": "dev", |
| }, |
| ) |
| for lang in languages |
| ] |
| + [ |
| datasets.SplitGenerator( |
| name=datasets.Split(f"test.{lang}"), |
| gen_kwargs={ |
| "archive": dl_manager.iter_archive(archive), |
| "data_path": f"{data_folder}/{_PATHS[name]['test'].format(lang)}", |
| "split": "test", |
| }, |
| ) |
| for lang in languages |
| ] |
| ) |
|
|
| def _generate_examples(self, archive, data_path, split=None): |
| keys = list(self._info().features.keys()) |
| src_f = tgt_f = None |
| for path, file in archive: |
| if self.config.name == "mlqa": |
| if path == data_path: |
| data = json.load(file) |
| for examples in data["data"]: |
| for example in examples["paragraphs"]: |
| context = example["context"] |
| for qa in example["qas"]: |
| question = qa["question"] |
| id_ = qa["id"] |
| answers = qa["answers"] |
| answers_start = [answer["answer_start"] for answer in answers] |
| answers_text = [answer["text"] for answer in answers] |
| yield id_, { |
| "context": context, |
| "question": question, |
| "answers": {"answer_start": answers_start, "text": answers_text}, |
| } |
| elif self.config.name in ["ner", "pos"]: |
| if path == data_path: |
| words = [] |
| result = [] |
| idx = -1 |
| for line in file: |
| line = line.decode("utf-8") |
| if line.strip() == "": |
| if len(words) > 0: |
| out_dict = {keys[0]: words, keys[1]: result} |
| words = [] |
| result = [] |
| idx += 1 |
| yield idx, out_dict |
| else: |
| splits = line.strip().split(" ") |
| words.append(splits[0]) |
| result.append(splits[1]) |
| elif self.config.name in ["ntg", "qg"]: |
| if path == data_path + ".src." + split: |
| src_f = [line.decode("utf-8") for line in file] |
| elif path == data_path + ".tgt." + split: |
| tgt_f = [line.decode("utf-8") for line in file] |
| if src_f and tgt_f: |
| for idx, (src_line, tgt_line) in enumerate(zip(src_f, tgt_f)): |
| yield idx, {keys[0]: src_line.strip(), keys[1]: tgt_line.strip()} |
| else: |
| _process_dict = { |
| "paws-x": {"0": "different", "1": "same"}, |
| "xnli": {"contradictory": "contradiction"}, |
| "qam": {"0": "False", "1": "True"}, |
| "wpr": {"0": "Bad", "1": "Fair", "2": "Good", "3": "Excellent", "4": "Perfect"}, |
| } |
|
|
| def _process(value): |
| if self.config.name in _process_dict and value in _process_dict[self.config.name]: |
| return _process_dict[self.config.name][value] |
| return value |
|
|
| if path == data_path: |
| for idx, line in enumerate(file): |
| line = line.decode("utf-8") |
| if data_path.split(".")[-1] == "tsv" and idx == 0: |
| continue |
| items = line.strip().split("\t") |
| yield idx, { |
| key: _process(value) |
| for key, value in zip(keys, items[1:] if self.config.name == "paws-x" else items) |
| } |
|
|