| from pathlib import Path |
| from typing import List |
|
|
| import datasets |
|
|
| from seacrowd.utils import schemas |
| from seacrowd.utils.configs import SEACrowdConfig |
| from seacrowd.utils.constants import Licenses, Tasks |
|
|
| _DATASETNAME = "sap_wat" |
|
|
| _LANGUAGES = ["eng", "ind", "zlm", "tha", "vie"] |
|
|
| _CITATION = """\ |
| @inproceedings{buschbeck-exel-2020-parallel, |
| title = "A Parallel Evaluation Data Set of Software Documentation with Document Structure Annotation", |
| author = "Buschbeck, Bianka and |
| Exel, Miriam", |
| editor = "Nakazawa, Toshiaki and |
| Nakayama, Hideki and |
| Ding, Chenchen and |
| Dabre, Raj and |
| Kunchukuttan, Anoop and |
| Pa, Win Pa and |
| Bojar, Ond{\v{r}}ej and |
| Parida, Shantipriya and |
| Goto, Isao and |
| Mino, Hidaya and |
| Manabe, Hiroshi and |
| Sudoh, Katsuhito and |
| Kurohashi, Sadao and |
| Bhattacharyya, Pushpak", |
| booktitle = "Proceedings of the 7th Workshop on Asian Translation", |
| month = dec, |
| year = "2020", |
| address = "Suzhou, China", |
| publisher = "Association for Computational Linguistics", |
| url = "https://aclanthology.org/2020.wat-1.20", |
| pages = "160--169", |
| abstract = "This paper accompanies the software documentation data set for machine translation, a parallel |
| evaluation data set of data originating from the SAP Help Portal, that we released to the machine translation |
| community for research purposes. It offers the possibility to tune and evaluate machine translation systems |
| in the domain of corporate software documentation and contributes to the availability of a wider range of |
| evaluation scenarios. The data set comprises of the language pairs English to Hindi, Indonesian, Malay and |
| Thai, and thus also increases the test coverage for the many low-resource language pairs. Unlike most evaluation |
| data sets that consist of plain parallel text, the segments in this data set come with additional metadata that |
| describes structural information of the document context. We provide insights into the origin and creation, the |
| particularities and characteristics of the data set as well as machine translation results.", |
| } |
| |
| """ |
|
|
| _DESCRIPTION = """The data set originates from the SAP Help Portal that contains documentation for SAP products and user |
| assistance for product-related questions. The data has been processed in a way that makes it suitable as development and |
| test data for machine translation purposes. The current language scope is English to Hindi, Indonesian, Japanese, Korean, |
| Malay, Thai, Vietnamese, Simplified Chinese and Traditional Chinese. For each language pair about 4k segments are available, |
| split into development and test data. The segments are provided in their document context and are annotated with additional |
| metadata from the document.""" |
|
|
| _HOMEPAGE = "https://github.com/SAP/software-documentation-data-set-for-machine-translation" |
|
|
| _LICENSE = Licenses.CC_BY_NC_4_0.value |
|
|
| _URLs = { |
| _DATASETNAME: "https://raw.githubusercontent.com/SAP/software-documentation-data-set-for-machine-translation/master/{split}_data/en{lang}/software_documentation.{split}.en{lang}.{appx}" |
| } |
|
|
| _SUPPORTED_TASKS = [ |
| Tasks.MACHINE_TRANSLATION |
| ] |
|
|
| _SOURCE_VERSION = "1.0.0" |
| _SEACROWD_VERSION = "2024.06.20" |
|
|
| _SUBSET = ["id", "ms", "th", "vi"] |
|
|
| _LOCAL = False |
|
|
| class SapWatDataset(datasets.GeneratorBasedBuilder): |
| """SAP WAT is a software documentation dataset for machine translation. The current language scope is English to Hindi, |
| Indonesian, Japanese, Korean, Malay, Thai, Vietnamese, Simplified Chinese and Traditional Chinese. Here, we only consider |
| EN-ID, EN-TH, EN-MS, EN-VI""" |
|
|
| BUILDER_CONFIGS = [ |
| SEACrowdConfig( |
| name=f"{_DATASETNAME}_en_{lang}_source", |
| version=datasets.Version(_SOURCE_VERSION), |
| description=f"SAP WAT source schema for EN-{lang.upper()}", |
| schema="source", |
| subset_id=f"{_DATASETNAME}_en_{lang}", |
| ) |
| for lang in _SUBSET] + [ |
| SEACrowdConfig( |
| name=f"{_DATASETNAME}_en_{lang}_seacrowd_t2t", |
| version=datasets.Version(_SEACROWD_VERSION), |
| description=f"SAP WAT SEACrowd schema for EN-{lang.upper()}", |
| schema="seacrowd_t2t", |
| subset_id=f"{_DATASETNAME}_en_{lang}", |
| ) |
| for lang in _SUBSET |
| ] |
|
|
| DEFAULT_CONFIG_NAME = "sap_wat_en_id_source" |
|
|
| def _info(self): |
| if self.config.schema == "source": |
| features = datasets.Features( |
| { |
| "id": datasets.Value("string"), |
| "text": datasets.Value("string"), |
| "label": datasets.Value("string") |
| } |
| ) |
| elif self.config.schema == "seacrowd_t2t": |
| features = schemas.text2text_features |
|
|
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=features, |
| homepage=_HOMEPAGE, |
| license=_LICENSE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators( |
| self, dl_manager: datasets.DownloadManager |
| ) -> List[datasets.SplitGenerator]: |
| lang = self.config.name.split("_")[3] |
|
|
| splits = {datasets.Split.VALIDATION: "dev", datasets.Split.TEST: "test"} |
| data_urls = { |
| split: _URLs[_DATASETNAME].format(split=splits[split], lang=lang, appx=lang) for split in splits |
| } |
| dl_paths = dl_manager.download(data_urls) |
|
|
| en_data_urls = { |
| split: _URLs[_DATASETNAME].format(split=splits[split], lang=lang, appx="en") for split in splits |
| } |
| en_dl_paths = dl_manager.download(en_data_urls) |
| return [ |
| datasets.SplitGenerator( |
| name=split, |
| gen_kwargs={"filepath": dl_paths[split], "en_filepath": en_dl_paths[split]}, |
| ) |
| for split in splits |
| ] |
|
|
| def _generate_examples(self, filepath: Path, en_filepath: Path): |
| with open(en_filepath, "r") as f: |
| lines_1 = f.readlines() |
| with open(filepath, "r") as f: |
| lines_2 = f.readlines() |
|
|
| if self.config.schema == "source": |
| for _id, (line_1, line_2) in enumerate(zip(lines_1, lines_2)): |
| ex = { |
| "id": _id, |
| "text": line_1.strip(), |
| "label": line_2.strip() |
| } |
| yield _id, ex |
|
|
| elif self.config.schema == "seacrowd_t2t": |
| lang = self.config.name.split("_")[3] |
| lang_name = _LANGUAGES[_SUBSET.index(lang)+1] |
|
|
| for _id, (line_1, line_2) in enumerate(zip(lines_1, lines_2)): |
| ex = { |
| "id": _id, |
| "text_1": line_1.strip(), |
| "text_2": line_2.strip(), |
| "text_1_name": 'eng', |
| "text_2_name": lang_name, |
| } |
| yield _id, ex |
| else: |
| raise ValueError(f"Invalid config: {self.config.name}") |