| import json |
| from typing import List |
|
|
| import datasets |
|
|
| from seacrowd.utils import schemas |
| from seacrowd.utils.configs import SEACrowdConfig |
| from seacrowd.utils.constants import Licenses, Tasks |
|
|
| _CITATION = """\ |
| @misc{fitzgerald2022massive, |
| title={MASSIVE: A 1M-Example Multilingual Natural Language Understanding Dataset with 51 Typologically-Diverse Languages}, |
| author={Jack FitzGerald and Christopher Hench and Charith Peris and Scott Mackie and Kay Rottmann and Ana Sanchez and Aaron |
| Nash and Liam Urbach and Vishesh Kakarala and Richa Singh and Swetha Ranganath and Laurie Crist and Misha Britan and Wouter |
| Leeuwis and Gokhan Tur and Prem Natarajan}, |
| year={2022}, |
| eprint={2204.08582}, |
| archivePrefix={arXiv}, |
| primaryClass={cs.CL} |
| } |
| @inproceedings{bastianelli-etal-2020-slurp, |
| title = "{SLURP}: A Spoken Language Understanding Resource Package", |
| author = "Bastianelli, Emanuele and |
| Vanzo, Andrea and |
| Swietojanski, Pawel and |
| Rieser, Verena", |
| booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
| month = nov, |
| year = "2020", |
| address = "Online", |
| publisher = "Association for Computational Linguistics", |
| url = "https://aclanthology.org/2020.emnlp-main.588", |
| doi = "10.18653/v1/2020.emnlp-main.588", |
| pages = "7252--7262", |
| abstract = "Spoken Language Understanding infers semantic meaning directly from audio data, and thus promises to |
| reduce error propagation and misunderstandings in end-user applications. However, publicly available SLU resources are limited. |
| In this paper, we release SLURP, a new SLU package containing the following: (1) A new challenging dataset in English spanning |
| 18 domains, which is substantially bigger and linguistically more diverse than existing datasets; (2) Competitive baselines |
| based on state-of-the-art NLU and ASR systems; (3) A new transparent metric for entity labelling which enables a detailed error |
| analysis for identifying potential areas of improvement. SLURP is available at https://github.com/pswietojanski/slurp." |
| } |
| """ |
| _DATASETNAME = "massive" |
| _DESCRIPTION = """\ |
| MASSIVE dataset—Multilingual Amazon Slu resource package (SLURP) for Slot-filling, Intent classification, and |
| Virtual assistant Evaluation. MASSIVE contains 1M realistic, parallel, labeled virtual assistant utterances |
| spanning 18 domains, 60 intents, and 55 slots. MASSIVE was created by tasking professional translators to |
| localize the English-only SLURP dataset into 50 typologically diverse languages, including 8 native languages |
| and 2 other languages mostly spoken in Southeast Asia. |
| """ |
| _HOMEPAGE = "https://github.com/alexa/massive" |
| _LICENSE = Licenses.CC_BY_4_0.value |
| _LOCAL = False |
| _LANGUAGES = ["ind", "jav", "khm", "zlm", "mya", "tha", "tgl", "vie"] |
|
|
| _URLS = { |
| _DATASETNAME: "https://amazon-massive-nlu-dataset.s3.amazonaws.com/amazon-massive-dataset-1.1.tar.gz", |
| } |
| _SUPPORTED_TASKS = [Tasks.INTENT_CLASSIFICATION, Tasks.SLOT_FILLING] |
| _SOURCE_VERSION = "1.1.0" |
| _SEACROWD_VERSION = "2024.06.20" |
|
|
| |
| _LANGS = [ |
| "af-ZA", |
| "am-ET", |
| "ar-SA", |
| "az-AZ", |
| "bn-BD", |
| "cy-GB", |
| "da-DK", |
| "de-DE", |
| "el-GR", |
| "en-US", |
| "es-ES", |
| "fa-IR", |
| "fi-FI", |
| "fr-FR", |
| "he-IL", |
| "hi-IN", |
| "hu-HU", |
| "hy-AM", |
| "id-ID", |
| "is-IS", |
| "it-IT", |
| "ja-JP", |
| "jv-ID", |
| "ka-GE", |
| "km-KH", |
| "kn-IN", |
| "ko-KR", |
| "lv-LV", |
| "ml-IN", |
| "mn-MN", |
| "ms-MY", |
| "my-MM", |
| "nb-NO", |
| "nl-NL", |
| "pl-PL", |
| "pt-PT", |
| "ro-RO", |
| "ru-RU", |
| "sl-SL", |
| "sq-AL", |
| "sv-SE", |
| "sw-KE", |
| "ta-IN", |
| "te-IN", |
| "th-TH", |
| "tl-PH", |
| "tr-TR", |
| "ur-PK", |
| "vi-VN", |
| "zh-CN", |
| "zh-TW", |
| ] |
| _SUBSETS = ["id-ID", "jv-ID", "km-KH", "ms-MY", "my-MM", "th-TH", "tl-PH", "vi-VN"] |
|
|
| _SCENARIOS = ["calendar", "recommendation", "social", "general", "news", "cooking", "iot", "email", "weather", "alarm", "transport", "lists", "takeaway", "play", "audio", "music", "qa", "datetime"] |
|
|
| _INTENTS = [ |
| "audio_volume_other", |
| "play_music", |
| "iot_hue_lighton", |
| "general_greet", |
| "calendar_set", |
| "audio_volume_down", |
| "social_query", |
| "audio_volume_mute", |
| "iot_wemo_on", |
| "iot_hue_lightup", |
| "audio_volume_up", |
| "iot_coffee", |
| "takeaway_query", |
| "qa_maths", |
| "play_game", |
| "cooking_query", |
| "iot_hue_lightdim", |
| "iot_wemo_off", |
| "music_settings", |
| "weather_query", |
| "news_query", |
| "alarm_remove", |
| "social_post", |
| "recommendation_events", |
| "transport_taxi", |
| "takeaway_order", |
| "music_query", |
| "calendar_query", |
| "lists_query", |
| "qa_currency", |
| "recommendation_movies", |
| "general_joke", |
| "recommendation_locations", |
| "email_querycontact", |
| "lists_remove", |
| "play_audiobook", |
| "email_addcontact", |
| "lists_createoradd", |
| "play_radio", |
| "qa_stock", |
| "alarm_query", |
| "email_sendemail", |
| "general_quirky", |
| "music_likeness", |
| "cooking_recipe", |
| "email_query", |
| "datetime_query", |
| "transport_traffic", |
| "play_podcasts", |
| "iot_hue_lightchange", |
| "calendar_remove", |
| "transport_query", |
| "transport_ticket", |
| "qa_factoid", |
| "iot_cleaning", |
| "alarm_set", |
| "datetime_convert", |
| "iot_hue_lightoff", |
| "qa_definition", |
| "music_dislikeness", |
| ] |
|
|
| _TAGS = [ |
| "O", |
| "B-food_type", |
| "B-movie_type", |
| "B-person", |
| "B-change_amount", |
| "I-relation", |
| "I-game_name", |
| "B-date", |
| "B-movie_name", |
| "I-person", |
| "I-place_name", |
| "I-podcast_descriptor", |
| "I-audiobook_name", |
| "B-email_folder", |
| "B-coffee_type", |
| "B-app_name", |
| "I-time", |
| "I-coffee_type", |
| "B-transport_agency", |
| "B-podcast_descriptor", |
| "I-playlist_name", |
| "B-media_type", |
| "B-song_name", |
| "I-music_descriptor", |
| "I-song_name", |
| "B-event_name", |
| "I-timeofday", |
| "B-alarm_type", |
| "B-cooking_type", |
| "I-business_name", |
| "I-color_type", |
| "B-podcast_name", |
| "I-personal_info", |
| "B-weather_descriptor", |
| "I-list_name", |
| "B-transport_descriptor", |
| "I-game_type", |
| "I-date", |
| "B-place_name", |
| "B-color_type", |
| "B-game_name", |
| "I-artist_name", |
| "I-drink_type", |
| "B-business_name", |
| "B-timeofday", |
| "B-sport_type", |
| "I-player_setting", |
| "I-transport_agency", |
| "B-game_type", |
| "B-player_setting", |
| "I-music_album", |
| "I-event_name", |
| "I-general_frequency", |
| "I-podcast_name", |
| "I-cooking_type", |
| "I-radio_name", |
| "I-joke_type", |
| "I-meal_type", |
| "I-transport_type", |
| "B-joke_type", |
| "B-time", |
| "B-order_type", |
| "B-business_type", |
| "B-general_frequency", |
| "I-food_type", |
| "I-time_zone", |
| "B-currency_name", |
| "B-time_zone", |
| "B-ingredient", |
| "B-house_place", |
| "B-audiobook_name", |
| "I-ingredient", |
| "I-media_type", |
| "I-news_topic", |
| "B-music_genre", |
| "I-definition_word", |
| "B-list_name", |
| "B-playlist_name", |
| "B-email_address", |
| "I-currency_name", |
| "I-movie_name", |
| "I-device_type", |
| "I-weather_descriptor", |
| "B-audiobook_author", |
| "I-audiobook_author", |
| "I-app_name", |
| "I-order_type", |
| "I-transport_name", |
| "B-radio_name", |
| "I-business_type", |
| "B-definition_word", |
| "B-artist_name", |
| "I-movie_type", |
| "B-transport_name", |
| "I-email_folder", |
| "B-music_album", |
| "I-house_place", |
| "I-music_genre", |
| "B-drink_type", |
| "I-alarm_type", |
| "B-music_descriptor", |
| "B-news_topic", |
| "B-meal_type", |
| "I-transport_descriptor", |
| "I-email_address", |
| "I-change_amount", |
| "B-device_type", |
| "B-transport_type", |
| "B-relation", |
| "I-sport_type", |
| "B-personal_info", |
| ] |
|
|
|
|
| class MASSIVEDataset(datasets.GeneratorBasedBuilder): |
| """MASSIVE datasets contains datasets to detect the intent from the text and fill the dialogue slots""" |
|
|
| BUILDER_CONFIGS = ( |
| [ |
| SEACrowdConfig( |
| name=f"massive_{subset}_source", |
| version=datasets.Version(_SOURCE_VERSION), |
| description=f"MASSIVE source schema for {subset}", |
| schema="source", |
| subset_id="massive_" + subset, |
| ) |
| for subset in _SUBSETS |
| ] |
| + [ |
| SEACrowdConfig( |
| name=f"massive_{subset}_seacrowd_text", |
| version=datasets.Version(_SEACROWD_VERSION), |
| description=f"MASSIVE Nusantara intent classification schema for {subset}", |
| schema="seacrowd_text", |
| subset_id="massive_intent_" + subset, |
| ) |
| for subset in _SUBSETS |
| ] |
| + [ |
| SEACrowdConfig( |
| name=f"massive_{subset}_seacrowd_seq_label", |
| version=datasets.Version(_SEACROWD_VERSION), |
| description=f"MASSIVE Nusantara slot filling schema for {subset}", |
| schema="seacrowd_seq_label", |
| subset_id="massive_slot_filling_" + subset, |
| ) |
| for subset in _SUBSETS |
| ] |
| + [ |
| SEACrowdConfig( |
| name="massive_source", |
| version=datasets.Version(_SOURCE_VERSION), |
| description="MASSIVE source schema", |
| schema="source", |
| subset_id="massive", |
| ), |
| SEACrowdConfig( |
| name="massive_seacrowd_text", |
| version=datasets.Version(_SEACROWD_VERSION), |
| description="MASSIVE Nusantara intent classification schema", |
| schema="seacrowd_text", |
| subset_id="massive_intent", |
| ), |
| SEACrowdConfig( |
| name="massive_seacrowd_seq_label", |
| version=datasets.Version(_SEACROWD_VERSION), |
| description="MASSIVE Nusantara slot filling schema", |
| schema="seacrowd_seq_label", |
| subset_id="massive_slot_filling", |
| ), |
| ] |
| ) |
|
|
| DEFAULT_CONFIG_NAME = "massive_id-ID_source" |
|
|
| def _info(self) -> datasets.DatasetInfo: |
| if self.config.schema == "source": |
| features = datasets.Features( |
| { |
| "id": datasets.Value("string"), |
| "locale": datasets.Value("string"), |
| "partition": datasets.Value("string"), |
| "scenario": datasets.features.ClassLabel(names=_SCENARIOS), |
| "intent": datasets.features.ClassLabel(names=_INTENTS), |
| "utt": datasets.Value("string"), |
| "annot_utt": datasets.Value("string"), |
| "tokens": datasets.Sequence(datasets.Value("string")), |
| "ner_tags": datasets.Sequence(datasets.features.ClassLabel(names=_TAGS)), |
| "worker_id": datasets.Value("string"), |
| "slot_method": datasets.Sequence( |
| { |
| "slot": datasets.Value("string"), |
| "method": datasets.Value("string"), |
| } |
| ), |
| "judgments": datasets.Sequence( |
| { |
| "worker_id": datasets.Value("string"), |
| "intent_score": datasets.Value("int8"), |
| "slots_score": datasets.Value("int8"), |
| "grammar_score": datasets.Value("int8"), |
| "spelling_score": datasets.Value("int8"), |
| "language_identification": datasets.Value("string"), |
| } |
| ), |
| } |
| ) |
| elif self.config.schema == "seacrowd_text": |
| features = schemas.text_features(label_names=_INTENTS) |
| elif self.config.schema == "seacrowd_seq_label": |
| features = schemas.seq_label_features(label_names=_TAGS) |
| else: |
| raise ValueError(f"Invalid config schema: {self.config.schema}") |
|
|
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=features, |
| homepage=_HOMEPAGE, |
| license=_LICENSE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager: datasets.DownloadManager) -> List[datasets.SplitGenerator]: |
| archive = dl_manager.download(_URLS[_DATASETNAME]) |
|
|
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={ |
| "files": dl_manager.iter_archive(archive), |
| "split": "train", |
| "lang": self.config.name, |
| }, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.VALIDATION, |
| gen_kwargs={ |
| "files": dl_manager.iter_archive(archive), |
| "split": "dev", |
| "lang": self.config.name, |
| }, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.TEST, |
| gen_kwargs={ |
| "files": dl_manager.iter_archive(archive), |
| "split": "test", |
| "lang": self.config.name, |
| }, |
| ), |
| ] |
|
|
| def _get_bio_format(self, text): |
| """This function is modified from https://huggingface.co/datasets/qanastek/MASSIVE/blob/main/MASSIVE.py""" |
| tags, tokens = [], [] |
|
|
| bio_mode = False |
| cpt_bio = 0 |
| current_tag = None |
|
|
| split_iter = iter(text.split(" ")) |
|
|
| for s in split_iter: |
| if s.startswith("["): |
| current_tag = s.strip("[") |
| bio_mode = True |
| cpt_bio += 1 |
| next(split_iter) |
| continue |
|
|
| elif s.endswith("]"): |
| bio_mode = False |
| if cpt_bio == 1: |
| prefix = "B-" |
| else: |
| prefix = "I-" |
| token = prefix + current_tag |
| word = s.strip("]") |
| current_tag = None |
| cpt_bio = 0 |
|
|
| else: |
| if bio_mode: |
| if cpt_bio == 1: |
| prefix = "B-" |
| else: |
| prefix = "I-" |
| token = prefix + current_tag |
| word = s |
| cpt_bio += 1 |
| else: |
| token = "O" |
| word = s |
|
|
| tags.append(token) |
| tokens.append(word) |
|
|
| return tokens, tags |
|
|
| def _generate_examples(self, files: list, split: str, lang: str): |
| _id = 0 |
|
|
| lang = lang.replace("massive_", "").replace("source", "").replace("seacrowd_text", "").replace("seacrowd_seq_label", "") |
|
|
| if not lang: |
| lang = _LANGS.copy() |
| else: |
| lang = [lang[:-1]] |
|
|
| |
|
|
| for path, f in files: |
| curr_lang = path.split(f"{_SOURCE_VERSION[:-2]}/data/")[-1].split(".jsonl")[0] |
|
|
| if not lang: |
| break |
| elif curr_lang in lang: |
| lang.remove(curr_lang) |
| else: |
| continue |
|
|
| |
| lines = f.read().decode(encoding="utf-8").split("\n") |
|
|
| for line in lines: |
| data = json.loads(line) |
|
|
| if data["partition"] != split: |
| continue |
|
|
| |
| if "slot_method" in data: |
| slot_method = [ |
| { |
| "slot": s["slot"], |
| "method": s["method"], |
| } |
| for s in data["slot_method"] |
| ] |
| else: |
| slot_method = [] |
|
|
| |
| if "judgments" in data: |
| judgments = [ |
| { |
| "worker_id": j["worker_id"], |
| "intent_score": j["intent_score"], |
| "slots_score": j["slots_score"], |
| "grammar_score": j["grammar_score"], |
| "spelling_score": j["spelling_score"], |
| "language_identification": j["language_identification"] if "language_identification" in j else "target", |
| } |
| for j in data["judgments"] |
| ] |
| else: |
| judgments = [] |
|
|
| if self.config.schema == "source": |
| tokens, tags = self._get_bio_format(data["annot_utt"]) |
|
|
| yield _id, { |
| "id": str(_id) + "_" + data["id"], |
| "locale": data["locale"], |
| "partition": data["partition"], |
| "scenario": data["scenario"], |
| "intent": data["intent"], |
| "utt": data["utt"], |
| "annot_utt": data["annot_utt"], |
| "tokens": tokens, |
| "ner_tags": tags, |
| "worker_id": data["worker_id"], |
| "slot_method": slot_method, |
| "judgments": judgments, |
| } |
|
|
| elif self.config.schema == "seacrowd_seq_label": |
| tokens, tags = self._get_bio_format(data["annot_utt"]) |
|
|
| yield _id, { |
| "id": str(_id) + "_" + data["id"], |
| "tokens": tokens, |
| "labels": tags, |
| } |
|
|
| elif self.config.schema == "seacrowd_text": |
| yield _id, { |
| "id": str(_id) + "_" + data["id"], |
| "text": data["utt"], |
| "label": data["intent"], |
| } |
|
|
| else: |
| raise ValueError(f"Invalid config: {self.config.name}") |
|
|
| _id += 1 |
|
|