FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /tasks /BitextMining /multilingual /IndicGenBenchFloresBitextMining.py
| from __future__ import annotations | |
| from typing import Any | |
| import datasets | |
| from mteb.abstasks.TaskMetadata import TaskMetadata | |
| from ....abstasks import AbsTaskBitextMining, CrosslingualTask | |
| _LANGUAGES = [ | |
| "ben-Beng", | |
| "guj_Gujr", | |
| "hin_Deva", | |
| "kan_Knda", | |
| "mal_Mlym", | |
| "mar_Deva", | |
| "tam_Taml", | |
| "tel_Telu", | |
| "urd_Arab", | |
| "asm_Beng", | |
| "bho_Deva", | |
| "nep_Deva", | |
| "ory_Orya", | |
| "pan_Guru", | |
| "pus_Arab", | |
| "san-Deva", | |
| "awa_Deva", | |
| "bgc_Deva", | |
| "bod_Tibt", | |
| "boy_Deva", | |
| "gbm_Deva", | |
| "gom_Deva", | |
| "hne_Deva", | |
| "raj_Deva", | |
| "mai_Deva", | |
| "mni_Mtei", | |
| "mup_Deva", | |
| "mwr_Deva", | |
| "sat_Olck", | |
| ] | |
| _ENG_LANGUAGE = ["eng-Latn"] | |
| _CODE_MAPPING = { | |
| "ben": "bn", | |
| "guj": "gu", | |
| "hin": "hi", | |
| "kan": "kn", | |
| "mal": "ml", | |
| "mar": "mr", | |
| "tam": "ta", | |
| "tel": "te", | |
| "urd": "ur", | |
| "asm": "as", | |
| "bho": "bho", | |
| "nep": "ne", | |
| "ory": "or", | |
| "pan": "pa", | |
| "pus": "ps", | |
| "san": "sa", | |
| "awa": "awa", | |
| "bgc": "bgc", | |
| "bod": "bo", | |
| "boy": "brx", | |
| "gbm": "gbm", | |
| "gom": "gom", | |
| "hne": "hne", | |
| "raj": "hoj", | |
| "mai": "mai", | |
| "mni": "mni", | |
| "mup": "mup", | |
| "mwr": "mwr", | |
| "sat": "sat", | |
| } | |
| _SPLIT = ["validation", "test"] | |
| def get_lang_pairs() -> dict[str, list[str]]: | |
| # add eng-> xx and xx -> eng lang pairs | |
| # Normalize language codes | |
| normalized_languages = [lang.replace("_", "-") for lang in _LANGUAGES] | |
| # Create dictionary for language pairs | |
| language_pairs = {} | |
| for lang in normalized_languages: | |
| lang_code = lang.split("-")[0] | |
| lang_to_eng_key = f"{lang_code}-eng" | |
| eng_to_lang_key = f"eng-{lang_code}" | |
| language_pairs[lang_to_eng_key] = [lang, _ENG_LANGUAGE[0]] | |
| language_pairs[eng_to_lang_key] = [_ENG_LANGUAGE[0], lang] | |
| return language_pairs | |
| _LANGUAGES_MAPPING = get_lang_pairs() | |
| class IndicGenBenchFloresBitextMining(AbsTaskBitextMining, CrosslingualTask): | |
| metadata = TaskMetadata( | |
| name="IndicGenBenchFloresBitextMining", | |
| dataset={ | |
| "path": "google/IndicGenBench_flores_in", | |
| "revision": "f8650438298df086750ff4973661bb58a201a5ee", | |
| "trust_remote_code": True, | |
| }, | |
| description="Flores-IN dataset is an extension of Flores dataset released as a part of the IndicGenBench by Google", | |
| reference="https://github.com/google-research-datasets/indic-gen-bench/", | |
| type="BitextMining", | |
| category="s2s", | |
| eval_splits=_SPLIT, | |
| eval_langs=_LANGUAGES_MAPPING, | |
| main_score="f1", | |
| date=("2023-10-01", "2024-05-01"), | |
| form=["written"], | |
| domains=["Web", "News"], | |
| task_subtypes=[], | |
| license="CC-BY-SA-4.0", | |
| socioeconomic_status="mixed", | |
| annotations_creators="expert-annotated", | |
| dialect=[], | |
| text_creation="human-translated and localized", | |
| bibtex_citation="""@misc{singh2024indicgenbench, | |
| title={IndicGenBench: A Multilingual Benchmark to Evaluate Generation Capabilities of LLMs on Indic Languages}, | |
| author={Harman Singh and Nitish Gupta and Shikhar Bharadwaj and Dinesh Tewari and Partha Talukdar}, | |
| year={2024}, | |
| eprint={2404.16816}, | |
| archivePrefix={arXiv}, | |
| primaryClass={cs.CL} | |
| }""", | |
| n_samples={"validation": 997, "test": 1012}, | |
| avg_character_length={"validation": 126.25, "test": 130.84}, | |
| ) | |
| def load_data(self, **kwargs: Any) -> None: | |
| """Load dataset from HuggingFace hub""" | |
| if self.data_loaded: | |
| return | |
| self.dataset = {} | |
| for lang in self.hf_subsets: | |
| langs = lang.split("-") | |
| source_lang = langs[0] | |
| target_lang = langs[1] | |
| if source_lang == "eng": | |
| coded_target_language = _CODE_MAPPING[target_lang] | |
| language = f"en_{coded_target_language}" | |
| else: | |
| coded_source_language = _CODE_MAPPING[source_lang] | |
| language = f"{coded_source_language}_en" | |
| self.dataset[lang] = datasets.load_dataset( | |
| **self.metadata_dict["dataset"], | |
| field="examples", | |
| data_files={ | |
| "validation": f"flores_{language}_dev.json", | |
| "test": f"flores_{language}_test.json", | |
| }, | |
| ) | |
| self.dataset_transform() | |
| self.data_loaded = True | |
| def dataset_transform(self) -> None: | |
| for lang in self.hf_subsets: | |
| for split in _SPLIT: | |
| self.dataset[lang][split] = self.dataset[lang][split].rename_columns( | |
| {"source": "sentence1", "target": "sentence2"} | |
| ) | |