FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /tasks /Retrieval /multilingual /CrossLingualSemanticDiscriminationWMT21.py
| from typing import Dict, List | |
| import datasets | |
| from mteb.abstasks import AbsTaskRetrieval, CrosslingualTask, TaskMetadata | |
| _LANGUAGES = { | |
| "wmt21.de.fr": ["deu-Latn", "fra-Latn"], | |
| "wmt21.fr.de": ["fra-Latn", "deu-Latn"], | |
| } | |
| def _build_lang_pair(langs: List[str]) -> str: | |
| """Builds a language pair separated by a dash. | |
| e.g., ['eng-Latn', 'deu-Latn'] -> 'eng-deu'. | |
| """ | |
| return langs[0].split("-")[0] + "-" + langs[1].split("-")[0] | |
| def extend_lang_pairs() -> Dict[str, List[str]]: | |
| eval_langs = {} | |
| for langs in _LANGUAGES.values(): | |
| lang_pair = _build_lang_pair(langs) | |
| eval_langs[lang_pair] = langs | |
| return eval_langs | |
| _EVAL_LANGS = extend_lang_pairs() | |
| class CrossLingualSemanticDiscriminationWMT21(AbsTaskRetrieval, CrosslingualTask): | |
| metadata = TaskMetadata( | |
| name="CrossLingualSemanticDiscriminationWMT21", | |
| dataset={ | |
| "path": "Andrianos/clsd_wmt19_21", | |
| "revision": "9627fbdb39b827ee5c066011ebe1e947cdb137bd", | |
| }, | |
| description="Evaluate a multilingual embedding model based on its ability to discriminate against the original parallel pair against challenging distractors - spawning from WMT21 DE-FR test set", | |
| reference="https://huggingface.co/datasets/Andrianos/clsd_wmt19_21", | |
| type="Retrieval", | |
| category="s2s", | |
| eval_splits=["test"], | |
| eval_langs=_EVAL_LANGS, | |
| main_score="recall_at_1", | |
| date=("2020-01-01", "2023-12-12"), | |
| form=["written"], | |
| domains=["News"], | |
| task_subtypes=["Cross-Lingual Semantic Discrimination"], | |
| license="CC BY-SA 4.0", | |
| socioeconomic_status="high", | |
| annotations_creators="derived", | |
| dialect=[], | |
| text_creation="LM-generated and verified", | |
| bibtex_citation="preprint_coming", | |
| n_samples={"test": 1786}, | |
| avg_character_length={"test": 159}, | |
| ) | |
| def __init__(self, **kwargs): | |
| self.num_of_distractors = 4 | |
| super().__init__(**kwargs) | |
| def load_data(self, **kwargs): | |
| """Generic data loader function for original clsd datasets with the format shown in "hf_dataset_link". | |
| Loading the hf dataset, it populates the following three variables to be used for retrieval evaluation. | |
| self.corpus | |
| self.queries | |
| self.relevant_docs | |
| Sets self.data_loaded to True. | |
| """ | |
| if self.data_loaded: | |
| return | |
| queries, corpus, relevant_docs = {}, {}, {} | |
| dataset_raw = {} | |
| for split in self.metadata.eval_splits: | |
| for hf_subset, langs in _LANGUAGES.items(): | |
| lang_pair = _build_lang_pair(langs) | |
| dataset_raw[lang_pair] = datasets.load_dataset( | |
| name=hf_subset, | |
| **self.metadata_dict["dataset"], | |
| )[split] | |
| queries[lang_pair] = {} | |
| corpus[lang_pair] = {} | |
| relevant_docs[lang_pair] = {} | |
| queries[lang_pair][split] = {} | |
| corpus[lang_pair][split] = {} | |
| relevant_docs[lang_pair][split] = {} | |
| # Generate unique IDs for queries and documents | |
| query_id_counter = 1 | |
| document_id_counter = 1 | |
| for row in dataset_raw[lang_pair]: | |
| query_text = row["Source"] | |
| positive_text = [row["Target"]] | |
| negative_texts = [ | |
| row[f"TargetAdv{str(i)}"] | |
| for i in range( | |
| 1, self.num_of_distractors + 1 | |
| ) # Four Distractors. Columns are named TargetAdv1-TargetAdv4 | |
| ] | |
| # Assign unique ID to the query | |
| query_id = f"Q{query_id_counter}" | |
| queries[lang_pair][split][query_id] = query_text | |
| query_id_counter += 1 | |
| # Add true parallel and distractors to corpus with unique id. | |
| for text in positive_text + negative_texts: | |
| doc_id = f"D{document_id_counter}" | |
| corpus[lang_pair][split][doc_id] = {"text": text} | |
| document_id_counter += 1 | |
| # Add relevant document information to relevant_docs for positive texts only | |
| if text in positive_text: | |
| if query_id not in relevant_docs[lang_pair][split]: | |
| relevant_docs[lang_pair][split][query_id] = {} | |
| relevant_docs[lang_pair][split][query_id][doc_id] = 1 | |
| self.corpus = datasets.DatasetDict(corpus) | |
| self.queries = datasets.DatasetDict(queries) | |
| self.relevant_docs = datasets.DatasetDict(relevant_docs) | |
| self.data_loaded = True | |