FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /tasks /Retrieval /multilingual /MintakaRetrieval.py
| from __future__ import annotations | |
| import datasets | |
| from mteb.abstasks.TaskMetadata import TaskMetadata | |
| from ....abstasks import MultilingualTask | |
| from ....abstasks.AbsTaskRetrieval import AbsTaskRetrieval | |
| _EVAL_SPLIT = "test" | |
| _LANGS = { | |
| "ar": ["ara-Arab"], | |
| "de": ["deu-Latn"], | |
| "es": ["spa-Latn"], | |
| "fr": ["fra-Latn"], | |
| "hi": ["hin-Deva"], | |
| "it": ["ita-Latn"], | |
| "ja": ["jpn-Hira"], | |
| "pt": ["por-Latn"], | |
| } | |
| def _load_mintaka_data( | |
| path: str, langs: list, split: str, cache_dir: str = None, revision: str = None | |
| ): | |
| queries = {lang: {split: {}} for lang in langs} | |
| corpus = {lang: {split: {}} for lang in langs} | |
| relevant_docs = {lang: {split: {}} for lang in langs} | |
| for lang in langs: | |
| data = datasets.load_dataset( | |
| path, | |
| lang, | |
| split=split, | |
| cache_dir=cache_dir, | |
| revision=revision, | |
| ) | |
| question_ids = { | |
| question: _id for _id, question in enumerate(set(data["question"])) | |
| } | |
| answer_ids = {answer: _id for _id, answer in enumerate(set(data["answer"]))} | |
| for row in data: | |
| question = row["question"] | |
| answer = row["answer"] | |
| query_id = f"Q{question_ids[question]}" | |
| queries[lang][split][query_id] = question | |
| doc_id = f"D{answer_ids[answer]}" | |
| corpus[lang][split][doc_id] = {"text": answer} | |
| if query_id not in relevant_docs[lang][split]: | |
| relevant_docs[lang][split][query_id] = {} | |
| relevant_docs[lang][split][query_id][doc_id] = 1 | |
| corpus = datasets.DatasetDict(corpus) | |
| queries = datasets.DatasetDict(queries) | |
| relevant_docs = datasets.DatasetDict(relevant_docs) | |
| return corpus, queries, relevant_docs | |
| class MintakaRetrieval(MultilingualTask, AbsTaskRetrieval): | |
| metadata = TaskMetadata( | |
| name="MintakaRetrieval", | |
| description="MintakaRetrieval", | |
| reference=None, | |
| dataset={ | |
| "path": "jinaai/mintakaqa", | |
| "revision": "efa78cc2f74bbcd21eff2261f9e13aebe40b814e", | |
| }, | |
| type="Retrieval", | |
| category="s2p", | |
| eval_splits=[_EVAL_SPLIT], | |
| eval_langs=_LANGS, | |
| main_score="ndcg_at_10", | |
| date=None, | |
| form=None, | |
| domains=None, | |
| task_subtypes=None, | |
| license=None, | |
| socioeconomic_status=None, | |
| annotations_creators=None, | |
| dialect=None, | |
| text_creation=None, | |
| bibtex_citation=None, | |
| n_samples=None, | |
| avg_character_length=None, | |
| ) | |
| def load_data(self, **kwargs): | |
| if self.data_loaded: | |
| return | |
| self.corpus, self.queries, self.relevant_docs = _load_mintaka_data( | |
| path=self.metadata_dict["dataset"]["path"], | |
| langs=self.metadata.eval_langs, | |
| split=self.metadata_dict["eval_splits"][0], | |
| cache_dir=kwargs.get("cache_dir", None), | |
| revision=self.metadata_dict["dataset"]["revision"], | |
| ) | |
| self.data_loaded = True | |