hc99's picture
Add files using upload-large-folder tool
73cc8d2 verified
raw
history blame
3.05 kB
from __future__ import annotations
import datasets
from mteb.abstasks.TaskMetadata import TaskMetadata
from ....abstasks import MultilingualTask
from ....abstasks.AbsTaskRetrieval import AbsTaskRetrieval
_EVAL_SPLIT = "test"
_LANGS = {
"ar": ["ara-Arab"],
"de": ["deu-Latn"],
"es": ["spa-Latn"],
"fr": ["fra-Latn"],
"hi": ["hin-Deva"],
"it": ["ita-Latn"],
"ja": ["jpn-Hira"],
"pt": ["por-Latn"],
}
def _load_mintaka_data(
path: str, langs: list, split: str, cache_dir: str = None, revision: str = None
):
queries = {lang: {split: {}} for lang in langs}
corpus = {lang: {split: {}} for lang in langs}
relevant_docs = {lang: {split: {}} for lang in langs}
for lang in langs:
data = datasets.load_dataset(
path,
lang,
split=split,
cache_dir=cache_dir,
revision=revision,
)
question_ids = {
question: _id for _id, question in enumerate(set(data["question"]))
}
answer_ids = {answer: _id for _id, answer in enumerate(set(data["answer"]))}
for row in data:
question = row["question"]
answer = row["answer"]
query_id = f"Q{question_ids[question]}"
queries[lang][split][query_id] = question
doc_id = f"D{answer_ids[answer]}"
corpus[lang][split][doc_id] = {"text": answer}
if query_id not in relevant_docs[lang][split]:
relevant_docs[lang][split][query_id] = {}
relevant_docs[lang][split][query_id][doc_id] = 1
corpus = datasets.DatasetDict(corpus)
queries = datasets.DatasetDict(queries)
relevant_docs = datasets.DatasetDict(relevant_docs)
return corpus, queries, relevant_docs
class MintakaRetrieval(MultilingualTask, AbsTaskRetrieval):
metadata = TaskMetadata(
name="MintakaRetrieval",
description="MintakaRetrieval",
reference=None,
dataset={
"path": "jinaai/mintakaqa",
"revision": "efa78cc2f74bbcd21eff2261f9e13aebe40b814e",
},
type="Retrieval",
category="s2p",
eval_splits=[_EVAL_SPLIT],
eval_langs=_LANGS,
main_score="ndcg_at_10",
date=None,
form=None,
domains=None,
task_subtypes=None,
license=None,
socioeconomic_status=None,
annotations_creators=None,
dialect=None,
text_creation=None,
bibtex_citation=None,
n_samples=None,
avg_character_length=None,
)
def load_data(self, **kwargs):
if self.data_loaded:
return
self.corpus, self.queries, self.relevant_docs = _load_mintaka_data(
path=self.metadata_dict["dataset"]["path"],
langs=self.metadata.eval_langs,
split=self.metadata_dict["eval_splits"][0],
cache_dir=kwargs.get("cache_dir", None),
revision=self.metadata_dict["dataset"]["revision"],
)
self.data_loaded = True