FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /tasks /Retrieval /multilingual /WikipediaRetrievalMultilingual.py
hc99's picture
Add files using upload-large-folder tool
73cc8d2 verified
raw
history blame
4.95 kB
from __future__ import annotations
from datasets import load_dataset
from mteb.abstasks.TaskMetadata import TaskMetadata
from ....abstasks import MultilingualTask
from ....abstasks.AbsTaskRetrieval import AbsTaskRetrieval
_EVAL_LANGS = {
"bg": ["bul-Cyrl"],
"bn": ["ben-Beng"],
"cs": ["ces-Latn"],
"da": ["dan-Latn"],
"de": ["deu-Latn"],
"en": ["eng-Latn"],
"fa": ["fas-Arab"],
"fi": ["fin-Latn"],
"hi": ["hin-Deva"],
"it": ["ita-Latn"],
"nl": ["nld-Latn"],
"pt": ["por-Latn"],
"ro": ["ron-Latn"],
"sr": ["srp-Cyrl"],
"no": ["nor-Latn"],
"sv": ["swe-Latn"],
}
# adapted from MIRACLRetrieval
def _load_data(
path: str,
langs: list,
split: str,
cache_dir: str = None,
revision_queries: str = None,
revision_corpus: str = None,
revision_qrels: str = None,
):
queries = {lang: {split: {}} for lang in langs}
corpus = {lang: {split: {}} for lang in langs}
qrels = {lang: {split: {}} for lang in langs}
for lang in langs:
queries_path = path
corpus_path = path.replace("queries", "corpus")
qrels_path = path.replace("queries", "qrels")
queries_lang = load_dataset(
queries_path,
lang,
split=split,
cache_dir=cache_dir,
revision=revision_queries,
)
corpus_lang = load_dataset(
corpus_path,
lang,
split=split,
cache_dir=cache_dir,
revision=revision_corpus,
)
qrels_lang = load_dataset(
qrels_path,
lang,
split=split,
cache_dir=cache_dir,
revision=revision_qrels,
)
# don't pass on titles to make task harder
corpus_lang_dict = {doc["_id"]: {"text": doc["text"]} for doc in corpus_lang}
queries_lang_dict = {
query["_id"]: {"text": query["text"]} for query in queries_lang
}
# qrels_lang_dict = {qrel["query-id"]: {qrel["corpus-id"]: qrel["score"]} for qrel in qrels_lang}
qrels_lang_dict = {}
for qrel in qrels_lang:
if qrel["score"] == 0.5:
continue
# score = 0 if qrel["score"] == 0.5 else qrel["score"]
# score = int(score)
score = int(qrel["score"])
qrels_lang_dict[qrel["query-id"]] = {qrel["corpus-id"]: score}
corpus[lang][split] = corpus_lang_dict
queries[lang][split] = queries_lang_dict
qrels[lang][split] = qrels_lang_dict
return corpus, queries, qrels
class WikipediaRetrievalMultilingual(MultilingualTask, AbsTaskRetrieval):
metadata = TaskMetadata(
name="WikipediaRetrievalMultilingual",
description="The dataset is derived from Cohere's wikipedia-2023-11 dataset and contains synthetically generated queries.",
reference="https://huggingface.co/datasets/ellamind/wikipedia-2023-11-retrieval-pt",
dataset={
"path": "ellamind/wikipedia-2023-11-retrieval-multilingual-queries",
"revision": "3b6ea595c94bac3448a2ad167ca2e06abd340d6e", # avoid validation error
"revision_corpus": "f20ac0c449c85358d3d5c72a95f92f1eddc98aa5",
"revision_qrels": "ec88a7bb2da034d538e98e3122d2c98530ca1c8d",
},
type="Retrieval",
category="s2p",
eval_splits=["test"],
eval_langs=_EVAL_LANGS,
main_score="ndcg_at_10",
date=("2023-11-01", "2024-05-15"),
form=["written"],
domains=["Encyclopaedic"],
task_subtypes=["Question answering", "Article retrieval"],
license="cc-by-sa-3.0",
socioeconomic_status="mixed",
annotations_creators="LM-generated",
dialect=[],
text_creation="LM-generated and verified",
bibtex_citation="",
n_samples={
"en": 1500,
"de": 1500,
"it": 1500,
"pt": 1500,
"nl": 1500,
"cs": 1500,
"ro": 1500,
"bg": 1500,
"sr": 1500,
"fi": 1500,
"da": 1500,
"fa": 1500,
"hi": 1500,
"bn": 1500,
"no": 1500,
"sv": 1500,
},
avg_character_length={"test": 452},
)
def load_data(self, **kwargs):
if self.data_loaded:
return
self.corpus, self.queries, self.relevant_docs = _load_data(
path=self.metadata_dict["dataset"]["path"],
langs=self.hf_subsets,
split=self.metadata_dict["eval_splits"][0],
cache_dir=kwargs.get("cache_dir", None),
revision_queries=self.metadata_dict["dataset"]["revision"],
revision_corpus=self.metadata_dict["dataset"]["revision_corpus"],
revision_qrels=self.metadata_dict["dataset"]["revision_qrels"],
)
self.data_loaded = True