hc99's picture
Add files using upload-large-folder tool
73cc8d2 verified
raw
history blame
2.53 kB
import datasets
from mteb.abstasks.TaskMetadata import TaskMetadata
from ....abstasks.AbsTaskRetrieval import AbsTaskRetrieval
class LEMBWikimQARetrieval(AbsTaskRetrieval):
_EVAL_SPLIT = "test"
metadata = TaskMetadata(
name="LEMBWikimQARetrieval",
dataset={
"path": "dwzhu/LongEmbed",
"revision": "6e346642246bfb4928c560ee08640dc84d074e8c",
"name": "2wikimqa",
},
reference="https://huggingface.co/datasets/dwzhu/LongEmbed",
description=("2wikimqa subset of dwzhu/LongEmbed dataset."),
type="Retrieval",
category="s2p",
eval_splits=[_EVAL_SPLIT],
eval_langs=["eng-Latn"],
main_score="ndcg_at_10",
date=("1950-01-01", "2019-12-31"),
form=["written"],
domains=["Encyclopaedic"],
task_subtypes=["Article retrieval"],
license="Not specified",
socioeconomic_status="medium",
annotations_creators="derived",
dialect=[],
text_creation="found",
bibtex_citation="""
@inproceedings{ho2020constructing,
title={Constructing A Multi-hop QA Dataset for Comprehensive Evaluation of Reasoning Steps},
author={Ho, Xanh and Nguyen, Anh-Khoa Duong and Sugawara, Saku and Aizawa, Akiko},
booktitle={Proceedings of the 28th International Conference on Computational Linguistics},
pages={6609--6625},
year={2020}
}
""",
n_samples={_EVAL_SPLIT: 500},
avg_character_length={_EVAL_SPLIT: 37513},
)
def load_data(self, **kwargs):
if self.data_loaded:
return
query_list = datasets.load_dataset(**self.metadata_dict["dataset"])[
"queries"
] # dict_keys(['qid', 'text'])
queries = {row["qid"]: row["text"] for row in query_list}
corpus_list = datasets.load_dataset(**self.metadata_dict["dataset"])[
"corpus"
] # dict_keys(['doc_id', 'text'])
corpus = {row["doc_id"]: {"text": row["text"]} for row in corpus_list}
qrels_list = datasets.load_dataset(**self.metadata_dict["dataset"])[
"qrels"
] # dict_keys(['qid', 'doc_id'])
qrels = {row["qid"]: {row["doc_id"]: 1} for row in qrels_list}
self.corpus = {self._EVAL_SPLIT: corpus}
self.queries = {self._EVAL_SPLIT: queries}
self.relevant_docs = {self._EVAL_SPLIT: qrels}
self.data_loaded = True