hc99's picture
Add files using upload-large-folder tool
73cc8d2 verified
raw
history blame
3.64 kB
import datasets
from mteb.abstasks.TaskMetadata import TaskMetadata
from ....abstasks.AbsTaskRetrieval import AbsTaskRetrieval
class LEMBPasskeyRetrieval(AbsTaskRetrieval):
_EVAL_SPLIT = [
"test_256",
"test_512",
"test_1024",
"test_2048",
"test_4096",
"test_8192",
"test_16384",
"test_32768",
]
metadata = TaskMetadata(
name="LEMBPasskeyRetrieval",
dataset={
"path": "dwzhu/LongEmbed",
"revision": "6e346642246bfb4928c560ee08640dc84d074e8c",
"name": "passkey",
},
reference="https://huggingface.co/datasets/dwzhu/LongEmbed",
description=("passkey subset of dwzhu/LongEmbed dataset."),
type="Retrieval",
category="s2p",
eval_splits=_EVAL_SPLIT,
eval_langs=["eng-Latn"],
main_score="ndcg_at_10",
date=("2000-01-01", "2023-12-31"),
form=["written"],
domains=["Fiction"],
task_subtypes=["Article retrieval"],
license="Not specified",
socioeconomic_status="low",
annotations_creators="derived",
dialect=[],
text_creation="found",
bibtex_citation="""
@article{zhu2024longembed,
title={LongEmbed: Extending Embedding Models for Long Context Retrieval},
author={Zhu, Dawei and Wang, Liang and Yang, Nan and Song, Yifan and Wu, Wenhao and Wei, Furu and Li, Sujian},
journal={arXiv preprint arXiv:2404.12096},
year={2024}
}
""",
n_samples={
"test_256": 150,
"test_512": 150,
"test_1024": 150,
"test_2048": 150,
"test_4096": 150,
"test_8192": 150,
"test_16384": 150,
"test_32768": 150,
},
avg_character_length={
"test_256": 914.9,
"test_512": 1823.0,
"test_1024": 3644.7,
"test_2048": 7280.0,
"test_4096": 14555.5,
"test_8192": 29108.1,
"test_16384": 58213.9,
"test_32768": 116417.9,
},
)
def load_data(self, **kwargs):
if self.data_loaded:
return
self.corpus = {}
self.queries = {}
self.relevant_docs = {}
for split in self._EVAL_SPLIT:
context_length = int(split.split("_")[1])
query_list = datasets.load_dataset(**self.metadata_dict["dataset"])[
"queries"
] # dict_keys(['qid', 'text'])
query_list = query_list.filter(
lambda x: x["context_length"] == context_length
)
queries = {row["qid"]: row["text"] for row in query_list}
corpus_list = datasets.load_dataset(**self.metadata_dict["dataset"])[
"corpus"
] # dict_keys(['doc_id', 'text'])
corpus_list = corpus_list.filter(
lambda x: x["context_length"] == context_length
)
corpus = {row["doc_id"]: {"text": row["text"]} for row in corpus_list}
qrels_list = datasets.load_dataset(**self.metadata_dict["dataset"])[
"qrels"
] # dict_keys(['qid', 'doc_id'])
qrels_list = qrels_list.filter(
lambda x: x["context_length"] == context_length
)
qrels = {row["qid"]: {row["doc_id"]: 1} for row in qrels_list}
self.corpus[split] = corpus
self.queries[split] = queries
self.relevant_docs[split] = qrels
self.data_loaded = True