| from __future__ import annotations |
|
|
| import datasets |
|
|
| from mteb.abstasks.TaskMetadata import TaskMetadata |
|
|
| from ....abstasks.AbsTaskRetrieval import AbsTaskRetrieval |
|
|
|
|
| class FQuADRetrieval(AbsTaskRetrieval): |
| _EVAL_SPLITS = ["test", "validation"] |
|
|
| metadata = TaskMetadata( |
| name="FQuADRetrieval", |
| description="This dataset has been built from the French SQuad dataset.", |
| reference="https://huggingface.co/datasets/manu/fquad2_test", |
| dataset={ |
| "path": "manu/fquad2_test", |
| "revision": "5384ce827bbc2156d46e6fcba83d75f8e6e1b4a6", |
| }, |
| type="Retrieval", |
| category="s2p", |
| eval_splits=_EVAL_SPLITS, |
| eval_langs=["fra-Latn"], |
| main_score="ndcg_at_10", |
| date=("2019-11-01", "2020-05-01"), |
| form=["written"], |
| domains=["Encyclopaedic"], |
| task_subtypes=["Article retrieval"], |
| license="apache-2.0", |
| socioeconomic_status="mixed", |
| annotations_creators="human-annotated", |
| dialect=[], |
| text_creation="created", |
| bibtex_citation="""@inproceedings{dhoffschmidt-etal-2020-fquad, |
| title = "{FQ}u{AD}: {F}rench Question Answering Dataset", |
| author = "d{'}Hoffschmidt, Martin and |
| Belblidia, Wacim and |
| Heinrich, Quentin and |
| Brendl{\'e}, Tom and |
| Vidal, Maxime", |
| editor = "Cohn, Trevor and |
| He, Yulan and |
| Liu, Yang", |
| booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2020", |
| month = nov, |
| year = "2020", |
| address = "Online", |
| publisher = "Association for Computational Linguistics", |
| url = "https://aclanthology.org/2020.findings-emnlp.107", |
| doi = "10.18653/v1/2020.findings-emnlp.107", |
| pages = "1193--1208", |
| }""", |
| n_samples={"test": 400, "validation": 100}, |
| avg_character_length={"test": 937, "validation": 930}, |
| ) |
|
|
| def load_data(self, **kwargs): |
| if self.data_loaded: |
| return |
| dataset_raw = datasets.load_dataset( |
| **self.metadata_dict["dataset"], |
| ) |
|
|
| |
| dataset_raw["validation"] = dataset_raw["valid_hasAns"] |
| del dataset_raw["valid_hasAns"] |
|
|
| dataset_raw["test"] = dataset_raw["test_hasAns"] |
| del dataset_raw["test_hasAns"] |
|
|
| |
| dataset_raw = dataset_raw.rename_column("context", "text") |
|
|
| self.queries = { |
| eval_split: { |
| str(i): q["question"] for i, q in enumerate(dataset_raw[eval_split]) |
| } |
| for eval_split in self.metadata_dict["eval_splits"] |
| } |
|
|
| self.corpus = { |
| eval_split: {str(row["title"]): row for row in dataset_raw[eval_split]} |
| for eval_split in self.metadata_dict["eval_splits"] |
| } |
|
|
| self.relevant_docs = { |
| eval_split: { |
| str(i): {str(q["title"]): 1} |
| for i, q in enumerate(dataset_raw[eval_split]) |
| } |
| for eval_split in self.metadata_dict["eval_splits"] |
| } |
|
|
| self.data_loaded = True |
|
|