from __future__ import annotations import json import datasets from mteb.abstasks.TaskMetadata import TaskMetadata from ....abstasks import MultilingualTask from ....abstasks.AbsTaskRetrieval import AbsTaskRetrieval _EVAL_SPLITS = ["dev", "test"] _LANGS = { # - "english": ["eng-Latn"], "french": ["fra-Latn"], } def _load_statcan_data( path: str, langs: list, splits: str, cache_dir: str = None, revision: str = None ): queries = {lang: {split: {} for split in splits} for lang in langs} corpus = {lang: {split: {} for split in splits} for lang in langs} relevant_docs = {lang: {split: {} for split in splits} for lang in langs} for split in splits: for lang in langs: query_table = datasets.load_dataset( path, f"queries_{lang}", split=split, cache_dir=cache_dir, revision=revision, ) corpus_table = datasets.load_dataset( path, "corpus", split=lang, cache_dir=cache_dir, revision=revision, ) for row in query_table: query = json.loads(row["query"]) query_id = row["query_id"] doc_id = row["doc_id"] queries[lang][split][query_id] = query if query_id not in relevant_docs[lang][split]: relevant_docs[lang][split][query_id] = {} relevant_docs[lang][split][query_id][doc_id] = 1 for row in corpus_table: doc_id = row["doc_id"] doc_content = row["doc"] corpus[lang][split][doc_id] = {"text": doc_content} corpus = datasets.DatasetDict(corpus) queries = datasets.DatasetDict(queries) relevant_docs = datasets.DatasetDict(relevant_docs) return corpus, queries, relevant_docs class StatcanDialogueDatasetRetrieval(MultilingualTask, AbsTaskRetrieval): metadata = TaskMetadata( name="StatcanDialogueDatasetRetrieval", description="A Dataset for Retrieving Data Tables through Conversations with Genuine Intents, available in English and French.", dataset={ "path": "McGill-NLP/statcan-dialogue-dataset-retrieval", "revision": "7a26938c93e99e0759a1df416896bb72527e2f33", }, type="Retrieval", category="s2p", eval_splits=_EVAL_SPLITS, eval_langs=_LANGS, main_score="recall_at_10", reference="https://mcgill-nlp.github.io/statcan-dialogue-dataset/", date=("2020-01-01", "2020-04-15"), form=["written"], domains=["Government", "Web"], task_subtypes=["Conversational retrieval"], license="https://huggingface.co/datasets/McGill-NLP/statcan-dialogue-dataset-retrieval/blob/main/LICENSE.md", socioeconomic_status="high", annotations_creators="derived", dialect=[], text_creation="found", bibtex_citation=""" @inproceedings{lu-etal-2023-statcan, title = "The {S}tat{C}an Dialogue Dataset: Retrieving Data Tables through Conversations with Genuine Intents", author = "Lu, Xing Han and Reddy, Siva and de Vries, Harm", booktitle = "Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics", month = may, year = "2023", address = "Dubrovnik, Croatia", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/2304.01412", pages = "2799--2829", } """, n_samples={"dev": 1000, "test": 1011, "corpus": 5907}, avg_character_length={"dev": 776.58, "test": 857.13, "corpus": 6806.97}, ) def load_data(self, **kwargs): if self.data_loaded: return self.corpus, self.queries, self.relevant_docs = _load_statcan_data( path=self.metadata_dict["dataset"]["path"], langs=list(_LANGS.keys()), splits=self.metadata_dict["eval_splits"], cache_dir=kwargs.get("cache_dir", None), revision=self.metadata_dict["dataset"]["revision"], ) self.data_loaded = True