from __future__ import annotations from mteb.abstasks.AbsTaskPairClassification import AbsTaskPairClassification from mteb.abstasks.TaskMetadata import TaskMetadata class CTKFactsNLI(AbsTaskPairClassification): metadata = TaskMetadata( name="CTKFactsNLI", dataset={ "path": "ctu-aic/ctkfacts_nli", "revision": "387ae4582c8054cb52ef57ef0941f19bd8012abf", }, description="Czech Natural Language Inference dataset of around 3K evidence-claim pairs labelled with SUPPORTS, REFUTES or NOT ENOUGH INFO veracity labels. Extracted from a round of fact-checking experiments.", reference="https://arxiv.org/abs/2201.11115", type="PairClassification", category="s2s", eval_splits=["validation", "test"], eval_langs=["ces-Latn"], main_score="ap", date=("2020-09-01", "2021-08-31"), # academic year 2020/2021 form=["written"], domains=["News"], task_subtypes=["Claim verification"], license="CC-BY-SA-3.0", socioeconomic_status="mixed", annotations_creators="human-annotated", dialect=[], text_creation="found", bibtex_citation="""@article{ullrich2023csfever, title={CsFEVER and CTKFacts: acquiring Czech data for fact verification}, author={Ullrich, Herbert and Drchal, Jan and R{\`y}par, Martin and Vincourov{\'a}, Hana and Moravec, V{\'a}clav}, journal={Language Resources and Evaluation}, volume={57}, number={4}, pages={1571--1605}, year={2023}, publisher={Springer} }""", n_samples={ "test": 375, "validation": 305, }, # after removing label 1=NOT ENOUGH INFO avg_character_length={"test": 225.62, "validation": 219.32}, ) def dataset_transform(self): _dataset = {} self.dataset.pop("train") # keep labels 0=REFUTES and 2=SUPPORTS, and map them as 0 and 1 for binary classification hf_dataset = self.dataset.filter(lambda x: x["label"] in [0, 2]) hf_dataset = hf_dataset.map( lambda example: {"label": 1 if example["label"] == 2 else 0} ) for split in self.metadata.eval_splits: _dataset[split] = [ { "sent1": hf_dataset[split]["evidence"], "sent2": hf_dataset[split]["claim"], "labels": hf_dataset[split]["label"], } ] self.dataset = _dataset