FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /tasks /Classification /dan /DdiscoCohesionClassification.py
hc99's picture
Add files using upload-large-folder tool
73cc8d2 verified
raw
history blame
2.94 kB
from __future__ import annotations
from mteb.abstasks.AbsTaskClassification import AbsTaskClassification
from mteb.abstasks.TaskMetadata import TaskMetadata
class DdiscoCohesionClassification(AbsTaskClassification):
metadata = TaskMetadata(
name="Ddisco",
dataset={
"path": "DDSC/ddisco",
"revision": "514ab557579fcfba538a4078d6d647248a0e6eb7",
},
description="A Danish Discourse dataset with values for coherence and source (Wikipedia or Reddit)",
reference="https://aclanthology.org/2022.lrec-1.260/",
type="Classification",
category="s2s",
eval_splits=["test"],
eval_langs=["dan-Latn"],
main_score="accuracy",
date=("2021-01-01", "2022-06-25"),
form=["written"],
domains=["Non-fiction", "Social"],
dialect=[],
task_subtypes=["Discourse coherence"],
license="cc-by-sa-3.0",
socioeconomic_status="high",
annotations_creators="expert-annotated",
text_creation="found",
bibtex_citation="""
@inproceedings{flansmose-mikkelsen-etal-2022-ddisco,
title = "{DD}is{C}o: A Discourse Coherence Dataset for {D}anish",
author = "Flansmose Mikkelsen, Linea and
Kinch, Oliver and
Jess Pedersen, Anders and
Lacroix, Oph{\'e}lie",
editor = "Calzolari, Nicoletta and
B{\'e}chet, Fr{\'e}d{\'e}ric and
Blache, Philippe and
Choukri, Khalid and
Cieri, Christopher and
Declerck, Thierry and
Goggi, Sara and
Isahara, Hitoshi and
Maegaard, Bente and
Mariani, Joseph and
Mazo, H{\'e}l{\`e}ne and
Odijk, Jan and
Piperidis, Stelios",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://aclanthology.org/2022.lrec-1.260",
pages = "2440--2445",
abstract = "To date, there has been no resource for studying discourse coherence on real-world Danish texts. Discourse coherence has mostly been approached with the assumption that incoherent texts can be represented by coherent texts in which sentences have been shuffled. However, incoherent real-world texts rarely resemble that. We thus present DDisCo, a dataset including text from the Danish Wikipedia and Reddit annotated for discourse coherence. We choose to annotate real-world texts instead of relying on artificially incoherent text for training and testing models. Then, we evaluate the performance of several methods, including neural networks, on the dataset.",
}
""",
n_samples=None,
avg_character_length=None,
)
def dataset_transform(self):
self.dataset = self.dataset.rename_columns({"rating": "label"}).remove_columns(
["domain"]
)