FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /tasks /Classification /multilingual /AmazonCounterfactualClassification.py
hc99's picture
Add files using upload-large-folder tool
83d24b2 verified
raw
history blame
3.63 kB
from __future__ import annotations
from mteb.abstasks.TaskMetadata import TaskMetadata
from ....abstasks import AbsTaskClassification, MultilingualTask
class AmazonCounterfactualClassification(MultilingualTask, AbsTaskClassification):
metadata = TaskMetadata(
name="AmazonCounterfactualClassification",
dataset={
"path": "mteb/amazon_counterfactual",
"revision": "e8379541af4e31359cca9fbcf4b00f2671dba205",
},
description=(
"A collection of Amazon customer reviews annotated for counterfactual detection pair classification."
),
reference="https://arxiv.org/abs/2104.06893",
category="s2s",
type="Classification",
eval_splits=["validation", "test"],
eval_langs={
"en-ext": ["eng-Latn"],
"en": ["eng-Latn"],
"de": ["deu-Latn"],
"ja": ["jpn-Jpan"],
},
main_score="accuracy",
date=(
"2018-01-01",
"2021-12-31",
), # Estimated range for the collection of Amazon reviews
form=["written"],
domains=["Reviews"],
task_subtypes=["Counterfactual Detection"],
license="CC BY 4.0",
socioeconomic_status="mixed",
annotations_creators="human-annotated",
dialect=[],
text_creation="found",
bibtex_citation="""@inproceedings{oneill-etal-2021-wish,
title = "{I} Wish {I} Would Have Loved This One, But {I} Didn{'}t {--} A Multilingual Dataset for Counterfactual Detection in Product Review",
author = "O{'}Neill, James and
Rozenshtein, Polina and
Kiryo, Ryuichi and
Kubota, Motoko and
Bollegala, Danushka",
editor = "Moens, Marie-Francine and
Huang, Xuanjing and
Specia, Lucia and
Yih, Scott Wen-tau",
booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2021",
address = "Online and Punta Cana, Dominican Republic",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.emnlp-main.568",
doi = "10.18653/v1/2021.emnlp-main.568",
pages = "7092--7108",
abstract = "Counterfactual statements describe events that did not or cannot take place. We consider the problem of counterfactual detection (CFD) in product reviews. For this purpose, we annotate a multilingual CFD dataset from Amazon product reviews covering counterfactual statements written in English, German, and Japanese languages. The dataset is unique as it contains counterfactuals in multiple languages, covers a new application area of e-commerce reviews, and provides high quality professional annotations. We train CFD models using different text representation methods and classifiers. We find that these models are robust against the selectional biases introduced due to cue phrase-based sentence selection. Moreover, our CFD dataset is compatible with prior datasets and can be merged to learn accurate CFD models. Applying machine translation on English counterfactual examples to create multilingual data performs poorly, demonstrating the language-specificity of this problem, which has been ignored so far.",
}""",
n_samples={"validation": 335, "test": 670},
avg_character_length={"validation": 109.2, "test": 106.1},
)
@property
def metadata_dict(self) -> dict[str, str]:
metadata_dict = super().metadata_dict
metadata_dict["n_experiments"] = 10
metadata_dict["samples_per_label"] = 32
return metadata_dict