hc99's picture
Add files using upload-large-folder tool
73cc8d2 verified
raw
history blame
6.44 kB
from __future__ import annotations
from mteb.abstasks import MultilingualTask
from mteb.abstasks.AbsTaskPairClassification import AbsTaskPairClassification
from mteb.abstasks.TaskMetadata import TaskMetadata
_LANGS = {
"ar": ["ara-Arab"],
"bg": ["bul-Cyrl"],
"de": ["deu-Latn"],
"el": ["ell-Grek"],
"en": ["eng-Latn"],
"es": ["spa-Latn"],
"fr": ["fra-Latn"],
"hi": ["hin-Deva"],
"ru": ["rus-Cyrl"],
"sw": ["swa-Latn"],
"th": ["tha-Thai"],
"tr": ["tur-Latn"],
"vi": ["vie-Latn"],
"zh": ["zho-Hans"],
}
class XNLI(MultilingualTask, AbsTaskPairClassification):
metadata = TaskMetadata(
name="XNLI",
dataset={
"path": "mteb/xnli",
"revision": "09698e0180d87dc247ca447d3a1248b931ac0cdb",
},
description="",
reference="https://aclanthology.org/D18-1269/",
category="s2s",
type="PairClassification",
eval_splits=["test", "validation"],
eval_langs=_LANGS,
main_score="ap",
date=("2018-01-01", "2018-11-04"),
form=["written"],
domains=["Non-fiction", "Fiction", "Government"],
task_subtypes=[],
license="Not specified",
socioeconomic_status="mixed",
annotations_creators="expert-annotated",
dialect=[],
text_creation="created",
bibtex_citation="""@InProceedings{conneau2018xnli,
author = {Conneau, Alexis
and Rinott, Ruty
and Lample, Guillaume
and Williams, Adina
and Bowman, Samuel R.
and Schwenk, Holger
and Stoyanov, Veselin},
title = {XNLI: Evaluating Cross-lingual Sentence Representations},
booktitle = {Proceedings of the 2018 Conference on Empirical Methods
in Natural Language Processing},
year = {2018},
publisher = {Association for Computational Linguistics},
location = {Brussels, Belgium},
}
""",
n_samples={"validation": 2163, "test": 2460},
avg_character_length={"validation": 106.5, "test": 106.5},
)
def dataset_transform(self):
_dataset = {}
for lang in self.hf_subsets:
_dataset[lang] = {}
self.dataset[lang] = self.stratified_subsampling(
self.dataset[lang], seed=self.seed, splits=self.metadata.eval_splits
)
for split in self.metadata.eval_splits:
# 0=entailment, 2=contradiction. Filter out neutral to match the task.
# Then map entailment as positive (1) and contradiction as negative (0).
hf_dataset = self.dataset[lang][split].filter(
lambda x: x["label"] in [0, 2]
)
hf_dataset = hf_dataset.map(
lambda example: {"label": 0 if example["label"] == 2 else 1}
)
_dataset[lang][split] = [
{
"sent1": hf_dataset["premise"],
"sent2": hf_dataset["hypothesis"],
"labels": hf_dataset["label"],
}
]
self.dataset = _dataset
_LANGS_2 = {
"punjabi": ["pan-Guru"],
"gujrati": ["guj-Gujr"],
"kannada": ["kan-Knda"],
"assamese": ["asm-Beng"],
"bengali": ["ben-Beng"],
"marathi": ["mar-Deva"],
"bhojpuri": ["bho-Deva"],
"odiya": ["ory-Orya"],
"sanskrit": ["san-Deva"],
"tamil": ["tam-Taml"],
"turkish": ["tur-Latn"],
"greek": ["ell-Grek"],
"russian": ["rus-Cyrl"],
}
class XNLIV2(MultilingualTask, AbsTaskPairClassification):
metadata = TaskMetadata(
name="XNLIV2",
dataset={
"path": "mteb/xnli2.0-multi-pair",
"revision": "5b7d477a8c62cdd18e2fed7e015497c20b4371ad",
},
description="""
This is subset of 'XNLI 2.0: Improving XNLI dataset and performance on Cross Lingual Understanding'
with languages that were not part of the original XNLI plus three (verified) languages that are not strongly covered in MTEB
""",
reference="https://arxiv.org/pdf/2301.06527",
category="s2s",
type="PairClassification",
eval_splits=["test"],
eval_langs=_LANGS_2,
main_score="ap",
date=("2018-01-01", "2018-11-04"),
form=["written"],
domains=["Non-fiction", "Fiction", "Government"],
task_subtypes=[],
license="Not specified",
socioeconomic_status="mixed",
annotations_creators="expert-annotated",
dialect=[],
text_creation="machine-translated and verified",
bibtex_citation="""@inproceedings{upadhyay2023xnli,
title={XNLI 2.0: Improving XNLI dataset and performance on Cross Lingual Understanding (XLU)},
author={Upadhyay, Ankit Kumar and Upadhya, Harsit Kumar},
booktitle={2023 IEEE 8th International Conference for Convergence in Technology (I2CT)},
pages={1--6},
year={2023},
organization={IEEE}
}
""",
n_samples={"test": 5010},
avg_character_length={"test": 80.06}, # average of premise and hypothesis
)
def dataset_transform(self):
_dataset = {}
for lang in self.langs:
_dataset[lang] = {}
self.dataset[lang] = self.stratified_subsampling(
self.dataset[lang], seed=self.seed, splits=self.metadata.eval_splits
)
for split in self.metadata.eval_splits:
# 0=entailment, 2=contradiction. Filter out neutral to match the task.
# Then map entailment as positive (1) and contradiction as negative (0).
hf_dataset = self.dataset[lang][split].filter(
lambda x: x["label"] in [0, 2]
)
hf_dataset = hf_dataset.map(
lambda example: {"label": 0 if example["label"] == 2 else 1}
)
_dataset[lang][split] = [
{
"sent1": hf_dataset["premise"],
"sent2": hf_dataset["hypothesis"],
"labels": hf_dataset["label"],
}
]
self.dataset = _dataset