hc99's picture
Add files using upload-large-folder tool
83d24b2 verified
raw
history blame
1.92 kB
from __future__ import annotations
from mteb.abstasks.TaskMetadata import TaskMetadata
from ....abstasks import AbsTaskClassification
_LANGUAGES = {
"ar": ["ara-Arab"],
"bg": ["bul-Cyrl"],
"de": ["deu-Latn"],
"el": ["ell-Grek"],
"en": ["eng-Latn"],
"es": ["spa-Latn"],
"fr": ["fra-Latn"],
"hi": ["hin-Deva"],
"it": ["ita-Latn"],
"ja": ["jpn-Jpan"],
"nl": ["nld-Latn"],
"pl": ["pol-Latn"],
"pt": ["por-Latn"],
"ru": ["rus-Cyrl"],
"sw": ["swa-Latn"],
"th": ["tha-Thai"],
"tr": ["tur-Latn"],
"ur": ["urd-Arab"],
"vi": ["vie-Latn"],
"zh": ["cmn-Hans"],
}
class LanguageClassification(AbsTaskClassification):
metadata = TaskMetadata(
name="LanguageClassification",
dataset={
"path": "papluca/language-identification",
"revision": "aa56583bf2bc52b0565770607d6fc3faebecf9e2",
},
description="A language identification dataset for 20 languages.",
reference="https://huggingface.co/datasets/papluca/language-identification",
category="s2s",
type="Classification",
eval_splits=["test"],
eval_langs=_LANGUAGES,
main_score="accuracy",
date=("2021-11-01", "2021-11-30"),
form=["written"],
domains=["Reviews", "Web", "Non-fiction", "Fiction", "Government"],
task_subtypes=["Language identification"],
license="Not specified",
socioeconomic_status="mixed",
annotations_creators="derived",
dialect=[],
text_creation="found",
bibtex_citation="",
n_samples={"test": 2048},
avg_character_length={"test": 107.8},
)
def dataset_transform(self) -> None:
self.dataset = self.dataset.rename_columns({"labels": "label"})
self.dataset = self.stratified_subsampling(
self.dataset, seed=self.seed, splits=["test"]
)