hc99's picture
Add files using upload-large-folder tool
83d24b2 verified
raw
history blame
2.56 kB
from __future__ import annotations
from collections import Counter
import datasets
from datasets import DatasetDict
from mteb.abstasks import AbsTaskClassification, MultilingualTask
from mteb.abstasks.TaskMetadata import TaskMetadata
class TurkicClassification(MultilingualTask, AbsTaskClassification):
metadata = TaskMetadata(
name="TurkicClassification",
description="A dataset of news classification in three Turkic languages.",
dataset={
"path": "Electrotubbie/classification_Turkic_languages",
"revision": "db1a67c1bdd54fbb8536af026dc8596f00f9c41d",
},
reference="https://huggingface.co/datasets/Electrotubbie/classification_Turkic_languages/",
type="Classification",
category="s2s",
eval_splits=["train"],
eval_langs={
"ky": ["kir-Cyrl"],
"kk": ["kaz-Cyrl"],
"ba": ["bak-Cyrl"],
},
main_score="accuracy",
date=("2023-02-16", "2023-09-03"),
form=["written"],
domains=["News"],
task_subtypes=["Topic classification"],
license="Not specified",
socioeconomic_status="low",
annotations_creators="derived",
dialect=[],
text_creation="found",
bibtex_citation="""
""",
n_samples={"train": 193056},
avg_character_length={"train": 1103.13},
)
def transform_data(self, dataset, lang):
dataset_lang = DatasetDict()
label_count = Counter(dataset["train"]["label"])
dataset_lang["train"] = dataset["train"].filter(
lambda example: example["lang"] == lang
and label_count[example["label"]] >= 20
)
dataset_lang = self.stratified_subsampling(
dataset_lang, seed=self.seed, splits=["train"]
)
return dataset_lang["train"]
def load_data(self, **kwargs):
"""Load dataset from HuggingFace hub"""
if self.data_loaded:
return
dataset = {}
metadata = self.metadata_dict.get("dataset", None)
full_dataset = datasets.load_dataset(**metadata)
full_dataset = full_dataset.rename_columns(
{"processed_text": "text", "category": "label"}
)
for lang in self.langs:
dataset[lang] = DatasetDict()
filtered_dataset = self.transform_data(full_dataset, lang)
dataset[lang]["train"] = filtered_dataset
self.dataset = dataset
self.dataset_transform()
self.data_loaded = True