hc99's picture
Add files using upload-large-folder tool
83d24b2 verified
raw
history blame
3.26 kB
from typing import Any
import datasets
from mteb.abstasks import AbsTaskClassification, MultilingualTask
from mteb.abstasks.TaskMetadata import TaskMetadata
class NaijaSenti(AbsTaskClassification, MultilingualTask):
metadata = TaskMetadata(
name="NaijaSenti",
description="NaijaSenti is the first large-scale human-annotated Twitter sentiment dataset for the four most widely spoken languages in Nigeria — Hausa, Igbo, Nigerian-Pidgin, and Yorùbá — consisting of around 30,000 annotated tweets per language, including a significant fraction of code-mixed tweets.",
reference="https://github.com/hausanlp/NaijaSenti",
dataset={
"path": "HausaNLP/NaijaSenti-Twitter",
"revision": "a3d0415a828178edf3466246f49cfcd83b946ab3",
},
type="Classification",
category="s2s",
eval_splits=["test"],
eval_langs={
"hau": ["hau-Latn"],
"ibo": ["ibo-Latn"],
"pcm": ["pcm-Latn"],
"yor": ["yor-Latn"],
},
main_score="accuracy",
date=("2022-05-01", "2023-05-08"),
form=["written"],
domains=["Social"],
task_subtypes=["Sentiment/Hate speech"],
license="CC-BY-4.0 license",
socioeconomic_status="low",
annotations_creators="expert-annotated",
dialect=[],
text_creation="found",
bibtex_citation="""
@inproceedings{muhammad-etal-2022-naijasenti,
title = "{N}aija{S}enti: A {N}igerian {T}witter Sentiment Corpus for Multilingual Sentiment Analysis",
author = "Muhammad, Shamsuddeen Hassan and
Adelani, David Ifeoluwa and
Ruder, Sebastian and
Ahmad, Ibrahim Sa{'}id and
Abdulmumin, Idris and
Bello, Bello Shehu and
Choudhury, Monojit and
Emezue, Chris Chinenye and
Abdullahi, Saheed Salahudeen and
Aremu, Anuoluwapo and
Jorge, Al{\'\i}pio and
Brazdil, Pavel",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://aclanthology.org/2022.lrec-1.63",
pages = "590--602",
}""",
n_samples={"test": 4800},
avg_character_length={"test": 72.81},
)
def load_data(self, **kwargs: Any) -> None:
"""Load dataset from HuggingFace hub"""
if self.data_loaded:
return
self.dataset = {}
for lang in self.hf_subsets:
self.dataset[lang] = datasets.load_dataset(
name=f"{lang}",
**self.metadata_dict["dataset"],
)
self.dataset[lang] = datasets.DatasetDict(
{
"train": self.dataset[lang]["train"],
"test": self.dataset[lang]["test"],
}
)
self.dataset_transform()
self.data_loaded = True
def dataset_transform(self) -> None:
for lang in self.hf_subsets:
self.dataset[lang] = self.dataset[lang].map(
lambda example: {"text": example["tweet"]}
)