hc99's picture
Add files using upload-large-folder tool
73cc8d2 verified
raw
history blame
3.07 kB
from __future__ import annotations
from typing import Any
import datasets
import numpy as np
from mteb.abstasks import AbsTaskClustering, MultilingualTask
from mteb.abstasks.TaskMetadata import TaskMetadata
_LANGUAGES = {
"as": ["asm-Beng"],
"bd": ["brx-Deva"],
"bn": ["ben-Beng"],
"gu": ["guj-Gujr"],
"hi": ["hin-Deva"],
"kn": ["kan-Knda"],
"ml": ["mal-Mlym"],
"mr": ["mar-Deva"],
"or": ["ory-Orya"],
"pa": ["pan-Guru"],
"ta": ["tam-Taml"],
"te": ["tel-Telu"],
"ur": ["urd-Arab"],
}
class IndicReviewsClusteringP2P(AbsTaskClustering, MultilingualTask):
metadata = TaskMetadata(
name="IndicReviewsClusteringP2P",
dataset={
"path": "ai4bharat/IndicSentiment",
"revision": "ccb472517ce32d103bba9d4f5df121ed5a6592a4",
},
description="Clustering of reviews from IndicSentiment dataset. Clustering of 14 sets on the generic categories label.",
reference="https://arxiv.org/abs/2212.05409",
type="Clustering",
category="p2p",
eval_splits=["test"],
eval_langs=_LANGUAGES,
main_score="v_measure",
date=("2022-08-01", "2022-12-20"),
form=["written"],
domains=["Reviews"],
task_subtypes=["Thematic clustering"],
license="CC0",
socioeconomic_status="mixed",
annotations_creators="human-annotated",
dialect=[],
text_creation="machine-translated and verified",
bibtex_citation="""@article{doddapaneni2022towards,
title = {Towards Leaving No Indic Language Behind: Building Monolingual Corpora, Benchmark and Models for Indic Languages},
author = {Sumanth Doddapaneni and Rahul Aralikatte and Gowtham Ramesh and Shreyansh Goyal and Mitesh M. Khapra and Anoop Kunchukuttan and Pratyush Kumar},
journal = {Annual Meeting of the Association for Computational Linguistics},
year = {2022},
doi = {10.18653/v1/2023.acl-long.693}
}""",
n_samples={"test": 1000},
avg_character_length={"test": 137.6},
)
def load_data(self, **kwargs: Any) -> None:
"""Load dataset from HuggingFace hub"""
if self.data_loaded:
return
self.dataset = {}
for lang in self.hf_subsets:
self.dataset[lang] = datasets.load_dataset(
name=f"translation-{lang}",
**self.metadata_dict["dataset"],
)
self.dataset_transform()
self.data_loaded = True
def dataset_transform(self) -> None:
for lang in self.hf_subsets:
self.dataset[lang].pop("validation")
texts = self.dataset[lang]["test"]["INDIC REVIEW"]
labels = self.dataset[lang]["test"]["GENERIC CATEGORIES"]
new_format = {
"sentences": [split.tolist() for split in np.array_split(texts, 5)],
"labels": [split.tolist() for split in np.array_split(labels, 5)],
}
self.dataset[lang]["test"] = datasets.Dataset.from_dict(new_format)