hc99's picture
Add files using upload-large-folder tool
83d24b2 verified
raw
history blame
2.77 kB
from __future__ import annotations
from mteb.abstasks import AbsTaskClassification
from mteb.abstasks.TaskMetadata import TaskMetadata
class NepaliNewsClassification(AbsTaskClassification):
metadata = TaskMetadata(
name="NepaliNewsClassification",
description="A Nepali dataset for 7500 news articles ",
reference="https://github.com/goru001/nlp-for-nepali",
dataset={
"path": "bpHigh/iNLTK_Nepali_News_Dataset",
"revision": "79125f20d858a08f71ec4923169a6545221725c4",
},
type="Classification",
category="s2s",
date=("2019-01-01", "2020-01-01"),
eval_splits=["train"],
eval_langs=["nep-Deva"],
main_score="accuracy",
form=["written"],
domains=["News"],
task_subtypes=["Topic classification"],
license="CC BY-SA 4.0",
socioeconomic_status="mixed",
annotations_creators="derived",
dialect=[],
text_creation="found",
bibtex_citation="""
@inproceedings{arora-2020-inltk,
title = "i{NLTK}: Natural Language Toolkit for Indic Languages",
author = "Arora, Gaurav",
editor = "Park, Eunjeong L. and
Hagiwara, Masato and
Milajevs, Dmitrijs and
Liu, Nelson F. and
Chauhan, Geeticka and
Tan, Liling",
booktitle = "Proceedings of Second Workshop for NLP Open Source Software (NLP-OSS)",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2020.nlposs-1.10",
doi = "10.18653/v1/2020.nlposs-1.10",
pages = "66--71",
abstract = "We present iNLTK, an open-source NLP library consisting of pre-trained language models and out-of-the-box support for Data Augmentation, Textual Similarity, Sentence Embeddings, Word Embeddings, Tokenization and Text Generation in 13 Indic Languages. By using pre-trained models from iNLTK for text classification on publicly available datasets, we significantly outperform previously reported results. On these datasets, we also show that by using pre-trained models and data augmentation from iNLTK, we can achieve more than 95{\%} of the previous best performance by using less than 10{\%} of the training data. iNLTK is already being widely used by the community and has 40,000+ downloads, 600+ stars and 100+ forks on GitHub.",
}
""",
n_samples={"train": 5975, "test": 1495},
avg_character_length={"train": 196.61, "test": 196.017},
)
def dataset_transform(self):
self.dataset = self.dataset.rename_column("paras", "text")
self.dataset = self.stratified_subsampling(
self.dataset, seed=self.seed, splits=["train"]
)