FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /tasks /Classification /fin /FinToxicityClassification.py
| from __future__ import annotations | |
| from mteb.abstasks import AbsTaskClassification | |
| from mteb.abstasks.TaskMetadata import TaskMetadata | |
| class FinToxicityClassification(AbsTaskClassification): | |
| metadata = TaskMetadata( | |
| name="FinToxicityClassification", | |
| description=""" | |
| This dataset is a DeepL -based machine translated version of the Jigsaw toxicity dataset for Finnish. The dataset is originally from a Kaggle competition https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge/data. | |
| The original dataset poses a multi-label text classification problem and includes the labels identity_attack, insult, obscene, severe_toxicity, threat and toxicity. | |
| Here adapted for toxicity classification, which is the most represented class. | |
| """, | |
| dataset={ | |
| "path": "TurkuNLP/jigsaw_toxicity_pred_fi", | |
| "revision": "6e7340e6be87124f319e25290778760c14df64d3", | |
| }, | |
| reference="https://aclanthology.org/2023.nodalida-1.68", | |
| type="Classification", | |
| category="s2s", | |
| eval_splits=["test"], | |
| eval_langs=["fin-Latn"], | |
| main_score="f1", | |
| date=("2023-03-13", "2023-09-25"), | |
| form=["written"], | |
| domains=["News"], | |
| task_subtypes=["Sentiment/Hate speech"], | |
| license="ccy-by-sa-4.0", | |
| socioeconomic_status="high", | |
| annotations_creators="derived", | |
| dialect=[], | |
| text_creation="machine-translated", | |
| bibtex_citation=""" | |
| @inproceedings{eskelinen-etal-2023-toxicity, | |
| title = "Toxicity Detection in {F}innish Using Machine Translation", | |
| author = "Eskelinen, Anni and | |
| Silvala, Laura and | |
| Ginter, Filip and | |
| Pyysalo, Sampo and | |
| Laippala, Veronika", | |
| booktitle = "Proceedings of the 24th Nordic Conference on Computational Linguistics (NoDaLiDa)", | |
| month = may, | |
| year = "2023", | |
| }""", | |
| n_samples={"train": 2048, "test": 2048}, | |
| avg_character_length={"train": 432.63, "test": 401.03}, | |
| ) | |
| def dataset_transform(self): | |
| self.dataset = self.dataset.rename_column("label_toxicity", "label") | |
| remove_cols = [ | |
| col | |
| for col in self.dataset["test"].column_names | |
| if col not in ["text", "label"] | |
| ] | |
| self.dataset = self.dataset.remove_columns(remove_cols) | |
| self.dataset = self.stratified_subsampling( | |
| self.dataset, seed=self.seed, splits=["train", "test"] | |
| ) | |