from __future__ import annotations from mteb.abstasks import AbsTaskClassification from mteb.abstasks.TaskMetadata import TaskMetadata _EVAL_SPLITS = ["test"] class ToxicChatClassification(AbsTaskClassification): metadata = TaskMetadata( name="ToxicChatClassification", description="""This dataset contains toxicity annotations on 10K user prompts collected from the Vicuna online demo. We utilize a human-AI collaborative annotation framework to guarantee the quality of annotation while maintaining a feasible annotation workload. The details of data collection, pre-processing, and annotation can be found in our paper. We believe that ToxicChat can be a valuable resource to drive further advancements toward building a safe and healthy environment for user-AI interactions. Only human annotated samples are selected here.""", reference="https://aclanthology.org/2023.findings-emnlp.311/", dataset={ "path": "lmsys/toxic-chat", "name": "toxicchat0124", "revision": "3e0319203c7162b9c9f8015b594441f979c199bc", }, type="Classification", category="s2s", eval_splits=_EVAL_SPLITS, eval_langs=["eng-Latn"], main_score="accuracy", date=("2023-10-26", "2024-01-31"), form=["written"], domains=["Constructed"], task_subtypes=["Sentiment/Hate speech"], license="cc-by-4.0", socioeconomic_status="high", annotations_creators="expert-annotated", dialect=[], text_creation="found", bibtex_citation="""@misc{lin2023toxicchat, title={ToxicChat: Unveiling Hidden Challenges of Toxicity Detection in Real-World User-AI Conversation}, author={Zi Lin and Zihan Wang and Yongqi Tong and Yangkun Wang and Yuxin Guo and Yujia Wang and Jingbo Shang}, year={2023}, eprint={2310.17389}, archivePrefix={arXiv}, primaryClass={cs.CL} }""", n_samples={"test": 1427}, avg_character_length={"test": 189.4}, ) def dataset_transform(self): keep_cols = ["user_input", "toxicity"] rename_dict = dict(zip(keep_cols, ["text", "label"])) remove_cols = [ col for col in self.dataset[_EVAL_SPLITS[0]].column_names if col not in keep_cols ] self.dataset = self.dataset.rename_columns(rename_dict) self.dataset = self.stratified_subsampling( self.dataset, seed=self.seed, splits=["test"] ) # only use human-annotated data self.dataset = self.dataset.filter(lambda x: x["human_annotation"]) self.dataset = self.dataset.remove_columns(remove_cols)