hc99's picture
Add files using upload-large-folder tool
73cc8d2 verified
raw
history blame
2.07 kB
from __future__ import annotations
from mteb.abstasks.TaskMetadata import TaskMetadata
from ....abstasks import AbsTaskClassification
class DBpediaClassification(AbsTaskClassification):
metadata = TaskMetadata(
name="DBpediaClassification",
description="DBpedia14 is a dataset of English texts from Wikipedia articles, categorized into 14 non-overlapping classes based on their DBpedia ontology.",
reference="https://arxiv.org/abs/1509.01626",
dataset={
"path": "fancyzhx/dbpedia_14",
"revision": "9abd46cf7fc8b4c64290f26993c540b92aa145ac",
},
type="Classification",
category="s2s",
eval_splits=["test"],
eval_langs=["eng-Latn"],
main_score="accuracy",
date=("2022-01-25", "2022-01-25"),
form=["written"],
domains=["Encyclopaedic"],
task_subtypes=["Topic classification"],
license="cc-by-sa-3.0",
socioeconomic_status="low",
annotations_creators="derived",
dialect=[],
text_creation="found",
bibtex_citation="""
@inproceedings{NIPS2015_250cf8b5,
author = {Zhang, Xiang and Zhao, Junbo and LeCun, Yann},
booktitle = {Advances in Neural Information Processing Systems},
editor = {C. Cortes and N. Lawrence and D. Lee and M. Sugiyama and R. Garnett},
pages = {},
publisher = {Curran Associates, Inc.},
title = {Character-level Convolutional Networks for Text Classification},
url = {https://proceedings.neurips.cc/paper_files/paper/2015/file/250cf8b51c773f3f8dc8b4be867a9a02-Paper.pdf},
volume = {28},
year = {2015}
}
""",
n_samples={"test": 70000},
avg_character_length={"test": 281.40},
)
def dataset_transform(self):
self.dataset = self.dataset.rename_column("content", "text")
self.dataset = self.stratified_subsampling(
self.dataset, seed=self.seed, splits=["train", "test"]
)