FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /tasks /Classification /eng /TweetSentimentExtractionClassification.py
hc99's picture
Add files using upload-large-folder tool
73cc8d2 verified
raw
history blame
1.69 kB
from __future__ import annotations
from mteb.abstasks.TaskMetadata import TaskMetadata
from ....abstasks import AbsTaskClassification
class TweetSentimentExtractionClassification(AbsTaskClassification):
metadata = TaskMetadata(
name="TweetSentimentExtractionClassification",
description="",
reference="https://www.kaggle.com/competitions/tweet-sentiment-extraction/overview",
dataset={
"path": "mteb/tweet_sentiment_extraction",
"revision": "d604517c81ca91fe16a244d1248fc021f9ecee7a",
},
type="Classification",
category="s2s",
eval_splits=["test"],
eval_langs=["eng-Latn"],
main_score="accuracy",
date=(
"2020-01-01",
"2020-12-31",
), # Estimated range for the collection of tweets
form=["written"],
domains=["Social"],
task_subtypes=["Sentiment/Hate speech"],
license="Not specified",
socioeconomic_status="mixed",
annotations_creators="human-annotated",
dialect=[],
text_creation="found",
bibtex_citation="""@misc{tweet-sentiment-extraction,
author = {Maggie, Phil Culliton, Wei Chen},
title = {Tweet Sentiment Extraction},
publisher = {Kaggle},
year = {2020},
url = {https://kaggle.com/competitions/tweet-sentiment-extraction}
}""",
n_samples={"test": 3534},
avg_character_length={"test": 67.8},
)
@property
def metadata_dict(self) -> dict[str, str]:
metadata_dict = dict(self.metadata)
metadata_dict["n_experiments"] = 10
metadata_dict["samples_per_label"] = 32
return metadata_dict