FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /tasks /Clustering /deu /BlurbsClusteringP2P.py
| from __future__ import annotations | |
| import itertools | |
| import numpy as np | |
| from datasets import Dataset, DatasetDict | |
| from mteb.abstasks.AbsTaskClustering import AbsTaskClustering | |
| from mteb.abstasks.AbsTaskClusteringFast import AbsTaskClusteringFast | |
| from mteb.abstasks.TaskMetadata import TaskMetadata | |
| NUM_SAMPLES = 2048 | |
| class BlurbsClusteringP2P(AbsTaskClustering): | |
| superseeded_by = "BlurbsClusteringP2P.v2" | |
| metadata = TaskMetadata( | |
| name="BlurbsClusteringP2P", | |
| description="Clustering of book titles+blurbs. Clustering of 28 sets, either on the main or secondary genre.", | |
| reference="https://www.inf.uni-hamburg.de/en/inst/ab/lt/resources/data/germeval-2019-hmc.html", | |
| dataset={ | |
| "path": "slvnwhrl/blurbs-clustering-p2p", | |
| "revision": "a2dd5b02a77de3466a3eaa98ae586b5610314496", | |
| }, | |
| type="Clustering", | |
| category="p2p", | |
| eval_splits=["test"], | |
| eval_langs=["deu-Latn"], | |
| main_score="v_measure", | |
| date=None, | |
| form=None, | |
| domains=None, | |
| task_subtypes=None, | |
| license=None, | |
| socioeconomic_status=None, | |
| annotations_creators=None, | |
| dialect=None, | |
| text_creation=None, | |
| bibtex_citation=None, | |
| n_samples={"test": 174637}, | |
| avg_character_length={"test": 664.09}, | |
| ) | |
| class BlurbsClusteringP2PFast(AbsTaskClusteringFast): | |
| # a faster version of BlurbsClusteringP2P, since it does not sample from the same distribution we can't use the AbsTaskClusteringFast, instead we | |
| # simply downsample each cluster. | |
| metadata = TaskMetadata( | |
| name="BlurbsClusteringP2P.v2", | |
| description="Clustering of book titles+blurbs. Clustering of 28 sets, either on the main or secondary genre.", | |
| reference="https://www.inf.uni-hamburg.de/en/inst/ab/lt/resources/data/germeval-2019-hmc.html", | |
| dataset={ | |
| "path": "slvnwhrl/blurbs-clustering-p2p", | |
| "revision": "a2dd5b02a77de3466a3eaa98ae586b5610314496", | |
| }, | |
| type="Clustering", | |
| category="p2p", | |
| eval_splits=["test"], | |
| eval_langs=["deu-Latn"], | |
| main_score="v_measure", | |
| date=( | |
| "1900-01-01", | |
| "2019-12-31", | |
| ), # since it is books it is likely to be from the 20th century -> paper from 2019 | |
| form=["written"], | |
| domains=["Fiction"], | |
| task_subtypes=["Thematic clustering"], | |
| license="cc-by-nc-4.0", | |
| socioeconomic_status="mixed", | |
| annotations_creators="derived", | |
| dialect=[], | |
| text_creation="found", | |
| bibtex_citation="""@inproceedings{Remus2019GermEval2T, | |
| title={GermEval 2019 Task 1: Hierarchical Classification of Blurbs}, | |
| author={Steffen Remus and Rami Aly and Chris Biemann}, | |
| booktitle={Conference on Natural Language Processing}, | |
| year={2019}, | |
| url={https://api.semanticscholar.org/CorpusID:208334484} | |
| }""", | |
| n_samples={"test": NUM_SAMPLES}, | |
| avg_character_length={"test": 664.09}, | |
| ) | |
| def dataset_transform(self): | |
| ds = dict() | |
| for split in self.metadata.eval_splits: | |
| labels = list(itertools.chain.from_iterable(self.dataset[split]["labels"])) | |
| sentences = list( | |
| itertools.chain.from_iterable(self.dataset[split]["sentences"]) | |
| ) | |
| # Remove sentences and labels with only 1 label example. | |
| unique_labels, counts = np.unique(labels, return_counts=True) | |
| solo_label_idx = np.where(counts == 1) | |
| solo_labels = unique_labels[solo_label_idx] | |
| for solo_label in solo_labels: | |
| loc = labels.index(solo_label) | |
| labels.pop(loc) | |
| sentences.pop(loc) | |
| ds[split] = Dataset.from_dict({"labels": labels, "sentences": sentences}) | |
| self.dataset = DatasetDict(ds) | |
| self.dataset = self.stratified_subsampling( | |
| self.dataset, | |
| self.seed, | |
| self.metadata.eval_splits, | |
| label="labels", | |
| n_samples=NUM_SAMPLES, | |
| ) | |