FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /abstasks /AbsTaskMultilabelClassification.py
hc99's picture
Add files using upload-large-folder tool
83d24b2 verified
raw
history blame
6.43 kB
from __future__ import annotations
import itertools
import logging
from collections import defaultdict
from typing import Any
import numpy as np
from sklearn.base import ClassifierMixin, clone
from sklearn.metrics import f1_score, label_ranking_average_precision_score
from sklearn.model_selection import train_test_split
from sklearn.neighbors import KNeighborsClassifier
from sklearn.preprocessing import MultiLabelBinarizer
from ..MTEBResults import HFSubset, ScoresDict
from .AbsTask import AbsTask
logger = logging.getLogger(__name__)
def evaluate_classifier(
embeddings_train: np.ndarray,
y_train: np.ndarray,
embeddings_test: np.ndarray,
y_test: np.ndarray,
classifier: ClassifierMixin,
):
scores = {}
classifier = clone(classifier)
classifier.fit(embeddings_train, y_train)
y_pred = classifier.predict(embeddings_test)
accuracy = classifier.score(embeddings_test, y_test)
f1 = f1_score(y_test, y_pred, average="macro")
scores["accuracy"] = accuracy
scores["f1"] = f1
lrap = label_ranking_average_precision_score(y_test, y_pred)
scores["lrap"] = lrap
return scores
class AbsTaskMultilabelClassification(AbsTask):
"""Abstract class for multioutput classification tasks
The similarity is computed between pairs and the results are ranked.
self.load_data() must generate a huggingface dataset with a split matching self.metadata_dict["eval_splits"], and assign it to self.dataset. It must contain the following columns:
text: str
label: list[Hashable]
"""
classifier = KNeighborsClassifier(n_neighbors=5)
def __init__(
self,
n_experiments=None,
samples_per_label=None,
batch_size=32,
**kwargs,
):
super().__init__(**kwargs)
self.batch_size = batch_size
# Bootstrap parameters
self.n_experiments = n_experiments or getattr(self, "n_experiments", 10)
self.samples_per_label = samples_per_label or getattr(
self, "samples_per_label", 8
)
# Run metadata validation by instantiating addressing the attribute
# This is quite hacky. Ideally, this would be done in the constructor of
# each concrete task, but then we have to duplicate the __init__ method's
# interface.
if hasattr(self, "metadata"):
self.metadata
def _add_main_score(self, scores):
scores["main_score"] = scores[self.metadata.main_score]
def evaluate(
self, model, eval_split="test", train_split="train", **kwargs
) -> dict[HFSubset, ScoresDict]:
if not self.data_loaded:
self.load_data()
scores = {}
hf_subsets = [l for l in self.dataset] if self.is_multilingual else ["default"]
for hf_subset in hf_subsets:
logger.info(
f"\nTask: {self.metadata.name}, split: {eval_split}, subset: {hf_subset}. Running..."
)
if hf_subset not in self.dataset and hf_subset == "default":
ds = self.dataset
else:
ds = self.dataset[hf_subset]
scores[hf_subset] = self._evaluate_subset(
model, ds, eval_split, train_split, **kwargs
)
self._add_main_score(scores[hf_subset])
return scores
def _evaluate_subset(
self, model, dataset, eval_split="test", train_split="train", **kwargs
) -> ScoresDict:
train_split = dataset[train_split]
eval_split = dataset[eval_split]
params = {
"classifier_type": type(self.classifier).__name__,
"classifier_params": self.classifier.get_params(),
"batch_size": self.batch_size,
}
params.update(kwargs)
scores = []
# Bootstrap sample indices from training set for each experiment
train_samples = []
for _ in range(self.n_experiments):
sample_indices, _ = self._undersample_data_indices(
train_split["label"], self.samples_per_label, None
)
train_samples.append(sample_indices)
# Encode all unique sentences at the indices
unique_train_indices = list(set(itertools.chain.from_iterable(train_samples)))
unique_train_sentences = train_split.select(unique_train_indices)["text"]
unique_train_embeddings = dict(
zip(unique_train_indices, model.encode(unique_train_sentences))
)
test_text = eval_split["text"]
binarizer = MultiLabelBinarizer()
y_test = binarizer.fit_transform(eval_split["label"])
# Stratified subsampling of test set to 2000 examples.
try:
if len(test_text) > 2000:
test_text, _, y_test, _ = train_test_split(
test_text, y_test, stratify=y_test, train_size=2000
)
except ValueError:
logger.warn("Couldn't subsample, continuing with the entire test set.")
X_test = model.encode(test_text)
for i_experiment, sample_indices in enumerate(train_samples):
logger.info(
"=" * 10
+ f" Experiment {i_experiment+1}/{self.n_experiments} "
+ "=" * 10
)
X_train = np.stack([unique_train_embeddings[idx] for idx in sample_indices])
y_train = train_split.select(sample_indices)["label"]
y_train = binarizer.transform(y_train)
scores_exp = evaluate_classifier(
X_train, y_train, X_test, y_test, self.classifier
)
scores.append(scores_exp)
avg_scores: dict[str, Any] = {
k: np.mean([s[k] for s in scores]) for k in scores[0].keys()
}
avg_scores["scores_per_experiment"] = scores
return avg_scores
def _undersample_data_indices(self, y, samples_per_label, idxs=None):
"""Undersample data to have samples_per_label samples of each label"""
sample_indices = []
if idxs is None:
idxs = np.arange(len(y))
np.random.shuffle(idxs)
label_counter = defaultdict(int)
for i in idxs:
if any((label_counter[label] < samples_per_label) for label in y[i]):
sample_indices.append(i)
for label in y[i]:
label_counter[label] += 1
return sample_indices, idxs