FEA-Bench / testbed /embeddings-benchmark__mteb /mteb /abstasks /AbsTaskClassification.py
hc99's picture
Add files using upload-large-folder tool
83d24b2 verified
raw
history blame
5.63 kB
from __future__ import annotations
import logging
from collections import defaultdict
from typing import Any
import numpy as np
from ..evaluation.evaluators import (
kNNClassificationEvaluator,
kNNClassificationEvaluatorPytorch,
logRegClassificationEvaluator,
)
from ..MTEBResults import HFSubset, ScoresDict
from .AbsTask import AbsTask
logger = logging.getLogger(__name__)
class AbsTaskClassification(AbsTask):
"""Abstract class for kNN classification tasks
The similarity is computed between pairs and the results are ranked.
self.load_data() must generate a huggingface dataset with a split matching self.metadata_dict["eval_splits"], and assign it to self.dataset. It
must contain the following columns:
text: str
label: int
"""
def __init__(
self,
method: str = "logReg",
n_experiments: int | None = None,
samples_per_label: int | None = None,
k: int = 3,
batch_size: int = 32,
**kwargs,
):
super().__init__(**kwargs)
self.batch_size = batch_size
self.method = method
# Bootstrap parameters
self.n_experiments: int = ( # type: ignore
n_experiments
if n_experiments is not None
else self.metadata_dict.get("n_experiments", 10)
)
self.samples_per_label: int = ( # type: ignore
samples_per_label
if samples_per_label is not None
else self.metadata_dict.get("samples_per_label", 8)
)
# kNN parameters
self.k = k
# Run metadata validation by instantiating addressing the attribute
# This is quite hacky. Ideally, this would be done in the constructor of
# each concrete task, but then we have to duplicate the __init__ method's
# interface.
if hasattr(self, "metadata"):
self.metadata
def _add_main_score(self, scores: dict[HFSubset, ScoresDict]) -> None:
scores["main_score"] = scores[self.metadata.main_score]
def evaluate(
self, model, eval_split="test", train_split="train", **kwargs
) -> dict[HFSubset, ScoresDict]:
if not self.data_loaded:
self.load_data()
scores = {}
hf_subsets = [l for l in self.dataset] if self.is_multilingual else ["default"]
for hf_subset in hf_subsets:
logger.info(
f"\nTask: {self.metadata.name}, split: {eval_split}, subset: {hf_subset}. Running..."
)
if hf_subset not in self.dataset and hf_subset == "default":
ds = self.dataset
else:
ds = self.dataset[hf_subset]
scores[hf_subset] = self._evaluate_subset(
model, ds, eval_split, train_split, **kwargs
)
self._add_main_score(scores[hf_subset])
return scores
def _evaluate_subset(
self, model, dataset, eval_split="test", train_split="train", **kwargs
) -> ScoresDict:
train_split = dataset[train_split]
eval_split = dataset[eval_split]
params = {"k": self.k, "batch_size": self.batch_size}
params.update(kwargs)
scores = []
test_cache, idxs = (
None,
None,
) # we store idxs to make the shuffling reproducible
for i in range(self.n_experiments):
logger.info(
"=" * 10 + f" Experiment {i+1}/{self.n_experiments} " + "=" * 10
)
# Bootstrap `self.samples_per_label` samples per label for each split
X_sampled, y_sampled, idxs = self._undersample_data(
train_split["text"], train_split["label"], self.samples_per_label, idxs
)
if self.method == "kNN":
evaluator = kNNClassificationEvaluator(
X_sampled,
y_sampled,
eval_split["text"],
eval_split["label"],
**params,
)
elif self.method == "kNN-pytorch":
evaluator = kNNClassificationEvaluatorPytorch(
X_sampled,
y_sampled,
eval_split["text"],
eval_split["label"],
**params,
)
elif self.method == "logReg":
evaluator = logRegClassificationEvaluator(
X_sampled,
y_sampled,
eval_split["text"],
eval_split["label"],
**params,
)
else:
raise ValueError(f"Method {self.method} not supported")
scores_exp, test_cache = evaluator(model, test_cache=test_cache)
scores.append(scores_exp)
avg_scores: dict[str, Any] = {
k: np.mean([s[k] for s in scores]) for k in scores[0].keys()
}
avg_scores["scores_per_experiment"] = scores
return avg_scores
def _undersample_data(self, X, y, samples_per_label: int, idxs=None):
"""Undersample data to have samples_per_label samples of each label"""
X_sampled = []
y_sampled = []
if idxs is None:
idxs = np.arange(len(y))
np.random.shuffle(idxs)
label_counter = defaultdict(int)
for i in idxs:
if label_counter[y[i]] < samples_per_label:
X_sampled.append(X[i])
y_sampled.append(y[i])
label_counter[y[i]] += 1
return X_sampled, y_sampled, idxs