hc99's picture
Add files using upload-large-folder tool
83d24b2 verified
raw
history blame
2.37 kB
from __future__ import annotations
import logging
from datasets import Dataset
from ..evaluation.evaluators import BitextMiningEvaluator
from ..MTEBResults import HFSubset, ScoresDict
from .AbsTask import AbsTask
logger = logging.getLogger(__name__)
class AbsTaskBitextMining(AbsTask):
"""Abstract class for BitextMining tasks
The similarity is computed between pairs and the results are ranked.
self.load_data() must generate a huggingface dataset with a split matching self.metadata_dict["eval_splits"], and assign it to self.dataset. It must contain the following columns:
id: str
sentence1: str
sentence2: str
"""
parallel_subsets = False
def __init__(self, **kwargs):
super().__init__(**kwargs)
def evaluate(self, model, split, **kwargs) -> dict[HFSubset, ScoresDict]:
if not self.data_loaded:
self.load_data()
hf_subsets = (
[l for l in self.dataset]
if self.is_multilingual or self.is_crosslingual
else ["default"]
)
scores = {}
if self.parallel_subsets:
scores["default"] = self._evaluate_subset(
model, self.dataset[split], parallel=True, **kwargs
)
else:
for hf_subet in hf_subsets:
logger.info(
f"\nTask: {self.metadata_dict['name']}, split: {split}, subset: {hf_subet}. Running..."
)
if hf_subet not in self.dataset and hf_subet == "default":
data_split = self.dataset[split]
else:
data_split = self.dataset[hf_subet][split]
scores[hf_subet] = self._evaluate_subset(
model, data_split, subsets=["sentence1", "sentence2"], **kwargs
)
return scores
def _evaluate_subset(
self, model, data_split: Dataset, parallel=False, **kwargs
) -> ScoresDict:
evaluator = BitextMiningEvaluator(data_split, **kwargs)
metrics = evaluator(model)
if parallel:
for v in metrics.values():
self._add_main_score(v)
else:
self._add_main_score(metrics)
return metrics
def _add_main_score(self, scores) -> None:
scores["main_score"] = scores[self.metadata.main_score]