from __future__ import annotations import logging import numpy as np from mteb.MTEBResults import ScoresDict from ..evaluation.evaluators import SummarizationEvaluator from .AbsTask import AbsTask logger = logging.getLogger(__name__) class AbsTaskSummarization(AbsTask): """Abstract class for summarization experiments. self.load_data() must generate a huggingface dataset with a split matching self.metadata_dict["eval_splits"], and assign it to self.dataset. It must contain the following columns: text: str human_summaries: list[str] machine_summaries: list[str] relevance: list[float] (the score of the machine generated summaries) """ def __init__(self, **kwargs): super().__init__(**kwargs) @property def min_score(self): return self.metadata_dict["min_score"] @property def max_score(self): return self.metadata_dict["max_score"] def _evaluate_subset(self, model, data_split, **kwargs) -> ScoresDict: normalized_scores = list( map( lambda x: (np.array(x) - self.min_score) / (self.max_score - self.min_score), data_split["relevance"], ) ) evaluator = SummarizationEvaluator( machine_summaries=data_split["machine_summaries"], human_summaries=data_split["human_summaries"], texts=data_split["text"], gold_scores=normalized_scores, **kwargs, ) scores = evaluator(model) self._add_main_score(scores) return scores def _add_main_score(self, scores: ScoresDict) -> None: m_score = self.metadata.main_score dist, metric = m_score.split("_") dist_mapping = {"cosine": "cos_sim"} scores["main_score"] = scores[dist_mapping.get(dist, dist)][metric]