#!/usr/bin/env python3 """Render current failure-taxonomy summary tables and SVG figures. The legacy renderer uses the 35-row preliminary evidence table. This renderer uses the post-PR193 paper-grade taxonomy surface so paper/deck figures can cite the same CSV as the current failure-analysis text. """ from __future__ import annotations import csv import html from collections import Counter from pathlib import Path ROOT = Path(__file__).resolve().parents[1] METRICS_DIR = ROOT / "results" / "metrics" FIGURES_DIR = ROOT / "results" / "figures" SOURCE_CSV = METRICS_DIR / "failure_taxonomy_current.csv" LABELS = { "low_task_completion": "Task completion", "low_data_retrieval_accuracy": "Data retrieval accuracy", "low_agent_sequence_correct": "Agent sequence correctness", "low_generalized_result_verification": "Result verification", } PALETTE = ["#1f6f78", "#c85a3a", "#6f7d2a", "#5b6aa8"] def read_rows() -> list[dict[str, str]]: with SOURCE_CSV.open(newline="") as f: return list(csv.DictReader(f)) def write_csv(path: Path, fieldnames: list[str], rows: list[dict[str, object]]) -> None: path.parent.mkdir(parents=True, exist_ok=True) with path.open("w", newline="") as f: writer = csv.DictWriter(f, fieldnames=fieldnames, lineterminator="\n") writer.writeheader() writer.writerows(rows) def pct(numerator: int, denominator: int) -> str: return f"{(100.0 * numerator / denominator):.1f}" if denominator else "0.0" def xml(value: object) -> str: return html.escape(str(value), quote=True) def write_auto_label_counts(rows: list[dict[str, str]]) -> list[dict[str, object]]: paper_failed = [r for r in rows if r["paper_eligible"] == "true"] counts = Counter(r["auto_taxonomy_label"] for r in paper_failed) total = len(paper_failed) out = [] for label, count in counts.most_common(): out.append( { "auto_taxonomy_label": label, "display_label": LABELS.get(label, label.replace("_", " ")), "rows": count, "percent_of_paper_failures": pct(count, total), "source_rows": total, "source_csv": "results/metrics/failure_taxonomy_current.csv", } ) write_csv( METRICS_DIR / "failure_taxonomy_current_auto_label_counts.csv", [ "auto_taxonomy_label", "display_label", "rows", "percent_of_paper_failures", "source_rows", "source_csv", ], out, ) return out def write_failed_dim_counts(rows: list[dict[str, str]]) -> list[dict[str, object]]: paper_failed = [r for r in rows if r["paper_eligible"] == "true"] counts = Counter(r["failed_dim_count"] for r in paper_failed) total = len(paper_failed) out = [] for failed_dim_count in sorted(counts, key=lambda v: int(v)): count = counts[failed_dim_count] out.append( { "failed_dim_count": failed_dim_count, "rows": count, "percent_of_paper_failures": pct(count, total), "source_rows": total, "source_csv": "results/metrics/failure_taxonomy_current.csv", } ) write_csv( METRICS_DIR / "failure_taxonomy_current_failed_dim_counts.csv", [ "failed_dim_count", "rows", "percent_of_paper_failures", "source_rows", "source_csv", ], out, ) return out def write_manual_audit_counts(rows: list[dict[str, str]]) -> list[dict[str, object]]: audited = [r for r in rows if r["audit_status"] == "manual_confirmed"] categories = [ ("audit_decision", "audit_decision"), ("berkeley_label", "berkeley_label"), ("failure_stage", "failure_stage"), ] out = [] for section, column in categories: counts = Counter(r[column] or "blank" for r in audited) for value, count in counts.most_common(): out.append( { "section": section, "value": value, "rows": count, "percent_of_manual_sample": pct(count, len(audited)), "source_rows": len(audited), "source_csv": "results/metrics/failure_taxonomy_current.csv", } ) write_csv( METRICS_DIR / "failure_taxonomy_current_manual_audit_counts.csv", [ "section", "value", "rows", "percent_of_manual_sample", "source_rows", "source_csv", ], out, ) return out def svg_auto_label_bar_chart(rows: list[dict[str, object]], path: Path) -> None: width, height = 1080, 520 left, top = 390, 118 bar_h, gap = 56, 24 max_count = max(int(r["rows"]) for r in rows) or 1 parts = [ f'', '', 'Current paper-grade failure taxonomy', '1,276 paper-eligible failed judge rows from failure_taxonomy_current.csv', 'Auto label is the highest-priority failed judge dimension; manual audit sample is tracked separately.', ] for i, row in enumerate(rows): y = top + i * (bar_h + gap) count = int(row["rows"]) bar_w = int((width - left - 150) * count / max_count) color = PALETTE[i % len(PALETTE)] percent = row["percent_of_paper_failures"] parts.extend( [ f'{xml(row["display_label"])}', f'', f'{count} ({percent}%)', ] ) parts.append("") path.write_text("\n".join(parts) + "\n") def main() -> None: rows = read_rows() FIGURES_DIR.mkdir(parents=True, exist_ok=True) auto_label_counts = write_auto_label_counts(rows) failed_dim_counts = write_failed_dim_counts(rows) manual_counts = write_manual_audit_counts(rows) svg_auto_label_bar_chart( auto_label_counts, FIGURES_DIR / "failure_taxonomy_current_auto_label_counts.svg", ) print( "Rendered " f"{len(auto_label_counts)} auto-label rows, " f"{len(failed_dim_counts)} failed-dim rows, " f"{len(manual_counts)} manual-audit rows." ) if __name__ == "__main__": main()