anonymousreview111's picture
Upload judgesense-benchmark dataset
8848f9e verified
# JudgeSense: A Benchmark for Prompt Sensitivity in LLM-as-a-Judge Systems
[![License: CC-BY-4.0](https://img.shields.io/badge/License-CC--BY--4.0-lightgrey.svg)](https://creativecommons.org/licenses/by/4.0/)
[![arXiv](https://img.shields.io/badge/arXiv-[REDACTED]-red.svg)]()
[![HuggingFace](https://img.shields.io/badge/dataset-HuggingFace-orange.svg)](https://huggingface.co/datasets/anonymousreview111/judgesense-benchmark)
---
## Overview
**JudgeSense** is a benchmark dataset of **500 hand-validated prompt pairs** for measuring prompt sensitivity in LLM-as-a-Judge evaluation systems. Each pair contains two differently phrased but semantically equivalent judge prompts applied to the same response, enabling rigorous measurement of how much a judge's decision changes due to prompt wording alone.
All 500 pairs were independently validated by two human annotators with full agreement: 500 confirmed semantically equivalent; 50 pairs involving Template 4 (polarity-inverted) were labeled non-equivalent by both annotators and excluded before publication.
The dataset covers four evaluation task types:
| Task | Source | Pairs | Labels |
|------|--------|-------|--------|
| **Factuality** | TruthfulQA | 125 | accurate / inaccurate |
| **Coherence** | SummEval | 125 | score_1 ... score_5 |
| **Preference** | MT-Bench | 125 | A / B |
| **Relevance** | BEIR | 125 | A / B |
---
## What This Enables
- **Prompt sensitivity evaluation** — measure how fragile a judge is to phrasing variation
- **LLM judge robustness benchmarking** — compare models on decision consistency
- **Detection of prompt-induced artifacts** — identify polarity inversions (T4) and other systematic biases
---
## Quick Start
```python
from utils.load_judgesense import load_task, load_all
from utils.compute_jss import compute_jss
# Load one task
pairs = load_task("factuality")
print(f"{len(pairs)} pairs loaded")
# Load all tasks
all_data = load_all()
# Compute JSS from your judge's decisions
jss = compute_jss(decisions_a, decisions_b)
print(f"JSS: {jss:.3f}")
```
Run the full example:
```bash
cd judgesense-benchmark
python examples/run_jss_example.py
```
---
## Dataset Schema
Each JSONL record has eight fields:
```json
{
"pair_id": "fact_001",
"task_type": "factuality",
"source_benchmark": "TruthfulQA",
"prompt_a": "Is this factually correct? Answer YES or NO only.\n\nResponse: ...",
"prompt_b": "Fact-check this response. Reply YES (correct) or NO (incorrect).\n\nResponse: ...",
"response_being_judged": "The Earth orbits around the Sun.",
"ground_truth_label": "accurate",
"semantic_equivalence_score": 1.0
}
```
---
## Metric: Judge Sensitivity Score (JSS)
JSS is the fraction of pairs where both prompt variants elicit the same decision from the judge:
```
JSS = (1/N) * sum( decisions_a[i] == decisions_b[i] )
```
- **JSS = 1.0** — perfectly consistent; the judge never changes its decision due to prompt phrasing
- **JSS = 0.0** — maximally sensitive; every decision flips between prompts
A high flip rate (= 1 - JSS) indicates the judge's apparent decisions are largely driven by prompt design rather than the content being evaluated.
---
## Benchmark Results (13 judges, pass-3)
### Coherence (most discriminating task)
| Model | JSS | Cohen's kappa |
|---|---|---|
| Claude Sonnet 4.5 | 0.99 | 0.986 |
| Qwen-2.5-72B | 0.92 | 0.842 |
| GPT-4o | 0.91 | 0.828 |
| GPT-5.5 | 0.83 | 0.694 |
| GPT-4o-mini | 0.78 | 0.627 |
| Claude Haiku 4.5 | 0.73 | 0.583 |
| Claude Opus 4.7 | 0.70 | 0.580 |
| LLaMA-3.1-70B | 0.55 | 0.338 |
| DeepSeek-R1 | 0.53 | 0.332 |
| Qwen 3.6 Flash | 0.51 | 0.372 |
| DeepSeek-V4 Flash | 0.50 | 0.349 |
| Mistral-7B | 0.48 | -0.082 |
| Gemini 2.5 Flash | 0.39 | -0.057 |
### Factuality (after T4 polarity correction)
| Model | JSS (raw) | JSS (corrected) | Delta |
|---|---|---|---|
| GPT-4o | 0.63 | 0.98 | +0.35 |
| GPT-4o-mini | 0.63 | 0.96 | +0.33 |
| Claude Haiku 4.5 | 0.63 | 0.97 | +0.34 |
| Claude Sonnet 4.5 | 0.63 | 0.97 | +0.34 |
| DeepSeek-R1 | 0.63 | 0.96 | +0.33 |
| LLaMA-3.1-70B | 0.63 | 0.99 | +0.36 |
| Gemini 2.5 Flash | 0.63 | 0.98 | +0.35 |
| Qwen-2.5-72B | 0.63 | 0.98 | +0.35 |
| Mistral-7B | 0.71 | 0.89 | +0.18 |
| GPT-5.5 | 0.63 | 0.98 | +0.35 |
| Claude Opus 4.7 | 0.63 | 0.99 | +0.36 |
| Qwen 3.6 Flash | 0.63 | 0.97 | +0.34 |
| DeepSeek-V4 Flash | 0.62 | 0.95 | +0.33 |
---
## Key Insights
> **Coherence JSS varies by more than 0.6 units across 13 judges and does not track model scale or recency.**
- Claude Opus 4.7 (0.70) scores lower than Claude Haiku 4.5 (0.73); GPT-5.5 (0.83) scores lower than GPT-4o (0.91)
- Factuality JSS ranges from 0.89 to 0.99 after T4 correction; residual variation reflects genuine model-level differences
- Preference and relevance JSS are degenerate (12 of 13 judges always select option A)
---
## Citation
If you use JudgeSense in your research, please cite the accompanying paper (details redacted for double-blind review).
---
## License
- **Dataset**: [CC-BY-4.0](https://creativecommons.org/licenses/by/4.0/)
- **Code**: MIT License
---
*Anonymous submission for double-blind review. All evaluations conducted on public benchmarks and APIs.*