# HWAI AI-detection ensemble — reproducibility benchmark > **Thesis.** Closed commercial benchmarks (Originality.ai, Winston AI, GPTZero > Premium) publish AUROC numbers on proprietary corpora you cannot audit. > HWAI publishes its **exact smoke corpus**, **exact evaluation script**, and > **exact calibration files** so anyone can reproduce our numbers in ≤10 min. ## What you need to reproduce 1. A running instance of `ml-services-hwai` (Hetzner CX43 ~$25/mo, or local Docker). 2. Our calibration file (`/opt/ml-services/calibration.json`, shipped in v1.10). 3. The eval script: `services/ml-services-hwai/scripts/eval_ensemble_corpus.py`. 4. The hand-curated corpus: **embedded in the eval script as `CORPUS`** (44 texts, EN + RU, human + AI). All four above are open in this repo. ## Reference numbers (v1.10, 2026-04-24, 44-text hand-curated OOD) | Lang | Ensemble AUROC | Brier | Detectors | |---|---|---|---| | EN | **0.770** (target: ≥0.80 post-weight-tuning) | ~0.18 | ai_detect + radar + binoculars + desklib | | RU | **0.837** | ~0.12 | ai_detect + radar + binoculars | Per-detector numbers: see `notes/contentos_ensemble_weights_tuning_2026-04-24.md`. ## How to reproduce ```bash # On any host with Python 3.11+ and access to ml-services: export ML_SERVICES_URL="http://:3300" export ML_SERVICES_API_KEY="" python3 services/ml-services-hwai/scripts/eval_ensemble_corpus.py # Output: /tmp/eval_ensemble_.json + .md ``` The output includes: - Per-detector raw + calibrated scores, per sample. - Per-lang AUROC + Brier (ensemble + individual detectors). - Confusion-matrix style pass/fail per sample. ## Why 44 texts? Smoke battery sized for fast iteration (< 2 min wall-time), **hand-picked to cover the failure modes** we've seen in production: - **EN human formal** (press release, court filing, product manual) — biggest false-positive risk for models that learned "AI = formal"). - **RU human news/journalism** — known FP mode for RADAR-Vicuna. - **EN + RU AI 2026-era** (Claude-4, Gemini 2.5, GPT-4o style) — the distribution shift that breaks 2022-era-trained detectors. - **Edge cases** (interview transcripts, casual parent notes, product reviews) where detectors commonly overreach. Full statistical-grade numbers on `n=750` OOD calibration split: `services/ml-services-hwai/corpus/cal_test.jsonl`. Run: ```bash python3 services/ml-services-hwai/scripts/eval_ensemble_corpus.py \ --from-cal-test services/ml-services-hwai/corpus/cal_test.jsonl --parallel 4 ``` ## Compare with external detectors Free-tier cross-check (Sapling AI, 50 req/day): ```bash export SAPLING_API_KEY="..." python3 services/ml-services-hwai/scripts/bench_competitors.py \ --out /tmp/bench_competitors.json ``` Produces side-by-side AUROC on identical corpus. ## What this benchmark does NOT claim - It is **not** a universal ranking of AI detectors. Different corpora + different objectives (academic integrity vs SEO vs marketing QA) produce different rankings. - It is **not** statistically powered for `p<0.05` differences of `<0.02` AUROC. For that, use `cal_test.jsonl` (n=750). - It is **not** stable under adversarial paraphrase attack (no detector is). ## What it DOES claim - Full reproducibility: bit-identical inputs → bit-identical scores. - Calibration stability: pinned baseline via `test_calibration_regression.py` auto-runs on every cal swap; rollback on drop >0.05 AUROC. - Honest distribution reporting: the corpus is public so you can audit whether it matches your use-case. ## Cost to run full reproducibility | Step | Cost | |---|---| | Spin up ml-services (Hetzner CX43 monthly) | $25 | | Run smoke eval | $0 | | Run `cal_test.jsonl` n=750 statistical eval | $0 (self-hosted) | | Sapling cross-check (free tier) | $0 | | **Total** | **$25 one-time hosting** | Commercial equivalents: - Originality.ai: $15 trial + per-call costs, no AUROC audit, closed corpus. - GPTZero Premium: $15/mo, closed corpus. - Winston AI: $29/mo, closed corpus. ## Cross-references - Corpus balance analysis: `notes/contentos_corpus_balance_analysis_2026-04-24.md` - Ensemble weights tuning: `notes/contentos_ensemble_weights_tuning_2026-04-24.md` - Fork #2 v1 → v2 progression: `notes/contentos_fork2_en_selfgen_spec.md` - Regression test pins: `services/ml-services-hwai/tests/test_calibration_regression.py` - 7-blocks gap assessment: `notes/contentos_7blocks_gap_assessment_2026-04-24.md`