Datasets:
File size: 3,174 Bytes
c1b249f 16702ad c1b249f 16702ad c1b249f 16702ad c1b249f 16702ad c1b249f 16702ad c1b249f 16702ad 1c1567d 16702ad 1c1567d 16702ad 1c1567d 16702ad 1c1567d 16702ad 1c1567d 16702ad 1c1567d 16702ad c1b249f 16702ad c1b249f 16702ad c1b249f 16702ad | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 | ---
language:
- en
- ru
license: mit
tags:
- ai-text-detection
- reproducibility
- bilingual
- adversarial-robustness
- calibration
---
# ContentOS — Reproducible Bilingual AI-Text-Detection Ensemble
**Pre-print v1.0 (2026-04-27)**
This repository contains the open pre-print and supporting artifacts for
ContentOS, a reproducible English+Russian AI-text-detection ensemble.
## Authors and affiliation
- **Gregory Shevchenko** — author, Humanswith.ai (founder)
- **Humanswith.ai team** — methodology, calibration, evaluation infrastructure
ContentOS is a Humanswith.ai product. This preprint is published under
the author's personal HuggingFace account; the supporting code repository
is maintained under the organization account (see "Code repository" below).
- Author profile: https://huggingface.co/gshevchenko
- Organization: https://humanswith.ai
- Contact for collaboration: open a Discussion on this dataset
## Code repository
Public benchmark + evaluation scripts:
**https://github.com/humanswith-ai/contentos-benchmark**
The repo includes regression test suite (8 pinned baselines, 0.05s),
streaming-CSV eval scripts (partial-tolerant), per-genre AUROC
analyzer, and the calibration JSON shape for v1.11 production state.
## Headline numbers (v1.11 production, 2026-04-29 measurement)
| Metric | EN | RU |
|---|---|---|
| OOD AUROC (176-sample expanded smoke) | **0.864** | **0.846** |
| Wrong-rate | 4% | 9% |
| p50 latency (EN ensemble) | **1.2 s** | — |
| Adversarial AUROC (n=300, OOD-paired) | **0.998** | — |
Earlier v1.0 paper reported 0.802 / 0.847 on the original 44-text
smoke battery; the 4× expanded battery with class balance per
(lang, genre) cell stabilized numbers upward. Per-genre details in
the [companion repo](https://github.com/humanswith-ai/contentos-benchmark).
## Files
- `paper.pdf` — full pre-print (~6,000 words, 9 sections + 5 appendices)
- `paper.html` — self-contained HTML version with embedded figures
- `paper.md` — source markdown
- `figures/` — 4 figures (PNG + SVG)
- `REPRODUCIBILITY.md` — open methodology, how to reproduce in 90 minutes
## Reproducibility
The full methodology and calibration corpus description are documented in
`REPRODUCIBILITY.md`, which is sufficient for independent re-implementation
of the ensemble.
A public mirror with the evaluation scripts (`eval_ensemble_corpus.py`,
8 pinned regression tests, atomic-swap deploy with 30-second rollback)
will be released within ~2 weeks following the v1.12 RU recalibration
chain. Target reproduction infrastructure: Hetzner CX43 (8 vCPU, no GPU,
~€14/month) or equivalent.
For early access before the public mirror, please open a discussion on
this dataset.
## Cite as
```bibtex
@misc{contentos2026,
title={ContentOS: A Reproducible Bilingual AI-Text-Detection Ensemble with Adversarial Robustness Evaluation},
author={Humanswith.ai team},
year={2026},
url={https://huggingface.co/datasets/gshevchenko/contentos-preprint},
}
```
## License
MIT for code, methodology, and corpus aggregation. Underlying data sources retain their original licenses (HC3, AINL-Eval-2025, ai-text-detection-pile).
|