ClarusC64's picture
Update README.md
98fde0f verified
---
language:
- en
license: mit
pretty_name: Clarus Clinical Stability Benchmark
tags:
- clarusc64
- stability-reasoning
- clinical-benchmark
- tabular-reasoning
- system-stability
- trajectory-analysis
---
# Benchmark Documentation
Core
- benchmark_structure.md
- benchmark_matrix.md
- datasets.md
Evaluation
- evaluation_framework.md
- transfer_matrix.md
- clarus_score.md
Robustness
- missing_data_protocol.md
- imbalance_protocol.md
- robustness_suite.md
Theory
- stability_manifold.md
- stability_topology.md
- stability_mechanisms.md
Results
- baseline_results.md
- leaderboard.md
# Clarus Clinical Stability Benchmark
The Clarus Clinical Stability Benchmark evaluates whether machine learning models can detect **latent instability in complex clinical systems**.
Most tabular benchmarks reward models for learning correlations within a single dataset.
The Clarus benchmark instead evaluates whether models can infer instability from **interacting proxy signals across multiple physiological and operational regimes**.
Each dataset represents a simplified regime in which instability emerges from multi-variable interaction rather than single-variable thresholds.
---
# Benchmark Concept
In real clinical systems, deterioration rarely occurs because one measurement crosses a threshold.
Instead, instability emerges when several components drift simultaneously.
Examples include:
- circulatory compensation failure
- microvascular perfusion loss
- metabolic energy collapse
- respiratory control failure
- endocrine dysregulation
- thermoregulatory breakdown
- coagulation instability
- hospital operational overload
Each dataset exposes a different regime while keeping the underlying structure similar:
**instability arises from interacting system signals.**
The generative rules that determine the labels are intentionally not published.
Models must infer instability from observable proxies.
---
# Included Datasets
| Stability Regime | Dataset |
|---|---|
| Hemodynamic collapse | ClarusC64/clinical-hemodynamic-collapse-v0.1 |
| Sepsis trajectory instability | ClarusC64/clinical-sepsis-trajectory-instability-v0.1 |
| Intervention delay failure | ClarusC64/clinical-intervention-delay-failure-v0.1 |
| Organ coupling cascade | ClarusC64/clinical-organ-coupling-cascade-v0.1 |
| Recovery window detection | ClarusC64/clinical-recovery-window-detection-v0.1 |
| Ventilation–Perfusion instability | ClarusC64/clinical-ventilation-perfusion-instability-v0.1 |
| Hemorrhage compensation collapse | ClarusC64/clinical-hemorrhage-compensation-collapse-v0.1 |
| Electrolyte instability | ClarusC64/clinical-electrolyte-instability-v0.1 |
| Microcirculation instability | ClarusC64/clinical-microcirculation-instability-v0.1 |
| Endocrine instability | ClarusC64/clinical-endocrine-instability-v0.1 |
| Thermoregulation instability | ClarusC64/clinical-thermoregulation-instability-v0.1 |
| Cellular energy instability | ClarusC64/clinical-cellular-energy-instability-v0.1 |
| Respiratory drive instability | ClarusC64/clinical-respiratory-drive-instability-v0.1 |
| Coagulation instability | ClarusC64/clinical-coagulation-instability-v0.1 |
| Hospital operational collapse | ClarusC64/clinical-hospital-operational-collapse-v0.1 |
Each dataset repository contains:
data/train.csv
data/test.csv
scorer.py
README.md
---
# Evaluation Protocol
Predictions must follow the format:
scenario_id,prediction
Example:
MC101,0
MC102,1
Evaluation is performed using the **scorer located in the dataset repository**.
Example:
python scorer.py --predictions predictions.csv --truth data/test.csv --output metrics.json
The `--truth` path refers to the dataset's local `data/test.csv` file.
Metrics reported include:
- accuracy
- precision
- recall
- f1
- confusion matrix
---
# Benchmark Tasks
The benchmark supports three evaluation settings.
## 1 Single-Dataset Evaluation
Train and test on the same dataset.
Purpose:
Measure baseline performance within a single stability regime.
---
## 2 Cross-Regime Transfer
Train on one regime and test on another.
Example:
Train → clinical-hemodynamic-collapse-v0.1
Test → clinical-microcirculation-instability-v0.1
Purpose:
Determine whether models learn **general stability reasoning** rather than dataset-specific correlations.
---
## 3 Multi-Regime Training
Train on multiple datasets simultaneously.
Evaluate performance across all regimes.
Purpose:
Test whether models can learn shared stability representations across physiological systems.
---
# Dataset Design Principles
The Clarus datasets follow several explicit design rules.
### No Single-Feature Dominance
No observable variable strongly predicts the label independently.
Target:
|correlation| < 0.30
---
### Interaction-Based Labels
Instability emerges from interactions between multiple variables rather than isolated thresholds.
---
### Adversarial Symmetry
Rows with nearly identical values may produce opposite labels.
This prevents trivial heuristics.
---
### Decoy Variables
Some variables appear meaningful but do not determine the label independently.
---
### Hidden Generative Logic
The dataset generator and rule equations are intentionally not published.
Models must infer instability from proxy signals.
---
# Baseline Results
Reference baseline experiments are provided in:
baseline_results.md
These establish approximate difficulty levels for common tabular models.
---
# Benchmark Architecture
The benchmark can be interpreted as observing a **shared stability manifold** through different clinical regimes.
Each dataset exposes a different control system while preserving the underlying concept of instability emerging from interacting signals.
Additional details are provided in:
stability_manifold.md
---
# Research Applications
The benchmark supports research into:
- system stability reasoning
- interaction-based tabular learning
- cross-domain generalization
- clinical early warning modeling
- infrastructure and system risk detection
---
Quick Start
# Quick Start
This example demonstrates how to evaluate a simple model on one Clarus dataset.
---
## 1 Install dependencies
Example environment:
pip install pandas scikit-learn
---
## 2 Load the dataset
train = data/train.csv
test = data/test.csv
---
## 3 Train a simple baseline model
Example using logistic regression:
import pandas as pd
from sklearn.linear_model import LogisticRegression
train = pd.read_csv("data/train.csv")
X = train.drop(columns=["scenario_id","label"])
y = train["label"]
model = LogisticRegression()
model.fit(X, y)
---
## 4 Generate predictions
test = pd.read_csv("data/test.csv")
X_test = test.drop(columns=["scenario_id","label"])
pred = model.predict(X_test)
out = pd.DataFrame({
"scenario_id": test["scenario_id"],
"prediction": pred
})
out.to_csv("predictions.csv", index=False)
---
## 5 Evaluate predictions
Run the official scorer:
python scorer.py --predictions predictions.csv --truth data/test.csv
The scorer returns:
- accuracy
- precision
- recall
- f1
- confusion matrix
# License
MIT