ClarusC64 commited on
Commit
3b25f8b
·
verified ·
1 Parent(s): 5607093

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +73 -59
README.md CHANGED
@@ -14,51 +14,45 @@ tags:
14
 
15
  # Clarus Clinical Stability Benchmark
16
 
17
- The Clarus Clinical Stability Benchmark evaluates whether machine learning models can infer latent system instability from interacting proxy signals across multiple physiological and operational regimes.
18
-
19
- Unlike traditional tabular benchmarks that reward threshold detection or single-feature correlations, Clarus datasets are designed so that instability emerges from interactions between variables.
20
-
21
- The benchmark therefore measures stability reasoning rather than pattern memorization.
22
-
23
  The Clarus Clinical Stability Benchmark evaluates whether machine learning models can detect **latent instability in complex clinical systems**.
24
 
25
- Most tabular benchmarks test pattern recognition from static variables.
26
-
27
- This benchmark instead evaluates **reasoning about interacting system signals** that determine whether a system remains stable or moves toward collapse.
28
 
29
- The datasets represent different physiological and operational regimes where instability emerges from **multi-variable interaction dynamics** rather than single-variable thresholds.
30
 
31
  ---
32
 
33
  # Benchmark Concept
34
 
35
- The core idea of the benchmark is **latent stability geometry**.
36
 
37
- In real clinical systems, instability arises when multiple interacting components drift simultaneously.
38
 
39
  Examples include:
40
 
41
- - circulatory compensation failure
42
- - microvascular perfusion loss
43
- - metabolic energy collapse
44
- - respiratory control failure
45
- - endocrine dysregulation
46
- - thermoregulatory breakdown
47
- - coagulation instability
48
- - hospital operational overload
49
 
50
- Each dataset captures one regime.
 
51
 
52
- The true generative logic is not published.
53
 
54
- Models must infer instability from **interacting proxy signals**.
55
 
56
  ---
57
 
58
  # Included Datasets
59
 
60
- | Regime | Dataset Repo |
61
- |------|------|
62
  | Hemodynamic collapse | ClarusC64/clinical-hemodynamic-collapse-v0.1 |
63
  | Sepsis trajectory instability | ClarusC64/clinical-sepsis-trajectory-instability-v0.1 |
64
  | Intervention delay failure | ClarusC64/clinical-intervention-delay-failure-v0.1 |
@@ -75,18 +69,20 @@ Models must infer instability from **interacting proxy signals**.
75
  | Coagulation instability | ClarusC64/clinical-coagulation-instability-v0.1 |
76
  | Hospital operational collapse | ClarusC64/clinical-hospital-operational-collapse-v0.1 |
77
 
78
- Each dataset includes:
79
-
80
- train.csv
81
- test.csv
82
- scorer.py
83
  README.md
84
 
 
85
  ---
86
 
87
  # Evaluation Protocol
88
 
89
- All datasets use the same prediction format:
 
 
90
  scenario_id,prediction
91
 
92
 
@@ -97,20 +93,23 @@ MC101,0
97
  MC102,1
98
 
99
 
100
- Predictions are evaluated with the official Clarus scorer:
 
 
101
 
102
 
103
  python scorer.py --predictions predictions.csv --truth data/test.csv --output metrics.json
104
 
105
 
106
- Metrics:
 
 
107
 
108
  - accuracy
109
  - precision
110
  - recall
111
  - f1
112
  - confusion matrix
113
- - dataset integrity diagnostics
114
 
115
  ---
116
 
@@ -122,7 +121,9 @@ The benchmark supports three evaluation settings.
122
 
123
  Train and test on the same dataset.
124
 
125
- This measures performance on a single stability regime.
 
 
126
 
127
  ---
128
 
@@ -132,15 +133,12 @@ Train on one regime and test on another.
132
 
133
  Example:
134
 
135
- Train:
136
-
137
- clinical-hemodynamic-collapse
138
 
139
- Test:
140
 
141
- clinical-microcirculation-instability
142
-
143
- Large performance drops indicate the model learned **surface patterns rather than stability reasoning**.
144
 
145
  ---
146
 
@@ -148,17 +146,21 @@ Large performance drops indicate the model learned **surface patterns rather tha
148
 
149
  Train on multiple datasets simultaneously.
150
 
151
- Evaluate whether models can learn **general stability reasoning across physiological systems**.
 
 
 
 
152
 
153
  ---
154
 
155
  # Dataset Design Principles
156
 
157
- The Clarus datasets follow several design rules.
158
 
159
  ### No Single-Feature Dominance
160
 
161
- No observable variable strongly predicts the label alone.
162
 
163
  Target:
164
 
@@ -168,7 +170,7 @@ Target:
168
 
169
  ### Interaction-Based Labels
170
 
171
- Instability emerges from interactions between multiple variables.
172
 
173
  ---
174
 
@@ -182,26 +184,40 @@ This prevents trivial heuristics.
182
 
183
  ### Decoy Variables
184
 
185
- Some variables appear meaningful but are not part of the label rule.
186
 
187
  ---
188
 
189
  ### Hidden Generative Logic
190
 
191
- The dataset generator and latent rule equations are not published.
192
 
193
  Models must infer instability from proxy signals.
194
 
195
  ---
196
 
197
- # Structural Note
 
 
198
 
199
- These datasets represent simplified proxies for real clinical system dynamics.
200
 
201
- They are designed for research into **latent stability reasoning**, not clinical decision support.
 
202
 
203
  ---
204
- - stability_manifold.md — explains the shared stability geometry across regimes
 
 
 
 
 
 
 
 
 
 
 
205
 
206
  ---
207
 
@@ -209,13 +225,11 @@ They are designed for research into **latent stability reasoning**, not clinical
209
 
210
  The benchmark supports research into:
211
 
212
- - stability reasoning in machine learning
213
- - interaction-based tabular reasoning
214
- - cross-domain system modeling
215
- - clinical early warning systems
216
- - multi-variable instability detection
217
-
218
- The Clarus benchmark evaluates whether models can detect latent instability across interacting system regimes rather than learning surface patterns within a single dataset.
219
 
220
  ---
221
 
 
14
 
15
  # Clarus Clinical Stability Benchmark
16
 
 
 
 
 
 
 
17
  The Clarus Clinical Stability Benchmark evaluates whether machine learning models can detect **latent instability in complex clinical systems**.
18
 
19
+ Most tabular benchmarks reward models for learning correlations within a single dataset.
20
+ The Clarus benchmark instead evaluates whether models can infer instability from **interacting proxy signals across multiple physiological and operational regimes**.
 
21
 
22
+ Each dataset represents a simplified regime in which instability emerges from multi-variable interaction rather than single-variable thresholds.
23
 
24
  ---
25
 
26
  # Benchmark Concept
27
 
28
+ In real clinical systems, deterioration rarely occurs because one measurement crosses a threshold.
29
 
30
+ Instead, instability emerges when several components drift simultaneously.
31
 
32
  Examples include:
33
 
34
+ - circulatory compensation failure
35
+ - microvascular perfusion loss
36
+ - metabolic energy collapse
37
+ - respiratory control failure
38
+ - endocrine dysregulation
39
+ - thermoregulatory breakdown
40
+ - coagulation instability
41
+ - hospital operational overload
42
 
43
+ Each dataset exposes a different regime while keeping the underlying structure similar:
44
+ **instability arises from interacting system signals.**
45
 
46
+ The generative rules that determine the labels are intentionally not published.
47
 
48
+ Models must infer instability from observable proxies.
49
 
50
  ---
51
 
52
  # Included Datasets
53
 
54
+ | Stability Regime | Dataset |
55
+ |---|---|
56
  | Hemodynamic collapse | ClarusC64/clinical-hemodynamic-collapse-v0.1 |
57
  | Sepsis trajectory instability | ClarusC64/clinical-sepsis-trajectory-instability-v0.1 |
58
  | Intervention delay failure | ClarusC64/clinical-intervention-delay-failure-v0.1 |
 
69
  | Coagulation instability | ClarusC64/clinical-coagulation-instability-v0.1 |
70
  | Hospital operational collapse | ClarusC64/clinical-hospital-operational-collapse-v0.1 |
71
 
72
+ Each dataset repository contains:
73
+ data/train.csv
74
+ data/test.csv
75
+ scorer.py
 
76
  README.md
77
 
78
+
79
  ---
80
 
81
  # Evaluation Protocol
82
 
83
+ Predictions must follow the format:
84
+
85
+
86
  scenario_id,prediction
87
 
88
 
 
93
  MC102,1
94
 
95
 
96
+ Evaluation is performed using the **scorer located in the dataset repository**.
97
+
98
+ Example:
99
 
100
 
101
  python scorer.py --predictions predictions.csv --truth data/test.csv --output metrics.json
102
 
103
 
104
+ The `--truth` path refers to the dataset's local `data/test.csv` file.
105
+
106
+ Metrics reported include:
107
 
108
  - accuracy
109
  - precision
110
  - recall
111
  - f1
112
  - confusion matrix
 
113
 
114
  ---
115
 
 
121
 
122
  Train and test on the same dataset.
123
 
124
+ Purpose:
125
+
126
+ Measure baseline performance within a single stability regime.
127
 
128
  ---
129
 
 
133
 
134
  Example:
135
 
136
+ Train → clinical-hemodynamic-collapse-v0.1
137
+ Test → clinical-microcirculation-instability-v0.1
 
138
 
139
+ Purpose:
140
 
141
+ Determine whether models learn **general stability reasoning** rather than dataset-specific correlations.
 
 
142
 
143
  ---
144
 
 
146
 
147
  Train on multiple datasets simultaneously.
148
 
149
+ Evaluate performance across all regimes.
150
+
151
+ Purpose:
152
+
153
+ Test whether models can learn shared stability representations across physiological systems.
154
 
155
  ---
156
 
157
  # Dataset Design Principles
158
 
159
+ The Clarus datasets follow several explicit design rules.
160
 
161
  ### No Single-Feature Dominance
162
 
163
+ No observable variable strongly predicts the label independently.
164
 
165
  Target:
166
 
 
170
 
171
  ### Interaction-Based Labels
172
 
173
+ Instability emerges from interactions between multiple variables rather than isolated thresholds.
174
 
175
  ---
176
 
 
184
 
185
  ### Decoy Variables
186
 
187
+ Some variables appear meaningful but do not determine the label independently.
188
 
189
  ---
190
 
191
  ### Hidden Generative Logic
192
 
193
+ The dataset generator and rule equations are intentionally not published.
194
 
195
  Models must infer instability from proxy signals.
196
 
197
  ---
198
 
199
+ # Baseline Results
200
+
201
+ Reference baseline experiments are provided in:
202
 
203
+ baseline_results.md
204
 
205
+
206
+ These establish approximate difficulty levels for common tabular models.
207
 
208
  ---
209
+
210
+ # Benchmark Architecture
211
+
212
+ The benchmark can be interpreted as observing a **shared stability manifold** through different clinical regimes.
213
+
214
+ Each dataset exposes a different control system while preserving the underlying concept of instability emerging from interacting signals.
215
+
216
+ Additional details are provided in:
217
+
218
+
219
+ stability_manifold.md
220
+
221
 
222
  ---
223
 
 
225
 
226
  The benchmark supports research into:
227
 
228
+ - system stability reasoning
229
+ - interaction-based tabular learning
230
+ - cross-domain generalization
231
+ - clinical early warning modeling
232
+ - infrastructure and system risk detection
 
 
233
 
234
  ---
235