ClarusC64 commited on
Commit
a91cb23
·
verified ·
1 Parent(s): 877c8e1

Create benchmark_scope.md

Browse files
Files changed (1) hide show
  1. benchmark_scope.md +133 -0
benchmark_scope.md ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Benchmark Scope
2
+
3
+ The Clarus Stability Benchmark evaluates whether machine learning systems can detect **latent instability dynamics** across complex systems.
4
+
5
+ Most tabular benchmarks measure a model’s ability to detect statistical correlations in datasets.
6
+
7
+ The Clarus benchmark instead focuses on **stability reasoning** — the ability to detect when interacting system variables are approaching instability.
8
+
9
+ The benchmark is built around the idea that many complex systems share common instability mechanisms even when their surface variables differ.
10
+
11
+ Examples include:
12
+
13
+ - clinical physiological systems
14
+ - molecular and protein systems
15
+ - quantum computing systems
16
+
17
+ In each domain, instability arises when interacting pressures exceed the system’s capacity to maintain stability.
18
+
19
+ The datasets in this benchmark expose only **observable proxy variables**.
20
+
21
+ The latent stability rules and generators used to produce the datasets are not included.
22
+
23
+ This design ensures that models must infer instability from interactions between variables rather than from explicit rules.
24
+
25
+ ---
26
+
27
+ # System Domains
28
+
29
+ The benchmark currently spans three system scales.
30
+
31
+ ## Clinical Systems
32
+
33
+ Datasets describing physiological instability.
34
+
35
+ Examples include:
36
+
37
+ - circulation and perfusion collapse
38
+ - respiratory control instability
39
+ - renal filtration failure
40
+ - endocrine feedback instability
41
+ - metabolic supply-demand imbalance
42
+
43
+ These datasets simulate clinical monitoring conditions where multiple physiological signals interact over time.
44
+
45
+ ---
46
+
47
+ ## Molecular Systems
48
+
49
+ Datasets describing molecular stability and protein behavior.
50
+
51
+ Examples include:
52
+
53
+ - protein folding pathway instability
54
+ - mutation-driven structural destabilization
55
+ - aggregation risk
56
+ - chaperone rescue window failure
57
+ - protein interface collapse
58
+ - conformational switching instability
59
+
60
+ These datasets test whether models can detect instability in molecular interaction networks.
61
+
62
+ ---
63
+
64
+ ## Quantum Systems
65
+
66
+ Datasets describing instability in quantum computing devices.
67
+
68
+ Examples include:
69
+
70
+ - coherence collapse
71
+ - gate sequence instability
72
+ - entanglement decay
73
+ - error correction failure
74
+ - control pulse instability
75
+
76
+ These datasets represent simplified stability conditions in noisy intermediate-scale quantum (NISQ) devices.
77
+
78
+ ---
79
+
80
+ # Benchmark Design Principles
81
+
82
+ The benchmark follows several design constraints.
83
+
84
+ ### No Single-Feature Dominance
85
+
86
+ Labels cannot be predicted using a single variable.
87
+
88
+ Instability emerges from **interactions between variables**.
89
+
90
+ ### Hidden Stability Geometry
91
+
92
+ Datasets expose only observable proxies.
93
+
94
+ The latent stability rules used to generate labels are not published.
95
+
96
+ ### Adversarial Symmetry
97
+
98
+ Datasets include examples with very similar values but different outcomes.
99
+
100
+ This prevents models from relying on simple thresholds.
101
+
102
+ ### Mixed Instability Mechanisms
103
+
104
+ Datasets include multiple instability regimes within the same domain.
105
+
106
+ ---
107
+
108
+ # Evaluation Philosophy
109
+
110
+ The benchmark evaluates models across several reasoning levels.
111
+
112
+ 1. single-dataset prediction
113
+ 2. within-domain transfer
114
+ 3. cross-domain transfer
115
+ 4. missing observation robustness
116
+ 5. class imbalance robustness
117
+
118
+ These evaluation tasks test whether models learn **general instability reasoning** rather than dataset-specific patterns.
119
+
120
+ ---
121
+
122
+ # Intended Use
123
+
124
+ The Clarus Stability Benchmark is designed for research into:
125
+
126
+ - machine learning reasoning over complex systems
127
+ - stability detection in noisy environments
128
+ - cross-domain generalization
129
+ - robustness to incomplete observations
130
+
131
+ The benchmark is not a simulator for clinical, molecular, or quantum systems.
132
+
133
+ Instead it provides compact tabular datasets that express stability dynamics through observable proxies.