ClarusC64 commited on
Commit
877c8e1
·
verified ·
1 Parent(s): a38c42c

Create evaluation_framework.md

Browse files
Files changed (1) hide show
  1. evaluation_framework.md +178 -0
evaluation_framework.md ADDED
@@ -0,0 +1,178 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Clarus Evaluation Framework
2
+
3
+ The Clarus Stability Benchmark evaluates whether machine learning models can detect **latent instability dynamics** rather than relying on simple correlations.
4
+
5
+ The evaluation framework measures model capability across multiple reasoning levels.
6
+
7
+ These levels test progressively more difficult forms of stability reasoning.
8
+
9
+ ---
10
+
11
+ # Evaluation Levels
12
+
13
+ ## Level 1 — Single Dataset Evaluation
14
+
15
+ Models are trained and evaluated on the same dataset.
16
+
17
+ Purpose:
18
+
19
+ Measure whether the model can detect instability patterns within a single system.
20
+
21
+ Procedure:
22
+
23
+ 1. Train model on `train.csv`
24
+ 2. Generate predictions for `test.csv`
25
+ 3. Evaluate using the dataset scorer
26
+
27
+ Metrics:
28
+
29
+ - accuracy
30
+ - precision
31
+ - recall
32
+ - f1 score
33
+ - confusion matrix
34
+
35
+ Limitations:
36
+
37
+ High performance at this level does not necessarily indicate true stability reasoning.
38
+
39
+ Models may rely on dataset-specific correlations.
40
+
41
+ ---
42
+
43
+ ## Level 2 — Within-Domain Transfer
44
+
45
+ Models are trained on one dataset and evaluated on a different dataset within the same domain.
46
+
47
+ Example:
48
+
49
+ Train on:
50
+
51
+ protein-folding-pathway-instability
52
+
53
+ Test on:
54
+
55
+ protein-aggregation-risk-instability
56
+
57
+ Purpose:
58
+
59
+ Evaluate whether models can generalize stability reasoning across related systems.
60
+
61
+ Evaluation procedure:
62
+
63
+ 1. Train on source dataset
64
+ 2. Predict on target dataset
65
+ 3. Score using target dataset scorer
66
+
67
+ ---
68
+
69
+ ## Level 3 — Cross-Domain Transfer
70
+
71
+ Models are trained on one system domain and evaluated on another.
72
+
73
+ Domains in the Clarus benchmark include:
74
+
75
+ - clinical systems
76
+ - molecular / protein systems
77
+ - quantum systems
78
+
79
+ Example transfer tasks:
80
+
81
+ | Train Domain | Test Domain |
82
+ |---|---|
83
+ | clinical | clinical |
84
+ | protein | protein |
85
+ | quantum | quantum |
86
+ | clinical | protein |
87
+ | protein | quantum |
88
+ | quantum | clinical |
89
+
90
+ Purpose:
91
+
92
+ Determine whether models learn general **stability geometry** rather than domain-specific patterns.
93
+
94
+ ---
95
+
96
+ # Robustness Evaluation
97
+
98
+ The benchmark includes robustness tests that simulate real-world system conditions.
99
+
100
+ ---
101
+
102
+ ## Missing Data Evaluation
103
+
104
+ Real systems often contain incomplete observations.
105
+
106
+ Trajectory datasets may include variants where timepoints are missing.
107
+
108
+ Variants include:
109
+
110
+ - missing t0
111
+ - missing t1
112
+ - missing t2
113
+ - random missing
114
+
115
+ Purpose:
116
+
117
+ Evaluate whether models can infer stability dynamics when observations are incomplete.
118
+
119
+ The prediction task remains unchanged.
120
+
121
+ ---
122
+
123
+ ## Class Imbalance Evaluation
124
+
125
+ Many real-world systems exhibit rare failure events.
126
+
127
+ Datasets may include variants with different class distributions.
128
+
129
+ Supported imbalance regimes:
130
+
131
+ - balanced (50 / 50)
132
+ - mild imbalance (70 / 30)
133
+ - severe imbalance (90 / 10)
134
+ - extreme imbalance (99 / 1)
135
+
136
+ Purpose:
137
+
138
+ Evaluate whether models detect true instability rather than relying on class prevalence.
139
+
140
+ Accuracy alone is insufficient under imbalance conditions.
141
+
142
+ Recommended metrics:
143
+
144
+ - precision
145
+ - recall
146
+ - F1 score
147
+
148
+ ---
149
+
150
+ # Transfer Stability Score
151
+
152
+ To summarize transfer performance, the benchmark defines the **Transfer Stability Score (TSS)**.
153
+
154
+ Definition:
155
+
156
+ TSS = mean F1 score across all transfer evaluation tasks.
157
+
158
+ Interpretation:
159
+
160
+ High TSS indicates that the model has learned stability reasoning that generalizes across systems.
161
+
162
+ Low TSS suggests the model relies on dataset-specific patterns.
163
+
164
+ ---
165
+
166
+ # Benchmark Objective
167
+
168
+ The Clarus benchmark evaluates whether models can detect instability dynamics across complex systems.
169
+
170
+ The benchmark tests five core capabilities:
171
+
172
+ 1. pattern detection within individual datasets
173
+ 2. interaction reasoning across variables
174
+ 3. trajectory reasoning over time
175
+ 4. robustness to incomplete observation
176
+ 5. cross-domain stability reasoning
177
+
178
+ Models that succeed across all levels demonstrate the ability to reason about **latent stability geometry** rather than simple statistical correlations.