# Clarus Evaluation Framework The Clarus Stability Benchmark evaluates whether machine learning models can detect **latent instability dynamics** rather than relying on simple correlations. The evaluation framework measures model capability across multiple reasoning levels. These levels test progressively more difficult forms of stability reasoning. --- # Evaluation Levels ## Level 1 — Single Dataset Evaluation Models are trained and evaluated on the same dataset. Purpose: Measure whether the model can detect instability patterns within a single system. Procedure: 1. Train model on `train.csv` 2. Generate predictions for `test.csv` 3. Evaluate using the dataset scorer Metrics: - accuracy - precision - recall - f1 score - confusion matrix Limitations: High performance at this level does not necessarily indicate true stability reasoning. Models may rely on dataset-specific correlations. --- ## Level 2 — Within-Domain Transfer Models are trained on one dataset and evaluated on a different dataset within the same domain. Example: Train on: protein-folding-pathway-instability Test on: protein-aggregation-risk-instability Purpose: Evaluate whether models can generalize stability reasoning across related systems. Evaluation procedure: 1. Train on source dataset 2. Predict on target dataset 3. Score using target dataset scorer --- ## Level 3 — Cross-Domain Transfer Models are trained on one system domain and evaluated on another. Domains in the Clarus benchmark include: - clinical systems - molecular / protein systems - quantum systems Example transfer tasks: | Train Domain | Test Domain | |---|---| | clinical | clinical | | protein | protein | | quantum | quantum | | clinical | protein | | protein | quantum | | quantum | clinical | Purpose: Determine whether models learn general **stability geometry** rather than domain-specific patterns. --- # Robustness Evaluation The benchmark includes robustness tests that simulate real-world system conditions. --- ## Missing Data Evaluation Real systems often contain incomplete observations. Trajectory datasets may include variants where timepoints are missing. Variants include: - missing t0 - missing t1 - missing t2 - random missing Purpose: Evaluate whether models can infer stability dynamics when observations are incomplete. The prediction task remains unchanged. --- ## Class Imbalance Evaluation Many real-world systems exhibit rare failure events. Datasets may include variants with different class distributions. Supported imbalance regimes: - balanced (50 / 50) - mild imbalance (70 / 30) - severe imbalance (90 / 10) - extreme imbalance (99 / 1) Purpose: Evaluate whether models detect true instability rather than relying on class prevalence. Accuracy alone is insufficient under imbalance conditions. Recommended metrics: - precision - recall - F1 score --- # Transfer Stability Score To summarize transfer performance, the benchmark defines the **Transfer Stability Score (TSS)**. Definition: TSS = mean F1 score across all transfer evaluation tasks. Interpretation: High TSS indicates that the model has learned stability reasoning that generalizes across systems. Low TSS suggests the model relies on dataset-specific patterns. --- # Benchmark Objective The Clarus benchmark evaluates whether models can detect instability dynamics across complex systems. The benchmark tests five core capabilities: 1. pattern detection within individual datasets 2. interaction reasoning across variables 3. trajectory reasoning over time 4. robustness to incomplete observation 5. cross-domain stability reasoning Models that succeed across all levels demonstrate the ability to reason about **latent stability geometry** rather than simple statistical correlations.