NegBioDB / experiment_results.md
jang1563's picture
NegBioDB final: 4 domains, fully audited
6d1bbc7

NegBioDB Experiment Results

Last updated: 2026-03-24

Cross-Domain Summary

Domain Negatives ML Status LLM L1 (acc) LLM L4 (MCC)
DTI (Drug-Target Interaction) 30,459,583 24/24 complete Gemini 1.000 (3-shot) ≤ 0.18 (near random)
CT (Clinical Trial Failure) 132,925 108/108 complete Gemini 0.68 Gemini 0.56
PPI (Protein-Protein Interaction) 2,229,670 54/54 complete ~1.000 (3-shot artifact) Llama 0.441
GE (Gene Essentiality / DepMap) 28,759,256 Seed 42 complete 4/5 models complete Pending (Llama)

Key insight: LLM discrimination ability (L4 MCC) increases with task complexity: DTI (0.05–0.18) < PPI (0.33–0.44) < CT (~0.48–0.56). All domains show evidence of training data contamination in PPI/CT.


DTI Domain (Drug-Target Interaction)

Status as of 2026-03-24: ML 24/24 complete, LLM 81/81 complete.

Database

  • 30,459,583 negative results, 5 splits (random/cold_compound/cold_target/scaffold/temporal)
  • 1,725,446 export rows (862,723 pos + 862,723 neg)
  • Source tiers: gold=818,611 / silver=198 / bronze=28,845,632 (PubChem dynamic tiers)

ML Results Summary

  • 3 models: DeepDTA, GraphDTA, DrugBAN × 5 splits × 2 negative types (NegBioDB + uniform_random)
  • NegBioDB negatives: Random split near-perfect AUROC (trivially separable)
  • Control negatives (uniform_random): Harder splits show meaningful degradation

LLM Results (81/81 complete)

Task Best Model Metric Value
L1 (MCQ) Gemini Accuracy (3-shot) 1.000
L4 (Discrim) Best model MCC ≤ 0.18 (all near random)

Key finding: DTI LLMs show near-random discrimination (L4 MCC ≤ 0.18). The binding/non-binding decision is too nuanced for zero-shot or few-shot LLMs. This is the hardest domain.


CT Domain (Clinical Trial Failure)

Status as of 2026-03-20: ML 108/108 complete, LLM 80/80 complete, L3 judge complete.

Database

  • 132,925 failure results from 216,987 trials
  • Tiers: gold 23,570 / silver 28,505 / bronze 60,223 / copper 20,627
  • 8 failure categories: safety > efficacy > enrollment > strategic > regulatory > design > other (PK=0)

ML Results (CT-M1: Binary Failure Prediction)

Aggregated over 3 seeds (source: results/ct_table_m1_aggregated.csv):

Model Split Negatives AUROC MCC
XGBoost random NegBioDB 1.000 1.000
GNN random NegBioDB 1.000 1.000
MLP random NegBioDB 1.000 0.992
XGBoost random degree_matched 0.844 0.553
MLP random degree_matched 0.801 0.454
GNN random degree_matched 0.758 0.440
XGBoost cold_condition NegBioDB 1.000 1.000
XGBoost cold_drug NegBioDB 1.000 0.999

Key finding: CT-M1 is trivially separable on NegBioDB negatives (AUROC=1.0 random/cold splits). Control negatives (degree_matched) reveal meaningful discrimination AUROC ~0.76–0.84.

ML Results (CT-M2: 7-way Failure Category)

Aggregated over 3 seeds (source: results/ct_table_m2_aggregated.csv):

Model Split Macro-F1 Weighted-F1 MCC
XGBoost random 0.510 0.751 0.637
XGBoost degree_balanced 0.521 0.758 0.645
XGBoost cold_condition 0.338 0.686 0.570
XGBoost cold_drug 0.414 0.683 0.555
XGBoost scaffold 0.193 0.567 0.374
XGBoost temporal 0.193 0.602 0.454
GNN random 0.468 0.672 0.526
MLP random 0.358 0.619 0.432

Key finding: XGBoost best for M2. Scaffold/temporal splits are hardest (macro-F1 ~0.19). Degree-balanced helps.

LLM Results (80/80 complete)

Task Gemini GPT-4o-mini Haiku Qwen-7B Llama-8B
L1 (5-way MCQ, acc) 0.68 0.64 0.66 0.65 0.63
L2 (extraction, field_f1) 0.75 0.73 0.48 0.81 0.77
L3 (reasoning, /5.0) ~4.7 ~4.6 ~4.4 ~4.5 ~4.6
L4 (discrim, MCC) 0.56 0.49 0.50 0.48 0.50

Key finding: CT LLMs show meaningful discrimination (MCC ~0.5). L3 ceiling effect — all models score 4.4–4.7/5.0 (judge too lenient; scores not discriminative). L2 Qwen/Llama outperform API models on field F1.


PPI Domain (Protein-Protein Interaction)

Status as of 2026-03-23: ML 54/54 complete (seeds 42/43/44), LLM 80/80 complete, L3 judged.

Database

  • 2,229,670 negative results in DB (IntAct 779 / HuRI 500K / hu.MAP 1.23M / STRING 500K); 2,220,786 export rows after split filtering
  • 61,728 positive pairs (HuRI Y2H), 18,412 proteins

ML Results (PPI-M1: Binary Non-interaction Prediction)

Aggregated over 3 seeds (source: results/ppi/table1_aggregated.md):

Model Split Negatives AUROC MCC LogAUC
SiameseCNN random NegBioDB 0.963 ± 0.000 0.794 ± 0.012 0.517 ± 0.018
PIPR random NegBioDB 0.964 ± 0.001 0.812 ± 0.006 0.519 ± 0.009
MLPFeatures random NegBioDB 0.962 ± 0.001 0.788 ± 0.003 0.567 ± 0.005
SiameseCNN cold_protein NegBioDB 0.873 ± 0.002 0.568 ± 0.019 0.314 ± 0.014
PIPR cold_protein NegBioDB 0.859 ± 0.008 0.565 ± 0.019 0.288 ± 0.010
MLPFeatures cold_protein NegBioDB 0.931 ± 0.001 0.706 ± 0.005 0.476 ± 0.005
SiameseCNN cold_both NegBioDB 0.585 ± 0.040 0.070 ± 0.004 0.037 ± 0.010
PIPR cold_both NegBioDB 0.409 ± 0.077 −0.018 ± 0.044 0.031 ± 0.019
MLPFeatures cold_both NegBioDB 0.950 ± 0.021 0.749 ± 0.043 0.595 ± 0.051
SiameseCNN random uniform_random 0.965 ± 0.001 0.806 ± 0.007 0.552 ± 0.002
PIPR random uniform_random 0.966 ± 0.000 0.810 ± 0.002 0.565 ± 0.005

Key findings:

  • MLPFeatures dominates cold splits: AUROC 0.95 on cold_both (vs PIPR collapse to 0.41)
  • PIPR catastrophic failure on cold_both: AUROC 0.41 (below random!)
  • Control negatives (uniform_random, degree_matched) inflate LogAUC by +0.04–0.05

LLM Results (80/80 complete, post-hoc fixes applied 2026-03-23)

Task Gemini GPT-4o-mini Haiku Qwen-7B Llama-8B
L1 (4-way MCQ, acc) zero-shot 0.75 0.75 0.75 0.75 0.75
L1 (4-way MCQ, acc) 3-shot 1.000 0.997 0.998 0.998 0.997
L2 (method_accuracy) 3-shot 1.000 ~0.94 ~0.08 ~0.94 ~0.08
L3 (reasoning, /5.0) zero-shot 4.4–4.7 4.4–4.7 4.3–4.7 4.3–4.7 4.3–4.7
L3 (reasoning, /5.0) 3-shot 3.1–3.7 3.1–3.7 3.1–3.7 3.1–3.7 3.1–3.7
L4 (discrim, MCC) zero-shot 0.43 0.430 0.38 0.36 0.441

Key findings:

  • L1: 3-shot near-perfect is an artifact (in-context examples reveal pattern format)
  • L2: Gemini/GPT/Qwen correctly identify interaction methods, Haiku/Llama fail
  • L3: zero-shot > 3-shot (gold reasoning examples degrade performance); structural reasoning collapses to ~1.2/5 in 3-shot
  • L4: All models show contamination (acc_pre_2015 >> acc_post_2020)

GE Domain (Gene Essentiality / DepMap)

Status as of 2026-03-24: ETL+ML complete, LLM 4/5 models done (Llama pending), L3 judged.

Database

  • 28,759,256 negative results (CRISPR 19.7M + RNAi 9.1M)
  • Final tiers: Gold 753,878 / Silver 18,608,686 / Bronze 9,396,692
  • 22,549,910 aggregated pairs (19,554 genes × 2,132 cell lines)

ML Results (Seed 42 complete; seeds 43/44 in progress)

5 splits: random/cold_gene/cold_cell_line/cold_both/degree_balanced Models: XGBoost, MLPFeatures Aggregated results (3 seeds) pending; individual seed 42 results available in results/ge/.

LLM Results (4/5 models complete — Llama pending)

Task Models Done Key Finding
L1 (4-way MCQ, 1,200 items) Haiku, Gemini, GPT, Qwen Results pending Llama
L2 (field extraction, 500 items) Haiku, Gemini, GPT, Qwen Results pending Llama
L3 (reasoning, 200 items, judged) Haiku, Gemini, GPT, Qwen zero-shot >> 3-shot (4.5 vs 2.5 overall mean)
L4 (discrimination, 475 items) Haiku, Gemini, GPT, Qwen Results pending Llama

Expected key finding: GE L4 MCC likely intermediate between PPI and DTI given DepMap is a public widely-studied dataset with high training data overlap.


Methodology Notes

ML Evaluation Protocol

  • Metrics: AUROC (primary), LogAUC (early enrichment), AUPRC, MCC, BEDROC
  • Seeds: 3 seeds (42, 43, 44) for statistical robustness (except GE seed 42 only so far)
  • Negative types: NegBioDB (structured negatives) vs control negatives (uniform_random, degree_matched)
  • Splits: Random → Cold (one entity unseen) → Cold-Both (both entities unseen, hardest)

LLM Evaluation Protocol

  • Models: Claude Haiku-4.5, Gemini 2.5-Flash, GPT-4o-mini, Qwen2.5-7B-Instruct, Llama-3.1-8B-Instruct
  • Configs: Zero-shot × 1 + 3-shot × 3 fewshot sets = 4 configs per (model × task)
  • L3 judge: Gemini 2.5-Flash LLM-as-judge, 4 dimensions × 5-point scale
  • L4 contamination test: Older vs newer entity pairs (pre-cutoff vs post-cutoff)