| --- |
| license: mit |
| task_categories: |
| - question-answering |
| - text-generation |
| language: |
| - en |
| tags: |
| - science |
| - physics |
| - biology |
| - chemistry |
| - experimental-prediction |
| - benchmark |
| size_categories: |
| - n<1K |
| --- |
| |
| # SciPredict: Can LLMs Predict the Outcomes of Research Experiments? |
|
|
| **Paper:** SciPredict: Can LLMs Predict the Outcomes of Research Experiments in Natural Sciences? |
|
|
| ## Overview |
|
|
| SciPredict is a benchmark evaluating whether AI systems can predict experimental outcomes in physics, biology, and chemistry. The dataset comprises **405 questions** derived from recently published empirical studies (post-March 2025), spanning **33 subdomains**. |
|
|
| ## Dataset Structure |
|
|
| - **Total Questions:** 405 (5,716 rows including model responses) |
| - **Domains:** Physics (9 subdomains), Chemistry (10 subdomains), Biology (14 subdomains) |
| - **Question Formats:** Multiple-choice (MCQ), Free-format, Numerical |
|
|
| ### Key Fields |
|
|
| - `DOMAIN`: Scientific domain (Physics, Biology, Chemistry) |
| - `FIELD`: Specific field within the domain |
| - `PQ_FORMAT`: Question format (MCQ, Free-Format, Numerical) |
| - `TITLE`: Paper title |
| - `URL`: Paper URL |
| - `PUBLISHING_DATE`: Publication date |
| - `EXPERIMENTAL_SETUP`: Description of the experimental configuration |
| - `MEASUREMENT_TAKEN`: What was measured in the experiment |
| - `OUTCOME_PREDICTION_QUESTION`: The prediction task |
| - `GTA`: Ground truth answer |
| - `BACKGROUND_KNOWLEDGE`: Expert-curated background knowledge |
| - `RELATED_PAPERS_DATA`: Related papers information |
|
|
| ## Key Findings |
|
|
| - **Model accuracy:** 14-26% (vs. ~20% human expert accuracy) |
| - **Poor calibration:** Models cannot distinguish reliable from unreliable predictions |
| - **Background knowledge helps:** Providing expert-curated context improves performance |
| - **Format matters:** Performance degrades from MCQ → Free-form → Numerical |
|
|