Add PRIOR_ART_ANALYSIS.md — comprehensive landscape of similar systems with high-school-level explanations
Browse files- PRIOR_ART_ANALYSIS.md +562 -0
PRIOR_ART_ANALYSIS.md
ADDED
|
@@ -0,0 +1,562 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# PhD Research OS — Prior Art Analysis
|
| 2 |
+
## Every System That Has Tried Something Similar, and How We Compare
|
| 3 |
+
|
| 4 |
+
**Date**: 2026-04-23
|
| 5 |
+
**Status**: COMPREHENSIVE — 15 systems analyzed across 6 capability areas
|
| 6 |
+
**Written for**: Anyone with a high school education
|
| 7 |
+
**Purpose**: Honest accounting of who has built what, so we know exactly where we stand
|
| 8 |
+
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
## What Is This Document?
|
| 12 |
+
|
| 13 |
+
Before you build anything, you should check if someone already built it. That's not pessimism — it's good science. If someone already solved a problem, you don't need to solve it again. You can use their solution and spend your time on the parts nobody has cracked yet.
|
| 14 |
+
|
| 15 |
+
We searched through research papers, open-source code, commercial products, and HuggingFace repositories to find every system that does something similar to PhD Research OS. We found 15 significant systems. Here is what each one does, where it beats us, where we beat it, and what we should steal (with credit).
|
| 16 |
+
|
| 17 |
+
**The bottom line**: Nobody has built the complete system we're building. But every piece of our system exists somewhere else in isolation. The novel part is the combination — and three specific capabilities that nobody else has published.
|
| 18 |
+
|
| 19 |
+
---
|
| 20 |
+
|
| 21 |
+
## Table of Contents
|
| 22 |
+
|
| 23 |
+
1. [The Scoreboard — How Every System Compares](#1-the-scoreboard)
|
| 24 |
+
2. [Tier 1: The Closest Rivals](#2-tier-1-closest-rivals)
|
| 25 |
+
3. [Tier 2: Best-in-Class Components](#3-tier-2-best-in-class-components)
|
| 26 |
+
4. [Tier 3: Specialist Tools We Should Use](#4-tier-3-specialist-tools)
|
| 27 |
+
5. [Tier 4: Commercial Systems for Context](#5-tier-4-commercial-systems)
|
| 28 |
+
6. [What Nobody Has Built Yet (Our Innovations)](#6-what-nobody-has-built-yet)
|
| 29 |
+
7. [The Honest Comparison Table](#7-the-honest-comparison-table)
|
| 30 |
+
8. [Essential Resources and Links](#8-essential-resources)
|
| 31 |
+
|
| 32 |
+
---
|
| 33 |
+
|
| 34 |
+
## 1. The Scoreboard — How Every System Compares {#1-the-scoreboard}
|
| 35 |
+
|
| 36 |
+
Our system has 5 core capabilities. Here's who else has each one:
|
| 37 |
+
|
| 38 |
+
| Capability | What It Means | Who Has It | Who Does It Best |
|
| 39 |
+
|------------|--------------|------------|-----------------|
|
| 40 |
+
| **1. Extract claims from papers** | Read a PDF and pull out the key findings | PaperQA2, FactReview, SciEx, AgentSLR | PaperQA2 (superhuman on benchmarks) |
|
| 41 |
+
| **2. Label each claim epistemically** | Is this a Fact? An Interpretation? A Hypothesis? | KGX3 (paper-level only) | **Nobody does claim-level** ← Us |
|
| 42 |
+
| **3. Build a typed knowledge graph** | Connect claims with labeled edges (supports, refutes, extends) | Paper Circle, SciIE/SciERC, ORKG | Paper Circle (multi-agent KG) |
|
| 43 |
+
| **4. Detect contradictions across papers** | Find when Paper A disagrees with Paper B | PaperQA2, CLAIRE, FactReview | PaperQA2 (2.34× better than PhDs) |
|
| 44 |
+
| **5. Compute calibrated confidence scores** | Use math formulas (not AI guessing) to rate trustworthiness | **Nobody does formula-based** | **Nobody** ← Us |
|
| 45 |
+
|
| 46 |
+
**The key insight**: Lots of systems can do capabilities 1, 3, and 4. A few can do capability 2 at the paper level. **Nobody** combines all 5 at the claim level. And **nobody** does capability 5 the way we do (code-computed, not LLM-guessed).
|
| 47 |
+
|
| 48 |
+
---
|
| 49 |
+
|
| 50 |
+
## 2. Tier 1: The Closest Rivals {#2-tier-1-closest-rivals}
|
| 51 |
+
|
| 52 |
+
These are the systems that come closest to what we're building. If you only read one section, read this one.
|
| 53 |
+
|
| 54 |
+
---
|
| 55 |
+
|
| 56 |
+
### 🥇 PaperQA2 — The Gold Standard for Scientific Literature AI
|
| 57 |
+
|
| 58 |
+
**Paper**: [Language Agents Achieve Superhuman Synthesis of Scientific Knowledge](https://arxiv.org/abs/2409.13740) (2024)
|
| 59 |
+
**Earlier version**: [PaperQA: Retrieval-Augmented Generative Agent](https://arxiv.org/abs/2312.07559) (2023)
|
| 60 |
+
**Code**: [`github.com/Future-House/paper-qa`](https://github.com/Future-House/paper-qa) — open source, `pip install paper-qa`
|
| 61 |
+
**Built by**: FutureHouse (well-funded AI research lab)
|
| 62 |
+
|
| 63 |
+
#### What It Does (The Analogy)
|
| 64 |
+
|
| 65 |
+
Imagine you're writing a research paper and you have a brilliant research assistant. You ask them a question like "Does protein X bind to receptor Y?" They go to the library, pull out 50 relevant papers, read the important sections, highlight the key sentences, and come back with a summary that includes page numbers for everything they claim. If two papers disagree, they tell you about the contradiction.
|
| 66 |
+
|
| 67 |
+
That's PaperQA2. It's a multi-step AI agent that:
|
| 68 |
+
1. **Searches** for relevant papers (like a librarian)
|
| 69 |
+
2. **Retrieves** the most important chunks from each paper (like highlighting)
|
| 70 |
+
3. **Reranks** those chunks by relevance using AI (like sorting your highlights)
|
| 71 |
+
4. **Summarizes** each chunk in context (like writing margin notes)
|
| 72 |
+
5. **Generates** a final answer with citations (like writing a paragraph with footnotes)
|
| 73 |
+
6. **Detects contradictions** across papers (like spotting when two witnesses disagree)
|
| 74 |
+
|
| 75 |
+
#### How Good Is It?
|
| 76 |
+
|
| 77 |
+
This is the scary part. PaperQA2 has been **benchmarked against real PhD-level scientists**:
|
| 78 |
+
|
| 79 |
+
| Task | PaperQA2 Score | Human PhD Score | Winner |
|
| 80 |
+
|------|---------------|-----------------|--------|
|
| 81 |
+
| Answer questions about papers (LitQA2) | 85.2% precision | 73.8% precision | 🤖 PaperQA2 |
|
| 82 |
+
| Find contradictions in protein biology | 2.34× more found | Baseline | 🤖 PaperQA2 |
|
| 83 |
+
| Write Wikipedia-quality articles (WikiCrow) | 86.1% precision | 71.2% precision | 🤖 PaperQA2 |
|
| 84 |
+
|
| 85 |
+
**Analogy**: If paper-reading were a sport, PaperQA2 would be the Olympic champion. We're not competing directly — we're building a different sport (epistemic analysis), but we should learn from the champion's training methods.
|
| 86 |
+
|
| 87 |
+
#### What It Does Better Than Us
|
| 88 |
+
|
| 89 |
+
| Feature | PaperQA2 | PhD Research OS | What This Means |
|
| 90 |
+
|---------|----------|-----------------|-----------------|
|
| 91 |
+
| PDF parsing | GROBID (professional, battle-tested) | PyMuPDF (basic) | Their input quality is much higher |
|
| 92 |
+
| Retrieval quality | Dense retrieval + LLM reranking (RCS) | No retrieval system | They find the right sections faster |
|
| 93 |
+
| Contradiction detection | Benchmarked, superhuman | Designed but not built | They can prove it works; we can't yet |
|
| 94 |
+
| "I don't know" ability | Refuses when evidence is insufficient | No refusal mechanism | They're honest about gaps; we aren't yet |
|
| 95 |
+
| Citation traversal | Follows reference chains automatically | Citation resolution designed | They do it; we plan to do it |
|
| 96 |
+
|
| 97 |
+
#### What We Do Better Than PaperQA2
|
| 98 |
+
|
| 99 |
+
| Feature | PhD Research OS | PaperQA2 | What This Means |
|
| 100 |
+
|---------|-----------------|----------|-----------------|
|
| 101 |
+
| Epistemic labels | Fact / Interpretation / Hypothesis | None — flat claims | We tell you WHAT KIND of claim it is |
|
| 102 |
+
| Knowledge graph | Typed edges (supports/refutes/extends) | No persistent graph | We build a web of connected knowledge |
|
| 103 |
+
| Calibrated scoring | 3-score code-computed formula | LLM-stated confidence | Our confidence numbers are auditable |
|
| 104 |
+
| Local-first privacy | Paper data never leaves your computer | Requires cloud API calls | We protect unpublished research |
|
| 105 |
+
| Human oversight design | Council votes, proposals, manual mode | Answer generation only | We're built for verification, not just answers |
|
| 106 |
+
|
| 107 |
+
#### The Big Lesson from PaperQA2
|
| 108 |
+
|
| 109 |
+
Their secret weapon is **RCS (Rerank + Contextual Summarize)**. Before answering a question, they have the AI rerank all retrieved chunks by relevance, then write a contextual summary of each chunk. This two-step filter removes noise before the final answer is generated.
|
| 110 |
+
|
| 111 |
+
**Analogy**: Imagine you're studying for an exam. Method A: read all your notes and try to answer from memory. Method B: first sort your notes by relevance to the question, then write a one-sentence summary of each relevant note, then answer using only the summaries. Method B is dramatically better. That's RCS.
|
| 112 |
+
|
| 113 |
+
**We should adopt this.** Our Layer 2 extraction pipeline would benefit enormously from an RCS-style pre-filtering step before the AI Council processes each section.
|
| 114 |
+
|
| 115 |
+
---
|
| 116 |
+
|
| 117 |
+
### 🥈 FactReview — The Closest to Claim-Level Verification
|
| 118 |
+
|
| 119 |
+
**Paper**: [FactReview: Evidence-Grounded Reviews with Literature Positioning](https://arxiv.org/abs/2604.04074) (2026)
|
| 120 |
+
|
| 121 |
+
#### What It Does (The Analogy)
|
| 122 |
+
|
| 123 |
+
Imagine a very thorough peer reviewer. They don't just read your paper — they go find all the papers you cited, plus papers you DIDN'T cite, and check: "Does the existing literature actually support what this paper claims?" For every claim, they write: "SUPPORTS — three papers agree" or "REFUTES — this contradicts Smith 2023" or "RELATED — similar work, but not directly comparable."
|
| 124 |
+
|
| 125 |
+
And here's the clever part: if your paper says "our model achieves 95% accuracy," FactReview actually **runs your code** to check if that's true. It does what most human reviewers can't do — it verifies empirical claims by execution.
|
| 126 |
+
|
| 127 |
+
#### What It Does Better Than Us
|
| 128 |
+
|
| 129 |
+
- **Execution-based verification**: It can actually run code to check if claimed results are real. We have no equivalent. This is like a teacher not just reading your math homework but re-doing the calculations.
|
| 130 |
+
- **Literature positioning**: For each claim, it retrieves relevant prior work and labels the relationship (SUPPORTS / REFUTES / RELATED). This is close to our knowledge graph edges.
|
| 131 |
+
|
| 132 |
+
#### What We Do Better
|
| 133 |
+
|
| 134 |
+
- **Persistent knowledge graph**: FactReview checks one paper at a time. It doesn't build a lasting web of knowledge that grows as you add more papers.
|
| 135 |
+
- **Epistemic classification**: FactReview labels relationships between claims (supports/refutes), but doesn't label the claims themselves (Fact vs Hypothesis).
|
| 136 |
+
- **Calibrated scoring**: FactReview says "supports" or "refutes" — binary. We give a 3-dimensional confidence score.
|
| 137 |
+
|
| 138 |
+
#### The Big Lesson from FactReview
|
| 139 |
+
|
| 140 |
+
**Dual evidence is better than single evidence.** They check claims against BOTH the manuscript itself AND external literature. We should adopt this: when extracting a claim, check it against both (a) the specific paper and (b) the existing knowledge graph.
|
| 141 |
+
|
| 142 |
+
---
|
| 143 |
+
|
| 144 |
+
### 🥉 KGX3 / iKuhn — The Closest Epistemic System
|
| 145 |
+
|
| 146 |
+
**Paper**: [A Novel Kuhnian Ontology for Epistemic Classification](https://arxiv.org/abs/2002.03531) (2020)
|
| 147 |
+
**System**: Formerly called Soph.io, then iKuhn, now KGX3 / Preprint Watch
|
| 148 |
+
|
| 149 |
+
#### What It Does (The Analogy)
|
| 150 |
+
|
| 151 |
+
Imagine a very opinionated professor who reads every paper and immediately says: "This paper confirms what we already knew," or "This paper challenges the established theory," or "This paper introduces a completely new method." They don't just read the claims — they classify the paper's ROLE in the field's history.
|
| 152 |
+
|
| 153 |
+
KGX3 does this automatically using a philosophy framework from Thomas Kuhn (a famous historian of science). It classifies every paper using a three-part code:
|
| 154 |
+
- **M** (Methodological): Did they reuse an old method, adapt one, or invent a new one?
|
| 155 |
+
- **N** (Observational): Were the results expected, unexpected, or truly anomalous?
|
| 156 |
+
- **P** (Positional): Does this confirm, extend, or challenge existing knowledge?
|
| 157 |
+
|
| 158 |
+
So a paper might be classified as `(Novel, Unexpected, Challenges)` — a new method found something nobody predicted that contradicts the consensus. That's a paradigm-shifting paper.
|
| 159 |
+
|
| 160 |
+
#### Why This Matters for Us
|
| 161 |
+
|
| 162 |
+
KGX3 is the ONLY system we found that does **formal epistemic classification** — labeling the knowledge-status of scientific content using rules and formulas, not AI guessing. That's exactly our philosophy!
|
| 163 |
+
|
| 164 |
+
**But** there's a critical difference: KGX3 classifies **whole papers**. We classify **individual claims** within papers. This is like the difference between rating a whole restaurant (KGX3) versus rating each dish on the menu (us). Our approach is much more granular and much harder.
|
| 165 |
+
|
| 166 |
+
#### The Big Lesson from KGX3
|
| 167 |
+
|
| 168 |
+
Their **deterministic language-game filters** are brilliant. Instead of asking an AI "is this a confirmation or a challenge?", they have specific word patterns and rules that detect each epistemic category. For example, phrases like "consistent with previous findings" trigger the "Confirms" filter, while "contrary to expectations" triggers the "Challenges" filter.
|
| 169 |
+
|
| 170 |
+
**The activation threshold (θ=0.7)** is validated across multiple scientific disciplines. This is exactly the kind of empirically-grounded rule that our scoring formulas need.
|
| 171 |
+
|
| 172 |
+
**We should adopt the principle**: define specific linguistic triggers for each epistemic category (Fact / Interpretation / Hypothesis) and use them as code-based validators, not just AI prompts.
|
| 173 |
+
|
| 174 |
+
---
|
| 175 |
+
|
| 176 |
+
### Paper Circle — The Closest Multi-Agent KG System
|
| 177 |
+
|
| 178 |
+
**Paper**: [Paper Circle: An Open-source Multi-agent Research Discovery Framework](https://arxiv.org/abs/2604.06170) (2025)
|
| 179 |
+
**Code**: [`github.com/MAXNORM8650/papercircle`](https://github.com/MAXNORM8650/papercircle)
|
| 180 |
+
|
| 181 |
+
#### What It Does (The Analogy)
|
| 182 |
+
|
| 183 |
+
Imagine a team of research assistants, each with a different specialty:
|
| 184 |
+
- **Assistant 1** reads papers and identifies key concepts
|
| 185 |
+
- **Assistant 2** identifies the methods used
|
| 186 |
+
- **Assistant 3** extracts experimental results
|
| 187 |
+
- **Assistant 4** links figures and tables to the right concepts
|
| 188 |
+
- **A manager** coordinates all four and checks that nothing was missed
|
| 189 |
+
|
| 190 |
+
Together, they build a knowledge graph — a web of connected concepts, methods, experiments, and datasets.
|
| 191 |
+
|
| 192 |
+
Paper Circle does exactly this with AI agents. Their architecture looks like this:
|
| 193 |
+
|
| 194 |
+
```
|
| 195 |
+
CodeAgent (the manager)
|
| 196 |
+
→ Concept Extractor (finds ideas and theories)
|
| 197 |
+
→ Method Extractor (finds algorithms and techniques)
|
| 198 |
+
→ Experiment Extractor (finds setups, datasets, metrics, results)
|
| 199 |
+
→ Linkage Agent (connects figures/tables to concepts/methods)
|
| 200 |
+
→ Coverage Checker (makes sure nothing important was missed)
|
| 201 |
+
```
|
| 202 |
+
|
| 203 |
+
#### What We Should Learn
|
| 204 |
+
|
| 205 |
+
1. **The Coverage Checker is genius.** After all agents have done their work, a final agent checks: "Did we miss any important sections? Are there figures without linked concepts? Are there methods mentioned in the text that we didn't extract?" This prevents the silent omission problem — where the system looks complete but actually missed something important.
|
| 206 |
+
|
| 207 |
+
2. **Their provenance model is solid.** Every node in their knowledge graph carries: source chunk IDs, page numbers, verification status, confidence scores, and timestamps. This is very close to our Layer 7 design.
|
| 208 |
+
|
| 209 |
+
#### Where We Differ
|
| 210 |
+
|
| 211 |
+
Paper Circle builds a **structural** knowledge graph (concepts, methods, experiments connected by "used-in" and "part-of" edges). We build an **epistemic** knowledge graph (claims connected by "supports," "refutes," and "extends" edges). Think of it this way: Paper Circle maps the **territory** of science (what exists). We map the **arguments** of science (what agrees, what disagrees, what's uncertain).
|
| 212 |
+
|
| 213 |
+
---
|
| 214 |
+
|
| 215 |
+
### AgentSLR — The Best End-to-End Literature Review System
|
| 216 |
+
|
| 217 |
+
**Paper**: [AgentSLR: Automating Systematic Literature Reviews with Agentic AI](https://arxiv.org/abs/2603.22327) (2025)
|
| 218 |
+
**Code**: [`github.com/oxrml/agentslr`](https://github.com/oxrml/agentslr)
|
| 219 |
+
|
| 220 |
+
#### What It Does (The Analogy)
|
| 221 |
+
|
| 222 |
+
Imagine you're a doctor who needs to know: "Does drug X help with disease Y?" To answer properly, you need to do a **systematic review** — find EVERY study about this question, check which ones are good quality, extract the key numbers from each one, and synthesize a conclusion.
|
| 223 |
+
|
| 224 |
+
This normally takes a team of researchers 7 weeks. AgentSLR does it in 20 hours. And it performs at human level.
|
| 225 |
+
|
| 226 |
+
#### The Numbers
|
| 227 |
+
|
| 228 |
+
- **58× faster** than human systematic reviews
|
| 229 |
+
- Applied to 9 WHO-priority pathogens (real-world medical task)
|
| 230 |
+
- Open-source with human-in-the-loop hooks
|
| 231 |
+
|
| 232 |
+
#### What We Should Learn
|
| 233 |
+
|
| 234 |
+
AgentSLR is the proof that AI can process scientific literature at scale AND match human quality. But it doesn't build knowledge graphs, doesn't classify epistemic status, and doesn't detect contradictions. It's a **pipeline** (find → screen → extract → report), not an **engine** (continuously building and updating a knowledge web).
|
| 235 |
+
|
| 236 |
+
**Our role is different**: AgentSLR answers one question at a time. We build a growing knowledge base that answers every future question about the papers you've already processed. Think of AgentSLR as a very fast research assistant. We're building a very smart filing cabinet.
|
| 237 |
+
|
| 238 |
+
---
|
| 239 |
+
|
| 240 |
+
### CLAIRE (Stanford) — The Best Contradiction Detector
|
| 241 |
+
|
| 242 |
+
**Paper**: [Detecting Corpus-Level Knowledge Inconsistencies with LLMs](https://arxiv.org/abs/2509.23233) (2024)
|
| 243 |
+
**Code**: [`github.com/stanford-oval/inconsistency-detection`](https://github.com/stanford-oval/inconsistency-detection)
|
| 244 |
+
|
| 245 |
+
#### What It Does (The Analogy)
|
| 246 |
+
|
| 247 |
+
Wikipedia has millions of articles. Some of them contradict each other — Article A says one thing, Article B says the opposite. Nobody has time to read all of Wikipedia looking for contradictions. CLAIRE does.
|
| 248 |
+
|
| 249 |
+
It works like a detective:
|
| 250 |
+
1. **Extract claims** from one article
|
| 251 |
+
2. **Search** the rest of the corpus for related content
|
| 252 |
+
3. **Compare** using AI reasoning — do these agree or disagree?
|
| 253 |
+
4. **Flag** contradictions for human editors to review
|
| 254 |
+
|
| 255 |
+
When tested with real Wikipedia editors, the editors said they were **more confident** and found **more issues** when using CLAIRE than working alone.
|
| 256 |
+
|
| 257 |
+
#### What We Should Learn
|
| 258 |
+
|
| 259 |
+
CLAIRE's **agentic RAG loop** (extract → retrieve → compare → flag) maps directly to our conflict detection pipeline. The key insight is that you should retrieve evidence BEFORE making a contradiction judgment — don't just compare two claims in isolation, but also look at what other sources say about both claims.
|
| 260 |
+
|
| 261 |
+
**Their domain is Wikipedia, not science.** Scientific contradictions are harder because methods matter — two papers might look contradictory but actually used different experimental conditions. We handle this with our method-compatibility layer. CLAIRE doesn't have this.
|
| 262 |
+
|
| 263 |
+
---
|
| 264 |
+
|
| 265 |
+
## 3. Tier 2: Best-in-Class Components {#3-tier-2-best-in-class-components}
|
| 266 |
+
|
| 267 |
+
These aren't competing systems — they're the best tools for specific sub-problems. We should use them directly.
|
| 268 |
+
|
| 269 |
+
---
|
| 270 |
+
|
| 271 |
+
### SciFact + MultiVerS — The Standard for Claim Verification
|
| 272 |
+
|
| 273 |
+
**Paper (SciFact)**: [Fact or Fiction: Verifying Scientific Claims](https://arxiv.org/abs/2004.14974) (2020)
|
| 274 |
+
**Paper (MultiVerS)**: [MultiVerS: Improving scientific claim verification](https://arxiv.org/abs/2112.01640) (2021)
|
| 275 |
+
**Dataset**: [`bigbio/scifact`](https://huggingface.co/datasets/bigbio/scifact) on HuggingFace
|
| 276 |
+
**Code**: [`github.com/allenai/multivers`](https://github.com/allenai/multivers)
|
| 277 |
+
**Built by**: AllenAI (the Allen Institute for AI — one of the best AI research labs in the world)
|
| 278 |
+
|
| 279 |
+
#### What It Is (The Analogy)
|
| 280 |
+
|
| 281 |
+
SciFact is like a standardized test for claim-checking AI. It contains 1,400 scientific claims written by experts, and for each claim, the "answer key" tells you which research abstracts SUPPORT it, REFUTE it, or provide NO INFORMATION.
|
| 282 |
+
|
| 283 |
+
MultiVerS is the best model for taking this test. It reads a claim AND a full research abstract at the same time, then outputs: (a) does this abstract support, refute, or have nothing to say about this claim? and (b) which specific sentences in the abstract are the evidence?
|
| 284 |
+
|
| 285 |
+
#### Why This Matters for Us
|
| 286 |
+
|
| 287 |
+
SciFact defines the SUPPORTS / REFUTES / NOINFO label scheme that our knowledge graph edges are based on. It's the industry standard. Our system should be **benchmarkable** against SciFact — if we can't match MultiVerS scores on this standard test, our extraction pipeline isn't good enough.
|
| 288 |
+
|
| 289 |
+
**How to use it**: [`bigbio/scifact`](https://huggingface.co/datasets/bigbio/scifact) is available on HuggingFace Hub. Each example has a claim text, evidence documents, labels (SUPPORTS/REFUTES/NOINFO), and rationale sentences.
|
| 290 |
+
|
| 291 |
+
---
|
| 292 |
+
|
| 293 |
+
### SciRIFF — The Best Training Data for Scientific Tasks
|
| 294 |
+
|
| 295 |
+
**Paper**: [SciRIFF: Enhancing LM Instruction-Following for Scientific Literature](https://arxiv.org/abs/2406.07835) (2024)
|
| 296 |
+
**Dataset**: [`allenai/SciRIFF`](https://huggingface.co/datasets/allenai/SciRIFF) on HuggingFace
|
| 297 |
+
**Built by**: AllenAI
|
| 298 |
+
|
| 299 |
+
#### What It Is (The Analogy)
|
| 300 |
+
|
| 301 |
+
Imagine you want to teach someone to be a research assistant. SciRIFF is the textbook. It contains **137,000 practice exercises** across **54 different scientific tasks**: extracting information, verifying claims, summarizing papers, answering questions, classifying text.
|
| 302 |
+
|
| 303 |
+
Each exercise was written by **human experts** — not generated by AI. This matters because AI-generated training data teaches the model AI-style shortcuts, not real scientific reasoning.
|
| 304 |
+
|
| 305 |
+
#### Why This Matters for Us
|
| 306 |
+
|
| 307 |
+
Our current training data is 1,900 synthetic examples generated by a Python script. SciRIFF has 137,000 expert-written examples. That's 72× more data, and it's real, not fake.
|
| 308 |
+
|
| 309 |
+
**Key finding from the SciRIFF paper**: Expert-written templates perform significantly better than GPT-4-generated templates. This validates our concern about synthetic training data — real human-written examples are genuinely better.
|
| 310 |
+
|
| 311 |
+
**Training recipe from the paper**: 5 epochs, learning rate 2e-5, BF16, batch size 128, sequence length 4096. 7B models reach ~70% on scientific tasks; you need 70B for 77%+.
|
| 312 |
+
|
| 313 |
+
---
|
| 314 |
+
|
| 315 |
+
### SPECTER2 — The Best Scientific Paper Embeddings
|
| 316 |
+
|
| 317 |
+
**Models**: [`allenai/specter2_base`](https://huggingface.co/allenai/specter2_base) + task-specific adapters on HuggingFace
|
| 318 |
+
**Built by**: AllenAI
|
| 319 |
+
|
| 320 |
+
#### What It Is (The Analogy)
|
| 321 |
+
|
| 322 |
+
When you hear a song, your brain instantly converts it into a "feeling" — happy, sad, energetic. SPECTER2 does the same thing for scientific text. It converts any scientific sentence, abstract, or paper into a list of numbers (called an "embedding") that captures its meaning.
|
| 323 |
+
|
| 324 |
+
Two sentences that mean the same thing — even if they use completely different words — will get similar numbers. "The detection limit was 0.8 fM" and "We achieved femtomolar-level sensitivity" would have very similar embeddings.
|
| 325 |
+
|
| 326 |
+
#### Why This Matters for Us
|
| 327 |
+
|
| 328 |
+
Our deduplication system (Layer 3) currently uses word overlap, which misses paraphrases completely. SPECTER2 would fix this instantly. It's specifically trained on scientific text, so it understands domain-specific terminology.
|
| 329 |
+
|
| 330 |
+
**Available adapters**:
|
| 331 |
+
| Adapter | HuggingFace ID | Use Case |
|
| 332 |
+
|---------|---------------|----------|
|
| 333 |
+
| Proximity/Retrieval | `allenai/specter2` | Finding similar claims (deduplication) |
|
| 334 |
+
| Classification | `allenai/specter2_classification` | Topic/field classification |
|
| 335 |
+
| Ad-hoc Search | `allenai/specter2_adhoc_query` | Searching claims by query |
|
| 336 |
+
|
| 337 |
+
---
|
| 338 |
+
|
| 339 |
+
### SciIE / SciERC — The Standard for Scientific KG Construction
|
| 340 |
+
|
| 341 |
+
**Paper**: [Multi-Task IE for Scientific Knowledge Graph Construction](https://arxiv.org/abs/1808.09602) (2018)
|
| 342 |
+
**Dataset**: SciERC — 500 annotated abstracts with entities, relations, and coreference
|
| 343 |
+
**Built by**: AllenAI
|
| 344 |
+
|
| 345 |
+
#### What It Is (The Analogy)
|
| 346 |
+
|
| 347 |
+
SciERC is like a labeled map of how scientific concepts connect to each other. In 500 research paper abstracts, human experts marked:
|
| 348 |
+
- **Entities**: What things are mentioned? (Tasks, Methods, Metrics, Materials)
|
| 349 |
+
- **Relations**: How do they connect? (used-for, feature-of, part-of, compare)
|
| 350 |
+
- **Coreference**: When do two mentions refer to the same thing?
|
| 351 |
+
|
| 352 |
+
SciIE is the AI model that learned from this map and can now automatically build knowledge graphs from new papers.
|
| 353 |
+
|
| 354 |
+
#### Why This Matters for Us
|
| 355 |
+
|
| 356 |
+
The SciERC relation types (used-for, feature-of, part-of) are **structural** — they describe how things relate physically or functionally. Our edge types (supports, refutes, extends) are **epistemic** — they describe how claims relate evidentially.
|
| 357 |
+
|
| 358 |
+
We need both. A complete knowledge graph should know that "Method A was used-for Task B" (structural) AND "Claim X from Paper 1 supports Claim Y from Paper 2" (epistemic). We should use SciERC's taxonomy as the starting point for our structural edges, alongside our existing epistemic edges.
|
| 359 |
+
|
| 360 |
+
---
|
| 361 |
+
|
| 362 |
+
### SciBERT & BiomedBERT — Domain-Specific Language Models
|
| 363 |
+
|
| 364 |
+
| Model | HuggingFace ID | What It's Good For |
|
| 365 |
+
|-------|---------------|-------------------|
|
| 366 |
+
| SciBERT | [`allenai/scibert_scivocab_uncased`](https://huggingface.co/allenai/scibert_scivocab_uncased) | General scientific NLP (NER, classification) |
|
| 367 |
+
| BiomedBERT | [`microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract-fulltext`](https://huggingface.co/microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract-fulltext) | Biomedical text (best on BLURB benchmark) |
|
| 368 |
+
| SciBERT-NLI | [`gsarti/scibert-nli`](https://huggingface.co/gsarti/scibert-nli) | Scientific entailment/contradiction detection |
|
| 369 |
+
|
| 370 |
+
#### Why These Matter for Us
|
| 371 |
+
|
| 372 |
+
These are smaller, faster models trained specifically on scientific text. They won't replace our main Qwen brain, but they can serve as **lightweight validators**. For example:
|
| 373 |
+
- Use SciBERT-NLI as a fast pre-filter for contradiction detection (check if two claims even MIGHT contradict before sending to the expensive main model)
|
| 374 |
+
- Use SciBERT for scientific NER to help with entity resolution (Layer 1)
|
| 375 |
+
- Use SPECTER2 + SciBERT together for claim embedding (Layer 3)
|
| 376 |
+
|
| 377 |
+
---
|
| 378 |
+
|
| 379 |
+
## 4. Tier 3: Specialist Tools We Should Use {#4-tier-3-specialist-tools}
|
| 380 |
+
|
| 381 |
+
### PDF Parsing
|
| 382 |
+
|
| 383 |
+
| Tool | What It Does | Analogy |
|
| 384 |
+
|------|-------------|---------|
|
| 385 |
+
| **Nougat** ([`facebook/nougat-base`](https://huggingface.co/facebook/nougat-base)) | Converts scientific PDFs to markdown with proper LaTeX equations | A translator who speaks both "PDF" and "structured text" and is especially good at math |
|
| 386 |
+
| **GROBID** ([`github.com/kermitt2/grobid`](https://github.com/kermitt2/grobid)) | Extracts authors, citations, references, and paper structure | A librarian who catalogs every detail about a book |
|
| 387 |
+
| **Marker** ([`github.com/VikParuchuri/marker`](https://github.com/VikParuchuri/marker)) | ML-based PDF → markdown with layout awareness | A reader who understands that columns, captions, and footnotes are different things |
|
| 388 |
+
|
| 389 |
+
**Our plan already includes all three.** The SYSTEM_DESIGN.md correctly specifies Marker + Nougat + GROBID. What we don't have yet is the actual integration code.
|
| 390 |
+
|
| 391 |
+
### Scientific NER (Named Entity Recognition)
|
| 392 |
+
|
| 393 |
+
| Tool | Paper | What It Does |
|
| 394 |
+
|------|-------|-------------|
|
| 395 |
+
| **GLiNER-biomed** | [arxiv:2504.00676](https://arxiv.org/abs/2504.00676) | Zero-shot biomedical entity recognition — can find ANY type of entity without being specifically trained for it |
|
| 396 |
+
| **SciER** | [arxiv:2410.21155](https://arxiv.org/abs/2410.21155) | Extracts datasets, methods, and tasks from full-text papers — directly feeds into knowledge graph nodes |
|
| 397 |
+
|
| 398 |
+
### Fact-Checking Systems
|
| 399 |
+
|
| 400 |
+
| System | Paper | Key Innovation |
|
| 401 |
+
|--------|-------|---------------|
|
| 402 |
+
| **ClaimVer** | [arxiv:2403.09724](https://arxiv.org/abs/2403.09724) | KG-backed claim verification with natural language explanations — generates human-readable reasoning for why a claim is true/false |
|
| 403 |
+
| **CLUE** | [arxiv:2505.17855](https://arxiv.org/abs/2505.17855) | Explains WHERE uncertainty comes from — inter-evidence conflicts vs. agreement. This is exactly what our 3-score system tries to capture |
|
| 404 |
+
| **PubHealth** | [arxiv:2010.09926](https://arxiv.org/abs/2010.09926) | Explainable fact-checking for health claims — proof that domain-specific fact-checking outperforms general-purpose |
|
| 405 |
+
|
| 406 |
+
### Confidence Calibration
|
| 407 |
+
|
| 408 |
+
| System | Paper | Key Innovation |
|
| 409 |
+
|--------|-------|---------------|
|
| 410 |
+
| **CritiCal** | [arxiv:2510.24505](https://arxiv.org/abs/2510.24505) | Natural language critiques improve LLM calibration — the AI writes WHY it's uncertain, and this improves the accuracy of its confidence scores |
|
| 411 |
+
| **MetaFaith** | [arxiv:2505.24858](https://arxiv.org/abs/2505.24858) | Prompt-based faithful uncertainty — teaches the model to express uncertainty using hedging words that actually correspond to its real confidence |
|
| 412 |
+
| **Calibration-Tuning** | [arxiv:2406.08391](https://arxiv.org/abs/2406.08391) | Fine-tune on correct/incorrect answer pairs → model learns when it's likely to be wrong |
|
| 413 |
+
|
| 414 |
+
### Systematic Review Tools
|
| 415 |
+
|
| 416 |
+
| Tool | Type | Best For |
|
| 417 |
+
|------|------|---------|
|
| 418 |
+
| **ASReview** ([`github.com/asreview/asreview`](https://github.com/asreview/asreview)) | Open-source | Active learning for paper screening — learns your preferences to surface the most relevant papers first |
|
| 419 |
+
| **LatteReview** ([arxiv:2501.05468](https://arxiv.org/abs/2501.05468)) | Open-source | Multi-agent systematic review framework with Pydantic validation |
|
| 420 |
+
| **Bio-SIEVE** ([arxiv:2308.06610](https://arxiv.org/abs/2308.06610)) | Open-source | Instruction-tuned LLaMA for abstract screening |
|
| 421 |
+
|
| 422 |
+
---
|
| 423 |
+
|
| 424 |
+
## 5. Tier 4: Commercial Systems for Context {#5-tier-4-commercial-systems}
|
| 425 |
+
|
| 426 |
+
These are commercial products that researchers actually use today. We're not competing with them directly — they answer questions, we build knowledge maps — but we should understand what researchers are already used to.
|
| 427 |
+
|
| 428 |
+
| System | What It Does | Strengths | What It Can't Do (That We Can) |
|
| 429 |
+
|--------|-------------|-----------|-------------------------------|
|
| 430 |
+
| **Semantic Scholar** ([semanticscholar.org](https://www.semanticscholar.org/)) | Citation graph, paper search, TLDR summaries, SPECTER embeddings | Massive database (200M+ papers), free API, open data | No claim extraction, no epistemic labels, no contradiction detection |
|
| 431 |
+
| **Elicit** ([elicit.com](https://elicit.com)) | AI research assistant — finds papers, extracts data, answers questions | Clean UI, popular with researchers | Benchmarked as inferior to PaperQA2; no knowledge graph, no epistemic classification |
|
| 432 |
+
| **Consensus** ([consensus.app](https://consensus.app)) | "Consensus meter" showing if papers agree | Good at showing agreement/disagreement at topic level | No claim-level analysis, no confidence decomposition, no public API |
|
| 433 |
+
| **ResearchRabbit** ([researchrabbit.ai](https://www.researchrabbit.ai/)) | Citation network visualization | Beautiful graph visualization, free | No claim analysis, no extraction, just citation mapping |
|
| 434 |
+
| **Litmaps** ([litmaps.com](https://www.litmaps.com/)) | Citation mapping with timeline | Good discovery tool | No analysis beyond citation connections |
|
| 435 |
+
| **ORKG** ([orkg.org](https://orkg.org)) | Open Research Knowledge Graph — human-contributed structured paper comparisons | Structured comparison tables created by researchers | Manual input only — no automated extraction |
|
| 436 |
+
| **SciScore** ([sciscore.com](https://www.sciscore.com/)) | Methods rigor assessment — checks if papers report blinding, randomization, statistics properly | Addresses an important research quality dimension | Very narrow focus — just methods reporting, not claims or knowledge |
|
| 437 |
+
| **Iris.ai** ([iris.ai](https://iris.ai)) | Evidence mapping, relevance scoring | Nice relevance visualizations | No epistemic classification, no public API |
|
| 438 |
+
|
| 439 |
+
---
|
| 440 |
+
|
| 441 |
+
## 6. What Nobody Has Built Yet (Our Innovations) {#6-what-nobody-has-built-yet}
|
| 442 |
+
|
| 443 |
+
After analyzing all 15 systems, here are the things that **exist nowhere else**:
|
| 444 |
+
|
| 445 |
+
### Innovation 1: Claim-Level Epistemic Classification
|
| 446 |
+
|
| 447 |
+
**What it is**: Labeling EACH CLAIM (not each paper) as Fact, Interpretation, or Hypothesis.
|
| 448 |
+
|
| 449 |
+
**Who comes close**: KGX3 classifies whole papers using a Kuhnian triplet (M, N, P). But a single paper contains dozens of claims — some are facts (direct measurements), some are interpretations (the authors' explanation of the measurements), and some are hypotheses (what the authors think might be true). KGX3 can't distinguish these within a single paper.
|
| 450 |
+
|
| 451 |
+
**Analogy**: KGX3 rates a whole restaurant (4 stars). We rate each dish on the menu (appetizer: 5 stars, soup: 3 stars, dessert: 4 stars). Both are useful, but ours gives you much more detail.
|
| 452 |
+
|
| 453 |
+
**Why nobody has done this**: It's HARD. The same sentence can be a Fact in one context and an Interpretation in another. "The treatment worked" is a Fact in the Results section (because it's backed by data) and an Interpretation in the Abstract (because it might overstate what the data actually shows). You need section-awareness, qualifier detection, statistical evidence extraction, and a rules-based engine to do this correctly.
|
| 454 |
+
|
| 455 |
+
### Innovation 2: Code-Computed Calibrated Confidence
|
| 456 |
+
|
| 457 |
+
**What it is**: Using math formulas in code (not AI guessing) to compute how much you should trust each claim. Our 3-score system (evidence quality × truth likelihood × qualifier strength) is computed by Python code, not by asking the AI "how confident are you?"
|
| 458 |
+
|
| 459 |
+
**Who comes close**: SAFE (Google's factuality system) checks claims against search results, but its scores are binary (supported/not supported). CritiCal and MetaFaith improve LLM calibration, but they still rely on the LLM's self-assessed confidence. KGX3 uses deterministic rules but only at paper level.
|
| 460 |
+
|
| 461 |
+
**Analogy**: Other systems ask a student "how well do you think you did on the test?" and trust their answer. We give the test to 3 different graders, multiply their scores using a fixed formula, and add penalty points for unclear handwriting. The student never touches the grade book.
|
| 462 |
+
|
| 463 |
+
**Why nobody has done this**: Most AI systems want to be end-to-end — input goes in, output comes out, no human-designed formulas in between. Our approach deliberately breaks this by inserting code-computed gates between AI outputs and final confidence numbers. This requires more engineering but produces auditable, reproducible scores.
|
| 464 |
+
|
| 465 |
+
### Innovation 3: The Integrated 7-Layer Local-First Pipeline
|
| 466 |
+
|
| 467 |
+
**What it is**: A complete system that does parse → resolve → extract → dedup → graph → score → evaluate → export, all running locally on your laptop without sending data to the cloud.
|
| 468 |
+
|
| 469 |
+
**Who comes close**: PaperQA2 does search → retrieve → summarize → answer, but requires cloud APIs and doesn't build a persistent knowledge graph. Paper Circle builds a knowledge graph but doesn't score claims or detect contradictions. AgentSLR does systematic reviews end-to-end but doesn't build any persistent knowledge structure.
|
| 470 |
+
|
| 471 |
+
**Analogy**: Other systems are like food delivery apps — each one brings you a different dish. PaperQA2 brings you answers. SciFact brings you verification. SPECTER2 brings you similarity scores. We're building a kitchen that can cook all these dishes and remember every recipe forever.
|
| 472 |
+
|
| 473 |
+
---
|
| 474 |
+
|
| 475 |
+
## 7. The Honest Comparison Table {#7-the-honest-comparison-table}
|
| 476 |
+
|
| 477 |
+
Here's the full truth. Green means better, yellow means comparable, red means worse.
|
| 478 |
+
|
| 479 |
+
| Feature | PhD Research OS | PaperQA2 | FactReview | KGX3 | Paper Circle | AgentSLR |
|
| 480 |
+
|---------|:-:|:-:|:-:|:-:|:-:|:-:|
|
| 481 |
+
| PDF parsing quality | 🔴 Basic | 🟢 GROBID | 🟡 Standard | N/A | 🟡 Standard | 🟡 Standard |
|
| 482 |
+
| Claim extraction | 🟡 Designed | 🟢 Proven superhuman | 🟢 Literature-grounded | 🔴 Paper-level only | 🟡 Concept-level | 🟡 Data extraction |
|
| 483 |
+
| Epistemic classification | 🟢 **Unique: claim-level** | 🔴 None | 🟡 Binary | 🟢 Paper-level | 🔴 None | 🔴 None |
|
| 484 |
+
| Knowledge graph | 🟢 **Typed epistemic edges** | 🔴 None | 🔴 None | 🔴 None | 🟢 Structural KG | 🔴 None |
|
| 485 |
+
| Contradiction detection | 🟡 Designed | 🟢 **Benchmarked superhuman** | 🟢 Literature-grounded | 🟡 Paradigm drift | 🔴 None | 🔴 None |
|
| 486 |
+
| Confidence scoring | 🟢 **Formula-based, 3-score** | 🔴 LLM-stated | 🔴 Binary | 🟡 Deterministic rules | 🟡 Basic confidence | 🔴 None |
|
| 487 |
+
| Human oversight | 🟢 **Council + proposals** | 🔴 None | 🟡 Reviewer-facing | 🔴 None | 🔴 None | 🟡 Human-in-loop |
|
| 488 |
+
| Privacy (local-first) | 🟢 **All local** | 🔴 Requires cloud | 🔴 Requires cloud | 🟡 Unknown | 🟡 Partial | 🔴 Requires cloud |
|
| 489 |
+
| Actually working today? | 🔴 **Prototype** | 🟢 **Production** | 🟡 Research | 🟡 Proprietary | 🟡 Open-source | 🟢 Open-source |
|
| 490 |
+
| Benchmarked against humans? | 🔴 **No** | 🟢 **Yes, superhuman** | 🟡 Partial | 🟡 Yes | 🔴 No | 🟢 Yes |
|
| 491 |
+
|
| 492 |
+
**The honest reading**: We have the best design for epistemic analysis and human oversight. PaperQA2 has the best working system. Our job is to close that gap — build what we've designed, using the best components from everyone else.
|
| 493 |
+
|
| 494 |
+
---
|
| 495 |
+
|
| 496 |
+
## 8. Essential Resources and Links {#8-essential-resources}
|
| 497 |
+
|
| 498 |
+
### Systems to Study (Code Available)
|
| 499 |
+
|
| 500 |
+
| System | Install / Repository | Priority |
|
| 501 |
+
|--------|---------------------|----------|
|
| 502 |
+
| PaperQA2 | `pip install paper-qa` · [`github.com/Future-House/paper-qa`](https://github.com/Future-House/paper-qa) | 🔴 Study immediately |
|
| 503 |
+
| Paper Circle | [`github.com/MAXNORM8650/papercircle`](https://github.com/MAXNORM8650/papercircle) | 🟠 Study for KG pattern |
|
| 504 |
+
| AgentSLR | [`github.com/oxrml/agentslr`](https://github.com/oxrml/agentslr) | 🟡 Study for pipeline pattern |
|
| 505 |
+
| CLAIRE | [`github.com/stanford-oval/inconsistency-detection`](https://github.com/stanford-oval/inconsistency-detection) | 🟡 Study for contradiction detection |
|
| 506 |
+
| ASReview | [`github.com/asreview/asreview`](https://github.com/asreview/asreview) | 🟡 Study for screening UX |
|
| 507 |
+
|
| 508 |
+
### Datasets to Use (Available on HuggingFace)
|
| 509 |
+
|
| 510 |
+
| Dataset | HuggingFace ID | What It's For | Priority |
|
| 511 |
+
|---------|---------------|---------------|----------|
|
| 512 |
+
| SciFact | [`bigbio/scifact`](https://huggingface.co/datasets/bigbio/scifact) | Benchmark for claim verification | 🔴 Use as evaluation benchmark |
|
| 513 |
+
| SciRIFF | [`allenai/SciRIFF`](https://huggingface.co/datasets/allenai/SciRIFF) | 137K expert training examples | 🔴 Use for training data |
|
| 514 |
+
| Evidence Inference | [`bigbio/evidence_inference`](https://huggingface.co/datasets/bigbio/evidence_inference) | RCT outcome extraction | 🟠 Use for clinical paper training |
|
| 515 |
+
| QASPER | [`allenai/qasper`](https://huggingface.co/datasets/allenai/qasper) | Full-text scientific QA | 🟡 Use for extraction training |
|
| 516 |
+
| SciRepEval | [`allenai/scirepeval`](https://huggingface.co/datasets/allenai/scirepeval) | 6M+ paper embedding triplets | 🟡 Use for embedding fine-tuning |
|
| 517 |
+
| CiteWorth | [`copenlu/citeworth`](https://huggingface.co/datasets/copenlu/citeworth) | Citation-worthy sentence detection | 🟡 Use for claim identification |
|
| 518 |
+
|
| 519 |
+
### Models to Use (Available on HuggingFace)
|
| 520 |
+
|
| 521 |
+
| Model | HuggingFace ID | What It's For | Priority |
|
| 522 |
+
|-------|---------------|---------------|----------|
|
| 523 |
+
| SPECTER2 | [`allenai/specter2_base`](https://huggingface.co/allenai/specter2_base) + adapters | Scientific text embeddings (Layer 3 dedup) | 🔴 Integrate immediately |
|
| 524 |
+
| Nougat | [`facebook/nougat-base`](https://huggingface.co/facebook/nougat-base) | PDF → markdown with LaTeX (Layer 0) | 🔴 Integrate immediately |
|
| 525 |
+
| SciBERT | [`allenai/scibert_scivocab_uncased`](https://huggingface.co/allenai/scibert_scivocab_uncased) | Scientific NER and classification | 🟠 Use for entity resolution |
|
| 526 |
+
| SciBERT-NLI | [`gsarti/scibert-nli`](https://huggingface.co/gsarti/scibert-nli) | Fast contradiction pre-filter | 🟠 Use in conflict detection |
|
| 527 |
+
| BiomedBERT | [`microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract-fulltext`](https://huggingface.co/microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract-fulltext) | Biomedical text processing | 🟡 Use for bio-domain papers |
|
| 528 |
+
|
| 529 |
+
### Key Papers to Read
|
| 530 |
+
|
| 531 |
+
| Paper | ArXiv ID | Why It Matters |
|
| 532 |
+
|-------|----------|---------------|
|
| 533 |
+
| PaperQA2 | [`2409.13740`](https://arxiv.org/abs/2409.13740) | Superhuman scientific QA — benchmark to beat |
|
| 534 |
+
| SciFact | [`2004.14974`](https://arxiv.org/abs/2004.14974) | Defines the claim verification task |
|
| 535 |
+
| SciRIFF | [`2406.07835`](https://arxiv.org/abs/2406.07835) | Best training data for scientific extraction |
|
| 536 |
+
| KGX3 / iKuhn | [`2002.03531`](https://arxiv.org/abs/2002.03531) | Only published epistemic classification system |
|
| 537 |
+
| Paper Circle | [`2604.06170`](https://arxiv.org/abs/2604.06170) | Multi-agent KG construction pattern |
|
| 538 |
+
| FactReview | [`2604.04074`](https://arxiv.org/abs/2604.04074) | Execution-based claim verification |
|
| 539 |
+
| CLAIRE | [`2509.23233`](https://arxiv.org/abs/2509.23233) | Corpus-level contradiction detection |
|
| 540 |
+
| AgentSLR | [`2603.22327`](https://arxiv.org/abs/2603.22327) | End-to-end systematic review pipeline |
|
| 541 |
+
| MultiVerS | [`2112.01640`](https://arxiv.org/abs/2112.01640) | Best model for SUPPORTS/REFUTES |
|
| 542 |
+
| SciERC | [`1808.09602`](https://arxiv.org/abs/1808.09602) | Scientific entity + relation extraction |
|
| 543 |
+
| CritiCal | [`2510.24505`](https://arxiv.org/abs/2510.24505) | Critique-based confidence calibration |
|
| 544 |
+
| CLUE | [`2505.17855`](https://arxiv.org/abs/2505.17855) | Uncertainty source explanation |
|
| 545 |
+
|
| 546 |
+
---
|
| 547 |
+
|
| 548 |
+
## How This Document Was Created
|
| 549 |
+
|
| 550 |
+
1. **Searched** academic databases (arXiv, Semantic Scholar, HuggingFace Papers) for every system matching our 5 core capabilities
|
| 551 |
+
2. **Read** the papers — not just abstracts, but methodology sections — to understand what each system actually does vs. what it claims to do
|
| 552 |
+
3. **Verified** code availability — checked GitHub repos and HuggingFace models to confirm they're real, not vaporware
|
| 553 |
+
4. **Compared** feature-by-feature against our SYSTEM_DESIGN.md
|
| 554 |
+
5. **Ranked** by closeness to our system and quality of implementation
|
| 555 |
+
6. **Wrote** everything in plain language so anyone can follow
|
| 556 |
+
|
| 557 |
+
**This is a living document.** As new systems appear, they should be added here.
|
| 558 |
+
|
| 559 |
+
---
|
| 560 |
+
|
| 561 |
+
*Analysis conducted 2026-04-23 using research across arXiv, HuggingFace Hub, GitHub, and Semantic Scholar.*
|
| 562 |
+
*Every system and resource link has been verified as of this date.*
|