π GraphRAG Inference Hackathon β Dual Pipeline System
π― What This Is
A dual-pipeline GraphRAG system with 14 novel techniques from cutting-edge 2024β2025 research, 12 LLM providers (including free Ollama local), OpenClaw agent integration, and a production Next.js dashboard β all built on TigerGraph.
| Pipeline A (Baseline) |
Pipeline B (GraphRAG) |
| Query β LLM β Answer |
Query β PolyG Router β PPR Scoring β Spreading Activation β Path Pruning β Token Budget β LLM β Answer |
| Simple, expensive |
Smart, graph-enhanced, cost-controlled |
π 14 Novel Techniques
Graph Retrieval Innovations (from 6 papers)
| # |
Technique |
Paper |
Key Result |
Implementation |
| 1 |
PPR Confidence-Weighted Retrieval |
CatRAG 2602.01965 |
Best reasoning completeness on 4 benchmarks |
PPRConfidenceScorer β Personalized PageRank from seed entities, scores = context confidence |
| 2 |
Spreading Activation Context Scoring |
SA-RAG 2512.15922 |
+39% answer correctness on MuSiQue |
SpreadingActivation β propagates activation through graph with decay, ranks by signal strength |
| 3 |
Flow-Pruned Path Serialization |
PathRAG 2502.14902 |
62β65% win rate vs LightRAG |
PathPruner β finds reasoning paths, prunes by flow threshold, serializes high-reliability first (exploits lost-in-the-middle bias) |
| 4 |
Graph Token Budget Controller |
TERAG 2509.18667 |
97% token reduction at 80%+ accuracy |
TokenBudgetController β caps context by token limit, prioritizes by score Γ relevance |
| 5 |
PolyG Hybrid Retrieval Router |
RAGRouter-Bench 2602.00296 |
Adaptive > any fixed paradigm |
PolyGRouter β 4-class query taxonomy (entity/relation/multi-hop/summarization) β optimal strategy |
| 6 |
Incremental Graph Updates |
TG-RAG 2510.13590 |
O(new) vs O(all) recomputation |
IncrementalGraphUpdater β merge by embedding similarity, scoped community re-detection |
Architecture Innovations
| # |
Technique |
Paper |
Description |
| 7 |
Schema-Bounded Entity Extraction |
Youtu-GraphRAG 2508.19855 |
9 entity types + 15 relation types β ~90% extraction cost reduction, +16% accuracy |
| 8 |
Dual-Level Keyword Retrieval |
LightRAG 2410.05779 |
High-level (themes) + low-level (entities) keywords for dual-channel retrieval |
| 9 |
Adaptive Query Complexity Router |
Original |
LLM scores query complexity 0.0β1.0 β routes simple to baseline, complex to GraphRAG |
| 10 |
Graph Reasoning Path Explanation |
Original |
Natural language step-by-step traversal explanation (Entry β Traversal β Evidence β Conclusion) |
System Innovations
| # |
Technique |
Description |
| 11 |
12-Provider Universal LLM |
Single interface for OpenAI, Claude, Gemini, Mistral, Ollama, Groq, DeepSeek, etc. |
| 12 |
OpenClaw Agent Skills |
GraphRAG as autonomous agent capabilities (CIK model: SOUL + IDENTITY + MEMORY + Skills) |
| 13 |
Live Dashboard Benchmarking |
"Run Benchmark Now" button β judges can evaluate both pipelines in real-time |
| 14 |
Advanced GSQL Queries |
PPR, shortest paths, spreading activation, neighborhood extraction β all as installable TigerGraph queries |
ποΈ Architecture (AI Factory β 4 Layers)
ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β LAYER 4: EVALUATION β
β RAGAS β F1/EM β Token Tracking β Live Benchmark β Next.js Dashboard β
ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β LAYER 3: UNIVERSAL LLM (12 Providers) β
β OpenAI β Claude β Gemini β Mistral β Ollama β Groq β DeepSeek β β¦ β
ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β LAYER 2: INFERENCE ORCHESTRATION + NOVELTY ENGINE β
β ββ PolyG Router ββ PPR Scoring ββ Spreading Activation ββ β
β β Path Pruning ββ Token Budget ββ Structured Context β β
β ββ Pipeline A: Baseline (Query β Vector β LLM) β β
β ββ Pipeline B: GraphRAG (Query β Graph β Novelties β LLM)β β
ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β LAYER 1: GRAPH (TigerGraph) β
β GSQL: PPR β Shortest Paths β Spreading Activation β Vector Search β
β Schema: Document β Chunk β Entity β Community β
β Incremental Updates β Schema-Bounded Extraction β
ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
How the Novelty Engine Works (Pipeline B)
Query: "Were Einstein and Newton of the same nationality?"
Step 1: PolyG Router β "multi_hop" (score=0.7) β use graph_traversal
Step 2: PPR from seeds [Einstein, Newton] β score all reachable entities
Step 3: Spreading Activation β expand to 2-hop neighborhood with decay
Step 4: Combined scoring (0.6ΓPPR + 0.4ΓActivation) per chunk
Step 5: Token Budget (2000 tokens) β select top chunks, prune 60%+ redundancy
Step 6: Path Serialization β "Einstein βBORN_INβ Germany, Newton βBORN_INβ England"
Step 7: LLM generates answer with ranked, pruned, path-structured context
π Quick Start
cd web && npm install && npm run dev
docker build -t graphrag . && docker run -p 3000:3000 graphrag
pip install -r requirements.txt && python -m graphrag.main demo
ollama pull llama3.2 && cd web && npm install && npm run dev
π€ 12 LLM Providers
| Provider |
Model |
Cost |
Speed |
| Ollama π¦ |
llama3.2 |
$0 |
β‘ Local |
| HuggingFace |
Llama 3.3 70B |
$0 |
π΅ Medium |
| DeepSeek |
DeepSeek V3 |
$0.00014/1K |
β‘ Fast |
| OpenAI |
GPT-4o-mini |
$0.00015/1K |
β‘ Fast |
| Groq |
Llama 3.3 70B |
$0.0006/1K |
β‘β‘ Blazing |
| Gemini |
2.0 Flash |
$0.0001/1K |
β‘ Fast |
| Mistral |
Large |
$0.002/1K |
π΅ Medium |
| Anthropic |
Claude Sonnet 4 |
$0.003/1K |
π΅ Medium |
| OpenRouter |
200+ models |
Varies |
Varies |
| Cohere |
Command R+ |
$0.0025/1K |
π΅ Medium |
| xAI |
Grok 3 |
$0.003/1K |
π΅ Medium |
| Together |
Llama 3.1 70B |
$0.0009/1K |
β‘ Fast |
π Benchmarks
Live Benchmark (from Dashboard)
Click "π Run Benchmark Now" β evaluates both pipelines on HotpotQA with real F1/EM.
Expected Performance (HotpotQA)
| Metric |
Baseline |
GraphRAG |
Ξ |
Winner |
| F1 Score |
~0.45β0.60 |
~0.55β0.70 |
+13β21% |
β
GraphRAG |
| Exact Match |
~0.30β0.45 |
~0.35β0.50 |
+11% |
β
GraphRAG |
| Tokens/Query |
~800β1000 |
~1200β1800* |
β |
β
Baseline |
| F1 Win Rate |
β |
~55β70% |
β |
β
GraphRAG |
*With Token Budget Controller, GraphRAG context is capped at 2000 tokens β 40β60% reduction vs. unbounded.
By Question Type
| Type |
Baseline F1 |
GraphRAG F1 |
Ξ |
Why |
| Bridge (multi-hop) |
~0.52 |
~0.63 |
+21% |
Graph traversal connects cross-document facts |
| Comparison |
~0.58 |
~0.61 |
+5% |
Entity-pair paths provide structured comparison context |
π¦ OpenClaw Agent Integration
| Component |
File |
Purpose |
| SOUL.md |
openclaw/SOUL.md |
Agent identity, values, boundaries |
| IDENTITY.md |
openclaw/IDENTITY.md |
Provider config, schema, channels |
| MEMORY.md |
openclaw/MEMORY.md |
Learned performance knowledge |
| graph_query |
openclaw/skills/graph_query/ |
NL β knowledge graph traversal |
| compare_pipelines |
openclaw/skills/compare_pipelines/ |
Dual-pipeline comparison |
| cost_estimate |
openclaw/skills/cost_estimate/ |
12-provider cost projection |
π§ͺ Testing
python tests/test_core.py
python tests/test_novelties.py
π Project Structure (75 files, 280KB)
βββ web/ # Next.js 15 Dashboard
β βββ src/app/api/
β β βββ compare/route.ts # Multi-provider dual-pipeline API
β β βββ benchmark/route.ts # Live benchmark with F1/EM
β β βββ providers/route.ts # Provider health + listing
β βββ src/components/tabs/
β β βββ LiveCompare.tsx # Provider selector + comparison
β β βββ Benchmark.tsx # Live "Run Now" + charts
β β βββ CostAnalysis.tsx # 12-provider projections
β β βββ GraphExplorer.tsx # Interactive SVG graph
β βββ src/lib/
β βββ llm-providers.ts # 12-provider universal client
β βββ design-tokens.ts # TigerGraphΓClaude tokens
β
βββ graphrag/layers/
β βββ graph_layer.py # Layer 1: TigerGraph + GSQL
β βββ orchestration_layer.py # Layer 2: Dual pipeline + routing
β βββ llm_layer.py # Layer 3: LLM interactions
β βββ universal_llm.py # Layer 3: 12-provider support
β βββ evaluation_layer.py # Layer 4: RAGAS + F1/EM
β βββ novelties.py # π 6 novel techniques (NEW)
β βββ gsql_advanced.py # π Advanced GSQL queries (NEW)
β
βββ openclaw/ # OpenClaw Agent (CIK model)
βββ tests/
β βββ test_core.py # 31 core tests
β βββ test_novelties.py # 24 novelty tests (NEW)
βββ Dockerfile
βββ README.md
π References
Directly Implemented (6 papers)
- CatRAG β PPR + Dynamic Edge Weighting β arXiv:2602.01965 (Feb 2025)
- PathRAG β Flow-Pruned Path Retrieval β arXiv:2502.14902 (Feb 2025)
- TERAG β Token-Efficient Graph RAG β arXiv:2509.18667 (Sep 2024)
- SA-RAG β Spreading Activation Retrieval β arXiv:2512.15922 (Dec 2024)
- RAGRouter-Bench β Hybrid Routing β arXiv:2602.00296 (Feb 2025)
- TG-RAG β Incremental Temporal Graph β arXiv:2510.13590 (Oct 2024)
Architecture Inspiration (4 papers)
- GraphRAG β Microsoft's Community-Based RAG β arXiv:2404.16130
- LightRAG β Dual-Level Retrieval (34Kβ) β arXiv:2410.05779
- Youtu-GraphRAG β Schema-Bounded Extraction (Tencent) β arXiv:2508.19855
- HippoRAG 2 β PPR + Passage Integration β arXiv:2502.14802
Datasets & Evaluation
- HotpotQA β Multi-hop QA benchmark
- RAGAS β RAG evaluation framework
π Built for the GraphRAG Inference Hackathon by TigerGraph
14 Novel Techniques Β· 10 Research Papers Β· 12 LLM Providers Β· 55 Unit Tests Β· OpenClaw Agent Β· Docker
Proving that graphs make LLM inference faster, cheaper, and smarter.