Spaces:
Running
Running
metadata
title: groundlens
emoji: π
colorFrom: yellow
colorTo: red
sdk: static
pinned: false
Groundlens
Geometric methods for LLM grounding verification. No second LLM. Deterministic. Same inputs, same scores, every time.
What we do
We detect LLM hallucinations using embedding geometry β not by asking another model to judge the output. Two metrics, each targeting a different failure mode:
- SGI (Semantic Grounding Index) β measures whether a response actually used the source material it was given. Built for RAG pipeline verification.
- DGI (Directional Grounding Index) β measures whether a response follows geometric patterns typical of grounded answers. Works without any source context.
Both methods run a single embedding call. Deterministic. Auditable by design.
Research
Three peer-reviewed papers form the foundation:
- Semantic Grounding Index (arXiv:2512.13771) β ratio-based grounding verification for RAG systems.
- A Geometric Taxonomy of Hallucinations (arXiv:2602.13224) β three-type hallucination classification with a confabulation benchmark.
- Rotational Dynamics of Factual Constraint Processing (arXiv:2603.13259) β transformers reject wrong answers via rotation, not rescaling. Phase transition at 1.6B parameters.
Use groundlens
| How | What |
|---|---|
| Python library | pip install groundlens β GitHub Β· Docs |
| MCP server | pip install groundlens-mcp β works with Claude Desktop, Cursor, Windsurf β GitHub |
| REST API | groundlens-api β hosted on this Space, Swagger docs at /docs |
| Interactive demo | groundlens-demo β try it without installing anything |
Philosophy
groundlens is verification triage, not truth detection. It tells you which responses earned the right to be trusted and which need human review. We publish our AUROC numbers even when they're unflattering. We document what we can't detect (Type III confabulations) as a theorem, not a footnote.