README / README.md
AI-that-works's picture
Update README.md
a78b883 verified
metadata
title: groundlens
emoji: πŸ“
colorFrom: yellow
colorTo: red
sdk: static
pinned: false

Groundlens

groundlens

Geometric methods for LLM grounding verification. No second LLM. Deterministic. Same inputs, same scores, every time.

What we do

We detect LLM hallucinations using embedding geometry β€” not by asking another model to judge the output. Two metrics, each targeting a different failure mode:

  • SGI (Semantic Grounding Index) β€” measures whether a response actually used the source material it was given. Built for RAG pipeline verification.
  • DGI (Directional Grounding Index) β€” measures whether a response follows geometric patterns typical of grounded answers. Works without any source context.

Both methods run a single embedding call. Deterministic. Auditable by design.

Research

Three peer-reviewed papers form the foundation:

  1. Semantic Grounding Index (arXiv:2512.13771) β€” ratio-based grounding verification for RAG systems.
  2. A Geometric Taxonomy of Hallucinations (arXiv:2602.13224) β€” three-type hallucination classification with a confabulation benchmark.
  3. Rotational Dynamics of Factual Constraint Processing (arXiv:2603.13259) β€” transformers reject wrong answers via rotation, not rescaling. Phase transition at 1.6B parameters.

Use groundlens

How What
Python library pip install groundlens β€” GitHub Β· Docs
MCP server pip install groundlens-mcp β€” works with Claude Desktop, Cursor, Windsurf β€” GitHub
REST API groundlens-api β€” hosted on this Space, Swagger docs at /docs
Interactive demo groundlens-demo β€” try it without installing anything

Philosophy

groundlens is verification triage, not truth detection. It tells you which responses earned the right to be trusted and which need human review. We publish our AUROC numbers even when they're unflattering. We document what we can't detect (Type III confabulations) as a theorem, not a footnote.

Links