| --- |
| license: cc-by-2.5 |
| language: |
| - en |
| metrics: |
| - accuracy |
| new_version: GrokciAI/Quidoris_Engine |
| repository: https://github.com/Grokci/Quidoris_Engine/tree/main |
| tags: |
| - chemistry |
| - biology |
| - finance |
| - music |
| - art |
| - code |
| - climate |
| - medical |
| - agent |
| - text-generation-inference |
| - document-question-answering |
| pipeline_tag: reinforcement-learning |
| library_name: fastai |
| --- |
| # Infra / local-first implementation details |
| - sqlite |
| - fts5 |
| - streaming |
| - sse |
| - local-inference |
| - byok |
|
|
| # Intended ecosystem compatibility |
| - openai-compatible |
| - huggingface |
| - mistral |
| - llama |
|
|
| # Product-ish / devtool signals |
| - devtools |
| - productivity |
| - citations |
| - interpretability |
| --- |
|
|
| # Core Idea |
|
|
|
|
| **Core idea:** *Don’t stuff your library into the model. Let the model interact with the library.* |
|
|
| Relationship Learning Models promise the ability to have unlimited context windows by providing an environmment in which they can "scan" a document library and retrieve |
| information from this library be taking small snapshots of the information, and either answering the prompt provided it from these snapshots, or, additionally call small agents |
| to assist it in returing their reply to the user's prompt. This model type promises to reduce content rot to zero. |
|
|
| The Quidoris Engine is an RLM environment. It is a harness that lives independently from an RLM but RLM's can dock with the harness and use its powerful environment |
| to run an unlimited context window. Right now LLM's are limited to a few million token context windows. |
| RLM's, in concert with the Quidoris Engine, have unlimited contect windows. Think 1000 documents or 10000 document context windows. This is what the |
| Quidoris Engine does for all RLM's. |
|
|
|
|
|
|
| # Model Card for Model ID |
|
|
| Quidoris Engine (RLM Harness) |
|
|
| Quidoris Engine is a **local-first** inference harness inspired by *Recursive Language Models (RLMs)*: |
| it treats long prompts and large document libraries as an **external environment**, letting models **search, read, cite, and recurse**—instead of cramming everything |
| into a single context window. |
|
|
| It runs as a **local daemon** with an HTTP API (SSE-friendly) and ships with a lightweight local web UI. |
|
|
|
|
|
|
| ## Model Details |
|
|
| ### Model Description |
|
|
| A RLM inference harness to allow any user to utilize RLM's locally on their own machines. |
|
|
| ## What this is |
|
|
| - A **local daemon** (Bun) + **SQLite/FTS5** index for fast retrieval |
| - A **provider-agnostic** harness (BYOK): Local CLI / Hugging Face / OpenAI-compatible providers |
| - A **UI Launcher** that can auto-start the daemon (since browsers can’t spawn local processes) |
| - An “evidence-first” workflow: **citations** and **evidence spotlight** are first-class |
|
|
| --- |
|
|
| ## Why it exists |
|
|
| Large models struggle with: |
| - **Context window limits** |
| - **Hallucination risk** when they can’t reference exact evidence |
| - **Cost blowups** when prompts get huge |
|
|
| Quidoris Engine makes “long context” feel interactive: |
| - keep the **source material outside** the model |
| - allow **programmatic access** via search/read APIs |
| - produce answers with **traceable evidence** |
|
|
| --- |
|
|
| ## Features |
|
|
| ### Library + Index |
| - Local SQLite index (`rlm_index.sqlite`) |
| - FTS5 search (chunked content for recall + precise snippets) |
| - Incremental indexing (re-index only when files change) |
|
|
| ### RLM-friendly runtime |
| - Designed for “action → observe → recurse” loops |
| - Planned/compatible with async/batched `llm_query` patterns |
| - SSE endpoints for streaming runs/steps (where available) |
|
|
| ### Local web UI |
| - Glassy login/boot screen |
| - “Enter the Engine” auto-starts the daemon via the local launcher |
| - Workspace scaffold (`/app`) for Library / Runs / Evidence |
|
|
| --- |
|
|
| ## Install |
|
|
| ### Requirements |
| - **Bun** (recommended) |
| - SQLite (bundled via `bun:sqlite`) |
|
|
| Clone / download from Hugging Face or GitHub. |
|
|
| --- |
|
|
| ## Quickstart (local) |
|
|
| ### 1) Build UI |
| ```bash |
| cd ui |
| bun install |
| bun run build |
| |
| |
| |
| - **Developed by:** Ken Isaacson, Sr. "Grokci" |
| - **Funded by [optional]:** Quidoris Research Group" |
| - **Shared by [optional]:** Quidoris Research Group" |
| - **Model type:** Relationship Learning Model Harness - for use with Relationship Learning Models and Large Language Models |
| - **Language(s) (NLP):** BUN, Python, SQLite |
| - **License:** cc-by-2.5 |
| - |
| ### Model Sources [optional] |
| |
| <!-- Provide the basic links for the model. --> |
| |
| - **Repository:** [More Information Needed] |
| - **Paper [optional]:** [More Information Needed] |
| - **Demo [optional]:** [More Information Needed] |
| |
| **Quidoris Engine is not a trained model**, it’s a *local RLM harness/engine + daemon + UI*. |
| |
| |
| --- |
| |
| ## Uses |
| |
| Quidoris Engine is intended to be used as a **local-first inference harness** that enables RLM-style workflows: treating large prompts and document libraries as an **external environment** the model can interact with via actions (search/read/cite), rather than loading everything into the model context. |
| |
| Foreseeable users: |
| |
| * Individual developers running local tooling |
| * Researchers prototyping RLM-like inference loops |
| * Teams running BYOK workflows on laptops/workstations |
| * Contributors extending adapters, ingestion, or UI |
| |
| People potentially affected: |
| |
| * The user (and any collaborators) whose documents are indexed and queried |
| * Anyone whose content is included in the local library (e.g., shared docs) |
| * Any downstream consumers of generated outputs (reports, emails, code, decisions) |
| |
| ### Direct Use |
| |
| Direct use means running Quidoris Engine locally as shipped: |
| |
| * Start the **launcher + UI** locally |
| * Auto-start the **daemon** |
| * Index and manage a local document library |
| * Use the daemon API to: |
| |
| * search local docs (FTS5) |
| * read snippets/regions |
| * generate model outputs with evidence/citations (as those endpoints are implemented) |
| * Use the UI to: |
| |
| * sign in locally |
| * enter the workspace (“Engine”) |
| * (planned) upload/manage docs, tags/folders, run tasks, view evidence spotlight |
| |
| Typical scenarios: |
| |
| * “I have a large folder of PDFs + notes. Let the model answer questions with citations.” |
| * “I want a controllable long-context tool that’s provider-agnostic.” |
| * “I want an RLM-like loop where the model chooses actions (search/read/subcall) and I can inspect evidence.” |
| |
| ### Downstream Use [optional] |
| |
| Downstream use means embedding Quidoris Engine into a larger system (still local-first) or using it as a substrate for an RLM agent/controller: |
| |
| * Integrating Quidoris Engine into an internal toolchain via `/v1/*` API |
| * Building a custom “RLM controller” that: |
| |
| * decides actions (search/read/transform/query-submodel) |
| * calls Quidoris Engine as the environment layer |
| * streams steps via SSE to a UI or logger |
| * Extending adapters for model providers (HF inference, local CLI, OpenAI-compatible servers) |
| * Building domain-specific ingestion pipelines (PDF→text, audio→transcript, etc.) |
| |
| ### Out-of-Scope Use |
| |
| Quidoris Engine is **not** intended for: |
| |
| * Running as a public hosted multi-tenant SaaS (this repo is local-first by design) |
| * Handling regulated/high-stakes workflows without additional controls (health, legal, finance) |
| * Serving as a security boundary between untrusted users and private documents (local indexing implies access) |
| * “Guaranteed correctness” answers: it can improve evidence access, but models can still misinterpret |
| * Real-time collaborative editing or “Google Drive-like” shared document management |
| * Malware, phishing, or abusive content generation (the tool should not be used to facilitate harm) |
| |
| --- |
| |
| ## Bias, Risks, and Limitations |
| |
| Technical limitations: |
| |
| * Retrieval is only as good as: |
| |
| * ingestion quality (OCR, parsing errors) |
| * chunking strategy |
| * query formulation |
| * SQLite FTS5 is lexical (not semantic embeddings by default). It can miss paraphrases unless augmented. |
| * Evidence spotlight/citations reduce hallucinations but don’t eliminate: |
| |
| * incorrect reasoning |
| * cherry-picked evidence |
| * quoting irrelevant snippets |
| * Provider behavior differs: different LLMs can produce different answers to identical evidence. |
| |
| Sociotechnical / user risks: |
| |
| * Private data risk: users may index sensitive documents; local-first reduces exposure but doesn’t remove it. |
| * “Trust from polish” risk: a beautiful UI can cause over-trust in outputs. |
| * Misuse risk: tool can be used to summarize or extract sensitive data if the library contains it. |
| * Attribution risk: users may treat retrieved snippets as definitive without verifying full context. |
| |
| System limitations (current repo state): |
| |
| * Many `/v1/*` endpoints in the OpenAPI spec may still be stubs returning `501` until implemented. |
| * Current auth is local-oriented and not a full enterprise IAM solution. |
| * Uploading non-text media (images/audio/video) requires ingestion pipelines not yet included by default. |
| |
| ### Recommendations |
| |
| * Treat outputs as *drafts* unless validated, especially for high-stakes decisions. |
| * Always inspect: |
| |
| * which documents were retrieved |
| * which chunks were used |
| * whether the snippet context is sufficient |
| * Add guardrails for sensitive libraries: |
| |
| * separate profiles/databases for different projects |
| * explicit “do not index” directories |
| * local encryption at rest if required |
| * If you enable remote access, lock it down: |
| |
| * bind to localhost only (default) |
| * require auth |
| * avoid exposing ports publicly |
| * For better recall, consider augmenting FTS with semantic retrieval later (optional plugin). |
| |
| --- |
| |
| ## How to Get Started with the Model |
| |
| > Quidoris Engine is a tool (daemon + UI), not a trained model. |
| |
| ### Quickstart (local UI + auto-start daemon) |
| |
| ```bash |
| # Build the UI |
| cd ui |
| bun install |
| bun run build |
|
|
| # Start the launcher (serves UI + proxies API + can spawn the daemon) |
| cd .. |
| bun run ui:launch |
| ``` |
| |
| Open: |
| |
| * [http://127.0.0.1:5173](http://127.0.0.1:5173) |
| |
| Click **Enter the Engine**: |
| |
| * launcher starts the daemon if needed |
| * UI waits for `/v1/health`, performs login, routes to `/app` |
| |
| ### Daemon only (no UI) |
| |
| ```bash |
| bun run quidoris-engine.ts daemon --port 8787 |
| # then call: |
| # GET http://127.0.0.1:8787/v1/health |
| ``` |
| |
| --- |
| |
| ## Training Details |
| |
| > Not applicable. Quidoris Engine is not a trained model; it is a local-first harness for running inference workflows against user-provided documents and user-selected model providers. |
| |
| ### Training Data |
| |
| Not applicable. |
| |
| ### Training Procedure |
| |
| Not applicable. |
| |
| #### Preprocessing [optional] |
| |
| Not applicable (except local document ingestion/indexing, which is not training). |
| |
| #### Training Hyperparameters |
| |
| * **Training regime:** Not applicable. |
| |
| #### Speeds, Sizes, Times [optional] |
| |
| Not applicable for “training.” |
| (Performance notes that *are* relevant would be indexing speed, query latency, daemon startup time, etc., which depend heavily on machine and library size.) |
| |
| --- |
| |
| ## Evaluation |
| |
| Quidoris Engine itself is not evaluated like a model, but it can be evaluated as a *system* (retrieval quality, citation accuracy, latency). |
| |
| ### Testing Data, Factors & Metrics |
| |
| #### Testing Data |
| |
| Not applicable as a model card training set; for system evaluation you can use: |
| |
| * a curated local doc set with known answers |
| * “needle-in-haystack” tests (known facts buried in large libraries) |
| * long-context QA benchmarks adapted to tool-based retrieval (optional) |
| |
| #### Factors |
| |
| Suggested evaluation factors: |
| |
| * document types: markdown, txt, html, pdf→text, code, emails |
| * library size: 100 docs vs 10k docs |
| * chunk sizes: 8KB vs 32KB |
| * query styles: keyword vs natural language |
| * provider differences: local vs hosted LLMs |
| |
| #### Metrics |
| |
| Suggested metrics for system-level evaluation: |
| |
| * Retrieval: |
| |
| * recall@k (did the correct doc appear in top k?) |
| * snippet precision (did the snippet contain the needed evidence?) |
| * Evidence correctness: |
| |
| * citation validity (does cited text actually support the claim?) |
| * Latency: |
| |
| * search time |
| * end-to-end “question→answer” time |
| * Stability: |
| |
| * incremental indexing correctness |
| * reproducibility across runs |
| |
| ### Results |
| |
| Not yet reported in this repository. |
| |
| #### Summary |
| |
| Quidoris Engine is an infrastructure layer; evaluation is best done per deployment and provider choice. |
| |
| --- |
| |
| ## Model Examination [optional] |
| |
| Not applicable as a trained model. |
| However, Quidoris Engine supports *system examination*: |
| |
| * evidence spotlight (inspect chunks used) |
| * step traces (for RLM-like action/observation loops) |
| * logs of search/read actions |
| |
| --- |
| |
| ## Environmental Impact |
| |
| Not applicable as a trained model. |
| |
| For local usage: |
| |
| * emissions are dominated by whichever model provider you choose (local GPU, hosted inference, etc.) |
| * Quidoris Engine overhead is primarily: |
| |
| * SQLite indexing (CPU) |
| * serving UI + HTTP daemon (minimal) |
| |
| If you want, you can report: |
| |
| * indexing energy/time on a reference machine |
| * typical daemon runtime overhead |
| |
| --- |
| |
| ## Technical Specifications [optional] |
| |
| ### Model Architecture and Objective |
| |
| Quidoris Engine is not a model architecture. It is an **inference harness** + **local environment** that: |
| |
| * stores a document library externally (SQLite + FTS5) |
| * exposes tools (`search`, `read`) via HTTP |
| * supports an agent/controller loop (RLM-style) that can recurse and cite evidence |
| * uses SSE-friendly API patterns for streaming steps/results |
| |
| ### Compute Infrastructure |
| |
| Runs locally on user machines. |
| |
| #### Hardware |
| |
| Typical: |
| |
| * CPU laptop/workstation |
| * optional GPU if running local models (not required for the engine itself) |
| |
| #### Software |
| |
| * Bun runtime |
| * SQLite via `bun:sqlite` |
| * UI: Vite + React (local web app) |
| * API: local HTTP daemon + launcher proxy |
| |
| --- |
| |
| ## Citation [optional] |
| |
| |
| **BibTeX:** |
| |
| ```bibtex |
| @software{quidoris_engine_2026, |
| author = {Quidoris Research Group}, |
| title = {Quidoris Engine: Local-first RLM Harness}, |
| year = {2026}, |
| url = {https://huggingface.co/Quidoris/Quidoris_Engine |
| } |
| ``` |
| |
| **APA:** |
| Quidoris Research Group. (2026). *Quidoris Engine: Local-first RLM Harness* [Software]. Hugging Face. [https://huggingface.co/](https://huggingface.co/)<ORG>/<REPO> |
| |
| --- |
| |
| ## Glossary [optional] |
| |
| * **RLM (Recursive Language Model):** an inference strategy where the model treats long prompts as an external environment and recursively operates over parts of it via tools/actions. |
| * **FTS5:** SQLite Full-Text Search extension for fast lexical search with ranking/snippets. |
| * **BYOK:** Bring Your Own Key — you supply API credentials to providers; tool does not centrally store them. |
| * **Evidence spotlight:** UI pattern that opens the exact retrieved snippet(s) supporting a claim. |
| * **SSE:** Server-Sent Events — HTTP streaming alternative to WebSockets for one-way event streams. |
| |
| --- |
| |
| ## More Information [optional] |
| |
| * OpenAPI spec: `openapi/openapi.yaml` |
| * API quickstart: `docs/api.md` |
| * DB schema: `docs/db_schema.sql` |
| * Error schema: `docs/errors.md` |
| * Pagination: `docs/pagination.md` |
| * SSE notes: `docs/sse.md` |
|
|
| --- |
|
|
| ## Model Card Authors [optional] |
|
|
| * Quidoris Research Group |
|
|
| ## Model Card Contact |
|
|
| * Project contact: **Ken (@grokci)** |
|
|
| --- |
|
|
| If you want, I can also rewrite this in the “official HF style” with a tighter tone and a short “TL;DR” at the top, plus a more HF-native tag set (`tools`, `agent`, `retrieval-augmented-generation`, etc.). |