Spaces:
Configuration error
Configuration error
fix: add sdk: static + sync to latest org card content
Browse files
README.md
CHANGED
|
@@ -1,7 +1,14 @@
|
|
| 1 |
---
|
| 2 |
title: Divinci AI
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
| 4 |
|
|
|
|
| 5 |
# Divinci AI
|
| 6 |
|
| 7 |
Feature-level interpretability artifacts for open transformers β built openly, validated empirically.
|
|
@@ -10,7 +17,6 @@ A **vindex** is a transformer's weights decompiled into a queryable feature data
|
|
| 10 |
|
| 11 |
Think of it as the model's index: the thing you search before you run it.
|
| 12 |
|
| 13 |
-
---
|
| 14 |
|
| 15 |
## Interactive viewer
|
| 16 |
|
|
@@ -20,7 +26,6 @@ Think of it as the model's index: the thing you search before you run it.
|
|
| 20 |
|
| 21 |
Pick any of 9 models from the dropdown. Toggle between the 3D cylinder spiral and a flat 2D circuit/network view. Hit **β Compare** to render the current model alongside Bonsai 1-bit, side-by-side β the contrast between fp16 structure (organized rings) and 1-bit dissolution (scattered cloud) is the most direct picture of what 1-bit training does to a transformer's internal organization that we know how to render. Search for entity features (`?q=paris&model=gemma-4-e2b`) to see real probe-derived activations light up across the layer stack β backed by a 5000-token offline-built search index.
|
| 22 |
|
| 23 |
-
---
|
| 24 |
|
| 25 |
## Published vindexes
|
| 26 |
|
|
@@ -39,7 +44,6 @@ Cross-family evidence in hand: **Gemma**, **Qwen3**, **Mistral**, **Llama**, **O
|
|
| 39 |
| **Bonsai 8B** | 1-bit (Qwen 3 base, post-quantized) | 8B | *vindex pending publish* | 0.429 | **C5 = 1** (circuit dissolved); var@64 = 0.093 |
|
| 40 |
| **BitNet b1.58-2B-4T** | 1-bit (Microsoft, native) | 2B | *vindex pending publish* | (Phase 2 pending) | **var@64 = 0.111** mean across 30 layers β n=2 confirmation of dissolution |
|
| 41 |
|
| 42 |
-
---
|
| 43 |
|
| 44 |
## What's a vindex?
|
| 45 |
|
|
@@ -49,7 +53,6 @@ Concretely: given a query like `"Paris β capital"`, a vindex walk returns the
|
|
| 49 |
|
| 50 |
LarQL (the toolchain that builds vindexes) is open-source: [github.com/chrishayuk/larql](https://github.com/chrishayuk/larql) | [github.com/Divinci-AI/larql](https://github.com/Divinci-AI/larql).
|
| 51 |
|
| 52 |
-
---
|
| 53 |
|
| 54 |
## Research
|
| 55 |
|
|
@@ -73,7 +76,6 @@ Mechanistic knowledge editing in transformer feature space. Includes a negative
|
|
| 73 |
|
| 74 |
Working notebooks: [github.com/Divinci-AI/server/tree/preview/notebooks](https://github.com/Divinci-AI/server/tree/preview/notebooks)
|
| 75 |
|
| 76 |
-
---
|
| 77 |
|
| 78 |
## Working in public
|
| 79 |
|
|
@@ -81,6 +83,5 @@ Every measurement in our papers traces back to a notebook and a commit. Negative
|
|
| 81 |
|
| 82 |
If you replicate a result and find a discrepancy, open an issue on the LarQL repo.
|
| 83 |
|
| 84 |
-
---
|
| 85 |
|
| 86 |
*Vindexes on this org are free for academic and research use (CC-BY-NC 4.0). Commercial licensing: mike@divinci.ai*
|
|
|
|
| 1 |
---
|
| 2 |
title: Divinci AI
|
| 3 |
+
emoji: π§
|
| 4 |
+
colorFrom: green
|
| 5 |
+
colorTo: yellow
|
| 6 |
+
sdk: static
|
| 7 |
+
pinned: false
|
| 8 |
+
short_description: Feature-level interpretability for open transformers
|
| 9 |
---
|
| 10 |
|
| 11 |
+
|
| 12 |
# Divinci AI
|
| 13 |
|
| 14 |
Feature-level interpretability artifacts for open transformers β built openly, validated empirically.
|
|
|
|
| 17 |
|
| 18 |
Think of it as the model's index: the thing you search before you run it.
|
| 19 |
|
|
|
|
| 20 |
|
| 21 |
## Interactive viewer
|
| 22 |
|
|
|
|
| 26 |
|
| 27 |
Pick any of 9 models from the dropdown. Toggle between the 3D cylinder spiral and a flat 2D circuit/network view. Hit **β Compare** to render the current model alongside Bonsai 1-bit, side-by-side β the contrast between fp16 structure (organized rings) and 1-bit dissolution (scattered cloud) is the most direct picture of what 1-bit training does to a transformer's internal organization that we know how to render. Search for entity features (`?q=paris&model=gemma-4-e2b`) to see real probe-derived activations light up across the layer stack β backed by a 5000-token offline-built search index.
|
| 28 |
|
|
|
|
| 29 |
|
| 30 |
## Published vindexes
|
| 31 |
|
|
|
|
| 44 |
| **Bonsai 8B** | 1-bit (Qwen 3 base, post-quantized) | 8B | *vindex pending publish* | 0.429 | **C5 = 1** (circuit dissolved); var@64 = 0.093 |
|
| 45 |
| **BitNet b1.58-2B-4T** | 1-bit (Microsoft, native) | 2B | *vindex pending publish* | (Phase 2 pending) | **var@64 = 0.111** mean across 30 layers β n=2 confirmation of dissolution |
|
| 46 |
|
|
|
|
| 47 |
|
| 48 |
## What's a vindex?
|
| 49 |
|
|
|
|
| 53 |
|
| 54 |
LarQL (the toolchain that builds vindexes) is open-source: [github.com/chrishayuk/larql](https://github.com/chrishayuk/larql) | [github.com/Divinci-AI/larql](https://github.com/Divinci-AI/larql).
|
| 55 |
|
|
|
|
| 56 |
|
| 57 |
## Research
|
| 58 |
|
|
|
|
| 76 |
|
| 77 |
Working notebooks: [github.com/Divinci-AI/server/tree/preview/notebooks](https://github.com/Divinci-AI/server/tree/preview/notebooks)
|
| 78 |
|
|
|
|
| 79 |
|
| 80 |
## Working in public
|
| 81 |
|
|
|
|
| 83 |
|
| 84 |
If you replicate a result and find a discrepancy, open an issue on the LarQL repo.
|
| 85 |
|
|
|
|
| 86 |
|
| 87 |
*Vindexes on this org are free for academic and research use (CC-BY-NC 4.0). Commercial licensing: mike@divinci.ai*
|