mikeumus-divincian commited on
Commit
55a4cc4
Β·
verified Β·
1 Parent(s): 8fceac7

fix: add sdk: static + sync to latest org card content

Browse files
Files changed (1) hide show
  1. README.md +7 -6
README.md CHANGED
@@ -1,7 +1,14 @@
1
  ---
2
  title: Divinci AI
 
 
 
 
 
 
3
  ---
4
 
 
5
  # Divinci AI
6
 
7
  Feature-level interpretability artifacts for open transformers β€” built openly, validated empirically.
@@ -10,7 +17,6 @@ A **vindex** is a transformer's weights decompiled into a queryable feature data
10
 
11
  Think of it as the model's index: the thing you search before you run it.
12
 
13
- ---
14
 
15
  ## Interactive viewer
16
 
@@ -20,7 +26,6 @@ Think of it as the model's index: the thing you search before you run it.
20
 
21
  Pick any of 9 models from the dropdown. Toggle between the 3D cylinder spiral and a flat 2D circuit/network view. Hit **β‡Œ Compare** to render the current model alongside Bonsai 1-bit, side-by-side β€” the contrast between fp16 structure (organized rings) and 1-bit dissolution (scattered cloud) is the most direct picture of what 1-bit training does to a transformer's internal organization that we know how to render. Search for entity features (`?q=paris&model=gemma-4-e2b`) to see real probe-derived activations light up across the layer stack β€” backed by a 5000-token offline-built search index.
22
 
23
- ---
24
 
25
  ## Published vindexes
26
 
@@ -39,7 +44,6 @@ Cross-family evidence in hand: **Gemma**, **Qwen3**, **Mistral**, **Llama**, **O
39
  | **Bonsai 8B** | 1-bit (Qwen 3 base, post-quantized) | 8B | *vindex pending publish* | 0.429 | **C5 = 1** (circuit dissolved); var@64 = 0.093 |
40
  | **BitNet b1.58-2B-4T** | 1-bit (Microsoft, native) | 2B | *vindex pending publish* | (Phase 2 pending) | **var@64 = 0.111** mean across 30 layers β€” n=2 confirmation of dissolution |
41
 
42
- ---
43
 
44
  ## What's a vindex?
45
 
@@ -49,7 +53,6 @@ Concretely: given a query like `"Paris β†’ capital"`, a vindex walk returns the
49
 
50
  LarQL (the toolchain that builds vindexes) is open-source: [github.com/chrishayuk/larql](https://github.com/chrishayuk/larql) | [github.com/Divinci-AI/larql](https://github.com/Divinci-AI/larql).
51
 
52
- ---
53
 
54
  ## Research
55
 
@@ -73,7 +76,6 @@ Mechanistic knowledge editing in transformer feature space. Includes a negative
73
 
74
  Working notebooks: [github.com/Divinci-AI/server/tree/preview/notebooks](https://github.com/Divinci-AI/server/tree/preview/notebooks)
75
 
76
- ---
77
 
78
  ## Working in public
79
 
@@ -81,6 +83,5 @@ Every measurement in our papers traces back to a notebook and a commit. Negative
81
 
82
  If you replicate a result and find a discrepancy, open an issue on the LarQL repo.
83
 
84
- ---
85
 
86
  *Vindexes on this org are free for academic and research use (CC-BY-NC 4.0). Commercial licensing: mike@divinci.ai*
 
1
  ---
2
  title: Divinci AI
3
+ emoji: 🧠
4
+ colorFrom: green
5
+ colorTo: yellow
6
+ sdk: static
7
+ pinned: false
8
+ short_description: Feature-level interpretability for open transformers
9
  ---
10
 
11
+
12
  # Divinci AI
13
 
14
  Feature-level interpretability artifacts for open transformers β€” built openly, validated empirically.
 
17
 
18
  Think of it as the model's index: the thing you search before you run it.
19
 
 
20
 
21
  ## Interactive viewer
22
 
 
26
 
27
  Pick any of 9 models from the dropdown. Toggle between the 3D cylinder spiral and a flat 2D circuit/network view. Hit **β‡Œ Compare** to render the current model alongside Bonsai 1-bit, side-by-side β€” the contrast between fp16 structure (organized rings) and 1-bit dissolution (scattered cloud) is the most direct picture of what 1-bit training does to a transformer's internal organization that we know how to render. Search for entity features (`?q=paris&model=gemma-4-e2b`) to see real probe-derived activations light up across the layer stack β€” backed by a 5000-token offline-built search index.
28
 
 
29
 
30
  ## Published vindexes
31
 
 
44
  | **Bonsai 8B** | 1-bit (Qwen 3 base, post-quantized) | 8B | *vindex pending publish* | 0.429 | **C5 = 1** (circuit dissolved); var@64 = 0.093 |
45
  | **BitNet b1.58-2B-4T** | 1-bit (Microsoft, native) | 2B | *vindex pending publish* | (Phase 2 pending) | **var@64 = 0.111** mean across 30 layers β€” n=2 confirmation of dissolution |
46
 
 
47
 
48
  ## What's a vindex?
49
 
 
53
 
54
  LarQL (the toolchain that builds vindexes) is open-source: [github.com/chrishayuk/larql](https://github.com/chrishayuk/larql) | [github.com/Divinci-AI/larql](https://github.com/Divinci-AI/larql).
55
 
 
56
 
57
  ## Research
58
 
 
76
 
77
  Working notebooks: [github.com/Divinci-AI/server/tree/preview/notebooks](https://github.com/Divinci-AI/server/tree/preview/notebooks)
78
 
 
79
 
80
  ## Working in public
81
 
 
83
 
84
  If you replicate a result and find a discrepancy, open an issue on the LarQL repo.
85
 
 
86
 
87
  *Vindexes on this org are free for academic and research use (CC-BY-NC 4.0). Commercial licensing: mike@divinci.ai*