mikeumus-divincian commited on
Commit
c8f0aac
·
verified ·
1 Parent(s): 683e392

Add Divinci-AI organization card

Browse files
Files changed (1) hide show
  1. README.md +54 -7
README.md CHANGED
@@ -1,10 +1,57 @@
1
  ---
2
- title: README
3
- emoji: 😻
4
- colorFrom: pink
5
- colorTo: green
6
- sdk: static
7
- pinned: false
8
  ---
9
 
10
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: Divinci AI
 
 
 
 
 
3
  ---
4
 
5
+ # Divinci AI
6
+
7
+ Feature-level interpretability artifacts for open transformers — built openly, validated empirically.
8
+
9
+ A **vindex** is a transformer's weights decompiled into a queryable feature database. It exposes the entity associations, circuit structure, and knowledge-editing surfaces that live inside a model's FFN layers — without requiring GPU inference for most operations.
10
+
11
+ Think of it as the model's index: the thing you search before you run it.
12
+
13
+ ---
14
+
15
+ ## Published vindexes
16
+
17
+ | Model | Architecture | Params | Vindex |
18
+ |-------|-------------|--------|--------|
19
+ | Gemma 4 E2B-it | Dense (Gemma 4) | 2B | [Divinci-AI/gemma-4-e2b-vindex](https://huggingface.co/Divinci-AI/gemma-4-e2b-vindex) |
20
+ | Qwen3.6-35B-A3B | MoE (Qwen3.6) | 35B / 3B active | [Divinci-AI/qwen3.6-35b-a3b-vindex](https://huggingface.co/Divinci-AI/qwen3.6-35b-a3b-vindex) |
21
+ | GPT-OSS 120B | MoE (OpenAI) | 120B / ~13B active | *building* |
22
+
23
+ Three organizations, three architectures: Gemma dense, Qwen MoE, OpenAI MoE.
24
+
25
+ ---
26
+
27
+ ## What's a vindex?
28
+
29
+ Standard model weights tell you *what* a model computes. A vindex tells you *where* it stores specific knowledge and *which features* need to change for a targeted edit.
30
+
31
+ Concretely: given a query like `"Paris → capital"`, a vindex walk returns the layers, feature directions, and token associations that encode that fact. A patch operation writes a rank-1 ΔW that suppresses or overwrites that association — compiled back to standard HuggingFace safetensors for inference.
32
+
33
+ LarQL (the toolchain that builds vindexes) is open-source: [github.com/chrishayuk/larql](https://github.com/chrishayuk/larql) | [github.com/Divinci-AI/larql](https://github.com/Divinci-AI/larql).
34
+
35
+ ---
36
+
37
+ ## Research
38
+
39
+ **Paper 1 — Architectural Invariants of Transformer Computation** *(arXiv forthcoming)*
40
+ Five properties measured across every model in this collection. Three hold within ±15% coefficient of variation across architectures, organizations, and scales. One collapses under 1-bit quantization. One scales monotonically with model size.
41
+
42
+ **Paper 2 — Constellation Edits** *(draft)*
43
+ Mechanistic knowledge editing in transformer feature space. Includes a negative result: why activation-space edits fail in 1-bit models, and what weight-space geometry reveals about why.
44
+
45
+ Working notebooks: [github.com/Divinci-AI/server/tree/preview/notebooks](https://github.com/Divinci-AI/server/tree/preview/notebooks)
46
+
47
+ ---
48
+
49
+ ## Working in public
50
+
51
+ Every measurement in our papers traces back to a notebook and a commit. Negative results ship alongside positive ones — the compensation mechanism that defeats knowledge editing in 1-bit models is in the notebooks, not buried in a supplement.
52
+
53
+ If you replicate a result and find a discrepancy, open an issue on the LarQL repo.
54
+
55
+ ---
56
+
57
+ *Vindexes on this org are free for academic and research use (CC-BY-NC 4.0). Commercial licensing: mike@divinci.ai*