TinmanLabSL commited on
Commit
4fe7b6d
·
verified ·
1 Parent(s): eef6293

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +86 -63
README.md CHANGED
@@ -1,63 +1,86 @@
1
- ---
2
- title: README
3
- emoji: 🔩
4
- colorFrom: gray
5
- colorTo: blue
6
- sdk: static
7
- pinned: false
8
- ---
9
-
10
- <div align="center">
11
-
12
- # Tinman Lab
13
-
14
- ### Autonomous Machines. Second-Order Systems.
15
-
16
- <sub>AGENT MEMORY · ADVERSARIAL SAFETY · AGENTIC ECONOMY · PERCEPTION SYSTEMS · APPLIED RESEARCH</sub>
17
-
18
- ---
19
-
20
- </div>
21
-
22
- ## Research
23
-
24
- ### Disposition Distillation a three-arc negative result
25
-
26
- We set out to train *behavioral dispositions* — self-verification, uncertainty
27
- acknowledgment, feedback integration — into sub-billion-parameter language
28
- models. Across three independent operator classes (SFT/DPO LoRA imitation,
29
- attention-head tempering on `o_proj`, and frozen-base hidden-state confidence
30
- sidecars), no operator moved judge-measured disposition without simultaneously
31
- damaging content quality or collapsing into stylistic mimicry. The failure is
32
- consistent across Qwen3-0.6B, Qwen3-1.7B, Qwen3.5-0.8B, Gemma 4 E2B, and
33
- SmolLM2-1.7B-Instruct.
34
-
35
- The contribution is the falsification, a two-failure-mode taxonomy for linear
36
- hidden-state probes, and a methodological pipeline that converts
37
- CV-on-same-distribution false positives into honest negatives.
38
-
39
- 📄 **Paper** — [arXiv:2604.11867](https://arxiv.org/abs/2604.11867)
40
- 🔧 **Artifacts** — [github.com/tinmanlabsl/disposition-distillation](https://github.com/tinmanlabsl/disposition-distillation)
41
-
42
- ## Open-source proofs of concept
43
-
44
- Apache 2.0. Internal product models are not listed here.
45
-
46
- ### Tinman SmolOmni (MLA)
47
-
48
- - [Tinman-SmolOmni-MLA-256M](https://huggingface.co/Tinman-Lab/Tinman-SmolOmni-MLA-256M)
49
- - [Tinman-SmolOmni-MLA-500M](https://huggingface.co/Tinman-Lab/Tinman-SmolOmni-MLA-500M)
50
- - [Tinman-SmolOmni-MLA-Toolkit](https://huggingface.co/Tinman-Lab/Tinman-SmolOmni-MLA-Toolkit)
51
-
52
- ### Tinman Companion (Gemma 4)
53
-
54
- - [Tinman-gemma4-companion-merged](https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-merged) — full precision
55
- - [Tinman-gemma4-companion-gguf](https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-gguf) — GGUF quantized for llama.cpp
56
- - [Tinman-gemma4-companion-litert-lm](https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-litert-lm) — LiteRT-LM for on-device deployment
57
- - [Tinman-gemma4-companion-sft](https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-sft) — SFT checkpoint
58
- - [Tinman-gemma4-companion-dpo](https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-dpo) — DPO checkpoint
59
-
60
- ## Links
61
-
62
- - [Website](https://tinmanlab.com)
63
- - [GitHub](https://github.com/tinmanlabsl/)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ title: README
3
+ emoji: 🔩
4
+ colorFrom: gray
5
+ colorTo: blue
6
+ sdk: static
7
+ pinned: false
8
+ ---
9
+ <div class="center">
10
+ <h1>Tinman Lab</h1>
11
+ <h3>Autonomous Machines. Second-Order Systems.</h3>
12
+ <div class="sub">AGENT MEMORY · ADVERSARIAL SAFETY · AGENTIC ECONOMY · PERCEPTION SYSTEMS · APPLIED RESEARCH</div>
13
+ </div>
14
+
15
+ <hr>
16
+
17
+ <p>We build on-device AI systems that reason, remember, and self-correct — small models designed to run autonomously at the edge with calibrated uncertainty and adversarial robustness.</p>
18
+
19
+ <h2>Research Areas</h2>
20
+
21
+ <ul>
22
+ <li><strong>Agent Memory</strong> — Encrypted semantic memory infrastructure for persistent agent context</li>
23
+ <li><strong>Adversarial Safety</strong> — Multi-agent stress-testing and trust verification for autonomous systems</li>
24
+ <li><strong>Perception Systems</strong>On-device vision, voice, and multimodal understanding</li>
25
+ <li><strong>Disposition Distillation</strong> — A three-arc study finding that imitation, attention-head tempering, and frozen-base sidecars all fail to move judge-measured disposition without damaging content quality at sub-billion scale (<a href="https://arxiv.org/abs/2604.11867">arXiv:2604.11867</a>).</li>
26
+ </ul>
27
+
28
+ <h2>Open-Source Releases</h2>
29
+
30
+ <p><strong>Tinman SmolOmni (MLA)</strong> small omnimodal models with multi-head latent attention.</p>
31
+
32
+ <table>
33
+ <thead>
34
+ <tr><th>Model</th><th>Description</th></tr>
35
+ </thead>
36
+ <tbody>
37
+ <tr><td><a href="https://huggingface.co/Tinman-Lab/Tinman-SmolOmni-MLA-256M">Tinman-SmolOmni-MLA-256M</a></td><td>256M parameter omnimodal</td></tr>
38
+ <tr><td><a href="https://huggingface.co/Tinman-Lab/Tinman-SmolOmni-MLA-500M">Tinman-SmolOmni-MLA-500M</a></td><td>500M parameter omnimodal</td></tr>
39
+ <tr><td><a href="https://huggingface.co/Tinman-Lab/Tinman-SmolOmni-MLA-Toolkit">Tinman-SmolOmni-MLA-Toolkit</a></td><td>Training and inference toolkit</td></tr>
40
+ </tbody>
41
+ </table>
42
+
43
+ <p><strong>Tinman Companion</strong> — Gemma 4 fine-tunes for on-device companion use cases.</p>
44
+
45
+ <table>
46
+ <thead>
47
+ <tr><th>Model</th><th>Description</th></tr>
48
+ </thead>
49
+ <tbody>
50
+ <tr><td><a href="https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-merged">Tinman-gemma4-companion-merged</a></td><td>Full-precision merged model</td></tr>
51
+ <tr><td><a href="https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-gguf">Tinman-gemma4-companion-gguf</a></td><td>GGUF quantized for llama.cpp</td></tr>
52
+ <tr><td><a href="https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-litert-lm">Tinman-gemma4-companion-litert-lm</a></td><td>LiteRT-LM for on-device deployment</td></tr>
53
+ <tr><td><a href="https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-sft">Tinman-gemma4-companion-sft</a></td><td>SFT checkpoint</td></tr>
54
+ <tr><td><a href="https://huggingface.co/Tinman-Lab/Tinman-gemma4-companion-dpo">Tinman-gemma4-companion-dpo</a></td><td>DPO checkpoint</td></tr>
55
+ </tbody>
56
+ </table>
57
+
58
+ <h2>Links</h2>
59
+
60
+ <p class="links">
61
+ <a href="https://tinmanlab.com" rel="nofollow">Website</a>
62
+ <a href="https://github.com/tinmanlabsl/" rel="nofollow">GitHub</a>
63
+ <a href="https://arxiv.org/abs/2604.11867" rel="nofollow">Paper (arXiv)</a>
64
+ <a href="https://github.com/tinmanlabsl/disposition-distillation" rel="nofollow">DD Artifacts</a>
65
+ </p>
66
+
67
+ <style>
68
+ body { font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, 'Helvetica Neue', Arial, sans-serif; max-width: 720px; margin: 0 auto; padding: 2rem 1.5rem; color: #1a1a2e; line-height: 1.6; background: #fff; }
69
+ .center { text-align: center; }
70
+ h1 { font-size: 2rem; margin-bottom: 0.2rem; font-weight: 700; }
71
+ h3 { font-size: 1.1rem; font-weight: 400; color: #333; margin-top: 0.2rem; margin-bottom: 0.3rem; }
72
+ .sub { font-size: 0.75rem; letter-spacing: 0.15em; color: #888; margin-bottom: 1.5rem; }
73
+ hr { border: none; border-top: 1px solid #e0e0e0; margin: 1.5rem 0; }
74
+ h2 { font-size: 1.2rem; margin-top: 1.8rem; margin-bottom: 0.8rem; }
75
+ p { margin-bottom: 0.8rem; }
76
+ ul { padding-left: 1.2rem; }
77
+ li { margin-bottom: 0.4rem; }
78
+ table { width: 100%; border-collapse: collapse; margin: 1rem 0; font-size: 0.95rem; }
79
+ th, td { text-align: left; padding: 0.5rem 0.8rem; border-bottom: 1px solid #eee; }
80
+ th { font-weight: 600; border-bottom: 2px solid #ddd; }
81
+ a { color: #2563eb; text-decoration: none; }
82
+ a:hover { text-decoration: underline; }
83
+ .links a { margin-right: 1.5rem; }
84
+ strong { font-weight: 600; }
85
+ </style>
86
+