File size: 2,343 Bytes
e9d2353
099ac1d
3eab76b
420915b
 
 
e9d2353
 
 
420915b
 
3eab76b
 
420915b
 
 
 
 
 
ac83565
61bbac7
ac01ee8
1999145
537bf72
420915b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3eab76b
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
---
title: README
emoji: 🐨
colorFrom: indigo
colorTo: purple
sdk: static
pinned: false
---

# Tripplet Research

## What Tripplet is.

An independent research collective exploring the frontier of open language models through **weight-space merging**, efficient fine-tuning, and architectural experimentation.

We focus on practical questions: how far can small-to-mid-scale models (7B–13B) be pushed on consumer hardware without sacrificing reasoning, tool use, or multilingual ability?

## Released Models

### 🚨 [Taipei 3.1](https://huggingface.co/tripplet-research/taipei3.1)
### 🎨 [Majuli 3.1](https://huggingface.co/tripplet-research/majuli3.1)
### 🧠 [Suzhou 3.1](https://huggingface.co/tripplet-research/suzhou3.1)
### ⭐ [Suzhou 3.2](https://huggingface.co/tripplet-research/suzhou3.2)
### 🤮 [Synthara Legacy](https://huggingface.co/tripplet-research/synthara-legacy)

An 8.95B SLERP merge combining agent/tool-use capabilities with a broad general-purpose base. Built on the Qwen3.5 architecture with mixed linear + full attention layers, 262K context window, and support for 29+ languages.

- **Method**: Gradient-blended SLERP across attention and MLP layers
- **Base models**: [kai-os/Carnice-9b](https://hf.co/kai-os/Carnice-9b) + [principled-intelligence/Qwen3.5-9B-text-only](https://hf.co/principled-intelligence/Qwen3.5-9B-text-only)
- **License**: Apache 2.0

## Research Directions

- **Weight-space merging** — SLERP, TIES, DARE, and layer-wise gradient blending across differently-tuned checkpoints
- **Efficient reasoning** — Chain-of-thought quality in 9B-class models with long context
- **Edge-capable agents** — Tool-using models that fit on consumer GPUs and Apple Silicon
- **Multilingual generalization** — Understanding how merging affects non-English performance

## Philosophy

We believe the most interesting capability gains at this scale come not from training larger models, but from recombining the weights of models that already know different things. Merging is cheap, repeatable, and surprisingly effective — and it deserves more rigorous study.

## License

All Tripplet Research releases are **Apache 2.0** unless explicitly noted otherwise.

---

*Tripplet Research ships under the Tripplet Models brand. Interested in collaborating? Open a discussion on any of our repos.*