File size: 8,969 Bytes
c37eea6
 
 
b0fe3a0
 
 
 
 
c37eea6
 
 
b0fe3a0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c37eea6
b0fe3a0
 
3403450
b0fe3a0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c37eea6
 
b0fe3a0
c37eea6
b0fe3a0
 
 
 
c37eea6
b0fe3a0
 
 
 
 
 
c37eea6
b0fe3a0
 
 
 
c37eea6
b0fe3a0
 
 
 
c37eea6
b0fe3a0
 
c37eea6
 
 
b0fe3a0
 
 
c37eea6
b0fe3a0
c37eea6
 
 
b0fe3a0
 
 
 
 
 
 
 
 
 
c37eea6
b0fe3a0
 
 
c37eea6
b0fe3a0
 
 
 
 
 
 
 
 
 
 
c37eea6
 
 
b0fe3a0
c37eea6
b0fe3a0
c37eea6
b0fe3a0
 
 
 
 
c37eea6
b0fe3a0
 
 
 
 
c37eea6
 
 
 
 
b0fe3a0
 
c37eea6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b0fe3a0
 
c37eea6
 
 
 
 
 
 
b0fe3a0
c37eea6
b0fe3a0
 
 
 
c37eea6
b0fe3a0
c37eea6
 
 
b0fe3a0
 
 
 
 
 
 
 
 
c37eea6
b0fe3a0
 
 
c37eea6
 
 
b0fe3a0
c37eea6
 
 
b0fe3a0
 
 
 
 
c37eea6
 
 
 
 
b0fe3a0
c37eea6
b0fe3a0
 
 
 
 
 
c37eea6
 
 
71d2b84
 
b0fe3a0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
---
license: apache-2.0
language:
  - en
  - zh
  - ko
  - ja
  - multilingual
library_name: transformers
pipeline_tag: text-generation
tags:
  - darwin
  - darwin-v7
  - evolutionary-merge
  - merge
  - mergekit
  - reasoning
  - advanced-reasoning
  - chain-of-thought
  - thinking
  - qwen3.6
  - qwen
  - claude-opus
  - distillation
  - multilingual
  - gpqa
  - benchmark
  - open-source
  - apache-2.0
  - hybrid-vigor
  - proto-agi
  - vidraft
  - eval-results
base_model:
  - Qwen/Qwen3.6-27B
  - rico03/Qwen3.6-27B-Claude-Opus-Reasoning-Distilled
base_model_relation: merge
model-index:
  - name: Darwin-28B-Opus
    results:
      - task:
          type: text-generation
          name: Graduate-Level Reasoning
        dataset:
          type: Idavidrein/gpqa
          name: GPQA Diamond
          config: gpqa_diamond
          split: train
        metrics:
          - type: accuracy
            value: 88.89
            name: Accuracy
            verified: false
---

# Darwin-28B-Opus β€” Qwen3.6-27B Γ— Opus-Distilled Evolutionary Merge

<p align="center">
  <a href="https://huggingface.co/FINAL-Bench/Darwin-28B-Opus"><img src="https://img.shields.io/badge/⭐_GPQA_Diamond-88.89%25_Darwin--28B--Opus-gold?style=for-the-badge" alt="GPQA"></a>
  <a href="https://huggingface.co/FINAL-Bench/Darwin-36B-Opus"><img src="https://img.shields.io/badge/🧬_Sibling-Darwin--36B--Opus_(88.4%25)-blue?style=for-the-badge" alt="36B"></a>
</p>

<p align="center">
  <a href="https://huggingface.co/FINAL-Bench/Darwin-4B-Genesis"><img src="https://img.shields.io/badge/🧬_Model-Darwin--4B--Genesis-blue?style=for-the-badge" alt="Genesis"></a>
  <a href="https://huggingface.co/FINAL-Bench/Darwin-9B-Opus"><img src="https://img.shields.io/badge/🧬_Model-Darwin--9B--Opus-blue?style=for-the-badge" alt="9B"></a>
  <a href="https://huggingface.co/FINAL-Bench/Darwin-9B-NEG"><img src="https://img.shields.io/badge/⚑_Model-Darwin--9B--NEG_(84.3%25)-purple?style=for-the-badge" alt="NEG"></a>
  <a href="https://huggingface.co/FINAL-Bench/Darwin-27B-Opus"><img src="https://img.shields.io/badge/🧬_Model-Darwin--27B--Opus_(86.9%25)-blue?style=for-the-badge" alt="27B"></a>
</p>

<p align="center">
  <a href="https://huggingface.co/FINAL-Bench/Darwin-31B-Opus"><img src="https://img.shields.io/badge/🧬_Model-Darwin--31B--Opus_(85.9%25)-blue?style=for-the-badge" alt="31B"></a>
  <a href="https://huggingface.co/FINAL-Bench/Darwin-36B-Opus"><img src="https://img.shields.io/badge/⭐_Model-Darwin--36B--Opus_(88.4%25)-blue?style=for-the-badge" alt="36B"></a>
</p>

<p align="center">
  <a href="https://huggingface.co/collections/FINAL-Bench/darwin-family"><img src="https://img.shields.io/badge/🏠_Darwin_Family-Collection-green?style=for-the-badge" alt="Family"></a>
  <a href="https://huggingface.co/spaces/FINAL-Bench/Leaderboard"><img src="https://img.shields.io/badge/πŸ†_FINAL_Bench-Leaderboard-green?style=for-the-badge" alt="FINAL Bench"></a>
</p>

> Qwen3.6-27B dense Β· 27.6B parameters Β· Hybrid Linear/Full Attention Β· BF16 Β· Thinking Mode Β· Apache 2.0
> **Darwin V7 evolutionary merge: Father Γ— Opus-distilled Mother β†’ 88.89% on GPQA Diamond (3-stage adaptive evaluation)**

---

## Abstract

**Darwin-28B-Opus** is the first reasoning model of the Darwin series built on the **Qwen3.6 generation** backbone. Produced by the Darwin V7 evolutionary breeding engine from two publicly available parents, it combines the strong bilingual reasoning of Qwen3.6-27B with Claude Opus 4-style chain-of-thought distilled behaviour.

On the **GPQA Diamond** graduate-level reasoning benchmark (198 PhD-level questions), Darwin-28B-Opus scores **88.89 %** under the standard 3-stage adaptive evaluation, slightly edging out its larger MoE sibling Darwin-36B-Opus (88.4 %) and clearly surpassing its Qwen3.5-generation counterpart Darwin-27B-Opus (86.9 %).

---

## 🧬 Model Lineage

| Role | Model | Role in the Merge |
|:---:|:---|:---|
| **Father (爢)** | [`Qwen/Qwen3.6-27B`](https://huggingface.co/Qwen/Qwen3.6-27B) | Qwen3.6 generation dense backbone with hybrid linear/full attention. |
| **Mother (母)** | [`rico03/Qwen3.6-27B-Claude-Opus-Reasoning-Distilled`](https://huggingface.co/rico03/Qwen3.6-27B-Claude-Opus-Reasoning-Distilled) | Claude Opus reasoning-distilled variant of the same backbone (Jackrong-style distillation, 14 k traces). |
| **Offspring** | **`Darwin-28B-Opus`** (this model) | Darwin V7 evolutionary merge; Qwen3.6 architecture retained, Opus reasoning style inherited. |

> **Why 28B?** The `28B` label denotes the Qwen3.6-generation member of the Darwin lineup (`+1` over the Qwen3.5-era `Darwin-27B-Opus`).
> The actual parameter count is **27.6 B**, and the architecture exactly follows Qwen3.6-27B.

---

## βš™οΈ Technical Specifications

| Component | Value |
|:---|:---|
| Architecture | `Qwen3_5ForConditionalGeneration` (Qwen3.6 generation, hybrid linear + full attention) |
| Parameters | **27.6 B** (BF16) |
| Hidden size | 5 120 |
| Intermediate size | 17 408 |
| Head dim | 256 |
| Layers | 64 (3 linear : 1 full attention, `full_attention_interval = 4`) |
| Precision | bfloat16 |
| Context length | Inherited from base (long-chain reasoning supported) |
| License | Apache 2.0 |

---

## πŸ† Benchmark β€” GPQA Diamond (198 questions)

Darwin-28B-Opus is evaluated under our standard **3-stage adaptive evaluation** protocol, identical to the protocol used across the Darwin series.

| Stage | Decoding Protocol | Cost | **Accuracy** |
|:---:|:---|:---:|:---:|
| **Stage 1** | Single-shot greedy baseline | 1Γ— | **74.75 %** (148 / 198) |
| **Stage 2** | Majority vote Γ—8 at temperature 0.7 on Stage-1 wrongs | 8Γ— | **83.84 %** (166 / 198) |
| **Stage 3** | Adaptive ensemble refinement (close-tie tiebreaker + iterative MTI on residual hard questions) | β‰ˆ 20Γ— | **πŸ₯‡ 88.89 %** (176 / 198) |

**Key performance indicators**:
- Stage 1 β†’ Stage 3: **+14.14 %p** through adaptive protocol
- vs Darwin-27B-Opus (86.9 %): **+1.99 %p**
- vs Darwin-36B-Opus (88.4 %): **+0.49 %p**
- vs Darwin-31B-Opus (85.9 %): **+2.99 %p**

---

## πŸš€ Usage

### Standard inference (Stage 1 baseline)

```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

tok = AutoTokenizer.from_pretrained(
    "FINAL-Bench/Darwin-28B-Opus",
    trust_remote_code=True,
)
model = AutoModelForCausalLM.from_pretrained(
    "FINAL-Bench/Darwin-28B-Opus",
    torch_dtype=torch.bfloat16,
    device_map="auto",
    trust_remote_code=True,
)

messages = [
    {"role": "user",
     "content": "Solve: If f(x) = xΒ³ βˆ’ 3x + 2, find all critical points and classify them."}
]
text = tok.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tok(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=2048, do_sample=False)
print(tok.decode(outputs[0][inputs.input_ids.shape[-1]:], skip_special_tokens=True))
```

### Enhanced accuracy (Stage 2-3 adaptive)

For leaderboard-grade accuracy, combine:
1. Stage 1 greedy baseline,
2. Stage 2 maj@8 temperature sampling on low-confidence answers,
3. Stage 3 adaptive refinement on still-disputed answers.

Reference implementation is provided in the Darwin-series evaluation harness.

---

## 🎯 Recommended Use-Cases

- **Graduate-level STEM reasoning** (GPQA / science qualifying exams)
- **Mathematical problem solving** (MATH, AIME-style problems)
- **Code generation and debugging** (HumanEval, MBPP)
- **Complex multi-step chain-of-thought tasks**
- **Bilingual reasoning** (strong English + Korean; also Chinese / Japanese)

## ⚠️ Limitations

- At 27.6 B parameters in bfloat16, full inference requires β‰ˆ 55 GB of VRAM (e.g., a single A100-80GB or B200).
- Optimised for English first, with secondary support for Korean, Chinese, and Japanese.
- Deep Opus-style reasoning traces tend to be verbose β€” control with `max_new_tokens` as needed.

---

## πŸ“š Citation

```bibtex
@misc{darwin28b_opus_2026,
  title  = {Darwin-28B-Opus: Evolutionary Merging of Qwen3.6-27B with Claude-Opus-Distilled Reasoning},
  author = {FINAL-Bench / Darwin Research Team},
  year   = {2026},
  howpublished = {\url{https://huggingface.co/FINAL-Bench/Darwin-28B-Opus}},
  note   = {Darwin V7 Β· Mother-centric Ratio Interpolation merge Β· 88.89 % GPQA Diamond (3-stage)}
}
```

---

## πŸ”— Related Darwin Models

- **Darwin-36B-Opus** β€” MoE 36B, Qwen3.6-35B-A3B Γ— Opus distilled, GPQA 88.4 %
- **Darwin-31B-Opus** β€” 31B dense, multilingual-strong reasoning, GPQA 85.9 %
- **Darwin-27B-Opus** β€” 27B dense (Qwen3.5 generation), GPQA 86.9 %
- **Darwin-9B-NEG** β€” 9B with Native Entropy Gating, GPQA 84.3 %
- **Darwin-9B-Opus** β€” the Qwen3.5-9B Darwin member
- **Darwin-4B-Genesis** β€” smallest Darwin member

---

This model is introduced in [Darwin Family](https://arxiv.org/abs/2605.14386).

*Darwin V7 Β· Qwen3.6 generation flagship Β· Sealed 2026-04-25 Β· FINAL-Bench*