File size: 3,208 Bytes
1c0db2c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
---
license: apache-2.0
language:
- ko
- en
library_name: transformers
base_model:
- FINAL-Bench/Darwin-9B-Opus
- FINAL-Bench/Darwin-9B-NEG
tags:
- darwin
- korean
- bilingual
- reasoning
- merged-model
- evolutionary-merge
pipeline_tag: text-generation
---

# Darwin-9B-KOREA

**ํ•œ๊ตญ์–ด/์˜์–ด ์ด์ค‘์–ธ์–ด ์ถ”๋ก ์— ์ตœ์ ํ™”๋œ 9B ํŒŒ๋ผ๋ฏธํ„ฐ ๋ชจ๋ธ (28B-KOREA์˜ ๊ฒฝ๋Ÿ‰ ์ž๋งค ๋ชจ๋ธ)**

VIDRAFT Darwin ์‹œ๋ฆฌ์ฆˆ์˜ PERFECT ๋ถ€๋ชจ ํŽ˜์–ด ๋จธ์ง€. Darwin-9B-Opus(์ •) ร— Darwin-9B-NEG(๋ฐ˜) ํŽ˜์–ด๋ฅผ per-layer ๋™์  ๋น„์œจ๋กœ ๊ฒฐํ•ฉํ•˜์—ฌ, 28B-KOREA ๋Œ€๋น„ 1/2.7 ์‚ฌ์ด์ฆˆ๋กœ ๊ฑฐ์˜ ๋™๋“ฑํ•œ ํ•œ๊ตญ์–ด ๋Šฅ๋ ฅ์„ ๋‹ฌ์„ฑ.

## ๋ถ€๋ชจ ๋ชจ๋ธ (PERFECT Pair)

| Role | Model | Strength |
|------|-------|----------|
| Father | [FINAL-Bench/Darwin-9B-Opus](https://huggingface.co/FINAL-Bench/Darwin-9B-Opus) | ์˜์–ด reasoning, ํ•œ๊ตญ์–ด ๋‹ต๋ณ€ |
| Mother | [FINAL-Bench/Darwin-9B-NEG](https://huggingface.co/FINAL-Bench/Darwin-9B-NEG) | Negative SFT, ์ •/๋ฐ˜ ํ•ฉ์„ฑ |

ํŽ˜์–ด ํ˜ธํ™˜์„ฑ: hidden=4096, intermediate=12288, layers=32, vocab=248320 โ€” 100% ์ผ์น˜ PERFECT pair.

## ๋จธ์ง€ ๋ฐฉ์‹

- **์•Œ๊ณ ๋ฆฌ์ฆ˜**: Per-layer linear interpolation (float32 โ†’ bfloat16)
- **t vector**: 32 ๋ ˆ์ด์–ด ๋™์  ๊ฐ€์ค‘์น˜ (mean t=0.513)
  - Golden Reasoning Layer (L0): t=0.90 (28B ํŽ˜์–ด๋Š” L47, 9B๋Š” L0์ด ์ž…๋ ฅ ์งํ›„ ๊ฐ€์žฅ reasoning critical)
  - Output Router (L31): t=0.53
- **์ฑ— ํ…œํ”Œ๋ฆฟ/ํ† ํฌ๋‚˜์ด์ €**: Father ๊ธฐ์ค€

## ํ‰๊ฐ€ ๊ฒฐ๊ณผ (35-sample 3-way bench, max_tokens=5120)

| ํ‰๊ฐ€ ํ•ญ๋ชฉ | Father (9B-Opus) | Mother (9B-NEG) | **KOREA (Child)** |
|---|---|---|---|
| ์ •ํ™•๋„ (29๊ฐœ ๊ฐ๊ด€์‹) | 93.1% | 93.1% | **93.1%** |
| ํ•œ๊ตญ์–ด ์ถœ๋ ฅ๋ฅ  (ํ•œ๊ตญ์–ด ์งˆ๋ฌธ 23๊ฐœ) | 91.3% | 87.0% | **91.3%** โญ |

**Win/Loss**: Father vs Child 0:0, Mother vs Child 0:0 โ€” **์™„์ „ ๋™๋ฅ ** (์ž์‹์ด ๋‘ ๋ถ€๋ชจ์™€ ๋™๊ธ‰)

**28B-KOREA ๋น„๊ต**: ์ •ํ™•๋„ -3.5%p (96.6% โ†’ 93.1%, ๋ชจ๋ธ ํฌ๊ธฐ ์ฐจ์ด), **ํ•œ๊ตญ์–ด ์ถœ๋ ฅ๋ฅ  ๋™์ผ 91.3%**.

## ์‚ฌ์šฉ ๊ถŒ์žฅ

- **๊ถŒ์žฅ max_tokens**: 1024 ์ด์ƒ (chain-of-thought ํŠน์„ฑ)
- **GPU ๋ฉ”๋ชจ๋ฆฌ**: ~20GB BF16 (๋‹จ์ผ H100/H200/A100 80GB ์ถฉ๋ถ„)
- **์‚ฌ๊ณ  ํŒจํ„ด**: ์˜์–ด reasoning ํ›„ ํ•œ๊ตญ์–ด ๋‹ต๋ณ€. ๋‹ต ์ •ํ™•๋„ ์šฐ์„ ์ด๋ฉด ๊ถŒ์žฅ.

## Quick Start

```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model = AutoModelForCausalLM.from_pretrained(
    "VIDraft/Darwin-9B-KOREA",
    torch_dtype=torch.bfloat16,
    device_map="auto",
    trust_remote_code=True,
)
tok = AutoTokenizer.from_pretrained("VIDraft/Darwin-9B-KOREA", trust_remote_code=True)

msgs = [{"role": "user", "content": "ํ•œ๊ตญ ์ตœ์ดˆ ํ•œ๊ธ€ ์ฐฝ์ œ ์ด์œ ๋ฅผ ํ•œ ๋ฌธ์žฅ์œผ๋กœ."}]
inputs = tok.apply_chat_template(msgs, return_tensors="pt", add_generation_prompt=True).to(model.device)
out = model.generate(inputs, max_new_tokens=1024, do_sample=False, pad_token_id=tok.eos_token_id)
print(tok.decode(out[0][inputs.shape[1]:], skip_special_tokens=True))
```

## License

Apache 2.0 (๋ถ€๋ชจ ๋ชจ๋ธ ์ƒ์†).

## Related Models

- **[VIDraft/Darwin-28B-KOREA](https://huggingface.co/VIDraft/Darwin-28B-KOREA)** โ€” ํฐ ์ž๋งค ๋ชจ๋ธ (52GB, 96.6% ์ •ํ™•๋„)

---

*VIDRAFT Darwin Factory โ€” Pair: Darwin-9B-Opus ร— Darwin-9B-NEG โ†’ Darwin-9B-KOREA*