Darwin-9B-KOREA / README.md
openfree's picture
Initial release: Darwin-9B-KOREA from PERFECT pair (Opus x NEG)
1c0db2c verified
---
license: apache-2.0
language:
- ko
- en
library_name: transformers
base_model:
- FINAL-Bench/Darwin-9B-Opus
- FINAL-Bench/Darwin-9B-NEG
tags:
- darwin
- korean
- bilingual
- reasoning
- merged-model
- evolutionary-merge
pipeline_tag: text-generation
---
# Darwin-9B-KOREA
**ํ•œ๊ตญ์–ด/์˜์–ด ์ด์ค‘์–ธ์–ด ์ถ”๋ก ์— ์ตœ์ ํ™”๋œ 9B ํŒŒ๋ผ๋ฏธํ„ฐ ๋ชจ๋ธ (28B-KOREA์˜ ๊ฒฝ๋Ÿ‰ ์ž๋งค ๋ชจ๋ธ)**
VIDRAFT Darwin ์‹œ๋ฆฌ์ฆˆ์˜ PERFECT ๋ถ€๋ชจ ํŽ˜์–ด ๋จธ์ง€. Darwin-9B-Opus(์ •) ร— Darwin-9B-NEG(๋ฐ˜) ํŽ˜์–ด๋ฅผ per-layer ๋™์  ๋น„์œจ๋กœ ๊ฒฐํ•ฉํ•˜์—ฌ, 28B-KOREA ๋Œ€๋น„ 1/2.7 ์‚ฌ์ด์ฆˆ๋กœ ๊ฑฐ์˜ ๋™๋“ฑํ•œ ํ•œ๊ตญ์–ด ๋Šฅ๋ ฅ์„ ๋‹ฌ์„ฑ.
## ๋ถ€๋ชจ ๋ชจ๋ธ (PERFECT Pair)
| Role | Model | Strength |
|------|-------|----------|
| Father | [FINAL-Bench/Darwin-9B-Opus](https://huggingface.co/FINAL-Bench/Darwin-9B-Opus) | ์˜์–ด reasoning, ํ•œ๊ตญ์–ด ๋‹ต๋ณ€ |
| Mother | [FINAL-Bench/Darwin-9B-NEG](https://huggingface.co/FINAL-Bench/Darwin-9B-NEG) | Negative SFT, ์ •/๋ฐ˜ ํ•ฉ์„ฑ |
ํŽ˜์–ด ํ˜ธํ™˜์„ฑ: hidden=4096, intermediate=12288, layers=32, vocab=248320 โ€” 100% ์ผ์น˜ PERFECT pair.
## ๋จธ์ง€ ๋ฐฉ์‹
- **์•Œ๊ณ ๋ฆฌ์ฆ˜**: Per-layer linear interpolation (float32 โ†’ bfloat16)
- **t vector**: 32 ๋ ˆ์ด์–ด ๋™์  ๊ฐ€์ค‘์น˜ (mean t=0.513)
- Golden Reasoning Layer (L0): t=0.90 (28B ํŽ˜์–ด๋Š” L47, 9B๋Š” L0์ด ์ž…๋ ฅ ์งํ›„ ๊ฐ€์žฅ reasoning critical)
- Output Router (L31): t=0.53
- **์ฑ— ํ…œํ”Œ๋ฆฟ/ํ† ํฌ๋‚˜์ด์ €**: Father ๊ธฐ์ค€
## ํ‰๊ฐ€ ๊ฒฐ๊ณผ (35-sample 3-way bench, max_tokens=5120)
| ํ‰๊ฐ€ ํ•ญ๋ชฉ | Father (9B-Opus) | Mother (9B-NEG) | **KOREA (Child)** |
|---|---|---|---|
| ์ •ํ™•๋„ (29๊ฐœ ๊ฐ๊ด€์‹) | 93.1% | 93.1% | **93.1%** |
| ํ•œ๊ตญ์–ด ์ถœ๋ ฅ๋ฅ  (ํ•œ๊ตญ์–ด ์งˆ๋ฌธ 23๊ฐœ) | 91.3% | 87.0% | **91.3%** โญ |
**Win/Loss**: Father vs Child 0:0, Mother vs Child 0:0 โ€” **์™„์ „ ๋™๋ฅ ** (์ž์‹์ด ๋‘ ๋ถ€๋ชจ์™€ ๋™๊ธ‰)
**28B-KOREA ๋น„๊ต**: ์ •ํ™•๋„ -3.5%p (96.6% โ†’ 93.1%, ๋ชจ๋ธ ํฌ๊ธฐ ์ฐจ์ด), **ํ•œ๊ตญ์–ด ์ถœ๋ ฅ๋ฅ  ๋™์ผ 91.3%**.
## ์‚ฌ์šฉ ๊ถŒ์žฅ
- **๊ถŒ์žฅ max_tokens**: 1024 ์ด์ƒ (chain-of-thought ํŠน์„ฑ)
- **GPU ๋ฉ”๋ชจ๋ฆฌ**: ~20GB BF16 (๋‹จ์ผ H100/H200/A100 80GB ์ถฉ๋ถ„)
- **์‚ฌ๊ณ  ํŒจํ„ด**: ์˜์–ด reasoning ํ›„ ํ•œ๊ตญ์–ด ๋‹ต๋ณ€. ๋‹ต ์ •ํ™•๋„ ์šฐ์„ ์ด๋ฉด ๊ถŒ์žฅ.
## Quick Start
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model = AutoModelForCausalLM.from_pretrained(
"VIDraft/Darwin-9B-KOREA",
torch_dtype=torch.bfloat16,
device_map="auto",
trust_remote_code=True,
)
tok = AutoTokenizer.from_pretrained("VIDraft/Darwin-9B-KOREA", trust_remote_code=True)
msgs = [{"role": "user", "content": "ํ•œ๊ตญ ์ตœ์ดˆ ํ•œ๊ธ€ ์ฐฝ์ œ ์ด์œ ๋ฅผ ํ•œ ๋ฌธ์žฅ์œผ๋กœ."}]
inputs = tok.apply_chat_template(msgs, return_tensors="pt", add_generation_prompt=True).to(model.device)
out = model.generate(inputs, max_new_tokens=1024, do_sample=False, pad_token_id=tok.eos_token_id)
print(tok.decode(out[0][inputs.shape[1]:], skip_special_tokens=True))
```
## License
Apache 2.0 (๋ถ€๋ชจ ๋ชจ๋ธ ์ƒ์†).
## Related Models
- **[VIDraft/Darwin-28B-KOREA](https://huggingface.co/VIDraft/Darwin-28B-KOREA)** โ€” ํฐ ์ž๋งค ๋ชจ๋ธ (52GB, 96.6% ์ •ํ™•๋„)
---
*VIDRAFT Darwin Factory โ€” Pair: Darwin-9B-Opus ร— Darwin-9B-NEG โ†’ Darwin-9B-KOREA*