Darwin-9B-KOREA / README.md
openfree's picture
Initial release: Darwin-9B-KOREA from PERFECT pair (Opus x NEG)
1c0db2c verified
metadata
license: apache-2.0
language:
  - ko
  - en
library_name: transformers
base_model:
  - FINAL-Bench/Darwin-9B-Opus
  - FINAL-Bench/Darwin-9B-NEG
tags:
  - darwin
  - korean
  - bilingual
  - reasoning
  - merged-model
  - evolutionary-merge
pipeline_tag: text-generation

Darwin-9B-KOREA

ํ•œ๊ตญ์–ด/์˜์–ด ์ด์ค‘์–ธ์–ด ์ถ”๋ก ์— ์ตœ์ ํ™”๋œ 9B ํŒŒ๋ผ๋ฏธํ„ฐ ๋ชจ๋ธ (28B-KOREA์˜ ๊ฒฝ๋Ÿ‰ ์ž๋งค ๋ชจ๋ธ)

VIDRAFT Darwin ์‹œ๋ฆฌ์ฆˆ์˜ PERFECT ๋ถ€๋ชจ ํŽ˜์–ด ๋จธ์ง€. Darwin-9B-Opus(์ •) ร— Darwin-9B-NEG(๋ฐ˜) ํŽ˜์–ด๋ฅผ per-layer ๋™์  ๋น„์œจ๋กœ ๊ฒฐํ•ฉํ•˜์—ฌ, 28B-KOREA ๋Œ€๋น„ 1/2.7 ์‚ฌ์ด์ฆˆ๋กœ ๊ฑฐ์˜ ๋™๋“ฑํ•œ ํ•œ๊ตญ์–ด ๋Šฅ๋ ฅ์„ ๋‹ฌ์„ฑ.

๋ถ€๋ชจ ๋ชจ๋ธ (PERFECT Pair)

Role Model Strength
Father FINAL-Bench/Darwin-9B-Opus ์˜์–ด reasoning, ํ•œ๊ตญ์–ด ๋‹ต๋ณ€
Mother FINAL-Bench/Darwin-9B-NEG Negative SFT, ์ •/๋ฐ˜ ํ•ฉ์„ฑ

ํŽ˜์–ด ํ˜ธํ™˜์„ฑ: hidden=4096, intermediate=12288, layers=32, vocab=248320 โ€” 100% ์ผ์น˜ PERFECT pair.

๋จธ์ง€ ๋ฐฉ์‹

  • ์•Œ๊ณ ๋ฆฌ์ฆ˜: Per-layer linear interpolation (float32 โ†’ bfloat16)
  • t vector: 32 ๋ ˆ์ด์–ด ๋™์  ๊ฐ€์ค‘์น˜ (mean t=0.513)
    • Golden Reasoning Layer (L0): t=0.90 (28B ํŽ˜์–ด๋Š” L47, 9B๋Š” L0์ด ์ž…๋ ฅ ์งํ›„ ๊ฐ€์žฅ reasoning critical)
    • Output Router (L31): t=0.53
  • ์ฑ— ํ…œํ”Œ๋ฆฟ/ํ† ํฌ๋‚˜์ด์ €: Father ๊ธฐ์ค€

ํ‰๊ฐ€ ๊ฒฐ๊ณผ (35-sample 3-way bench, max_tokens=5120)

ํ‰๊ฐ€ ํ•ญ๋ชฉ Father (9B-Opus) Mother (9B-NEG) KOREA (Child)
์ •ํ™•๋„ (29๊ฐœ ๊ฐ๊ด€์‹) 93.1% 93.1% 93.1%
ํ•œ๊ตญ์–ด ์ถœ๋ ฅ๋ฅ  (ํ•œ๊ตญ์–ด ์งˆ๋ฌธ 23๊ฐœ) 91.3% 87.0% 91.3% โญ

Win/Loss: Father vs Child 0:0, Mother vs Child 0:0 โ€” ์™„์ „ ๋™๋ฅ  (์ž์‹์ด ๋‘ ๋ถ€๋ชจ์™€ ๋™๊ธ‰)

28B-KOREA ๋น„๊ต: ์ •ํ™•๋„ -3.5%p (96.6% โ†’ 93.1%, ๋ชจ๋ธ ํฌ๊ธฐ ์ฐจ์ด), ํ•œ๊ตญ์–ด ์ถœ๋ ฅ๋ฅ  ๋™์ผ 91.3%.

์‚ฌ์šฉ ๊ถŒ์žฅ

  • ๊ถŒ์žฅ max_tokens: 1024 ์ด์ƒ (chain-of-thought ํŠน์„ฑ)
  • GPU ๋ฉ”๋ชจ๋ฆฌ: ~20GB BF16 (๋‹จ์ผ H100/H200/A100 80GB ์ถฉ๋ถ„)
  • ์‚ฌ๊ณ  ํŒจํ„ด: ์˜์–ด reasoning ํ›„ ํ•œ๊ตญ์–ด ๋‹ต๋ณ€. ๋‹ต ์ •ํ™•๋„ ์šฐ์„ ์ด๋ฉด ๊ถŒ์žฅ.

Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model = AutoModelForCausalLM.from_pretrained(
    "VIDraft/Darwin-9B-KOREA",
    torch_dtype=torch.bfloat16,
    device_map="auto",
    trust_remote_code=True,
)
tok = AutoTokenizer.from_pretrained("VIDraft/Darwin-9B-KOREA", trust_remote_code=True)

msgs = [{"role": "user", "content": "ํ•œ๊ตญ ์ตœ์ดˆ ํ•œ๊ธ€ ์ฐฝ์ œ ์ด์œ ๋ฅผ ํ•œ ๋ฌธ์žฅ์œผ๋กœ."}]
inputs = tok.apply_chat_template(msgs, return_tensors="pt", add_generation_prompt=True).to(model.device)
out = model.generate(inputs, max_new_tokens=1024, do_sample=False, pad_token_id=tok.eos_token_id)
print(tok.decode(out[0][inputs.shape[1]:], skip_special_tokens=True))

License

Apache 2.0 (๋ถ€๋ชจ ๋ชจ๋ธ ์ƒ์†).

Related Models


VIDRAFT Darwin Factory โ€” Pair: Darwin-9B-Opus ร— Darwin-9B-NEG โ†’ Darwin-9B-KOREA