Rogue-28B-MIX / README.md
ginigen-ai's picture
fix lineage: Qwen-3.6-27B (great-gf) / Darwin-3.6-28B (gf) / Darwin-28B-KR (father), narrow boxes, no trade secrets
8f9b2f1 verified
---
license: apache-2.0
language:
- ko
- en
library_name: transformers
tags:
- korean
- reasoning
- multimodal
- mix
base_model:
- FINAL-Bench/Darwin-28B-KR
- NewenAI/QuettaLLMs-27B-Koreasoner-V3
---
# Rogue-28B-MIX
ν•œκ΅­μ–΄ reasoning + multimodal mix λͺ¨λΈ.
---
## πŸ›οΈ κ°€λ¬Έ 계보
```
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚ 증쑰뢀 (Great-Grandfather) β”‚
β”‚ Qwen-3.6-27B β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
β”‚
β–Ό
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚ μ‘°λΆ€ (Grandfather) β”‚
β”‚ Darwin-3.6-28B β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
β”‚
β–Ό
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚ μ•„λΉ  (Father) β”‚
β”‚ FINAL-Bench/Darwin-28B-KR β”‚
β”‚ - ν•œκ΅­μ–΄ νŠΉν™” reasoning λͺ¨λΈ β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
β”‚
Γ—Γ— ꡐ배 Γ—Γ—
β”‚
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚ μ—„λ§ˆ (Mother) β”‚
β”‚ NewenAI/QuettaLLMs-27B-Koreasoner-V3 β”‚
β”‚ - K-AI Leaderboard 1μœ„ β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
β”‚
β–Ό
╔══════════════════════════════════════════╗
β•‘ μžμ‹ (Child) β€” λ³Έ λͺ¨λΈ β•‘
β•‘ ginigen-ai/Rogue-28B-MIX β•‘
β•‘ β•‘
β•‘ - μΉœκ°€μ˜ reasoning κ³„μŠΉ β•‘
β•‘ - μ™Έκ°€μ˜ ν•œκ΅­μ–΄ K-AI 지식 κ³„μŠΉ β•‘
β•‘ - <think> μΆ”λ‘  트레이슀 보쑴 β•‘
β•‘ - λ©€ν‹°λͺ¨λ‹¬ ν—€λ“œ 보쑴 β•‘
β•šβ•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•β•
```
---
## πŸŽ“ ν•™μŠ΅ κ°œμš”
1. μΉœκ°€ Γ— μ™Έκ°€ λͺ¨λΈ κ°€μ€‘μΉ˜ λ¨Έμ§€
2. K-AI 도메인 λ°μ΄ν„°λ‘œ μΆ”κ°€ SFT
3. K-AI Leaderboard Docker ν˜Έν™˜ ν˜•μ‹ μ •λΉ„
---
## πŸ“Š 평가
ν•œκ΅­μ–΄ 곡개 10 데이터셋, 100문제 Γ— 1 seed.
| Dataset | Rogue-28B-MIX | μ—„λ§ˆ(Quetta) |
|:---|---:|---:|
| CLIcK | 84% | 85% |
| KMMLU History | **48%** πŸ† | 45% |
| KMMLU Law | 25% | 26% |
| KMMLU Health | **81%** πŸ† | 80% |
| HAERAE GK | 63% | 66% |
| HAERAE History | 89% | 90% |
| HAERAE Linguistics | 90% | 95% |
| KoBEST Hellaswag | 95% | 97% |
| KoBEST COPA | 98% | 99% |
| KoBEST BoolQ | 97% | 97% |
| **Macro Avg** | **77.0%** | **78.0%** |
K-AI Leaderboard 핡심 μΉ΄ν…Œκ³ λ¦¬(μ˜λ£ŒΒ·μ—­μ‚¬)μ—μ„œ μ—„λ§ˆ μΆ”μ›”.
---
## 🎯 μ‚¬μš©λ²•
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_id = "ginigen-ai/Rogue-28B-MIX"
tokenizer = AutoTokenizer.from_pretrained(
model_id, trust_remote_code=True
)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto",
trust_remote_code=True,
)
prompt = "ν•œκ΅­μ˜ 좔석에 λŒ€ν•΄ μ„€λͺ…ν•΄μ£Όμ„Έμš”."
messages = [{"role": "user", "content": prompt}]
inputs = tokenizer.apply_chat_template(
messages, return_tensors="pt", add_generation_prompt=True
)
out = model.generate(
inputs.to(model.device),
max_new_tokens=512,
do_sample=False,
)
print(tokenizer.decode(out[0], skip_special_tokens=False))
```
---
## πŸ› οΈ 사양
- νŒŒλΌλ―Έν„°: 28B (multimodal)
- μ–‘μžν™”: bf16
- μ»¨ν…μŠ€νŠΈ: 8K (ν™•μž₯ κ°€λŠ₯)
- μ–Έμ–΄: ν•œκ΅­μ–΄ + μ˜μ–΄
- μΆ”λ‘ : `<think>` reasoning trace
- License: Apache 2.0
---
## 🀝 좜처
- μ•„λΉ : [FINAL-Bench/Darwin-28B-KR](https://huggingface.co/FINAL-Bench/Darwin-28B-KR)
- μ—„λ§ˆ: [NewenAI/QuettaLLMs-27B-Koreasoner-V3](https://huggingface.co/NewenAI/QuettaLLMs-27B-Koreasoner-V3)