Upload sft_config.yaml with huggingface_hub
Browse files- sft_config.yaml +69 -0
sft_config.yaml
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
|
| 2 |
+
# LUNA 100M β SFT Fine-Tuning Configuration
|
| 3 |
+
# βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
|
| 4 |
+
|
| 5 |
+
auto_config: true
|
| 6 |
+
|
| 7 |
+
# ββ Paths βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
|
| 8 |
+
hf_model_repo: "ASTERIZER/LUNA-100M" # HF model repo with pretrained latest.pt
|
| 9 |
+
hf_dataset_repo: "ASTERIZER/Luna_Dataset" # HF dataset repo with SFT data
|
| 10 |
+
pretrained_ckpt: "Base/out/pretrain/luna_100m/latest.pt" # local path (auto-downloaded)
|
| 11 |
+
train_json: "Base/Datasets/sft_clean/train.json"
|
| 12 |
+
val_json: "Base/Datasets/sft_clean/val.json"
|
| 13 |
+
out_dir: "Base/out/sft/luna_100m_sft"
|
| 14 |
+
tokenizer_dir: "Base/checkpoints/EleutherAI/pythia-160m"
|
| 15 |
+
|
| 16 |
+
# ββ Model (must match pretrained β do NOT change) ββββββββββββββββββββββββββββ
|
| 17 |
+
model:
|
| 18 |
+
vocab_size: 50304
|
| 19 |
+
seq_len: 1024
|
| 20 |
+
n_layer: 10
|
| 21 |
+
n_embd: 768
|
| 22 |
+
n_head: 12
|
| 23 |
+
|
| 24 |
+
# ββ Training ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
|
| 25 |
+
train:
|
| 26 |
+
epochs: 2
|
| 27 |
+
max_tokens: 0 # 0 = use epochs instead
|
| 28 |
+
lr_warmup_steps: 200
|
| 29 |
+
save_interval: 500
|
| 30 |
+
log_interval: 10
|
| 31 |
+
eval_interval: 500 # run validation + eval prompts every N steps
|
| 32 |
+
max_norm: 1.0
|
| 33 |
+
|
| 34 |
+
# ββ Optimizer (lower LR for fine-tuning) ββββββββββββββββββββββββββββββββββββββ
|
| 35 |
+
optimizer:
|
| 36 |
+
lr: 1.5e-5
|
| 37 |
+
min_lr: 1.0e-6
|
| 38 |
+
weight_decay: 0.01
|
| 39 |
+
betas: [0.9, 0.95]
|
| 40 |
+
eps: 1.0e-8
|
| 41 |
+
|
| 42 |
+
# ββ Batch sizing ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
|
| 43 |
+
batch:
|
| 44 |
+
global_batch: 64
|
| 45 |
+
micro_batch: 8
|
| 46 |
+
grad_accum: 8
|
| 47 |
+
|
| 48 |
+
# ββ DataLoader ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
|
| 49 |
+
dataloader:
|
| 50 |
+
num_workers: 4
|
| 51 |
+
pin_memory: true
|
| 52 |
+
|
| 53 |
+
# ββ Hardware ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
|
| 54 |
+
hardware:
|
| 55 |
+
precision: "bf16"
|
| 56 |
+
compile: false
|
| 57 |
+
|
| 58 |
+
# ββ Eval prompts (run after each checkpoint) ββββββββββββββββββββββββββββββββββ
|
| 59 |
+
eval_prompts:
|
| 60 |
+
- "Who are you?"
|
| 61 |
+
- "What is your name?"
|
| 62 |
+
- "Who created you?"
|
| 63 |
+
- "Who is Asterizer?"
|
| 64 |
+
- "What is the capital of France?"
|
| 65 |
+
- "Explain photosynthesis in simple terms."
|
| 66 |
+
- "Write a short poem about the moon."
|
| 67 |
+
- "What is 25 times 4?"
|
| 68 |
+
- "Translate 'hello' to Spanish."
|
| 69 |
+
- "Summarize the theory of evolution in two sentences."
|