Upload rag_mcp_lora_config.yaml with huggingface_hub
Browse files- rag_mcp_lora_config.yaml +63 -0
rag_mcp_lora_config.yaml
ADDED
|
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
auto_config: true
|
| 2 |
+
|
| 3 |
+
hf_model_repo: "ASTERIZER/LUNA-100M"
|
| 4 |
+
hf_model_file: "sft_v1/final/model.pth"
|
| 5 |
+
pretrained_ckpt: "Base/out/input_models/luna_sft_v1/model.pth"
|
| 6 |
+
|
| 7 |
+
train_json: "Base/Datasets/rag_mcp_sft/train.json"
|
| 8 |
+
val_json: "Base/Datasets/rag_mcp_sft/val.json"
|
| 9 |
+
out_dir: "Base/out/sft/rag_mcp_lora"
|
| 10 |
+
tokenizer_dir: "Base/checkpoints/EleutherAI/pythia-160m"
|
| 11 |
+
|
| 12 |
+
model:
|
| 13 |
+
vocab_size: 50304
|
| 14 |
+
seq_len: 1024
|
| 15 |
+
n_layer: 10
|
| 16 |
+
n_embd: 768
|
| 17 |
+
n_head: 12
|
| 18 |
+
|
| 19 |
+
train:
|
| 20 |
+
epochs: 2
|
| 21 |
+
lr_warmup_steps: 100
|
| 22 |
+
save_interval: 250
|
| 23 |
+
log_interval: 10
|
| 24 |
+
eval_interval: 250
|
| 25 |
+
max_norm: 1.0
|
| 26 |
+
|
| 27 |
+
optimizer:
|
| 28 |
+
lr: 2.0e-4
|
| 29 |
+
min_lr: 2.0e-5
|
| 30 |
+
weight_decay: 0.0
|
| 31 |
+
betas: [0.9, 0.95]
|
| 32 |
+
eps: 1.0e-8
|
| 33 |
+
|
| 34 |
+
batch:
|
| 35 |
+
global_batch: 64
|
| 36 |
+
micro_batch: 8
|
| 37 |
+
grad_accum: 8
|
| 38 |
+
auto_probe_batch: true
|
| 39 |
+
probe_safety: 0.94
|
| 40 |
+
|
| 41 |
+
dataloader:
|
| 42 |
+
num_workers: 4
|
| 43 |
+
pin_memory: true
|
| 44 |
+
|
| 45 |
+
hardware:
|
| 46 |
+
precision: "bf16"
|
| 47 |
+
|
| 48 |
+
lora:
|
| 49 |
+
rank: 16
|
| 50 |
+
alpha: 32
|
| 51 |
+
dropout: 0.05
|
| 52 |
+
target_modules:
|
| 53 |
+
- "attn.c_attn"
|
| 54 |
+
- "attn.c_proj"
|
| 55 |
+
- "mlp.fc"
|
| 56 |
+
- "mlp.proj"
|
| 57 |
+
|
| 58 |
+
eval_prompts:
|
| 59 |
+
- "Explain what retrieval-augmented generation is and why teams use it."
|
| 60 |
+
- "Describe the core parts of MCP for an engineer who has never used it."
|
| 61 |
+
- "Compare RAG and MCP without confusing their roles."
|
| 62 |
+
- "What are good practices for grounding answers with retrieved context?"
|
| 63 |
+
- "How can an MCP server expose retrieval capabilities to an AI host?"
|