ASTERIZER commited on
Commit
a783c2e
·
verified ·
1 Parent(s): 8e239f6

Upload rag_mcp_lora_config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. rag_mcp_lora_config.yaml +63 -0
rag_mcp_lora_config.yaml ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ auto_config: true
2
+
3
+ hf_model_repo: "ASTERIZER/LUNA-100M"
4
+ hf_model_file: "sft_v1/final/model.pth"
5
+ pretrained_ckpt: "Base/out/input_models/luna_sft_v1/model.pth"
6
+
7
+ train_json: "Base/Datasets/rag_mcp_sft/train.json"
8
+ val_json: "Base/Datasets/rag_mcp_sft/val.json"
9
+ out_dir: "Base/out/sft/rag_mcp_lora"
10
+ tokenizer_dir: "Base/checkpoints/EleutherAI/pythia-160m"
11
+
12
+ model:
13
+ vocab_size: 50304
14
+ seq_len: 1024
15
+ n_layer: 10
16
+ n_embd: 768
17
+ n_head: 12
18
+
19
+ train:
20
+ epochs: 2
21
+ lr_warmup_steps: 100
22
+ save_interval: 250
23
+ log_interval: 10
24
+ eval_interval: 250
25
+ max_norm: 1.0
26
+
27
+ optimizer:
28
+ lr: 2.0e-4
29
+ min_lr: 2.0e-5
30
+ weight_decay: 0.0
31
+ betas: [0.9, 0.95]
32
+ eps: 1.0e-8
33
+
34
+ batch:
35
+ global_batch: 64
36
+ micro_batch: 8
37
+ grad_accum: 8
38
+ auto_probe_batch: true
39
+ probe_safety: 0.94
40
+
41
+ dataloader:
42
+ num_workers: 4
43
+ pin_memory: true
44
+
45
+ hardware:
46
+ precision: "bf16"
47
+
48
+ lora:
49
+ rank: 16
50
+ alpha: 32
51
+ dropout: 0.05
52
+ target_modules:
53
+ - "attn.c_attn"
54
+ - "attn.c_proj"
55
+ - "mlp.fc"
56
+ - "mlp.proj"
57
+
58
+ eval_prompts:
59
+ - "Explain what retrieval-augmented generation is and why teams use it."
60
+ - "Describe the core parts of MCP for an engineer who has never used it."
61
+ - "Compare RAG and MCP without confusing their roles."
62
+ - "What are good practices for grounding answers with retrieved context?"
63
+ - "How can an MCP server expose retrieval capabilities to an AI host?"