Jeff Moe commited on
Commit
6748efb
·
1 Parent(s): ba42d51

Training configuration

Browse files
Files changed (1) hide show
  1. training_config.yaml +58 -0
training_config.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # LibreHPS-4B training config — frozen for the v1.1 weights release.
2
+ # This is an archive of the harness config used to produce the
3
+ # released weights and is provided for reproducibility. It is NOT the
4
+ # config you would use for new training runs (paths point inside the
5
+ # original training tree).
6
+ model:
7
+ name_or_path: "Qwen/Qwen3.5-4B-Base"
8
+ is_moe: false
9
+ reward_token: "<|Reward|>"
10
+ dropout: 0.05
11
+
12
+ data:
13
+ index_parquet: "librehps/dataset/index.parquet"
14
+ num_video_frames: 6
15
+ modalities: null # all of image | video | scalar_image | scalar_video
16
+ max_length: 2048
17
+
18
+ optimizer:
19
+ visual_lr: 2.0e-6
20
+ merger_lr: 2.0e-6
21
+ rm_head_lr: 2.0e-6
22
+ special_token_lr: 2.0e-6
23
+ weight_decay: 0.0
24
+ betas: [0.9, 0.95]
25
+
26
+ scheduler:
27
+ kind: "constant_with_warmup"
28
+ warmup_ratio: 0.05
29
+
30
+ training:
31
+ global_batch_size: 384
32
+ per_gpu_micro_batch: 12
33
+ gradient_accumulation: 4
34
+ epochs: 1
35
+ lambda_reg: 0.1
36
+ lambda_router: 0.0 # dense — router loss inactive
37
+ seed: 42
38
+ # Cap optimiser steps for bounded runs. null = run to ``epochs``
39
+ # completion. Overridable via ``--smoke-steps N`` CLI or
40
+ # ``LIBREHPS_SMOKE_STEPS=N`` env var.
41
+ max_train_steps: null
42
+
43
+ precision:
44
+ kind: "bf16"
45
+
46
+ deepspeed:
47
+ config_path: "librehps/config/ds_config/zero3.json"
48
+
49
+ fa4:
50
+ cache_enabled: true
51
+ cache_dir: "/home/train/.cache/flash_attn_cute_dsl"
52
+
53
+ logging:
54
+ every_steps: 10
55
+
56
+ checkpoint:
57
+ dir: "checkpoints/LibreHPS_4B"
58
+ every_steps: 500