han1823123123 commited on
Commit
5acdc5b
·
verified ·
1 Parent(s): 0f6b2c1

Upload training_logs/agentic_mlc_08__seed1.json with huggingface_hub

Browse files
training_logs/agentic_mlc_08__seed1.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 119.20391082763672,
4
+ 20.994279861450195,
5
+ 19.80500602722168,
6
+ 19.49486541748047,
7
+ 19.16708755493164,
8
+ 19.014822006225586,
9
+ 18.84629249572754,
10
+ 18.766942977905273,
11
+ 18.680843353271484,
12
+ 18.581817626953125,
13
+ 18.56205177307129,
14
+ 18.516429901123047,
15
+ 18.45138931274414,
16
+ 18.425264358520508,
17
+ 18.345462799072266,
18
+ 18.30092430114746
19
+ ],
20
+ "l0": [
21
+ 500.0,
22
+ 490.520751953125,
23
+ 488.123779296875,
24
+ 490.2802734375,
25
+ 487.699462890625,
26
+ 490.568115234375,
27
+ 488.890869140625,
28
+ 489.780029296875,
29
+ 490.360595703125,
30
+ 490.09716796875,
31
+ 490.591064453125,
32
+ 489.838134765625,
33
+ 491.20263671875,
34
+ 490.468017578125,
35
+ 488.822509765625,
36
+ 489.115478515625
37
+ ],
38
+ "steps_logged": [
39
+ 0,
40
+ 200,
41
+ 400,
42
+ 600,
43
+ 800,
44
+ 1000,
45
+ 1200,
46
+ 1400,
47
+ 1600,
48
+ 1800,
49
+ 2000,
50
+ 2200,
51
+ 2400,
52
+ 2600,
53
+ 2800,
54
+ 3000
55
+ ],
56
+ "final_step": 3000,
57
+ "converged": true,
58
+ "plateau_last": 0.014966912061519877,
59
+ "elapsed_s": 1596.3369405269623,
60
+ "h": 3686,
61
+ "gamma": 0.5,
62
+ "n_scales": 3,
63
+ "alpha": null,
64
+ "row": 6,
65
+ "arch_id": "agentic_mlc_08",
66
+ "arch": "agentic_mlc_08",
67
+ "group": 1,
68
+ "src_class": "MLCContrastiveMultiscale",
69
+ "src_module": "src.architectures.mlc_contrastive_multiscale",
70
+ "T": null,
71
+ "T_max": null,
72
+ "t_sample": null,
73
+ "n_layers": 5,
74
+ "k_win": 500,
75
+ "k_pos": 100,
76
+ "shifts": [
77
+ 1,
78
+ 2,
79
+ 3
80
+ ],
81
+ "seed": 1,
82
+ "d_in": 2304,
83
+ "d_sae": 18432,
84
+ "subject_model": "google/gemma-2-2b",
85
+ "anchor_layer": 12,
86
+ "mlc_layers": [
87
+ 10,
88
+ 11,
89
+ 12,
90
+ 13,
91
+ 14
92
+ ],
93
+ "phase": "phase7_unification",
94
+ "run_id": "agentic_mlc_08__seed1"
95
+ }