han1823123123 commited on
Commit
2cf884b
·
verified ·
1 Parent(s): e1ea882

Upload training_logs/agentic_mlc_08__seed2.json with huggingface_hub

Browse files
training_logs/agentic_mlc_08__seed2.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 114.07154083251953,
4
+ 21.01778221130371,
5
+ 19.882240295410156,
6
+ 19.427143096923828,
7
+ 19.17346954345703,
8
+ 19.071523666381836,
9
+ 18.934507369995117,
10
+ 18.770261764526367,
11
+ 18.760887145996094,
12
+ 18.614261627197266,
13
+ 18.597003936767578,
14
+ 18.489173889160156,
15
+ 18.47693634033203,
16
+ 18.44523811340332,
17
+ 18.417991638183594,
18
+ 18.37946891784668
19
+ ],
20
+ "l0": [
21
+ 500.0,
22
+ 490.521484375,
23
+ 489.621337890625,
24
+ 489.908203125,
25
+ 490.220458984375,
26
+ 489.6064453125,
27
+ 490.60498046875,
28
+ 490.625732421875,
29
+ 490.88037109375,
30
+ 487.94921875,
31
+ 487.245361328125,
32
+ 489.518798828125,
33
+ 487.82470703125,
34
+ 489.97265625,
35
+ 489.576904296875,
36
+ 489.735595703125
37
+ ],
38
+ "steps_logged": [
39
+ 0,
40
+ 200,
41
+ 400,
42
+ 600,
43
+ 800,
44
+ 1000,
45
+ 1200,
46
+ 1400,
47
+ 1600,
48
+ 1800,
49
+ 2000,
50
+ 2200,
51
+ 2400,
52
+ 2600,
53
+ 2800,
54
+ 3000
55
+ ],
56
+ "final_step": 3000,
57
+ "converged": true,
58
+ "plateau_last": 0.01567208781682553,
59
+ "elapsed_s": 1637.4347159862518,
60
+ "h": 3686,
61
+ "gamma": 0.5,
62
+ "n_scales": 3,
63
+ "alpha": null,
64
+ "row": 6,
65
+ "arch_id": "agentic_mlc_08",
66
+ "arch": "agentic_mlc_08",
67
+ "group": 1,
68
+ "src_class": "MLCContrastiveMultiscale",
69
+ "src_module": "src.architectures.mlc_contrastive_multiscale",
70
+ "T": null,
71
+ "T_max": null,
72
+ "t_sample": null,
73
+ "n_layers": 5,
74
+ "k_win": 500,
75
+ "k_pos": 100,
76
+ "shifts": [
77
+ 1,
78
+ 2,
79
+ 3
80
+ ],
81
+ "seed": 2,
82
+ "d_in": 2304,
83
+ "d_sae": 18432,
84
+ "subject_model": "google/gemma-2-2b",
85
+ "anchor_layer": 12,
86
+ "mlc_layers": [
87
+ 10,
88
+ 11,
89
+ 12,
90
+ 13,
91
+ 14
92
+ ],
93
+ "phase": "phase7_unification",
94
+ "run_id": "agentic_mlc_08__seed2"
95
+ }