han1823123123 commited on
Commit
ef8b238
·
verified ·
1 Parent(s): 21d9271

Upload training_logs/it_topk_sae__seed42.json with huggingface_hub

Browse files
training_logs/it_topk_sae__seed42.json ADDED
@@ -0,0 +1,135 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 330472.625,
4
+ 8282.25,
5
+ 5019.365234375,
6
+ 4627.8935546875,
7
+ 4344.2958984375,
8
+ 4165.61767578125,
9
+ 4061.419921875,
10
+ 3949.7724609375,
11
+ 3856.04345703125,
12
+ 3789.70751953125,
13
+ 3722.8681640625,
14
+ 3660.104736328125,
15
+ 3587.613525390625,
16
+ 3542.554443359375,
17
+ 3522.66162109375,
18
+ 3441.0146484375,
19
+ 3461.027099609375,
20
+ 3424.27490234375,
21
+ 3358.95263671875,
22
+ 3321.433837890625,
23
+ 3334.40966796875,
24
+ 3294.07080078125,
25
+ 3255.5068359375,
26
+ 3267.88232421875,
27
+ 3261.734619140625,
28
+ 3204.3544921875,
29
+ 3187.548828125,
30
+ 3197.526123046875,
31
+ 3196.59716796875,
32
+ 3219.7607421875,
33
+ 3181.529052734375
34
+ ],
35
+ "l0": [
36
+ 500.0,
37
+ 500.0,
38
+ 500.0,
39
+ 500.0,
40
+ 500.0,
41
+ 500.0,
42
+ 500.0,
43
+ 500.0,
44
+ 500.0,
45
+ 500.0,
46
+ 500.0,
47
+ 500.0,
48
+ 500.0,
49
+ 500.0,
50
+ 500.0,
51
+ 500.0,
52
+ 499.912109375,
53
+ 499.718017578125,
54
+ 499.6640625,
55
+ 499.5517578125,
56
+ 499.44384765625,
57
+ 499.33203125,
58
+ 499.3232421875,
59
+ 499.622314453125,
60
+ 499.333984375,
61
+ 499.10009765625,
62
+ 499.09423828125,
63
+ 498.9716796875,
64
+ 499.114501953125,
65
+ 499.11376953125,
66
+ 499.015625
67
+ ],
68
+ "steps_logged": [
69
+ 0,
70
+ 200,
71
+ 400,
72
+ 600,
73
+ 800,
74
+ 1000,
75
+ 1200,
76
+ 1400,
77
+ 1600,
78
+ 1800,
79
+ 2000,
80
+ 2200,
81
+ 2400,
82
+ 2600,
83
+ 2800,
84
+ 3000,
85
+ 3200,
86
+ 3400,
87
+ 3600,
88
+ 3800,
89
+ 4000,
90
+ 4200,
91
+ 4400,
92
+ 4600,
93
+ 4800,
94
+ 5000,
95
+ 5200,
96
+ 5400,
97
+ 5600,
98
+ 5800,
99
+ 6000
100
+ ],
101
+ "final_step": 6000,
102
+ "converged": true,
103
+ "plateau_last": 0.018459560435451345,
104
+ "elapsed_s": 707.0378153324127,
105
+ "row": 1,
106
+ "arch_id": "topk_sae",
107
+ "arch": "topk_sae",
108
+ "group": 1,
109
+ "src_class": "TopKSAE",
110
+ "src_module": "src.architectures.topk_sae",
111
+ "T": 1,
112
+ "T_max": null,
113
+ "t_sample": null,
114
+ "n_layers": null,
115
+ "k_win": 500,
116
+ "k_pos": 500,
117
+ "shifts": null,
118
+ "alpha": null,
119
+ "gamma": null,
120
+ "n_scales": null,
121
+ "seed": 42,
122
+ "d_in": 2304,
123
+ "d_sae": 18432,
124
+ "subject_model": "google/gemma-2-2b-it",
125
+ "anchor_layer": 13,
126
+ "mlc_layers": [
127
+ 11,
128
+ 12,
129
+ 13,
130
+ 14,
131
+ 15
132
+ ],
133
+ "phase": "phase7_unification",
134
+ "run_id": "it_topk_sae__seed42"
135
+ }