han1823123123 commited on
Commit
5d0de7a
·
verified ·
1 Parent(s): c982f87

Upload training_logs/it_txc_bare_antidead_t5__seed42.json with huggingface_hub

Browse files
training_logs/it_txc_bare_antidead_t5__seed42.json ADDED
@@ -0,0 +1,117 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 33886.93359375,
4
+ 9228.7294921875,
5
+ 8243.6083984375,
6
+ 7657.58740234375,
7
+ 7218.82666015625,
8
+ 6937.9375,
9
+ 6732.0859375,
10
+ 6602.2431640625,
11
+ 6442.50146484375,
12
+ 6351.95263671875,
13
+ 6310.92822265625,
14
+ 6234.79248046875,
15
+ 6191.4921875,
16
+ 6122.57421875,
17
+ 6067.0712890625,
18
+ 6071.34228515625,
19
+ 6052.97998046875,
20
+ 6047.341796875,
21
+ 5969.74072265625,
22
+ 5957.29296875,
23
+ 5905.70947265625,
24
+ 5913.67724609375,
25
+ 5909.44677734375,
26
+ 5882.01123046875,
27
+ 5887.45263671875
28
+ ],
29
+ "l0": [
30
+ 500.0,
31
+ 500.0,
32
+ 500.0,
33
+ 500.0,
34
+ 500.0,
35
+ 500.0,
36
+ 500.0,
37
+ 500.0,
38
+ 500.0,
39
+ 500.0,
40
+ 500.0,
41
+ 500.0,
42
+ 500.0,
43
+ 500.0,
44
+ 500.0,
45
+ 500.0,
46
+ 500.0,
47
+ 500.0,
48
+ 500.0,
49
+ 500.0,
50
+ 500.0,
51
+ 500.0,
52
+ 500.0,
53
+ 500.0,
54
+ 500.0
55
+ ],
56
+ "steps_logged": [
57
+ 0,
58
+ 200,
59
+ 400,
60
+ 600,
61
+ 800,
62
+ 1000,
63
+ 1200,
64
+ 1400,
65
+ 1600,
66
+ 1800,
67
+ 2000,
68
+ 2200,
69
+ 2400,
70
+ 2600,
71
+ 2800,
72
+ 3000,
73
+ 3200,
74
+ 3400,
75
+ 3600,
76
+ 3800,
77
+ 4000,
78
+ 4200,
79
+ 4400,
80
+ 4600,
81
+ 4800
82
+ ],
83
+ "final_step": 4800,
84
+ "converged": true,
85
+ "plateau_last": 0.019947719849343955,
86
+ "elapsed_s": 3513.2716693878174,
87
+ "row": 9,
88
+ "arch_id": "txc_bare_antidead_t5",
89
+ "arch": "txc_bare_antidead_t5",
90
+ "group": 2,
91
+ "src_class": "TXCBareAntidead",
92
+ "src_module": "src.architectures.txc_bare_antidead",
93
+ "T": 5,
94
+ "T_max": null,
95
+ "t_sample": null,
96
+ "n_layers": null,
97
+ "k_win": 500,
98
+ "k_pos": 100,
99
+ "shifts": null,
100
+ "alpha": null,
101
+ "gamma": null,
102
+ "n_scales": null,
103
+ "seed": 42,
104
+ "d_in": 2304,
105
+ "d_sae": 18432,
106
+ "subject_model": "google/gemma-2-2b-it",
107
+ "anchor_layer": 13,
108
+ "mlc_layers": [
109
+ 11,
110
+ 12,
111
+ 13,
112
+ 14,
113
+ 15
114
+ ],
115
+ "phase": "phase7_unification",
116
+ "run_id": "it_txc_bare_antidead_t5__seed42"
117
+ }