han1823123123 commited on
Commit
253d0b5
·
verified ·
1 Parent(s): 2bca034

Upload training_logs/txc_bare_antidead_t7__seed42.json with huggingface_hub

Browse files
training_logs/txc_bare_antidead_t7__seed42.json ADDED
@@ -0,0 +1,117 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 12940.1435546875,
4
+ 6024.25830078125,
5
+ 5384.02587890625,
6
+ 4980.40087890625,
7
+ 4739.42822265625,
8
+ 4588.5595703125,
9
+ 4476.72509765625,
10
+ 4370.6455078125,
11
+ 4302.85546875,
12
+ 4238.1455078125,
13
+ 4165.4375,
14
+ 4114.576171875,
15
+ 4093.58447265625,
16
+ 4081.149169921875,
17
+ 4058.395751953125,
18
+ 4029.7861328125,
19
+ 4018.369140625,
20
+ 4001.9296875,
21
+ 3967.4404296875,
22
+ 3964.6044921875,
23
+ 3977.429931640625,
24
+ 3921.758056640625,
25
+ 3934.429931640625,
26
+ 3929.673095703125
27
+ ],
28
+ "l0": [
29
+ 500.0,
30
+ 500.0,
31
+ 500.0,
32
+ 500.0,
33
+ 500.0,
34
+ 500.0,
35
+ 500.0,
36
+ 500.0,
37
+ 500.0,
38
+ 500.0,
39
+ 500.0,
40
+ 500.0,
41
+ 500.0,
42
+ 500.0,
43
+ 500.0,
44
+ 500.0,
45
+ 500.0,
46
+ 500.0,
47
+ 500.0,
48
+ 500.0,
49
+ 500.0,
50
+ 500.0,
51
+ 500.0,
52
+ 500.0
53
+ ],
54
+ "steps_logged": [
55
+ 0,
56
+ 200,
57
+ 400,
58
+ 600,
59
+ 800,
60
+ 1000,
61
+ 1200,
62
+ 1400,
63
+ 1600,
64
+ 1800,
65
+ 2000,
66
+ 2200,
67
+ 2400,
68
+ 2600,
69
+ 2800,
70
+ 3000,
71
+ 3200,
72
+ 3400,
73
+ 3600,
74
+ 3800,
75
+ 4000,
76
+ 4200,
77
+ 4400,
78
+ 4600
79
+ ],
80
+ "final_step": 4600,
81
+ "converged": true,
82
+ "plateau_last": 0.01733547528374738,
83
+ "elapsed_s": 1535.4529547691345,
84
+ "T": 7,
85
+ "src_class": "TXCBareAntidead",
86
+ "n_seqs_used": 24000,
87
+ "ctx_used": 64,
88
+ "ctx_slice_direction": "last",
89
+ "row": 607,
90
+ "arch_id": "txc_bare_antidead_t7",
91
+ "arch": "txc_bare_antidead_t7",
92
+ "group": 99,
93
+ "src_module": "src.architectures.txc_bare_antidead",
94
+ "T_max": null,
95
+ "t_sample": null,
96
+ "n_layers": null,
97
+ "k_win": 500,
98
+ "k_pos": 71,
99
+ "shifts": null,
100
+ "alpha": null,
101
+ "gamma": null,
102
+ "n_scales": null,
103
+ "seed": 42,
104
+ "d_in": 2304,
105
+ "d_sae": 18432,
106
+ "subject_model": "google/gemma-2-2b",
107
+ "anchor_layer": 12,
108
+ "mlc_layers": [
109
+ 10,
110
+ 11,
111
+ 12,
112
+ 13,
113
+ 14
114
+ ],
115
+ "phase": "phase7_unification",
116
+ "run_id": "txc_bare_antidead_t7__seed42"
117
+ }