han1823123123 commited on
Commit
1f792f7
·
verified ·
1 Parent(s): 12ec049

Upload training_logs/txc_bare_antidead_t7__seed1.json with huggingface_hub

Browse files
training_logs/txc_bare_antidead_t7__seed1.json ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 13045.2060546875,
4
+ 6025.482421875,
5
+ 5369.08349609375,
6
+ 4985.1376953125,
7
+ 4745.466796875,
8
+ 4597.77294921875,
9
+ 4483.83935546875,
10
+ 4358.7001953125,
11
+ 4316.10888671875,
12
+ 4220.833984375,
13
+ 4204.65771484375,
14
+ 4150.03466796875,
15
+ 4130.4462890625,
16
+ 4085.968994140625,
17
+ 4055.129150390625,
18
+ 4038.227783203125,
19
+ 4038.641845703125,
20
+ 3989.169921875,
21
+ 3996.6240234375,
22
+ 3990.182373046875,
23
+ 3966.92041015625,
24
+ 3955.350341796875,
25
+ 3943.477783203125
26
+ ],
27
+ "l0": [
28
+ 500.0,
29
+ 500.0,
30
+ 500.0,
31
+ 500.0,
32
+ 500.0,
33
+ 500.0,
34
+ 500.0,
35
+ 500.0,
36
+ 500.0,
37
+ 500.0,
38
+ 500.0,
39
+ 500.0,
40
+ 500.0,
41
+ 500.0,
42
+ 500.0,
43
+ 500.0,
44
+ 500.0,
45
+ 500.0,
46
+ 500.0,
47
+ 500.0,
48
+ 500.0,
49
+ 500.0,
50
+ 500.0
51
+ ],
52
+ "steps_logged": [
53
+ 0,
54
+ 200,
55
+ 400,
56
+ 600,
57
+ 800,
58
+ 1000,
59
+ 1200,
60
+ 1400,
61
+ 1600,
62
+ 1800,
63
+ 2000,
64
+ 2200,
65
+ 2400,
66
+ 2600,
67
+ 2800,
68
+ 3000,
69
+ 3200,
70
+ 3400,
71
+ 3600,
72
+ 3800,
73
+ 4000,
74
+ 4200,
75
+ 4400
76
+ ],
77
+ "final_step": 4400,
78
+ "converged": true,
79
+ "plateau_last": 0.017547401765571555,
80
+ "elapsed_s": 1522.3476676940918,
81
+ "T": 7,
82
+ "src_class": "TXCBareAntidead",
83
+ "n_seqs_used": 24000,
84
+ "ctx_used": 64,
85
+ "ctx_slice_direction": "last",
86
+ "row": 607,
87
+ "arch_id": "txc_bare_antidead_t7",
88
+ "arch": "txc_bare_antidead_t7",
89
+ "group": 99,
90
+ "src_module": "src.architectures.txc_bare_antidead",
91
+ "T_max": null,
92
+ "t_sample": null,
93
+ "n_layers": null,
94
+ "k_win": 500,
95
+ "k_pos": 71,
96
+ "shifts": null,
97
+ "alpha": null,
98
+ "gamma": null,
99
+ "n_scales": null,
100
+ "seed": 1,
101
+ "d_in": 2304,
102
+ "d_sae": 18432,
103
+ "subject_model": "google/gemma-2-2b",
104
+ "anchor_layer": 12,
105
+ "mlc_layers": [
106
+ 10,
107
+ 11,
108
+ 12,
109
+ 13,
110
+ 14
111
+ ],
112
+ "phase": "phase7_unification",
113
+ "run_id": "txc_bare_antidead_t7__seed1"
114
+ }