han1823123123 commited on
Commit
5fca702
·
verified ·
1 Parent(s): ba09465

Upload training_logs/txc_bare_antidead_t3_kpos20__seed42.json with huggingface_hub

Browse files
training_logs/txc_bare_antidead_t3_kpos20__seed42.json ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "arch_id": "txc_bare_antidead_t3_kpos20",
3
+ "src_class": "TXCBareAntidead",
4
+ "src_module": "src.architectures.txc_bare_antidead",
5
+ "T": 3,
6
+ "T_max": null,
7
+ "t_sample": null,
8
+ "k_win": 60,
9
+ "k_pos": 20,
10
+ "shifts": null,
11
+ "alpha": null,
12
+ "gamma": null,
13
+ "n_scales": null,
14
+ "n_layers": null,
15
+ "mlc_layers": null,
16
+ "d_in": 2304,
17
+ "d_sae": 18432,
18
+ "subject_model": "google/gemma-2-2b",
19
+ "anchor_layer": 12,
20
+ "hook_name": null,
21
+ "seed": 42,
22
+ "phase": "phase7_unification",
23
+ "group": 2,
24
+ "recipe": "TXCBareAntidead at k_pos=20 (W Phase 1 sweep)",
25
+ "purpose": "W Phase 1 sweep cell \u2014 sparsity-matched TXC",
26
+ "batch_size": 4096,
27
+ "lr": 0.0003,
28
+ "max_steps": 25000,
29
+ "elapsed_s": 1988.125037908554,
30
+ "final_step": 4600,
31
+ "converged": true,
32
+ "plateau_last": 0.018206182183690075,
33
+ "loss": [
34
+ 25338.6484375,
35
+ 6515.59521484375,
36
+ 5791.5185546875,
37
+ 5459.94677734375,
38
+ 5239.92822265625,
39
+ 5088.38671875,
40
+ 4971.3212890625,
41
+ 4863.9306640625,
42
+ 4865.43505859375,
43
+ 4825.353515625,
44
+ 4766.38037109375,
45
+ 4717.3115234375,
46
+ 4656.666015625,
47
+ 4698.89599609375,
48
+ 4653.7685546875,
49
+ 4597.822265625,
50
+ 4593.3125,
51
+ 4595.89697265625,
52
+ 4549.24951171875,
53
+ 4528.59716796875,
54
+ 4517.05078125,
55
+ 4505.091796875,
56
+ 4531.8037109375,
57
+ 4488.9453125
58
+ ],
59
+ "l0": [
60
+ 60.0,
61
+ 60.0,
62
+ 60.0,
63
+ 60.0,
64
+ 60.0,
65
+ 60.0,
66
+ 60.0,
67
+ 60.0,
68
+ 60.0,
69
+ 60.0,
70
+ 60.0,
71
+ 60.0,
72
+ 60.0,
73
+ 60.0,
74
+ 60.0,
75
+ 60.0,
76
+ 60.0,
77
+ 60.0,
78
+ 60.0,
79
+ 60.0,
80
+ 60.0,
81
+ 60.0,
82
+ 60.0,
83
+ 60.0
84
+ ],
85
+ "steps_logged": [
86
+ 0,
87
+ 200,
88
+ 400,
89
+ 600,
90
+ 800,
91
+ 1000,
92
+ 1200,
93
+ 1400,
94
+ 1600,
95
+ 1800,
96
+ 2000,
97
+ 2200,
98
+ 2400,
99
+ 2600,
100
+ 2800,
101
+ 3000,
102
+ 3200,
103
+ 3400,
104
+ 3600,
105
+ 3800,
106
+ 4000,
107
+ 4200,
108
+ 4400,
109
+ 4600
110
+ ],
111
+ "n_train_seqs": 24000
112
+ }