han1823123123 commited on
Commit
7ffa813
·
verified ·
1 Parent(s): dd296b2

Upload training_logs/txc_bare_antidead_t10__seed2.json with huggingface_hub

Browse files
training_logs/txc_bare_antidead_t10__seed2.json ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 17932.51953125,
4
+ 6549.25341796875,
5
+ 5851.27294921875,
6
+ 5451.41162109375,
7
+ 5218.31884765625,
8
+ 5049.38671875,
9
+ 4955.45458984375,
10
+ 4895.53515625,
11
+ 4821.29931640625,
12
+ 4780.52099609375,
13
+ 4681.20849609375,
14
+ 4695.91015625,
15
+ 4656.79638671875,
16
+ 4640.0078125,
17
+ 4638.40869140625,
18
+ 4604.4765625,
19
+ 4589.45361328125,
20
+ 4576.60400390625,
21
+ 4611.01318359375
22
+ ],
23
+ "l0": [
24
+ 500.0,
25
+ 495.078125,
26
+ 495.338134765625,
27
+ 494.291015625,
28
+ 493.177490234375,
29
+ 493.44189453125,
30
+ 492.522216796875,
31
+ 494.070556640625,
32
+ 493.125,
33
+ 492.378662109375,
34
+ 490.884765625,
35
+ 491.9541015625,
36
+ 491.320068359375,
37
+ 492.12548828125,
38
+ 491.37109375,
39
+ 492.783935546875,
40
+ 490.42724609375,
41
+ 492.625732421875,
42
+ 492.26904296875
43
+ ],
44
+ "steps_logged": [
45
+ 0,
46
+ 200,
47
+ 400,
48
+ 600,
49
+ 800,
50
+ 1000,
51
+ 1200,
52
+ 1400,
53
+ 1600,
54
+ 1800,
55
+ 2000,
56
+ 2200,
57
+ 2400,
58
+ 2600,
59
+ 2800,
60
+ 3000,
61
+ 3200,
62
+ 3400,
63
+ 3600
64
+ ],
65
+ "final_step": 3600,
66
+ "converged": true,
67
+ "plateau_last": 0.01852475359428169,
68
+ "elapsed_s": 2157.621589899063,
69
+ "row": 10,
70
+ "arch_id": "txc_bare_antidead_t10",
71
+ "arch": "txc_bare_antidead_t10",
72
+ "group": 2,
73
+ "src_class": "TXCBareAntidead",
74
+ "src_module": "src.architectures.txc_bare_antidead",
75
+ "T": 10,
76
+ "T_max": null,
77
+ "t_sample": null,
78
+ "n_layers": null,
79
+ "k_win": 500,
80
+ "k_pos": 50,
81
+ "shifts": null,
82
+ "alpha": null,
83
+ "gamma": null,
84
+ "n_scales": null,
85
+ "seed": 2,
86
+ "d_in": 2304,
87
+ "d_sae": 18432,
88
+ "subject_model": "google/gemma-2-2b",
89
+ "anchor_layer": 12,
90
+ "mlc_layers": [
91
+ 10,
92
+ 11,
93
+ 12,
94
+ 13,
95
+ 14
96
+ ],
97
+ "phase": "phase7_unification",
98
+ "run_id": "txc_bare_antidead_t10__seed2"
99
+ }