han1823123123 commited on
Commit
8716e3d
·
verified ·
1 Parent(s): 9b4eb4a

Upload training_logs/txc_bare_antidead_t20__seed42.json with huggingface_hub

Browse files
training_logs/txc_bare_antidead_t20__seed42.json ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 16041.8544921875,
4
+ 7564.35546875,
5
+ 6897.26123046875,
6
+ 6578.85546875,
7
+ 6413.78271484375,
8
+ 6272.0361328125,
9
+ 6199.66357421875,
10
+ 6122.4013671875,
11
+ 6090.73583984375,
12
+ 6034.8017578125,
13
+ 6002.14697265625,
14
+ 5981.08349609375,
15
+ 5931.5400390625,
16
+ 5924.2861328125,
17
+ 5922.11669921875,
18
+ 5899.1552734375,
19
+ 5884.80615234375,
20
+ 5886.90966796875
21
+ ],
22
+ "l0": [
23
+ 500.0,
24
+ 496.969970703125,
25
+ 495.277099609375,
26
+ 496.025390625,
27
+ 495.86962890625,
28
+ 495.411376953125,
29
+ 495.5185546875,
30
+ 494.61865234375,
31
+ 494.830078125,
32
+ 495.408447265625,
33
+ 495.050537109375,
34
+ 494.5859375,
35
+ 492.891357421875,
36
+ 493.76318359375,
37
+ 494.0830078125,
38
+ 494.323974609375,
39
+ 494.130859375,
40
+ 494.554931640625
41
+ ],
42
+ "steps_logged": [
43
+ 0,
44
+ 200,
45
+ 400,
46
+ 600,
47
+ 800,
48
+ 1000,
49
+ 1200,
50
+ 1400,
51
+ 1600,
52
+ 1800,
53
+ 2000,
54
+ 2200,
55
+ 2400,
56
+ 2600,
57
+ 2800,
58
+ 3000,
59
+ 3200,
60
+ 3400
61
+ ],
62
+ "final_step": 3400,
63
+ "converged": true,
64
+ "plateau_last": 0.017411079069201706,
65
+ "elapsed_s": 3833.5945932865143,
66
+ "row": 11,
67
+ "arch_id": "txc_bare_antidead_t20",
68
+ "arch": "txc_bare_antidead_t20",
69
+ "group": 2,
70
+ "src_class": "TXCBareAntidead",
71
+ "src_module": "src.architectures.txc_bare_antidead",
72
+ "T": 20,
73
+ "T_max": null,
74
+ "t_sample": null,
75
+ "n_layers": null,
76
+ "k_win": 500,
77
+ "k_pos": 25,
78
+ "shifts": null,
79
+ "alpha": null,
80
+ "gamma": null,
81
+ "n_scales": null,
82
+ "seed": 42,
83
+ "d_in": 2304,
84
+ "d_sae": 18432,
85
+ "subject_model": "google/gemma-2-2b",
86
+ "anchor_layer": 12,
87
+ "mlc_layers": [
88
+ 10,
89
+ 11,
90
+ 12,
91
+ 13,
92
+ 14
93
+ ],
94
+ "phase": "phase7_unification",
95
+ "run_id": "txc_bare_antidead_t20__seed42"
96
+ }