han1823123123 commited on
Commit
baac68e
·
verified ·
1 Parent(s): 027bfda

Upload training_logs/phase5b_subseq_track2__seed1.json with huggingface_hub

Browse files
training_logs/phase5b_subseq_track2__seed1.json ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 17265.51953125,
4
+ 6290.14990234375,
5
+ 5699.16748046875,
6
+ 5454.59326171875,
7
+ 5321.16943359375,
8
+ 5182.515625,
9
+ 5067.52978515625,
10
+ 5077.9296875,
11
+ 5035.59326171875,
12
+ 4954.22119140625,
13
+ 4938.888671875,
14
+ 4913.4189453125,
15
+ 4926.28271484375,
16
+ 4898.189453125,
17
+ 4879.82421875,
18
+ 4825.37255859375,
19
+ 4854.55419921875,
20
+ 4866.50390625
21
+ ],
22
+ "l0": [
23
+ 500.0,
24
+ 497.40380859375,
25
+ 497.3095703125,
26
+ 496.5732421875,
27
+ 496.7822265625,
28
+ 497.09375,
29
+ 496.07421875,
30
+ 496.8115234375,
31
+ 496.672119140625,
32
+ 496.0234375,
33
+ 494.92138671875,
34
+ 495.0595703125,
35
+ 495.5595703125,
36
+ 495.54443359375,
37
+ 495.055908203125,
38
+ 494.145751953125,
39
+ 493.975341796875,
40
+ 494.615478515625
41
+ ],
42
+ "steps_logged": [
43
+ 0,
44
+ 200,
45
+ 400,
46
+ 600,
47
+ 800,
48
+ 1000,
49
+ 1200,
50
+ 1400,
51
+ 1600,
52
+ 1800,
53
+ 2000,
54
+ 2200,
55
+ 2400,
56
+ 2600,
57
+ 2800,
58
+ 3000,
59
+ 3200,
60
+ 3400
61
+ ],
62
+ "final_step": 3400,
63
+ "converged": true,
64
+ "plateau_last": 0.01792446679831454,
65
+ "elapsed_s": 2074.7629668712616,
66
+ "T_max": 10,
67
+ "t_sample": 5,
68
+ "row": 12,
69
+ "arch_id": "phase5b_subseq_track2",
70
+ "arch": "phase5b_subseq_track2",
71
+ "group": 2,
72
+ "src_class": "SubseqTXCBareAntidead",
73
+ "src_module": "src.architectures.phase5b_subseq_sampling_txcdr",
74
+ "T": null,
75
+ "n_layers": null,
76
+ "k_win": 500,
77
+ "k_pos": 100,
78
+ "shifts": null,
79
+ "alpha": null,
80
+ "gamma": null,
81
+ "n_scales": null,
82
+ "seed": 1,
83
+ "d_in": 2304,
84
+ "d_sae": 18432,
85
+ "subject_model": "google/gemma-2-2b",
86
+ "anchor_layer": 12,
87
+ "mlc_layers": [
88
+ 10,
89
+ 11,
90
+ 12,
91
+ 13,
92
+ 14
93
+ ],
94
+ "phase": "phase7_unification",
95
+ "run_id": "phase5b_subseq_track2__seed1"
96
+ }