han1823123123 commited on
Commit
0514beb
·
verified ·
1 Parent(s): ffec72a

Upload training_logs/txcdr_t5__seed2.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. training_logs/txcdr_t5__seed2.json +123 -0
training_logs/txcdr_t5__seed2.json ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 48092.453125,
4
+ 6306.45263671875,
5
+ 5469.94482421875,
6
+ 5074.91748046875,
7
+ 4788.94482421875,
8
+ 4618.28857421875,
9
+ 4547.6591796875,
10
+ 4418.791015625,
11
+ 4324.10595703125,
12
+ 4253.98388671875,
13
+ 4233.44091796875,
14
+ 4145.1748046875,
15
+ 4100.9345703125,
16
+ 4104.19677734375,
17
+ 4074.540771484375,
18
+ 4028.160888671875,
19
+ 4010.862548828125,
20
+ 3970.176513671875,
21
+ 3949.996826171875,
22
+ 3920.809326171875,
23
+ 3915.040771484375,
24
+ 3875.427001953125,
25
+ 3878.01806640625,
26
+ 3858.763427734375,
27
+ 3883.25048828125,
28
+ 3866.01416015625,
29
+ 3835.19384765625
30
+ ],
31
+ "l0": [
32
+ 500.0,
33
+ 495.21484375,
34
+ 496.008056640625,
35
+ 495.939453125,
36
+ 494.549560546875,
37
+ 493.9580078125,
38
+ 494.4990234375,
39
+ 494.210205078125,
40
+ 493.7119140625,
41
+ 493.146728515625,
42
+ 494.04443359375,
43
+ 493.61376953125,
44
+ 492.7265625,
45
+ 493.489501953125,
46
+ 492.774658203125,
47
+ 492.2578125,
48
+ 491.93896484375,
49
+ 493.103515625,
50
+ 493.124755859375,
51
+ 491.602783203125,
52
+ 491.806884765625,
53
+ 491.767333984375,
54
+ 492.539794921875,
55
+ 492.02978515625,
56
+ 491.913330078125,
57
+ 492.790283203125,
58
+ 491.43603515625
59
+ ],
60
+ "steps_logged": [
61
+ 0,
62
+ 200,
63
+ 400,
64
+ 600,
65
+ 800,
66
+ 1000,
67
+ 1200,
68
+ 1400,
69
+ 1600,
70
+ 1800,
71
+ 2000,
72
+ 2200,
73
+ 2400,
74
+ 2600,
75
+ 2800,
76
+ 3000,
77
+ 3200,
78
+ 3400,
79
+ 3600,
80
+ 3800,
81
+ 4000,
82
+ 4200,
83
+ 4400,
84
+ 4600,
85
+ 4800,
86
+ 5000,
87
+ 5200
88
+ ],
89
+ "final_step": 5200,
90
+ "converged": true,
91
+ "plateau_last": 0.01580170808955213,
92
+ "elapsed_s": 1014.807079076767,
93
+ "row": 16,
94
+ "arch_id": "txcdr_t5",
95
+ "arch": "txcdr_t5",
96
+ "group": 3,
97
+ "src_class": "TemporalCrosscoder",
98
+ "src_module": "src.architectures.crosscoder",
99
+ "T": 5,
100
+ "T_max": null,
101
+ "t_sample": null,
102
+ "n_layers": null,
103
+ "k_win": 500,
104
+ "k_pos": 100,
105
+ "shifts": null,
106
+ "alpha": null,
107
+ "gamma": null,
108
+ "n_scales": null,
109
+ "seed": 2,
110
+ "d_in": 2304,
111
+ "d_sae": 18432,
112
+ "subject_model": "google/gemma-2-2b",
113
+ "anchor_layer": 12,
114
+ "mlc_layers": [
115
+ 10,
116
+ 11,
117
+ 12,
118
+ 13,
119
+ 14
120
+ ],
121
+ "phase": "phase7_unification",
122
+ "run_id": "txcdr_t5__seed2"
123
+ }