han1823123123 commited on
Commit
9a3c8f1
·
verified ·
1 Parent(s): 7ccb4cb

Upload training_logs/mlc__seed2.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. training_logs/mlc__seed2.json +105 -0
training_logs/mlc__seed2.json ADDED
@@ -0,0 +1,105 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 108364.0234375,
4
+ 4026.425048828125,
5
+ 3392.5234375,
6
+ 3163.514892578125,
7
+ 3035.2373046875,
8
+ 2914.15185546875,
9
+ 2829.30126953125,
10
+ 2792.417724609375,
11
+ 2742.762451171875,
12
+ 2776.29541015625,
13
+ 2653.72314453125,
14
+ 2669.077880859375,
15
+ 2629.986572265625,
16
+ 2664.68994140625,
17
+ 2601.24609375,
18
+ 2580.51806640625,
19
+ 2639.58642578125,
20
+ 2563.909912109375,
21
+ 2561.493408203125,
22
+ 2594.917724609375,
23
+ 2565.618408203125
24
+ ],
25
+ "l0": [
26
+ 500.0,
27
+ 492.274658203125,
28
+ 494.14306640625,
29
+ 492.90283203125,
30
+ 493.268310546875,
31
+ 492.36572265625,
32
+ 493.102294921875,
33
+ 493.1220703125,
34
+ 491.611083984375,
35
+ 493.525146484375,
36
+ 490.8408203125,
37
+ 490.330810546875,
38
+ 488.69091796875,
39
+ 491.362548828125,
40
+ 491.734375,
41
+ 490.404052734375,
42
+ 489.929443359375,
43
+ 489.762451171875,
44
+ 491.277099609375,
45
+ 489.604736328125,
46
+ 489.1552734375
47
+ ],
48
+ "steps_logged": [
49
+ 0,
50
+ 200,
51
+ 400,
52
+ 600,
53
+ 800,
54
+ 1000,
55
+ 1200,
56
+ 1400,
57
+ 1600,
58
+ 1800,
59
+ 2000,
60
+ 2200,
61
+ 2400,
62
+ 2600,
63
+ 2800,
64
+ 3000,
65
+ 3200,
66
+ 3400,
67
+ 3600,
68
+ 3800,
69
+ 4000
70
+ ],
71
+ "final_step": 4000,
72
+ "converged": true,
73
+ "plateau_last": 0.016735184303764403,
74
+ "elapsed_s": 768.920223236084,
75
+ "row": 4,
76
+ "arch_id": "mlc",
77
+ "arch": "mlc",
78
+ "group": 1,
79
+ "src_class": "MultiLayerCrosscoder",
80
+ "src_module": "src.architectures.mlc",
81
+ "T": null,
82
+ "T_max": null,
83
+ "t_sample": null,
84
+ "n_layers": 5,
85
+ "k_win": 500,
86
+ "k_pos": 100,
87
+ "shifts": null,
88
+ "alpha": null,
89
+ "gamma": null,
90
+ "n_scales": null,
91
+ "seed": 2,
92
+ "d_in": 2304,
93
+ "d_sae": 18432,
94
+ "subject_model": "google/gemma-2-2b",
95
+ "anchor_layer": 12,
96
+ "mlc_layers": [
97
+ 10,
98
+ 11,
99
+ 12,
100
+ 13,
101
+ 14
102
+ ],
103
+ "phase": "phase7_unification",
104
+ "run_id": "mlc__seed2"
105
+ }