han1823123123 commited on
Commit
b1a8be0
·
verified ·
1 Parent(s): 7c5fb58

Upload training_logs/phase57_partB_h8_bare_multidistance_t4__seed42.json with huggingface_hub

Browse files
training_logs/phase57_partB_h8_bare_multidistance_t4__seed42.json ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 104747.3515625,
4
+ 31749.99609375,
5
+ 28715.025390625,
6
+ 26984.966796875,
7
+ 25926.0,
8
+ 25149.37890625,
9
+ 24708.50390625,
10
+ 24328.4375,
11
+ 23765.94921875,
12
+ 23726.056640625,
13
+ 23436.99609375,
14
+ 23265.38671875,
15
+ 23184.435546875,
16
+ 22998.498046875,
17
+ 23105.54296875,
18
+ 22940.259765625,
19
+ 22808.177734375,
20
+ 22701.306640625,
21
+ 22758.66015625
22
+ ],
23
+ "l0": [
24
+ 500.0,
25
+ 496.94921875,
26
+ 494.61865234375,
27
+ 494.24755859375,
28
+ 494.36376953125,
29
+ 494.5859375,
30
+ 495.0576171875,
31
+ 494.3623046875,
32
+ 493.201171875,
33
+ 493.9169921875,
34
+ 493.85107421875,
35
+ 492.978271484375,
36
+ 493.305908203125,
37
+ 492.621826171875,
38
+ 494.5302734375,
39
+ 493.551513671875,
40
+ 492.848388671875,
41
+ 493.04638671875,
42
+ 492.834228515625
43
+ ],
44
+ "steps_logged": [
45
+ 0,
46
+ 200,
47
+ 400,
48
+ 600,
49
+ 800,
50
+ 1000,
51
+ 1200,
52
+ 1400,
53
+ 1600,
54
+ 1800,
55
+ 2000,
56
+ 2200,
57
+ 2400,
58
+ 2600,
59
+ 2800,
60
+ 3000,
61
+ 3200,
62
+ 3400,
63
+ 3600
64
+ ],
65
+ "final_step": 3600,
66
+ "converged": true,
67
+ "plateau_last": 0.01970155844341609,
68
+ "elapsed_s": 2152.387626171112,
69
+ "shifts": [
70
+ 1,
71
+ 2
72
+ ],
73
+ "matryoshka_h_size": 3686,
74
+ "alpha": null,
75
+ "row": 31,
76
+ "arch_id": "phase57_partB_h8_bare_multidistance_t4",
77
+ "arch": "phase57_partB_h8_bare_multidistance_t4",
78
+ "group": 4,
79
+ "src_class": "TXCBareMultiDistanceContrastiveAntidead",
80
+ "src_module": "src.architectures.txc_bare_multidistance_contrastive_antidead",
81
+ "T": 4,
82
+ "T_max": null,
83
+ "t_sample": null,
84
+ "n_layers": null,
85
+ "k_win": 500,
86
+ "k_pos": 125,
87
+ "gamma": null,
88
+ "n_scales": null,
89
+ "seed": 42,
90
+ "d_in": 2304,
91
+ "d_sae": 18432,
92
+ "subject_model": "google/gemma-2-2b",
93
+ "anchor_layer": 12,
94
+ "mlc_layers": [
95
+ 10,
96
+ 11,
97
+ 12,
98
+ 13,
99
+ 14
100
+ ],
101
+ "phase": "phase7_unification",
102
+ "run_id": "phase57_partB_h8_bare_multidistance_t4__seed42"
103
+ }