han1823123123 commited on
Commit
769b4ca
·
verified ·
1 Parent(s): f35d66c

Upload training_logs/phase57_partB_h8_bare_multidistance_t10__seed42.json with huggingface_hub

Browse files
training_logs/phase57_partB_h8_bare_multidistance_t10__seed42.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 105003.7109375,
4
+ 52358.4921875,
5
+ 47925.06640625,
6
+ 45527.77734375,
7
+ 44065.27734375,
8
+ 43462.57421875,
9
+ 42782.03515625,
10
+ 42197.1640625,
11
+ 41981.99609375,
12
+ 41768.98046875,
13
+ 41431.15234375,
14
+ 41320.86328125,
15
+ 41095.10546875,
16
+ 41000.9765625,
17
+ 40877.00390625,
18
+ 40746.453125,
19
+ 40753.51171875,
20
+ 40709.9140625
21
+ ],
22
+ "l0": [
23
+ 500.0,
24
+ 500.0,
25
+ 500.0,
26
+ 500.0,
27
+ 500.0,
28
+ 500.0,
29
+ 500.0,
30
+ 500.0,
31
+ 500.0,
32
+ 500.0,
33
+ 500.0,
34
+ 500.0,
35
+ 500.0,
36
+ 500.0,
37
+ 500.0,
38
+ 500.0,
39
+ 500.0,
40
+ 500.0
41
+ ],
42
+ "steps_logged": [
43
+ 0,
44
+ 200,
45
+ 400,
46
+ 600,
47
+ 800,
48
+ 1000,
49
+ 1200,
50
+ 1400,
51
+ 1600,
52
+ 1800,
53
+ 2000,
54
+ 2200,
55
+ 2400,
56
+ 2600,
57
+ 2800,
58
+ 3000,
59
+ 3200,
60
+ 3400
61
+ ],
62
+ "final_step": 3400,
63
+ "converged": true,
64
+ "plateau_last": 0.016908817185128506,
65
+ "elapsed_s": 24616.253779649734,
66
+ "T": 10,
67
+ "shifts": [
68
+ 1,
69
+ 2,
70
+ 5
71
+ ],
72
+ "matryoshka_h_size": 3686,
73
+ "alpha": 1.0,
74
+ "src_class": "TXCBareMultiDistanceContrastiveAntidead",
75
+ "n_seqs_used": 24000,
76
+ "ctx_used": 64,
77
+ "ctx_slice_direction": "last",
78
+ "row": 37,
79
+ "arch_id": "phase57_partB_h8_bare_multidistance_t10",
80
+ "arch": "phase57_partB_h8_bare_multidistance_t10",
81
+ "group": 4,
82
+ "src_module": "src.architectures.txc_bare_multidistance_contrastive_antidead",
83
+ "T_max": null,
84
+ "t_sample": null,
85
+ "n_layers": null,
86
+ "k_win": 500,
87
+ "k_pos": 50,
88
+ "gamma": null,
89
+ "n_scales": null,
90
+ "seed": 42,
91
+ "d_in": 2304,
92
+ "d_sae": 18432,
93
+ "subject_model": "google/gemma-2-2b",
94
+ "anchor_layer": 12,
95
+ "mlc_layers": [
96
+ 10,
97
+ 11,
98
+ 12,
99
+ 13,
100
+ 14
101
+ ],
102
+ "phase": "phase7_unification",
103
+ "run_id": "phase57_partB_h8_bare_multidistance_t10__seed42"
104
+ }