han1823123123 commited on
Commit
d05f802
·
verified ·
1 Parent(s): e591c0c

Upload training_logs/phase57_partB_h8_bare_multidistance_t8__seed1.json with huggingface_hub

Browse files
training_logs/phase57_partB_h8_bare_multidistance_t8__seed1.json ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 127731.703125,
4
+ 50203.98046875,
5
+ 45579.3359375,
6
+ 43645.43359375,
7
+ 41998.22265625,
8
+ 41584.6171875,
9
+ 40820.59375,
10
+ 40344.14453125,
11
+ 40248.71484375,
12
+ 39754.06640625,
13
+ 39695.66015625,
14
+ 39567.66015625,
15
+ 39196.078125,
16
+ 39148.4765625,
17
+ 38974.421875,
18
+ 38756.3125,
19
+ 38870.921875,
20
+ 38766.66015625
21
+ ],
22
+ "l0": [
23
+ 500.0,
24
+ 495.25390625,
25
+ 494.4990234375,
26
+ 494.58203125,
27
+ 491.463623046875,
28
+ 492.6103515625,
29
+ 492.31201171875,
30
+ 492.677490234375,
31
+ 493.69140625,
32
+ 492.987060546875,
33
+ 493.145263671875,
34
+ 494.060546875,
35
+ 492.019287109375,
36
+ 492.97021484375,
37
+ 492.91845703125,
38
+ 491.835693359375,
39
+ 492.138916015625,
40
+ 492.42333984375
41
+ ],
42
+ "steps_logged": [
43
+ 0,
44
+ 200,
45
+ 400,
46
+ 600,
47
+ 800,
48
+ 1000,
49
+ 1200,
50
+ 1400,
51
+ 1600,
52
+ 1800,
53
+ 2000,
54
+ 2200,
55
+ 2400,
56
+ 2600,
57
+ 2800,
58
+ 3000,
59
+ 3200,
60
+ 3400
61
+ ],
62
+ "final_step": 3400,
63
+ "converged": true,
64
+ "plateau_last": 0.01987979132831461,
65
+ "elapsed_s": 5185.838559150696,
66
+ "shifts": [
67
+ 1,
68
+ 2,
69
+ 4
70
+ ],
71
+ "matryoshka_h_size": 3686,
72
+ "alpha": null,
73
+ "row": 35,
74
+ "arch_id": "phase57_partB_h8_bare_multidistance_t8",
75
+ "arch": "phase57_partB_h8_bare_multidistance_t8",
76
+ "group": 4,
77
+ "src_class": "TXCBareMultiDistanceContrastiveAntidead",
78
+ "src_module": "src.architectures.txc_bare_multidistance_contrastive_antidead",
79
+ "T": 8,
80
+ "T_max": null,
81
+ "t_sample": null,
82
+ "n_layers": null,
83
+ "k_win": 500,
84
+ "k_pos": 62,
85
+ "gamma": null,
86
+ "n_scales": null,
87
+ "seed": 1,
88
+ "d_in": 2304,
89
+ "d_sae": 18432,
90
+ "subject_model": "google/gemma-2-2b",
91
+ "anchor_layer": 12,
92
+ "mlc_layers": [
93
+ 10,
94
+ 11,
95
+ 12,
96
+ 13,
97
+ 14
98
+ ],
99
+ "phase": "phase7_unification",
100
+ "run_id": "phase57_partB_h8_bare_multidistance_t8__seed1"
101
+ }