han1823123123 commited on
Commit
f1c0b1b
·
verified ·
1 Parent(s): 5ebe50c

Upload training_logs/hill_h8_T5_shifts32only__seed42.json with huggingface_hub

Browse files
training_logs/hill_h8_T5_shifts32only__seed42.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 73915.5078125,
4
+ 22923.2109375,
5
+ 20817.060546875,
6
+ 19601.046875,
7
+ 18815.34765625,
8
+ 18333.5625,
9
+ 18001.43359375,
10
+ 17645.74609375,
11
+ 17508.25390625,
12
+ 17347.92578125,
13
+ 17155.791015625,
14
+ 17035.326171875,
15
+ 17006.482421875,
16
+ 16944.908203125,
17
+ 16904.462890625,
18
+ 16740.8515625,
19
+ 16779.978515625,
20
+ 16667.26171875,
21
+ 16698.97265625
22
+ ],
23
+ "l0": [
24
+ 500.0,
25
+ 498.84033203125,
26
+ 498.359375,
27
+ 498.796875,
28
+ 498.160888671875,
29
+ 498.933349609375,
30
+ 498.59814453125,
31
+ 498.33056640625,
32
+ 498.47265625,
33
+ 498.2353515625,
34
+ 498.01513671875,
35
+ 498.154296875,
36
+ 498.423828125,
37
+ 497.996826171875,
38
+ 498.9462890625,
39
+ 498.221923828125,
40
+ 498.77880859375,
41
+ 498.54052734375,
42
+ 497.891845703125
43
+ ],
44
+ "steps_logged": [
45
+ 0,
46
+ 200,
47
+ 400,
48
+ 600,
49
+ 800,
50
+ 1000,
51
+ 1200,
52
+ 1400,
53
+ 1600,
54
+ 1800,
55
+ 2000,
56
+ 2200,
57
+ 2400,
58
+ 2600,
59
+ 2800,
60
+ 3000,
61
+ 3200,
62
+ 3400,
63
+ 3600
64
+ ],
65
+ "final_step": 3600,
66
+ "converged": true,
67
+ "plateau_last": 0.019872472025036062,
68
+ "elapsed_s": 1750.1614429950714,
69
+ "shifts": [
70
+ 32
71
+ ],
72
+ "matryoshka_h_size": 3686,
73
+ "alpha": 1.0,
74
+ "row": 201,
75
+ "arch_id": "hill_h8_T5_shifts32only",
76
+ "arch": "hill_h8_T5_shifts32only",
77
+ "group": 99,
78
+ "src_class": "TXCBareMultiDistanceContrastiveAntidead",
79
+ "src_module": "src.architectures.txc_bare_multidistance_contrastive_antidead",
80
+ "T": 5,
81
+ "T_max": null,
82
+ "t_sample": null,
83
+ "n_layers": null,
84
+ "k_win": 500,
85
+ "k_pos": 100,
86
+ "gamma": null,
87
+ "n_scales": null,
88
+ "seed": 42,
89
+ "d_in": 2304,
90
+ "d_sae": 18432,
91
+ "subject_model": "google/gemma-2-2b",
92
+ "anchor_layer": 12,
93
+ "mlc_layers": [
94
+ 10,
95
+ 11,
96
+ 12,
97
+ 13,
98
+ 14
99
+ ],
100
+ "phase": "phase7_unification",
101
+ "run_id": "hill_h8_T5_shifts32only__seed42"
102
+ }