han1823123123 commited on
Commit
be15dc9
·
verified ·
1 Parent(s): 0e081b6

Upload training_logs/agentic_mlc_08__seed42.json with huggingface_hub

Browse files
training_logs/agentic_mlc_08__seed42.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "loss": [
3
+ 109.431396484375,
4
+ 21.006454467773438,
5
+ 19.819128036499023,
6
+ 19.44778060913086,
7
+ 19.251941680908203,
8
+ 18.98621368408203,
9
+ 18.978458404541016,
10
+ 18.821090698242188,
11
+ 18.753286361694336,
12
+ 18.617267608642578,
13
+ 18.577377319335938,
14
+ 18.555763244628906,
15
+ 18.540782928466797,
16
+ 18.459762573242188,
17
+ 18.414371490478516,
18
+ 18.41324234008789
19
+ ],
20
+ "l0": [
21
+ 500.0,
22
+ 489.573486328125,
23
+ 488.43603515625,
24
+ 489.99951171875,
25
+ 490.673828125,
26
+ 489.634765625,
27
+ 491.376220703125,
28
+ 489.533935546875,
29
+ 488.439453125,
30
+ 488.724609375,
31
+ 488.857666015625,
32
+ 489.613525390625,
33
+ 489.6533203125,
34
+ 489.018798828125,
35
+ 489.099365234375,
36
+ 489.8125
37
+ ],
38
+ "steps_logged": [
39
+ 0,
40
+ 200,
41
+ 400,
42
+ 600,
43
+ 800,
44
+ 1000,
45
+ 1200,
46
+ 1400,
47
+ 1600,
48
+ 1800,
49
+ 2000,
50
+ 2200,
51
+ 2400,
52
+ 2600,
53
+ 2800,
54
+ 3000
55
+ ],
56
+ "final_step": 3000,
57
+ "converged": true,
58
+ "plateau_last": 0.014545007608135013,
59
+ "elapsed_s": 1539.747231245041,
60
+ "h": 3686,
61
+ "gamma": 0.5,
62
+ "n_scales": 3,
63
+ "alpha": null,
64
+ "row": 6,
65
+ "arch_id": "agentic_mlc_08",
66
+ "arch": "agentic_mlc_08",
67
+ "group": 1,
68
+ "src_class": "MLCContrastiveMultiscale",
69
+ "src_module": "src.architectures.mlc_contrastive_multiscale",
70
+ "T": null,
71
+ "T_max": null,
72
+ "t_sample": null,
73
+ "n_layers": 5,
74
+ "k_win": 500,
75
+ "k_pos": 100,
76
+ "shifts": [
77
+ 1,
78
+ 2,
79
+ 3
80
+ ],
81
+ "seed": 42,
82
+ "d_in": 2304,
83
+ "d_sae": 18432,
84
+ "subject_model": "google/gemma-2-2b",
85
+ "anchor_layer": 12,
86
+ "mlc_layers": [
87
+ 10,
88
+ 11,
89
+ 12,
90
+ 13,
91
+ 14
92
+ ],
93
+ "phase": "phase7_unification",
94
+ "run_id": "agentic_mlc_08__seed42"
95
+ }