diff --git "a/New/REG/wandb/run-20260326_123101-m3lli51t/files/output.log" "b/New/REG/wandb/run-20260326_123101-m3lli51t/files/output.log" new file mode 100644--- /dev/null +++ "b/New/REG/wandb/run-20260326_123101-m3lli51t/files/output.log" @@ -0,0 +1,7853 @@ +Steps: 0%| | 1/1000000 [00:02<778:45:47, 2.80s/it][2026-03-26 12:31:06] Generating EMA samples done. +[2026-03-26 12:31:06] Step: 1, Training Logs: loss_final: 1.831830, loss_mean: 1.710702, proj_loss: 0.003202, loss_mean_cls: 0.117926, grad_norm: 1.587149 +Steps: 0%| | 2/1000000 [00:03<359:22:05, 1.29s/it, grad_norm=1.59, loss_final=1.83, loss_mean=1.71, loss_mean_cls=0.118, proj_loss=0.0032][2026-03-26 12:31:06] Step: 2, Training Logs: loss_final: 1.774854, loss_mean: 1.671099, proj_loss: -0.014662, loss_mean_cls: 0.118417, grad_norm: 0.830605 +Steps: 0%| | 3/1000000 [00:03<226:38:58, 1.23it/s, grad_norm=0.831, loss_final=1.77, loss_mean=1.67, loss_mean_cls=0.118, proj_loss=-0.0147][2026-03-26 12:31:06] Step: 3, Training Logs: loss_final: 1.763657, loss_mean: 1.675607, proj_loss: -0.030356, loss_mean_cls: 0.118406, grad_norm: 0.879399 +Steps: 0%| | 4/1000000 [00:03<164:05:54, 1.69it/s, grad_norm=0.879, loss_final=1.76, loss_mean=1.68, loss_mean_cls=0.118, proj_loss=-0.0304][2026-03-26 12:31:06] Step: 4, Training Logs: loss_final: 1.779609, loss_mean: 1.700004, proj_loss: -0.038070, loss_mean_cls: 0.117676, grad_norm: 0.819437 +Steps: 0%| | 5/1000000 [00:03<129:34:04, 2.14it/s, grad_norm=0.819, loss_final=1.78, loss_mean=1.7, loss_mean_cls=0.118, proj_loss=-0.0381][2026-03-26 12:31:07] Step: 5, Training Logs: loss_final: 1.750054, loss_mean: 1.675290, proj_loss: -0.043013, loss_mean_cls: 0.117777, grad_norm: 1.021779 +Steps: 0%| | 6/1000000 [00:04<108:44:40, 2.55it/s, grad_norm=1.02, loss_final=1.75, loss_mean=1.68, loss_mean_cls=0.118, proj_loss=-0.043][2026-03-26 12:31:07] Step: 6, Training Logs: loss_final: 1.713697, loss_mean: 1.641904, proj_loss: -0.046146, loss_mean_cls: 0.117939, grad_norm: 0.895714 +Steps: 0%| | 7/1000000 [00:04<95:43:08, 2.90it/s, grad_norm=0.896, loss_final=1.71, loss_mean=1.64, loss_mean_cls=0.118, proj_loss=-0.0461] [2026-03-26 12:31:07] Step: 7, Training Logs: loss_final: 1.700274, loss_mean: 1.631343, proj_loss: -0.048905, loss_mean_cls: 0.117836, grad_norm: 1.212205 +Steps: 0%| | 8/1000000 [00:04<87:02:09, 3.19it/s, grad_norm=1.21, loss_final=1.7, loss_mean=1.63, loss_mean_cls=0.118, proj_loss=-0.0489][2026-03-26 12:31:07] Step: 8, Training Logs: loss_final: 1.689372, loss_mean: 1.623324, proj_loss: -0.051387, loss_mean_cls: 0.117435, grad_norm: 1.398131 +Steps: 0%| | 9/1000000 [00:04<81:12:12, 3.42it/s, grad_norm=1.4, loss_final=1.69, loss_mean=1.62, loss_mean_cls=0.117, proj_loss=-0.0514][2026-03-26 12:31:08] Step: 9, Training Logs: loss_final: 1.729895, loss_mean: 1.664244, proj_loss: -0.052055, loss_mean_cls: 0.117706, grad_norm: 2.476503 +Steps: 0%| | 10/1000000 [00:05<77:13:14, 3.60it/s, grad_norm=2.48, loss_final=1.73, loss_mean=1.66, loss_mean_cls=0.118, proj_loss=-0.0521][2026-03-26 12:31:08] Step: 10, Training Logs: loss_final: 1.641072, loss_mean: 1.573964, proj_loss: -0.049600, loss_mean_cls: 0.116707, grad_norm: 1.163038 +Steps: 0%| | 11/1000000 [00:05<74:29:56, 3.73it/s, grad_norm=1.16, loss_final=1.64, loss_mean=1.57, loss_mean_cls=0.117, proj_loss=-0.0496][2026-03-26 12:31:08] Step: 11, Training Logs: loss_final: 1.704696, loss_mean: 1.637629, proj_loss: -0.050654, loss_mean_cls: 0.117720, grad_norm: 2.780670 +Steps: 0%| | 12/1000000 [00:05<72:38:18, 3.82it/s, grad_norm=2.78, loss_final=1.7, loss_mean=1.64, loss_mean_cls=0.118, proj_loss=-0.0507][2026-03-26 12:31:08] Step: 12, Training Logs: loss_final: 1.635992, loss_mean: 1.573394, proj_loss: -0.054477, loss_mean_cls: 0.117075, grad_norm: 1.375333 +Steps: 0%| | 13/1000000 [00:05<71:21:20, 3.89it/s, grad_norm=1.38, loss_final=1.64, loss_mean=1.57, loss_mean_cls=0.117, proj_loss=-0.0545][2026-03-26 12:31:09] Step: 13, Training Logs: loss_final: 1.646804, loss_mean: 1.584257, proj_loss: -0.055227, loss_mean_cls: 0.117773, grad_norm: 1.557633 +Steps: 0%| | 14/1000000 [00:05<70:23:29, 3.95it/s, grad_norm=1.56, loss_final=1.65, loss_mean=1.58, loss_mean_cls=0.118, proj_loss=-0.0552][2026-03-26 12:31:09] Step: 14, Training Logs: loss_final: 1.641286, loss_mean: 1.579234, proj_loss: -0.055001, loss_mean_cls: 0.117053, grad_norm: 1.542598 +Steps: 0%| | 15/1000000 [00:06<69:47:16, 3.98it/s, grad_norm=1.54, loss_final=1.64, loss_mean=1.58, loss_mean_cls=0.117, proj_loss=-0.055][2026-03-26 12:31:09] Step: 15, Training Logs: loss_final: 1.635570, loss_mean: 1.571396, proj_loss: -0.052898, loss_mean_cls: 0.117071, grad_norm: 1.258048 +Steps: 0%| | 16/1000000 [00:06<69:19:04, 4.01it/s, grad_norm=1.26, loss_final=1.64, loss_mean=1.57, loss_mean_cls=0.117, proj_loss=-0.0529][2026-03-26 12:31:09] Step: 16, Training Logs: loss_final: 1.610049, loss_mean: 1.547075, proj_loss: -0.053394, loss_mean_cls: 0.116368, grad_norm: 1.060012 +Steps: 0%| | 17/1000000 [00:06<69:01:26, 4.02it/s, grad_norm=1.06, loss_final=1.61, loss_mean=1.55, loss_mean_cls=0.116, proj_loss=-0.0534][2026-03-26 12:31:10] Step: 17, Training Logs: loss_final: 1.593763, loss_mean: 1.531993, proj_loss: -0.055101, loss_mean_cls: 0.116872, grad_norm: 1.147782 +Steps: 0%| | 18/1000000 [00:06<68:49:30, 4.04it/s, grad_norm=1.15, loss_final=1.59, loss_mean=1.53, loss_mean_cls=0.117, proj_loss=-0.0551][2026-03-26 12:31:10] Step: 18, Training Logs: loss_final: 1.584808, loss_mean: 1.524330, proj_loss: -0.056157, loss_mean_cls: 0.116635, grad_norm: 1.176332 +Steps: 0%| | 19/1000000 [00:07<68:42:41, 4.04it/s, grad_norm=1.18, loss_final=1.58, loss_mean=1.52, loss_mean_cls=0.117, proj_loss=-0.0562][2026-03-26 12:31:10] Step: 19, Training Logs: loss_final: 1.571087, loss_mean: 1.509044, proj_loss: -0.054372, loss_mean_cls: 0.116414, grad_norm: 1.131125 +Steps: 0%| | 20/1000000 [00:07<68:35:14, 4.05it/s, grad_norm=1.13, loss_final=1.57, loss_mean=1.51, loss_mean_cls=0.116, proj_loss=-0.0544][2026-03-26 12:31:10] Step: 20, Training Logs: loss_final: 1.539737, loss_mean: 1.477966, proj_loss: -0.054994, loss_mean_cls: 0.116765, grad_norm: 1.075183 +Steps: 0%| | 21/1000000 [00:07<68:32:34, 4.05it/s, grad_norm=1.08, loss_final=1.54, loss_mean=1.48, loss_mean_cls=0.117, proj_loss=-0.055][2026-03-26 12:31:11] Step: 21, Training Logs: loss_final: 1.528002, loss_mean: 1.466739, proj_loss: -0.055123, loss_mean_cls: 0.116386, grad_norm: 1.036196 +Steps: 0%| | 22/1000000 [00:07<68:32:08, 4.05it/s, grad_norm=1.04, loss_final=1.53, loss_mean=1.47, loss_mean_cls=0.116, proj_loss=-0.0551][2026-03-26 12:31:11] Step: 22, Training Logs: loss_final: 1.532516, loss_mean: 1.470622, proj_loss: -0.054418, loss_mean_cls: 0.116312, grad_norm: 1.676948 +Steps: 0%| | 23/1000000 [00:08<68:30:57, 4.05it/s, grad_norm=1.68, loss_final=1.53, loss_mean=1.47, loss_mean_cls=0.116, proj_loss=-0.0544][2026-03-26 12:31:11] Step: 23, Training Logs: loss_final: 1.564712, loss_mean: 1.507069, proj_loss: -0.058813, loss_mean_cls: 0.116455, grad_norm: 2.156650 +Steps: 0%| | 24/1000000 [00:08<68:29:12, 4.06it/s, grad_norm=2.16, loss_final=1.56, loss_mean=1.51, loss_mean_cls=0.116, proj_loss=-0.0588][2026-03-26 12:31:11] Step: 24, Training Logs: loss_final: 1.540663, loss_mean: 1.479752, proj_loss: -0.054639, loss_mean_cls: 0.115550, grad_norm: 2.211862 +Steps: 0%| | 25/1000000 [00:08<68:29:52, 4.06it/s, grad_norm=2.21, loss_final=1.54, loss_mean=1.48, loss_mean_cls=0.116, proj_loss=-0.0546][2026-03-26 12:31:12] Step: 25, Training Logs: loss_final: 1.502737, loss_mean: 1.441529, proj_loss: -0.054073, loss_mean_cls: 0.115280, grad_norm: 1.594505 +Steps: 0%| | 26/1000000 [00:08<68:33:40, 4.05it/s, grad_norm=1.59, loss_final=1.5, loss_mean=1.44, loss_mean_cls=0.115, proj_loss=-0.0541][2026-03-26 12:31:12] Step: 26, Training Logs: loss_final: 1.523026, loss_mean: 1.463264, proj_loss: -0.056279, loss_mean_cls: 0.116042, grad_norm: 2.588172 +Steps: 0%| | 27/1000000 [00:09<68:34:44, 4.05it/s, grad_norm=2.59, loss_final=1.52, loss_mean=1.46, loss_mean_cls=0.116, proj_loss=-0.0563][2026-03-26 12:31:12] Step: 27, Training Logs: loss_final: 1.528182, loss_mean: 1.467721, proj_loss: -0.054857, loss_mean_cls: 0.115318, grad_norm: 2.285323 +Steps: 0%| | 28/1000000 [00:09<68:33:03, 4.05it/s, grad_norm=2.29, loss_final=1.53, loss_mean=1.47, loss_mean_cls=0.115, proj_loss=-0.0549][2026-03-26 12:31:12] Step: 28, Training Logs: loss_final: 1.493315, loss_mean: 1.434541, proj_loss: -0.056577, loss_mean_cls: 0.115351, grad_norm: 1.795185 +Steps: 0%| | 29/1000000 [00:09<68:29:23, 4.06it/s, grad_norm=1.8, loss_final=1.49, loss_mean=1.43, loss_mean_cls=0.115, proj_loss=-0.0566][2026-03-26 12:31:13] Step: 29, Training Logs: loss_final: 1.495071, loss_mean: 1.438045, proj_loss: -0.058663, loss_mean_cls: 0.115689, grad_norm: 1.410272 +Steps: 0%| | 30/1000000 [00:09<68:28:56, 4.06it/s, grad_norm=1.41, loss_final=1.5, loss_mean=1.44, loss_mean_cls=0.116, proj_loss=-0.0587][2026-03-26 12:31:13] Step: 30, Training Logs: loss_final: 1.473940, loss_mean: 1.414495, proj_loss: -0.055426, loss_mean_cls: 0.114871, grad_norm: 1.097311 +Steps: 0%| | 31/1000000 [00:10<68:28:12, 4.06it/s, grad_norm=1.1, loss_final=1.47, loss_mean=1.41, loss_mean_cls=0.115, proj_loss=-0.0554][2026-03-26 12:31:13] Step: 31, Training Logs: loss_final: 1.485892, loss_mean: 1.429290, proj_loss: -0.058327, loss_mean_cls: 0.114930, grad_norm: 1.315569 +Steps: 0%| | 32/1000000 [00:10<68:27:50, 4.06it/s, grad_norm=1.32, loss_final=1.49, loss_mean=1.43, loss_mean_cls=0.115, proj_loss=-0.0583][2026-03-26 12:31:13] Step: 32, Training Logs: loss_final: 1.464506, loss_mean: 1.405257, proj_loss: -0.055400, loss_mean_cls: 0.114649, grad_norm: 1.289816 +Steps: 0%| | 33/1000000 [00:10<68:25:45, 4.06it/s, grad_norm=1.29, loss_final=1.46, loss_mean=1.41, loss_mean_cls=0.115, proj_loss=-0.0554][2026-03-26 12:31:14] Step: 33, Training Logs: loss_final: 1.490027, loss_mean: 1.431066, proj_loss: -0.055938, loss_mean_cls: 0.114899, grad_norm: 1.356932 +Steps: 0%| | 34/1000000 [00:10<68:25:54, 4.06it/s, grad_norm=1.36, loss_final=1.49, loss_mean=1.43, loss_mean_cls=0.115, proj_loss=-0.0559][2026-03-26 12:31:14] Step: 34, Training Logs: loss_final: 1.456502, loss_mean: 1.398506, proj_loss: -0.056388, loss_mean_cls: 0.114384, grad_norm: 0.953684 +Steps: 0%| | 35/1000000 [00:11<68:26:43, 4.06it/s, grad_norm=0.954, loss_final=1.46, loss_mean=1.4, loss_mean_cls=0.114, proj_loss=-0.0564][2026-03-26 12:31:14] Step: 35, Training Logs: loss_final: 1.451401, loss_mean: 1.392304, proj_loss: -0.055721, loss_mean_cls: 0.114818, grad_norm: 0.901062 +Steps: 0%| | 36/1000000 [00:11<68:29:00, 4.06it/s, grad_norm=0.901, loss_final=1.45, loss_mean=1.39, loss_mean_cls=0.115, proj_loss=-0.0557][2026-03-26 12:31:14] Step: 36, Training Logs: loss_final: 1.459791, loss_mean: 1.402153, proj_loss: -0.056932, loss_mean_cls: 0.114569, grad_norm: 0.771318 +Steps: 0%| | 37/1000000 [00:11<68:29:59, 4.06it/s, grad_norm=0.771, loss_final=1.46, loss_mean=1.4, loss_mean_cls=0.115, proj_loss=-0.0569][2026-03-26 12:31:15] Step: 37, Training Logs: loss_final: 1.438417, loss_mean: 1.381726, proj_loss: -0.057720, loss_mean_cls: 0.114412, grad_norm: 0.765226 +Steps: 0%| | 38/1000000 [00:11<68:30:21, 4.05it/s, grad_norm=0.765, loss_final=1.44, loss_mean=1.38, loss_mean_cls=0.114, proj_loss=-0.0577][2026-03-26 12:31:15] Step: 38, Training Logs: loss_final: 1.460572, loss_mean: 1.401469, proj_loss: -0.055347, loss_mean_cls: 0.114449, grad_norm: 0.781913 +Steps: 0%| | 39/1000000 [00:12<68:28:52, 4.06it/s, grad_norm=0.782, loss_final=1.46, loss_mean=1.4, loss_mean_cls=0.114, proj_loss=-0.0553][2026-03-26 12:31:15] Step: 39, Training Logs: loss_final: 1.437020, loss_mean: 1.379368, proj_loss: -0.057175, loss_mean_cls: 0.114827, grad_norm: 0.681366 +Steps: 0%| | 40/1000000 [00:12<68:46:46, 4.04it/s, grad_norm=0.681, loss_final=1.44, loss_mean=1.38, loss_mean_cls=0.115, proj_loss=-0.0572][2026-03-26 12:31:15] Step: 40, Training Logs: loss_final: 1.436018, loss_mean: 1.378294, proj_loss: -0.056293, loss_mean_cls: 0.114017, grad_norm: 0.598676 +Steps: 0%| | 41/1000000 [00:12<68:34:41, 4.05it/s, grad_norm=0.599, loss_final=1.44, loss_mean=1.38, loss_mean_cls=0.114, proj_loss=-0.0563][2026-03-26 12:31:16] Step: 41, Training Logs: loss_final: 1.429434, loss_mean: 1.372004, proj_loss: -0.056637, loss_mean_cls: 0.114067, grad_norm: 0.666882 +Steps: 0%| | 42/1000000 [00:12<68:33:40, 4.05it/s, grad_norm=0.667, loss_final=1.43, loss_mean=1.37, loss_mean_cls=0.114, proj_loss=-0.0566][2026-03-26 12:31:16] Step: 42, Training Logs: loss_final: 1.414209, loss_mean: 1.353422, proj_loss: -0.053293, loss_mean_cls: 0.114080, grad_norm: 0.525751 +Steps: 0%| | 43/1000000 [00:13<68:34:49, 4.05it/s, grad_norm=0.526, loss_final=1.41, loss_mean=1.35, loss_mean_cls=0.114, proj_loss=-0.0533][2026-03-26 12:31:16] Step: 43, Training Logs: loss_final: 1.412690, loss_mean: 1.355159, proj_loss: -0.056555, loss_mean_cls: 0.114086, grad_norm: 0.538756 +Steps: 0%| | 44/1000000 [00:13<68:32:54, 4.05it/s, grad_norm=0.539, loss_final=1.41, loss_mean=1.36, loss_mean_cls=0.114, proj_loss=-0.0566][2026-03-26 12:31:16] Step: 44, Training Logs: loss_final: 1.407236, loss_mean: 1.349214, proj_loss: -0.055287, loss_mean_cls: 0.113309, grad_norm: 0.568122 +Steps: 0%| | 45/1000000 [00:13<68:28:25, 4.06it/s, grad_norm=0.568, loss_final=1.41, loss_mean=1.35, loss_mean_cls=0.113, proj_loss=-0.0553][2026-03-26 12:31:17] Step: 45, Training Logs: loss_final: 1.389176, loss_mean: 1.331999, proj_loss: -0.056370, loss_mean_cls: 0.113547, grad_norm: 0.575899 +Steps: 0%| | 46/1000000 [00:13<68:29:56, 4.06it/s, grad_norm=0.576, loss_final=1.39, loss_mean=1.33, loss_mean_cls=0.114, proj_loss=-0.0564][2026-03-26 12:31:17] Step: 46, Training Logs: loss_final: 1.423756, loss_mean: 1.365782, proj_loss: -0.055219, loss_mean_cls: 0.113194, grad_norm: 0.514916 +Steps: 0%| | 47/1000000 [00:14<68:31:56, 4.05it/s, grad_norm=0.515, loss_final=1.42, loss_mean=1.37, loss_mean_cls=0.113, proj_loss=-0.0552][2026-03-26 12:31:17] Step: 47, Training Logs: loss_final: 1.372617, loss_mean: 1.315681, proj_loss: -0.056606, loss_mean_cls: 0.113542, grad_norm: 0.702552 +Steps: 0%| | 48/1000000 [00:14<68:30:08, 4.05it/s, grad_norm=0.703, loss_final=1.37, loss_mean=1.32, loss_mean_cls=0.114, proj_loss=-0.0566][2026-03-26 12:31:17] Step: 48, Training Logs: loss_final: 1.368565, loss_mean: 1.310973, proj_loss: -0.055468, loss_mean_cls: 0.113059, grad_norm: 0.671382 +Steps: 0%| | 49/1000000 [00:14<68:30:20, 4.05it/s, grad_norm=0.671, loss_final=1.37, loss_mean=1.31, loss_mean_cls=0.113, proj_loss=-0.0555][2026-03-26 12:31:18] Step: 49, Training Logs: loss_final: 1.350346, loss_mean: 1.294427, proj_loss: -0.057010, loss_mean_cls: 0.112928, grad_norm: 0.838704 +Steps: 0%| | 50/1000000 [00:14<68:28:55, 4.06it/s, grad_norm=0.839, loss_final=1.35, loss_mean=1.29, loss_mean_cls=0.113, proj_loss=-0.057][2026-03-26 12:31:18] Step: 50, Training Logs: loss_final: 1.528777, loss_mean: 1.470094, proj_loss: -0.055211, loss_mean_cls: 0.113894, grad_norm: 5.495419 +Steps: 0%| | 51/1000000 [00:15<68:29:54, 4.06it/s, grad_norm=5.5, loss_final=1.53, loss_mean=1.47, loss_mean_cls=0.114, proj_loss=-0.0552][2026-03-26 12:31:18] Step: 51, Training Logs: loss_final: 1.545767, loss_mean: 1.488336, proj_loss: -0.056170, loss_mean_cls: 0.113601, grad_norm: 5.703853 +Steps: 0%| | 52/1000000 [00:15<68:28:41, 4.06it/s, grad_norm=5.7, loss_final=1.55, loss_mean=1.49, loss_mean_cls=0.114, proj_loss=-0.0562][2026-03-26 12:31:18] Step: 52, Training Logs: loss_final: 1.529312, loss_mean: 1.469831, proj_loss: -0.054023, loss_mean_cls: 0.113503, grad_norm: 4.364698 +Steps: 0%| | 53/1000000 [00:15<68:28:32, 4.06it/s, grad_norm=4.36, loss_final=1.53, loss_mean=1.47, loss_mean_cls=0.114, proj_loss=-0.054][2026-03-26 12:31:19] Step: 53, Training Logs: loss_final: 1.485708, loss_mean: 1.429670, proj_loss: -0.056806, loss_mean_cls: 0.112844, grad_norm: 2.956588 +Steps: 0%| | 54/1000000 [00:15<68:32:38, 4.05it/s, grad_norm=2.96, loss_final=1.49, loss_mean=1.43, loss_mean_cls=0.113, proj_loss=-0.0568][2026-03-26 12:31:19] Step: 54, Training Logs: loss_final: 1.448938, loss_mean: 1.389676, proj_loss: -0.053657, loss_mean_cls: 0.112919, grad_norm: 2.338463 +Steps: 0%| | 55/1000000 [00:16<68:31:34, 4.05it/s, grad_norm=2.34, loss_final=1.45, loss_mean=1.39, loss_mean_cls=0.113, proj_loss=-0.0537][2026-03-26 12:31:19] Step: 55, Training Logs: loss_final: 1.372547, loss_mean: 1.312310, proj_loss: -0.052790, loss_mean_cls: 0.113026, grad_norm: 1.420881 +Steps: 0%| | 56/1000000 [00:16<68:29:29, 4.06it/s, grad_norm=1.42, loss_final=1.37, loss_mean=1.31, loss_mean_cls=0.113, proj_loss=-0.0528][2026-03-26 12:31:19] Step: 56, Training Logs: loss_final: 1.360455, loss_mean: 1.305232, proj_loss: -0.056781, loss_mean_cls: 0.112003, grad_norm: 1.270558 +Steps: 0%| | 57/1000000 [00:16<68:26:27, 4.06it/s, grad_norm=1.27, loss_final=1.36, loss_mean=1.31, loss_mean_cls=0.112, proj_loss=-0.0568][2026-03-26 12:31:19] Step: 57, Training Logs: loss_final: 1.341354, loss_mean: 1.284813, proj_loss: -0.056390, loss_mean_cls: 0.112931, grad_norm: 1.480417 +Steps: 0%| | 58/1000000 [00:16<68:28:45, 4.06it/s, grad_norm=1.48, loss_final=1.34, loss_mean=1.28, loss_mean_cls=0.113, proj_loss=-0.0564][2026-03-26 12:31:20] Step: 58, Training Logs: loss_final: 1.363635, loss_mean: 1.305964, proj_loss: -0.054406, loss_mean_cls: 0.112077, grad_norm: 1.406665 +Steps: 0%| | 59/1000000 [00:17<68:28:11, 4.06it/s, grad_norm=1.41, loss_final=1.36, loss_mean=1.31, loss_mean_cls=0.112, proj_loss=-0.0544][2026-03-26 12:31:20] Step: 59, Training Logs: loss_final: 1.331269, loss_mean: 1.275587, proj_loss: -0.057196, loss_mean_cls: 0.112878, grad_norm: 0.836532 +Steps: 0%| | 60/1000000 [00:17<68:25:37, 4.06it/s, grad_norm=0.837, loss_final=1.33, loss_mean=1.28, loss_mean_cls=0.113, proj_loss=-0.0572][2026-03-26 12:31:20] Step: 60, Training Logs: loss_final: 1.333368, loss_mean: 1.277109, proj_loss: -0.056473, loss_mean_cls: 0.112731, grad_norm: 1.154376 +Steps: 0%| | 61/1000000 [00:17<68:24:43, 4.06it/s, grad_norm=1.15, loss_final=1.33, loss_mean=1.28, loss_mean_cls=0.113, proj_loss=-0.0565][2026-03-26 12:31:20] Step: 61, Training Logs: loss_final: 1.310154, loss_mean: 1.253334, proj_loss: -0.055764, loss_mean_cls: 0.112585, grad_norm: 0.950036 +Steps: 0%| | 62/1000000 [00:17<68:26:09, 4.06it/s, grad_norm=0.95, loss_final=1.31, loss_mean=1.25, loss_mean_cls=0.113, proj_loss=-0.0558][2026-03-26 12:31:21] Step: 62, Training Logs: loss_final: 1.335962, loss_mean: 1.278448, proj_loss: -0.054452, loss_mean_cls: 0.111967, grad_norm: 0.817743 +Steps: 0%| | 63/1000000 [00:18<68:26:11, 4.06it/s, grad_norm=0.818, loss_final=1.34, loss_mean=1.28, loss_mean_cls=0.112, proj_loss=-0.0545][2026-03-26 12:31:21] Step: 63, Training Logs: loss_final: 1.336019, loss_mean: 1.280428, proj_loss: -0.056032, loss_mean_cls: 0.111623, grad_norm: 0.711049 +Steps: 0%| | 64/1000000 [00:18<68:24:54, 4.06it/s, grad_norm=0.711, loss_final=1.34, loss_mean=1.28, loss_mean_cls=0.112, proj_loss=-0.056][2026-03-26 12:31:21] Step: 64, Training Logs: loss_final: 1.304106, loss_mean: 1.247556, proj_loss: -0.055629, loss_mean_cls: 0.112179, grad_norm: 0.901901 +Steps: 0%| | 65/1000000 [00:18<68:25:00, 4.06it/s, grad_norm=0.902, loss_final=1.3, loss_mean=1.25, loss_mean_cls=0.112, proj_loss=-0.0556][2026-03-26 12:31:21] Step: 65, Training Logs: loss_final: 1.268339, loss_mean: 1.212154, proj_loss: -0.055885, loss_mean_cls: 0.112071, grad_norm: 0.903228 +Steps: 0%| | 66/1000000 [00:18<68:32:01, 4.05it/s, grad_norm=0.903, loss_final=1.27, loss_mean=1.21, loss_mean_cls=0.112, proj_loss=-0.0559][2026-03-26 12:31:22] Step: 66, Training Logs: loss_final: 1.293975, loss_mean: 1.241298, proj_loss: -0.059217, loss_mean_cls: 0.111893, grad_norm: 0.830619 +Steps: 0%| | 67/1000000 [00:19<68:29:39, 4.06it/s, grad_norm=0.831, loss_final=1.29, loss_mean=1.24, loss_mean_cls=0.112, proj_loss=-0.0592][2026-03-26 12:31:22] Step: 67, Training Logs: loss_final: 1.291745, loss_mean: 1.234950, proj_loss: -0.054458, loss_mean_cls: 0.111253, grad_norm: 0.928774 +Steps: 0%| | 68/1000000 [00:19<68:30:16, 4.05it/s, grad_norm=0.929, loss_final=1.29, loss_mean=1.23, loss_mean_cls=0.111, proj_loss=-0.0545][2026-03-26 12:31:22] Step: 68, Training Logs: loss_final: 1.277053, loss_mean: 1.224011, proj_loss: -0.057731, loss_mean_cls: 0.110774, grad_norm: 0.713197 +Steps: 0%| | 69/1000000 [00:19<68:33:32, 4.05it/s, grad_norm=0.713, loss_final=1.28, loss_mean=1.22, loss_mean_cls=0.111, proj_loss=-0.0577][2026-03-26 12:31:22] Step: 69, Training Logs: loss_final: 1.273278, loss_mean: 1.217088, proj_loss: -0.055009, loss_mean_cls: 0.111198, grad_norm: 0.877212 +Steps: 0%| | 70/1000000 [00:19<68:32:12, 4.05it/s, grad_norm=0.877, loss_final=1.27, loss_mean=1.22, loss_mean_cls=0.111, proj_loss=-0.055][2026-03-26 12:31:23] Step: 70, Training Logs: loss_final: 1.245440, loss_mean: 1.191837, proj_loss: -0.056853, loss_mean_cls: 0.110457, grad_norm: 0.904337 +Steps: 0%| | 71/1000000 [00:20<68:32:26, 4.05it/s, grad_norm=0.904, loss_final=1.25, loss_mean=1.19, loss_mean_cls=0.11, proj_loss=-0.0569][2026-03-26 12:31:23] Step: 71, Training Logs: loss_final: 1.285351, loss_mean: 1.230825, proj_loss: -0.056440, loss_mean_cls: 0.110966, grad_norm: 1.481927 +Steps: 0%| | 72/1000000 [00:20<68:31:10, 4.05it/s, grad_norm=1.48, loss_final=1.29, loss_mean=1.23, loss_mean_cls=0.111, proj_loss=-0.0564][2026-03-26 12:31:23] Step: 72, Training Logs: loss_final: 1.220358, loss_mean: 1.164082, proj_loss: -0.055503, loss_mean_cls: 0.111779, grad_norm: 1.145216 +Steps: 0%| | 73/1000000 [00:20<68:32:18, 4.05it/s, grad_norm=1.15, loss_final=1.22, loss_mean=1.16, loss_mean_cls=0.112, proj_loss=-0.0555][2026-03-26 12:31:23] Step: 73, Training Logs: loss_final: 1.222753, loss_mean: 1.167938, proj_loss: -0.055525, loss_mean_cls: 0.110340, grad_norm: 1.036319 +Steps: 0%| | 74/1000000 [00:20<68:31:20, 4.05it/s, grad_norm=1.04, loss_final=1.22, loss_mean=1.17, loss_mean_cls=0.11, proj_loss=-0.0555][2026-03-26 12:31:24] Step: 74, Training Logs: loss_final: 1.289667, loss_mean: 1.234167, proj_loss: -0.056104, loss_mean_cls: 0.111604, grad_norm: 3.436272 +Steps: 0%| | 75/1000000 [00:21<68:29:26, 4.06it/s, grad_norm=3.44, loss_final=1.29, loss_mean=1.23, loss_mean_cls=0.112, proj_loss=-0.0561][2026-03-26 12:31:24] Step: 75, Training Logs: loss_final: 1.272270, loss_mean: 1.218367, proj_loss: -0.057167, loss_mean_cls: 0.111071, grad_norm: 2.859861 +Steps: 0%| | 76/1000000 [00:21<68:29:44, 4.06it/s, grad_norm=2.86, loss_final=1.27, loss_mean=1.22, loss_mean_cls=0.111, proj_loss=-0.0572][2026-03-26 12:31:24] Step: 76, Training Logs: loss_final: 1.246748, loss_mean: 1.190513, proj_loss: -0.055522, loss_mean_cls: 0.111757, grad_norm: 1.770109 +Steps: 0%| | 77/1000000 [00:21<68:32:50, 4.05it/s, grad_norm=1.77, loss_final=1.25, loss_mean=1.19, loss_mean_cls=0.112, proj_loss=-0.0555][2026-03-26 12:31:24] Step: 77, Training Logs: loss_final: 1.269541, loss_mean: 1.216386, proj_loss: -0.056962, loss_mean_cls: 0.110117, grad_norm: 1.632940 +Steps: 0%| | 78/1000000 [00:21<68:30:19, 4.05it/s, grad_norm=1.63, loss_final=1.27, loss_mean=1.22, loss_mean_cls=0.11, proj_loss=-0.057][2026-03-26 12:31:25] Step: 78, Training Logs: loss_final: 1.224685, loss_mean: 1.170564, proj_loss: -0.057046, loss_mean_cls: 0.111167, grad_norm: 1.036261 +Steps: 0%| | 79/1000000 [00:22<68:30:29, 4.05it/s, grad_norm=1.04, loss_final=1.22, loss_mean=1.17, loss_mean_cls=0.111, proj_loss=-0.057][2026-03-26 12:31:25] Step: 79, Training Logs: loss_final: 1.175630, loss_mean: 1.122066, proj_loss: -0.056768, loss_mean_cls: 0.110332, grad_norm: 0.952835 +Steps: 0%| | 80/1000000 [00:22<68:31:51, 4.05it/s, grad_norm=0.953, loss_final=1.18, loss_mean=1.12, loss_mean_cls=0.11, proj_loss=-0.0568][2026-03-26 12:31:25] Step: 80, Training Logs: loss_final: 1.215047, loss_mean: 1.160970, proj_loss: -0.056096, loss_mean_cls: 0.110173, grad_norm: 1.286654 +Steps: 0%| | 81/1000000 [00:22<68:31:56, 4.05it/s, grad_norm=1.29, loss_final=1.22, loss_mean=1.16, loss_mean_cls=0.11, proj_loss=-0.0561][2026-03-26 12:31:25] Step: 81, Training Logs: loss_final: 1.258787, loss_mean: 1.206628, proj_loss: -0.058400, loss_mean_cls: 0.110559, grad_norm: 1.849372 +Steps: 0%| | 82/1000000 [00:22<68:32:03, 4.05it/s, grad_norm=1.85, loss_final=1.26, loss_mean=1.21, loss_mean_cls=0.111, proj_loss=-0.0584][2026-03-26 12:31:26] Step: 82, Training Logs: loss_final: 1.188258, loss_mean: 1.134863, proj_loss: -0.057671, loss_mean_cls: 0.111065, grad_norm: 0.915613 +Steps: 0%| | 83/1000000 [00:23<68:33:35, 4.05it/s, grad_norm=0.916, loss_final=1.19, loss_mean=1.13, loss_mean_cls=0.111, proj_loss=-0.0577][2026-03-26 12:31:26] Step: 83, Training Logs: loss_final: 1.221325, loss_mean: 1.165980, proj_loss: -0.055194, loss_mean_cls: 0.110539, grad_norm: 1.247545 +Steps: 0%| | 84/1000000 [00:23<68:35:30, 4.05it/s, grad_norm=1.25, loss_final=1.22, loss_mean=1.17, loss_mean_cls=0.111, proj_loss=-0.0552][2026-03-26 12:31:26] Step: 84, Training Logs: loss_final: 1.221409, loss_mean: 1.168155, proj_loss: -0.057535, loss_mean_cls: 0.110790, grad_norm: 1.124083 +Steps: 0%| | 85/1000000 [00:23<68:34:41, 4.05it/s, grad_norm=1.12, loss_final=1.22, loss_mean=1.17, loss_mean_cls=0.111, proj_loss=-0.0575][2026-03-26 12:31:26] Step: 85, Training Logs: loss_final: 1.232134, loss_mean: 1.176702, proj_loss: -0.054494, loss_mean_cls: 0.109926, grad_norm: 0.972285 +Steps: 0%| | 86/1000000 [00:23<68:36:51, 4.05it/s, grad_norm=0.972, loss_final=1.23, loss_mean=1.18, loss_mean_cls=0.11, proj_loss=-0.0545][2026-03-26 12:31:27] Step: 86, Training Logs: loss_final: 1.238377, loss_mean: 1.185558, proj_loss: -0.056352, loss_mean_cls: 0.109171, grad_norm: 0.942116 +Steps: 0%| | 87/1000000 [00:23<68:37:34, 4.05it/s, grad_norm=0.942, loss_final=1.24, loss_mean=1.19, loss_mean_cls=0.109, proj_loss=-0.0564][2026-03-26 12:31:27] Step: 87, Training Logs: loss_final: 1.200011, loss_mean: 1.148143, proj_loss: -0.056962, loss_mean_cls: 0.108830, grad_norm: 0.713559 +Steps: 0%| | 88/1000000 [00:24<68:37:05, 4.05it/s, grad_norm=0.714, loss_final=1.2, loss_mean=1.15, loss_mean_cls=0.109, proj_loss=-0.057][2026-03-26 12:31:27] Step: 88, Training Logs: loss_final: 1.206771, loss_mean: 1.153441, proj_loss: -0.056850, loss_mean_cls: 0.110180, grad_norm: 0.685349 +Steps: 0%| | 89/1000000 [00:24<68:33:05, 4.05it/s, grad_norm=0.685, loss_final=1.21, loss_mean=1.15, loss_mean_cls=0.11, proj_loss=-0.0569][2026-03-26 12:31:27] Step: 89, Training Logs: loss_final: 1.192860, loss_mean: 1.141545, proj_loss: -0.058279, loss_mean_cls: 0.109593, grad_norm: 1.346806 +Steps: 0%| | 90/1000000 [00:24<68:33:33, 4.05it/s, grad_norm=1.35, loss_final=1.19, loss_mean=1.14, loss_mean_cls=0.11, proj_loss=-0.0583][2026-03-26 12:31:28] Step: 90, Training Logs: loss_final: 1.166398, loss_mean: 1.114082, proj_loss: -0.057711, loss_mean_cls: 0.110027, grad_norm: 1.251912 +Steps: 0%| | 91/1000000 [00:24<68:30:42, 4.05it/s, grad_norm=1.25, loss_final=1.17, loss_mean=1.11, loss_mean_cls=0.11, proj_loss=-0.0577][2026-03-26 12:31:28] Step: 91, Training Logs: loss_final: 1.170440, loss_mean: 1.118997, proj_loss: -0.057921, loss_mean_cls: 0.109364, grad_norm: 1.367671 +Steps: 0%| | 92/1000000 [00:25<68:36:45, 4.05it/s, grad_norm=1.37, loss_final=1.17, loss_mean=1.12, loss_mean_cls=0.109, proj_loss=-0.0579][2026-03-26 12:31:28] Step: 92, Training Logs: loss_final: 1.163484, loss_mean: 1.112976, proj_loss: -0.058365, loss_mean_cls: 0.108873, grad_norm: 0.763432 +Steps: 0%| | 93/1000000 [00:25<68:32:45, 4.05it/s, grad_norm=0.763, loss_final=1.16, loss_mean=1.11, loss_mean_cls=0.109, proj_loss=-0.0584][2026-03-26 12:31:28] Step: 93, Training Logs: loss_final: 1.178396, loss_mean: 1.127202, proj_loss: -0.058560, loss_mean_cls: 0.109754, grad_norm: 1.000902 +Steps: 0%| | 94/1000000 [00:25<68:47:21, 4.04it/s, grad_norm=1, loss_final=1.18, loss_mean=1.13, loss_mean_cls=0.11, proj_loss=-0.0586][2026-03-26 12:31:29] Step: 94, Training Logs: loss_final: 1.175437, loss_mean: 1.126130, proj_loss: -0.059551, loss_mean_cls: 0.108859, grad_norm: 0.860972 +Steps: 0%| | 95/1000000 [00:25<68:44:29, 4.04it/s, grad_norm=0.861, loss_final=1.18, loss_mean=1.13, loss_mean_cls=0.109, proj_loss=-0.0596][2026-03-26 12:31:29] Step: 95, Training Logs: loss_final: 1.152767, loss_mean: 1.097090, proj_loss: -0.054695, loss_mean_cls: 0.110372, grad_norm: 0.598102 +Steps: 0%| | 96/1000000 [00:26<68:41:21, 4.04it/s, grad_norm=0.598, loss_final=1.15, loss_mean=1.1, loss_mean_cls=0.11, proj_loss=-0.0547][2026-03-26 12:31:29] Step: 96, Training Logs: loss_final: 1.155106, loss_mean: 1.103692, proj_loss: -0.057617, loss_mean_cls: 0.109031, grad_norm: 0.606026 +Steps: 0%| | 97/1000000 [00:26<68:38:11, 4.05it/s, grad_norm=0.606, loss_final=1.16, loss_mean=1.1, loss_mean_cls=0.109, proj_loss=-0.0576][2026-03-26 12:31:29] Step: 97, Training Logs: loss_final: 1.166581, loss_mean: 1.115146, proj_loss: -0.057432, loss_mean_cls: 0.108867, grad_norm: 0.512892 +Steps: 0%| | 98/1000000 [00:26<68:40:48, 4.04it/s, grad_norm=0.513, loss_final=1.17, loss_mean=1.12, loss_mean_cls=0.109, proj_loss=-0.0574][2026-03-26 12:31:30] Step: 98, Training Logs: loss_final: 1.170160, loss_mean: 1.119404, proj_loss: -0.057705, loss_mean_cls: 0.108461, grad_norm: 0.631401 +Steps: 0%| | 99/1000000 [00:26<68:37:16, 4.05it/s, grad_norm=0.631, loss_final=1.17, loss_mean=1.12, loss_mean_cls=0.108, proj_loss=-0.0577][2026-03-26 12:31:30] Step: 99, Training Logs: loss_final: 1.134508, loss_mean: 1.081762, proj_loss: -0.056654, loss_mean_cls: 0.109401, grad_norm: 0.552610 +Steps: 0%| | 100/1000000 [00:27<68:41:03, 4.04it/s, grad_norm=0.553, loss_final=1.13, loss_mean=1.08, loss_mean_cls=0.109, proj_loss=-0.0567][2026-03-26 12:31:30] Step: 100, Training Logs: loss_final: 1.149402, loss_mean: 1.098173, proj_loss: -0.058181, loss_mean_cls: 0.109411, grad_norm: 0.651845 +Steps: 0%| | 101/1000000 [00:27<68:36:54, 4.05it/s, grad_norm=0.652, loss_final=1.15, loss_mean=1.1, loss_mean_cls=0.109, proj_loss=-0.0582][2026-03-26 12:31:30] Step: 101, Training Logs: loss_final: 1.134636, loss_mean: 1.082225, proj_loss: -0.055971, loss_mean_cls: 0.108382, grad_norm: 0.537758 +Steps: 0%| | 102/1000000 [00:27<68:36:53, 4.05it/s, grad_norm=0.538, loss_final=1.13, loss_mean=1.08, loss_mean_cls=0.108, proj_loss=-0.056][2026-03-26 12:31:31] Step: 102, Training Logs: loss_final: 1.109870, loss_mean: 1.055017, proj_loss: -0.055033, loss_mean_cls: 0.109885, grad_norm: 0.559971 +Steps: 0%| | 103/1000000 [00:27<68:37:26, 4.05it/s, grad_norm=0.56, loss_final=1.11, loss_mean=1.06, loss_mean_cls=0.11, proj_loss=-0.055][2026-03-26 12:31:31] Step: 103, Training Logs: loss_final: 1.129662, loss_mean: 1.079390, proj_loss: -0.058598, loss_mean_cls: 0.108870, grad_norm: 0.566579 +Steps: 0%| | 104/1000000 [00:28<68:35:33, 4.05it/s, grad_norm=0.567, loss_final=1.13, loss_mean=1.08, loss_mean_cls=0.109, proj_loss=-0.0586][2026-03-26 12:31:31] Step: 104, Training Logs: loss_final: 1.115328, loss_mean: 1.064379, proj_loss: -0.057644, loss_mean_cls: 0.108592, grad_norm: 0.379436 +Steps: 0%| | 105/1000000 [00:28<68:35:50, 4.05it/s, grad_norm=0.379, loss_final=1.12, loss_mean=1.06, loss_mean_cls=0.109, proj_loss=-0.0576][2026-03-26 12:31:31] Step: 105, Training Logs: loss_final: 1.121608, loss_mean: 1.066465, proj_loss: -0.054576, loss_mean_cls: 0.109719, grad_norm: 0.642051 +Steps: 0%| | 106/1000000 [00:28<68:36:10, 4.05it/s, grad_norm=0.642, loss_final=1.12, loss_mean=1.07, loss_mean_cls=0.11, proj_loss=-0.0546][2026-03-26 12:31:32] Step: 106, Training Logs: loss_final: 1.114737, loss_mean: 1.064877, proj_loss: -0.059202, loss_mean_cls: 0.109062, grad_norm: 0.511954 +Steps: 0%| | 107/1000000 [00:28<68:34:38, 4.05it/s, grad_norm=0.512, loss_final=1.11, loss_mean=1.06, loss_mean_cls=0.109, proj_loss=-0.0592][2026-03-26 12:31:32] Step: 107, Training Logs: loss_final: 1.118678, loss_mean: 1.068141, proj_loss: -0.057795, loss_mean_cls: 0.108331, grad_norm: 0.508057 +Steps: 0%| | 108/1000000 [00:29<68:33:25, 4.05it/s, grad_norm=0.508, loss_final=1.12, loss_mean=1.07, loss_mean_cls=0.108, proj_loss=-0.0578][2026-03-26 12:31:32] Step: 108, Training Logs: loss_final: 1.097240, loss_mean: 1.049366, proj_loss: -0.059131, loss_mean_cls: 0.107005, grad_norm: 0.424846 +Steps: 0%| | 109/1000000 [00:29<68:37:37, 4.05it/s, grad_norm=0.425, loss_final=1.1, loss_mean=1.05, loss_mean_cls=0.107, proj_loss=-0.0591][2026-03-26 12:31:32] Step: 109, Training Logs: loss_final: 1.125956, loss_mean: 1.074565, proj_loss: -0.056749, loss_mean_cls: 0.108140, grad_norm: 0.478142 +Steps: 0%| | 110/1000000 [00:29<68:36:14, 4.05it/s, grad_norm=0.478, loss_final=1.13, loss_mean=1.07, loss_mean_cls=0.108, proj_loss=-0.0567][2026-03-26 12:31:33] Step: 110, Training Logs: loss_final: 1.091241, loss_mean: 1.040059, proj_loss: -0.057226, loss_mean_cls: 0.108408, grad_norm: 0.411122 +Steps: 0%| | 111/1000000 [00:29<68:34:57, 4.05it/s, grad_norm=0.411, loss_final=1.09, loss_mean=1.04, loss_mean_cls=0.108, proj_loss=-0.0572][2026-03-26 12:31:33] Step: 111, Training Logs: loss_final: 1.091349, loss_mean: 1.040080, proj_loss: -0.057852, loss_mean_cls: 0.109121, grad_norm: 0.431137 +Steps: 0%| | 112/1000000 [00:30<68:35:31, 4.05it/s, grad_norm=0.431, loss_final=1.09, loss_mean=1.04, loss_mean_cls=0.109, proj_loss=-0.0579][2026-03-26 12:31:33] Step: 112, Training Logs: loss_final: 1.083756, loss_mean: 1.033347, proj_loss: -0.058504, loss_mean_cls: 0.108913, grad_norm: 0.527481 +Steps: 0%| | 113/1000000 [00:30<68:35:56, 4.05it/s, grad_norm=0.527, loss_final=1.08, loss_mean=1.03, loss_mean_cls=0.109, proj_loss=-0.0585][2026-03-26 12:31:33] Step: 113, Training Logs: loss_final: 1.081557, loss_mean: 1.032119, proj_loss: -0.058913, loss_mean_cls: 0.108351, grad_norm: 0.333903 +Steps: 0%| | 114/1000000 [00:30<68:36:46, 4.05it/s, grad_norm=0.334, loss_final=1.08, loss_mean=1.03, loss_mean_cls=0.108, proj_loss=-0.0589][2026-03-26 12:31:34] Step: 114, Training Logs: loss_final: 1.095187, loss_mean: 1.042371, proj_loss: -0.055702, loss_mean_cls: 0.108518, grad_norm: 0.635450 +Steps: 0%| | 115/1000000 [00:30<68:37:38, 4.05it/s, grad_norm=0.635, loss_final=1.1, loss_mean=1.04, loss_mean_cls=0.109, proj_loss=-0.0557][2026-03-26 12:31:34] Step: 115, Training Logs: loss_final: 1.110024, loss_mean: 1.061346, proj_loss: -0.059385, loss_mean_cls: 0.108064, grad_norm: 0.570668 +Steps: 0%| | 116/1000000 [00:31<68:39:33, 4.05it/s, grad_norm=0.571, loss_final=1.11, loss_mean=1.06, loss_mean_cls=0.108, proj_loss=-0.0594][2026-03-26 12:31:34] Step: 116, Training Logs: loss_final: 1.111026, loss_mean: 1.059694, proj_loss: -0.056001, loss_mean_cls: 0.107332, grad_norm: 0.461403 +Steps: 0%| | 117/1000000 [00:31<68:38:48, 4.05it/s, grad_norm=0.461, loss_final=1.11, loss_mean=1.06, loss_mean_cls=0.107, proj_loss=-0.056][2026-03-26 12:31:34] Step: 117, Training Logs: loss_final: 1.092383, loss_mean: 1.041848, proj_loss: -0.056481, loss_mean_cls: 0.107015, grad_norm: 0.561504 +Steps: 0%| | 118/1000000 [00:31<68:35:36, 4.05it/s, grad_norm=0.562, loss_final=1.09, loss_mean=1.04, loss_mean_cls=0.107, proj_loss=-0.0565][2026-03-26 12:31:35] Step: 118, Training Logs: loss_final: 1.097607, loss_mean: 1.046873, proj_loss: -0.056681, loss_mean_cls: 0.107415, grad_norm: 0.336896 +Steps: 0%| | 119/1000000 [00:31<68:39:55, 4.04it/s, grad_norm=0.337, loss_final=1.1, loss_mean=1.05, loss_mean_cls=0.107, proj_loss=-0.0567][2026-03-26 12:31:35] Step: 119, Training Logs: loss_final: 1.101323, loss_mean: 1.050875, proj_loss: -0.057353, loss_mean_cls: 0.107802, grad_norm: 0.512678 +Steps: 0%| | 120/1000000 [00:32<68:45:30, 4.04it/s, grad_norm=0.513, loss_final=1.1, loss_mean=1.05, loss_mean_cls=0.108, proj_loss=-0.0574][2026-03-26 12:31:35] Step: 120, Training Logs: loss_final: 1.090719, loss_mean: 1.040358, proj_loss: -0.057579, loss_mean_cls: 0.107940, grad_norm: 0.647719 +Steps: 0%| | 121/1000000 [00:32<68:45:37, 4.04it/s, grad_norm=0.648, loss_final=1.09, loss_mean=1.04, loss_mean_cls=0.108, proj_loss=-0.0576][2026-03-26 12:31:35] Step: 121, Training Logs: loss_final: 1.100960, loss_mean: 1.050641, proj_loss: -0.056450, loss_mean_cls: 0.106769, grad_norm: 0.740847 +Steps: 0%| | 122/1000000 [00:32<68:48:54, 4.04it/s, grad_norm=0.741, loss_final=1.1, loss_mean=1.05, loss_mean_cls=0.107, proj_loss=-0.0564][2026-03-26 12:31:36] Step: 122, Training Logs: loss_final: 1.094331, loss_mean: 1.043633, proj_loss: -0.057119, loss_mean_cls: 0.107816, grad_norm: 0.703549 +Steps: 0%| | 123/1000000 [00:32<68:47:09, 4.04it/s, grad_norm=0.704, loss_final=1.09, loss_mean=1.04, loss_mean_cls=0.108, proj_loss=-0.0571][2026-03-26 12:31:36] Step: 123, Training Logs: loss_final: 1.074613, loss_mean: 1.023621, proj_loss: -0.057242, loss_mean_cls: 0.108233, grad_norm: 0.777318 +Steps: 0%| | 124/1000000 [00:33<68:44:58, 4.04it/s, grad_norm=0.777, loss_final=1.07, loss_mean=1.02, loss_mean_cls=0.108, proj_loss=-0.0572][2026-03-26 12:31:36] Step: 124, Training Logs: loss_final: 1.100855, loss_mean: 1.049433, proj_loss: -0.056038, loss_mean_cls: 0.107459, grad_norm: 0.896371 +Steps: 0%| | 125/1000000 [00:33<68:42:29, 4.04it/s, grad_norm=0.896, loss_final=1.1, loss_mean=1.05, loss_mean_cls=0.107, proj_loss=-0.056][2026-03-26 12:31:36] Step: 125, Training Logs: loss_final: 1.067755, loss_mean: 1.013157, proj_loss: -0.053211, loss_mean_cls: 0.107808, grad_norm: 0.432621 +Steps: 0%| | 126/1000000 [00:33<68:44:34, 4.04it/s, grad_norm=0.433, loss_final=1.07, loss_mean=1.01, loss_mean_cls=0.108, proj_loss=-0.0532][2026-03-26 12:31:37] Step: 126, Training Logs: loss_final: 1.080500, loss_mean: 1.032326, proj_loss: -0.057975, loss_mean_cls: 0.106150, grad_norm: 0.608260 +Steps: 0%| | 127/1000000 [00:33<68:40:52, 4.04it/s, grad_norm=0.608, loss_final=1.08, loss_mean=1.03, loss_mean_cls=0.106, proj_loss=-0.058][2026-03-26 12:31:37] Step: 127, Training Logs: loss_final: 1.047237, loss_mean: 0.993018, proj_loss: -0.054335, loss_mean_cls: 0.108555, grad_norm: 0.472418 +Steps: 0%| | 128/1000000 [00:34<68:41:31, 4.04it/s, grad_norm=0.472, loss_final=1.05, loss_mean=0.993, loss_mean_cls=0.109, proj_loss=-0.0543][2026-03-26 12:31:37] Step: 128, Training Logs: loss_final: 1.080488, loss_mean: 1.032159, proj_loss: -0.058632, loss_mean_cls: 0.106961, grad_norm: 0.573020 +Steps: 0%| | 129/1000000 [00:34<68:39:54, 4.04it/s, grad_norm=0.573, loss_final=1.08, loss_mean=1.03, loss_mean_cls=0.107, proj_loss=-0.0586][2026-03-26 12:31:37] Step: 129, Training Logs: loss_final: 1.101775, loss_mean: 1.052473, proj_loss: -0.057114, loss_mean_cls: 0.106416, grad_norm: 0.466397 +Steps: 0%| | 130/1000000 [00:34<68:37:07, 4.05it/s, grad_norm=0.466, loss_final=1.1, loss_mean=1.05, loss_mean_cls=0.106, proj_loss=-0.0571][2026-03-26 12:31:38] Step: 130, Training Logs: loss_final: 1.072294, loss_mean: 1.019119, proj_loss: -0.054501, loss_mean_cls: 0.107676, grad_norm: 0.452210 +Steps: 0%| | 131/1000000 [00:34<68:36:50, 4.05it/s, grad_norm=0.452, loss_final=1.07, loss_mean=1.02, loss_mean_cls=0.108, proj_loss=-0.0545][2026-03-26 12:31:38] Step: 131, Training Logs: loss_final: 1.074273, loss_mean: 1.024463, proj_loss: -0.057019, loss_mean_cls: 0.106829, grad_norm: 0.666487 +Steps: 0%| | 132/1000000 [00:35<68:34:50, 4.05it/s, grad_norm=0.666, loss_final=1.07, loss_mean=1.02, loss_mean_cls=0.107, proj_loss=-0.057][2026-03-26 12:31:38] Step: 132, Training Logs: loss_final: 1.077711, loss_mean: 1.028686, proj_loss: -0.057624, loss_mean_cls: 0.106649, grad_norm: 0.490701 +Steps: 0%| | 133/1000000 [00:35<68:33:41, 4.05it/s, grad_norm=0.491, loss_final=1.08, loss_mean=1.03, loss_mean_cls=0.107, proj_loss=-0.0576][2026-03-26 12:31:38] Step: 133, Training Logs: loss_final: 1.089618, loss_mean: 1.041807, proj_loss: -0.057296, loss_mean_cls: 0.105107, grad_norm: 0.478508 +Steps: 0%| | 134/1000000 [00:35<68:34:15, 4.05it/s, grad_norm=0.479, loss_final=1.09, loss_mean=1.04, loss_mean_cls=0.105, proj_loss=-0.0573][2026-03-26 12:31:39] Step: 134, Training Logs: loss_final: 1.054472, loss_mean: 1.004816, proj_loss: -0.057369, loss_mean_cls: 0.107025, grad_norm: 0.447318 +Steps: 0%| | 135/1000000 [00:35<68:38:44, 4.05it/s, grad_norm=0.447, loss_final=1.05, loss_mean=1, loss_mean_cls=0.107, proj_loss=-0.0574][2026-03-26 12:31:39] Step: 135, Training Logs: loss_final: 1.079052, loss_mean: 1.032238, proj_loss: -0.058380, loss_mean_cls: 0.105194, grad_norm: 0.609700 +Steps: 0%| | 136/1000000 [00:36<68:39:38, 4.05it/s, grad_norm=0.61, loss_final=1.08, loss_mean=1.03, loss_mean_cls=0.105, proj_loss=-0.0584][2026-03-26 12:31:39] Step: 136, Training Logs: loss_final: 1.037418, loss_mean: 0.987586, proj_loss: -0.057885, loss_mean_cls: 0.107716, grad_norm: 0.488968 +Steps: 0%| | 137/1000000 [00:36<68:41:03, 4.04it/s, grad_norm=0.489, loss_final=1.04, loss_mean=0.988, loss_mean_cls=0.108, proj_loss=-0.0579][2026-03-26 12:31:39] Step: 137, Training Logs: loss_final: 1.047126, loss_mean: 0.997487, proj_loss: -0.057481, loss_mean_cls: 0.107119, grad_norm: 0.576666 +Steps: 0%| | 138/1000000 [00:36<68:41:04, 4.04it/s, grad_norm=0.577, loss_final=1.05, loss_mean=0.997, loss_mean_cls=0.107, proj_loss=-0.0575][2026-03-26 12:31:39] Step: 138, Training Logs: loss_final: 1.068466, loss_mean: 1.019352, proj_loss: -0.057216, loss_mean_cls: 0.106330, grad_norm: 0.592344 +Steps: 0%| | 139/1000000 [00:36<68:44:13, 4.04it/s, grad_norm=0.592, loss_final=1.07, loss_mean=1.02, loss_mean_cls=0.106, proj_loss=-0.0572][2026-03-26 12:31:40] Step: 139, Training Logs: loss_final: 1.047829, loss_mean: 0.998686, proj_loss: -0.056623, loss_mean_cls: 0.105767, grad_norm: 0.512050 +Steps: 0%| | 140/1000000 [00:37<68:43:21, 4.04it/s, grad_norm=0.512, loss_final=1.05, loss_mean=0.999, loss_mean_cls=0.106, proj_loss=-0.0566][2026-03-26 12:31:40] Step: 140, Training Logs: loss_final: 1.050622, loss_mean: 1.001688, proj_loss: -0.057124, loss_mean_cls: 0.106058, grad_norm: 0.731481 +Steps: 0%| | 141/1000000 [00:37<68:42:15, 4.04it/s, grad_norm=0.731, loss_final=1.05, loss_mean=1, loss_mean_cls=0.106, proj_loss=-0.0571][2026-03-26 12:31:40] Step: 141, Training Logs: loss_final: 1.045911, loss_mean: 0.995527, proj_loss: -0.056203, loss_mean_cls: 0.106587, grad_norm: 0.559935 +Steps: 0%| | 142/1000000 [00:37<68:42:28, 4.04it/s, grad_norm=0.56, loss_final=1.05, loss_mean=0.996, loss_mean_cls=0.107, proj_loss=-0.0562][2026-03-26 12:31:40] Step: 142, Training Logs: loss_final: 1.028471, loss_mean: 0.978568, proj_loss: -0.057450, loss_mean_cls: 0.107353, grad_norm: 0.668990 +Steps: 0%| | 143/1000000 [00:37<68:40:27, 4.04it/s, grad_norm=0.669, loss_final=1.03, loss_mean=0.979, loss_mean_cls=0.107, proj_loss=-0.0575][2026-03-26 12:31:41] Step: 143, Training Logs: loss_final: 1.025816, loss_mean: 0.977182, proj_loss: -0.058640, loss_mean_cls: 0.107274, grad_norm: 0.517329 +Steps: 0%| | 144/1000000 [00:38<68:42:18, 4.04it/s, grad_norm=0.517, loss_final=1.03, loss_mean=0.977, loss_mean_cls=0.107, proj_loss=-0.0586][2026-03-26 12:31:41] Step: 144, Training Logs: loss_final: 1.056857, loss_mean: 1.008102, proj_loss: -0.057124, loss_mean_cls: 0.105879, grad_norm: 0.872304 +Steps: 0%| | 145/1000000 [00:38<68:41:33, 4.04it/s, grad_norm=0.872, loss_final=1.06, loss_mean=1.01, loss_mean_cls=0.106, proj_loss=-0.0571][2026-03-26 12:31:41] Step: 145, Training Logs: loss_final: 1.061423, loss_mean: 1.012796, proj_loss: -0.056338, loss_mean_cls: 0.104965, grad_norm: 0.973191 +Steps: 0%| | 146/1000000 [00:38<68:43:17, 4.04it/s, grad_norm=0.973, loss_final=1.06, loss_mean=1.01, loss_mean_cls=0.105, proj_loss=-0.0563][2026-03-26 12:31:41] Step: 146, Training Logs: loss_final: 1.029460, loss_mean: 0.979104, proj_loss: -0.056039, loss_mean_cls: 0.106395, grad_norm: 0.700436 +Steps: 0%| | 147/1000000 [00:38<68:40:42, 4.04it/s, grad_norm=0.7, loss_final=1.03, loss_mean=0.979, loss_mean_cls=0.106, proj_loss=-0.056][2026-03-26 12:31:42] Step: 147, Training Logs: loss_final: 1.087413, loss_mean: 1.038025, proj_loss: -0.056654, loss_mean_cls: 0.106042, grad_norm: 0.956897 +Steps: 0%| | 148/1000000 [00:39<68:39:47, 4.04it/s, grad_norm=0.957, loss_final=1.09, loss_mean=1.04, loss_mean_cls=0.106, proj_loss=-0.0567][2026-03-26 12:31:42] Step: 148, Training Logs: loss_final: 1.036939, loss_mean: 0.987621, proj_loss: -0.057066, loss_mean_cls: 0.106384, grad_norm: 0.558876 +Steps: 0%| | 149/1000000 [00:39<68:47:20, 4.04it/s, grad_norm=0.559, loss_final=1.04, loss_mean=0.988, loss_mean_cls=0.106, proj_loss=-0.0571][2026-03-26 12:31:42] Step: 149, Training Logs: loss_final: 1.056204, loss_mean: 1.007969, proj_loss: -0.057384, loss_mean_cls: 0.105618, grad_norm: 0.742868 +Steps: 0%| | 150/1000000 [00:39<68:52:18, 4.03it/s, grad_norm=0.743, loss_final=1.06, loss_mean=1.01, loss_mean_cls=0.106, proj_loss=-0.0574][2026-03-26 12:31:42] Step: 150, Training Logs: loss_final: 1.058900, loss_mean: 1.010366, proj_loss: -0.056838, loss_mean_cls: 0.105372, grad_norm: 0.739378 +Steps: 0%| | 151/1000000 [00:39<68:57:09, 4.03it/s, grad_norm=0.739, loss_final=1.06, loss_mean=1.01, loss_mean_cls=0.105, proj_loss=-0.0568][2026-03-26 12:31:43] Step: 151, Training Logs: loss_final: 1.045360, loss_mean: 0.995290, proj_loss: -0.054630, loss_mean_cls: 0.104701, grad_norm: 0.664568 +Steps: 0%| | 152/1000000 [00:40<68:59:59, 4.03it/s, grad_norm=0.665, loss_final=1.05, loss_mean=0.995, loss_mean_cls=0.105, proj_loss=-0.0546][2026-03-26 12:31:43] Step: 152, Training Logs: loss_final: 1.030435, loss_mean: 0.979209, proj_loss: -0.054913, loss_mean_cls: 0.106139, grad_norm: 0.607004 +Steps: 0%| | 153/1000000 [00:40<68:58:09, 4.03it/s, grad_norm=0.607, loss_final=1.03, loss_mean=0.979, loss_mean_cls=0.106, proj_loss=-0.0549][2026-03-26 12:31:43] Step: 153, Training Logs: loss_final: 1.027020, loss_mean: 0.979031, proj_loss: -0.058208, loss_mean_cls: 0.106197, grad_norm: 0.634553 +Steps: 0%| | 154/1000000 [00:40<69:00:11, 4.02it/s, grad_norm=0.635, loss_final=1.03, loss_mean=0.979, loss_mean_cls=0.106, proj_loss=-0.0582][2026-03-26 12:31:43] Step: 154, Training Logs: loss_final: 1.032125, loss_mean: 0.982889, proj_loss: -0.055725, loss_mean_cls: 0.104961, grad_norm: 0.593822 +Steps: 0%| | 155/1000000 [00:40<68:57:45, 4.03it/s, grad_norm=0.594, loss_final=1.03, loss_mean=0.983, loss_mean_cls=0.105, proj_loss=-0.0557][2026-03-26 12:31:44] Step: 155, Training Logs: loss_final: 1.067990, loss_mean: 1.020609, proj_loss: -0.056638, loss_mean_cls: 0.104019, grad_norm: 0.657880 +Steps: 0%| | 156/1000000 [00:41<68:58:36, 4.03it/s, grad_norm=0.658, loss_final=1.07, loss_mean=1.02, loss_mean_cls=0.104, proj_loss=-0.0566][2026-03-26 12:31:44] Step: 156, Training Logs: loss_final: 1.054000, loss_mean: 1.006099, proj_loss: -0.057281, loss_mean_cls: 0.105182, grad_norm: 0.605363 +Steps: 0%| | 157/1000000 [00:41<68:59:49, 4.03it/s, grad_norm=0.605, loss_final=1.05, loss_mean=1.01, loss_mean_cls=0.105, proj_loss=-0.0573][2026-03-26 12:31:44] Step: 157, Training Logs: loss_final: 1.053500, loss_mean: 1.007918, proj_loss: -0.058201, loss_mean_cls: 0.103783, grad_norm: 0.722253 +Steps: 0%| | 158/1000000 [00:41<69:00:30, 4.02it/s, grad_norm=0.722, loss_final=1.05, loss_mean=1.01, loss_mean_cls=0.104, proj_loss=-0.0582][2026-03-26 12:31:44] Step: 158, Training Logs: loss_final: 1.032212, loss_mean: 0.985206, proj_loss: -0.058161, loss_mean_cls: 0.105167, grad_norm: 0.563597 +Steps: 0%| | 159/1000000 [00:41<69:00:42, 4.02it/s, grad_norm=0.564, loss_final=1.03, loss_mean=0.985, loss_mean_cls=0.105, proj_loss=-0.0582][2026-03-26 12:31:45] Step: 159, Training Logs: loss_final: 1.052639, loss_mean: 1.002140, proj_loss: -0.054765, loss_mean_cls: 0.105265, grad_norm: 0.733957 +Steps: 0%| | 160/1000000 [00:42<68:59:41, 4.03it/s, grad_norm=0.734, loss_final=1.05, loss_mean=1, loss_mean_cls=0.105, proj_loss=-0.0548][2026-03-26 12:31:45] Step: 160, Training Logs: loss_final: 1.030410, loss_mean: 0.984424, proj_loss: -0.058626, loss_mean_cls: 0.104613, grad_norm: 0.594558 +Steps: 0%| | 161/1000000 [00:42<69:06:47, 4.02it/s, grad_norm=0.595, loss_final=1.03, loss_mean=0.984, loss_mean_cls=0.105, proj_loss=-0.0586][2026-03-26 12:31:45] Step: 161, Training Logs: loss_final: 1.049261, loss_mean: 1.003627, proj_loss: -0.058980, loss_mean_cls: 0.104614, grad_norm: 1.096347 +Steps: 0%| | 162/1000000 [00:42<88:49:41, 3.13it/s, grad_norm=1.1, loss_final=1.05, loss_mean=1, loss_mean_cls=0.105, proj_loss=-0.059][2026-03-26 12:31:46] Step: 162, Training Logs: loss_final: 1.051320, loss_mean: 1.005750, proj_loss: -0.058691, loss_mean_cls: 0.104261, grad_norm: 0.491767 +Steps: 0%| | 163/1000000 [00:43<82:54:48, 3.35it/s, grad_norm=0.492, loss_final=1.05, loss_mean=1.01, loss_mean_cls=0.104, proj_loss=-0.0587][2026-03-26 12:31:46] Step: 163, Training Logs: loss_final: 1.035672, loss_mean: 0.988019, proj_loss: -0.057354, loss_mean_cls: 0.105007, grad_norm: 0.738863 +Steps: 0%| | 164/1000000 [00:43<78:38:15, 3.53it/s, grad_norm=0.739, loss_final=1.04, loss_mean=0.988, loss_mean_cls=0.105, proj_loss=-0.0574][2026-03-26 12:31:46] Step: 164, Training Logs: loss_final: 1.042202, loss_mean: 0.998165, proj_loss: -0.059133, loss_mean_cls: 0.103171, grad_norm: 0.787208 +Steps: 0%| | 165/1000000 [00:43<75:37:43, 3.67it/s, grad_norm=0.787, loss_final=1.04, loss_mean=0.998, loss_mean_cls=0.103, proj_loss=-0.0591][2026-03-26 12:31:46] Step: 165, Training Logs: loss_final: 1.013956, loss_mean: 0.965827, proj_loss: -0.057061, loss_mean_cls: 0.105191, grad_norm: 0.800582 +Steps: 0%| | 166/1000000 [00:43<73:29:05, 3.78it/s, grad_norm=0.801, loss_final=1.01, loss_mean=0.966, loss_mean_cls=0.105, proj_loss=-0.0571][2026-03-26 12:31:47] Step: 166, Training Logs: loss_final: 1.026864, loss_mean: 0.978646, proj_loss: -0.056784, loss_mean_cls: 0.105003, grad_norm: 1.069103 +Steps: 0%| | 167/1000000 [00:44<72:02:26, 3.86it/s, grad_norm=1.07, loss_final=1.03, loss_mean=0.979, loss_mean_cls=0.105, proj_loss=-0.0568][2026-03-26 12:31:47] Step: 167, Training Logs: loss_final: 1.031582, loss_mean: 0.985558, proj_loss: -0.058774, loss_mean_cls: 0.104799, grad_norm: 0.572788 +Steps: 0%| | 168/1000000 [00:44<70:59:40, 3.91it/s, grad_norm=0.573, loss_final=1.03, loss_mean=0.986, loss_mean_cls=0.105, proj_loss=-0.0588][2026-03-26 12:31:47] Step: 168, Training Logs: loss_final: 1.038177, loss_mean: 0.989839, proj_loss: -0.055580, loss_mean_cls: 0.103918, grad_norm: 1.078343 +Steps: 0%| | 169/1000000 [00:44<70:18:31, 3.95it/s, grad_norm=1.08, loss_final=1.04, loss_mean=0.99, loss_mean_cls=0.104, proj_loss=-0.0556][2026-03-26 12:31:47] Step: 169, Training Logs: loss_final: 1.024175, loss_mean: 0.976353, proj_loss: -0.057148, loss_mean_cls: 0.104971, grad_norm: 0.644929 +Steps: 0%| | 170/1000000 [00:44<69:47:41, 3.98it/s, grad_norm=0.645, loss_final=1.02, loss_mean=0.976, loss_mean_cls=0.105, proj_loss=-0.0571][2026-03-26 12:31:48] Step: 170, Training Logs: loss_final: 1.059798, loss_mean: 1.012884, proj_loss: -0.056802, loss_mean_cls: 0.103715, grad_norm: 0.961506 +Steps: 0%| | 171/1000000 [00:45<69:26:12, 4.00it/s, grad_norm=0.962, loss_final=1.06, loss_mean=1.01, loss_mean_cls=0.104, proj_loss=-0.0568][2026-03-26 12:31:48] Step: 171, Training Logs: loss_final: 1.029175, loss_mean: 0.981992, proj_loss: -0.057173, loss_mean_cls: 0.104357, grad_norm: 0.645632 +Steps: 0%| | 172/1000000 [00:45<69:10:46, 4.01it/s, grad_norm=0.646, loss_final=1.03, loss_mean=0.982, loss_mean_cls=0.104, proj_loss=-0.0572][2026-03-26 12:31:48] Step: 172, Training Logs: loss_final: 1.017420, loss_mean: 0.970494, proj_loss: -0.057777, loss_mean_cls: 0.104703, grad_norm: 0.823365 +Steps: 0%| | 173/1000000 [00:45<69:10:08, 4.02it/s, grad_norm=0.823, loss_final=1.02, loss_mean=0.97, loss_mean_cls=0.105, proj_loss=-0.0578][2026-03-26 12:31:48] Step: 173, Training Logs: loss_final: 1.016136, loss_mean: 0.968377, proj_loss: -0.056997, loss_mean_cls: 0.104756, grad_norm: 0.868837 +Steps: 0%| | 174/1000000 [00:45<68:58:17, 4.03it/s, grad_norm=0.869, loss_final=1.02, loss_mean=0.968, loss_mean_cls=0.105, proj_loss=-0.057][2026-03-26 12:31:49] Step: 174, Training Logs: loss_final: 1.059633, loss_mean: 1.011309, proj_loss: -0.055249, loss_mean_cls: 0.103573, grad_norm: 0.929197 +Steps: 0%| | 175/1000000 [00:46<68:51:29, 4.03it/s, grad_norm=0.929, loss_final=1.06, loss_mean=1.01, loss_mean_cls=0.104, proj_loss=-0.0552][2026-03-26 12:31:49] Step: 175, Training Logs: loss_final: 1.056512, loss_mean: 1.009780, proj_loss: -0.056741, loss_mean_cls: 0.103473, grad_norm: 0.969713 +Steps: 0%| | 176/1000000 [00:46<68:48:15, 4.04it/s, grad_norm=0.97, loss_final=1.06, loss_mean=1.01, loss_mean_cls=0.103, proj_loss=-0.0567][2026-03-26 12:31:49] Step: 176, Training Logs: loss_final: 1.025866, loss_mean: 0.980405, proj_loss: -0.058846, loss_mean_cls: 0.104308, grad_norm: 0.696566 +Steps: 0%| | 177/1000000 [00:46<68:45:15, 4.04it/s, grad_norm=0.697, loss_final=1.03, loss_mean=0.98, loss_mean_cls=0.104, proj_loss=-0.0588][2026-03-26 12:31:49] Step: 177, Training Logs: loss_final: 1.045044, loss_mean: 0.996350, proj_loss: -0.054167, loss_mean_cls: 0.102861, grad_norm: 0.839030 +Steps: 0%| | 178/1000000 [00:46<68:44:53, 4.04it/s, grad_norm=0.839, loss_final=1.05, loss_mean=0.996, loss_mean_cls=0.103, proj_loss=-0.0542][2026-03-26 12:31:50] Step: 178, Training Logs: loss_final: 1.011556, loss_mean: 0.966363, proj_loss: -0.058602, loss_mean_cls: 0.103796, grad_norm: 0.800694 +Steps: 0%| | 179/1000000 [00:46<68:40:44, 4.04it/s, grad_norm=0.801, loss_final=1.01, loss_mean=0.966, loss_mean_cls=0.104, proj_loss=-0.0586][2026-03-26 12:31:50] Step: 179, Training Logs: loss_final: 1.037298, loss_mean: 0.990838, proj_loss: -0.056409, loss_mean_cls: 0.102869, grad_norm: 0.700006 +Steps: 0%| | 180/1000000 [00:47<68:46:29, 4.04it/s, grad_norm=0.7, loss_final=1.04, loss_mean=0.991, loss_mean_cls=0.103, proj_loss=-0.0564][2026-03-26 12:31:50] Step: 180, Training Logs: loss_final: 1.008286, loss_mean: 0.958345, proj_loss: -0.054923, loss_mean_cls: 0.104864, grad_norm: 0.739139 +Steps: 0%| | 181/1000000 [00:47<68:46:51, 4.04it/s, grad_norm=0.739, loss_final=1.01, loss_mean=0.958, loss_mean_cls=0.105, proj_loss=-0.0549][2026-03-26 12:31:50] Step: 181, Training Logs: loss_final: 1.027792, loss_mean: 0.981764, proj_loss: -0.057155, loss_mean_cls: 0.103182, grad_norm: 0.875367 +Steps: 0%| | 182/1000000 [00:47<68:49:32, 4.04it/s, grad_norm=0.875, loss_final=1.03, loss_mean=0.982, loss_mean_cls=0.103, proj_loss=-0.0572][2026-03-26 12:31:51] Step: 182, Training Logs: loss_final: 1.017327, loss_mean: 0.970175, proj_loss: -0.056668, loss_mean_cls: 0.103820, grad_norm: 0.843896 +Steps: 0%| | 183/1000000 [00:47<68:48:41, 4.04it/s, grad_norm=0.844, loss_final=1.02, loss_mean=0.97, loss_mean_cls=0.104, proj_loss=-0.0567][2026-03-26 12:31:51] Step: 183, Training Logs: loss_final: 1.031032, loss_mean: 0.982272, proj_loss: -0.054766, loss_mean_cls: 0.103526, grad_norm: 0.942891 +Steps: 0%| | 184/1000000 [00:48<68:51:48, 4.03it/s, grad_norm=0.943, loss_final=1.03, loss_mean=0.982, loss_mean_cls=0.104, proj_loss=-0.0548][2026-03-26 12:31:51] Step: 184, Training Logs: loss_final: 1.008994, loss_mean: 0.961337, proj_loss: -0.057157, loss_mean_cls: 0.104813, grad_norm: 0.850634 +Steps: 0%| | 185/1000000 [00:48<68:53:07, 4.03it/s, grad_norm=0.851, loss_final=1.01, loss_mean=0.961, loss_mean_cls=0.105, proj_loss=-0.0572][2026-03-26 12:31:51] Step: 185, Training Logs: loss_final: 1.023440, loss_mean: 0.979218, proj_loss: -0.058590, loss_mean_cls: 0.102812, grad_norm: 0.815850 +Steps: 0%| | 186/1000000 [00:48<68:54:22, 4.03it/s, grad_norm=0.816, loss_final=1.02, loss_mean=0.979, loss_mean_cls=0.103, proj_loss=-0.0586][2026-03-26 12:31:52] Step: 186, Training Logs: loss_final: 1.030619, loss_mean: 0.984677, proj_loss: -0.056762, loss_mean_cls: 0.102704, grad_norm: 0.471126 +Steps: 0%| | 187/1000000 [00:48<68:48:37, 4.04it/s, grad_norm=0.471, loss_final=1.03, loss_mean=0.985, loss_mean_cls=0.103, proj_loss=-0.0568][2026-03-26 12:31:52] Step: 187, Training Logs: loss_final: 1.040574, loss_mean: 0.999010, proj_loss: -0.059644, loss_mean_cls: 0.101208, grad_norm: 0.663741 +Steps: 0%| | 188/1000000 [00:49<68:51:08, 4.03it/s, grad_norm=0.664, loss_final=1.04, loss_mean=0.999, loss_mean_cls=0.101, proj_loss=-0.0596][2026-03-26 12:31:52] Step: 188, Training Logs: loss_final: 1.007815, loss_mean: 0.963298, proj_loss: -0.057983, loss_mean_cls: 0.102501, grad_norm: 0.517400 +Steps: 0%| | 189/1000000 [00:49<68:52:14, 4.03it/s, grad_norm=0.517, loss_final=1.01, loss_mean=0.963, loss_mean_cls=0.103, proj_loss=-0.058][2026-03-26 12:31:52] Step: 189, Training Logs: loss_final: 1.019086, loss_mean: 0.970895, proj_loss: -0.055862, loss_mean_cls: 0.104053, grad_norm: 0.696305 +Steps: 0%| | 190/1000000 [00:49<68:54:02, 4.03it/s, grad_norm=0.696, loss_final=1.02, loss_mean=0.971, loss_mean_cls=0.104, proj_loss=-0.0559][2026-03-26 12:31:53] Step: 190, Training Logs: loss_final: 1.031935, loss_mean: 0.987947, proj_loss: -0.058935, loss_mean_cls: 0.102923, grad_norm: 0.567594 +Steps: 0%| | 191/1000000 [00:49<68:52:33, 4.03it/s, grad_norm=0.568, loss_final=1.03, loss_mean=0.988, loss_mean_cls=0.103, proj_loss=-0.0589][2026-03-26 12:31:53] Step: 191, Training Logs: loss_final: 1.013610, loss_mean: 0.968988, proj_loss: -0.058106, loss_mean_cls: 0.102729, grad_norm: 0.598907 +Steps: 0%| | 192/1000000 [00:50<68:55:55, 4.03it/s, grad_norm=0.599, loss_final=1.01, loss_mean=0.969, loss_mean_cls=0.103, proj_loss=-0.0581][2026-03-26 12:31:53] Step: 192, Training Logs: loss_final: 1.014650, loss_mean: 0.968925, proj_loss: -0.057350, loss_mean_cls: 0.103075, grad_norm: 1.094147 +Steps: 0%| | 193/1000000 [00:50<68:50:00, 4.03it/s, grad_norm=1.09, loss_final=1.01, loss_mean=0.969, loss_mean_cls=0.103, proj_loss=-0.0573][2026-03-26 12:31:53] Step: 193, Training Logs: loss_final: 1.021936, loss_mean: 0.977287, proj_loss: -0.057351, loss_mean_cls: 0.102000, grad_norm: 1.078733 +Steps: 0%| | 194/1000000 [00:50<68:44:53, 4.04it/s, grad_norm=1.08, loss_final=1.02, loss_mean=0.977, loss_mean_cls=0.102, proj_loss=-0.0574][2026-03-26 12:31:54] Step: 194, Training Logs: loss_final: 1.020281, loss_mean: 0.973942, proj_loss: -0.057671, loss_mean_cls: 0.104010, grad_norm: 1.051165 +Steps: 0%| | 195/1000000 [00:50<68:40:54, 4.04it/s, grad_norm=1.05, loss_final=1.02, loss_mean=0.974, loss_mean_cls=0.104, proj_loss=-0.0577][2026-03-26 12:31:54] Step: 195, Training Logs: loss_final: 1.029278, loss_mean: 0.984605, proj_loss: -0.057741, loss_mean_cls: 0.102415, grad_norm: 1.156877 +Steps: 0%| | 196/1000000 [00:51<68:40:14, 4.04it/s, grad_norm=1.16, loss_final=1.03, loss_mean=0.985, loss_mean_cls=0.102, proj_loss=-0.0577][2026-03-26 12:31:54] Step: 196, Training Logs: loss_final: 1.047506, loss_mean: 1.003654, proj_loss: -0.056959, loss_mean_cls: 0.100811, grad_norm: 0.892999 +Steps: 0%| | 197/1000000 [00:51<68:39:36, 4.04it/s, grad_norm=0.893, loss_final=1.05, loss_mean=1, loss_mean_cls=0.101, proj_loss=-0.057][2026-03-26 12:31:54] Step: 197, Training Logs: loss_final: 1.049217, loss_mean: 1.004677, proj_loss: -0.056258, loss_mean_cls: 0.100798, grad_norm: 0.907164 +Steps: 0%| | 198/1000000 [00:51<68:44:26, 4.04it/s, grad_norm=0.907, loss_final=1.05, loss_mean=1, loss_mean_cls=0.101, proj_loss=-0.0563][2026-03-26 12:31:55] Step: 198, Training Logs: loss_final: 1.042606, loss_mean: 1.000546, proj_loss: -0.058271, loss_mean_cls: 0.100332, grad_norm: 1.096825 +Steps: 0%| | 199/1000000 [00:51<68:44:05, 4.04it/s, grad_norm=1.1, loss_final=1.04, loss_mean=1, loss_mean_cls=0.1, proj_loss=-0.0583][2026-03-26 12:31:55] Step: 199, Training Logs: loss_final: 1.025483, loss_mean: 0.979758, proj_loss: -0.056419, loss_mean_cls: 0.102144, grad_norm: 0.641371 +Steps: 0%| | 200/1000000 [00:52<68:43:36, 4.04it/s, grad_norm=0.641, loss_final=1.03, loss_mean=0.98, loss_mean_cls=0.102, proj_loss=-0.0564][2026-03-26 12:31:55] Step: 200, Training Logs: loss_final: 1.023498, loss_mean: 0.979453, proj_loss: -0.058488, loss_mean_cls: 0.102533, grad_norm: 1.036608 +Steps: 0%| | 201/1000000 [00:52<68:42:36, 4.04it/s, grad_norm=1.04, loss_final=1.02, loss_mean=0.979, loss_mean_cls=0.103, proj_loss=-0.0585][2026-03-26 12:31:55] Step: 201, Training Logs: loss_final: 1.017378, loss_mean: 0.972656, proj_loss: -0.057526, loss_mean_cls: 0.102248, grad_norm: 0.730632 +Steps: 0%| | 202/1000000 [00:52<68:42:40, 4.04it/s, grad_norm=0.731, loss_final=1.02, loss_mean=0.973, loss_mean_cls=0.102, proj_loss=-0.0575][2026-03-26 12:31:56] Step: 202, Training Logs: loss_final: 1.026805, loss_mean: 0.981401, proj_loss: -0.056441, loss_mean_cls: 0.101845, grad_norm: 0.955476 +Steps: 0%| | 203/1000000 [00:52<68:41:08, 4.04it/s, grad_norm=0.955, loss_final=1.03, loss_mean=0.981, loss_mean_cls=0.102, proj_loss=-0.0564][2026-03-26 12:31:56] Step: 203, Training Logs: loss_final: 1.020149, loss_mean: 0.976582, proj_loss: -0.059103, loss_mean_cls: 0.102670, grad_norm: 0.887529 +Steps: 0%| | 204/1000000 [00:53<68:39:14, 4.05it/s, grad_norm=0.888, loss_final=1.02, loss_mean=0.977, loss_mean_cls=0.103, proj_loss=-0.0591][2026-03-26 12:31:56] Step: 204, Training Logs: loss_final: 1.011456, loss_mean: 0.965522, proj_loss: -0.056242, loss_mean_cls: 0.102176, grad_norm: 0.922335 +Steps: 0%| | 205/1000000 [00:53<68:38:21, 4.05it/s, grad_norm=0.922, loss_final=1.01, loss_mean=0.966, loss_mean_cls=0.102, proj_loss=-0.0562][2026-03-26 12:31:56] Step: 205, Training Logs: loss_final: 1.013912, loss_mean: 0.967234, proj_loss: -0.056388, loss_mean_cls: 0.103065, grad_norm: 0.873407 +Steps: 0%| | 206/1000000 [00:53<68:39:54, 4.04it/s, grad_norm=0.873, loss_final=1.01, loss_mean=0.967, loss_mean_cls=0.103, proj_loss=-0.0564][2026-03-26 12:31:57] Step: 206, Training Logs: loss_final: 1.011353, loss_mean: 0.969064, proj_loss: -0.059253, loss_mean_cls: 0.101542, grad_norm: 0.686942 +Steps: 0%| | 207/1000000 [00:53<68:40:31, 4.04it/s, grad_norm=0.687, loss_final=1.01, loss_mean=0.969, loss_mean_cls=0.102, proj_loss=-0.0593][2026-03-26 12:31:57] Step: 207, Training Logs: loss_final: 0.997543, loss_mean: 0.954508, proj_loss: -0.059389, loss_mean_cls: 0.102424, grad_norm: 0.896265 +Steps: 0%| | 208/1000000 [00:54<68:39:12, 4.05it/s, grad_norm=0.896, loss_final=0.998, loss_mean=0.955, loss_mean_cls=0.102, proj_loss=-0.0594][2026-03-26 12:31:57] Step: 208, Training Logs: loss_final: 0.991150, loss_mean: 0.944803, proj_loss: -0.055989, loss_mean_cls: 0.102336, grad_norm: 0.571996 +Steps: 0%| | 209/1000000 [00:54<68:39:20, 4.05it/s, grad_norm=0.572, loss_final=0.991, loss_mean=0.945, loss_mean_cls=0.102, proj_loss=-0.056][2026-03-26 12:31:57] Step: 209, Training Logs: loss_final: 1.028124, loss_mean: 0.982140, proj_loss: -0.055675, loss_mean_cls: 0.101660, grad_norm: 0.745863 +Steps: 0%| | 210/1000000 [00:54<68:41:02, 4.04it/s, grad_norm=0.746, loss_final=1.03, loss_mean=0.982, loss_mean_cls=0.102, proj_loss=-0.0557][2026-03-26 12:31:58] Step: 210, Training Logs: loss_final: 0.972709, loss_mean: 0.925228, proj_loss: -0.057377, loss_mean_cls: 0.104857, grad_norm: 0.548025 +Steps: 0%| | 211/1000000 [00:54<68:43:34, 4.04it/s, grad_norm=0.548, loss_final=0.973, loss_mean=0.925, loss_mean_cls=0.105, proj_loss=-0.0574][2026-03-26 12:31:58] Step: 211, Training Logs: loss_final: 0.998090, loss_mean: 0.955177, proj_loss: -0.059020, loss_mean_cls: 0.101933, grad_norm: 1.036514 +Steps: 0%| | 212/1000000 [00:55<68:45:48, 4.04it/s, grad_norm=1.04, loss_final=0.998, loss_mean=0.955, loss_mean_cls=0.102, proj_loss=-0.059][2026-03-26 12:31:58] Step: 212, Training Logs: loss_final: 1.002998, loss_mean: 0.959711, proj_loss: -0.058263, loss_mean_cls: 0.101549, grad_norm: 0.668954 +Steps: 0%| | 213/1000000 [00:55<68:46:56, 4.04it/s, grad_norm=0.669, loss_final=1, loss_mean=0.96, loss_mean_cls=0.102, proj_loss=-0.0583][2026-03-26 12:31:58] Step: 213, Training Logs: loss_final: 1.007599, loss_mean: 0.964525, proj_loss: -0.059374, loss_mean_cls: 0.102448, grad_norm: 0.673296 +Steps: 0%| | 214/1000000 [00:55<68:46:30, 4.04it/s, grad_norm=0.673, loss_final=1.01, loss_mean=0.965, loss_mean_cls=0.102, proj_loss=-0.0594][2026-03-26 12:31:59] Step: 214, Training Logs: loss_final: 0.994557, loss_mean: 0.947571, proj_loss: -0.056062, loss_mean_cls: 0.103048, grad_norm: 1.097052 +Steps: 0%| | 215/1000000 [00:55<68:43:48, 4.04it/s, grad_norm=1.1, loss_final=0.995, loss_mean=0.948, loss_mean_cls=0.103, proj_loss=-0.0561][2026-03-26 12:31:59] Step: 215, Training Logs: loss_final: 1.013132, loss_mean: 0.967865, proj_loss: -0.056826, loss_mean_cls: 0.102093, grad_norm: 0.984973 +Steps: 0%| | 216/1000000 [00:56<68:41:35, 4.04it/s, grad_norm=0.985, loss_final=1.01, loss_mean=0.968, loss_mean_cls=0.102, proj_loss=-0.0568][2026-03-26 12:31:59] Step: 216, Training Logs: loss_final: 1.011439, loss_mean: 0.967272, proj_loss: -0.057349, loss_mean_cls: 0.101516, grad_norm: 0.800923 +Steps: 0%| | 217/1000000 [00:56<68:39:48, 4.04it/s, grad_norm=0.801, loss_final=1.01, loss_mean=0.967, loss_mean_cls=0.102, proj_loss=-0.0573][2026-03-26 12:31:59] Step: 217, Training Logs: loss_final: 0.997501, loss_mean: 0.957141, proj_loss: -0.060478, loss_mean_cls: 0.100837, grad_norm: 0.839226 +Steps: 0%| | 218/1000000 [00:56<68:39:11, 4.05it/s, grad_norm=0.839, loss_final=0.998, loss_mean=0.957, loss_mean_cls=0.101, proj_loss=-0.0605][2026-03-26 12:32:00] Step: 218, Training Logs: loss_final: 0.997823, loss_mean: 0.955163, proj_loss: -0.058832, loss_mean_cls: 0.101492, grad_norm: 0.686176 +Steps: 0%| | 219/1000000 [00:56<68:36:41, 4.05it/s, grad_norm=0.686, loss_final=0.998, loss_mean=0.955, loss_mean_cls=0.101, proj_loss=-0.0588][2026-03-26 12:32:00] Step: 219, Training Logs: loss_final: 1.021472, loss_mean: 0.975971, proj_loss: -0.056807, loss_mean_cls: 0.102308, grad_norm: 0.945811 +Steps: 0%| | 220/1000000 [00:57<68:39:20, 4.05it/s, grad_norm=0.946, loss_final=1.02, loss_mean=0.976, loss_mean_cls=0.102, proj_loss=-0.0568][2026-03-26 12:32:00] Step: 220, Training Logs: loss_final: 1.010488, loss_mean: 0.966560, proj_loss: -0.057525, loss_mean_cls: 0.101454, grad_norm: 0.610917 +Steps: 0%| | 221/1000000 [00:57<68:38:57, 4.05it/s, grad_norm=0.611, loss_final=1.01, loss_mean=0.967, loss_mean_cls=0.101, proj_loss=-0.0575][2026-03-26 12:32:00] Step: 221, Training Logs: loss_final: 0.976439, loss_mean: 0.931482, proj_loss: -0.057786, loss_mean_cls: 0.102743, grad_norm: 0.800497 +Steps: 0%| | 222/1000000 [00:57<68:43:27, 4.04it/s, grad_norm=0.8, loss_final=0.976, loss_mean=0.931, loss_mean_cls=0.103, proj_loss=-0.0578][2026-03-26 12:32:01] Step: 222, Training Logs: loss_final: 1.013234, loss_mean: 0.970239, proj_loss: -0.057575, loss_mean_cls: 0.100570, grad_norm: 0.444801 +Steps: 0%| | 223/1000000 [00:57<68:41:04, 4.04it/s, grad_norm=0.445, loss_final=1.01, loss_mean=0.97, loss_mean_cls=0.101, proj_loss=-0.0576][2026-03-26 12:32:01] Step: 223, Training Logs: loss_final: 0.981616, loss_mean: 0.935141, proj_loss: -0.056820, loss_mean_cls: 0.103296, grad_norm: 0.850086 +Steps: 0%| | 224/1000000 [00:58<68:43:03, 4.04it/s, grad_norm=0.85, loss_final=0.982, loss_mean=0.935, loss_mean_cls=0.103, proj_loss=-0.0568][2026-03-26 12:32:01] Step: 224, Training Logs: loss_final: 1.006476, loss_mean: 0.961221, proj_loss: -0.055930, loss_mean_cls: 0.101186, grad_norm: 0.728671 +Steps: 0%| | 225/1000000 [00:58<68:40:01, 4.04it/s, grad_norm=0.729, loss_final=1.01, loss_mean=0.961, loss_mean_cls=0.101, proj_loss=-0.0559][2026-03-26 12:32:01] Step: 225, Training Logs: loss_final: 0.993288, loss_mean: 0.947160, proj_loss: -0.056042, loss_mean_cls: 0.102170, grad_norm: 0.774063 +Steps: 0%| | 226/1000000 [00:58<68:52:19, 4.03it/s, grad_norm=0.774, loss_final=0.993, loss_mean=0.947, loss_mean_cls=0.102, proj_loss=-0.056][2026-03-26 12:32:02] Step: 226, Training Logs: loss_final: 0.998891, loss_mean: 0.953966, proj_loss: -0.057051, loss_mean_cls: 0.101977, grad_norm: 0.658739 +Steps: 0%| | 227/1000000 [00:58<68:36:52, 4.05it/s, grad_norm=0.659, loss_final=0.999, loss_mean=0.954, loss_mean_cls=0.102, proj_loss=-0.0571][2026-03-26 12:32:02] Step: 227, Training Logs: loss_final: 1.026876, loss_mean: 0.983965, proj_loss: -0.056732, loss_mean_cls: 0.099644, grad_norm: 0.774402 +Steps: 0%| | 228/1000000 [00:59<68:38:54, 4.05it/s, grad_norm=0.774, loss_final=1.03, loss_mean=0.984, loss_mean_cls=0.0996, proj_loss=-0.0567][2026-03-26 12:32:02] Step: 228, Training Logs: loss_final: 1.009487, loss_mean: 0.964037, proj_loss: -0.056870, loss_mean_cls: 0.102319, grad_norm: 0.657773 +Steps: 0%| | 229/1000000 [00:59<68:41:03, 4.04it/s, grad_norm=0.658, loss_final=1.01, loss_mean=0.964, loss_mean_cls=0.102, proj_loss=-0.0569][2026-03-26 12:32:02] Step: 229, Training Logs: loss_final: 0.984555, loss_mean: 0.941725, proj_loss: -0.058287, loss_mean_cls: 0.101117, grad_norm: 0.673736 +Steps: 0%| | 230/1000000 [00:59<68:43:27, 4.04it/s, grad_norm=0.674, loss_final=0.985, loss_mean=0.942, loss_mean_cls=0.101, proj_loss=-0.0583][2026-03-26 12:32:03] Step: 230, Training Logs: loss_final: 1.006507, loss_mean: 0.964571, proj_loss: -0.057090, loss_mean_cls: 0.099026, grad_norm: 0.823767 +Steps: 0%| | 231/1000000 [00:59<68:43:17, 4.04it/s, grad_norm=0.824, loss_final=1.01, loss_mean=0.965, loss_mean_cls=0.099, proj_loss=-0.0571][2026-03-26 12:32:03] Step: 231, Training Logs: loss_final: 1.009156, loss_mean: 0.967872, proj_loss: -0.059701, loss_mean_cls: 0.100985, grad_norm: 0.733705 +Steps: 0%| | 232/1000000 [01:00<68:42:42, 4.04it/s, grad_norm=0.734, loss_final=1.01, loss_mean=0.968, loss_mean_cls=0.101, proj_loss=-0.0597][2026-03-26 12:32:03] Step: 232, Training Logs: loss_final: 1.005023, loss_mean: 0.962199, proj_loss: -0.057101, loss_mean_cls: 0.099926, grad_norm: 1.203185 +Steps: 0%| | 233/1000000 [01:00<68:41:22, 4.04it/s, grad_norm=1.2, loss_final=1.01, loss_mean=0.962, loss_mean_cls=0.0999, proj_loss=-0.0571][2026-03-26 12:32:03] Step: 233, Training Logs: loss_final: 1.003059, loss_mean: 0.961346, proj_loss: -0.059722, loss_mean_cls: 0.101435, grad_norm: 0.941105 +Steps: 0%| | 234/1000000 [01:00<68:44:04, 4.04it/s, grad_norm=0.941, loss_final=1, loss_mean=0.961, loss_mean_cls=0.101, proj_loss=-0.0597][2026-03-26 12:32:04] Step: 234, Training Logs: loss_final: 1.002296, loss_mean: 0.959454, proj_loss: -0.057353, loss_mean_cls: 0.100196, grad_norm: 0.923537 +Steps: 0%| | 235/1000000 [01:00<68:43:21, 4.04it/s, grad_norm=0.924, loss_final=1, loss_mean=0.959, loss_mean_cls=0.1, proj_loss=-0.0574][2026-03-26 12:32:04] Step: 235, Training Logs: loss_final: 1.016214, loss_mean: 0.977302, proj_loss: -0.060105, loss_mean_cls: 0.099016, grad_norm: 1.090529 +Steps: 0%| | 236/1000000 [01:01<68:43:07, 4.04it/s, grad_norm=1.09, loss_final=1.02, loss_mean=0.977, loss_mean_cls=0.099, proj_loss=-0.0601][2026-03-26 12:32:04] Step: 236, Training Logs: loss_final: 0.978583, loss_mean: 0.937199, proj_loss: -0.060666, loss_mean_cls: 0.102050, grad_norm: 0.825940 +Steps: 0%| | 237/1000000 [01:01<68:43:29, 4.04it/s, grad_norm=0.826, loss_final=0.979, loss_mean=0.937, loss_mean_cls=0.102, proj_loss=-0.0607][2026-03-26 12:32:04] Step: 237, Training Logs: loss_final: 0.988920, loss_mean: 0.945597, proj_loss: -0.059319, loss_mean_cls: 0.102643, grad_norm: 1.195363 +Steps: 0%| | 238/1000000 [01:01<68:42:32, 4.04it/s, grad_norm=1.2, loss_final=0.989, loss_mean=0.946, loss_mean_cls=0.103, proj_loss=-0.0593][2026-03-26 12:32:04] Step: 238, Training Logs: loss_final: 0.988141, loss_mean: 0.946697, proj_loss: -0.058785, loss_mean_cls: 0.100229, grad_norm: 0.805799 +Steps: 0%| | 239/1000000 [01:01<68:44:58, 4.04it/s, grad_norm=0.806, loss_final=0.988, loss_mean=0.947, loss_mean_cls=0.1, proj_loss=-0.0588][2026-03-26 12:32:05] Step: 239, Training Logs: loss_final: 1.005688, loss_mean: 0.964243, proj_loss: -0.059284, loss_mean_cls: 0.100730, grad_norm: 0.783813 +Steps: 0%| | 240/1000000 [01:02<68:41:32, 4.04it/s, grad_norm=0.784, loss_final=1.01, loss_mean=0.964, loss_mean_cls=0.101, proj_loss=-0.0593][2026-03-26 12:32:05] Step: 240, Training Logs: loss_final: 0.985381, loss_mean: 0.941380, proj_loss: -0.058072, loss_mean_cls: 0.102074, grad_norm: 1.071268 +Steps: 0%| | 241/1000000 [01:02<68:40:53, 4.04it/s, grad_norm=1.07, loss_final=0.985, loss_mean=0.941, loss_mean_cls=0.102, proj_loss=-0.0581][2026-03-26 12:32:05] Step: 241, Training Logs: loss_final: 1.011900, loss_mean: 0.968049, proj_loss: -0.056812, loss_mean_cls: 0.100663, grad_norm: 0.989999 +Steps: 0%| | 242/1000000 [01:02<68:39:01, 4.05it/s, grad_norm=0.99, loss_final=1.01, loss_mean=0.968, loss_mean_cls=0.101, proj_loss=-0.0568][2026-03-26 12:32:05] Step: 242, Training Logs: loss_final: 1.003254, loss_mean: 0.959794, proj_loss: -0.057534, loss_mean_cls: 0.100994, grad_norm: 1.084171 +Steps: 0%| | 243/1000000 [01:02<68:38:35, 4.05it/s, grad_norm=1.08, loss_final=1, loss_mean=0.96, loss_mean_cls=0.101, proj_loss=-0.0575][2026-03-26 12:32:06] Step: 243, Training Logs: loss_final: 0.993067, loss_mean: 0.951656, proj_loss: -0.060007, loss_mean_cls: 0.101417, grad_norm: 0.847015 +Steps: 0%| | 244/1000000 [01:03<68:37:23, 4.05it/s, grad_norm=0.847, loss_final=0.993, loss_mean=0.952, loss_mean_cls=0.101, proj_loss=-0.06][2026-03-26 12:32:06] Step: 244, Training Logs: loss_final: 0.995453, loss_mean: 0.956487, proj_loss: -0.061031, loss_mean_cls: 0.099997, grad_norm: 1.048733 +Steps: 0%| | 245/1000000 [01:03<68:40:09, 4.04it/s, grad_norm=1.05, loss_final=0.995, loss_mean=0.956, loss_mean_cls=0.1, proj_loss=-0.061][2026-03-26 12:32:06] Step: 245, Training Logs: loss_final: 0.990811, loss_mean: 0.947793, proj_loss: -0.056902, loss_mean_cls: 0.099920, grad_norm: 0.709384 +Steps: 0%| | 246/1000000 [01:03<68:39:12, 4.05it/s, grad_norm=0.709, loss_final=0.991, loss_mean=0.948, loss_mean_cls=0.0999, proj_loss=-0.0569][2026-03-26 12:32:06] Step: 246, Training Logs: loss_final: 0.986912, loss_mean: 0.944664, proj_loss: -0.058584, loss_mean_cls: 0.100832, grad_norm: 1.026528 +Steps: 0%| | 247/1000000 [01:03<68:37:55, 4.05it/s, grad_norm=1.03, loss_final=0.987, loss_mean=0.945, loss_mean_cls=0.101, proj_loss=-0.0586][2026-03-26 12:32:07] Step: 247, Training Logs: loss_final: 0.991710, loss_mean: 0.948777, proj_loss: -0.057333, loss_mean_cls: 0.100266, grad_norm: 0.744696 +Steps: 0%| | 248/1000000 [01:04<68:38:14, 4.05it/s, grad_norm=0.745, loss_final=0.992, loss_mean=0.949, loss_mean_cls=0.1, proj_loss=-0.0573][2026-03-26 12:32:07] Step: 248, Training Logs: loss_final: 1.006303, loss_mean: 0.963129, proj_loss: -0.056158, loss_mean_cls: 0.099332, grad_norm: 0.731050 +Steps: 0%| | 249/1000000 [01:04<68:39:31, 4.04it/s, grad_norm=0.731, loss_final=1.01, loss_mean=0.963, loss_mean_cls=0.0993, proj_loss=-0.0562][2026-03-26 12:32:07] Step: 249, Training Logs: loss_final: 0.994813, loss_mean: 0.952721, proj_loss: -0.058365, loss_mean_cls: 0.100457, grad_norm: 1.007301 +Steps: 0%| | 250/1000000 [01:04<68:40:54, 4.04it/s, grad_norm=1.01, loss_final=0.995, loss_mean=0.953, loss_mean_cls=0.1, proj_loss=-0.0584][2026-03-26 12:32:07] Step: 250, Training Logs: loss_final: 1.005918, loss_mean: 0.964527, proj_loss: -0.057385, loss_mean_cls: 0.098776, grad_norm: 0.972881 +Steps: 0%| | 251/1000000 [01:04<68:39:59, 4.04it/s, grad_norm=0.973, loss_final=1.01, loss_mean=0.965, loss_mean_cls=0.0988, proj_loss=-0.0574][2026-03-26 12:32:08] Step: 251, Training Logs: loss_final: 1.000923, loss_mean: 0.961606, proj_loss: -0.059849, loss_mean_cls: 0.099166, grad_norm: 0.841588 +Steps: 0%| | 252/1000000 [01:05<68:40:28, 4.04it/s, grad_norm=0.842, loss_final=1, loss_mean=0.962, loss_mean_cls=0.0992, proj_loss=-0.0598][2026-03-26 12:32:08] Step: 252, Training Logs: loss_final: 1.018871, loss_mean: 0.981339, proj_loss: -0.060997, loss_mean_cls: 0.098528, grad_norm: 1.308803 +Steps: 0%| | 253/1000000 [01:05<68:39:09, 4.05it/s, grad_norm=1.31, loss_final=1.02, loss_mean=0.981, loss_mean_cls=0.0985, proj_loss=-0.061][2026-03-26 12:32:08] Step: 253, Training Logs: loss_final: 1.005546, loss_mean: 0.962419, proj_loss: -0.056838, loss_mean_cls: 0.099965, grad_norm: 1.077965 +Steps: 0%| | 254/1000000 [01:05<68:44:26, 4.04it/s, grad_norm=1.08, loss_final=1.01, loss_mean=0.962, loss_mean_cls=0.1, proj_loss=-0.0568][2026-03-26 12:32:08] Step: 254, Training Logs: loss_final: 1.007622, loss_mean: 0.967179, proj_loss: -0.059525, loss_mean_cls: 0.099968, grad_norm: 1.262148 +Steps: 0%| | 255/1000000 [01:05<73:41:07, 3.77it/s, grad_norm=1.26, loss_final=1.01, loss_mean=0.967, loss_mean_cls=0.1, proj_loss=-0.0595][2026-03-26 12:32:09] Step: 255, Training Logs: loss_final: 0.992815, loss_mean: 0.948507, proj_loss: -0.055231, loss_mean_cls: 0.099538, grad_norm: 0.951955 +Steps: 0%| | 256/1000000 [01:06<84:34:11, 3.28it/s, grad_norm=0.952, loss_final=0.993, loss_mean=0.949, loss_mean_cls=0.0995, proj_loss=-0.0552][2026-03-26 12:32:09] Step: 256, Training Logs: loss_final: 1.024616, loss_mean: 0.983502, proj_loss: -0.056958, loss_mean_cls: 0.098071, grad_norm: 1.185046 +Steps: 0%| | 257/1000000 [01:06<79:43:27, 3.48it/s, grad_norm=1.19, loss_final=1.02, loss_mean=0.984, loss_mean_cls=0.0981, proj_loss=-0.057][2026-03-26 12:32:09] Step: 257, Training Logs: loss_final: 0.989977, loss_mean: 0.949178, proj_loss: -0.058857, loss_mean_cls: 0.099656, grad_norm: 1.045894 +Steps: 0%| | 258/1000000 [01:06<76:27:30, 3.63it/s, grad_norm=1.05, loss_final=0.99, loss_mean=0.949, loss_mean_cls=0.0997, proj_loss=-0.0589][2026-03-26 12:32:10] Step: 258, Training Logs: loss_final: 1.019442, loss_mean: 0.977184, proj_loss: -0.056756, loss_mean_cls: 0.099014, grad_norm: 1.038551 +Steps: 0%| | 259/1000000 [01:07<74:44:27, 3.72it/s, grad_norm=1.04, loss_final=1.02, loss_mean=0.977, loss_mean_cls=0.099, proj_loss=-0.0568][2026-03-26 12:32:10] Step: 259, Training Logs: loss_final: 0.972934, loss_mean: 0.930265, proj_loss: -0.057984, loss_mean_cls: 0.100653, grad_norm: 1.327625 +Steps: 0%| | 260/1000000 [01:07<72:56:14, 3.81it/s, grad_norm=1.33, loss_final=0.973, loss_mean=0.93, loss_mean_cls=0.101, proj_loss=-0.058][2026-03-26 12:32:10] Step: 260, Training Logs: loss_final: 1.011267, loss_mean: 0.965571, proj_loss: -0.054975, loss_mean_cls: 0.100671, grad_norm: 1.325637 +Steps: 0%| | 261/1000000 [01:07<71:37:13, 3.88it/s, grad_norm=1.33, loss_final=1.01, loss_mean=0.966, loss_mean_cls=0.101, proj_loss=-0.055][2026-03-26 12:32:10] Step: 261, Training Logs: loss_final: 0.983396, loss_mean: 0.944338, proj_loss: -0.059574, loss_mean_cls: 0.098633, grad_norm: 1.042559 +Steps: 0%| | 262/1000000 [01:07<70:45:10, 3.92it/s, grad_norm=1.04, loss_final=0.983, loss_mean=0.944, loss_mean_cls=0.0986, proj_loss=-0.0596][2026-03-26 12:32:11] Step: 262, Training Logs: loss_final: 1.021115, loss_mean: 0.982293, proj_loss: -0.059727, loss_mean_cls: 0.098549, grad_norm: 1.145604 +Steps: 0%| | 263/1000000 [01:07<70:07:36, 3.96it/s, grad_norm=1.15, loss_final=1.02, loss_mean=0.982, loss_mean_cls=0.0985, proj_loss=-0.0597][2026-03-26 12:32:11] Step: 263, Training Logs: loss_final: 0.993264, loss_mean: 0.949015, proj_loss: -0.056583, loss_mean_cls: 0.100831, grad_norm: 0.767269 +Steps: 0%| | 264/1000000 [01:08<69:42:52, 3.98it/s, grad_norm=0.767, loss_final=0.993, loss_mean=0.949, loss_mean_cls=0.101, proj_loss=-0.0566][2026-03-26 12:32:11] Step: 264, Training Logs: loss_final: 0.980340, loss_mean: 0.937571, proj_loss: -0.057913, loss_mean_cls: 0.100682, grad_norm: 0.803277 +Steps: 0%| | 265/1000000 [01:08<69:21:36, 4.00it/s, grad_norm=0.803, loss_final=0.98, loss_mean=0.938, loss_mean_cls=0.101, proj_loss=-0.0579][2026-03-26 12:32:11] Step: 265, Training Logs: loss_final: 0.978879, loss_mean: 0.938932, proj_loss: -0.060792, loss_mean_cls: 0.100738, grad_norm: 1.006107 +Steps: 0%| | 266/1000000 [01:08<69:07:49, 4.02it/s, grad_norm=1.01, loss_final=0.979, loss_mean=0.939, loss_mean_cls=0.101, proj_loss=-0.0608][2026-03-26 12:32:12] Step: 266, Training Logs: loss_final: 1.009625, loss_mean: 0.967958, proj_loss: -0.057142, loss_mean_cls: 0.098810, grad_norm: 1.102213 +Steps: 0%| | 267/1000000 [01:08<68:58:05, 4.03it/s, grad_norm=1.1, loss_final=1.01, loss_mean=0.968, loss_mean_cls=0.0988, proj_loss=-0.0571][2026-03-26 12:32:12] Step: 267, Training Logs: loss_final: 0.994545, loss_mean: 0.952768, proj_loss: -0.057972, loss_mean_cls: 0.099749, grad_norm: 0.827322 +Steps: 0%| | 268/1000000 [01:09<69:08:31, 4.02it/s, grad_norm=0.827, loss_final=0.995, loss_mean=0.953, loss_mean_cls=0.0997, proj_loss=-0.058][2026-03-26 12:32:12] Step: 268, Training Logs: loss_final: 0.994354, loss_mean: 0.951172, proj_loss: -0.057492, loss_mean_cls: 0.100675, grad_norm: 0.912814 +Steps: 0%| | 269/1000000 [01:09<68:58:03, 4.03it/s, grad_norm=0.913, loss_final=0.994, loss_mean=0.951, loss_mean_cls=0.101, proj_loss=-0.0575][2026-03-26 12:32:12] Step: 269, Training Logs: loss_final: 0.986864, loss_mean: 0.947194, proj_loss: -0.059606, loss_mean_cls: 0.099276, grad_norm: 0.740584 +Steps: 0%| | 270/1000000 [01:09<68:53:24, 4.03it/s, grad_norm=0.741, loss_final=0.987, loss_mean=0.947, loss_mean_cls=0.0993, proj_loss=-0.0596][2026-03-26 12:32:13] Step: 270, Training Logs: loss_final: 0.986233, loss_mean: 0.948508, proj_loss: -0.061212, loss_mean_cls: 0.098936, grad_norm: 0.909525 +Steps: 0%| | 271/1000000 [01:09<68:48:31, 4.04it/s, grad_norm=0.91, loss_final=0.986, loss_mean=0.949, loss_mean_cls=0.0989, proj_loss=-0.0612][2026-03-26 12:32:13] Step: 271, Training Logs: loss_final: 0.991420, loss_mean: 0.952304, proj_loss: -0.059785, loss_mean_cls: 0.098901, grad_norm: 1.075006 +Steps: 0%| | 272/1000000 [01:10<68:47:32, 4.04it/s, grad_norm=1.08, loss_final=0.991, loss_mean=0.952, loss_mean_cls=0.0989, proj_loss=-0.0598][2026-03-26 12:32:13] Step: 272, Training Logs: loss_final: 0.972129, loss_mean: 0.929190, proj_loss: -0.057623, loss_mean_cls: 0.100562, grad_norm: 0.793513 +Steps: 0%| | 273/1000000 [01:10<68:46:38, 4.04it/s, grad_norm=0.794, loss_final=0.972, loss_mean=0.929, loss_mean_cls=0.101, proj_loss=-0.0576][2026-03-26 12:32:13] Step: 273, Training Logs: loss_final: 0.995215, loss_mean: 0.955730, proj_loss: -0.058800, loss_mean_cls: 0.098285, grad_norm: 1.009724 +Steps: 0%| | 274/1000000 [01:10<68:45:51, 4.04it/s, grad_norm=1.01, loss_final=0.995, loss_mean=0.956, loss_mean_cls=0.0983, proj_loss=-0.0588][2026-03-26 12:32:14] Step: 274, Training Logs: loss_final: 0.986987, loss_mean: 0.942885, proj_loss: -0.055892, loss_mean_cls: 0.099993, grad_norm: 0.698293 +Steps: 0%| | 275/1000000 [01:10<69:14:22, 4.01it/s, grad_norm=0.698, loss_final=0.987, loss_mean=0.943, loss_mean_cls=0.1, proj_loss=-0.0559][2026-03-26 12:32:14] Step: 275, Training Logs: loss_final: 1.001350, loss_mean: 0.958546, proj_loss: -0.056966, loss_mean_cls: 0.099770, grad_norm: 1.111159 +Steps: 0%| | 276/1000000 [01:11<69:05:31, 4.02it/s, grad_norm=1.11, loss_final=1, loss_mean=0.959, loss_mean_cls=0.0998, proj_loss=-0.057][2026-03-26 12:32:14] Step: 276, Training Logs: loss_final: 0.981626, loss_mean: 0.939604, proj_loss: -0.058006, loss_mean_cls: 0.100028, grad_norm: 0.901767 +Steps: 0%| | 277/1000000 [01:11<68:59:15, 4.03it/s, grad_norm=0.902, loss_final=0.982, loss_mean=0.94, loss_mean_cls=0.1, proj_loss=-0.058][2026-03-26 12:32:14] Step: 277, Training Logs: loss_final: 0.978743, loss_mean: 0.938305, proj_loss: -0.059293, loss_mean_cls: 0.099731, grad_norm: 0.658229 +Steps: 0%| | 278/1000000 [01:11<68:53:18, 4.03it/s, grad_norm=0.658, loss_final=0.979, loss_mean=0.938, loss_mean_cls=0.0997, proj_loss=-0.0593][2026-03-26 12:32:15] Step: 278, Training Logs: loss_final: 0.991895, loss_mean: 0.948205, proj_loss: -0.056716, loss_mean_cls: 0.100406, grad_norm: 0.895596 +Steps: 0%| | 279/1000000 [01:11<68:51:12, 4.03it/s, grad_norm=0.896, loss_final=0.992, loss_mean=0.948, loss_mean_cls=0.1, proj_loss=-0.0567][2026-03-26 12:32:15] Step: 279, Training Logs: loss_final: 0.996108, loss_mean: 0.956508, proj_loss: -0.058449, loss_mean_cls: 0.098049, grad_norm: 1.040697 +Steps: 0%| | 280/1000000 [01:12<68:47:01, 4.04it/s, grad_norm=1.04, loss_final=0.996, loss_mean=0.957, loss_mean_cls=0.098, proj_loss=-0.0584][2026-03-26 12:32:15] Step: 280, Training Logs: loss_final: 0.975133, loss_mean: 0.933949, proj_loss: -0.057634, loss_mean_cls: 0.098817, grad_norm: 0.550882 +Steps: 0%| | 281/1000000 [01:12<68:46:12, 4.04it/s, grad_norm=0.551, loss_final=0.975, loss_mean=0.934, loss_mean_cls=0.0988, proj_loss=-0.0576][2026-03-26 12:32:15] Step: 281, Training Logs: loss_final: 0.993694, loss_mean: 0.952943, proj_loss: -0.058422, loss_mean_cls: 0.099173, grad_norm: 1.014107 +Steps: 0%| | 282/1000000 [01:12<68:44:34, 4.04it/s, grad_norm=1.01, loss_final=0.994, loss_mean=0.953, loss_mean_cls=0.0992, proj_loss=-0.0584][2026-03-26 12:32:16] Step: 282, Training Logs: loss_final: 0.990724, loss_mean: 0.948912, proj_loss: -0.057612, loss_mean_cls: 0.099424, grad_norm: 1.019749 +Steps: 0%| | 283/1000000 [01:12<68:48:31, 4.04it/s, grad_norm=1.02, loss_final=0.991, loss_mean=0.949, loss_mean_cls=0.0994, proj_loss=-0.0576][2026-03-26 12:32:16] Step: 283, Training Logs: loss_final: 0.963277, loss_mean: 0.922188, proj_loss: -0.058724, loss_mean_cls: 0.099814, grad_norm: 0.783848 +Steps: 0%| | 284/1000000 [01:13<68:46:31, 4.04it/s, grad_norm=0.784, loss_final=0.963, loss_mean=0.922, loss_mean_cls=0.0998, proj_loss=-0.0587][2026-03-26 12:32:16] Step: 284, Training Logs: loss_final: 0.986826, loss_mean: 0.946674, proj_loss: -0.057905, loss_mean_cls: 0.098057, grad_norm: 1.102141 +Steps: 0%| | 285/1000000 [01:13<68:44:58, 4.04it/s, grad_norm=1.1, loss_final=0.987, loss_mean=0.947, loss_mean_cls=0.0981, proj_loss=-0.0579][2026-03-26 12:32:16] Step: 285, Training Logs: loss_final: 0.988226, loss_mean: 0.945465, proj_loss: -0.055758, loss_mean_cls: 0.098519, grad_norm: 1.006833 +Steps: 0%| | 286/1000000 [01:13<69:00:55, 4.02it/s, grad_norm=1.01, loss_final=0.988, loss_mean=0.945, loss_mean_cls=0.0985, proj_loss=-0.0558][2026-03-26 12:32:17] Step: 286, Training Logs: loss_final: 0.988611, loss_mean: 0.948256, proj_loss: -0.058156, loss_mean_cls: 0.098511, grad_norm: 0.632333 +Steps: 0%| | 287/1000000 [01:13<69:13:57, 4.01it/s, grad_norm=0.632, loss_final=0.989, loss_mean=0.948, loss_mean_cls=0.0985, proj_loss=-0.0582][2026-03-26 12:32:17] Step: 287, Training Logs: loss_final: 0.970673, loss_mean: 0.932679, proj_loss: -0.060345, loss_mean_cls: 0.098339, grad_norm: 0.765817 +Steps: 0%| | 288/1000000 [01:14<69:06:22, 4.02it/s, grad_norm=0.766, loss_final=0.971, loss_mean=0.933, loss_mean_cls=0.0983, proj_loss=-0.0603][2026-03-26 12:32:17] Step: 288, Training Logs: loss_final: 1.000619, loss_mean: 0.961202, proj_loss: -0.058122, loss_mean_cls: 0.097539, grad_norm: 1.028930 +Steps: 0%| | 289/1000000 [01:14<68:58:36, 4.03it/s, grad_norm=1.03, loss_final=1, loss_mean=0.961, loss_mean_cls=0.0975, proj_loss=-0.0581][2026-03-26 12:32:17] Step: 289, Training Logs: loss_final: 0.990765, loss_mean: 0.949296, proj_loss: -0.057860, loss_mean_cls: 0.099330, grad_norm: 0.564966 +Steps: 0%| | 290/1000000 [01:14<68:54:24, 4.03it/s, grad_norm=0.565, loss_final=0.991, loss_mean=0.949, loss_mean_cls=0.0993, proj_loss=-0.0579][2026-03-26 12:32:18] Step: 290, Training Logs: loss_final: 0.979909, loss_mean: 0.943508, proj_loss: -0.060834, loss_mean_cls: 0.097235, grad_norm: 0.673326 +Steps: 0%| | 291/1000000 [01:14<68:52:54, 4.03it/s, grad_norm=0.673, loss_final=0.98, loss_mean=0.944, loss_mean_cls=0.0972, proj_loss=-0.0608][2026-03-26 12:32:18] Step: 291, Training Logs: loss_final: 0.973956, loss_mean: 0.935428, proj_loss: -0.059296, loss_mean_cls: 0.097824, grad_norm: 0.750775 +Steps: 0%| | 292/1000000 [01:15<68:49:06, 4.04it/s, grad_norm=0.751, loss_final=0.974, loss_mean=0.935, loss_mean_cls=0.0978, proj_loss=-0.0593][2026-03-26 12:32:18] Step: 292, Training Logs: loss_final: 0.981551, loss_mean: 0.939989, proj_loss: -0.058548, loss_mean_cls: 0.100109, grad_norm: 0.777168 +Steps: 0%| | 293/1000000 [01:15<68:49:16, 4.04it/s, grad_norm=0.777, loss_final=0.982, loss_mean=0.94, loss_mean_cls=0.1, proj_loss=-0.0585][2026-03-26 12:32:18] Step: 293, Training Logs: loss_final: 0.973046, loss_mean: 0.932364, proj_loss: -0.058453, loss_mean_cls: 0.099136, grad_norm: 1.086967 +Steps: 0%| | 294/1000000 [01:15<68:44:26, 4.04it/s, grad_norm=1.09, loss_final=0.973, loss_mean=0.932, loss_mean_cls=0.0991, proj_loss=-0.0585][2026-03-26 12:32:19] Step: 294, Training Logs: loss_final: 1.000331, loss_mean: 0.957751, proj_loss: -0.055745, loss_mean_cls: 0.098325, grad_norm: 0.858071 +Steps: 0%| | 295/1000000 [01:15<68:43:03, 4.04it/s, grad_norm=0.858, loss_final=1, loss_mean=0.958, loss_mean_cls=0.0983, proj_loss=-0.0557][2026-03-26 12:32:19] Step: 295, Training Logs: loss_final: 0.986585, loss_mean: 0.946819, proj_loss: -0.058582, loss_mean_cls: 0.098349, grad_norm: 1.196973 +Steps: 0%| | 296/1000000 [01:16<68:42:34, 4.04it/s, grad_norm=1.2, loss_final=0.987, loss_mean=0.947, loss_mean_cls=0.0983, proj_loss=-0.0586][2026-03-26 12:32:19] Step: 296, Training Logs: loss_final: 0.989954, loss_mean: 0.953976, proj_loss: -0.060762, loss_mean_cls: 0.096740, grad_norm: 0.496121 +Steps: 0%| | 297/1000000 [01:16<106:09:29, 2.62it/s, grad_norm=0.496, loss_final=0.99, loss_mean=0.954, loss_mean_cls=0.0967, proj_loss=-0.0608][2026-03-26 12:32:20] Step: 297, Training Logs: loss_final: 0.986456, loss_mean: 0.945679, proj_loss: -0.058529, loss_mean_cls: 0.099307, grad_norm: 1.059560 +Steps: 0%| | 298/1000000 [01:17<94:52:28, 2.93it/s, grad_norm=1.06, loss_final=0.986, loss_mean=0.946, loss_mean_cls=0.0993, proj_loss=-0.0585] [2026-03-26 12:32:20] Step: 298, Training Logs: loss_final: 0.986627, loss_mean: 0.947036, proj_loss: -0.057947, loss_mean_cls: 0.097538, grad_norm: 0.756703 +Steps: 0%| | 299/1000000 [01:17<87:00:25, 3.19it/s, grad_norm=0.757, loss_final=0.987, loss_mean=0.947, loss_mean_cls=0.0975, proj_loss=-0.0579][2026-03-26 12:32:20] Step: 299, Training Logs: loss_final: 0.992831, loss_mean: 0.953316, proj_loss: -0.058194, loss_mean_cls: 0.097709, grad_norm: 0.976663 +Steps: 0%| | 300/1000000 [01:17<81:30:14, 3.41it/s, grad_norm=0.977, loss_final=0.993, loss_mean=0.953, loss_mean_cls=0.0977, proj_loss=-0.0582][2026-03-26 12:32:21] Step: 300, Training Logs: loss_final: 0.995491, loss_mean: 0.957289, proj_loss: -0.058005, loss_mean_cls: 0.096207, grad_norm: 0.762918 +Steps: 0%| | 301/1000000 [01:17<77:41:03, 3.57it/s, grad_norm=0.763, loss_final=0.995, loss_mean=0.957, loss_mean_cls=0.0962, proj_loss=-0.058][2026-03-26 12:32:21] Step: 301, Training Logs: loss_final: 0.972935, loss_mean: 0.935746, proj_loss: -0.059969, loss_mean_cls: 0.097158, grad_norm: 0.565393 +Steps: 0%| | 302/1000000 [01:18<75:00:07, 3.70it/s, grad_norm=0.565, loss_final=0.973, loss_mean=0.936, loss_mean_cls=0.0972, proj_loss=-0.06][2026-03-26 12:32:21] Step: 302, Training Logs: loss_final: 0.980464, loss_mean: 0.940112, proj_loss: -0.056666, loss_mean_cls: 0.097018, grad_norm: 0.640596 +Steps: 0%| | 303/1000000 [01:18<73:05:28, 3.80it/s, grad_norm=0.641, loss_final=0.98, loss_mean=0.94, loss_mean_cls=0.097, proj_loss=-0.0567][2026-03-26 12:32:21] Step: 303, Training Logs: loss_final: 0.974377, loss_mean: 0.936467, proj_loss: -0.059766, loss_mean_cls: 0.097676, grad_norm: 0.669195 +Steps: 0%| | 304/1000000 [01:18<71:45:54, 3.87it/s, grad_norm=0.669, loss_final=0.974, loss_mean=0.936, loss_mean_cls=0.0977, proj_loss=-0.0598][2026-03-26 12:32:21] Step: 304, Training Logs: loss_final: 0.983466, loss_mean: 0.944091, proj_loss: -0.057554, loss_mean_cls: 0.096929, grad_norm: 0.680904 +Steps: 0%| | 305/1000000 [01:18<70:49:31, 3.92it/s, grad_norm=0.681, loss_final=0.983, loss_mean=0.944, loss_mean_cls=0.0969, proj_loss=-0.0576][2026-03-26 12:32:22] Step: 305, Training Logs: loss_final: 0.979573, loss_mean: 0.938937, proj_loss: -0.056710, loss_mean_cls: 0.097345, grad_norm: 0.743392 +Steps: 0%| | 306/1000000 [01:19<70:10:02, 3.96it/s, grad_norm=0.743, loss_final=0.98, loss_mean=0.939, loss_mean_cls=0.0973, proj_loss=-0.0567][2026-03-26 12:32:22] Step: 306, Training Logs: loss_final: 0.990891, loss_mean: 0.951872, proj_loss: -0.057371, loss_mean_cls: 0.096390, grad_norm: 0.868185 +Steps: 0%| | 307/1000000 [01:19<69:44:26, 3.98it/s, grad_norm=0.868, loss_final=0.991, loss_mean=0.952, loss_mean_cls=0.0964, proj_loss=-0.0574][2026-03-26 12:32:22] Step: 307, Training Logs: loss_final: 0.959404, loss_mean: 0.921842, proj_loss: -0.059707, loss_mean_cls: 0.097269, grad_norm: 0.604670 +Steps: 0%| | 308/1000000 [01:19<69:24:59, 4.00it/s, grad_norm=0.605, loss_final=0.959, loss_mean=0.922, loss_mean_cls=0.0973, proj_loss=-0.0597][2026-03-26 12:32:22] Step: 308, Training Logs: loss_final: 0.968923, loss_mean: 0.930226, proj_loss: -0.057853, loss_mean_cls: 0.096550, grad_norm: 0.568940 +Steps: 0%| | 309/1000000 [01:19<69:13:54, 4.01it/s, grad_norm=0.569, loss_final=0.969, loss_mean=0.93, loss_mean_cls=0.0966, proj_loss=-0.0579][2026-03-26 12:32:23] Step: 309, Training Logs: loss_final: 0.973727, loss_mean: 0.934653, proj_loss: -0.058002, loss_mean_cls: 0.097075, grad_norm: 0.584207 +Steps: 0%| | 310/1000000 [01:20<91:03:36, 3.05it/s, grad_norm=0.584, loss_final=0.974, loss_mean=0.935, loss_mean_cls=0.0971, proj_loss=-0.058][2026-03-26 12:32:23] Step: 310, Training Logs: loss_final: 0.991983, loss_mean: 0.956769, proj_loss: -0.060448, loss_mean_cls: 0.095663, grad_norm: 0.913269 +Steps: 0%| | 311/1000000 [01:20<84:19:45, 3.29it/s, grad_norm=0.913, loss_final=0.992, loss_mean=0.957, loss_mean_cls=0.0957, proj_loss=-0.0604][2026-03-26 12:32:23] Step: 311, Training Logs: loss_final: 0.994248, loss_mean: 0.957794, proj_loss: -0.059178, loss_mean_cls: 0.095631, grad_norm: 0.646613 +Steps: 0%| | 312/1000000 [01:20<79:40:24, 3.49it/s, grad_norm=0.647, loss_final=0.994, loss_mean=0.958, loss_mean_cls=0.0956, proj_loss=-0.0592][2026-03-26 12:32:24] Step: 312, Training Logs: loss_final: 0.972140, loss_mean: 0.932302, proj_loss: -0.057819, loss_mean_cls: 0.097657, grad_norm: 0.631791 +Steps: 0%| | 313/1000000 [01:21<76:22:02, 3.64it/s, grad_norm=0.632, loss_final=0.972, loss_mean=0.932, loss_mean_cls=0.0977, proj_loss=-0.0578][2026-03-26 12:32:24] Step: 313, Training Logs: loss_final: 0.971353, loss_mean: 0.930580, proj_loss: -0.056625, loss_mean_cls: 0.097399, grad_norm: 0.780317 +Steps: 0%| | 314/1000000 [01:21<74:06:37, 3.75it/s, grad_norm=0.78, loss_final=0.971, loss_mean=0.931, loss_mean_cls=0.0974, proj_loss=-0.0566][2026-03-26 12:32:24] Step: 314, Training Logs: loss_final: 0.997982, loss_mean: 0.961877, proj_loss: -0.059728, loss_mean_cls: 0.095832, grad_norm: 1.137487 +Steps: 0%| | 315/1000000 [01:21<72:30:07, 3.83it/s, grad_norm=1.14, loss_final=0.998, loss_mean=0.962, loss_mean_cls=0.0958, proj_loss=-0.0597][2026-03-26 12:32:24] Step: 315, Training Logs: loss_final: 0.965903, loss_mean: 0.928316, proj_loss: -0.060693, loss_mean_cls: 0.098280, grad_norm: 1.028791 +Steps: 0%| | 316/1000000 [01:21<71:21:11, 3.89it/s, grad_norm=1.03, loss_final=0.966, loss_mean=0.928, loss_mean_cls=0.0983, proj_loss=-0.0607][2026-03-26 12:32:25] Step: 316, Training Logs: loss_final: 0.971787, loss_mean: 0.934083, proj_loss: -0.059599, loss_mean_cls: 0.097304, grad_norm: 1.007314 +Steps: 0%| | 317/1000000 [01:22<70:34:13, 3.93it/s, grad_norm=1.01, loss_final=0.972, loss_mean=0.934, loss_mean_cls=0.0973, proj_loss=-0.0596][2026-03-26 12:32:25] Step: 317, Training Logs: loss_final: 0.993669, loss_mean: 0.960738, proj_loss: -0.061117, loss_mean_cls: 0.094048, grad_norm: 0.723226 +Steps: 0%| | 318/1000000 [01:22<70:06:39, 3.96it/s, grad_norm=0.723, loss_final=0.994, loss_mean=0.961, loss_mean_cls=0.094, proj_loss=-0.0611][2026-03-26 12:32:25] Step: 318, Training Logs: loss_final: 0.954272, loss_mean: 0.916672, proj_loss: -0.058726, loss_mean_cls: 0.096326, grad_norm: 0.957483 +Steps: 0%| | 319/1000000 [01:22<69:40:39, 3.99it/s, grad_norm=0.957, loss_final=0.954, loss_mean=0.917, loss_mean_cls=0.0963, proj_loss=-0.0587][2026-03-26 12:32:25] Step: 319, Training Logs: loss_final: 0.979939, loss_mean: 0.940182, proj_loss: -0.057549, loss_mean_cls: 0.097306, grad_norm: 0.995893 +Steps: 0%| | 320/1000000 [01:22<69:25:19, 4.00it/s, grad_norm=0.996, loss_final=0.98, loss_mean=0.94, loss_mean_cls=0.0973, proj_loss=-0.0575][2026-03-26 12:32:26] Step: 320, Training Logs: loss_final: 0.987625, loss_mean: 0.946461, proj_loss: -0.056967, loss_mean_cls: 0.098131, grad_norm: 1.177954 +Steps: 0%| | 321/1000000 [01:23<69:12:15, 4.01it/s, grad_norm=1.18, loss_final=0.988, loss_mean=0.946, loss_mean_cls=0.0981, proj_loss=-0.057][2026-03-26 12:32:26] Step: 321, Training Logs: loss_final: 0.977528, loss_mean: 0.940133, proj_loss: -0.058415, loss_mean_cls: 0.095810, grad_norm: 0.694324 +Steps: 0%| | 322/1000000 [01:23<69:02:42, 4.02it/s, grad_norm=0.694, loss_final=0.978, loss_mean=0.94, loss_mean_cls=0.0958, proj_loss=-0.0584][2026-03-26 12:32:26] Step: 322, Training Logs: loss_final: 0.987167, loss_mean: 0.947536, proj_loss: -0.057764, loss_mean_cls: 0.097395, grad_norm: 0.908996 +Steps: 0%| | 323/1000000 [01:23<68:55:45, 4.03it/s, grad_norm=0.909, loss_final=0.987, loss_mean=0.948, loss_mean_cls=0.0974, proj_loss=-0.0578][2026-03-26 12:32:26] Step: 323, Training Logs: loss_final: 0.974244, loss_mean: 0.935659, proj_loss: -0.058670, loss_mean_cls: 0.097255, grad_norm: 1.250319 +Steps: 0%| | 324/1000000 [01:23<68:50:58, 4.03it/s, grad_norm=1.25, loss_final=0.974, loss_mean=0.936, loss_mean_cls=0.0973, proj_loss=-0.0587][2026-03-26 12:32:27] Step: 324, Training Logs: loss_final: 0.980883, loss_mean: 0.945333, proj_loss: -0.060655, loss_mean_cls: 0.096205, grad_norm: 0.868396 +Steps: 0%| | 325/1000000 [01:24<68:49:57, 4.03it/s, grad_norm=0.868, loss_final=0.981, loss_mean=0.945, loss_mean_cls=0.0962, proj_loss=-0.0607][2026-03-26 12:32:27] Step: 325, Training Logs: loss_final: 0.993653, loss_mean: 0.954988, proj_loss: -0.057439, loss_mean_cls: 0.096104, grad_norm: 0.973391 +Steps: 0%| | 326/1000000 [01:24<68:50:14, 4.03it/s, grad_norm=0.973, loss_final=0.994, loss_mean=0.955, loss_mean_cls=0.0961, proj_loss=-0.0574][2026-03-26 12:32:27] Step: 326, Training Logs: loss_final: 0.967760, loss_mean: 0.924994, proj_loss: -0.056133, loss_mean_cls: 0.098898, grad_norm: 1.101567 +Steps: 0%| | 327/1000000 [01:24<68:48:40, 4.04it/s, grad_norm=1.1, loss_final=0.968, loss_mean=0.925, loss_mean_cls=0.0989, proj_loss=-0.0561][2026-03-26 12:32:27] Step: 327, Training Logs: loss_final: 0.979715, loss_mean: 0.941959, proj_loss: -0.058407, loss_mean_cls: 0.096163, grad_norm: 0.885649 +Steps: 0%| | 328/1000000 [01:24<68:47:55, 4.04it/s, grad_norm=0.886, loss_final=0.98, loss_mean=0.942, loss_mean_cls=0.0962, proj_loss=-0.0584][2026-03-26 12:32:28] Step: 328, Training Logs: loss_final: 0.968544, loss_mean: 0.930652, proj_loss: -0.058014, loss_mean_cls: 0.095906, grad_norm: 0.855483 +Steps: 0%| | 329/1000000 [01:25<68:46:13, 4.04it/s, grad_norm=0.855, loss_final=0.969, loss_mean=0.931, loss_mean_cls=0.0959, proj_loss=-0.058][2026-03-26 12:32:28] Step: 329, Training Logs: loss_final: 0.980694, loss_mean: 0.943677, proj_loss: -0.058776, loss_mean_cls: 0.095792, grad_norm: 1.032594 +Steps: 0%| | 330/1000000 [01:25<68:44:28, 4.04it/s, grad_norm=1.03, loss_final=0.981, loss_mean=0.944, loss_mean_cls=0.0958, proj_loss=-0.0588][2026-03-26 12:32:28] Step: 330, Training Logs: loss_final: 0.953122, loss_mean: 0.912742, proj_loss: -0.058648, loss_mean_cls: 0.099028, grad_norm: 0.958204 +Steps: 0%| | 331/1000000 [01:25<68:44:41, 4.04it/s, grad_norm=0.958, loss_final=0.953, loss_mean=0.913, loss_mean_cls=0.099, proj_loss=-0.0586][2026-03-26 12:32:28] Step: 331, Training Logs: loss_final: 0.989100, loss_mean: 0.955247, proj_loss: -0.061621, loss_mean_cls: 0.095474, grad_norm: 0.948178 +Steps: 0%| | 332/1000000 [01:25<68:48:03, 4.04it/s, grad_norm=0.948, loss_final=0.989, loss_mean=0.955, loss_mean_cls=0.0955, proj_loss=-0.0616][2026-03-26 12:32:29] Step: 332, Training Logs: loss_final: 0.977777, loss_mean: 0.943190, proj_loss: -0.059974, loss_mean_cls: 0.094562, grad_norm: 0.516064 +Steps: 0%| | 333/1000000 [01:26<68:49:23, 4.03it/s, grad_norm=0.516, loss_final=0.978, loss_mean=0.943, loss_mean_cls=0.0946, proj_loss=-0.06][2026-03-26 12:32:29] Step: 333, Training Logs: loss_final: 0.971033, loss_mean: 0.932698, proj_loss: -0.058641, loss_mean_cls: 0.096976, grad_norm: 0.822070 +Steps: 0%| | 334/1000000 [01:26<68:47:03, 4.04it/s, grad_norm=0.822, loss_final=0.971, loss_mean=0.933, loss_mean_cls=0.097, proj_loss=-0.0586][2026-03-26 12:32:29] Step: 334, Training Logs: loss_final: 0.989305, loss_mean: 0.952394, proj_loss: -0.059226, loss_mean_cls: 0.096137, grad_norm: 0.689624 +Steps: 0%| | 335/1000000 [01:26<68:48:21, 4.04it/s, grad_norm=0.69, loss_final=0.989, loss_mean=0.952, loss_mean_cls=0.0961, proj_loss=-0.0592][2026-03-26 12:32:29] Step: 335, Training Logs: loss_final: 0.950438, loss_mean: 0.911332, proj_loss: -0.058388, loss_mean_cls: 0.097494, grad_norm: 0.619024 +Steps: 0%| | 336/1000000 [01:26<68:48:04, 4.04it/s, grad_norm=0.619, loss_final=0.95, loss_mean=0.911, loss_mean_cls=0.0975, proj_loss=-0.0584][2026-03-26 12:32:30] Step: 336, Training Logs: loss_final: 0.980081, loss_mean: 0.945710, proj_loss: -0.059802, loss_mean_cls: 0.094173, grad_norm: 0.913568 +Steps: 0%| | 337/1000000 [01:27<68:47:18, 4.04it/s, grad_norm=0.914, loss_final=0.98, loss_mean=0.946, loss_mean_cls=0.0942, proj_loss=-0.0598][2026-03-26 12:32:30] Step: 337, Training Logs: loss_final: 0.959860, loss_mean: 0.921331, proj_loss: -0.058880, loss_mean_cls: 0.097409, grad_norm: 0.945853 +Steps: 0%| | 338/1000000 [01:27<68:46:13, 4.04it/s, grad_norm=0.946, loss_final=0.96, loss_mean=0.921, loss_mean_cls=0.0974, proj_loss=-0.0589][2026-03-26 12:32:30] Step: 338, Training Logs: loss_final: 0.996890, loss_mean: 0.961259, proj_loss: -0.058683, loss_mean_cls: 0.094313, grad_norm: 1.018520 +Steps: 0%| | 339/1000000 [01:27<68:46:09, 4.04it/s, grad_norm=1.02, loss_final=0.997, loss_mean=0.961, loss_mean_cls=0.0943, proj_loss=-0.0587][2026-03-26 12:32:30] Step: 339, Training Logs: loss_final: 0.949175, loss_mean: 0.913320, proj_loss: -0.061024, loss_mean_cls: 0.096879, grad_norm: 1.054598 +Steps: 0%| | 340/1000000 [01:27<68:47:08, 4.04it/s, grad_norm=1.05, loss_final=0.949, loss_mean=0.913, loss_mean_cls=0.0969, proj_loss=-0.061][2026-03-26 12:32:31] Step: 340, Training Logs: loss_final: 0.957431, loss_mean: 0.917898, proj_loss: -0.057336, loss_mean_cls: 0.096869, grad_norm: 0.906872 +Steps: 0%| | 341/1000000 [01:28<68:52:22, 4.03it/s, grad_norm=0.907, loss_final=0.957, loss_mean=0.918, loss_mean_cls=0.0969, proj_loss=-0.0573][2026-03-26 12:32:31] Step: 341, Training Logs: loss_final: 0.974826, loss_mean: 0.939626, proj_loss: -0.060440, loss_mean_cls: 0.095640, grad_norm: 0.917959 +Steps: 0%| | 342/1000000 [01:28<68:48:29, 4.04it/s, grad_norm=0.918, loss_final=0.975, loss_mean=0.94, loss_mean_cls=0.0956, proj_loss=-0.0604][2026-03-26 12:32:31] Step: 342, Training Logs: loss_final: 0.972647, loss_mean: 0.938696, proj_loss: -0.060789, loss_mean_cls: 0.094740, grad_norm: 1.208978 +Steps: 0%| | 343/1000000 [01:28<68:47:24, 4.04it/s, grad_norm=1.21, loss_final=0.973, loss_mean=0.939, loss_mean_cls=0.0947, proj_loss=-0.0608][2026-03-26 12:32:31] Step: 343, Training Logs: loss_final: 0.971566, loss_mean: 0.936859, proj_loss: -0.059510, loss_mean_cls: 0.094217, grad_norm: 0.705757 +Steps: 0%| | 344/1000000 [01:28<68:47:30, 4.04it/s, grad_norm=0.706, loss_final=0.972, loss_mean=0.937, loss_mean_cls=0.0942, proj_loss=-0.0595][2026-03-26 12:32:32] Step: 344, Training Logs: loss_final: 0.953052, loss_mean: 0.915850, proj_loss: -0.059945, loss_mean_cls: 0.097148, grad_norm: 1.139799 +Steps: 0%| | 345/1000000 [01:29<68:48:47, 4.04it/s, grad_norm=1.14, loss_final=0.953, loss_mean=0.916, loss_mean_cls=0.0971, proj_loss=-0.0599][2026-03-26 12:32:32] Step: 345, Training Logs: loss_final: 0.957886, loss_mean: 0.924656, proj_loss: -0.061742, loss_mean_cls: 0.094971, grad_norm: 0.607647 +Steps: 0%| | 346/1000000 [01:29<68:50:37, 4.03it/s, grad_norm=0.608, loss_final=0.958, loss_mean=0.925, loss_mean_cls=0.095, proj_loss=-0.0617][2026-03-26 12:32:32] Step: 346, Training Logs: loss_final: 0.971133, loss_mean: 0.934231, proj_loss: -0.059093, loss_mean_cls: 0.095995, grad_norm: 0.865365 +Steps: 0%| | 347/1000000 [01:29<69:04:29, 4.02it/s, grad_norm=0.865, loss_final=0.971, loss_mean=0.934, loss_mean_cls=0.096, proj_loss=-0.0591][2026-03-26 12:32:32] Step: 347, Training Logs: loss_final: 0.968399, loss_mean: 0.933130, proj_loss: -0.060792, loss_mean_cls: 0.096062, grad_norm: 0.701452 +Steps: 0%| | 348/1000000 [01:29<68:58:43, 4.03it/s, grad_norm=0.701, loss_final=0.968, loss_mean=0.933, loss_mean_cls=0.0961, proj_loss=-0.0608][2026-03-26 12:32:33] Step: 348, Training Logs: loss_final: 0.973296, loss_mean: 0.938913, proj_loss: -0.060402, loss_mean_cls: 0.094785, grad_norm: 0.909943 +Steps: 0%| | 349/1000000 [01:30<68:55:51, 4.03it/s, grad_norm=0.91, loss_final=0.973, loss_mean=0.939, loss_mean_cls=0.0948, proj_loss=-0.0604][2026-03-26 12:32:33] Step: 349, Training Logs: loss_final: 0.990313, loss_mean: 0.958213, proj_loss: -0.062356, loss_mean_cls: 0.094457, grad_norm: 0.706893 +Steps: 0%| | 350/1000000 [01:30<68:52:07, 4.03it/s, grad_norm=0.707, loss_final=0.99, loss_mean=0.958, loss_mean_cls=0.0945, proj_loss=-0.0624][2026-03-26 12:32:33] Step: 350, Training Logs: loss_final: 0.939510, loss_mean: 0.904307, proj_loss: -0.061636, loss_mean_cls: 0.096838, grad_norm: 0.877991 +Steps: 0%| | 351/1000000 [01:30<68:51:51, 4.03it/s, grad_norm=0.878, loss_final=0.94, loss_mean=0.904, loss_mean_cls=0.0968, proj_loss=-0.0616][2026-03-26 12:32:33] Step: 351, Training Logs: loss_final: 0.968714, loss_mean: 0.935258, proj_loss: -0.061353, loss_mean_cls: 0.094809, grad_norm: 0.743212 +Steps: 0%| | 352/1000000 [01:30<68:48:28, 4.04it/s, grad_norm=0.743, loss_final=0.969, loss_mean=0.935, loss_mean_cls=0.0948, proj_loss=-0.0614][2026-03-26 12:32:34] Step: 352, Training Logs: loss_final: 0.948356, loss_mean: 0.913849, proj_loss: -0.061868, loss_mean_cls: 0.096375, grad_norm: 0.928566 +Steps: 0%| | 353/1000000 [01:31<70:08:35, 3.96it/s, grad_norm=0.929, loss_final=0.948, loss_mean=0.914, loss_mean_cls=0.0964, proj_loss=-0.0619][2026-03-26 12:32:34] Step: 353, Training Logs: loss_final: 0.972928, loss_mean: 0.937405, proj_loss: -0.058977, loss_mean_cls: 0.094499, grad_norm: 0.902424 +Steps: 0%| | 354/1000000 [01:31<69:55:30, 3.97it/s, grad_norm=0.902, loss_final=0.973, loss_mean=0.937, loss_mean_cls=0.0945, proj_loss=-0.059][2026-03-26 12:32:34] Step: 354, Training Logs: loss_final: 0.951401, loss_mean: 0.915800, proj_loss: -0.060787, loss_mean_cls: 0.096388, grad_norm: 0.897284 +Steps: 0%| | 355/1000000 [01:31<69:35:23, 3.99it/s, grad_norm=0.897, loss_final=0.951, loss_mean=0.916, loss_mean_cls=0.0964, proj_loss=-0.0608][2026-03-26 12:32:34] Step: 355, Training Logs: loss_final: 0.976292, loss_mean: 0.942041, proj_loss: -0.059366, loss_mean_cls: 0.093616, grad_norm: 0.866879 +Steps: 0%| | 356/1000000 [01:31<69:23:30, 4.00it/s, grad_norm=0.867, loss_final=0.976, loss_mean=0.942, loss_mean_cls=0.0936, proj_loss=-0.0594][2026-03-26 12:32:35] Step: 356, Training Logs: loss_final: 0.939919, loss_mean: 0.901638, proj_loss: -0.059961, loss_mean_cls: 0.098242, grad_norm: 1.110362 +Steps: 0%| | 357/1000000 [01:32<69:11:14, 4.01it/s, grad_norm=1.11, loss_final=0.94, loss_mean=0.902, loss_mean_cls=0.0982, proj_loss=-0.06][2026-03-26 12:32:35] Step: 357, Training Logs: loss_final: 0.963302, loss_mean: 0.930078, proj_loss: -0.061121, loss_mean_cls: 0.094345, grad_norm: 0.737158 +Steps: 0%| | 358/1000000 [01:32<69:02:03, 4.02it/s, grad_norm=0.737, loss_final=0.963, loss_mean=0.93, loss_mean_cls=0.0943, proj_loss=-0.0611][2026-03-26 12:32:35] Step: 358, Training Logs: loss_final: 0.965756, loss_mean: 0.929180, proj_loss: -0.059791, loss_mean_cls: 0.096368, grad_norm: 0.734212 +Steps: 0%| | 359/1000000 [01:32<68:59:34, 4.02it/s, grad_norm=0.734, loss_final=0.966, loss_mean=0.929, loss_mean_cls=0.0964, proj_loss=-0.0598][2026-03-26 12:32:35] Step: 359, Training Logs: loss_final: 0.952311, loss_mean: 0.919599, proj_loss: -0.061965, loss_mean_cls: 0.094677, grad_norm: 1.029808 +Steps: 0%| | 360/1000000 [01:32<68:54:16, 4.03it/s, grad_norm=1.03, loss_final=0.952, loss_mean=0.92, loss_mean_cls=0.0947, proj_loss=-0.062][2026-03-26 12:32:36] Step: 360, Training Logs: loss_final: 0.971809, loss_mean: 0.939301, proj_loss: -0.061691, loss_mean_cls: 0.094200, grad_norm: 1.148289 +Steps: 0%| | 361/1000000 [01:33<68:52:22, 4.03it/s, grad_norm=1.15, loss_final=0.972, loss_mean=0.939, loss_mean_cls=0.0942, proj_loss=-0.0617][2026-03-26 12:32:36] Step: 361, Training Logs: loss_final: 0.956710, loss_mean: 0.920153, proj_loss: -0.060704, loss_mean_cls: 0.097260, grad_norm: 1.520403 +Steps: 0%| | 362/1000000 [01:33<68:52:41, 4.03it/s, grad_norm=1.52, loss_final=0.957, loss_mean=0.92, loss_mean_cls=0.0973, proj_loss=-0.0607][2026-03-26 12:32:36] Step: 362, Training Logs: loss_final: 0.963595, loss_mean: 0.931543, proj_loss: -0.062273, loss_mean_cls: 0.094325, grad_norm: 1.296332 +Steps: 0%| | 363/1000000 [01:33<68:50:04, 4.03it/s, grad_norm=1.3, loss_final=0.964, loss_mean=0.932, loss_mean_cls=0.0943, proj_loss=-0.0623][2026-03-26 12:32:36] Step: 363, Training Logs: loss_final: 0.959192, loss_mean: 0.925981, proj_loss: -0.061871, loss_mean_cls: 0.095082, grad_norm: 1.544958 +Steps: 0%| | 364/1000000 [01:33<68:52:38, 4.03it/s, grad_norm=1.54, loss_final=0.959, loss_mean=0.926, loss_mean_cls=0.0951, proj_loss=-0.0619][2026-03-26 12:32:37] Step: 364, Training Logs: loss_final: 0.967446, loss_mean: 0.936962, proj_loss: -0.063893, loss_mean_cls: 0.094377, grad_norm: 1.265578 +Steps: 0%| | 365/1000000 [01:33<68:53:57, 4.03it/s, grad_norm=1.27, loss_final=0.967, loss_mean=0.937, loss_mean_cls=0.0944, proj_loss=-0.0639][2026-03-26 12:32:37] Step: 365, Training Logs: loss_final: 0.952718, loss_mean: 0.918184, proj_loss: -0.061472, loss_mean_cls: 0.096006, grad_norm: 0.846024 +Steps: 0%| | 366/1000000 [01:34<68:53:01, 4.03it/s, grad_norm=0.846, loss_final=0.953, loss_mean=0.918, loss_mean_cls=0.096, proj_loss=-0.0615][2026-03-26 12:32:37] Step: 366, Training Logs: loss_final: 0.982892, loss_mean: 0.949254, proj_loss: -0.059607, loss_mean_cls: 0.093245, grad_norm: 1.279920 +Steps: 0%| | 367/1000000 [01:34<68:51:33, 4.03it/s, grad_norm=1.28, loss_final=0.983, loss_mean=0.949, loss_mean_cls=0.0932, proj_loss=-0.0596][2026-03-26 12:32:37] Step: 367, Training Logs: loss_final: 0.951445, loss_mean: 0.920499, proj_loss: -0.062574, loss_mean_cls: 0.093520, grad_norm: 0.784874 +Steps: 0%| | 368/1000000 [01:34<68:50:48, 4.03it/s, grad_norm=0.785, loss_final=0.951, loss_mean=0.92, loss_mean_cls=0.0935, proj_loss=-0.0626][2026-03-26 12:32:38] Step: 368, Training Logs: loss_final: 0.955815, loss_mean: 0.925198, proj_loss: -0.063369, loss_mean_cls: 0.093986, grad_norm: 1.303871 +Steps: 0%| | 369/1000000 [01:34<68:45:29, 4.04it/s, grad_norm=1.3, loss_final=0.956, loss_mean=0.925, loss_mean_cls=0.094, proj_loss=-0.0634][2026-03-26 12:32:38] Step: 369, Training Logs: loss_final: 0.947885, loss_mean: 0.914334, proj_loss: -0.062816, loss_mean_cls: 0.096367, grad_norm: 0.809083 +Steps: 0%| | 370/1000000 [01:35<68:44:42, 4.04it/s, grad_norm=0.809, loss_final=0.948, loss_mean=0.914, loss_mean_cls=0.0964, proj_loss=-0.0628][2026-03-26 12:32:38] Step: 370, Training Logs: loss_final: 0.963672, loss_mean: 0.932732, proj_loss: -0.063638, loss_mean_cls: 0.094578, grad_norm: 1.047808 +Steps: 0%| | 371/1000000 [01:35<68:41:40, 4.04it/s, grad_norm=1.05, loss_final=0.964, loss_mean=0.933, loss_mean_cls=0.0946, proj_loss=-0.0636][2026-03-26 12:32:38] Step: 371, Training Logs: loss_final: 0.961507, loss_mean: 0.927056, proj_loss: -0.060789, loss_mean_cls: 0.095240, grad_norm: 0.912434 +Steps: 0%| | 372/1000000 [01:35<68:41:36, 4.04it/s, grad_norm=0.912, loss_final=0.962, loss_mean=0.927, loss_mean_cls=0.0952, proj_loss=-0.0608][2026-03-26 12:32:39] Step: 372, Training Logs: loss_final: 0.961875, loss_mean: 0.931439, proj_loss: -0.064925, loss_mean_cls: 0.095361, grad_norm: 1.162064 +Steps: 0%| | 373/1000000 [01:35<68:40:33, 4.04it/s, grad_norm=1.16, loss_final=0.962, loss_mean=0.931, loss_mean_cls=0.0954, proj_loss=-0.0649][2026-03-26 12:32:39] Step: 373, Training Logs: loss_final: 0.948309, loss_mean: 0.920274, proj_loss: -0.065714, loss_mean_cls: 0.093750, grad_norm: 0.923702 +Steps: 0%| | 374/1000000 [01:36<68:42:27, 4.04it/s, grad_norm=0.924, loss_final=0.948, loss_mean=0.92, loss_mean_cls=0.0937, proj_loss=-0.0657][2026-03-26 12:32:39] Step: 374, Training Logs: loss_final: 0.966720, loss_mean: 0.935342, proj_loss: -0.063851, loss_mean_cls: 0.095229, grad_norm: 1.228301 +Steps: 0%| | 375/1000000 [01:36<68:44:08, 4.04it/s, grad_norm=1.23, loss_final=0.967, loss_mean=0.935, loss_mean_cls=0.0952, proj_loss=-0.0639][2026-03-26 12:32:39] Step: 375, Training Logs: loss_final: 0.976550, loss_mean: 0.946257, proj_loss: -0.062463, loss_mean_cls: 0.092757, grad_norm: 0.928498 +Steps: 0%| | 376/1000000 [01:36<68:46:02, 4.04it/s, grad_norm=0.928, loss_final=0.977, loss_mean=0.946, loss_mean_cls=0.0928, proj_loss=-0.0625][2026-03-26 12:32:40] Step: 376, Training Logs: loss_final: 0.947943, loss_mean: 0.917900, proj_loss: -0.064577, loss_mean_cls: 0.094620, grad_norm: 1.020114 +Steps: 0%| | 377/1000000 [01:36<68:45:09, 4.04it/s, grad_norm=1.02, loss_final=0.948, loss_mean=0.918, loss_mean_cls=0.0946, proj_loss=-0.0646][2026-03-26 12:32:40] Step: 377, Training Logs: loss_final: 0.980930, loss_mean: 0.951148, proj_loss: -0.063299, loss_mean_cls: 0.093081, grad_norm: 0.996992 +Steps: 0%| | 378/1000000 [01:37<68:45:04, 4.04it/s, grad_norm=0.997, loss_final=0.981, loss_mean=0.951, loss_mean_cls=0.0931, proj_loss=-0.0633][2026-03-26 12:32:40] Step: 378, Training Logs: loss_final: 0.968472, loss_mean: 0.939864, proj_loss: -0.064063, loss_mean_cls: 0.092671, grad_norm: 0.672573 +Steps: 0%| | 379/1000000 [01:37<68:44:14, 4.04it/s, grad_norm=0.673, loss_final=0.968, loss_mean=0.94, loss_mean_cls=0.0927, proj_loss=-0.0641][2026-03-26 12:32:40] Step: 379, Training Logs: loss_final: 0.961877, loss_mean: 0.933621, proj_loss: -0.066170, loss_mean_cls: 0.094425, grad_norm: 0.974544 +Steps: 0%| | 380/1000000 [01:37<68:44:22, 4.04it/s, grad_norm=0.975, loss_final=0.962, loss_mean=0.934, loss_mean_cls=0.0944, proj_loss=-0.0662][2026-03-26 12:32:41] Step: 380, Training Logs: loss_final: 0.947361, loss_mean: 0.917449, proj_loss: -0.065681, loss_mean_cls: 0.095593, grad_norm: 0.831001 +Steps: 0%| | 381/1000000 [01:37<68:43:43, 4.04it/s, grad_norm=0.831, loss_final=0.947, loss_mean=0.917, loss_mean_cls=0.0956, proj_loss=-0.0657][2026-03-26 12:32:41] Step: 381, Training Logs: loss_final: 0.929772, loss_mean: 0.896638, proj_loss: -0.064105, loss_mean_cls: 0.097239, grad_norm: 1.080806 +Steps: 0%| | 382/1000000 [01:38<68:44:21, 4.04it/s, grad_norm=1.08, loss_final=0.93, loss_mean=0.897, loss_mean_cls=0.0972, proj_loss=-0.0641][2026-03-26 12:32:41] Step: 382, Training Logs: loss_final: 0.961574, loss_mean: 0.933333, proj_loss: -0.065842, loss_mean_cls: 0.094083, grad_norm: 0.926316 +Steps: 0%| | 383/1000000 [01:38<68:43:54, 4.04it/s, grad_norm=0.926, loss_final=0.962, loss_mean=0.933, loss_mean_cls=0.0941, proj_loss=-0.0658][2026-03-26 12:32:41] Step: 383, Training Logs: loss_final: 0.963462, loss_mean: 0.936265, proj_loss: -0.066489, loss_mean_cls: 0.093686, grad_norm: 1.048921 +Steps: 0%| | 384/1000000 [01:38<68:45:37, 4.04it/s, grad_norm=1.05, loss_final=0.963, loss_mean=0.936, loss_mean_cls=0.0937, proj_loss=-0.0665][2026-03-26 12:32:42] Step: 384, Training Logs: loss_final: 0.970258, loss_mean: 0.941024, proj_loss: -0.065650, loss_mean_cls: 0.094885, grad_norm: 0.905657 +Steps: 0%| | 385/1000000 [01:38<68:47:55, 4.04it/s, grad_norm=0.906, loss_final=0.97, loss_mean=0.941, loss_mean_cls=0.0949, proj_loss=-0.0657][2026-03-26 12:32:42] Step: 385, Training Logs: loss_final: 0.957928, loss_mean: 0.929032, proj_loss: -0.065241, loss_mean_cls: 0.094136, grad_norm: 1.172701 +Steps: 0%| | 386/1000000 [01:39<68:44:05, 4.04it/s, grad_norm=1.17, loss_final=0.958, loss_mean=0.929, loss_mean_cls=0.0941, proj_loss=-0.0652][2026-03-26 12:32:42] Step: 386, Training Logs: loss_final: 0.939844, loss_mean: 0.913947, proj_loss: -0.068499, loss_mean_cls: 0.094396, grad_norm: 0.947779 +Steps: 0%| | 387/1000000 [01:39<68:43:48, 4.04it/s, grad_norm=0.948, loss_final=0.94, loss_mean=0.914, loss_mean_cls=0.0944, proj_loss=-0.0685][2026-03-26 12:32:42] Step: 387, Training Logs: loss_final: 0.937712, loss_mean: 0.911314, proj_loss: -0.069460, loss_mean_cls: 0.095858, grad_norm: 0.863876 +Steps: 0%| | 388/1000000 [01:39<68:41:50, 4.04it/s, grad_norm=0.864, loss_final=0.938, loss_mean=0.911, loss_mean_cls=0.0959, proj_loss=-0.0695][2026-03-26 12:32:43] Step: 388, Training Logs: loss_final: 0.953520, loss_mean: 0.925279, proj_loss: -0.065178, loss_mean_cls: 0.093418, grad_norm: 0.720599 +Steps: 0%| | 389/1000000 [01:39<68:45:31, 4.04it/s, grad_norm=0.721, loss_final=0.954, loss_mean=0.925, loss_mean_cls=0.0934, proj_loss=-0.0652][2026-03-26 12:32:43] Step: 389, Training Logs: loss_final: 0.965072, loss_mean: 0.937788, proj_loss: -0.066447, loss_mean_cls: 0.093731, grad_norm: 0.927520 +Steps: 0%| | 390/1000000 [01:40<68:44:56, 4.04it/s, grad_norm=0.928, loss_final=0.965, loss_mean=0.938, loss_mean_cls=0.0937, proj_loss=-0.0664][2026-03-26 12:32:43] Step: 390, Training Logs: loss_final: 0.932952, loss_mean: 0.904997, proj_loss: -0.067380, loss_mean_cls: 0.095335, grad_norm: 0.826062 +Steps: 0%| | 391/1000000 [01:40<68:42:59, 4.04it/s, grad_norm=0.826, loss_final=0.933, loss_mean=0.905, loss_mean_cls=0.0953, proj_loss=-0.0674][2026-03-26 12:32:43] Step: 391, Training Logs: loss_final: 0.958809, loss_mean: 0.934706, proj_loss: -0.070561, loss_mean_cls: 0.094664, grad_norm: 1.215514 +Steps: 0%| | 392/1000000 [01:40<68:43:49, 4.04it/s, grad_norm=1.22, loss_final=0.959, loss_mean=0.935, loss_mean_cls=0.0947, proj_loss=-0.0706][2026-03-26 12:32:44] Step: 392, Training Logs: loss_final: 0.932826, loss_mean: 0.904736, proj_loss: -0.067427, loss_mean_cls: 0.095517, grad_norm: 0.651382 +Steps: 0%| | 393/1000000 [01:40<68:43:05, 4.04it/s, grad_norm=0.651, loss_final=0.933, loss_mean=0.905, loss_mean_cls=0.0955, proj_loss=-0.0674][2026-03-26 12:32:44] Step: 393, Training Logs: loss_final: 0.954353, loss_mean: 0.931120, proj_loss: -0.070141, loss_mean_cls: 0.093374, grad_norm: 0.895906 +Steps: 0%| | 394/1000000 [01:41<68:44:18, 4.04it/s, grad_norm=0.896, loss_final=0.954, loss_mean=0.931, loss_mean_cls=0.0934, proj_loss=-0.0701][2026-03-26 12:32:44] Step: 394, Training Logs: loss_final: 0.930727, loss_mean: 0.905707, proj_loss: -0.069970, loss_mean_cls: 0.094990, grad_norm: 0.899356 +Steps: 0%| | 395/1000000 [01:41<68:44:15, 4.04it/s, grad_norm=0.899, loss_final=0.931, loss_mean=0.906, loss_mean_cls=0.095, proj_loss=-0.07][2026-03-26 12:32:44] Step: 395, Training Logs: loss_final: 0.952216, loss_mean: 0.927127, proj_loss: -0.069158, loss_mean_cls: 0.094247, grad_norm: 1.156296 +Steps: 0%| | 396/1000000 [01:41<68:43:51, 4.04it/s, grad_norm=1.16, loss_final=0.952, loss_mean=0.927, loss_mean_cls=0.0942, proj_loss=-0.0692][2026-03-26 12:32:45] Step: 396, Training Logs: loss_final: 0.938697, loss_mean: 0.909361, proj_loss: -0.066381, loss_mean_cls: 0.095718, grad_norm: 0.570524 +Steps: 0%| | 397/1000000 [01:41<68:43:29, 4.04it/s, grad_norm=0.571, loss_final=0.939, loss_mean=0.909, loss_mean_cls=0.0957, proj_loss=-0.0664][2026-03-26 12:32:45] Step: 397, Training Logs: loss_final: 0.947474, loss_mean: 0.923166, proj_loss: -0.070597, loss_mean_cls: 0.094905, grad_norm: 0.711728 +Steps: 0%| | 398/1000000 [01:42<68:43:22, 4.04it/s, grad_norm=0.712, loss_final=0.947, loss_mean=0.923, loss_mean_cls=0.0949, proj_loss=-0.0706][2026-03-26 12:32:45] Step: 398, Training Logs: loss_final: 0.927158, loss_mean: 0.903343, proj_loss: -0.070166, loss_mean_cls: 0.093981, grad_norm: 0.590612 +Steps: 0%| | 399/1000000 [01:42<68:42:16, 4.04it/s, grad_norm=0.591, loss_final=0.927, loss_mean=0.903, loss_mean_cls=0.094, proj_loss=-0.0702][2026-03-26 12:32:45] Step: 399, Training Logs: loss_final: 0.941669, loss_mean: 0.917680, proj_loss: -0.070649, loss_mean_cls: 0.094638, grad_norm: 0.800125 +Steps: 0%| | 400/1000000 [01:42<68:45:12, 4.04it/s, grad_norm=0.8, loss_final=0.942, loss_mean=0.918, loss_mean_cls=0.0946, proj_loss=-0.0706][2026-03-26 12:32:46] Step: 400, Training Logs: loss_final: 0.947262, loss_mean: 0.921733, proj_loss: -0.068512, loss_mean_cls: 0.094041, grad_norm: 0.887344 +Steps: 0%| | 401/1000000 [01:42<68:44:21, 4.04it/s, grad_norm=0.887, loss_final=0.947, loss_mean=0.922, loss_mean_cls=0.094, proj_loss=-0.0685][2026-03-26 12:32:46] Step: 401, Training Logs: loss_final: 0.949284, loss_mean: 0.926072, proj_loss: -0.070019, loss_mean_cls: 0.093230, grad_norm: 0.588558 +Steps: 0%| | 402/1000000 [01:43<68:44:20, 4.04it/s, grad_norm=0.589, loss_final=0.949, loss_mean=0.926, loss_mean_cls=0.0932, proj_loss=-0.07][2026-03-26 12:32:46] Step: 402, Training Logs: loss_final: 0.958678, loss_mean: 0.939297, proj_loss: -0.072234, loss_mean_cls: 0.091616, grad_norm: 1.121605 +Steps: 0%| | 403/1000000 [01:43<68:42:28, 4.04it/s, grad_norm=1.12, loss_final=0.959, loss_mean=0.939, loss_mean_cls=0.0916, proj_loss=-0.0722][2026-03-26 12:32:46] Step: 403, Training Logs: loss_final: 0.953129, loss_mean: 0.925118, proj_loss: -0.066602, loss_mean_cls: 0.094613, grad_norm: 1.088594 +Steps: 0%| | 404/1000000 [01:43<68:49:48, 4.03it/s, grad_norm=1.09, loss_final=0.953, loss_mean=0.925, loss_mean_cls=0.0946, proj_loss=-0.0666][2026-03-26 12:32:47] Step: 404, Training Logs: loss_final: 0.939545, loss_mean: 0.914792, proj_loss: -0.070116, loss_mean_cls: 0.094869, grad_norm: 1.158259 +Steps: 0%| | 405/1000000 [01:43<68:45:30, 4.04it/s, grad_norm=1.16, loss_final=0.94, loss_mean=0.915, loss_mean_cls=0.0949, proj_loss=-0.0701][2026-03-26 12:32:47] Step: 405, Training Logs: loss_final: 0.941582, loss_mean: 0.917620, proj_loss: -0.070003, loss_mean_cls: 0.093965, grad_norm: 0.774004 +Steps: 0%| | 406/1000000 [01:44<68:44:48, 4.04it/s, grad_norm=0.774, loss_final=0.942, loss_mean=0.918, loss_mean_cls=0.094, proj_loss=-0.07][2026-03-26 12:32:47] Step: 406, Training Logs: loss_final: 0.965814, loss_mean: 0.943036, proj_loss: -0.069846, loss_mean_cls: 0.092624, grad_norm: 1.860089 +Steps: 0%| | 407/1000000 [01:44<68:42:56, 4.04it/s, grad_norm=1.86, loss_final=0.966, loss_mean=0.943, loss_mean_cls=0.0926, proj_loss=-0.0698][2026-03-26 12:32:47] Step: 407, Training Logs: loss_final: 0.954215, loss_mean: 0.933159, proj_loss: -0.072002, loss_mean_cls: 0.093058, grad_norm: 0.849717 +Steps: 0%| | 408/1000000 [01:44<68:44:42, 4.04it/s, grad_norm=0.85, loss_final=0.954, loss_mean=0.933, loss_mean_cls=0.0931, proj_loss=-0.072][2026-03-26 12:32:48] Step: 408, Training Logs: loss_final: 0.942488, loss_mean: 0.920945, proj_loss: -0.071382, loss_mean_cls: 0.092926, grad_norm: 1.550707 +Steps: 0%| | 409/1000000 [01:44<68:40:52, 4.04it/s, grad_norm=1.55, loss_final=0.942, loss_mean=0.921, loss_mean_cls=0.0929, proj_loss=-0.0714][2026-03-26 12:32:48] Step: 409, Training Logs: loss_final: 0.937095, loss_mean: 0.910991, proj_loss: -0.068186, loss_mean_cls: 0.094290, grad_norm: 0.988069 +Steps: 0%| | 410/1000000 [01:45<68:42:39, 4.04it/s, grad_norm=0.988, loss_final=0.937, loss_mean=0.911, loss_mean_cls=0.0943, proj_loss=-0.0682][2026-03-26 12:32:48] Step: 410, Training Logs: loss_final: 0.957049, loss_mean: 0.937063, proj_loss: -0.070991, loss_mean_cls: 0.090977, grad_norm: 1.457527 +Steps: 0%| | 411/1000000 [01:45<68:39:39, 4.04it/s, grad_norm=1.46, loss_final=0.957, loss_mean=0.937, loss_mean_cls=0.091, proj_loss=-0.071][2026-03-26 12:32:48] Step: 411, Training Logs: loss_final: 0.951377, loss_mean: 0.932589, proj_loss: -0.072538, loss_mean_cls: 0.091326, grad_norm: 1.195244 +Steps: 0%| | 412/1000000 [01:45<68:42:36, 4.04it/s, grad_norm=1.2, loss_final=0.951, loss_mean=0.933, loss_mean_cls=0.0913, proj_loss=-0.0725][2026-03-26 12:32:49] Step: 412, Training Logs: loss_final: 0.929829, loss_mean: 0.904743, proj_loss: -0.071019, loss_mean_cls: 0.096106, grad_norm: 1.064008 +Steps: 0%| | 413/1000000 [01:45<68:44:52, 4.04it/s, grad_norm=1.06, loss_final=0.93, loss_mean=0.905, loss_mean_cls=0.0961, proj_loss=-0.071][2026-03-26 12:32:49] Step: 413, Training Logs: loss_final: 0.942425, loss_mean: 0.916583, proj_loss: -0.068610, loss_mean_cls: 0.094452, grad_norm: 0.959373 +Steps: 0%| | 414/1000000 [01:46<68:47:21, 4.04it/s, grad_norm=0.959, loss_final=0.942, loss_mean=0.917, loss_mean_cls=0.0945, proj_loss=-0.0686][2026-03-26 12:32:49] Step: 414, Training Logs: loss_final: 0.971553, loss_mean: 0.950608, proj_loss: -0.071467, loss_mean_cls: 0.092412, grad_norm: 0.977226 +Steps: 0%| | 415/1000000 [01:46<68:46:39, 4.04it/s, grad_norm=0.977, loss_final=0.972, loss_mean=0.951, loss_mean_cls=0.0924, proj_loss=-0.0715][2026-03-26 12:32:49] Step: 415, Training Logs: loss_final: 0.956796, loss_mean: 0.932939, proj_loss: -0.069283, loss_mean_cls: 0.093140, grad_norm: 0.867599 +Steps: 0%| | 416/1000000 [01:46<68:46:33, 4.04it/s, grad_norm=0.868, loss_final=0.957, loss_mean=0.933, loss_mean_cls=0.0931, proj_loss=-0.0693][2026-03-26 12:32:50] Step: 416, Training Logs: loss_final: 0.959126, loss_mean: 0.940532, proj_loss: -0.072940, loss_mean_cls: 0.091534, grad_norm: 1.220545 +Steps: 0%| | 417/1000000 [01:46<68:45:17, 4.04it/s, grad_norm=1.22, loss_final=0.959, loss_mean=0.941, loss_mean_cls=0.0915, proj_loss=-0.0729][2026-03-26 12:32:50] Step: 417, Training Logs: loss_final: 0.931903, loss_mean: 0.907659, proj_loss: -0.071049, loss_mean_cls: 0.095293, grad_norm: 0.846425 +Steps: 0%| | 418/1000000 [01:47<68:50:17, 4.03it/s, grad_norm=0.846, loss_final=0.932, loss_mean=0.908, loss_mean_cls=0.0953, proj_loss=-0.071][2026-03-26 12:32:50] Step: 418, Training Logs: loss_final: 0.937170, loss_mean: 0.915802, proj_loss: -0.072757, loss_mean_cls: 0.094126, grad_norm: 1.155763 +Steps: 0%| | 419/1000000 [01:47<68:52:51, 4.03it/s, grad_norm=1.16, loss_final=0.937, loss_mean=0.916, loss_mean_cls=0.0941, proj_loss=-0.0728][2026-03-26 12:32:50] Step: 419, Training Logs: loss_final: 0.960104, loss_mean: 0.937258, proj_loss: -0.071775, loss_mean_cls: 0.094620, grad_norm: 0.751360 +Steps: 0%| | 420/1000000 [01:47<68:55:18, 4.03it/s, grad_norm=0.751, loss_final=0.96, loss_mean=0.937, loss_mean_cls=0.0946, proj_loss=-0.0718][2026-03-26 12:32:51] Step: 420, Training Logs: loss_final: 0.942952, loss_mean: 0.918391, proj_loss: -0.070321, loss_mean_cls: 0.094882, grad_norm: 0.933231 +Steps: 0%| | 421/1000000 [01:47<68:55:52, 4.03it/s, grad_norm=0.933, loss_final=0.943, loss_mean=0.918, loss_mean_cls=0.0949, proj_loss=-0.0703][2026-03-26 12:32:51] Step: 421, Training Logs: loss_final: 0.938178, loss_mean: 0.916841, proj_loss: -0.071290, loss_mean_cls: 0.092627, grad_norm: 0.883586 +Steps: 0%| | 422/1000000 [01:48<68:58:04, 4.03it/s, grad_norm=0.884, loss_final=0.938, loss_mean=0.917, loss_mean_cls=0.0926, proj_loss=-0.0713][2026-03-26 12:32:51] Step: 422, Training Logs: loss_final: 0.947665, loss_mean: 0.931105, proj_loss: -0.074482, loss_mean_cls: 0.091042, grad_norm: 0.940980 +Steps: 0%| | 423/1000000 [01:48<68:59:26, 4.02it/s, grad_norm=0.941, loss_final=0.948, loss_mean=0.931, loss_mean_cls=0.091, proj_loss=-0.0745][2026-03-26 12:32:51] Step: 423, Training Logs: loss_final: 0.948093, loss_mean: 0.928835, proj_loss: -0.074084, loss_mean_cls: 0.093343, grad_norm: 0.947795 +Steps: 0%| | 424/1000000 [01:48<69:00:19, 4.02it/s, grad_norm=0.948, loss_final=0.948, loss_mean=0.929, loss_mean_cls=0.0933, proj_loss=-0.0741][2026-03-26 12:32:52] Step: 424, Training Logs: loss_final: 0.969444, loss_mean: 0.953289, proj_loss: -0.073314, loss_mean_cls: 0.089469, grad_norm: 0.709136 +Steps: 0%| | 425/1000000 [01:48<69:01:59, 4.02it/s, grad_norm=0.709, loss_final=0.969, loss_mean=0.953, loss_mean_cls=0.0895, proj_loss=-0.0733][2026-03-26 12:32:52] Step: 425, Training Logs: loss_final: 0.939860, loss_mean: 0.921725, proj_loss: -0.075261, loss_mean_cls: 0.093396, grad_norm: 1.149891 +Steps: 0%| | 426/1000000 [01:49<68:56:25, 4.03it/s, grad_norm=1.15, loss_final=0.94, loss_mean=0.922, loss_mean_cls=0.0934, proj_loss=-0.0753][2026-03-26 12:32:52] Step: 426, Training Logs: loss_final: 0.941353, loss_mean: 0.923296, proj_loss: -0.073094, loss_mean_cls: 0.091152, grad_norm: 0.657582 +Steps: 0%| | 427/1000000 [01:49<68:56:08, 4.03it/s, grad_norm=0.658, loss_final=0.941, loss_mean=0.923, loss_mean_cls=0.0912, proj_loss=-0.0731][2026-03-26 12:32:52] Step: 427, Training Logs: loss_final: 0.926625, loss_mean: 0.905451, proj_loss: -0.072144, loss_mean_cls: 0.093317, grad_norm: 0.690076 +Steps: 0%| | 428/1000000 [01:49<68:54:31, 4.03it/s, grad_norm=0.69, loss_final=0.927, loss_mean=0.905, loss_mean_cls=0.0933, proj_loss=-0.0721][2026-03-26 12:32:52] Step: 428, Training Logs: loss_final: 0.934484, loss_mean: 0.913044, proj_loss: -0.071139, loss_mean_cls: 0.092579, grad_norm: 0.777756 +Steps: 0%| | 429/1000000 [01:49<68:57:19, 4.03it/s, grad_norm=0.778, loss_final=0.934, loss_mean=0.913, loss_mean_cls=0.0926, proj_loss=-0.0711][2026-03-26 12:32:53] Step: 429, Training Logs: loss_final: 0.944020, loss_mean: 0.926098, proj_loss: -0.074322, loss_mean_cls: 0.092244, grad_norm: 0.575530 +Steps: 0%| | 430/1000000 [01:50<68:59:22, 4.02it/s, grad_norm=0.576, loss_final=0.944, loss_mean=0.926, loss_mean_cls=0.0922, proj_loss=-0.0743][2026-03-26 12:32:53] Step: 430, Training Logs: loss_final: 0.930984, loss_mean: 0.910969, proj_loss: -0.072271, loss_mean_cls: 0.092286, grad_norm: 0.825428 +Steps: 0%| | 431/1000000 [01:50<69:00:51, 4.02it/s, grad_norm=0.825, loss_final=0.931, loss_mean=0.911, loss_mean_cls=0.0923, proj_loss=-0.0723][2026-03-26 12:32:53] Step: 431, Training Logs: loss_final: 0.969337, loss_mean: 0.952128, proj_loss: -0.072420, loss_mean_cls: 0.089630, grad_norm: 0.440833 +Steps: 0%| | 432/1000000 [01:50<69:00:32, 4.02it/s, grad_norm=0.441, loss_final=0.969, loss_mean=0.952, loss_mean_cls=0.0896, proj_loss=-0.0724][2026-03-26 12:32:53] Step: 432, Training Logs: loss_final: 0.932275, loss_mean: 0.908903, proj_loss: -0.069877, loss_mean_cls: 0.093250, grad_norm: 0.951635 +Steps: 0%| | 433/1000000 [01:50<68:57:11, 4.03it/s, grad_norm=0.952, loss_final=0.932, loss_mean=0.909, loss_mean_cls=0.0932, proj_loss=-0.0699][2026-03-26 12:32:54] Step: 433, Training Logs: loss_final: 0.943494, loss_mean: 0.920707, proj_loss: -0.069528, loss_mean_cls: 0.092316, grad_norm: 0.835305 +Steps: 0%| | 434/1000000 [01:51<68:54:13, 4.03it/s, grad_norm=0.835, loss_final=0.943, loss_mean=0.921, loss_mean_cls=0.0923, proj_loss=-0.0695][2026-03-26 12:32:54] Step: 434, Training Logs: loss_final: 0.935211, loss_mean: 0.918395, proj_loss: -0.075661, loss_mean_cls: 0.092477, grad_norm: 0.785236 +Steps: 0%| | 435/1000000 [01:51<68:58:29, 4.03it/s, grad_norm=0.785, loss_final=0.935, loss_mean=0.918, loss_mean_cls=0.0925, proj_loss=-0.0757][2026-03-26 12:32:54] Step: 435, Training Logs: loss_final: 0.940965, loss_mean: 0.914778, proj_loss: -0.067792, loss_mean_cls: 0.093978, grad_norm: 0.542633 +Steps: 0%| | 436/1000000 [01:51<68:56:29, 4.03it/s, grad_norm=0.543, loss_final=0.941, loss_mean=0.915, loss_mean_cls=0.094, proj_loss=-0.0678][2026-03-26 12:32:54] Step: 436, Training Logs: loss_final: 0.938931, loss_mean: 0.920587, proj_loss: -0.072839, loss_mean_cls: 0.091182, grad_norm: 0.791480 +Steps: 0%| | 437/1000000 [01:51<68:53:51, 4.03it/s, grad_norm=0.791, loss_final=0.939, loss_mean=0.921, loss_mean_cls=0.0912, proj_loss=-0.0728][2026-03-26 12:32:55] Step: 437, Training Logs: loss_final: 0.921438, loss_mean: 0.907625, proj_loss: -0.078283, loss_mean_cls: 0.092096, grad_norm: 1.045191 +Steps: 0%| | 438/1000000 [01:52<68:54:44, 4.03it/s, grad_norm=1.05, loss_final=0.921, loss_mean=0.908, loss_mean_cls=0.0921, proj_loss=-0.0783][2026-03-26 12:32:55] Step: 438, Training Logs: loss_final: 0.937452, loss_mean: 0.920694, proj_loss: -0.074707, loss_mean_cls: 0.091465, grad_norm: 0.811324 +Steps: 0%| | 439/1000000 [01:52<68:56:17, 4.03it/s, grad_norm=0.811, loss_final=0.937, loss_mean=0.921, loss_mean_cls=0.0915, proj_loss=-0.0747][2026-03-26 12:32:55] Step: 439, Training Logs: loss_final: 0.945026, loss_mean: 0.930196, proj_loss: -0.074880, loss_mean_cls: 0.089710, grad_norm: 0.984050 +Steps: 0%| | 440/1000000 [01:52<68:50:10, 4.03it/s, grad_norm=0.984, loss_final=0.945, loss_mean=0.93, loss_mean_cls=0.0897, proj_loss=-0.0749][2026-03-26 12:32:55] Step: 440, Training Logs: loss_final: 0.941390, loss_mean: 0.924168, proj_loss: -0.073825, loss_mean_cls: 0.091047, grad_norm: 0.711068 +Steps: 0%| | 441/1000000 [01:52<68:50:53, 4.03it/s, grad_norm=0.711, loss_final=0.941, loss_mean=0.924, loss_mean_cls=0.091, proj_loss=-0.0738][2026-03-26 12:32:56] Step: 441, Training Logs: loss_final: 0.945572, loss_mean: 0.927673, proj_loss: -0.073774, loss_mean_cls: 0.091673, grad_norm: 0.753394 +Steps: 0%| | 442/1000000 [01:53<68:43:56, 4.04it/s, grad_norm=0.753, loss_final=0.946, loss_mean=0.928, loss_mean_cls=0.0917, proj_loss=-0.0738][2026-03-26 12:32:56] Step: 442, Training Logs: loss_final: 0.957468, loss_mean: 0.941595, proj_loss: -0.075330, loss_mean_cls: 0.091202, grad_norm: 1.072621 +Steps: 0%| | 443/1000000 [01:53<68:44:38, 4.04it/s, grad_norm=1.07, loss_final=0.957, loss_mean=0.942, loss_mean_cls=0.0912, proj_loss=-0.0753][2026-03-26 12:32:56] Step: 443, Training Logs: loss_final: 0.931388, loss_mean: 0.912547, proj_loss: -0.073586, loss_mean_cls: 0.092427, grad_norm: 0.607360 +Steps: 0%| | 444/1000000 [01:53<68:42:38, 4.04it/s, grad_norm=0.607, loss_final=0.931, loss_mean=0.913, loss_mean_cls=0.0924, proj_loss=-0.0736][2026-03-26 12:32:56] Step: 444, Training Logs: loss_final: 0.929467, loss_mean: 0.913911, proj_loss: -0.077389, loss_mean_cls: 0.092946, grad_norm: 1.002862 +Steps: 0%| | 445/1000000 [01:53<68:43:47, 4.04it/s, grad_norm=1, loss_final=0.929, loss_mean=0.914, loss_mean_cls=0.0929, proj_loss=-0.0774][2026-03-26 12:32:57] Step: 445, Training Logs: loss_final: 0.942839, loss_mean: 0.925642, proj_loss: -0.075463, loss_mean_cls: 0.092659, grad_norm: 1.395024 +Steps: 0%| | 446/1000000 [01:54<68:44:42, 4.04it/s, grad_norm=1.4, loss_final=0.943, loss_mean=0.926, loss_mean_cls=0.0927, proj_loss=-0.0755][2026-03-26 12:32:57] Step: 446, Training Logs: loss_final: 0.926895, loss_mean: 0.907491, proj_loss: -0.073708, loss_mean_cls: 0.093111, grad_norm: 0.892229 +Steps: 0%| | 447/1000000 [01:54<68:49:12, 4.03it/s, grad_norm=0.892, loss_final=0.927, loss_mean=0.907, loss_mean_cls=0.0931, proj_loss=-0.0737][2026-03-26 12:32:57] Step: 447, Training Logs: loss_final: 0.926978, loss_mean: 0.902371, proj_loss: -0.069979, loss_mean_cls: 0.094586, grad_norm: 1.356788 +Steps: 0%| | 448/1000000 [01:54<68:46:11, 4.04it/s, grad_norm=1.36, loss_final=0.927, loss_mean=0.902, loss_mean_cls=0.0946, proj_loss=-0.07][2026-03-26 12:32:57] Step: 448, Training Logs: loss_final: 0.949302, loss_mean: 0.934084, proj_loss: -0.074430, loss_mean_cls: 0.089648, grad_norm: 1.011136 +Steps: 0%| | 449/1000000 [01:54<68:44:36, 4.04it/s, grad_norm=1.01, loss_final=0.949, loss_mean=0.934, loss_mean_cls=0.0896, proj_loss=-0.0744][2026-03-26 12:32:58] Step: 449, Training Logs: loss_final: 0.951159, loss_mean: 0.940485, proj_loss: -0.078063, loss_mean_cls: 0.088737, grad_norm: 1.151049 +Steps: 0%| | 450/1000000 [01:55<68:45:57, 4.04it/s, grad_norm=1.15, loss_final=0.951, loss_mean=0.94, loss_mean_cls=0.0887, proj_loss=-0.0781][2026-03-26 12:32:58] Step: 450, Training Logs: loss_final: 0.940795, loss_mean: 0.925427, proj_loss: -0.075217, loss_mean_cls: 0.090585, grad_norm: 1.089298 +Steps: 0%| | 451/1000000 [01:55<68:42:16, 4.04it/s, grad_norm=1.09, loss_final=0.941, loss_mean=0.925, loss_mean_cls=0.0906, proj_loss=-0.0752][2026-03-26 12:32:58] Step: 451, Training Logs: loss_final: 0.945210, loss_mean: 0.929403, proj_loss: -0.075684, loss_mean_cls: 0.091491, grad_norm: 0.625609 +Steps: 0%| | 452/1000000 [01:55<68:42:29, 4.04it/s, grad_norm=0.626, loss_final=0.945, loss_mean=0.929, loss_mean_cls=0.0915, proj_loss=-0.0757][2026-03-26 12:32:58] Step: 452, Training Logs: loss_final: 0.944826, loss_mean: 0.925392, proj_loss: -0.072282, loss_mean_cls: 0.091715, grad_norm: 1.239029 +Steps: 0%| | 453/1000000 [01:55<68:41:08, 4.04it/s, grad_norm=1.24, loss_final=0.945, loss_mean=0.925, loss_mean_cls=0.0917, proj_loss=-0.0723][2026-03-26 12:32:59] Step: 453, Training Logs: loss_final: 0.946454, loss_mean: 0.928948, proj_loss: -0.075627, loss_mean_cls: 0.093134, grad_norm: 1.094389 +Steps: 0%| | 454/1000000 [01:56<68:42:38, 4.04it/s, grad_norm=1.09, loss_final=0.946, loss_mean=0.929, loss_mean_cls=0.0931, proj_loss=-0.0756][2026-03-26 12:32:59] Step: 454, Training Logs: loss_final: 0.955351, loss_mean: 0.936005, proj_loss: -0.072653, loss_mean_cls: 0.092000, grad_norm: 0.883869 +Steps: 0%| | 455/1000000 [01:56<68:38:11, 4.05it/s, grad_norm=0.884, loss_final=0.955, loss_mean=0.936, loss_mean_cls=0.092, proj_loss=-0.0727][2026-03-26 12:32:59] Step: 455, Training Logs: loss_final: 0.940507, loss_mean: 0.924791, proj_loss: -0.075629, loss_mean_cls: 0.091345, grad_norm: 0.854470 +Steps: 0%| | 456/1000000 [01:56<68:39:37, 4.04it/s, grad_norm=0.854, loss_final=0.941, loss_mean=0.925, loss_mean_cls=0.0913, proj_loss=-0.0756][2026-03-26 12:32:59] Step: 456, Training Logs: loss_final: 0.924301, loss_mean: 0.907731, proj_loss: -0.076105, loss_mean_cls: 0.092675, grad_norm: 0.685292 +Steps: 0%| | 457/1000000 [01:56<68:42:12, 4.04it/s, grad_norm=0.685, loss_final=0.924, loss_mean=0.908, loss_mean_cls=0.0927, proj_loss=-0.0761][2026-03-26 12:33:00] Step: 457, Training Logs: loss_final: 0.924928, loss_mean: 0.908737, proj_loss: -0.075850, loss_mean_cls: 0.092041, grad_norm: 0.776079 +Steps: 0%| | 458/1000000 [01:57<68:40:16, 4.04it/s, grad_norm=0.776, loss_final=0.925, loss_mean=0.909, loss_mean_cls=0.092, proj_loss=-0.0759][2026-03-26 12:33:00] Step: 458, Training Logs: loss_final: 0.934830, loss_mean: 0.919473, proj_loss: -0.076267, loss_mean_cls: 0.091623, grad_norm: 0.793863 +Steps: 0%| | 459/1000000 [01:57<68:37:37, 4.05it/s, grad_norm=0.794, loss_final=0.935, loss_mean=0.919, loss_mean_cls=0.0916, proj_loss=-0.0763][2026-03-26 12:33:00] Step: 459, Training Logs: loss_final: 0.936312, loss_mean: 0.920752, proj_loss: -0.075777, loss_mean_cls: 0.091338, grad_norm: 1.214958 +Steps: 0%| | 460/1000000 [01:57<68:38:41, 4.04it/s, grad_norm=1.21, loss_final=0.936, loss_mean=0.921, loss_mean_cls=0.0913, proj_loss=-0.0758][2026-03-26 12:33:00] Step: 460, Training Logs: loss_final: 0.931532, loss_mean: 0.915761, proj_loss: -0.077025, loss_mean_cls: 0.092796, grad_norm: 0.524263 +Steps: 0%| | 461/1000000 [01:57<68:37:10, 4.05it/s, grad_norm=0.524, loss_final=0.932, loss_mean=0.916, loss_mean_cls=0.0928, proj_loss=-0.077][2026-03-26 12:33:01] Step: 461, Training Logs: loss_final: 0.921111, loss_mean: 0.904385, proj_loss: -0.076505, loss_mean_cls: 0.093231, grad_norm: 0.935109 +Steps: 0%| | 462/1000000 [01:58<68:40:39, 4.04it/s, grad_norm=0.935, loss_final=0.921, loss_mean=0.904, loss_mean_cls=0.0932, proj_loss=-0.0765][2026-03-26 12:33:01] Step: 462, Training Logs: loss_final: 0.944126, loss_mean: 0.925596, proj_loss: -0.074309, loss_mean_cls: 0.092838, grad_norm: 1.115015 +Steps: 0%| | 463/1000000 [01:58<68:42:39, 4.04it/s, grad_norm=1.12, loss_final=0.944, loss_mean=0.926, loss_mean_cls=0.0928, proj_loss=-0.0743][2026-03-26 12:33:01] Step: 463, Training Logs: loss_final: 0.936919, loss_mean: 0.925302, proj_loss: -0.077997, loss_mean_cls: 0.089614, grad_norm: 0.825607 +Steps: 0%| | 464/1000000 [01:58<68:46:08, 4.04it/s, grad_norm=0.826, loss_final=0.937, loss_mean=0.925, loss_mean_cls=0.0896, proj_loss=-0.078][2026-03-26 12:33:01] Step: 464, Training Logs: loss_final: 0.904013, loss_mean: 0.891953, proj_loss: -0.080657, loss_mean_cls: 0.092717, grad_norm: 0.735744 +Steps: 0%| | 465/1000000 [01:58<68:45:26, 4.04it/s, grad_norm=0.736, loss_final=0.904, loss_mean=0.892, loss_mean_cls=0.0927, proj_loss=-0.0807][2026-03-26 12:33:02] Step: 465, Training Logs: loss_final: 0.951164, loss_mean: 0.939970, proj_loss: -0.077634, loss_mean_cls: 0.088828, grad_norm: 1.206003 +Steps: 0%| | 466/1000000 [01:59<68:48:06, 4.04it/s, grad_norm=1.21, loss_final=0.951, loss_mean=0.94, loss_mean_cls=0.0888, proj_loss=-0.0776][2026-03-26 12:33:02] Step: 466, Training Logs: loss_final: 0.944732, loss_mean: 0.930165, proj_loss: -0.076326, loss_mean_cls: 0.090893, grad_norm: 0.722064 +Steps: 0%| | 467/1000000 [01:59<68:44:20, 4.04it/s, grad_norm=0.722, loss_final=0.945, loss_mean=0.93, loss_mean_cls=0.0909, proj_loss=-0.0763][2026-03-26 12:33:02] Step: 467, Training Logs: loss_final: 0.927726, loss_mean: 0.911858, proj_loss: -0.076706, loss_mean_cls: 0.092574, grad_norm: 1.106638 +Steps: 0%| | 468/1000000 [01:59<68:45:54, 4.04it/s, grad_norm=1.11, loss_final=0.928, loss_mean=0.912, loss_mean_cls=0.0926, proj_loss=-0.0767][2026-03-26 12:33:02] Step: 468, Training Logs: loss_final: 0.934944, loss_mean: 0.924443, proj_loss: -0.080806, loss_mean_cls: 0.091307, grad_norm: 1.096277 +Steps: 0%| | 469/1000000 [01:59<68:46:06, 4.04it/s, grad_norm=1.1, loss_final=0.935, loss_mean=0.924, loss_mean_cls=0.0913, proj_loss=-0.0808][2026-03-26 12:33:03] Step: 469, Training Logs: loss_final: 0.922004, loss_mean: 0.909482, proj_loss: -0.079068, loss_mean_cls: 0.091590, grad_norm: 0.785522 +Steps: 0%| | 470/1000000 [02:00<68:46:10, 4.04it/s, grad_norm=0.786, loss_final=0.922, loss_mean=0.909, loss_mean_cls=0.0916, proj_loss=-0.0791][2026-03-26 12:33:03] Step: 470, Training Logs: loss_final: 0.926421, loss_mean: 0.913781, proj_loss: -0.078847, loss_mean_cls: 0.091487, grad_norm: 0.962209 +Steps: 0%| | 471/1000000 [02:00<70:06:36, 3.96it/s, grad_norm=0.962, loss_final=0.926, loss_mean=0.914, loss_mean_cls=0.0915, proj_loss=-0.0788][2026-03-26 12:33:03] Step: 471, Training Logs: loss_final: 0.912004, loss_mean: 0.899084, proj_loss: -0.079241, loss_mean_cls: 0.092160, grad_norm: 0.898818 +Steps: 0%| | 472/1000000 [02:00<69:38:19, 3.99it/s, grad_norm=0.899, loss_final=0.912, loss_mean=0.899, loss_mean_cls=0.0922, proj_loss=-0.0792][2026-03-26 12:33:03] Step: 472, Training Logs: loss_final: 0.918869, loss_mean: 0.905143, proj_loss: -0.078153, loss_mean_cls: 0.091879, grad_norm: 0.525482 +Steps: 0%| | 473/1000000 [02:00<69:31:13, 3.99it/s, grad_norm=0.525, loss_final=0.919, loss_mean=0.905, loss_mean_cls=0.0919, proj_loss=-0.0782][2026-03-26 12:33:04] Step: 473, Training Logs: loss_final: 0.931920, loss_mean: 0.919501, proj_loss: -0.079622, loss_mean_cls: 0.092040, grad_norm: 0.795339 +Steps: 0%| | 474/1000000 [02:01<69:17:23, 4.01it/s, grad_norm=0.795, loss_final=0.932, loss_mean=0.92, loss_mean_cls=0.092, proj_loss=-0.0796][2026-03-26 12:33:04] Step: 474, Training Logs: loss_final: 0.941014, loss_mean: 0.930640, proj_loss: -0.080020, loss_mean_cls: 0.090394, grad_norm: 0.376308 +Steps: 0%| | 475/1000000 [02:01<69:07:12, 4.02it/s, grad_norm=0.376, loss_final=0.941, loss_mean=0.931, loss_mean_cls=0.0904, proj_loss=-0.08][2026-03-26 12:33:04] Step: 475, Training Logs: loss_final: 0.921053, loss_mean: 0.907522, proj_loss: -0.078750, loss_mean_cls: 0.092280, grad_norm: 0.856080 +Steps: 0%| | 476/1000000 [02:01<68:59:19, 4.02it/s, grad_norm=0.856, loss_final=0.921, loss_mean=0.908, loss_mean_cls=0.0923, proj_loss=-0.0787][2026-03-26 12:33:04] Step: 476, Training Logs: loss_final: 0.944330, loss_mean: 0.934826, proj_loss: -0.081077, loss_mean_cls: 0.090582, grad_norm: 0.564409 +Steps: 0%| | 477/1000000 [02:01<68:51:02, 4.03it/s, grad_norm=0.564, loss_final=0.944, loss_mean=0.935, loss_mean_cls=0.0906, proj_loss=-0.0811][2026-03-26 12:33:05] Step: 477, Training Logs: loss_final: 0.934376, loss_mean: 0.925861, proj_loss: -0.082763, loss_mean_cls: 0.091279, grad_norm: 0.966937 +Steps: 0%| | 478/1000000 [02:02<68:47:32, 4.04it/s, grad_norm=0.967, loss_final=0.934, loss_mean=0.926, loss_mean_cls=0.0913, proj_loss=-0.0828][2026-03-26 12:33:05] Step: 478, Training Logs: loss_final: 0.917335, loss_mean: 0.904983, proj_loss: -0.079739, loss_mean_cls: 0.092091, grad_norm: 0.846318 +Steps: 0%| | 479/1000000 [02:02<68:49:12, 4.03it/s, grad_norm=0.846, loss_final=0.917, loss_mean=0.905, loss_mean_cls=0.0921, proj_loss=-0.0797][2026-03-26 12:33:05] Step: 479, Training Logs: loss_final: 0.917869, loss_mean: 0.907629, proj_loss: -0.081629, loss_mean_cls: 0.091869, grad_norm: 1.055029 +Steps: 0%| | 480/1000000 [02:02<68:48:13, 4.04it/s, grad_norm=1.06, loss_final=0.918, loss_mean=0.908, loss_mean_cls=0.0919, proj_loss=-0.0816][2026-03-26 12:33:05] Step: 480, Training Logs: loss_final: 0.914161, loss_mean: 0.901797, proj_loss: -0.079872, loss_mean_cls: 0.092235, grad_norm: 0.714486 +Steps: 0%| | 481/1000000 [02:02<68:47:19, 4.04it/s, grad_norm=0.714, loss_final=0.914, loss_mean=0.902, loss_mean_cls=0.0922, proj_loss=-0.0799][2026-03-26 12:33:06] Step: 481, Training Logs: loss_final: 0.915056, loss_mean: 0.903562, proj_loss: -0.080745, loss_mean_cls: 0.092239, grad_norm: 1.002460 +Steps: 0%| | 482/1000000 [02:02<68:43:00, 4.04it/s, grad_norm=1, loss_final=0.915, loss_mean=0.904, loss_mean_cls=0.0922, proj_loss=-0.0807][2026-03-26 12:33:06] Step: 482, Training Logs: loss_final: 0.921440, loss_mean: 0.915393, proj_loss: -0.084713, loss_mean_cls: 0.090760, grad_norm: 0.493002 +Steps: 0%| | 483/1000000 [02:03<68:45:22, 4.04it/s, grad_norm=0.493, loss_final=0.921, loss_mean=0.915, loss_mean_cls=0.0908, proj_loss=-0.0847][2026-03-26 12:33:06] Step: 483, Training Logs: loss_final: 0.917179, loss_mean: 0.906275, proj_loss: -0.079236, loss_mean_cls: 0.090140, grad_norm: 0.959927 +Steps: 0%| | 484/1000000 [02:03<68:43:55, 4.04it/s, grad_norm=0.96, loss_final=0.917, loss_mean=0.906, loss_mean_cls=0.0901, proj_loss=-0.0792][2026-03-26 12:33:06] Step: 484, Training Logs: loss_final: 0.912586, loss_mean: 0.904824, proj_loss: -0.083719, loss_mean_cls: 0.091480, grad_norm: 1.062528 +Steps: 0%| | 485/1000000 [02:03<68:45:02, 4.04it/s, grad_norm=1.06, loss_final=0.913, loss_mean=0.905, loss_mean_cls=0.0915, proj_loss=-0.0837][2026-03-26 12:33:07] Step: 485, Training Logs: loss_final: 0.911882, loss_mean: 0.902366, proj_loss: -0.083031, loss_mean_cls: 0.092548, grad_norm: 1.164581 +Steps: 0%| | 486/1000000 [02:03<68:48:33, 4.03it/s, grad_norm=1.16, loss_final=0.912, loss_mean=0.902, loss_mean_cls=0.0925, proj_loss=-0.083][2026-03-26 12:33:07] Step: 486, Training Logs: loss_final: 0.930021, loss_mean: 0.919133, proj_loss: -0.080385, loss_mean_cls: 0.091273, grad_norm: 0.923461 +Steps: 0%| | 487/1000000 [02:04<68:51:51, 4.03it/s, grad_norm=0.923, loss_final=0.93, loss_mean=0.919, loss_mean_cls=0.0913, proj_loss=-0.0804][2026-03-26 12:33:07] Step: 487, Training Logs: loss_final: 0.949757, loss_mean: 0.939885, proj_loss: -0.079140, loss_mean_cls: 0.089012, grad_norm: 0.866302 +Steps: 0%| | 488/1000000 [02:04<68:55:34, 4.03it/s, grad_norm=0.866, loss_final=0.95, loss_mean=0.94, loss_mean_cls=0.089, proj_loss=-0.0791][2026-03-26 12:33:07] Step: 488, Training Logs: loss_final: 0.891810, loss_mean: 0.883580, proj_loss: -0.084685, loss_mean_cls: 0.092915, grad_norm: 0.836278 +Steps: 0%| | 489/1000000 [02:04<68:54:32, 4.03it/s, grad_norm=0.836, loss_final=0.892, loss_mean=0.884, loss_mean_cls=0.0929, proj_loss=-0.0847][2026-03-26 12:33:08] Step: 489, Training Logs: loss_final: 0.931517, loss_mean: 0.921550, proj_loss: -0.080801, loss_mean_cls: 0.090767, grad_norm: 0.849123 +Steps: 0%| | 490/1000000 [02:04<68:57:12, 4.03it/s, grad_norm=0.849, loss_final=0.932, loss_mean=0.922, loss_mean_cls=0.0908, proj_loss=-0.0808][2026-03-26 12:33:08] Step: 490, Training Logs: loss_final: 0.929950, loss_mean: 0.916498, proj_loss: -0.077953, loss_mean_cls: 0.091405, grad_norm: 1.365064 +Steps: 0%| | 491/1000000 [02:05<68:56:09, 4.03it/s, grad_norm=1.37, loss_final=0.93, loss_mean=0.916, loss_mean_cls=0.0914, proj_loss=-0.078][2026-03-26 12:33:08] Step: 491, Training Logs: loss_final: 0.948958, loss_mean: 0.941494, proj_loss: -0.082699, loss_mean_cls: 0.090163, grad_norm: 0.628506 +Steps: 0%| | 492/1000000 [02:05<68:59:52, 4.02it/s, grad_norm=0.629, loss_final=0.949, loss_mean=0.941, loss_mean_cls=0.0902, proj_loss=-0.0827][2026-03-26 12:33:08] Step: 492, Training Logs: loss_final: 0.921489, loss_mean: 0.912172, proj_loss: -0.081903, loss_mean_cls: 0.091220, grad_norm: 1.102094 +Steps: 0%| | 493/1000000 [02:05<69:01:44, 4.02it/s, grad_norm=1.1, loss_final=0.921, loss_mean=0.912, loss_mean_cls=0.0912, proj_loss=-0.0819][2026-03-26 12:33:09] Step: 493, Training Logs: loss_final: 0.925975, loss_mean: 0.918879, proj_loss: -0.084499, loss_mean_cls: 0.091595, grad_norm: 1.248804 +Steps: 0%| | 494/1000000 [02:05<68:54:02, 4.03it/s, grad_norm=1.25, loss_final=0.926, loss_mean=0.919, loss_mean_cls=0.0916, proj_loss=-0.0845][2026-03-26 12:33:09] Step: 494, Training Logs: loss_final: 0.931541, loss_mean: 0.926138, proj_loss: -0.083435, loss_mean_cls: 0.088838, grad_norm: 1.137338 +Steps: 0%| | 495/1000000 [02:06<69:00:06, 4.02it/s, grad_norm=1.14, loss_final=0.932, loss_mean=0.926, loss_mean_cls=0.0888, proj_loss=-0.0834][2026-03-26 12:33:09] Step: 495, Training Logs: loss_final: 0.925495, loss_mean: 0.918745, proj_loss: -0.083494, loss_mean_cls: 0.090244, grad_norm: 1.089026 +Steps: 0%| | 496/1000000 [02:06<68:58:42, 4.03it/s, grad_norm=1.09, loss_final=0.925, loss_mean=0.919, loss_mean_cls=0.0902, proj_loss=-0.0835][2026-03-26 12:33:09] Step: 496, Training Logs: loss_final: 0.940602, loss_mean: 0.937743, proj_loss: -0.084331, loss_mean_cls: 0.087190, grad_norm: 0.737297 +Steps: 0%| | 497/1000000 [02:06<68:59:38, 4.02it/s, grad_norm=0.737, loss_final=0.941, loss_mean=0.938, loss_mean_cls=0.0872, proj_loss=-0.0843][2026-03-26 12:33:10] Step: 497, Training Logs: loss_final: 0.916204, loss_mean: 0.910975, proj_loss: -0.084754, loss_mean_cls: 0.089984, grad_norm: 0.837046 +Steps: 0%| | 498/1000000 [02:06<68:59:58, 4.02it/s, grad_norm=0.837, loss_final=0.916, loss_mean=0.911, loss_mean_cls=0.09, proj_loss=-0.0848][2026-03-26 12:33:10] Step: 498, Training Logs: loss_final: 0.934876, loss_mean: 0.933581, proj_loss: -0.086928, loss_mean_cls: 0.088222, grad_norm: 0.699164 +Steps: 0%| | 499/1000000 [02:07<69:00:04, 4.02it/s, grad_norm=0.699, loss_final=0.935, loss_mean=0.934, loss_mean_cls=0.0882, proj_loss=-0.0869][2026-03-26 12:33:10] Step: 499, Training Logs: loss_final: 0.926161, loss_mean: 0.919881, proj_loss: -0.085008, loss_mean_cls: 0.091287, grad_norm: 1.189578 +Steps: 0%| | 500/1000000 [02:07<68:59:59, 4.02it/s, grad_norm=1.19, loss_final=0.926, loss_mean=0.92, loss_mean_cls=0.0913, proj_loss=-0.085][2026-03-26 12:33:10] Step: 500, Training Logs: loss_final: 0.911040, loss_mean: 0.902704, proj_loss: -0.083298, loss_mean_cls: 0.091633, grad_norm: 0.687825 +Steps: 0%| | 501/1000000 [02:07<68:58:37, 4.03it/s, grad_norm=0.688, loss_final=0.911, loss_mean=0.903, loss_mean_cls=0.0916, proj_loss=-0.0833][2026-03-26 12:33:11] Step: 501, Training Logs: loss_final: 0.939269, loss_mean: 0.933076, proj_loss: -0.083060, loss_mean_cls: 0.089252, grad_norm: 0.763203 +Steps: 0%| | 502/1000000 [02:07<69:00:02, 4.02it/s, grad_norm=0.763, loss_final=0.939, loss_mean=0.933, loss_mean_cls=0.0893, proj_loss=-0.0831][2026-03-26 12:33:11] Step: 502, Training Logs: loss_final: 0.916848, loss_mean: 0.910087, proj_loss: -0.083972, loss_mean_cls: 0.090733, grad_norm: 0.772783 +Steps: 0%| | 503/1000000 [02:08<68:53:34, 4.03it/s, grad_norm=0.773, loss_final=0.917, loss_mean=0.91, loss_mean_cls=0.0907, proj_loss=-0.084][2026-03-26 12:33:11] Step: 503, Training Logs: loss_final: 0.909969, loss_mean: 0.904876, proj_loss: -0.084153, loss_mean_cls: 0.089246, grad_norm: 0.892310 +Steps: 0%| | 504/1000000 [02:08<68:49:53, 4.03it/s, grad_norm=0.892, loss_final=0.91, loss_mean=0.905, loss_mean_cls=0.0892, proj_loss=-0.0842][2026-03-26 12:33:11] Step: 504, Training Logs: loss_final: 0.935077, loss_mean: 0.927995, proj_loss: -0.083303, loss_mean_cls: 0.090385, grad_norm: 1.002171 +Steps: 0%| | 505/1000000 [02:08<68:51:56, 4.03it/s, grad_norm=1, loss_final=0.935, loss_mean=0.928, loss_mean_cls=0.0904, proj_loss=-0.0833][2026-03-26 12:33:12] Step: 505, Training Logs: loss_final: 0.925269, loss_mean: 0.923098, proj_loss: -0.086505, loss_mean_cls: 0.088677, grad_norm: 0.600044 +Steps: 0%| | 506/1000000 [02:08<68:50:00, 4.03it/s, grad_norm=0.6, loss_final=0.925, loss_mean=0.923, loss_mean_cls=0.0887, proj_loss=-0.0865][2026-03-26 12:33:12] Step: 506, Training Logs: loss_final: 0.905582, loss_mean: 0.898994, proj_loss: -0.084240, loss_mean_cls: 0.090827, grad_norm: 0.839953 +Steps: 0%| | 507/1000000 [02:09<68:55:17, 4.03it/s, grad_norm=0.84, loss_final=0.906, loss_mean=0.899, loss_mean_cls=0.0908, proj_loss=-0.0842][2026-03-26 12:33:12] Step: 507, Training Logs: loss_final: 0.914504, loss_mean: 0.909708, proj_loss: -0.084919, loss_mean_cls: 0.089715, grad_norm: 0.998060 +Steps: 0%| | 508/1000000 [02:09<71:21:11, 3.89it/s, grad_norm=0.998, loss_final=0.915, loss_mean=0.91, loss_mean_cls=0.0897, proj_loss=-0.0849][2026-03-26 12:33:12] Step: 508, Training Logs: loss_final: 0.912581, loss_mean: 0.906473, proj_loss: -0.084328, loss_mean_cls: 0.090436, grad_norm: 0.584099 +Steps: 0%| | 509/1000000 [02:09<70:30:55, 3.94it/s, grad_norm=0.584, loss_final=0.913, loss_mean=0.906, loss_mean_cls=0.0904, proj_loss=-0.0843][2026-03-26 12:33:13] Step: 509, Training Logs: loss_final: 0.902957, loss_mean: 0.899079, proj_loss: -0.087343, loss_mean_cls: 0.091221, grad_norm: 0.951284 +Steps: 0%| | 510/1000000 [02:09<69:54:34, 3.97it/s, grad_norm=0.951, loss_final=0.903, loss_mean=0.899, loss_mean_cls=0.0912, proj_loss=-0.0873][2026-03-26 12:33:13] Step: 510, Training Logs: loss_final: 0.904078, loss_mean: 0.897093, proj_loss: -0.084751, loss_mean_cls: 0.091736, grad_norm: 0.455076 +Steps: 0%| | 511/1000000 [02:10<69:38:18, 3.99it/s, grad_norm=0.455, loss_final=0.904, loss_mean=0.897, loss_mean_cls=0.0917, proj_loss=-0.0848][2026-03-26 12:33:13] Step: 511, Training Logs: loss_final: 0.901537, loss_mean: 0.898350, proj_loss: -0.087922, loss_mean_cls: 0.091109, grad_norm: 0.825463 +Steps: 0%| | 512/1000000 [02:10<69:23:28, 4.00it/s, grad_norm=0.825, loss_final=0.902, loss_mean=0.898, loss_mean_cls=0.0911, proj_loss=-0.0879][2026-03-26 12:33:13] Step: 512, Training Logs: loss_final: 0.932737, loss_mean: 0.927224, proj_loss: -0.083400, loss_mean_cls: 0.088913, grad_norm: 0.942898 +Steps: 0%| | 513/1000000 [02:10<69:10:45, 4.01it/s, grad_norm=0.943, loss_final=0.933, loss_mean=0.927, loss_mean_cls=0.0889, proj_loss=-0.0834][2026-03-26 12:33:14] Step: 513, Training Logs: loss_final: 0.919693, loss_mean: 0.915804, proj_loss: -0.085662, loss_mean_cls: 0.089551, grad_norm: 0.927648 +Steps: 0%| | 514/1000000 [02:10<69:05:16, 4.02it/s, grad_norm=0.928, loss_final=0.92, loss_mean=0.916, loss_mean_cls=0.0896, proj_loss=-0.0857][2026-03-26 12:33:14] Step: 514, Training Logs: loss_final: 0.884082, loss_mean: 0.874614, proj_loss: -0.084799, loss_mean_cls: 0.094268, grad_norm: 0.778576 +Steps: 0%| | 515/1000000 [02:11<69:07:06, 4.02it/s, grad_norm=0.779, loss_final=0.884, loss_mean=0.875, loss_mean_cls=0.0943, proj_loss=-0.0848][2026-03-26 12:33:14] Step: 515, Training Logs: loss_final: 0.924644, loss_mean: 0.919017, proj_loss: -0.084979, loss_mean_cls: 0.090607, grad_norm: 0.511041 +Steps: 0%| | 516/1000000 [02:11<69:02:10, 4.02it/s, grad_norm=0.511, loss_final=0.925, loss_mean=0.919, loss_mean_cls=0.0906, proj_loss=-0.085][2026-03-26 12:33:14] Step: 516, Training Logs: loss_final: 0.898089, loss_mean: 0.887908, proj_loss: -0.082178, loss_mean_cls: 0.092359, grad_norm: 0.967360 +Steps: 0%| | 517/1000000 [02:11<68:54:04, 4.03it/s, grad_norm=0.967, loss_final=0.898, loss_mean=0.888, loss_mean_cls=0.0924, proj_loss=-0.0822][2026-03-26 12:33:15] Step: 517, Training Logs: loss_final: 0.940722, loss_mean: 0.936606, proj_loss: -0.083919, loss_mean_cls: 0.088035, grad_norm: 1.237571 +Steps: 0%| | 518/1000000 [02:11<68:52:37, 4.03it/s, grad_norm=1.24, loss_final=0.941, loss_mean=0.937, loss_mean_cls=0.088, proj_loss=-0.0839][2026-03-26 12:33:15] Step: 518, Training Logs: loss_final: 0.935150, loss_mean: 0.938818, proj_loss: -0.090335, loss_mean_cls: 0.086667, grad_norm: 0.617628 +Steps: 0%| | 519/1000000 [02:12<68:48:03, 4.04it/s, grad_norm=0.618, loss_final=0.935, loss_mean=0.939, loss_mean_cls=0.0867, proj_loss=-0.0903][2026-03-26 12:33:15] Step: 519, Training Logs: loss_final: 0.914580, loss_mean: 0.913023, proj_loss: -0.088680, loss_mean_cls: 0.090237, grad_norm: 0.923509 +Steps: 0%| | 520/1000000 [02:12<69:23:20, 4.00it/s, grad_norm=0.924, loss_final=0.915, loss_mean=0.913, loss_mean_cls=0.0902, proj_loss=-0.0887][2026-03-26 12:33:15] Step: 520, Training Logs: loss_final: 0.921522, loss_mean: 0.920933, proj_loss: -0.089323, loss_mean_cls: 0.089912, grad_norm: 1.032585 +Steps: 0%| | 521/1000000 [02:12<69:11:14, 4.01it/s, grad_norm=1.03, loss_final=0.922, loss_mean=0.921, loss_mean_cls=0.0899, proj_loss=-0.0893][2026-03-26 12:33:16] Step: 521, Training Logs: loss_final: 0.906107, loss_mean: 0.900967, proj_loss: -0.086895, loss_mean_cls: 0.092035, grad_norm: 0.901073 +Steps: 0%| | 522/1000000 [02:12<69:04:59, 4.02it/s, grad_norm=0.901, loss_final=0.906, loss_mean=0.901, loss_mean_cls=0.092, proj_loss=-0.0869][2026-03-26 12:33:16] Step: 522, Training Logs: loss_final: 0.914346, loss_mean: 0.910641, proj_loss: -0.087581, loss_mean_cls: 0.091287, grad_norm: 0.825407 +Steps: 0%| | 523/1000000 [02:13<68:57:16, 4.03it/s, grad_norm=0.825, loss_final=0.914, loss_mean=0.911, loss_mean_cls=0.0913, proj_loss=-0.0876][2026-03-26 12:33:16] Step: 523, Training Logs: loss_final: 0.897823, loss_mean: 0.887441, proj_loss: -0.083141, loss_mean_cls: 0.093524, grad_norm: 1.012766 +Steps: 0%| | 524/1000000 [02:13<68:53:49, 4.03it/s, grad_norm=1.01, loss_final=0.898, loss_mean=0.887, loss_mean_cls=0.0935, proj_loss=-0.0831][2026-03-26 12:33:16] Step: 524, Training Logs: loss_final: 0.922028, loss_mean: 0.923167, proj_loss: -0.089711, loss_mean_cls: 0.088573, grad_norm: 0.818106 +Steps: 0%| | 525/1000000 [02:13<68:45:25, 4.04it/s, grad_norm=0.818, loss_final=0.922, loss_mean=0.923, loss_mean_cls=0.0886, proj_loss=-0.0897][2026-03-26 12:33:17] Step: 525, Training Logs: loss_final: 0.905333, loss_mean: 0.901363, proj_loss: -0.087551, loss_mean_cls: 0.091521, grad_norm: 0.874414 +Steps: 0%| | 526/1000000 [02:13<68:47:19, 4.04it/s, grad_norm=0.874, loss_final=0.905, loss_mean=0.901, loss_mean_cls=0.0915, proj_loss=-0.0876][2026-03-26 12:33:17] Step: 526, Training Logs: loss_final: 0.910733, loss_mean: 0.908694, proj_loss: -0.088249, loss_mean_cls: 0.090288, grad_norm: 0.877222 +Steps: 0%| | 527/1000000 [02:14<68:46:31, 4.04it/s, grad_norm=0.877, loss_final=0.911, loss_mean=0.909, loss_mean_cls=0.0903, proj_loss=-0.0882][2026-03-26 12:33:17] Step: 527, Training Logs: loss_final: 0.916549, loss_mean: 0.916161, proj_loss: -0.089101, loss_mean_cls: 0.089489, grad_norm: 1.202938 +Steps: 0%| | 528/1000000 [02:14<68:46:36, 4.04it/s, grad_norm=1.2, loss_final=0.917, loss_mean=0.916, loss_mean_cls=0.0895, proj_loss=-0.0891][2026-03-26 12:33:17] Step: 528, Training Logs: loss_final: 0.910237, loss_mean: 0.910098, proj_loss: -0.089650, loss_mean_cls: 0.089789, grad_norm: 0.662663 +Steps: 0%| | 529/1000000 [02:14<68:41:21, 4.04it/s, grad_norm=0.663, loss_final=0.91, loss_mean=0.91, loss_mean_cls=0.0898, proj_loss=-0.0896][2026-03-26 12:33:18] Step: 529, Training Logs: loss_final: 0.924975, loss_mean: 0.924330, proj_loss: -0.088661, loss_mean_cls: 0.089307, grad_norm: 0.967660 +Steps: 0%| | 530/1000000 [02:14<68:40:12, 4.04it/s, grad_norm=0.968, loss_final=0.925, loss_mean=0.924, loss_mean_cls=0.0893, proj_loss=-0.0887][2026-03-26 12:33:18] Step: 530, Training Logs: loss_final: 0.910272, loss_mean: 0.909389, proj_loss: -0.089278, loss_mean_cls: 0.090161, grad_norm: 0.599244 +Steps: 0%| | 531/1000000 [02:15<68:38:48, 4.04it/s, grad_norm=0.599, loss_final=0.91, loss_mean=0.909, loss_mean_cls=0.0902, proj_loss=-0.0893][2026-03-26 12:33:18] Step: 531, Training Logs: loss_final: 0.901985, loss_mean: 0.901523, proj_loss: -0.090439, loss_mean_cls: 0.090901, grad_norm: 0.812302 +Steps: 0%| | 532/1000000 [02:15<68:38:04, 4.05it/s, grad_norm=0.812, loss_final=0.902, loss_mean=0.902, loss_mean_cls=0.0909, proj_loss=-0.0904][2026-03-26 12:33:18] Step: 532, Training Logs: loss_final: 0.917732, loss_mean: 0.921353, proj_loss: -0.093821, loss_mean_cls: 0.090201, grad_norm: 0.621560 +Steps: 0%| | 533/1000000 [02:15<68:36:09, 4.05it/s, grad_norm=0.622, loss_final=0.918, loss_mean=0.921, loss_mean_cls=0.0902, proj_loss=-0.0938][2026-03-26 12:33:19] Step: 533, Training Logs: loss_final: 0.905940, loss_mean: 0.904175, proj_loss: -0.088688, loss_mean_cls: 0.090452, grad_norm: 0.813458 +Steps: 0%| | 534/1000000 [02:15<68:35:37, 4.05it/s, grad_norm=0.813, loss_final=0.906, loss_mean=0.904, loss_mean_cls=0.0905, proj_loss=-0.0887][2026-03-26 12:33:19] Step: 534, Training Logs: loss_final: 0.897958, loss_mean: 0.898176, proj_loss: -0.089573, loss_mean_cls: 0.089355, grad_norm: 0.694430 +Steps: 0%| | 535/1000000 [02:16<68:37:56, 4.05it/s, grad_norm=0.694, loss_final=0.898, loss_mean=0.898, loss_mean_cls=0.0894, proj_loss=-0.0896][2026-03-26 12:33:19] Step: 535, Training Logs: loss_final: 0.926684, loss_mean: 0.931643, proj_loss: -0.092472, loss_mean_cls: 0.087513, grad_norm: 0.977427 +Steps: 0%| | 536/1000000 [02:16<68:38:00, 4.05it/s, grad_norm=0.977, loss_final=0.927, loss_mean=0.932, loss_mean_cls=0.0875, proj_loss=-0.0925][2026-03-26 12:33:19] Step: 536, Training Logs: loss_final: 0.905180, loss_mean: 0.904050, proj_loss: -0.089722, loss_mean_cls: 0.090853, grad_norm: 0.590290 +Steps: 0%| | 537/1000000 [02:16<68:40:17, 4.04it/s, grad_norm=0.59, loss_final=0.905, loss_mean=0.904, loss_mean_cls=0.0909, proj_loss=-0.0897][2026-03-26 12:33:20] Step: 537, Training Logs: loss_final: 0.894763, loss_mean: 0.892534, proj_loss: -0.089544, loss_mean_cls: 0.091773, grad_norm: 1.016889 +Steps: 0%| | 538/1000000 [02:16<68:40:37, 4.04it/s, grad_norm=1.02, loss_final=0.895, loss_mean=0.893, loss_mean_cls=0.0918, proj_loss=-0.0895][2026-03-26 12:33:20] Step: 538, Training Logs: loss_final: 0.913964, loss_mean: 0.914880, proj_loss: -0.090820, loss_mean_cls: 0.089904, grad_norm: 0.659975 +Steps: 0%| | 539/1000000 [02:17<68:42:03, 4.04it/s, grad_norm=0.66, loss_final=0.914, loss_mean=0.915, loss_mean_cls=0.0899, proj_loss=-0.0908][2026-03-26 12:33:20] Step: 539, Training Logs: loss_final: 0.904055, loss_mean: 0.905963, proj_loss: -0.091824, loss_mean_cls: 0.089916, grad_norm: 0.823219 +Steps: 0%| | 540/1000000 [02:17<68:41:07, 4.04it/s, grad_norm=0.823, loss_final=0.904, loss_mean=0.906, loss_mean_cls=0.0899, proj_loss=-0.0918][2026-03-26 12:33:20] Step: 540, Training Logs: loss_final: 0.907664, loss_mean: 0.914611, proj_loss: -0.094773, loss_mean_cls: 0.087827, grad_norm: 0.745015 +Steps: 0%| | 541/1000000 [02:17<68:46:17, 4.04it/s, grad_norm=0.745, loss_final=0.908, loss_mean=0.915, loss_mean_cls=0.0878, proj_loss=-0.0948][2026-03-26 12:33:21] Step: 541, Training Logs: loss_final: 0.890856, loss_mean: 0.894633, proj_loss: -0.092485, loss_mean_cls: 0.088708, grad_norm: 0.721815 +Steps: 0%| | 542/1000000 [02:17<68:45:43, 4.04it/s, grad_norm=0.722, loss_final=0.891, loss_mean=0.895, loss_mean_cls=0.0887, proj_loss=-0.0925][2026-03-26 12:33:21] Step: 542, Training Logs: loss_final: 0.917880, loss_mean: 0.919669, proj_loss: -0.090890, loss_mean_cls: 0.089100, grad_norm: 0.726091 +Steps: 0%| | 543/1000000 [02:18<68:43:53, 4.04it/s, grad_norm=0.726, loss_final=0.918, loss_mean=0.92, loss_mean_cls=0.0891, proj_loss=-0.0909][2026-03-26 12:33:21] Step: 543, Training Logs: loss_final: 0.892306, loss_mean: 0.891870, proj_loss: -0.091863, loss_mean_cls: 0.092298, grad_norm: 1.070594 +Steps: 0%| | 544/1000000 [02:18<68:45:00, 4.04it/s, grad_norm=1.07, loss_final=0.892, loss_mean=0.892, loss_mean_cls=0.0923, proj_loss=-0.0919][2026-03-26 12:33:21] Step: 544, Training Logs: loss_final: 0.908152, loss_mean: 0.913689, proj_loss: -0.094598, loss_mean_cls: 0.089060, grad_norm: 1.092224 +Steps: 0%| | 545/1000000 [02:18<68:44:22, 4.04it/s, grad_norm=1.09, loss_final=0.908, loss_mean=0.914, loss_mean_cls=0.0891, proj_loss=-0.0946][2026-03-26 12:33:22] Step: 545, Training Logs: loss_final: 0.915384, loss_mean: 0.913844, proj_loss: -0.089180, loss_mean_cls: 0.090719, grad_norm: 1.183913 +Steps: 0%| | 546/1000000 [02:18<68:44:23, 4.04it/s, grad_norm=1.18, loss_final=0.915, loss_mean=0.914, loss_mean_cls=0.0907, proj_loss=-0.0892][2026-03-26 12:33:22] Step: 546, Training Logs: loss_final: 0.914878, loss_mean: 0.917967, proj_loss: -0.092135, loss_mean_cls: 0.089045, grad_norm: 0.742442 +Steps: 0%| | 547/1000000 [02:19<68:41:24, 4.04it/s, grad_norm=0.742, loss_final=0.915, loss_mean=0.918, loss_mean_cls=0.089, proj_loss=-0.0921][2026-03-26 12:33:22] Step: 547, Training Logs: loss_final: 0.910105, loss_mean: 0.917032, proj_loss: -0.096351, loss_mean_cls: 0.089424, grad_norm: 1.088664 +Steps: 0%| | 548/1000000 [02:19<69:35:39, 3.99it/s, grad_norm=1.09, loss_final=0.91, loss_mean=0.917, loss_mean_cls=0.0894, proj_loss=-0.0964][2026-03-26 12:33:22] Step: 548, Training Logs: loss_final: 0.888676, loss_mean: 0.888027, proj_loss: -0.091121, loss_mean_cls: 0.091770, grad_norm: 1.037862 +Steps: 0%| | 549/1000000 [02:19<69:26:44, 4.00it/s, grad_norm=1.04, loss_final=0.889, loss_mean=0.888, loss_mean_cls=0.0918, proj_loss=-0.0911][2026-03-26 12:33:23] Step: 549, Training Logs: loss_final: 0.896954, loss_mean: 0.902165, proj_loss: -0.094101, loss_mean_cls: 0.088890, grad_norm: 0.745828 +Steps: 0%| | 550/1000000 [02:19<69:15:25, 4.01it/s, grad_norm=0.746, loss_final=0.897, loss_mean=0.902, loss_mean_cls=0.0889, proj_loss=-0.0941][2026-03-26 12:33:23] Step: 550, Training Logs: loss_final: 0.892719, loss_mean: 0.900676, proj_loss: -0.097272, loss_mean_cls: 0.089315, grad_norm: 0.562294 +Steps: 0%| | 551/1000000 [02:20<69:06:48, 4.02it/s, grad_norm=0.562, loss_final=0.893, loss_mean=0.901, loss_mean_cls=0.0893, proj_loss=-0.0973][2026-03-26 12:33:23] Step: 551, Training Logs: loss_final: 0.910249, loss_mean: 0.916176, proj_loss: -0.094792, loss_mean_cls: 0.088864, grad_norm: 0.786492 +Steps: 0%| | 552/1000000 [02:20<68:58:26, 4.03it/s, grad_norm=0.786, loss_final=0.91, loss_mean=0.916, loss_mean_cls=0.0889, proj_loss=-0.0948][2026-03-26 12:33:23] Step: 552, Training Logs: loss_final: 0.918950, loss_mean: 0.929198, proj_loss: -0.097575, loss_mean_cls: 0.087326, grad_norm: 0.592937 +Steps: 0%| | 553/1000000 [02:20<68:52:54, 4.03it/s, grad_norm=0.593, loss_final=0.919, loss_mean=0.929, loss_mean_cls=0.0873, proj_loss=-0.0976][2026-03-26 12:33:24] Step: 553, Training Logs: loss_final: 0.890236, loss_mean: 0.894189, proj_loss: -0.093753, loss_mean_cls: 0.089801, grad_norm: 0.849514 +Steps: 0%| | 554/1000000 [02:20<68:47:36, 4.04it/s, grad_norm=0.85, loss_final=0.89, loss_mean=0.894, loss_mean_cls=0.0898, proj_loss=-0.0938][2026-03-26 12:33:24] Step: 554, Training Logs: loss_final: 0.877132, loss_mean: 0.878932, proj_loss: -0.093230, loss_mean_cls: 0.091429, grad_norm: 0.867433 +Steps: 0%| | 555/1000000 [02:21<68:44:12, 4.04it/s, grad_norm=0.867, loss_final=0.877, loss_mean=0.879, loss_mean_cls=0.0914, proj_loss=-0.0932][2026-03-26 12:33:24] Step: 555, Training Logs: loss_final: 0.895866, loss_mean: 0.903981, proj_loss: -0.096651, loss_mean_cls: 0.088536, grad_norm: 1.116677 +Steps: 0%| | 556/1000000 [02:21<68:45:25, 4.04it/s, grad_norm=1.12, loss_final=0.896, loss_mean=0.904, loss_mean_cls=0.0885, proj_loss=-0.0967][2026-03-26 12:33:24] Step: 556, Training Logs: loss_final: 0.899860, loss_mean: 0.910805, proj_loss: -0.097702, loss_mean_cls: 0.086757, grad_norm: 0.833821 +Steps: 0%| | 557/1000000 [02:21<68:44:11, 4.04it/s, grad_norm=0.834, loss_final=0.9, loss_mean=0.911, loss_mean_cls=0.0868, proj_loss=-0.0977][2026-03-26 12:33:25] Step: 557, Training Logs: loss_final: 0.888301, loss_mean: 0.894486, proj_loss: -0.096854, loss_mean_cls: 0.090669, grad_norm: 1.064261 +Steps: 0%| | 558/1000000 [02:21<68:44:23, 4.04it/s, grad_norm=1.06, loss_final=0.888, loss_mean=0.894, loss_mean_cls=0.0907, proj_loss=-0.0969][2026-03-26 12:33:25] Step: 558, Training Logs: loss_final: 0.900214, loss_mean: 0.903697, proj_loss: -0.094190, loss_mean_cls: 0.090707, grad_norm: 0.944536 +Steps: 0%| | 559/1000000 [02:22<68:44:15, 4.04it/s, grad_norm=0.945, loss_final=0.9, loss_mean=0.904, loss_mean_cls=0.0907, proj_loss=-0.0942][2026-03-26 12:33:25] Step: 559, Training Logs: loss_final: 0.902075, loss_mean: 0.912433, proj_loss: -0.098479, loss_mean_cls: 0.088121, grad_norm: 0.739404 +Steps: 0%| | 560/1000000 [02:22<68:44:51, 4.04it/s, grad_norm=0.739, loss_final=0.902, loss_mean=0.912, loss_mean_cls=0.0881, proj_loss=-0.0985][2026-03-26 12:33:25] Step: 560, Training Logs: loss_final: 0.892931, loss_mean: 0.899194, proj_loss: -0.096705, loss_mean_cls: 0.090442, grad_norm: 0.823030 +Steps: 0%| | 561/1000000 [02:22<71:27:04, 3.89it/s, grad_norm=0.823, loss_final=0.893, loss_mean=0.899, loss_mean_cls=0.0904, proj_loss=-0.0967][2026-03-26 12:33:26] Step: 561, Training Logs: loss_final: 0.886981, loss_mean: 0.891642, proj_loss: -0.096111, loss_mean_cls: 0.091449, grad_norm: 0.579070 +Steps: 0%| | 562/1000000 [02:22<71:42:06, 3.87it/s, grad_norm=0.579, loss_final=0.887, loss_mean=0.892, loss_mean_cls=0.0914, proj_loss=-0.0961][2026-03-26 12:33:26] Step: 562, Training Logs: loss_final: 0.909065, loss_mean: 0.910941, proj_loss: -0.092692, loss_mean_cls: 0.090815, grad_norm: 0.747384 +Steps: 0%| | 563/1000000 [02:23<70:47:13, 3.92it/s, grad_norm=0.747, loss_final=0.909, loss_mean=0.911, loss_mean_cls=0.0908, proj_loss=-0.0927][2026-03-26 12:33:26] Step: 563, Training Logs: loss_final: 0.890010, loss_mean: 0.896423, proj_loss: -0.095623, loss_mean_cls: 0.089209, grad_norm: 0.632657 +Steps: 0%| | 564/1000000 [02:23<70:10:17, 3.96it/s, grad_norm=0.633, loss_final=0.89, loss_mean=0.896, loss_mean_cls=0.0892, proj_loss=-0.0956][2026-03-26 12:33:26] Step: 564, Training Logs: loss_final: 0.894659, loss_mean: 0.901562, proj_loss: -0.096769, loss_mean_cls: 0.089866, grad_norm: 0.769940 +Steps: 0%| | 565/1000000 [02:23<69:45:15, 3.98it/s, grad_norm=0.77, loss_final=0.895, loss_mean=0.902, loss_mean_cls=0.0899, proj_loss=-0.0968][2026-03-26 12:33:27] Step: 565, Training Logs: loss_final: 0.892419, loss_mean: 0.902339, proj_loss: -0.100117, loss_mean_cls: 0.090198, grad_norm: 1.032249 +Steps: 0%| | 566/1000000 [02:23<69:24:57, 4.00it/s, grad_norm=1.03, loss_final=0.892, loss_mean=0.902, loss_mean_cls=0.0902, proj_loss=-0.1][2026-03-26 12:33:27] Step: 566, Training Logs: loss_final: 0.893375, loss_mean: 0.902781, proj_loss: -0.100365, loss_mean_cls: 0.090959, grad_norm: 0.894692 +Steps: 0%| | 567/1000000 [02:24<69:12:09, 4.01it/s, grad_norm=0.895, loss_final=0.893, loss_mean=0.903, loss_mean_cls=0.091, proj_loss=-0.1][2026-03-26 12:33:27] Step: 567, Training Logs: loss_final: 0.874369, loss_mean: 0.876271, proj_loss: -0.094561, loss_mean_cls: 0.092659, grad_norm: 0.699464 +Steps: 0%| | 568/1000000 [02:24<69:06:06, 4.02it/s, grad_norm=0.699, loss_final=0.874, loss_mean=0.876, loss_mean_cls=0.0927, proj_loss=-0.0946][2026-03-26 12:33:27] Step: 568, Training Logs: loss_final: 0.907189, loss_mean: 0.913125, proj_loss: -0.095428, loss_mean_cls: 0.089492, grad_norm: 0.637123 +Steps: 0%| | 569/1000000 [02:24<69:00:19, 4.02it/s, grad_norm=0.637, loss_final=0.907, loss_mean=0.913, loss_mean_cls=0.0895, proj_loss=-0.0954][2026-03-26 12:33:28] Step: 569, Training Logs: loss_final: 0.876461, loss_mean: 0.883109, proj_loss: -0.097781, loss_mean_cls: 0.091133, grad_norm: 0.913654 +Steps: 0%| | 570/1000000 [02:24<68:53:19, 4.03it/s, grad_norm=0.914, loss_final=0.876, loss_mean=0.883, loss_mean_cls=0.0911, proj_loss=-0.0978][2026-03-26 12:33:28] Step: 570, Training Logs: loss_final: 0.891301, loss_mean: 0.900280, proj_loss: -0.098101, loss_mean_cls: 0.089122, grad_norm: 1.001783 +Steps: 0%| | 571/1000000 [02:25<68:48:12, 4.03it/s, grad_norm=1, loss_final=0.891, loss_mean=0.9, loss_mean_cls=0.0891, proj_loss=-0.0981][2026-03-26 12:33:28] Step: 571, Training Logs: loss_final: 0.912156, loss_mean: 0.925139, proj_loss: -0.099106, loss_mean_cls: 0.086122, grad_norm: 0.972801 +Steps: 0%| | 572/1000000 [02:25<68:46:16, 4.04it/s, grad_norm=0.973, loss_final=0.912, loss_mean=0.925, loss_mean_cls=0.0861, proj_loss=-0.0991][2026-03-26 12:33:28] Step: 572, Training Logs: loss_final: 0.887314, loss_mean: 0.898383, proj_loss: -0.100680, loss_mean_cls: 0.089611, grad_norm: 1.222481 +Steps: 0%| | 573/1000000 [02:25<68:49:02, 4.03it/s, grad_norm=1.22, loss_final=0.887, loss_mean=0.898, loss_mean_cls=0.0896, proj_loss=-0.101][2026-03-26 12:33:29] Step: 573, Training Logs: loss_final: 0.876574, loss_mean: 0.881761, proj_loss: -0.097881, loss_mean_cls: 0.092694, grad_norm: 0.863602 +Steps: 0%| | 574/1000000 [02:25<68:45:37, 4.04it/s, grad_norm=0.864, loss_final=0.877, loss_mean=0.882, loss_mean_cls=0.0927, proj_loss=-0.0979][2026-03-26 12:33:29] Step: 574, Training Logs: loss_final: 0.876279, loss_mean: 0.885270, proj_loss: -0.100191, loss_mean_cls: 0.091200, grad_norm: 1.111458 +Steps: 0%| | 575/1000000 [02:26<68:45:42, 4.04it/s, grad_norm=1.11, loss_final=0.876, loss_mean=0.885, loss_mean_cls=0.0912, proj_loss=-0.1][2026-03-26 12:33:29] Step: 575, Training Logs: loss_final: 0.883809, loss_mean: 0.895448, proj_loss: -0.101923, loss_mean_cls: 0.090284, grad_norm: 0.852407 +Steps: 0%| | 576/1000000 [02:26<68:46:08, 4.04it/s, grad_norm=0.852, loss_final=0.884, loss_mean=0.895, loss_mean_cls=0.0903, proj_loss=-0.102][2026-03-26 12:33:29] Step: 576, Training Logs: loss_final: 0.914021, loss_mean: 0.927472, proj_loss: -0.100613, loss_mean_cls: 0.087162, grad_norm: 1.405070 +Steps: 0%| | 577/1000000 [02:26<68:46:59, 4.04it/s, grad_norm=1.41, loss_final=0.914, loss_mean=0.927, loss_mean_cls=0.0872, proj_loss=-0.101][2026-03-26 12:33:30] Step: 577, Training Logs: loss_final: 0.877570, loss_mean: 0.896372, proj_loss: -0.106791, loss_mean_cls: 0.087989, grad_norm: 1.290731 +Steps: 0%| | 578/1000000 [02:26<68:47:01, 4.04it/s, grad_norm=1.29, loss_final=0.878, loss_mean=0.896, loss_mean_cls=0.088, proj_loss=-0.107][2026-03-26 12:33:30] Step: 578, Training Logs: loss_final: 0.871310, loss_mean: 0.882087, proj_loss: -0.101322, loss_mean_cls: 0.090544, grad_norm: 0.597123 +Steps: 0%| | 579/1000000 [02:27<68:43:52, 4.04it/s, grad_norm=0.597, loss_final=0.871, loss_mean=0.882, loss_mean_cls=0.0905, proj_loss=-0.101][2026-03-26 12:33:30] Step: 579, Training Logs: loss_final: 0.885583, loss_mean: 0.900469, proj_loss: -0.105077, loss_mean_cls: 0.090191, grad_norm: 1.476694 +Steps: 0%| | 580/1000000 [02:27<68:43:20, 4.04it/s, grad_norm=1.48, loss_final=0.886, loss_mean=0.9, loss_mean_cls=0.0902, proj_loss=-0.105][2026-03-26 12:33:30] Step: 580, Training Logs: loss_final: 0.889808, loss_mean: 0.903566, proj_loss: -0.103969, loss_mean_cls: 0.090212, grad_norm: 0.910858 +Steps: 0%| | 581/1000000 [02:27<68:40:44, 4.04it/s, grad_norm=0.911, loss_final=0.89, loss_mean=0.904, loss_mean_cls=0.0902, proj_loss=-0.104][2026-03-26 12:33:31] Step: 581, Training Logs: loss_final: 0.883499, loss_mean: 0.895426, proj_loss: -0.101365, loss_mean_cls: 0.089439, grad_norm: 1.047186 +Steps: 0%| | 582/1000000 [02:27<68:39:04, 4.04it/s, grad_norm=1.05, loss_final=0.883, loss_mean=0.895, loss_mean_cls=0.0894, proj_loss=-0.101][2026-03-26 12:33:31] Step: 582, Training Logs: loss_final: 0.889335, loss_mean: 0.907771, proj_loss: -0.105887, loss_mean_cls: 0.087452, grad_norm: 0.819693 +Steps: 0%| | 583/1000000 [02:28<68:36:39, 4.05it/s, grad_norm=0.82, loss_final=0.889, loss_mean=0.908, loss_mean_cls=0.0875, proj_loss=-0.106][2026-03-26 12:33:31] Step: 583, Training Logs: loss_final: 0.898732, loss_mean: 0.914710, proj_loss: -0.104244, loss_mean_cls: 0.088267, grad_norm: 0.758291 +Steps: 0%| | 584/1000000 [02:28<68:43:30, 4.04it/s, grad_norm=0.758, loss_final=0.899, loss_mean=0.915, loss_mean_cls=0.0883, proj_loss=-0.104][2026-03-26 12:33:31] Step: 584, Training Logs: loss_final: 0.865719, loss_mean: 0.878335, proj_loss: -0.102180, loss_mean_cls: 0.089565, grad_norm: 0.673079 +Steps: 0%| | 585/1000000 [02:28<68:42:17, 4.04it/s, grad_norm=0.673, loss_final=0.866, loss_mean=0.878, loss_mean_cls=0.0896, proj_loss=-0.102][2026-03-26 12:33:31] Step: 585, Training Logs: loss_final: 0.888314, loss_mean: 0.906142, proj_loss: -0.106680, loss_mean_cls: 0.088853, grad_norm: 0.874638 +Steps: 0%| | 586/1000000 [02:28<68:45:03, 4.04it/s, grad_norm=0.875, loss_final=0.888, loss_mean=0.906, loss_mean_cls=0.0889, proj_loss=-0.107][2026-03-26 12:33:32] Step: 586, Training Logs: loss_final: 0.900502, loss_mean: 0.916792, proj_loss: -0.104798, loss_mean_cls: 0.088508, grad_norm: 1.001004 +Steps: 0%| | 587/1000000 [02:29<68:44:47, 4.04it/s, grad_norm=1, loss_final=0.901, loss_mean=0.917, loss_mean_cls=0.0885, proj_loss=-0.105][2026-03-26 12:33:32] Step: 587, Training Logs: loss_final: 0.884866, loss_mean: 0.899851, proj_loss: -0.105184, loss_mean_cls: 0.090199, grad_norm: 1.015693 +Steps: 0%| | 588/1000000 [02:29<68:48:13, 4.03it/s, grad_norm=1.02, loss_final=0.885, loss_mean=0.9, loss_mean_cls=0.0902, proj_loss=-0.105][2026-03-26 12:33:32] Step: 588, Training Logs: loss_final: 0.896361, loss_mean: 0.915406, proj_loss: -0.107832, loss_mean_cls: 0.088787, grad_norm: 0.514327 +Steps: 0%| | 589/1000000 [02:29<68:44:43, 4.04it/s, grad_norm=0.514, loss_final=0.896, loss_mean=0.915, loss_mean_cls=0.0888, proj_loss=-0.108][2026-03-26 12:33:32] Step: 589, Training Logs: loss_final: 0.868655, loss_mean: 0.883067, proj_loss: -0.105779, loss_mean_cls: 0.091366, grad_norm: 0.611275 +Steps: 0%| | 590/1000000 [02:29<68:44:56, 4.04it/s, grad_norm=0.611, loss_final=0.869, loss_mean=0.883, loss_mean_cls=0.0914, proj_loss=-0.106][2026-03-26 12:33:33] Step: 590, Training Logs: loss_final: 0.871835, loss_mean: 0.889763, proj_loss: -0.108172, loss_mean_cls: 0.090245, grad_norm: 0.940646 +Steps: 0%| | 591/1000000 [02:30<68:42:34, 4.04it/s, grad_norm=0.941, loss_final=0.872, loss_mean=0.89, loss_mean_cls=0.0902, proj_loss=-0.108][2026-03-26 12:33:33] Step: 591, Training Logs: loss_final: 0.877626, loss_mean: 0.901109, proj_loss: -0.110864, loss_mean_cls: 0.087381, grad_norm: 0.837806 +Steps: 0%| | 592/1000000 [02:30<68:40:46, 4.04it/s, grad_norm=0.838, loss_final=0.878, loss_mean=0.901, loss_mean_cls=0.0874, proj_loss=-0.111][2026-03-26 12:33:33] Step: 592, Training Logs: loss_final: 0.874102, loss_mean: 0.894214, proj_loss: -0.109189, loss_mean_cls: 0.089076, grad_norm: 0.989442 +Steps: 0%| | 593/1000000 [02:30<68:39:57, 4.04it/s, grad_norm=0.989, loss_final=0.874, loss_mean=0.894, loss_mean_cls=0.0891, proj_loss=-0.109][2026-03-26 12:33:33] Step: 593, Training Logs: loss_final: 0.891992, loss_mean: 0.914107, proj_loss: -0.110574, loss_mean_cls: 0.088459, grad_norm: 1.527584 +Steps: 0%| | 594/1000000 [02:30<68:59:16, 4.02it/s, grad_norm=1.53, loss_final=0.892, loss_mean=0.914, loss_mean_cls=0.0885, proj_loss=-0.111][2026-03-26 12:33:34] Step: 594, Training Logs: loss_final: 0.869256, loss_mean: 0.885622, proj_loss: -0.107636, loss_mean_cls: 0.091271, grad_norm: 0.827941 +Steps: 0%| | 595/1000000 [02:31<68:54:56, 4.03it/s, grad_norm=0.828, loss_final=0.869, loss_mean=0.886, loss_mean_cls=0.0913, proj_loss=-0.108][2026-03-26 12:33:34] Step: 595, Training Logs: loss_final: 0.878392, loss_mean: 0.901379, proj_loss: -0.112255, loss_mean_cls: 0.089268, grad_norm: 1.280934 +Steps: 0%| | 596/1000000 [02:31<68:56:02, 4.03it/s, grad_norm=1.28, loss_final=0.878, loss_mean=0.901, loss_mean_cls=0.0893, proj_loss=-0.112][2026-03-26 12:33:34] Step: 596, Training Logs: loss_final: 0.896026, loss_mean: 0.920742, proj_loss: -0.112726, loss_mean_cls: 0.088010, grad_norm: 0.966416 +Steps: 0%| | 597/1000000 [02:31<68:52:38, 4.03it/s, grad_norm=0.966, loss_final=0.896, loss_mean=0.921, loss_mean_cls=0.088, proj_loss=-0.113][2026-03-26 12:33:34] Step: 597, Training Logs: loss_final: 0.890151, loss_mean: 0.912367, proj_loss: -0.111121, loss_mean_cls: 0.088905, grad_norm: 0.965084 +Steps: 0%| | 598/1000000 [02:31<68:49:02, 4.03it/s, grad_norm=0.965, loss_final=0.89, loss_mean=0.912, loss_mean_cls=0.0889, proj_loss=-0.111][2026-03-26 12:33:35] Step: 598, Training Logs: loss_final: 0.899819, loss_mean: 0.920064, proj_loss: -0.107695, loss_mean_cls: 0.087449, grad_norm: 1.170464 +Steps: 0%| | 599/1000000 [02:32<68:46:52, 4.04it/s, grad_norm=1.17, loss_final=0.9, loss_mean=0.92, loss_mean_cls=0.0874, proj_loss=-0.108][2026-03-26 12:33:35] Step: 599, Training Logs: loss_final: 0.861945, loss_mean: 0.882009, proj_loss: -0.111101, loss_mean_cls: 0.091037, grad_norm: 0.769068 +Steps: 0%| | 600/1000000 [02:32<68:48:11, 4.03it/s, grad_norm=0.769, loss_final=0.862, loss_mean=0.882, loss_mean_cls=0.091, proj_loss=-0.111][2026-03-26 12:33:35] Step: 600, Training Logs: loss_final: 0.892841, loss_mean: 0.918575, proj_loss: -0.114296, loss_mean_cls: 0.088562, grad_norm: 1.311424 +Steps: 0%| | 601/1000000 [02:32<68:47:24, 4.04it/s, grad_norm=1.31, loss_final=0.893, loss_mean=0.919, loss_mean_cls=0.0886, proj_loss=-0.114][2026-03-26 12:33:35] Step: 601, Training Logs: loss_final: 0.865575, loss_mean: 0.887724, proj_loss: -0.111545, loss_mean_cls: 0.089397, grad_norm: 0.879056 +Steps: 0%| | 602/1000000 [02:32<68:45:11, 4.04it/s, grad_norm=0.879, loss_final=0.866, loss_mean=0.888, loss_mean_cls=0.0894, proj_loss=-0.112][2026-03-26 12:33:36] Step: 602, Training Logs: loss_final: 0.885343, loss_mean: 0.911404, proj_loss: -0.114237, loss_mean_cls: 0.088176, grad_norm: 0.714984 +Steps: 0%| | 603/1000000 [02:33<68:47:02, 4.04it/s, grad_norm=0.715, loss_final=0.885, loss_mean=0.911, loss_mean_cls=0.0882, proj_loss=-0.114][2026-03-26 12:33:36] Step: 603, Training Logs: loss_final: 0.879609, loss_mean: 0.900916, proj_loss: -0.112370, loss_mean_cls: 0.091063, grad_norm: 1.107264 +Steps: 0%| | 604/1000000 [02:33<68:44:09, 4.04it/s, grad_norm=1.11, loss_final=0.88, loss_mean=0.901, loss_mean_cls=0.0911, proj_loss=-0.112][2026-03-26 12:33:36] Step: 604, Training Logs: loss_final: 0.876829, loss_mean: 0.907640, proj_loss: -0.119562, loss_mean_cls: 0.088751, grad_norm: 0.536574 +Steps: 0%| | 605/1000000 [02:33<68:44:16, 4.04it/s, grad_norm=0.537, loss_final=0.877, loss_mean=0.908, loss_mean_cls=0.0888, proj_loss=-0.12][2026-03-26 12:33:36] Step: 605, Training Logs: loss_final: 0.860769, loss_mean: 0.890741, proj_loss: -0.119996, loss_mean_cls: 0.090023, grad_norm: 0.814256 +Steps: 0%| | 606/1000000 [02:33<68:41:41, 4.04it/s, grad_norm=0.814, loss_final=0.861, loss_mean=0.891, loss_mean_cls=0.09, proj_loss=-0.12][2026-03-26 12:33:37] Step: 606, Training Logs: loss_final: 0.879975, loss_mean: 0.911479, proj_loss: -0.118752, loss_mean_cls: 0.087248, grad_norm: 0.858224 +Steps: 0%| | 607/1000000 [02:34<68:41:32, 4.04it/s, grad_norm=0.858, loss_final=0.88, loss_mean=0.911, loss_mean_cls=0.0872, proj_loss=-0.119][2026-03-26 12:33:37] Step: 607, Training Logs: loss_final: 0.857142, loss_mean: 0.885409, proj_loss: -0.118584, loss_mean_cls: 0.090318, grad_norm: 0.888430 +Steps: 0%| | 608/1000000 [02:34<68:41:29, 4.04it/s, grad_norm=0.888, loss_final=0.857, loss_mean=0.885, loss_mean_cls=0.0903, proj_loss=-0.119][2026-03-26 12:33:37] Step: 608, Training Logs: loss_final: 0.864777, loss_mean: 0.896466, proj_loss: -0.121053, loss_mean_cls: 0.089363, grad_norm: 0.691679 +Steps: 0%| | 609/1000000 [02:34<68:40:27, 4.04it/s, grad_norm=0.692, loss_final=0.865, loss_mean=0.896, loss_mean_cls=0.0894, proj_loss=-0.121][2026-03-26 12:33:37] Step: 609, Training Logs: loss_final: 0.862708, loss_mean: 0.893319, proj_loss: -0.120763, loss_mean_cls: 0.090152, grad_norm: 1.087572 +Steps: 0%| | 610/1000000 [02:34<68:39:43, 4.04it/s, grad_norm=1.09, loss_final=0.863, loss_mean=0.893, loss_mean_cls=0.0902, proj_loss=-0.121][2026-03-26 12:33:38] Step: 610, Training Logs: loss_final: 0.870039, loss_mean: 0.906669, proj_loss: -0.124944, loss_mean_cls: 0.088315, grad_norm: 0.670060 +Steps: 0%| | 611/1000000 [02:35<68:41:06, 4.04it/s, grad_norm=0.67, loss_final=0.87, loss_mean=0.907, loss_mean_cls=0.0883, proj_loss=-0.125][2026-03-26 12:33:38] Step: 611, Training Logs: loss_final: 0.858142, loss_mean: 0.894759, proj_loss: -0.125486, loss_mean_cls: 0.088869, grad_norm: 1.070298 +Steps: 0%| | 612/1000000 [02:35<68:42:15, 4.04it/s, grad_norm=1.07, loss_final=0.858, loss_mean=0.895, loss_mean_cls=0.0889, proj_loss=-0.125][2026-03-26 12:33:38] Step: 612, Training Logs: loss_final: 0.872394, loss_mean: 0.912928, proj_loss: -0.129447, loss_mean_cls: 0.088913, grad_norm: 0.893775 +Steps: 0%| | 613/1000000 [02:35<68:40:35, 4.04it/s, grad_norm=0.894, loss_final=0.872, loss_mean=0.913, loss_mean_cls=0.0889, proj_loss=-0.129][2026-03-26 12:33:38] Step: 613, Training Logs: loss_final: 0.874569, loss_mean: 0.913242, proj_loss: -0.126610, loss_mean_cls: 0.087937, grad_norm: 0.863737 +Steps: 0%| | 614/1000000 [02:35<68:40:54, 4.04it/s, grad_norm=0.864, loss_final=0.875, loss_mean=0.913, loss_mean_cls=0.0879, proj_loss=-0.127][2026-03-26 12:33:39] Step: 614, Training Logs: loss_final: 0.836454, loss_mean: 0.872634, proj_loss: -0.127767, loss_mean_cls: 0.091587, grad_norm: 0.814739 +Steps: 0%| | 615/1000000 [02:36<68:43:29, 4.04it/s, grad_norm=0.815, loss_final=0.836, loss_mean=0.873, loss_mean_cls=0.0916, proj_loss=-0.128][2026-03-26 12:33:39] Step: 615, Training Logs: loss_final: 0.854416, loss_mean: 0.895514, proj_loss: -0.130650, loss_mean_cls: 0.089552, grad_norm: 0.753335 +Steps: 0%| | 616/1000000 [02:36<68:38:47, 4.04it/s, grad_norm=0.753, loss_final=0.854, loss_mean=0.896, loss_mean_cls=0.0896, proj_loss=-0.131][2026-03-26 12:33:39] Step: 616, Training Logs: loss_final: 0.868590, loss_mean: 0.912710, proj_loss: -0.131313, loss_mean_cls: 0.087193, grad_norm: 0.824301 +Steps: 0%| | 617/1000000 [02:36<68:38:59, 4.04it/s, grad_norm=0.824, loss_final=0.869, loss_mean=0.913, loss_mean_cls=0.0872, proj_loss=-0.131][2026-03-26 12:33:39] Step: 617, Training Logs: loss_final: 0.855932, loss_mean: 0.891981, proj_loss: -0.126231, loss_mean_cls: 0.090182, grad_norm: 1.053235 +Steps: 0%| | 618/1000000 [02:36<68:39:18, 4.04it/s, grad_norm=1.05, loss_final=0.856, loss_mean=0.892, loss_mean_cls=0.0902, proj_loss=-0.126][2026-03-26 12:33:40] Step: 618, Training Logs: loss_final: 0.846795, loss_mean: 0.890365, proj_loss: -0.133459, loss_mean_cls: 0.089889, grad_norm: 0.547888 +Steps: 0%| | 619/1000000 [02:37<68:40:02, 4.04it/s, grad_norm=0.548, loss_final=0.847, loss_mean=0.89, loss_mean_cls=0.0899, proj_loss=-0.133][2026-03-26 12:33:40] Step: 619, Training Logs: loss_final: 0.851004, loss_mean: 0.893173, proj_loss: -0.132700, loss_mean_cls: 0.090531, grad_norm: 0.931155 +Steps: 0%| | 620/1000000 [02:37<68:37:49, 4.04it/s, grad_norm=0.931, loss_final=0.851, loss_mean=0.893, loss_mean_cls=0.0905, proj_loss=-0.133][2026-03-26 12:33:40] Step: 620, Training Logs: loss_final: 0.851620, loss_mean: 0.893776, proj_loss: -0.132077, loss_mean_cls: 0.089922, grad_norm: 1.007702 +Steps: 0%| | 621/1000000 [02:37<68:37:09, 4.05it/s, grad_norm=1.01, loss_final=0.852, loss_mean=0.894, loss_mean_cls=0.0899, proj_loss=-0.132][2026-03-26 12:33:40] Step: 621, Training Logs: loss_final: 0.850058, loss_mean: 0.892616, proj_loss: -0.133321, loss_mean_cls: 0.090763, grad_norm: 0.754161 +Steps: 0%| | 622/1000000 [02:37<68:36:27, 4.05it/s, grad_norm=0.754, loss_final=0.85, loss_mean=0.893, loss_mean_cls=0.0908, proj_loss=-0.133][2026-03-26 12:33:41] Step: 622, Training Logs: loss_final: 0.864470, loss_mean: 0.911066, proj_loss: -0.136236, loss_mean_cls: 0.089641, grad_norm: 0.881606 +Steps: 0%| | 623/1000000 [02:38<68:37:58, 4.04it/s, grad_norm=0.882, loss_final=0.864, loss_mean=0.911, loss_mean_cls=0.0896, proj_loss=-0.136][2026-03-26 12:33:41] Step: 623, Training Logs: loss_final: 0.843654, loss_mean: 0.893944, proj_loss: -0.140236, loss_mean_cls: 0.089946, grad_norm: 1.004259 +Steps: 0%| | 624/1000000 [02:38<68:39:11, 4.04it/s, grad_norm=1, loss_final=0.844, loss_mean=0.894, loss_mean_cls=0.0899, proj_loss=-0.14][2026-03-26 12:33:41] Step: 624, Training Logs: loss_final: 0.841439, loss_mean: 0.888836, proj_loss: -0.137544, loss_mean_cls: 0.090147, grad_norm: 0.917068 +Steps: 0%| | 625/1000000 [02:38<68:45:13, 4.04it/s, grad_norm=0.917, loss_final=0.841, loss_mean=0.889, loss_mean_cls=0.0901, proj_loss=-0.138][2026-03-26 12:33:41] Step: 625, Training Logs: loss_final: 0.851531, loss_mean: 0.907453, proj_loss: -0.144776, loss_mean_cls: 0.088854, grad_norm: 0.551723 +Steps: 0%| | 626/1000000 [02:38<68:46:58, 4.04it/s, grad_norm=0.552, loss_final=0.852, loss_mean=0.907, loss_mean_cls=0.0889, proj_loss=-0.145][2026-03-26 12:33:42] Step: 626, Training Logs: loss_final: 0.822615, loss_mean: 0.874171, proj_loss: -0.141638, loss_mean_cls: 0.090082, grad_norm: 0.815918 +Steps: 0%| | 627/1000000 [02:39<68:45:12, 4.04it/s, grad_norm=0.816, loss_final=0.823, loss_mean=0.874, loss_mean_cls=0.0901, proj_loss=-0.142][2026-03-26 12:33:42] Step: 627, Training Logs: loss_final: 0.839018, loss_mean: 0.897245, proj_loss: -0.147380, loss_mean_cls: 0.089153, grad_norm: 1.053321 +Steps: 0%| | 628/1000000 [02:39<68:45:14, 4.04it/s, grad_norm=1.05, loss_final=0.839, loss_mean=0.897, loss_mean_cls=0.0892, proj_loss=-0.147][2026-03-26 12:33:42] Step: 628, Training Logs: loss_final: 0.836670, loss_mean: 0.898210, proj_loss: -0.149749, loss_mean_cls: 0.088209, grad_norm: 0.701293 +Steps: 0%| | 629/1000000 [02:39<68:46:41, 4.04it/s, grad_norm=0.701, loss_final=0.837, loss_mean=0.898, loss_mean_cls=0.0882, proj_loss=-0.15][2026-03-26 12:33:42] Step: 629, Training Logs: loss_final: 0.840627, loss_mean: 0.900154, proj_loss: -0.148439, loss_mean_cls: 0.088912, grad_norm: 0.832492 +Steps: 0%| | 630/1000000 [02:39<68:46:06, 4.04it/s, grad_norm=0.832, loss_final=0.841, loss_mean=0.9, loss_mean_cls=0.0889, proj_loss=-0.148][2026-03-26 12:33:43] Step: 630, Training Logs: loss_final: 0.853539, loss_mean: 0.910656, proj_loss: -0.146087, loss_mean_cls: 0.088969, grad_norm: 0.835975 +Steps: 0%| | 631/1000000 [02:39<68:42:57, 4.04it/s, grad_norm=0.836, loss_final=0.854, loss_mean=0.911, loss_mean_cls=0.089, proj_loss=-0.146][2026-03-26 12:33:43] Step: 631, Training Logs: loss_final: 0.833950, loss_mean: 0.898628, proj_loss: -0.152347, loss_mean_cls: 0.087669, grad_norm: 0.723172 +Steps: 0%| | 632/1000000 [02:40<68:42:32, 4.04it/s, grad_norm=0.723, loss_final=0.834, loss_mean=0.899, loss_mean_cls=0.0877, proj_loss=-0.152][2026-03-26 12:33:43] Step: 632, Training Logs: loss_final: 0.834987, loss_mean: 0.894666, proj_loss: -0.150119, loss_mean_cls: 0.090440, grad_norm: 1.075904 +Steps: 0%| | 633/1000000 [02:40<68:35:58, 4.05it/s, grad_norm=1.08, loss_final=0.835, loss_mean=0.895, loss_mean_cls=0.0904, proj_loss=-0.15][2026-03-26 12:33:43] Step: 633, Training Logs: loss_final: 0.826757, loss_mean: 0.883433, proj_loss: -0.148084, loss_mean_cls: 0.091409, grad_norm: 0.575704 +Steps: 0%| | 634/1000000 [02:40<68:39:02, 4.04it/s, grad_norm=0.576, loss_final=0.827, loss_mean=0.883, loss_mean_cls=0.0914, proj_loss=-0.148][2026-03-26 12:33:44] Step: 634, Training Logs: loss_final: 0.798787, loss_mean: 0.867203, proj_loss: -0.158204, loss_mean_cls: 0.089788, grad_norm: 0.869029 +Steps: 0%| | 635/1000000 [02:40<68:39:43, 4.04it/s, grad_norm=0.869, loss_final=0.799, loss_mean=0.867, loss_mean_cls=0.0898, proj_loss=-0.158][2026-03-26 12:33:44] Step: 635, Training Logs: loss_final: 0.832127, loss_mean: 0.900009, proj_loss: -0.156154, loss_mean_cls: 0.088271, grad_norm: 0.676001 +Steps: 0%| | 636/1000000 [02:41<68:38:04, 4.04it/s, grad_norm=0.676, loss_final=0.832, loss_mean=0.9, loss_mean_cls=0.0883, proj_loss=-0.156][2026-03-26 12:33:44] Step: 636, Training Logs: loss_final: 0.835742, loss_mean: 0.908679, proj_loss: -0.160068, loss_mean_cls: 0.087131, grad_norm: 0.981709 +Steps: 0%| | 637/1000000 [02:41<68:38:14, 4.04it/s, grad_norm=0.982, loss_final=0.836, loss_mean=0.909, loss_mean_cls=0.0871, proj_loss=-0.16][2026-03-26 12:33:44] Step: 637, Training Logs: loss_final: 0.825965, loss_mean: 0.905401, proj_loss: -0.167048, loss_mean_cls: 0.087612, grad_norm: 0.561994 +Steps: 0%| | 638/1000000 [02:41<68:38:27, 4.04it/s, grad_norm=0.562, loss_final=0.826, loss_mean=0.905, loss_mean_cls=0.0876, proj_loss=-0.167][2026-03-26 12:33:45] Step: 638, Training Logs: loss_final: 0.854967, loss_mean: 0.928295, proj_loss: -0.161191, loss_mean_cls: 0.087863, grad_norm: 1.094758 +Steps: 0%| | 639/1000000 [02:41<69:33:48, 3.99it/s, grad_norm=1.09, loss_final=0.855, loss_mean=0.928, loss_mean_cls=0.0879, proj_loss=-0.161][2026-03-26 12:33:45] Step: 639, Training Logs: loss_final: 0.798887, loss_mean: 0.868829, proj_loss: -0.161516, loss_mean_cls: 0.091573, grad_norm: 0.462063 +Steps: 0%| | 640/1000000 [02:42<69:22:40, 4.00it/s, grad_norm=0.462, loss_final=0.799, loss_mean=0.869, loss_mean_cls=0.0916, proj_loss=-0.162][2026-03-26 12:33:45] Step: 640, Training Logs: loss_final: 0.825588, loss_mean: 0.906968, proj_loss: -0.169518, loss_mean_cls: 0.088138, grad_norm: 1.053439 +Steps: 0%| | 641/1000000 [02:42<69:07:22, 4.02it/s, grad_norm=1.05, loss_final=0.826, loss_mean=0.907, loss_mean_cls=0.0881, proj_loss=-0.17][2026-03-26 12:33:45] Step: 641, Training Logs: loss_final: 0.825577, loss_mean: 0.897894, proj_loss: -0.162084, loss_mean_cls: 0.089767, grad_norm: 1.107048 +Steps: 0%| | 642/1000000 [02:42<68:57:30, 4.03it/s, grad_norm=1.11, loss_final=0.826, loss_mean=0.898, loss_mean_cls=0.0898, proj_loss=-0.162][2026-03-26 12:33:46] Step: 642, Training Logs: loss_final: 0.831099, loss_mean: 0.924269, proj_loss: -0.179797, loss_mean_cls: 0.086627, grad_norm: 0.970554 +Steps: 0%| | 643/1000000 [02:42<68:51:23, 4.03it/s, grad_norm=0.971, loss_final=0.831, loss_mean=0.924, loss_mean_cls=0.0866, proj_loss=-0.18][2026-03-26 12:33:46] Step: 643, Training Logs: loss_final: 0.840786, loss_mean: 0.926975, proj_loss: -0.173064, loss_mean_cls: 0.086876, grad_norm: 0.545398 +Steps: 0%| | 644/1000000 [02:43<68:47:29, 4.04it/s, grad_norm=0.545, loss_final=0.841, loss_mean=0.927, loss_mean_cls=0.0869, proj_loss=-0.173][2026-03-26 12:33:46] Step: 644, Training Logs: loss_final: 0.820760, loss_mean: 0.909705, proj_loss: -0.176707, loss_mean_cls: 0.087762, grad_norm: 0.864666 +Steps: 0%| | 645/1000000 [02:43<68:45:24, 4.04it/s, grad_norm=0.865, loss_final=0.821, loss_mean=0.91, loss_mean_cls=0.0878, proj_loss=-0.177][2026-03-26 12:33:46] Step: 645, Training Logs: loss_final: 0.820022, loss_mean: 0.909487, proj_loss: -0.177026, loss_mean_cls: 0.087562, grad_norm: 0.630639 +Steps: 0%| | 646/1000000 [02:43<68:42:29, 4.04it/s, grad_norm=0.631, loss_final=0.82, loss_mean=0.909, loss_mean_cls=0.0876, proj_loss=-0.177][2026-03-26 12:33:47] Step: 646, Training Logs: loss_final: 0.808147, loss_mean: 0.903496, proj_loss: -0.182946, loss_mean_cls: 0.087597, grad_norm: 0.743004 +Steps: 0%| | 647/1000000 [02:43<68:42:01, 4.04it/s, grad_norm=0.743, loss_final=0.808, loss_mean=0.903, loss_mean_cls=0.0876, proj_loss=-0.183][2026-03-26 12:33:47] Step: 647, Training Logs: loss_final: 0.821601, loss_mean: 0.917308, proj_loss: -0.182944, loss_mean_cls: 0.087237, grad_norm: 0.613329 +Steps: 0%| | 648/1000000 [02:44<68:40:32, 4.04it/s, grad_norm=0.613, loss_final=0.822, loss_mean=0.917, loss_mean_cls=0.0872, proj_loss=-0.183][2026-03-26 12:33:47] Step: 648, Training Logs: loss_final: 0.808353, loss_mean: 0.891364, proj_loss: -0.173425, loss_mean_cls: 0.090414, grad_norm: 0.779334 +Steps: 0%| | 649/1000000 [02:44<68:39:43, 4.04it/s, grad_norm=0.779, loss_final=0.808, loss_mean=0.891, loss_mean_cls=0.0904, proj_loss=-0.173][2026-03-26 12:33:47] Step: 649, Training Logs: loss_final: 0.811516, loss_mean: 0.906626, proj_loss: -0.182979, loss_mean_cls: 0.087869, grad_norm: 1.333471 +Steps: 0%| | 650/1000000 [02:44<68:40:40, 4.04it/s, grad_norm=1.33, loss_final=0.812, loss_mean=0.907, loss_mean_cls=0.0879, proj_loss=-0.183][2026-03-26 12:33:48] Step: 650, Training Logs: loss_final: 0.809150, loss_mean: 0.908248, proj_loss: -0.186424, loss_mean_cls: 0.087326, grad_norm: 0.595754 +Steps: 0%| | 651/1000000 [02:44<68:42:25, 4.04it/s, grad_norm=0.596, loss_final=0.809, loss_mean=0.908, loss_mean_cls=0.0873, proj_loss=-0.186][2026-03-26 12:33:48] Step: 651, Training Logs: loss_final: 0.796590, loss_mean: 0.887703, proj_loss: -0.181014, loss_mean_cls: 0.089901, grad_norm: 1.102123 +Steps: 0%| | 652/1000000 [02:45<68:41:01, 4.04it/s, grad_norm=1.1, loss_final=0.797, loss_mean=0.888, loss_mean_cls=0.0899, proj_loss=-0.181][2026-03-26 12:33:48] Step: 652, Training Logs: loss_final: 0.806734, loss_mean: 0.898308, proj_loss: -0.181864, loss_mean_cls: 0.090290, grad_norm: 0.442461 +Steps: 0%| | 653/1000000 [02:45<68:44:01, 4.04it/s, grad_norm=0.442, loss_final=0.807, loss_mean=0.898, loss_mean_cls=0.0903, proj_loss=-0.182][2026-03-26 12:33:48] Step: 653, Training Logs: loss_final: 0.805688, loss_mean: 0.905165, proj_loss: -0.187471, loss_mean_cls: 0.087994, grad_norm: 1.213329 +Steps: 0%| | 654/1000000 [02:45<68:42:19, 4.04it/s, grad_norm=1.21, loss_final=0.806, loss_mean=0.905, loss_mean_cls=0.088, proj_loss=-0.187][2026-03-26 12:33:49] Step: 654, Training Logs: loss_final: 0.788537, loss_mean: 0.886679, proj_loss: -0.188251, loss_mean_cls: 0.090109, grad_norm: 0.408487 +Steps: 0%| | 655/1000000 [02:45<68:45:29, 4.04it/s, grad_norm=0.408, loss_final=0.789, loss_mean=0.887, loss_mean_cls=0.0901, proj_loss=-0.188][2026-03-26 12:33:49] Step: 655, Training Logs: loss_final: 0.804052, loss_mean: 0.902827, proj_loss: -0.188184, loss_mean_cls: 0.089409, grad_norm: 1.271014 +Steps: 0%| | 656/1000000 [02:46<68:42:34, 4.04it/s, grad_norm=1.27, loss_final=0.804, loss_mean=0.903, loss_mean_cls=0.0894, proj_loss=-0.188][2026-03-26 12:33:49] Step: 656, Training Logs: loss_final: 0.787566, loss_mean: 0.880154, proj_loss: -0.184112, loss_mean_cls: 0.091524, grad_norm: 0.572211 +Steps: 0%| | 657/1000000 [02:46<68:43:40, 4.04it/s, grad_norm=0.572, loss_final=0.788, loss_mean=0.88, loss_mean_cls=0.0915, proj_loss=-0.184][2026-03-26 12:33:49] Step: 657, Training Logs: loss_final: 0.805983, loss_mean: 0.920178, proj_loss: -0.200172, loss_mean_cls: 0.085977, grad_norm: 1.119224 +Steps: 0%| | 658/1000000 [02:46<68:41:09, 4.04it/s, grad_norm=1.12, loss_final=0.806, loss_mean=0.92, loss_mean_cls=0.086, proj_loss=-0.2][2026-03-26 12:33:50] Step: 658, Training Logs: loss_final: 0.802874, loss_mean: 0.898746, proj_loss: -0.186713, loss_mean_cls: 0.090841, grad_norm: 0.736029 +Steps: 0%| | 659/1000000 [02:46<68:40:57, 4.04it/s, grad_norm=0.736, loss_final=0.803, loss_mean=0.899, loss_mean_cls=0.0908, proj_loss=-0.187][2026-03-26 12:33:50] Step: 659, Training Logs: loss_final: 0.789656, loss_mean: 0.893733, proj_loss: -0.193871, loss_mean_cls: 0.089795, grad_norm: 1.089957 +Steps: 0%| | 660/1000000 [02:47<68:41:13, 4.04it/s, grad_norm=1.09, loss_final=0.79, loss_mean=0.894, loss_mean_cls=0.0898, proj_loss=-0.194][2026-03-26 12:33:50] Step: 660, Training Logs: loss_final: 0.799002, loss_mean: 0.916473, proj_loss: -0.203684, loss_mean_cls: 0.086214, grad_norm: 0.614755 +Steps: 0%| | 661/1000000 [02:47<68:37:45, 4.04it/s, grad_norm=0.615, loss_final=0.799, loss_mean=0.916, loss_mean_cls=0.0862, proj_loss=-0.204][2026-03-26 12:33:50] Step: 661, Training Logs: loss_final: 0.771212, loss_mean: 0.884396, proj_loss: -0.202382, loss_mean_cls: 0.089197, grad_norm: 0.934616 +Steps: 0%| | 662/1000000 [02:47<68:38:01, 4.04it/s, grad_norm=0.935, loss_final=0.771, loss_mean=0.884, loss_mean_cls=0.0892, proj_loss=-0.202][2026-03-26 12:33:51] Step: 662, Training Logs: loss_final: 0.780868, loss_mean: 0.900609, proj_loss: -0.206787, loss_mean_cls: 0.087045, grad_norm: 0.786675 +Steps: 0%| | 663/1000000 [02:47<68:39:42, 4.04it/s, grad_norm=0.787, loss_final=0.781, loss_mean=0.901, loss_mean_cls=0.087, proj_loss=-0.207][2026-03-26 12:33:51] Step: 663, Training Logs: loss_final: 0.798639, loss_mean: 0.921564, proj_loss: -0.208679, loss_mean_cls: 0.085754, grad_norm: 0.807896 +Steps: 0%| | 664/1000000 [02:48<68:37:30, 4.05it/s, grad_norm=0.808, loss_final=0.799, loss_mean=0.922, loss_mean_cls=0.0858, proj_loss=-0.209][2026-03-26 12:33:51] Step: 664, Training Logs: loss_final: 0.779856, loss_mean: 0.898608, proj_loss: -0.206493, loss_mean_cls: 0.087741, grad_norm: 0.727298 +Steps: 0%| | 665/1000000 [02:48<68:40:02, 4.04it/s, grad_norm=0.727, loss_final=0.78, loss_mean=0.899, loss_mean_cls=0.0877, proj_loss=-0.206][2026-03-26 12:33:51] Step: 665, Training Logs: loss_final: 0.773633, loss_mean: 0.884074, proj_loss: -0.199799, loss_mean_cls: 0.089357, grad_norm: 0.694067 +Steps: 0%| | 666/1000000 [02:48<68:43:42, 4.04it/s, grad_norm=0.694, loss_final=0.774, loss_mean=0.884, loss_mean_cls=0.0894, proj_loss=-0.2][2026-03-26 12:33:52] Step: 666, Training Logs: loss_final: 0.782267, loss_mean: 0.901367, proj_loss: -0.205993, loss_mean_cls: 0.086893, grad_norm: 0.862858 +Steps: 0%| | 667/1000000 [02:48<68:42:30, 4.04it/s, grad_norm=0.863, loss_final=0.782, loss_mean=0.901, loss_mean_cls=0.0869, proj_loss=-0.206][2026-03-26 12:33:52] Step: 667, Training Logs: loss_final: 0.783613, loss_mean: 0.909741, proj_loss: -0.213581, loss_mean_cls: 0.087452, grad_norm: 0.890019 +Steps: 0%| | 668/1000000 [02:49<68:42:36, 4.04it/s, grad_norm=0.89, loss_final=0.784, loss_mean=0.91, loss_mean_cls=0.0875, proj_loss=-0.214][2026-03-26 12:33:52] Step: 668, Training Logs: loss_final: 0.789546, loss_mean: 0.909610, proj_loss: -0.208305, loss_mean_cls: 0.088241, grad_norm: 0.892861 +Steps: 0%| | 669/1000000 [02:49<68:41:23, 4.04it/s, grad_norm=0.893, loss_final=0.79, loss_mean=0.91, loss_mean_cls=0.0882, proj_loss=-0.208][2026-03-26 12:33:52] Step: 669, Training Logs: loss_final: 0.774517, loss_mean: 0.887313, proj_loss: -0.202396, loss_mean_cls: 0.089601, grad_norm: 1.284951 +Steps: 0%| | 670/1000000 [02:49<68:46:04, 4.04it/s, grad_norm=1.28, loss_final=0.775, loss_mean=0.887, loss_mean_cls=0.0896, proj_loss=-0.202][2026-03-26 12:33:53] Step: 670, Training Logs: loss_final: 0.776232, loss_mean: 0.904021, proj_loss: -0.214885, loss_mean_cls: 0.087096, grad_norm: 0.550569 +Steps: 0%| | 671/1000000 [02:49<68:44:01, 4.04it/s, grad_norm=0.551, loss_final=0.776, loss_mean=0.904, loss_mean_cls=0.0871, proj_loss=-0.215][2026-03-26 12:33:53] Step: 671, Training Logs: loss_final: 0.775953, loss_mean: 0.904961, proj_loss: -0.216569, loss_mean_cls: 0.087561, grad_norm: 0.689483 +Steps: 0%| | 672/1000000 [02:50<68:40:40, 4.04it/s, grad_norm=0.689, loss_final=0.776, loss_mean=0.905, loss_mean_cls=0.0876, proj_loss=-0.217][2026-03-26 12:33:53] Step: 672, Training Logs: loss_final: 0.768940, loss_mean: 0.890789, proj_loss: -0.210813, loss_mean_cls: 0.088964, grad_norm: 0.753367 +Steps: 0%| | 673/1000000 [02:50<68:39:04, 4.04it/s, grad_norm=0.753, loss_final=0.769, loss_mean=0.891, loss_mean_cls=0.089, proj_loss=-0.211][2026-03-26 12:33:53] Step: 673, Training Logs: loss_final: 0.788315, loss_mean: 0.914994, proj_loss: -0.213646, loss_mean_cls: 0.086967, grad_norm: 0.889172 +Steps: 0%| | 674/1000000 [02:50<68:47:25, 4.04it/s, grad_norm=0.889, loss_final=0.788, loss_mean=0.915, loss_mean_cls=0.087, proj_loss=-0.214][2026-03-26 12:33:54] Step: 674, Training Logs: loss_final: 0.758586, loss_mean: 0.876535, proj_loss: -0.208124, loss_mean_cls: 0.090176, grad_norm: 0.558888 +Steps: 0%| | 675/1000000 [02:50<68:43:51, 4.04it/s, grad_norm=0.559, loss_final=0.759, loss_mean=0.877, loss_mean_cls=0.0902, proj_loss=-0.208][2026-03-26 12:33:54] Step: 675, Training Logs: loss_final: 0.769254, loss_mean: 0.893238, proj_loss: -0.213887, loss_mean_cls: 0.089904, grad_norm: 0.880892 +Steps: 0%| | 676/1000000 [02:51<68:44:09, 4.04it/s, grad_norm=0.881, loss_final=0.769, loss_mean=0.893, loss_mean_cls=0.0899, proj_loss=-0.214][2026-03-26 12:33:54] Step: 676, Training Logs: loss_final: 0.787639, loss_mean: 0.919786, proj_loss: -0.218564, loss_mean_cls: 0.086416, grad_norm: 0.747750 +Steps: 0%| | 677/1000000 [02:51<68:42:39, 4.04it/s, grad_norm=0.748, loss_final=0.788, loss_mean=0.92, loss_mean_cls=0.0864, proj_loss=-0.219][2026-03-26 12:33:54] Step: 677, Training Logs: loss_final: 0.776099, loss_mean: 0.909901, proj_loss: -0.220405, loss_mean_cls: 0.086604, grad_norm: 0.725681 +Steps: 0%| | 678/1000000 [02:51<68:43:23, 4.04it/s, grad_norm=0.726, loss_final=0.776, loss_mean=0.91, loss_mean_cls=0.0866, proj_loss=-0.22][2026-03-26 12:33:55] Step: 678, Training Logs: loss_final: 0.764266, loss_mean: 0.896196, proj_loss: -0.219544, loss_mean_cls: 0.087614, grad_norm: 0.999543 +Steps: 0%| | 679/1000000 [02:51<68:42:14, 4.04it/s, grad_norm=1, loss_final=0.764, loss_mean=0.896, loss_mean_cls=0.0876, proj_loss=-0.22][2026-03-26 12:33:55] Step: 679, Training Logs: loss_final: 0.771227, loss_mean: 0.894322, proj_loss: -0.213500, loss_mean_cls: 0.090406, grad_norm: 0.776393 +Steps: 0%| | 680/1000000 [02:52<68:42:19, 4.04it/s, grad_norm=0.776, loss_final=0.771, loss_mean=0.894, loss_mean_cls=0.0904, proj_loss=-0.214][2026-03-26 12:33:55] Step: 680, Training Logs: loss_final: 0.764524, loss_mean: 0.897817, proj_loss: -0.221449, loss_mean_cls: 0.088156, grad_norm: 1.061844 +Steps: 0%| | 681/1000000 [02:52<68:38:49, 4.04it/s, grad_norm=1.06, loss_final=0.765, loss_mean=0.898, loss_mean_cls=0.0882, proj_loss=-0.221][2026-03-26 12:33:55] Step: 681, Training Logs: loss_final: 0.775650, loss_mean: 0.903328, proj_loss: -0.216981, loss_mean_cls: 0.089304, grad_norm: 0.992711 +Steps: 0%| | 682/1000000 [02:52<68:40:31, 4.04it/s, grad_norm=0.993, loss_final=0.776, loss_mean=0.903, loss_mean_cls=0.0893, proj_loss=-0.217][2026-03-26 12:33:56] Step: 682, Training Logs: loss_final: 0.778156, loss_mean: 0.910980, proj_loss: -0.220898, loss_mean_cls: 0.088074, grad_norm: 0.990379 +Steps: 0%| | 683/1000000 [02:52<68:42:37, 4.04it/s, grad_norm=0.99, loss_final=0.778, loss_mean=0.911, loss_mean_cls=0.0881, proj_loss=-0.221][2026-03-26 12:33:56] Step: 683, Training Logs: loss_final: 0.778733, loss_mean: 0.918371, proj_loss: -0.227190, loss_mean_cls: 0.087551, grad_norm: 0.773872 +Steps: 0%| | 684/1000000 [02:53<68:43:01, 4.04it/s, grad_norm=0.774, loss_final=0.779, loss_mean=0.918, loss_mean_cls=0.0876, proj_loss=-0.227][2026-03-26 12:33:56] Step: 684, Training Logs: loss_final: 0.762302, loss_mean: 0.903128, proj_loss: -0.228901, loss_mean_cls: 0.088075, grad_norm: 0.942447 +Steps: 0%| | 685/1000000 [02:53<68:42:20, 4.04it/s, grad_norm=0.942, loss_final=0.762, loss_mean=0.903, loss_mean_cls=0.0881, proj_loss=-0.229][2026-03-26 12:33:56] Step: 685, Training Logs: loss_final: 0.749187, loss_mean: 0.878570, proj_loss: -0.219300, loss_mean_cls: 0.089917, grad_norm: 0.686358 +Steps: 0%| | 686/1000000 [02:53<68:45:21, 4.04it/s, grad_norm=0.686, loss_final=0.749, loss_mean=0.879, loss_mean_cls=0.0899, proj_loss=-0.219][2026-03-26 12:33:57] Step: 686, Training Logs: loss_final: 0.769746, loss_mean: 0.907349, proj_loss: -0.226976, loss_mean_cls: 0.089373, grad_norm: 1.097441 +Steps: 0%| | 687/1000000 [02:53<68:46:04, 4.04it/s, grad_norm=1.1, loss_final=0.77, loss_mean=0.907, loss_mean_cls=0.0894, proj_loss=-0.227][2026-03-26 12:33:57] Step: 687, Training Logs: loss_final: 0.764239, loss_mean: 0.908656, proj_loss: -0.230857, loss_mean_cls: 0.086440, grad_norm: 1.055716 +Steps: 0%| | 688/1000000 [02:54<68:44:09, 4.04it/s, grad_norm=1.06, loss_final=0.764, loss_mean=0.909, loss_mean_cls=0.0864, proj_loss=-0.231][2026-03-26 12:33:57] Step: 688, Training Logs: loss_final: 0.751112, loss_mean: 0.895287, proj_loss: -0.231664, loss_mean_cls: 0.087490, grad_norm: 0.794902 +Steps: 0%| | 689/1000000 [02:54<68:45:07, 4.04it/s, grad_norm=0.795, loss_final=0.751, loss_mean=0.895, loss_mean_cls=0.0875, proj_loss=-0.232][2026-03-26 12:33:57] Step: 689, Training Logs: loss_final: 0.754640, loss_mean: 0.891369, proj_loss: -0.225604, loss_mean_cls: 0.088875, grad_norm: 1.133097 +Steps: 0%| | 690/1000000 [02:54<68:44:15, 4.04it/s, grad_norm=1.13, loss_final=0.755, loss_mean=0.891, loss_mean_cls=0.0889, proj_loss=-0.226][2026-03-26 12:33:57] Step: 690, Training Logs: loss_final: 0.756475, loss_mean: 0.899805, proj_loss: -0.232324, loss_mean_cls: 0.088993, grad_norm: 0.631941 +Steps: 0%| | 691/1000000 [02:54<68:48:38, 4.03it/s, grad_norm=0.632, loss_final=0.756, loss_mean=0.9, loss_mean_cls=0.089, proj_loss=-0.232][2026-03-26 12:33:58] Step: 691, Training Logs: loss_final: 0.735881, loss_mean: 0.871699, proj_loss: -0.226149, loss_mean_cls: 0.090331, grad_norm: 0.956983 +Steps: 0%| | 692/1000000 [02:55<68:44:26, 4.04it/s, grad_norm=0.957, loss_final=0.736, loss_mean=0.872, loss_mean_cls=0.0903, proj_loss=-0.226][2026-03-26 12:33:58] Step: 692, Training Logs: loss_final: 0.754619, loss_mean: 0.904625, proj_loss: -0.237398, loss_mean_cls: 0.087392, grad_norm: 0.703185 +Steps: 0%| | 693/1000000 [02:55<68:42:30, 4.04it/s, grad_norm=0.703, loss_final=0.755, loss_mean=0.905, loss_mean_cls=0.0874, proj_loss=-0.237][2026-03-26 12:33:58] Step: 693, Training Logs: loss_final: 0.763325, loss_mean: 0.912647, proj_loss: -0.236284, loss_mean_cls: 0.086962, grad_norm: 0.974848 +Steps: 0%| | 694/1000000 [02:55<68:40:45, 4.04it/s, grad_norm=0.975, loss_final=0.763, loss_mean=0.913, loss_mean_cls=0.087, proj_loss=-0.236][2026-03-26 12:33:58] Step: 694, Training Logs: loss_final: 0.745866, loss_mean: 0.888887, proj_loss: -0.232386, loss_mean_cls: 0.089364, grad_norm: 0.451928 +Steps: 0%| | 695/1000000 [02:55<68:42:42, 4.04it/s, grad_norm=0.452, loss_final=0.746, loss_mean=0.889, loss_mean_cls=0.0894, proj_loss=-0.232][2026-03-26 12:33:59] Step: 695, Training Logs: loss_final: 0.754953, loss_mean: 0.900128, proj_loss: -0.234355, loss_mean_cls: 0.089181, grad_norm: 0.698542 +Steps: 0%| | 696/1000000 [02:56<68:38:40, 4.04it/s, grad_norm=0.699, loss_final=0.755, loss_mean=0.9, loss_mean_cls=0.0892, proj_loss=-0.234][2026-03-26 12:33:59] Step: 696, Training Logs: loss_final: 0.743671, loss_mean: 0.887056, proj_loss: -0.233130, loss_mean_cls: 0.089745, grad_norm: 0.621495 +Steps: 0%| | 697/1000000 [02:56<68:36:58, 4.05it/s, grad_norm=0.621, loss_final=0.744, loss_mean=0.887, loss_mean_cls=0.0897, proj_loss=-0.233][2026-03-26 12:33:59] Step: 697, Training Logs: loss_final: 0.728269, loss_mean: 0.871763, proj_loss: -0.234033, loss_mean_cls: 0.090539, grad_norm: 0.763117 +Steps: 0%| | 698/1000000 [02:56<68:38:35, 4.04it/s, grad_norm=0.763, loss_final=0.728, loss_mean=0.872, loss_mean_cls=0.0905, proj_loss=-0.234][2026-03-26 12:33:59] Step: 698, Training Logs: loss_final: 0.740820, loss_mean: 0.884044, proj_loss: -0.233557, loss_mean_cls: 0.090332, grad_norm: 0.605081 +Steps: 0%| | 699/1000000 [02:56<68:40:54, 4.04it/s, grad_norm=0.605, loss_final=0.741, loss_mean=0.884, loss_mean_cls=0.0903, proj_loss=-0.234][2026-03-26 12:34:00] Step: 699, Training Logs: loss_final: 0.737073, loss_mean: 0.887565, proj_loss: -0.239312, loss_mean_cls: 0.088820, grad_norm: 1.081179 +Steps: 0%| | 700/1000000 [02:57<68:42:01, 4.04it/s, grad_norm=1.08, loss_final=0.737, loss_mean=0.888, loss_mean_cls=0.0888, proj_loss=-0.239][2026-03-26 12:34:00] Step: 700, Training Logs: loss_final: 0.754948, loss_mean: 0.913176, proj_loss: -0.244538, loss_mean_cls: 0.086310, grad_norm: 1.016992 +Steps: 0%| | 701/1000000 [02:57<68:41:24, 4.04it/s, grad_norm=1.02, loss_final=0.755, loss_mean=0.913, loss_mean_cls=0.0863, proj_loss=-0.245][2026-03-26 12:34:00] Step: 701, Training Logs: loss_final: 0.762328, loss_mean: 0.924089, proj_loss: -0.247372, loss_mean_cls: 0.085611, grad_norm: 0.955231 +Steps: 0%| | 702/1000000 [02:57<68:40:46, 4.04it/s, grad_norm=0.955, loss_final=0.762, loss_mean=0.924, loss_mean_cls=0.0856, proj_loss=-0.247][2026-03-26 12:34:00] Step: 702, Training Logs: loss_final: 0.728918, loss_mean: 0.877762, proj_loss: -0.237885, loss_mean_cls: 0.089041, grad_norm: 0.840738 +Steps: 0%| | 703/1000000 [02:57<68:41:38, 4.04it/s, grad_norm=0.841, loss_final=0.729, loss_mean=0.878, loss_mean_cls=0.089, proj_loss=-0.238][2026-03-26 12:34:01] Step: 703, Training Logs: loss_final: 0.723167, loss_mean: 0.873119, proj_loss: -0.240541, loss_mean_cls: 0.090589, grad_norm: 0.698517 +Steps: 0%| | 704/1000000 [02:58<68:42:06, 4.04it/s, grad_norm=0.699, loss_final=0.723, loss_mean=0.873, loss_mean_cls=0.0906, proj_loss=-0.241][2026-03-26 12:34:01] Step: 704, Training Logs: loss_final: 0.731156, loss_mean: 0.894198, proj_loss: -0.249977, loss_mean_cls: 0.086936, grad_norm: 0.720244 +Steps: 0%| | 705/1000000 [02:58<68:41:12, 4.04it/s, grad_norm=0.72, loss_final=0.731, loss_mean=0.894, loss_mean_cls=0.0869, proj_loss=-0.25][2026-03-26 12:34:01] Step: 705, Training Logs: loss_final: 0.733990, loss_mean: 0.887879, proj_loss: -0.242608, loss_mean_cls: 0.088719, grad_norm: 0.704867 +Steps: 0%| | 706/1000000 [02:58<68:46:51, 4.04it/s, grad_norm=0.705, loss_final=0.734, loss_mean=0.888, loss_mean_cls=0.0887, proj_loss=-0.243][2026-03-26 12:34:01] Step: 706, Training Logs: loss_final: 0.718628, loss_mean: 0.866407, proj_loss: -0.237650, loss_mean_cls: 0.089871, grad_norm: 0.806147 +Steps: 0%| | 707/1000000 [02:58<68:46:41, 4.04it/s, grad_norm=0.806, loss_final=0.719, loss_mean=0.866, loss_mean_cls=0.0899, proj_loss=-0.238][2026-03-26 12:34:02] Step: 707, Training Logs: loss_final: 0.720488, loss_mean: 0.880941, proj_loss: -0.248028, loss_mean_cls: 0.087575, grad_norm: 0.683935 +Steps: 0%| | 708/1000000 [02:59<68:46:28, 4.04it/s, grad_norm=0.684, loss_final=0.72, loss_mean=0.881, loss_mean_cls=0.0876, proj_loss=-0.248][2026-03-26 12:34:02] Step: 708, Training Logs: loss_final: 0.729346, loss_mean: 0.882155, proj_loss: -0.241938, loss_mean_cls: 0.089129, grad_norm: 0.785862 +Steps: 0%| | 709/1000000 [02:59<68:43:23, 4.04it/s, grad_norm=0.786, loss_final=0.729, loss_mean=0.882, loss_mean_cls=0.0891, proj_loss=-0.242][2026-03-26 12:34:02] Step: 709, Training Logs: loss_final: 0.719688, loss_mean: 0.879901, proj_loss: -0.248713, loss_mean_cls: 0.088501, grad_norm: 0.618181 +Steps: 0%| | 710/1000000 [02:59<68:46:23, 4.04it/s, grad_norm=0.618, loss_final=0.72, loss_mean=0.88, loss_mean_cls=0.0885, proj_loss=-0.249][2026-03-26 12:34:02] Step: 710, Training Logs: loss_final: 0.729193, loss_mean: 0.887823, proj_loss: -0.246495, loss_mean_cls: 0.087865, grad_norm: 0.636426 +Steps: 0%| | 711/1000000 [02:59<68:46:06, 4.04it/s, grad_norm=0.636, loss_final=0.729, loss_mean=0.888, loss_mean_cls=0.0879, proj_loss=-0.246][2026-03-26 12:34:03] Step: 711, Training Logs: loss_final: 0.729474, loss_mean: 0.882331, proj_loss: -0.242162, loss_mean_cls: 0.089305, grad_norm: 0.842347 +Steps: 0%| | 712/1000000 [03:00<68:45:34, 4.04it/s, grad_norm=0.842, loss_final=0.729, loss_mean=0.882, loss_mean_cls=0.0893, proj_loss=-0.242][2026-03-26 12:34:03] Step: 712, Training Logs: loss_final: 0.746637, loss_mean: 0.899580, proj_loss: -0.242566, loss_mean_cls: 0.089623, grad_norm: 0.981900 +Steps: 0%| | 713/1000000 [03:00<68:44:32, 4.04it/s, grad_norm=0.982, loss_final=0.747, loss_mean=0.9, loss_mean_cls=0.0896, proj_loss=-0.243][2026-03-26 12:34:03] Step: 713, Training Logs: loss_final: 0.736160, loss_mean: 0.896125, proj_loss: -0.247296, loss_mean_cls: 0.087331, grad_norm: 0.806082 +Steps: 0%| | 714/1000000 [03:00<68:51:47, 4.03it/s, grad_norm=0.806, loss_final=0.736, loss_mean=0.896, loss_mean_cls=0.0873, proj_loss=-0.247][2026-03-26 12:34:03] Step: 714, Training Logs: loss_final: 0.736050, loss_mean: 0.904696, proj_loss: -0.255180, loss_mean_cls: 0.086533, grad_norm: 0.786964 +Steps: 0%| | 715/1000000 [03:00<68:49:48, 4.03it/s, grad_norm=0.787, loss_final=0.736, loss_mean=0.905, loss_mean_cls=0.0865, proj_loss=-0.255][2026-03-26 12:34:04] Step: 715, Training Logs: loss_final: 0.735029, loss_mean: 0.896587, proj_loss: -0.249341, loss_mean_cls: 0.087782, grad_norm: 0.792503 +Steps: 0%| | 716/1000000 [03:01<68:49:30, 4.03it/s, grad_norm=0.793, loss_final=0.735, loss_mean=0.897, loss_mean_cls=0.0878, proj_loss=-0.249][2026-03-26 12:34:04] Step: 716, Training Logs: loss_final: 0.742863, loss_mean: 0.909210, proj_loss: -0.254139, loss_mean_cls: 0.087791, grad_norm: 0.656601 +Steps: 0%| | 717/1000000 [03:01<68:47:06, 4.04it/s, grad_norm=0.657, loss_final=0.743, loss_mean=0.909, loss_mean_cls=0.0878, proj_loss=-0.254][2026-03-26 12:34:04] Step: 717, Training Logs: loss_final: 0.727072, loss_mean: 0.890721, proj_loss: -0.251938, loss_mean_cls: 0.088289, grad_norm: 0.841896 +Steps: 0%| | 718/1000000 [03:01<68:45:25, 4.04it/s, grad_norm=0.842, loss_final=0.727, loss_mean=0.891, loss_mean_cls=0.0883, proj_loss=-0.252][2026-03-26 12:34:04] Step: 718, Training Logs: loss_final: 0.741372, loss_mean: 0.907890, proj_loss: -0.253176, loss_mean_cls: 0.086658, grad_norm: 0.566268 +Steps: 0%| | 719/1000000 [03:01<68:41:12, 4.04it/s, grad_norm=0.566, loss_final=0.741, loss_mean=0.908, loss_mean_cls=0.0867, proj_loss=-0.253][2026-03-26 12:34:05] Step: 719, Training Logs: loss_final: 0.716520, loss_mean: 0.884675, proj_loss: -0.255672, loss_mean_cls: 0.087517, grad_norm: 0.650225 +Steps: 0%| | 720/1000000 [03:02<68:46:34, 4.04it/s, grad_norm=0.65, loss_final=0.717, loss_mean=0.885, loss_mean_cls=0.0875, proj_loss=-0.256][2026-03-26 12:34:05] Step: 720, Training Logs: loss_final: 0.723289, loss_mean: 0.896491, proj_loss: -0.259966, loss_mean_cls: 0.086764, grad_norm: 0.703133 +Steps: 0%| | 721/1000000 [03:02<68:44:15, 4.04it/s, grad_norm=0.703, loss_final=0.723, loss_mean=0.896, loss_mean_cls=0.0868, proj_loss=-0.26][2026-03-26 12:34:05] Step: 721, Training Logs: loss_final: 0.737914, loss_mean: 0.903392, proj_loss: -0.253093, loss_mean_cls: 0.087615, grad_norm: 0.559085 +Steps: 0%| | 722/1000000 [03:02<68:42:35, 4.04it/s, grad_norm=0.559, loss_final=0.738, loss_mean=0.903, loss_mean_cls=0.0876, proj_loss=-0.253][2026-03-26 12:34:05] Step: 722, Training Logs: loss_final: 0.715593, loss_mean: 0.867282, proj_loss: -0.242686, loss_mean_cls: 0.090997, grad_norm: 0.824649 +Steps: 0%| | 723/1000000 [03:02<68:38:41, 4.04it/s, grad_norm=0.825, loss_final=0.716, loss_mean=0.867, loss_mean_cls=0.091, proj_loss=-0.243][2026-03-26 12:34:06] Step: 723, Training Logs: loss_final: 0.721190, loss_mean: 0.897420, proj_loss: -0.262394, loss_mean_cls: 0.086164, grad_norm: 0.667359 +Steps: 0%| | 724/1000000 [03:03<68:38:16, 4.04it/s, grad_norm=0.667, loss_final=0.721, loss_mean=0.897, loss_mean_cls=0.0862, proj_loss=-0.262][2026-03-26 12:34:06] Step: 724, Training Logs: loss_final: 0.736352, loss_mean: 0.916371, proj_loss: -0.265134, loss_mean_cls: 0.085114, grad_norm: 0.826377 +Steps: 0%| | 725/1000000 [03:03<68:39:12, 4.04it/s, grad_norm=0.826, loss_final=0.736, loss_mean=0.916, loss_mean_cls=0.0851, proj_loss=-0.265][2026-03-26 12:34:06] Step: 725, Training Logs: loss_final: 0.735857, loss_mean: 0.906855, proj_loss: -0.258478, loss_mean_cls: 0.087479, grad_norm: 0.516780 +Steps: 0%| | 726/1000000 [03:03<68:40:01, 4.04it/s, grad_norm=0.517, loss_final=0.736, loss_mean=0.907, loss_mean_cls=0.0875, proj_loss=-0.258][2026-03-26 12:34:06] Step: 726, Training Logs: loss_final: 0.737402, loss_mean: 0.910123, proj_loss: -0.259630, loss_mean_cls: 0.086910, grad_norm: 1.229131 +Steps: 0%| | 727/1000000 [03:03<68:42:41, 4.04it/s, grad_norm=1.23, loss_final=0.737, loss_mean=0.91, loss_mean_cls=0.0869, proj_loss=-0.26][2026-03-26 12:34:07] Step: 727, Training Logs: loss_final: 0.729329, loss_mean: 0.912388, proj_loss: -0.268100, loss_mean_cls: 0.085041, grad_norm: 0.701096 +Steps: 0%| | 728/1000000 [03:04<68:40:14, 4.04it/s, grad_norm=0.701, loss_final=0.729, loss_mean=0.912, loss_mean_cls=0.085, proj_loss=-0.268][2026-03-26 12:34:07] Step: 728, Training Logs: loss_final: 0.731273, loss_mean: 0.894589, proj_loss: -0.252599, loss_mean_cls: 0.089283, grad_norm: 0.779492 +Steps: 0%| | 729/1000000 [03:04<68:40:08, 4.04it/s, grad_norm=0.779, loss_final=0.731, loss_mean=0.895, loss_mean_cls=0.0893, proj_loss=-0.253][2026-03-26 12:34:07] Step: 729, Training Logs: loss_final: 0.728844, loss_mean: 0.901214, proj_loss: -0.260862, loss_mean_cls: 0.088492, grad_norm: 0.714537 +Steps: 0%| | 730/1000000 [03:04<68:39:12, 4.04it/s, grad_norm=0.715, loss_final=0.729, loss_mean=0.901, loss_mean_cls=0.0885, proj_loss=-0.261][2026-03-26 12:34:07] Step: 730, Training Logs: loss_final: 0.710459, loss_mean: 0.895177, proj_loss: -0.270182, loss_mean_cls: 0.085464, grad_norm: 0.819666 +Steps: 0%| | 731/1000000 [03:04<68:44:18, 4.04it/s, grad_norm=0.82, loss_final=0.71, loss_mean=0.895, loss_mean_cls=0.0855, proj_loss=-0.27][2026-03-26 12:34:08] Step: 731, Training Logs: loss_final: 0.715345, loss_mean: 0.891263, proj_loss: -0.262586, loss_mean_cls: 0.086668, grad_norm: 0.733564 +Steps: 0%| | 732/1000000 [03:05<68:43:05, 4.04it/s, grad_norm=0.734, loss_final=0.715, loss_mean=0.891, loss_mean_cls=0.0867, proj_loss=-0.263][2026-03-26 12:34:08] Step: 732, Training Logs: loss_final: 0.724071, loss_mean: 0.894630, proj_loss: -0.258614, loss_mean_cls: 0.088055, grad_norm: 0.573418 +Steps: 0%| | 733/1000000 [03:05<68:42:29, 4.04it/s, grad_norm=0.573, loss_final=0.724, loss_mean=0.895, loss_mean_cls=0.0881, proj_loss=-0.259][2026-03-26 12:34:08] Step: 733, Training Logs: loss_final: 0.732623, loss_mean: 0.918575, proj_loss: -0.270464, loss_mean_cls: 0.084512, grad_norm: 0.815880 +Steps: 0%| | 734/1000000 [03:05<68:47:36, 4.03it/s, grad_norm=0.816, loss_final=0.733, loss_mean=0.919, loss_mean_cls=0.0845, proj_loss=-0.27][2026-03-26 12:34:08] Step: 734, Training Logs: loss_final: 0.715837, loss_mean: 0.897621, proj_loss: -0.268340, loss_mean_cls: 0.086556, grad_norm: 0.742887 +Steps: 0%| | 735/1000000 [03:05<69:42:56, 3.98it/s, grad_norm=0.743, loss_final=0.716, loss_mean=0.898, loss_mean_cls=0.0866, proj_loss=-0.268][2026-03-26 12:34:09] Step: 735, Training Logs: loss_final: 0.712724, loss_mean: 0.876935, proj_loss: -0.253449, loss_mean_cls: 0.089238, grad_norm: 0.812621 +Steps: 0%| | 736/1000000 [03:06<69:25:22, 4.00it/s, grad_norm=0.813, loss_final=0.713, loss_mean=0.877, loss_mean_cls=0.0892, proj_loss=-0.253][2026-03-26 12:34:09] Step: 736, Training Logs: loss_final: 0.720094, loss_mean: 0.906111, proj_loss: -0.270696, loss_mean_cls: 0.084678, grad_norm: 0.576468 +Steps: 0%| | 737/1000000 [03:06<69:12:01, 4.01it/s, grad_norm=0.576, loss_final=0.72, loss_mean=0.906, loss_mean_cls=0.0847, proj_loss=-0.271][2026-03-26 12:34:09] Step: 737, Training Logs: loss_final: 0.711635, loss_mean: 0.887192, proj_loss: -0.263300, loss_mean_cls: 0.087743, grad_norm: 0.824955 +Steps: 0%| | 738/1000000 [03:06<68:59:46, 4.02it/s, grad_norm=0.825, loss_final=0.712, loss_mean=0.887, loss_mean_cls=0.0877, proj_loss=-0.263][2026-03-26 12:34:09] Step: 738, Training Logs: loss_final: 0.716467, loss_mean: 0.897293, proj_loss: -0.267852, loss_mean_cls: 0.087026, grad_norm: 0.902528 +Steps: 0%| | 739/1000000 [03:06<68:53:22, 4.03it/s, grad_norm=0.903, loss_final=0.716, loss_mean=0.897, loss_mean_cls=0.087, proj_loss=-0.268][2026-03-26 12:34:10] Step: 739, Training Logs: loss_final: 0.708807, loss_mean: 0.902456, proj_loss: -0.278346, loss_mean_cls: 0.084697, grad_norm: 1.102866 +Steps: 0%| | 740/1000000 [03:06<68:49:04, 4.03it/s, grad_norm=1.1, loss_final=0.709, loss_mean=0.902, loss_mean_cls=0.0847, proj_loss=-0.278][2026-03-26 12:34:10] Step: 740, Training Logs: loss_final: 0.689653, loss_mean: 0.861105, proj_loss: -0.261387, loss_mean_cls: 0.089936, grad_norm: 0.716727 +Steps: 0%| | 741/1000000 [03:07<68:44:09, 4.04it/s, grad_norm=0.717, loss_final=0.69, loss_mean=0.861, loss_mean_cls=0.0899, proj_loss=-0.261][2026-03-26 12:34:10] Step: 741, Training Logs: loss_final: 0.708553, loss_mean: 0.882364, proj_loss: -0.262173, loss_mean_cls: 0.088363, grad_norm: 0.672798 +Steps: 0%| | 742/1000000 [03:07<68:40:29, 4.04it/s, grad_norm=0.673, loss_final=0.709, loss_mean=0.882, loss_mean_cls=0.0884, proj_loss=-0.262][2026-03-26 12:34:10] Step: 742, Training Logs: loss_final: 0.725740, loss_mean: 0.901793, proj_loss: -0.263618, loss_mean_cls: 0.087564, grad_norm: 0.672914 +Steps: 0%| | 743/1000000 [03:07<68:40:14, 4.04it/s, grad_norm=0.673, loss_final=0.726, loss_mean=0.902, loss_mean_cls=0.0876, proj_loss=-0.264][2026-03-26 12:34:11] Step: 743, Training Logs: loss_final: 0.710554, loss_mean: 0.903870, proj_loss: -0.278046, loss_mean_cls: 0.084730, grad_norm: 0.512395 +Steps: 0%| | 744/1000000 [03:07<68:40:35, 4.04it/s, grad_norm=0.512, loss_final=0.711, loss_mean=0.904, loss_mean_cls=0.0847, proj_loss=-0.278][2026-03-26 12:34:11] Step: 744, Training Logs: loss_final: 0.717786, loss_mean: 0.900255, proj_loss: -0.269116, loss_mean_cls: 0.086647, grad_norm: 0.663611 +Steps: 0%| | 745/1000000 [03:08<68:39:58, 4.04it/s, grad_norm=0.664, loss_final=0.718, loss_mean=0.9, loss_mean_cls=0.0866, proj_loss=-0.269][2026-03-26 12:34:11] Step: 745, Training Logs: loss_final: 0.716173, loss_mean: 0.906266, proj_loss: -0.275474, loss_mean_cls: 0.085382, grad_norm: 0.632018 +Steps: 0%| | 746/1000000 [03:08<68:37:46, 4.04it/s, grad_norm=0.632, loss_final=0.716, loss_mean=0.906, loss_mean_cls=0.0854, proj_loss=-0.275][2026-03-26 12:34:11] Step: 746, Training Logs: loss_final: 0.696287, loss_mean: 0.884889, proj_loss: -0.275045, loss_mean_cls: 0.086443, grad_norm: 0.720523 +Steps: 0%| | 747/1000000 [03:08<68:40:31, 4.04it/s, grad_norm=0.721, loss_final=0.696, loss_mean=0.885, loss_mean_cls=0.0864, proj_loss=-0.275][2026-03-26 12:34:12] Step: 747, Training Logs: loss_final: 0.706332, loss_mean: 0.890363, proj_loss: -0.270979, loss_mean_cls: 0.086948, grad_norm: 0.784285 +Steps: 0%| | 748/1000000 [03:08<68:48:02, 4.03it/s, grad_norm=0.784, loss_final=0.706, loss_mean=0.89, loss_mean_cls=0.0869, proj_loss=-0.271][2026-03-26 12:34:12] Step: 748, Training Logs: loss_final: 0.703561, loss_mean: 0.881719, proj_loss: -0.266296, loss_mean_cls: 0.088137, grad_norm: 0.496286 +Steps: 0%| | 749/1000000 [03:09<68:44:51, 4.04it/s, grad_norm=0.496, loss_final=0.704, loss_mean=0.882, loss_mean_cls=0.0881, proj_loss=-0.266][2026-03-26 12:34:12] Step: 749, Training Logs: loss_final: 0.700599, loss_mean: 0.880994, proj_loss: -0.268437, loss_mean_cls: 0.088043, grad_norm: 0.898764 +Steps: 0%| | 750/1000000 [03:09<68:48:49, 4.03it/s, grad_norm=0.899, loss_final=0.701, loss_mean=0.881, loss_mean_cls=0.088, proj_loss=-0.268][2026-03-26 12:34:12] Step: 750, Training Logs: loss_final: 0.718701, loss_mean: 0.893936, proj_loss: -0.265210, loss_mean_cls: 0.089976, grad_norm: 0.855435 +Steps: 0%| | 751/1000000 [03:09<68:46:03, 4.04it/s, grad_norm=0.855, loss_final=0.719, loss_mean=0.894, loss_mean_cls=0.09, proj_loss=-0.265][2026-03-26 12:34:13] Step: 751, Training Logs: loss_final: 0.692976, loss_mean: 0.871935, proj_loss: -0.266966, loss_mean_cls: 0.088007, grad_norm: 0.799038 +Steps: 0%| | 752/1000000 [03:09<68:45:36, 4.04it/s, grad_norm=0.799, loss_final=0.693, loss_mean=0.872, loss_mean_cls=0.088, proj_loss=-0.267][2026-03-26 12:34:13] Step: 752, Training Logs: loss_final: 0.705246, loss_mean: 0.891321, proj_loss: -0.273436, loss_mean_cls: 0.087361, grad_norm: 0.733896 +Steps: 0%| | 753/1000000 [03:10<68:47:15, 4.04it/s, grad_norm=0.734, loss_final=0.705, loss_mean=0.891, loss_mean_cls=0.0874, proj_loss=-0.273][2026-03-26 12:34:13] Step: 753, Training Logs: loss_final: 0.699978, loss_mean: 0.888678, proj_loss: -0.275581, loss_mean_cls: 0.086881, grad_norm: 0.574696 +Steps: 0%| | 754/1000000 [03:10<68:42:36, 4.04it/s, grad_norm=0.575, loss_final=0.7, loss_mean=0.889, loss_mean_cls=0.0869, proj_loss=-0.276][2026-03-26 12:34:13] Step: 754, Training Logs: loss_final: 0.693638, loss_mean: 0.871385, proj_loss: -0.267195, loss_mean_cls: 0.089448, grad_norm: 0.667449 +Steps: 0%| | 755/1000000 [03:10<68:40:50, 4.04it/s, grad_norm=0.667, loss_final=0.694, loss_mean=0.871, loss_mean_cls=0.0894, proj_loss=-0.267][2026-03-26 12:34:14] Step: 755, Training Logs: loss_final: 0.703011, loss_mean: 0.900847, proj_loss: -0.284061, loss_mean_cls: 0.086226, grad_norm: 0.506681 +Steps: 0%| | 756/1000000 [03:10<68:39:23, 4.04it/s, grad_norm=0.507, loss_final=0.703, loss_mean=0.901, loss_mean_cls=0.0862, proj_loss=-0.284][2026-03-26 12:34:14] Step: 756, Training Logs: loss_final: 0.716237, loss_mean: 0.898651, proj_loss: -0.269629, loss_mean_cls: 0.087215, grad_norm: 1.096449 +Steps: 0%| | 757/1000000 [03:11<68:38:06, 4.04it/s, grad_norm=1.1, loss_final=0.716, loss_mean=0.899, loss_mean_cls=0.0872, proj_loss=-0.27][2026-03-26 12:34:14] Step: 757, Training Logs: loss_final: 0.709369, loss_mean: 0.897120, proj_loss: -0.273910, loss_mean_cls: 0.086159, grad_norm: 0.947561 +Steps: 0%| | 758/1000000 [03:11<68:40:13, 4.04it/s, grad_norm=0.948, loss_final=0.709, loss_mean=0.897, loss_mean_cls=0.0862, proj_loss=-0.274][2026-03-26 12:34:14] Step: 758, Training Logs: loss_final: 0.696751, loss_mean: 0.880908, proj_loss: -0.271467, loss_mean_cls: 0.087310, grad_norm: 1.019352 +Steps: 0%| | 759/1000000 [03:11<68:52:48, 4.03it/s, grad_norm=1.02, loss_final=0.697, loss_mean=0.881, loss_mean_cls=0.0873, proj_loss=-0.271][2026-03-26 12:34:15] Step: 759, Training Logs: loss_final: 0.706047, loss_mean: 0.895595, proj_loss: -0.274998, loss_mean_cls: 0.085450, grad_norm: 0.937930 +Steps: 0%| | 760/1000000 [03:11<71:22:00, 3.89it/s, grad_norm=0.938, loss_final=0.706, loss_mean=0.896, loss_mean_cls=0.0855, proj_loss=-0.275][2026-03-26 12:34:15] Step: 760, Training Logs: loss_final: 0.702166, loss_mean: 0.900277, proj_loss: -0.284421, loss_mean_cls: 0.086310, grad_norm: 1.166219 +Steps: 0%| | 761/1000000 [03:12<70:33:19, 3.93it/s, grad_norm=1.17, loss_final=0.702, loss_mean=0.9, loss_mean_cls=0.0863, proj_loss=-0.284][2026-03-26 12:34:15] Step: 761, Training Logs: loss_final: 0.712497, loss_mean: 0.905488, proj_loss: -0.279726, loss_mean_cls: 0.086736, grad_norm: 0.697059 +Steps: 0%| | 762/1000000 [03:12<70:00:39, 3.96it/s, grad_norm=0.697, loss_final=0.712, loss_mean=0.905, loss_mean_cls=0.0867, proj_loss=-0.28][2026-03-26 12:34:15] Step: 762, Training Logs: loss_final: 0.705913, loss_mean: 0.893064, proj_loss: -0.275097, loss_mean_cls: 0.087947, grad_norm: 0.966233 +Steps: 0%| | 763/1000000 [03:12<70:11:47, 3.95it/s, grad_norm=0.966, loss_final=0.706, loss_mean=0.893, loss_mean_cls=0.0879, proj_loss=-0.275][2026-03-26 12:34:16] Step: 763, Training Logs: loss_final: 0.702643, loss_mean: 0.905963, proj_loss: -0.288943, loss_mean_cls: 0.085623, grad_norm: 0.688869 +Steps: 0%| | 764/1000000 [03:12<69:43:27, 3.98it/s, grad_norm=0.689, loss_final=0.703, loss_mean=0.906, loss_mean_cls=0.0856, proj_loss=-0.289][2026-03-26 12:34:16] Step: 764, Training Logs: loss_final: 0.679287, loss_mean: 0.858143, proj_loss: -0.270173, loss_mean_cls: 0.091318, grad_norm: 0.689869 +Steps: 0%| | 765/1000000 [03:13<69:22:20, 4.00it/s, grad_norm=0.69, loss_final=0.679, loss_mean=0.858, loss_mean_cls=0.0913, proj_loss=-0.27][2026-03-26 12:34:16] Step: 765, Training Logs: loss_final: 0.695145, loss_mean: 0.881853, proj_loss: -0.274235, loss_mean_cls: 0.087527, grad_norm: 0.801302 +Steps: 0%| | 766/1000000 [03:13<69:10:07, 4.01it/s, grad_norm=0.801, loss_final=0.695, loss_mean=0.882, loss_mean_cls=0.0875, proj_loss=-0.274][2026-03-26 12:34:16] Step: 766, Training Logs: loss_final: 0.697740, loss_mean: 0.894238, proj_loss: -0.281582, loss_mean_cls: 0.085084, grad_norm: 0.594505 +Steps: 0%| | 767/1000000 [03:13<68:58:47, 4.02it/s, grad_norm=0.595, loss_final=0.698, loss_mean=0.894, loss_mean_cls=0.0851, proj_loss=-0.282][2026-03-26 12:34:17] Step: 767, Training Logs: loss_final: 0.702218, loss_mean: 0.892829, proj_loss: -0.278135, loss_mean_cls: 0.087524, grad_norm: 0.880142 +Steps: 0%| | 768/1000000 [03:13<68:55:01, 4.03it/s, grad_norm=0.88, loss_final=0.702, loss_mean=0.893, loss_mean_cls=0.0875, proj_loss=-0.278][2026-03-26 12:34:17] Step: 768, Training Logs: loss_final: 0.688819, loss_mean: 0.882693, proj_loss: -0.280892, loss_mean_cls: 0.087018, grad_norm: 0.722418 +Steps: 0%| | 769/1000000 [03:14<68:49:12, 4.03it/s, grad_norm=0.722, loss_final=0.689, loss_mean=0.883, loss_mean_cls=0.087, proj_loss=-0.281][2026-03-26 12:34:17] Step: 769, Training Logs: loss_final: 0.698053, loss_mean: 0.868726, proj_loss: -0.262500, loss_mean_cls: 0.091828, grad_norm: 0.596277 +Steps: 0%| | 770/1000000 [03:14<68:46:53, 4.04it/s, grad_norm=0.596, loss_final=0.698, loss_mean=0.869, loss_mean_cls=0.0918, proj_loss=-0.263][2026-03-26 12:34:17] Step: 770, Training Logs: loss_final: 0.697989, loss_mean: 0.892496, proj_loss: -0.282033, loss_mean_cls: 0.087526, grad_norm: 0.715021 +Steps: 0%| | 771/1000000 [03:14<68:43:45, 4.04it/s, grad_norm=0.715, loss_final=0.698, loss_mean=0.892, loss_mean_cls=0.0875, proj_loss=-0.282][2026-03-26 12:34:18] Step: 771, Training Logs: loss_final: 0.698307, loss_mean: 0.902603, proj_loss: -0.289950, loss_mean_cls: 0.085653, grad_norm: 0.783140 +Steps: 0%| | 772/1000000 [03:14<68:43:05, 4.04it/s, grad_norm=0.783, loss_final=0.698, loss_mean=0.903, loss_mean_cls=0.0857, proj_loss=-0.29][2026-03-26 12:34:18] Step: 772, Training Logs: loss_final: 0.704836, loss_mean: 0.894169, proj_loss: -0.277918, loss_mean_cls: 0.088585, grad_norm: 1.073066 +Steps: 0%| | 773/1000000 [03:15<68:40:39, 4.04it/s, grad_norm=1.07, loss_final=0.705, loss_mean=0.894, loss_mean_cls=0.0886, proj_loss=-0.278][2026-03-26 12:34:18] Step: 773, Training Logs: loss_final: 0.698017, loss_mean: 0.898002, proj_loss: -0.285584, loss_mean_cls: 0.085600, grad_norm: 0.574242 +Steps: 0%| | 774/1000000 [03:15<68:41:40, 4.04it/s, grad_norm=0.574, loss_final=0.698, loss_mean=0.898, loss_mean_cls=0.0856, proj_loss=-0.286][2026-03-26 12:34:18] Step: 774, Training Logs: loss_final: 0.696658, loss_mean: 0.911976, proj_loss: -0.298150, loss_mean_cls: 0.082832, grad_norm: 0.890799 +Steps: 0%| | 775/1000000 [03:15<68:40:19, 4.04it/s, grad_norm=0.891, loss_final=0.697, loss_mean=0.912, loss_mean_cls=0.0828, proj_loss=-0.298][2026-03-26 12:34:19] Step: 775, Training Logs: loss_final: 0.702987, loss_mean: 0.902010, proj_loss: -0.285533, loss_mean_cls: 0.086509, grad_norm: 1.191976 +Steps: 0%| | 776/1000000 [03:15<68:39:54, 4.04it/s, grad_norm=1.19, loss_final=0.703, loss_mean=0.902, loss_mean_cls=0.0865, proj_loss=-0.286][2026-03-26 12:34:19] Step: 776, Training Logs: loss_final: 0.688740, loss_mean: 0.887511, proj_loss: -0.285926, loss_mean_cls: 0.087155, grad_norm: 0.725493 +Steps: 0%| | 777/1000000 [03:16<68:33:04, 4.05it/s, grad_norm=0.725, loss_final=0.689, loss_mean=0.888, loss_mean_cls=0.0872, proj_loss=-0.286][2026-03-26 12:34:19] Step: 777, Training Logs: loss_final: 0.687622, loss_mean: 0.891472, proj_loss: -0.289865, loss_mean_cls: 0.086015, grad_norm: 0.883302 +Steps: 0%| | 778/1000000 [03:16<68:35:15, 4.05it/s, grad_norm=0.883, loss_final=0.688, loss_mean=0.891, loss_mean_cls=0.086, proj_loss=-0.29][2026-03-26 12:34:19] Step: 778, Training Logs: loss_final: 0.696714, loss_mean: 0.900475, proj_loss: -0.289103, loss_mean_cls: 0.085341, grad_norm: 1.250824 +Steps: 0%| | 779/1000000 [03:16<68:36:20, 4.05it/s, grad_norm=1.25, loss_final=0.697, loss_mean=0.9, loss_mean_cls=0.0853, proj_loss=-0.289][2026-03-26 12:34:20] Step: 779, Training Logs: loss_final: 0.709054, loss_mean: 0.912014, proj_loss: -0.288566, loss_mean_cls: 0.085606, grad_norm: 0.476460 +Steps: 0%| | 780/1000000 [03:16<68:38:15, 4.04it/s, grad_norm=0.476, loss_final=0.709, loss_mean=0.912, loss_mean_cls=0.0856, proj_loss=-0.289][2026-03-26 12:34:20] Step: 780, Training Logs: loss_final: 0.708162, loss_mean: 0.914542, proj_loss: -0.291138, loss_mean_cls: 0.084758, grad_norm: 1.208392 +Steps: 0%| | 781/1000000 [03:17<68:35:01, 4.05it/s, grad_norm=1.21, loss_final=0.708, loss_mean=0.915, loss_mean_cls=0.0848, proj_loss=-0.291][2026-03-26 12:34:20] Step: 781, Training Logs: loss_final: 0.704684, loss_mean: 0.896194, proj_loss: -0.280353, loss_mean_cls: 0.088844, grad_norm: 0.646748 +Steps: 0%| | 782/1000000 [03:17<68:37:38, 4.04it/s, grad_norm=0.647, loss_final=0.705, loss_mean=0.896, loss_mean_cls=0.0888, proj_loss=-0.28][2026-03-26 12:34:20] Step: 782, Training Logs: loss_final: 0.701843, loss_mean: 0.903648, proj_loss: -0.286901, loss_mean_cls: 0.085096, grad_norm: 0.961500 +Steps: 0%| | 783/1000000 [03:17<68:37:51, 4.04it/s, grad_norm=0.962, loss_final=0.702, loss_mean=0.904, loss_mean_cls=0.0851, proj_loss=-0.287][2026-03-26 12:34:21] Step: 783, Training Logs: loss_final: 0.689129, loss_mean: 0.898317, proj_loss: -0.294469, loss_mean_cls: 0.085282, grad_norm: 1.364235 +Steps: 0%| | 784/1000000 [03:17<68:34:48, 4.05it/s, grad_norm=1.36, loss_final=0.689, loss_mean=0.898, loss_mean_cls=0.0853, proj_loss=-0.294][2026-03-26 12:34:21] Step: 784, Training Logs: loss_final: 0.704474, loss_mean: 0.901493, proj_loss: -0.284163, loss_mean_cls: 0.087144, grad_norm: 0.891715 +Steps: 0%| | 785/1000000 [03:18<68:36:36, 4.05it/s, grad_norm=0.892, loss_final=0.704, loss_mean=0.901, loss_mean_cls=0.0871, proj_loss=-0.284][2026-03-26 12:34:21] Step: 785, Training Logs: loss_final: 0.690872, loss_mean: 0.887253, proj_loss: -0.283637, loss_mean_cls: 0.087256, grad_norm: 1.238953 +Steps: 0%| | 786/1000000 [03:18<68:39:34, 4.04it/s, grad_norm=1.24, loss_final=0.691, loss_mean=0.887, loss_mean_cls=0.0873, proj_loss=-0.284][2026-03-26 12:34:21] Step: 786, Training Logs: loss_final: 0.699451, loss_mean: 0.892548, proj_loss: -0.280617, loss_mean_cls: 0.087520, grad_norm: 0.972260 +Steps: 0%| | 787/1000000 [03:18<68:46:09, 4.04it/s, grad_norm=0.972, loss_final=0.699, loss_mean=0.893, loss_mean_cls=0.0875, proj_loss=-0.281][2026-03-26 12:34:22] Step: 787, Training Logs: loss_final: 0.698146, loss_mean: 0.904601, proj_loss: -0.291631, loss_mean_cls: 0.085177, grad_norm: 1.299410 +Steps: 0%| | 788/1000000 [03:18<68:45:01, 4.04it/s, grad_norm=1.3, loss_final=0.698, loss_mean=0.905, loss_mean_cls=0.0852, proj_loss=-0.292][2026-03-26 12:34:22] Step: 788, Training Logs: loss_final: 0.699852, loss_mean: 0.907827, proj_loss: -0.292079, loss_mean_cls: 0.084104, grad_norm: 0.731752 +Steps: 0%| | 789/1000000 [03:19<68:43:48, 4.04it/s, grad_norm=0.732, loss_final=0.7, loss_mean=0.908, loss_mean_cls=0.0841, proj_loss=-0.292][2026-03-26 12:34:22] Step: 789, Training Logs: loss_final: 0.678287, loss_mean: 0.879376, proj_loss: -0.287601, loss_mean_cls: 0.086512, grad_norm: 0.760493 +Steps: 0%| | 790/1000000 [03:19<68:42:41, 4.04it/s, grad_norm=0.76, loss_final=0.678, loss_mean=0.879, loss_mean_cls=0.0865, proj_loss=-0.288][2026-03-26 12:34:22] Step: 790, Training Logs: loss_final: 0.692636, loss_mean: 0.904718, proj_loss: -0.296709, loss_mean_cls: 0.084627, grad_norm: 0.862993 +Steps: 0%| | 791/1000000 [03:19<68:42:27, 4.04it/s, grad_norm=0.863, loss_final=0.693, loss_mean=0.905, loss_mean_cls=0.0846, proj_loss=-0.297][2026-03-26 12:34:23] Step: 791, Training Logs: loss_final: 0.690469, loss_mean: 0.889556, proj_loss: -0.285980, loss_mean_cls: 0.086894, grad_norm: 0.473418 +Steps: 0%| | 792/1000000 [03:19<68:41:33, 4.04it/s, grad_norm=0.473, loss_final=0.69, loss_mean=0.89, loss_mean_cls=0.0869, proj_loss=-0.286][2026-03-26 12:34:23] Step: 792, Training Logs: loss_final: 0.694793, loss_mean: 0.894263, proj_loss: -0.285517, loss_mean_cls: 0.086047, grad_norm: 0.604175 +Steps: 0%| | 793/1000000 [03:20<68:40:44, 4.04it/s, grad_norm=0.604, loss_final=0.695, loss_mean=0.894, loss_mean_cls=0.086, proj_loss=-0.286][2026-03-26 12:34:23] Step: 793, Training Logs: loss_final: 0.690787, loss_mean: 0.886834, proj_loss: -0.284847, loss_mean_cls: 0.088800, grad_norm: 0.801828 +Steps: 0%| | 794/1000000 [03:20<68:37:13, 4.04it/s, grad_norm=0.802, loss_final=0.691, loss_mean=0.887, loss_mean_cls=0.0888, proj_loss=-0.285][2026-03-26 12:34:23] Step: 794, Training Logs: loss_final: 0.700649, loss_mean: 0.904868, proj_loss: -0.290254, loss_mean_cls: 0.086034, grad_norm: 0.688377 +Steps: 0%| | 795/1000000 [03:20<68:34:13, 4.05it/s, grad_norm=0.688, loss_final=0.701, loss_mean=0.905, loss_mean_cls=0.086, proj_loss=-0.29][2026-03-26 12:34:24] Step: 795, Training Logs: loss_final: 0.684706, loss_mean: 0.887058, proj_loss: -0.289133, loss_mean_cls: 0.086781, grad_norm: 0.775633 +Steps: 0%| | 796/1000000 [03:20<68:32:17, 4.05it/s, grad_norm=0.776, loss_final=0.685, loss_mean=0.887, loss_mean_cls=0.0868, proj_loss=-0.289][2026-03-26 12:34:24] Step: 796, Training Logs: loss_final: 0.686836, loss_mean: 0.888388, proj_loss: -0.288712, loss_mean_cls: 0.087160, grad_norm: 0.614851 +Steps: 0%| | 797/1000000 [03:21<68:33:55, 4.05it/s, grad_norm=0.615, loss_final=0.687, loss_mean=0.888, loss_mean_cls=0.0872, proj_loss=-0.289][2026-03-26 12:34:24] Step: 797, Training Logs: loss_final: 0.680772, loss_mean: 0.887025, proj_loss: -0.293272, loss_mean_cls: 0.087019, grad_norm: 0.812501 +Steps: 0%| | 798/1000000 [03:21<68:34:19, 4.05it/s, grad_norm=0.813, loss_final=0.681, loss_mean=0.887, loss_mean_cls=0.087, proj_loss=-0.293][2026-03-26 12:34:24] Step: 798, Training Logs: loss_final: 0.676296, loss_mean: 0.886310, proj_loss: -0.295905, loss_mean_cls: 0.085890, grad_norm: 0.721094 +Steps: 0%| | 799/1000000 [03:21<68:36:10, 4.05it/s, grad_norm=0.721, loss_final=0.676, loss_mean=0.886, loss_mean_cls=0.0859, proj_loss=-0.296][2026-03-26 12:34:25] Step: 799, Training Logs: loss_final: 0.676639, loss_mean: 0.889893, proj_loss: -0.298326, loss_mean_cls: 0.085072, grad_norm: 0.618646 +Steps: 0%| | 800/1000000 [03:21<69:34:38, 3.99it/s, grad_norm=0.619, loss_final=0.677, loss_mean=0.89, loss_mean_cls=0.0851, proj_loss=-0.298][2026-03-26 12:34:25] Step: 800, Training Logs: loss_final: 0.665973, loss_mean: 0.876351, proj_loss: -0.297027, loss_mean_cls: 0.086649, grad_norm: 0.648053 +Steps: 0%| | 801/1000000 [03:22<69:20:40, 4.00it/s, grad_norm=0.648, loss_final=0.666, loss_mean=0.876, loss_mean_cls=0.0866, proj_loss=-0.297][2026-03-26 12:34:25] Step: 801, Training Logs: loss_final: 0.693594, loss_mean: 0.902652, proj_loss: -0.294907, loss_mean_cls: 0.085849, grad_norm: 0.668914 +Steps: 0%| | 802/1000000 [03:22<69:09:09, 4.01it/s, grad_norm=0.669, loss_final=0.694, loss_mean=0.903, loss_mean_cls=0.0858, proj_loss=-0.295][2026-03-26 12:34:25] Step: 802, Training Logs: loss_final: 0.696404, loss_mean: 0.910811, proj_loss: -0.298296, loss_mean_cls: 0.083889, grad_norm: 0.630664 +Steps: 0%| | 803/1000000 [03:22<68:56:30, 4.03it/s, grad_norm=0.631, loss_final=0.696, loss_mean=0.911, loss_mean_cls=0.0839, proj_loss=-0.298][2026-03-26 12:34:26] Step: 803, Training Logs: loss_final: 0.681958, loss_mean: 0.894098, proj_loss: -0.297478, loss_mean_cls: 0.085339, grad_norm: 0.833440 +Steps: 0%| | 804/1000000 [03:22<68:51:40, 4.03it/s, grad_norm=0.833, loss_final=0.682, loss_mean=0.894, loss_mean_cls=0.0853, proj_loss=-0.297][2026-03-26 12:34:26] Step: 804, Training Logs: loss_final: 0.683829, loss_mean: 0.888882, proj_loss: -0.292434, loss_mean_cls: 0.087382, grad_norm: 0.849240 +Steps: 0%| | 805/1000000 [03:23<68:46:50, 4.04it/s, grad_norm=0.849, loss_final=0.684, loss_mean=0.889, loss_mean_cls=0.0874, proj_loss=-0.292][2026-03-26 12:34:26] Step: 805, Training Logs: loss_final: 0.679223, loss_mean: 0.895641, proj_loss: -0.300139, loss_mean_cls: 0.083721, grad_norm: 0.768023 +Steps: 0%| | 806/1000000 [03:23<68:46:06, 4.04it/s, grad_norm=0.768, loss_final=0.679, loss_mean=0.896, loss_mean_cls=0.0837, proj_loss=-0.3][2026-03-26 12:34:26] Step: 806, Training Logs: loss_final: 0.690158, loss_mean: 0.899095, proj_loss: -0.295328, loss_mean_cls: 0.086390, grad_norm: 0.443222 +Steps: 0%| | 807/1000000 [03:23<68:44:00, 4.04it/s, grad_norm=0.443, loss_final=0.69, loss_mean=0.899, loss_mean_cls=0.0864, proj_loss=-0.295][2026-03-26 12:34:27] Step: 807, Training Logs: loss_final: 0.693960, loss_mean: 0.904209, proj_loss: -0.296514, loss_mean_cls: 0.086265, grad_norm: 0.947275 +Steps: 0%| | 808/1000000 [03:23<68:42:34, 4.04it/s, grad_norm=0.947, loss_final=0.694, loss_mean=0.904, loss_mean_cls=0.0863, proj_loss=-0.297][2026-03-26 12:34:27] Step: 808, Training Logs: loss_final: 0.702618, loss_mean: 0.902564, proj_loss: -0.286791, loss_mean_cls: 0.086845, grad_norm: 0.592053 +Steps: 0%| | 809/1000000 [03:24<68:40:46, 4.04it/s, grad_norm=0.592, loss_final=0.703, loss_mean=0.903, loss_mean_cls=0.0868, proj_loss=-0.287][2026-03-26 12:34:27] Step: 809, Training Logs: loss_final: 0.675300, loss_mean: 0.871505, proj_loss: -0.285385, loss_mean_cls: 0.089180, grad_norm: 0.883590 +Steps: 0%| | 810/1000000 [03:24<68:42:27, 4.04it/s, grad_norm=0.884, loss_final=0.675, loss_mean=0.872, loss_mean_cls=0.0892, proj_loss=-0.285][2026-03-26 12:34:27] Step: 810, Training Logs: loss_final: 0.670174, loss_mean: 0.871016, proj_loss: -0.288668, loss_mean_cls: 0.087826, grad_norm: 0.566552 +Steps: 0%| | 811/1000000 [03:24<68:41:47, 4.04it/s, grad_norm=0.567, loss_final=0.67, loss_mean=0.871, loss_mean_cls=0.0878, proj_loss=-0.289][2026-03-26 12:34:27] Step: 811, Training Logs: loss_final: 0.689570, loss_mean: 0.890227, proj_loss: -0.288232, loss_mean_cls: 0.087575, grad_norm: 0.713904 +Steps: 0%| | 812/1000000 [03:24<68:43:52, 4.04it/s, grad_norm=0.714, loss_final=0.69, loss_mean=0.89, loss_mean_cls=0.0876, proj_loss=-0.288][2026-03-26 12:34:28] Step: 812, Training Logs: loss_final: 0.686783, loss_mean: 0.902717, proj_loss: -0.300116, loss_mean_cls: 0.084182, grad_norm: 0.468720 +Steps: 0%| | 813/1000000 [03:25<68:40:59, 4.04it/s, grad_norm=0.469, loss_final=0.687, loss_mean=0.903, loss_mean_cls=0.0842, proj_loss=-0.3][2026-03-26 12:34:28] Step: 813, Training Logs: loss_final: 0.675732, loss_mean: 0.876832, proj_loss: -0.287968, loss_mean_cls: 0.086868, grad_norm: 0.501045 +Steps: 0%| | 814/1000000 [03:25<68:41:26, 4.04it/s, grad_norm=0.501, loss_final=0.676, loss_mean=0.877, loss_mean_cls=0.0869, proj_loss=-0.288][2026-03-26 12:34:28] Step: 814, Training Logs: loss_final: 0.679593, loss_mean: 0.885562, proj_loss: -0.292772, loss_mean_cls: 0.086802, grad_norm: 0.486900 +Steps: 0%| | 815/1000000 [03:25<68:38:25, 4.04it/s, grad_norm=0.487, loss_final=0.68, loss_mean=0.886, loss_mean_cls=0.0868, proj_loss=-0.293][2026-03-26 12:34:28] Step: 815, Training Logs: loss_final: 0.673219, loss_mean: 0.882211, proj_loss: -0.295535, loss_mean_cls: 0.086542, grad_norm: 0.522326 +Steps: 0%| | 816/1000000 [03:25<68:37:41, 4.04it/s, grad_norm=0.522, loss_final=0.673, loss_mean=0.882, loss_mean_cls=0.0865, proj_loss=-0.296][2026-03-26 12:34:29] Step: 816, Training Logs: loss_final: 0.666005, loss_mean: 0.867068, proj_loss: -0.288645, loss_mean_cls: 0.087582, grad_norm: 0.509854 +Steps: 0%| | 817/1000000 [03:26<68:36:08, 4.05it/s, grad_norm=0.51, loss_final=0.666, loss_mean=0.867, loss_mean_cls=0.0876, proj_loss=-0.289][2026-03-26 12:34:29] Step: 817, Training Logs: loss_final: 0.675014, loss_mean: 0.874279, proj_loss: -0.288807, loss_mean_cls: 0.089541, grad_norm: 0.384232 +Steps: 0%| | 818/1000000 [03:26<68:38:24, 4.04it/s, grad_norm=0.384, loss_final=0.675, loss_mean=0.874, loss_mean_cls=0.0895, proj_loss=-0.289][2026-03-26 12:34:29] Step: 818, Training Logs: loss_final: 0.694740, loss_mean: 0.896994, proj_loss: -0.288698, loss_mean_cls: 0.086444, grad_norm: 0.768723 +Steps: 0%| | 819/1000000 [03:26<68:35:34, 4.05it/s, grad_norm=0.769, loss_final=0.695, loss_mean=0.897, loss_mean_cls=0.0864, proj_loss=-0.289][2026-03-26 12:34:29] Step: 819, Training Logs: loss_final: 0.682308, loss_mean: 0.886775, proj_loss: -0.291229, loss_mean_cls: 0.086761, grad_norm: 0.509716 +Steps: 0%| | 820/1000000 [03:26<68:37:29, 4.04it/s, grad_norm=0.51, loss_final=0.682, loss_mean=0.887, loss_mean_cls=0.0868, proj_loss=-0.291][2026-03-26 12:34:30] Step: 820, Training Logs: loss_final: 0.655889, loss_mean: 0.857687, proj_loss: -0.289420, loss_mean_cls: 0.087623, grad_norm: 0.612004 +Steps: 0%| | 821/1000000 [03:27<68:36:20, 4.05it/s, grad_norm=0.612, loss_final=0.656, loss_mean=0.858, loss_mean_cls=0.0876, proj_loss=-0.289][2026-03-26 12:34:30] Step: 821, Training Logs: loss_final: 0.677170, loss_mean: 0.880318, proj_loss: -0.290417, loss_mean_cls: 0.087269, grad_norm: 0.603841 +Steps: 0%| | 822/1000000 [03:27<68:39:13, 4.04it/s, grad_norm=0.604, loss_final=0.677, loss_mean=0.88, loss_mean_cls=0.0873, proj_loss=-0.29][2026-03-26 12:34:30] Step: 822, Training Logs: loss_final: 0.666269, loss_mean: 0.876607, proj_loss: -0.297382, loss_mean_cls: 0.087045, grad_norm: 0.598083 +Steps: 0%| | 823/1000000 [03:27<68:42:51, 4.04it/s, grad_norm=0.598, loss_final=0.666, loss_mean=0.877, loss_mean_cls=0.087, proj_loss=-0.297][2026-03-26 12:34:30] Step: 823, Training Logs: loss_final: 0.672466, loss_mean: 0.893118, proj_loss: -0.305579, loss_mean_cls: 0.084927, grad_norm: 0.625404 +Steps: 0%| | 824/1000000 [03:27<68:45:18, 4.04it/s, grad_norm=0.625, loss_final=0.672, loss_mean=0.893, loss_mean_cls=0.0849, proj_loss=-0.306][2026-03-26 12:34:31] Step: 824, Training Logs: loss_final: 0.689347, loss_mean: 0.899322, proj_loss: -0.296627, loss_mean_cls: 0.086652, grad_norm: 0.726622 +Steps: 0%| | 825/1000000 [03:28<68:42:49, 4.04it/s, grad_norm=0.727, loss_final=0.689, loss_mean=0.899, loss_mean_cls=0.0867, proj_loss=-0.297][2026-03-26 12:34:31] Step: 825, Training Logs: loss_final: 0.676620, loss_mean: 0.870298, proj_loss: -0.282998, loss_mean_cls: 0.089321, grad_norm: 0.677764 +Steps: 0%| | 826/1000000 [03:28<68:40:30, 4.04it/s, grad_norm=0.678, loss_final=0.677, loss_mean=0.87, loss_mean_cls=0.0893, proj_loss=-0.283][2026-03-26 12:34:31] Step: 826, Training Logs: loss_final: 0.685561, loss_mean: 0.904031, proj_loss: -0.303699, loss_mean_cls: 0.085229, grad_norm: 0.856169 +Steps: 0%| | 827/1000000 [03:28<68:40:14, 4.04it/s, grad_norm=0.856, loss_final=0.686, loss_mean=0.904, loss_mean_cls=0.0852, proj_loss=-0.304][2026-03-26 12:34:31] Step: 827, Training Logs: loss_final: 0.687746, loss_mean: 0.907184, proj_loss: -0.303705, loss_mean_cls: 0.084267, grad_norm: 0.533932 +Steps: 0%| | 828/1000000 [03:28<68:42:03, 4.04it/s, grad_norm=0.534, loss_final=0.688, loss_mean=0.907, loss_mean_cls=0.0843, proj_loss=-0.304][2026-03-26 12:34:32] Step: 828, Training Logs: loss_final: 0.673619, loss_mean: 0.879734, proj_loss: -0.294122, loss_mean_cls: 0.088008, grad_norm: 0.369978 +Steps: 0%| | 829/1000000 [03:29<68:42:50, 4.04it/s, grad_norm=0.37, loss_final=0.674, loss_mean=0.88, loss_mean_cls=0.088, proj_loss=-0.294][2026-03-26 12:34:32] Step: 829, Training Logs: loss_final: 0.670558, loss_mean: 0.888023, proj_loss: -0.303372, loss_mean_cls: 0.085907, grad_norm: 0.637574 +Steps: 0%| | 830/1000000 [03:29<68:42:51, 4.04it/s, grad_norm=0.638, loss_final=0.671, loss_mean=0.888, loss_mean_cls=0.0859, proj_loss=-0.303][2026-03-26 12:34:32] Step: 830, Training Logs: loss_final: 0.671516, loss_mean: 0.882031, proj_loss: -0.297434, loss_mean_cls: 0.086919, grad_norm: 0.551268 +Steps: 0%| | 831/1000000 [03:29<68:42:37, 4.04it/s, grad_norm=0.551, loss_final=0.672, loss_mean=0.882, loss_mean_cls=0.0869, proj_loss=-0.297][2026-03-26 12:34:32] Step: 831, Training Logs: loss_final: 0.665285, loss_mean: 0.879320, proj_loss: -0.300477, loss_mean_cls: 0.086442, grad_norm: 0.404501 +Steps: 0%| | 832/1000000 [03:29<68:41:35, 4.04it/s, grad_norm=0.405, loss_final=0.665, loss_mean=0.879, loss_mean_cls=0.0864, proj_loss=-0.3][2026-03-26 12:34:33] Step: 832, Training Logs: loss_final: 0.683866, loss_mean: 0.906703, proj_loss: -0.307296, loss_mean_cls: 0.084459, grad_norm: 0.528358 +Steps: 0%| | 833/1000000 [03:30<68:41:27, 4.04it/s, grad_norm=0.528, loss_final=0.684, loss_mean=0.907, loss_mean_cls=0.0845, proj_loss=-0.307][2026-03-26 12:34:33] Step: 833, Training Logs: loss_final: 0.661610, loss_mean: 0.872841, proj_loss: -0.297533, loss_mean_cls: 0.086302, grad_norm: 0.392951 +Steps: 0%| | 834/1000000 [03:30<68:39:40, 4.04it/s, grad_norm=0.393, loss_final=0.662, loss_mean=0.873, loss_mean_cls=0.0863, proj_loss=-0.298][2026-03-26 12:34:33] Step: 834, Training Logs: loss_final: 0.691407, loss_mean: 0.897703, proj_loss: -0.292174, loss_mean_cls: 0.085878, grad_norm: 0.714978 +Steps: 0%| | 835/1000000 [03:30<68:38:47, 4.04it/s, grad_norm=0.715, loss_final=0.691, loss_mean=0.898, loss_mean_cls=0.0859, proj_loss=-0.292][2026-03-26 12:34:33] Step: 835, Training Logs: loss_final: 0.668832, loss_mean: 0.891745, proj_loss: -0.307875, loss_mean_cls: 0.084962, grad_norm: 0.594982 +Steps: 0%| | 836/1000000 [03:30<68:39:03, 4.04it/s, grad_norm=0.595, loss_final=0.669, loss_mean=0.892, loss_mean_cls=0.085, proj_loss=-0.308][2026-03-26 12:34:34] Step: 836, Training Logs: loss_final: 0.682286, loss_mean: 0.888091, proj_loss: -0.293754, loss_mean_cls: 0.087949, grad_norm: 0.474721 +Steps: 0%| | 837/1000000 [03:31<68:39:38, 4.04it/s, grad_norm=0.475, loss_final=0.682, loss_mean=0.888, loss_mean_cls=0.0879, proj_loss=-0.294][2026-03-26 12:34:34] Step: 837, Training Logs: loss_final: 0.668747, loss_mean: 0.867482, proj_loss: -0.287865, loss_mean_cls: 0.089129, grad_norm: 0.471385 +Steps: 0%| | 838/1000000 [03:31<68:38:35, 4.04it/s, grad_norm=0.471, loss_final=0.669, loss_mean=0.867, loss_mean_cls=0.0891, proj_loss=-0.288][2026-03-26 12:34:34] Step: 838, Training Logs: loss_final: 0.687896, loss_mean: 0.900620, proj_loss: -0.298549, loss_mean_cls: 0.085825, grad_norm: 0.577949 +Steps: 0%| | 839/1000000 [03:31<68:39:08, 4.04it/s, grad_norm=0.578, loss_final=0.688, loss_mean=0.901, loss_mean_cls=0.0858, proj_loss=-0.299][2026-03-26 12:34:34] Step: 839, Training Logs: loss_final: 0.666764, loss_mean: 0.888462, proj_loss: -0.306619, loss_mean_cls: 0.084921, grad_norm: 0.609036 +Steps: 0%| | 840/1000000 [03:31<68:40:17, 4.04it/s, grad_norm=0.609, loss_final=0.667, loss_mean=0.888, loss_mean_cls=0.0849, proj_loss=-0.307][2026-03-26 12:34:35] Step: 840, Training Logs: loss_final: 0.661018, loss_mean: 0.880182, proj_loss: -0.305529, loss_mean_cls: 0.086365, grad_norm: 0.559662 +Steps: 0%| | 841/1000000 [03:32<68:42:16, 4.04it/s, grad_norm=0.56, loss_final=0.661, loss_mean=0.88, loss_mean_cls=0.0864, proj_loss=-0.306][2026-03-26 12:34:35] Step: 841, Training Logs: loss_final: 0.669058, loss_mean: 0.888409, proj_loss: -0.304593, loss_mean_cls: 0.085243, grad_norm: 0.709872 +Steps: 0%| | 842/1000000 [03:32<68:42:31, 4.04it/s, grad_norm=0.71, loss_final=0.669, loss_mean=0.888, loss_mean_cls=0.0852, proj_loss=-0.305][2026-03-26 12:34:35] Step: 842, Training Logs: loss_final: 0.650676, loss_mean: 0.869464, proj_loss: -0.304804, loss_mean_cls: 0.086016, grad_norm: 0.481040 +Steps: 0%| | 843/1000000 [03:32<68:41:13, 4.04it/s, grad_norm=0.481, loss_final=0.651, loss_mean=0.869, loss_mean_cls=0.086, proj_loss=-0.305][2026-03-26 12:34:35] Step: 843, Training Logs: loss_final: 0.669746, loss_mean: 0.887623, proj_loss: -0.304624, loss_mean_cls: 0.086747, grad_norm: 0.962730 +Steps: 0%| | 844/1000000 [03:32<68:41:27, 4.04it/s, grad_norm=0.963, loss_final=0.67, loss_mean=0.888, loss_mean_cls=0.0867, proj_loss=-0.305][2026-03-26 12:34:36] Step: 844, Training Logs: loss_final: 0.669313, loss_mean: 0.891693, proj_loss: -0.307814, loss_mean_cls: 0.085434, grad_norm: 0.589665 +Steps: 0%| | 845/1000000 [03:33<68:41:06, 4.04it/s, grad_norm=0.59, loss_final=0.669, loss_mean=0.892, loss_mean_cls=0.0854, proj_loss=-0.308][2026-03-26 12:34:36] Step: 845, Training Logs: loss_final: 0.671705, loss_mean: 0.884296, proj_loss: -0.300506, loss_mean_cls: 0.087915, grad_norm: 0.774365 +Steps: 0%| | 846/1000000 [03:33<68:41:02, 4.04it/s, grad_norm=0.774, loss_final=0.672, loss_mean=0.884, loss_mean_cls=0.0879, proj_loss=-0.301][2026-03-26 12:34:36] Step: 846, Training Logs: loss_final: 0.676356, loss_mean: 0.887264, proj_loss: -0.298273, loss_mean_cls: 0.087365, grad_norm: 0.783990 +Steps: 0%| | 847/1000000 [03:33<68:42:35, 4.04it/s, grad_norm=0.784, loss_final=0.676, loss_mean=0.887, loss_mean_cls=0.0874, proj_loss=-0.298][2026-03-26 12:34:36] Step: 847, Training Logs: loss_final: 0.678277, loss_mean: 0.897891, proj_loss: -0.306315, loss_mean_cls: 0.086701, grad_norm: 0.868286 +Steps: 0%| | 848/1000000 [03:33<68:58:50, 4.02it/s, grad_norm=0.868, loss_final=0.678, loss_mean=0.898, loss_mean_cls=0.0867, proj_loss=-0.306][2026-03-26 12:34:37] Step: 848, Training Logs: loss_final: 0.685910, loss_mean: 0.897680, proj_loss: -0.298036, loss_mean_cls: 0.086266, grad_norm: 0.768985 +Steps: 0%| | 849/1000000 [03:34<68:52:35, 4.03it/s, grad_norm=0.769, loss_final=0.686, loss_mean=0.898, loss_mean_cls=0.0863, proj_loss=-0.298][2026-03-26 12:34:37] Step: 849, Training Logs: loss_final: 0.661806, loss_mean: 0.878699, proj_loss: -0.302254, loss_mean_cls: 0.085361, grad_norm: 0.631726 +Steps: 0%| | 850/1000000 [03:34<68:49:08, 4.03it/s, grad_norm=0.632, loss_final=0.662, loss_mean=0.879, loss_mean_cls=0.0854, proj_loss=-0.302][2026-03-26 12:34:37] Step: 850, Training Logs: loss_final: 0.669798, loss_mean: 0.876855, proj_loss: -0.294526, loss_mean_cls: 0.087470, grad_norm: 1.108143 +Steps: 0%| | 851/1000000 [03:34<68:46:14, 4.04it/s, grad_norm=1.11, loss_final=0.67, loss_mean=0.877, loss_mean_cls=0.0875, proj_loss=-0.295][2026-03-26 12:34:37] Step: 851, Training Logs: loss_final: 0.687041, loss_mean: 0.903670, proj_loss: -0.301560, loss_mean_cls: 0.084930, grad_norm: 0.748580 +Steps: 0%| | 852/1000000 [03:34<68:44:05, 4.04it/s, grad_norm=0.749, loss_final=0.687, loss_mean=0.904, loss_mean_cls=0.0849, proj_loss=-0.302][2026-03-26 12:34:38] Step: 852, Training Logs: loss_final: 0.686787, loss_mean: 0.900746, proj_loss: -0.300232, loss_mean_cls: 0.086272, grad_norm: 0.855792 +Steps: 0%| | 853/1000000 [03:35<68:43:24, 4.04it/s, grad_norm=0.856, loss_final=0.687, loss_mean=0.901, loss_mean_cls=0.0863, proj_loss=-0.3][2026-03-26 12:34:38] Step: 853, Training Logs: loss_final: 0.659522, loss_mean: 0.870284, proj_loss: -0.299698, loss_mean_cls: 0.088936, grad_norm: 0.662395 +Steps: 0%| | 854/1000000 [03:35<68:41:58, 4.04it/s, grad_norm=0.662, loss_final=0.66, loss_mean=0.87, loss_mean_cls=0.0889, proj_loss=-0.3][2026-03-26 12:34:38] Step: 854, Training Logs: loss_final: 0.688260, loss_mean: 0.898787, proj_loss: -0.297346, loss_mean_cls: 0.086819, grad_norm: 0.775748 +Steps: 0%| | 855/1000000 [03:35<68:44:53, 4.04it/s, grad_norm=0.776, loss_final=0.688, loss_mean=0.899, loss_mean_cls=0.0868, proj_loss=-0.297][2026-03-26 12:34:38] Step: 855, Training Logs: loss_final: 0.671151, loss_mean: 0.894073, proj_loss: -0.308706, loss_mean_cls: 0.085784, grad_norm: 0.544061 +Steps: 0%| | 856/1000000 [03:35<68:46:34, 4.04it/s, grad_norm=0.544, loss_final=0.671, loss_mean=0.894, loss_mean_cls=0.0858, proj_loss=-0.309][2026-03-26 12:34:39] Step: 856, Training Logs: loss_final: 0.670350, loss_mean: 0.874133, proj_loss: -0.291902, loss_mean_cls: 0.088119, grad_norm: 0.930213 +Steps: 0%| | 857/1000000 [03:35<68:43:50, 4.04it/s, grad_norm=0.93, loss_final=0.67, loss_mean=0.874, loss_mean_cls=0.0881, proj_loss=-0.292][2026-03-26 12:34:39] Step: 857, Training Logs: loss_final: 0.647555, loss_mean: 0.863314, proj_loss: -0.302676, loss_mean_cls: 0.086917, grad_norm: 0.636457 +Steps: 0%| | 858/1000000 [03:36<68:51:19, 4.03it/s, grad_norm=0.636, loss_final=0.648, loss_mean=0.863, loss_mean_cls=0.0869, proj_loss=-0.303][2026-03-26 12:34:39] Step: 858, Training Logs: loss_final: 0.677033, loss_mean: 0.870977, proj_loss: -0.283552, loss_mean_cls: 0.089608, grad_norm: 0.907993 +Steps: 0%| | 859/1000000 [03:36<68:48:14, 4.03it/s, grad_norm=0.908, loss_final=0.677, loss_mean=0.871, loss_mean_cls=0.0896, proj_loss=-0.284][2026-03-26 12:34:39] Step: 859, Training Logs: loss_final: 0.662331, loss_mean: 0.870667, proj_loss: -0.296583, loss_mean_cls: 0.088247, grad_norm: 0.759349 +Steps: 0%| | 860/1000000 [03:36<69:56:06, 3.97it/s, grad_norm=0.759, loss_final=0.662, loss_mean=0.871, loss_mean_cls=0.0882, proj_loss=-0.297][2026-03-26 12:34:40] Step: 860, Training Logs: loss_final: 0.666205, loss_mean: 0.881708, proj_loss: -0.302094, loss_mean_cls: 0.086591, grad_norm: 0.695152 +Steps: 0%| | 861/1000000 [03:37<70:28:28, 3.94it/s, grad_norm=0.695, loss_final=0.666, loss_mean=0.882, loss_mean_cls=0.0866, proj_loss=-0.302][2026-03-26 12:34:40] Step: 861, Training Logs: loss_final: 0.671133, loss_mean: 0.880320, proj_loss: -0.296341, loss_mean_cls: 0.087154, grad_norm: 0.666873 +Steps: 0%| | 862/1000000 [03:37<69:56:31, 3.97it/s, grad_norm=0.667, loss_final=0.671, loss_mean=0.88, loss_mean_cls=0.0872, proj_loss=-0.296][2026-03-26 12:34:40] Step: 862, Training Logs: loss_final: 0.670292, loss_mean: 0.890439, proj_loss: -0.305499, loss_mean_cls: 0.085351, grad_norm: 0.411981 +Steps: 0%| | 863/1000000 [03:37<69:31:35, 3.99it/s, grad_norm=0.412, loss_final=0.67, loss_mean=0.89, loss_mean_cls=0.0854, proj_loss=-0.305][2026-03-26 12:34:40] Step: 863, Training Logs: loss_final: 0.669662, loss_mean: 0.873364, proj_loss: -0.292155, loss_mean_cls: 0.088453, grad_norm: 0.730683 +Steps: 0%| | 864/1000000 [03:37<69:14:34, 4.01it/s, grad_norm=0.731, loss_final=0.67, loss_mean=0.873, loss_mean_cls=0.0885, proj_loss=-0.292][2026-03-26 12:34:41] Step: 864, Training Logs: loss_final: 0.663773, loss_mean: 0.883694, proj_loss: -0.304713, loss_mean_cls: 0.084792, grad_norm: 0.524396 +Steps: 0%| | 865/1000000 [03:37<69:05:06, 4.02it/s, grad_norm=0.524, loss_final=0.664, loss_mean=0.884, loss_mean_cls=0.0848, proj_loss=-0.305][2026-03-26 12:34:41] Step: 865, Training Logs: loss_final: 0.679746, loss_mean: 0.898702, proj_loss: -0.304717, loss_mean_cls: 0.085761, grad_norm: 0.600264 +Steps: 0%| | 866/1000000 [03:38<68:55:44, 4.03it/s, grad_norm=0.6, loss_final=0.68, loss_mean=0.899, loss_mean_cls=0.0858, proj_loss=-0.305][2026-03-26 12:34:41] Step: 866, Training Logs: loss_final: 0.672702, loss_mean: 0.889611, proj_loss: -0.302992, loss_mean_cls: 0.086083, grad_norm: 0.508195 +Steps: 0%| | 867/1000000 [03:38<68:47:48, 4.03it/s, grad_norm=0.508, loss_final=0.673, loss_mean=0.89, loss_mean_cls=0.0861, proj_loss=-0.303][2026-03-26 12:34:41] Step: 867, Training Logs: loss_final: 0.661505, loss_mean: 0.885510, proj_loss: -0.309367, loss_mean_cls: 0.085362, grad_norm: 0.571539 +Steps: 0%| | 868/1000000 [03:38<68:43:45, 4.04it/s, grad_norm=0.572, loss_final=0.662, loss_mean=0.886, loss_mean_cls=0.0854, proj_loss=-0.309][2026-03-26 12:34:42] Step: 868, Training Logs: loss_final: 0.655129, loss_mean: 0.864437, proj_loss: -0.297375, loss_mean_cls: 0.088067, grad_norm: 0.575331 +Steps: 0%| | 869/1000000 [03:38<68:40:42, 4.04it/s, grad_norm=0.575, loss_final=0.655, loss_mean=0.864, loss_mean_cls=0.0881, proj_loss=-0.297][2026-03-26 12:34:42] Step: 869, Training Logs: loss_final: 0.670051, loss_mean: 0.867793, proj_loss: -0.287654, loss_mean_cls: 0.089913, grad_norm: 0.651023 +Steps: 0%| | 870/1000000 [03:39<68:40:59, 4.04it/s, grad_norm=0.651, loss_final=0.67, loss_mean=0.868, loss_mean_cls=0.0899, proj_loss=-0.288][2026-03-26 12:34:42] Step: 870, Training Logs: loss_final: 0.663452, loss_mean: 0.884917, proj_loss: -0.308208, loss_mean_cls: 0.086744, grad_norm: 0.924023 +Steps: 0%| | 871/1000000 [03:39<68:39:30, 4.04it/s, grad_norm=0.924, loss_final=0.663, loss_mean=0.885, loss_mean_cls=0.0867, proj_loss=-0.308][2026-03-26 12:34:42] Step: 871, Training Logs: loss_final: 0.670941, loss_mean: 0.893909, proj_loss: -0.309181, loss_mean_cls: 0.086213, grad_norm: 0.813729 +Steps: 0%| | 872/1000000 [03:39<68:39:29, 4.04it/s, grad_norm=0.814, loss_final=0.671, loss_mean=0.894, loss_mean_cls=0.0862, proj_loss=-0.309][2026-03-26 12:34:43] Step: 872, Training Logs: loss_final: 0.669458, loss_mean: 0.893236, proj_loss: -0.309508, loss_mean_cls: 0.085731, grad_norm: 0.568562 +Steps: 0%| | 873/1000000 [03:39<68:40:13, 4.04it/s, grad_norm=0.569, loss_final=0.669, loss_mean=0.893, loss_mean_cls=0.0857, proj_loss=-0.31][2026-03-26 12:34:43] Step: 873, Training Logs: loss_final: 0.666913, loss_mean: 0.892326, proj_loss: -0.310709, loss_mean_cls: 0.085295, grad_norm: 0.780698 +Steps: 0%| | 874/1000000 [03:40<68:44:10, 4.04it/s, grad_norm=0.781, loss_final=0.667, loss_mean=0.892, loss_mean_cls=0.0853, proj_loss=-0.311][2026-03-26 12:34:43] Step: 874, Training Logs: loss_final: 0.673195, loss_mean: 0.889844, proj_loss: -0.303891, loss_mean_cls: 0.087242, grad_norm: 0.524510 +Steps: 0%| | 875/1000000 [03:40<68:41:17, 4.04it/s, grad_norm=0.525, loss_final=0.673, loss_mean=0.89, loss_mean_cls=0.0872, proj_loss=-0.304][2026-03-26 12:34:43] Step: 875, Training Logs: loss_final: 0.671098, loss_mean: 0.881361, proj_loss: -0.297750, loss_mean_cls: 0.087487, grad_norm: 0.741587 +Steps: 0%| | 876/1000000 [03:40<68:40:56, 4.04it/s, grad_norm=0.742, loss_final=0.671, loss_mean=0.881, loss_mean_cls=0.0875, proj_loss=-0.298][2026-03-26 12:34:44] Step: 876, Training Logs: loss_final: 0.675719, loss_mean: 0.890097, proj_loss: -0.302064, loss_mean_cls: 0.087687, grad_norm: 0.612031 +Steps: 0%| | 877/1000000 [03:40<68:39:27, 4.04it/s, grad_norm=0.612, loss_final=0.676, loss_mean=0.89, loss_mean_cls=0.0877, proj_loss=-0.302][2026-03-26 12:34:44] Step: 877, Training Logs: loss_final: 0.648913, loss_mean: 0.884858, proj_loss: -0.320577, loss_mean_cls: 0.084633, grad_norm: 0.689144 +Steps: 0%| | 878/1000000 [03:41<68:40:48, 4.04it/s, grad_norm=0.689, loss_final=0.649, loss_mean=0.885, loss_mean_cls=0.0846, proj_loss=-0.321][2026-03-26 12:34:44] Step: 878, Training Logs: loss_final: 0.674273, loss_mean: 0.888351, proj_loss: -0.299850, loss_mean_cls: 0.085771, grad_norm: 0.759476 +Steps: 0%| | 879/1000000 [03:41<68:40:16, 4.04it/s, grad_norm=0.759, loss_final=0.674, loss_mean=0.888, loss_mean_cls=0.0858, proj_loss=-0.3][2026-03-26 12:34:44] Step: 879, Training Logs: loss_final: 0.670037, loss_mean: 0.892712, proj_loss: -0.309017, loss_mean_cls: 0.086342, grad_norm: 0.825466 +Steps: 0%| | 880/1000000 [03:41<68:39:54, 4.04it/s, grad_norm=0.825, loss_final=0.67, loss_mean=0.893, loss_mean_cls=0.0863, proj_loss=-0.309][2026-03-26 12:34:45] Step: 880, Training Logs: loss_final: 0.679185, loss_mean: 0.889874, proj_loss: -0.298521, loss_mean_cls: 0.087831, grad_norm: 1.006853 +Steps: 0%| | 881/1000000 [03:41<68:40:57, 4.04it/s, grad_norm=1.01, loss_final=0.679, loss_mean=0.89, loss_mean_cls=0.0878, proj_loss=-0.299][2026-03-26 12:34:45] Step: 881, Training Logs: loss_final: 0.677157, loss_mean: 0.897534, proj_loss: -0.305457, loss_mean_cls: 0.085080, grad_norm: 0.956432 +Steps: 0%| | 882/1000000 [03:42<68:41:00, 4.04it/s, grad_norm=0.956, loss_final=0.677, loss_mean=0.898, loss_mean_cls=0.0851, proj_loss=-0.305][2026-03-26 12:34:45] Step: 882, Training Logs: loss_final: 0.652976, loss_mean: 0.868968, proj_loss: -0.304039, loss_mean_cls: 0.088047, grad_norm: 0.930905 +Steps: 0%| | 883/1000000 [03:42<68:50:32, 4.03it/s, grad_norm=0.931, loss_final=0.653, loss_mean=0.869, loss_mean_cls=0.088, proj_loss=-0.304][2026-03-26 12:34:45] Step: 883, Training Logs: loss_final: 0.655235, loss_mean: 0.875426, proj_loss: -0.306106, loss_mean_cls: 0.085915, grad_norm: 0.944779 +Steps: 0%| | 884/1000000 [03:42<68:48:13, 4.03it/s, grad_norm=0.945, loss_final=0.655, loss_mean=0.875, loss_mean_cls=0.0859, proj_loss=-0.306][2026-03-26 12:34:46] Step: 884, Training Logs: loss_final: 0.682106, loss_mean: 0.898502, proj_loss: -0.303748, loss_mean_cls: 0.087353, grad_norm: 0.968919 +Steps: 0%| | 885/1000000 [03:42<68:47:06, 4.03it/s, grad_norm=0.969, loss_final=0.682, loss_mean=0.899, loss_mean_cls=0.0874, proj_loss=-0.304][2026-03-26 12:34:46] Step: 885, Training Logs: loss_final: 0.653675, loss_mean: 0.875088, proj_loss: -0.307361, loss_mean_cls: 0.085948, grad_norm: 0.438239 +Steps: 0%| | 886/1000000 [03:43<68:44:09, 4.04it/s, grad_norm=0.438, loss_final=0.654, loss_mean=0.875, loss_mean_cls=0.0859, proj_loss=-0.307][2026-03-26 12:34:46] Step: 886, Training Logs: loss_final: 0.644918, loss_mean: 0.855227, proj_loss: -0.298657, loss_mean_cls: 0.088349, grad_norm: 0.771842 +Steps: 0%| | 887/1000000 [03:43<68:44:41, 4.04it/s, grad_norm=0.772, loss_final=0.645, loss_mean=0.855, loss_mean_cls=0.0883, proj_loss=-0.299][2026-03-26 12:34:46] Step: 887, Training Logs: loss_final: 0.661361, loss_mean: 0.894776, proj_loss: -0.317441, loss_mean_cls: 0.084027, grad_norm: 0.877615 +Steps: 0%| | 888/1000000 [03:43<68:40:44, 4.04it/s, grad_norm=0.878, loss_final=0.661, loss_mean=0.895, loss_mean_cls=0.084, proj_loss=-0.317][2026-03-26 12:34:47] Step: 888, Training Logs: loss_final: 0.661006, loss_mean: 0.874442, proj_loss: -0.301737, loss_mean_cls: 0.088301, grad_norm: 0.895828 +Steps: 0%| | 889/1000000 [03:43<68:47:11, 4.03it/s, grad_norm=0.896, loss_final=0.661, loss_mean=0.874, loss_mean_cls=0.0883, proj_loss=-0.302][2026-03-26 12:34:47] Step: 889, Training Logs: loss_final: 0.668471, loss_mean: 0.894128, proj_loss: -0.310637, loss_mean_cls: 0.084980, grad_norm: 0.564807 +Steps: 0%| | 890/1000000 [03:44<68:35:58, 4.05it/s, grad_norm=0.565, loss_final=0.668, loss_mean=0.894, loss_mean_cls=0.085, proj_loss=-0.311][2026-03-26 12:34:47] Step: 890, Training Logs: loss_final: 0.673221, loss_mean: 0.898994, proj_loss: -0.311889, loss_mean_cls: 0.086116, grad_norm: 1.033971 +Steps: 0%| | 891/1000000 [03:44<68:38:36, 4.04it/s, grad_norm=1.03, loss_final=0.673, loss_mean=0.899, loss_mean_cls=0.0861, proj_loss=-0.312][2026-03-26 12:34:47] Step: 891, Training Logs: loss_final: 0.668003, loss_mean: 0.893915, proj_loss: -0.312133, loss_mean_cls: 0.086221, grad_norm: 0.866754 +Steps: 0%| | 892/1000000 [03:44<68:38:44, 4.04it/s, grad_norm=0.867, loss_final=0.668, loss_mean=0.894, loss_mean_cls=0.0862, proj_loss=-0.312][2026-03-26 12:34:48] Step: 892, Training Logs: loss_final: 0.685314, loss_mean: 0.908452, proj_loss: -0.307958, loss_mean_cls: 0.084820, grad_norm: 0.609705 +Steps: 0%| | 893/1000000 [03:44<68:39:30, 4.04it/s, grad_norm=0.61, loss_final=0.685, loss_mean=0.908, loss_mean_cls=0.0848, proj_loss=-0.308][2026-03-26 12:34:48] Step: 893, Training Logs: loss_final: 0.662438, loss_mean: 0.891991, proj_loss: -0.314589, loss_mean_cls: 0.085036, grad_norm: 0.739784 +Steps: 0%| | 894/1000000 [03:45<68:36:05, 4.05it/s, grad_norm=0.74, loss_final=0.662, loss_mean=0.892, loss_mean_cls=0.085, proj_loss=-0.315][2026-03-26 12:34:48] Step: 894, Training Logs: loss_final: 0.665497, loss_mean: 0.875265, proj_loss: -0.297480, loss_mean_cls: 0.087713, grad_norm: 0.677758 +Steps: 0%| | 895/1000000 [03:45<68:36:47, 4.04it/s, grad_norm=0.678, loss_final=0.665, loss_mean=0.875, loss_mean_cls=0.0877, proj_loss=-0.297][2026-03-26 12:34:48] Step: 895, Training Logs: loss_final: 0.674893, loss_mean: 0.894232, proj_loss: -0.305651, loss_mean_cls: 0.086312, grad_norm: 0.487635 +Steps: 0%| | 896/1000000 [03:45<68:37:32, 4.04it/s, grad_norm=0.488, loss_final=0.675, loss_mean=0.894, loss_mean_cls=0.0863, proj_loss=-0.306][2026-03-26 12:34:49] Step: 896, Training Logs: loss_final: 0.654372, loss_mean: 0.888404, proj_loss: -0.319051, loss_mean_cls: 0.085019, grad_norm: 0.694589 +Steps: 0%| | 897/1000000 [03:45<68:38:09, 4.04it/s, grad_norm=0.695, loss_final=0.654, loss_mean=0.888, loss_mean_cls=0.085, proj_loss=-0.319][2026-03-26 12:34:49] Step: 897, Training Logs: loss_final: 0.652977, loss_mean: 0.880426, proj_loss: -0.313232, loss_mean_cls: 0.085783, grad_norm: 0.643873 +Steps: 0%| | 898/1000000 [03:46<68:40:07, 4.04it/s, grad_norm=0.644, loss_final=0.653, loss_mean=0.88, loss_mean_cls=0.0858, proj_loss=-0.313][2026-03-26 12:34:49] Step: 898, Training Logs: loss_final: 0.657325, loss_mean: 0.886831, proj_loss: -0.315577, loss_mean_cls: 0.086071, grad_norm: 0.565621 +Steps: 0%| | 899/1000000 [03:46<68:38:26, 4.04it/s, grad_norm=0.566, loss_final=0.657, loss_mean=0.887, loss_mean_cls=0.0861, proj_loss=-0.316][2026-03-26 12:34:49] Step: 899, Training Logs: loss_final: 0.661518, loss_mean: 0.898394, proj_loss: -0.321884, loss_mean_cls: 0.085008, grad_norm: 0.673891 +Steps: 0%| | 900/1000000 [03:46<68:40:07, 4.04it/s, grad_norm=0.674, loss_final=0.662, loss_mean=0.898, loss_mean_cls=0.085, proj_loss=-0.322][2026-03-26 12:34:50] Step: 900, Training Logs: loss_final: 0.652118, loss_mean: 0.886932, proj_loss: -0.318476, loss_mean_cls: 0.083663, grad_norm: 0.718444 +Steps: 0%| | 901/1000000 [03:46<68:40:59, 4.04it/s, grad_norm=0.718, loss_final=0.652, loss_mean=0.887, loss_mean_cls=0.0837, proj_loss=-0.318][2026-03-26 12:34:50] Step: 901, Training Logs: loss_final: 0.664959, loss_mean: 0.893443, proj_loss: -0.314379, loss_mean_cls: 0.085895, grad_norm: 0.485548 +Steps: 0%| | 902/1000000 [03:47<68:43:20, 4.04it/s, grad_norm=0.486, loss_final=0.665, loss_mean=0.893, loss_mean_cls=0.0859, proj_loss=-0.314][2026-03-26 12:34:50] Step: 902, Training Logs: loss_final: 0.652597, loss_mean: 0.869490, proj_loss: -0.303310, loss_mean_cls: 0.086417, grad_norm: 0.487294 +Steps: 0%| | 903/1000000 [03:47<68:43:55, 4.04it/s, grad_norm=0.487, loss_final=0.653, loss_mean=0.869, loss_mean_cls=0.0864, proj_loss=-0.303][2026-03-26 12:34:50] Step: 903, Training Logs: loss_final: 0.652201, loss_mean: 0.864061, proj_loss: -0.301281, loss_mean_cls: 0.089421, grad_norm: 0.697277 +Steps: 0%| | 904/1000000 [03:47<68:43:36, 4.04it/s, grad_norm=0.697, loss_final=0.652, loss_mean=0.864, loss_mean_cls=0.0894, proj_loss=-0.301][2026-03-26 12:34:51] Step: 904, Training Logs: loss_final: 0.664743, loss_mean: 0.879818, proj_loss: -0.303110, loss_mean_cls: 0.088035, grad_norm: 0.458665 +Steps: 0%| | 905/1000000 [03:47<68:39:38, 4.04it/s, grad_norm=0.459, loss_final=0.665, loss_mean=0.88, loss_mean_cls=0.088, proj_loss=-0.303][2026-03-26 12:34:51] Step: 905, Training Logs: loss_final: 0.658730, loss_mean: 0.904056, proj_loss: -0.328049, loss_mean_cls: 0.082722, grad_norm: 0.588495 +Steps: 0%| | 906/1000000 [03:48<68:41:15, 4.04it/s, grad_norm=0.588, loss_final=0.659, loss_mean=0.904, loss_mean_cls=0.0827, proj_loss=-0.328][2026-03-26 12:34:51] Step: 906, Training Logs: loss_final: 0.658806, loss_mean: 0.892914, proj_loss: -0.318439, loss_mean_cls: 0.084331, grad_norm: 0.392986 +Steps: 0%| | 907/1000000 [03:48<68:42:18, 4.04it/s, grad_norm=0.393, loss_final=0.659, loss_mean=0.893, loss_mean_cls=0.0843, proj_loss=-0.318][2026-03-26 12:34:51] Step: 907, Training Logs: loss_final: 0.663946, loss_mean: 0.878386, proj_loss: -0.301387, loss_mean_cls: 0.086947, grad_norm: 0.512175 +Steps: 0%| | 908/1000000 [03:48<68:45:14, 4.04it/s, grad_norm=0.512, loss_final=0.664, loss_mean=0.878, loss_mean_cls=0.0869, proj_loss=-0.301][2026-03-26 12:34:52] Step: 908, Training Logs: loss_final: 0.653349, loss_mean: 0.866260, proj_loss: -0.301637, loss_mean_cls: 0.088726, grad_norm: 0.658341 +Steps: 0%| | 909/1000000 [03:48<68:41:09, 4.04it/s, grad_norm=0.658, loss_final=0.653, loss_mean=0.866, loss_mean_cls=0.0887, proj_loss=-0.302][2026-03-26 12:34:52] Step: 909, Training Logs: loss_final: 0.654645, loss_mean: 0.875533, proj_loss: -0.307577, loss_mean_cls: 0.086689, grad_norm: 0.616088 +Steps: 0%| | 910/1000000 [03:49<68:42:37, 4.04it/s, grad_norm=0.616, loss_final=0.655, loss_mean=0.876, loss_mean_cls=0.0867, proj_loss=-0.308][2026-03-26 12:34:52] Step: 910, Training Logs: loss_final: 0.660523, loss_mean: 0.887948, proj_loss: -0.312470, loss_mean_cls: 0.085045, grad_norm: 0.845319 +Steps: 0%| | 911/1000000 [03:49<68:40:59, 4.04it/s, grad_norm=0.845, loss_final=0.661, loss_mean=0.888, loss_mean_cls=0.085, proj_loss=-0.312][2026-03-26 12:34:52] Step: 911, Training Logs: loss_final: 0.656856, loss_mean: 0.883067, proj_loss: -0.312149, loss_mean_cls: 0.085938, grad_norm: 0.412593 +Steps: 0%| | 912/1000000 [03:49<68:41:50, 4.04it/s, grad_norm=0.413, loss_final=0.657, loss_mean=0.883, loss_mean_cls=0.0859, proj_loss=-0.312][2026-03-26 12:34:53] Step: 912, Training Logs: loss_final: 0.659945, loss_mean: 0.887569, proj_loss: -0.313949, loss_mean_cls: 0.086326, grad_norm: 0.825675 +Steps: 0%| | 913/1000000 [03:49<68:39:48, 4.04it/s, grad_norm=0.826, loss_final=0.66, loss_mean=0.888, loss_mean_cls=0.0863, proj_loss=-0.314][2026-03-26 12:34:53] Step: 913, Training Logs: loss_final: 0.675611, loss_mean: 0.897786, proj_loss: -0.307185, loss_mean_cls: 0.085009, grad_norm: 0.575294 +Steps: 0%| | 914/1000000 [03:50<68:40:54, 4.04it/s, grad_norm=0.575, loss_final=0.676, loss_mean=0.898, loss_mean_cls=0.085, proj_loss=-0.307][2026-03-26 12:34:53] Step: 914, Training Logs: loss_final: 0.649072, loss_mean: 0.869422, proj_loss: -0.307380, loss_mean_cls: 0.087029, grad_norm: 0.824065 +Steps: 0%| | 915/1000000 [03:50<68:40:43, 4.04it/s, grad_norm=0.824, loss_final=0.649, loss_mean=0.869, loss_mean_cls=0.087, proj_loss=-0.307][2026-03-26 12:34:53] Step: 915, Training Logs: loss_final: 0.662780, loss_mean: 0.890342, proj_loss: -0.312155, loss_mean_cls: 0.084592, grad_norm: 0.667563 +Steps: 0%| | 916/1000000 [03:50<68:40:48, 4.04it/s, grad_norm=0.668, loss_final=0.663, loss_mean=0.89, loss_mean_cls=0.0846, proj_loss=-0.312][2026-03-26 12:34:54] Step: 916, Training Logs: loss_final: 0.661582, loss_mean: 0.884415, proj_loss: -0.309055, loss_mean_cls: 0.086222, grad_norm: 0.725062 +Steps: 0%| | 917/1000000 [03:50<68:39:29, 4.04it/s, grad_norm=0.725, loss_final=0.662, loss_mean=0.884, loss_mean_cls=0.0862, proj_loss=-0.309][2026-03-26 12:34:54] Step: 917, Training Logs: loss_final: 0.665419, loss_mean: 0.883683, proj_loss: -0.305154, loss_mean_cls: 0.086890, grad_norm: 0.784678 +Steps: 0%| | 918/1000000 [03:51<68:42:07, 4.04it/s, grad_norm=0.785, loss_final=0.665, loss_mean=0.884, loss_mean_cls=0.0869, proj_loss=-0.305][2026-03-26 12:34:54] Step: 918, Training Logs: loss_final: 0.644437, loss_mean: 0.872478, proj_loss: -0.313385, loss_mean_cls: 0.085344, grad_norm: 0.465403 +Steps: 0%| | 919/1000000 [03:51<68:42:19, 4.04it/s, grad_norm=0.465, loss_final=0.644, loss_mean=0.872, loss_mean_cls=0.0853, proj_loss=-0.313][2026-03-26 12:34:54] Step: 919, Training Logs: loss_final: 0.638748, loss_mean: 0.874118, proj_loss: -0.321099, loss_mean_cls: 0.085729, grad_norm: 0.620001 +Steps: 0%| | 920/1000000 [03:51<68:40:25, 4.04it/s, grad_norm=0.62, loss_final=0.639, loss_mean=0.874, loss_mean_cls=0.0857, proj_loss=-0.321][2026-03-26 12:34:55] Step: 920, Training Logs: loss_final: 0.663295, loss_mean: 0.884306, proj_loss: -0.307859, loss_mean_cls: 0.086848, grad_norm: 0.693834 +Steps: 0%| | 921/1000000 [03:51<68:40:40, 4.04it/s, grad_norm=0.694, loss_final=0.663, loss_mean=0.884, loss_mean_cls=0.0868, proj_loss=-0.308][2026-03-26 12:34:55] Step: 921, Training Logs: loss_final: 0.667211, loss_mean: 0.891794, proj_loss: -0.310687, loss_mean_cls: 0.086104, grad_norm: 0.686151 +Steps: 0%| | 922/1000000 [03:52<68:41:16, 4.04it/s, grad_norm=0.686, loss_final=0.667, loss_mean=0.892, loss_mean_cls=0.0861, proj_loss=-0.311][2026-03-26 12:34:55] Step: 922, Training Logs: loss_final: 0.652872, loss_mean: 0.877692, proj_loss: -0.311114, loss_mean_cls: 0.086295, grad_norm: 0.777512 +Steps: 0%| | 923/1000000 [03:52<68:41:47, 4.04it/s, grad_norm=0.778, loss_final=0.653, loss_mean=0.878, loss_mean_cls=0.0863, proj_loss=-0.311][2026-03-26 12:34:55] Step: 923, Training Logs: loss_final: 0.657765, loss_mean: 0.888115, proj_loss: -0.314512, loss_mean_cls: 0.084163, grad_norm: 0.878540 +Steps: 0%| | 924/1000000 [03:52<68:41:23, 4.04it/s, grad_norm=0.879, loss_final=0.658, loss_mean=0.888, loss_mean_cls=0.0842, proj_loss=-0.315][2026-03-26 12:34:55] Step: 924, Training Logs: loss_final: 0.664013, loss_mean: 0.882089, proj_loss: -0.305515, loss_mean_cls: 0.087440, grad_norm: 0.962356 +Steps: 0%| | 925/1000000 [03:52<68:40:50, 4.04it/s, grad_norm=0.962, loss_final=0.664, loss_mean=0.882, loss_mean_cls=0.0874, proj_loss=-0.306][2026-03-26 12:34:56] Step: 925, Training Logs: loss_final: 0.656754, loss_mean: 0.880074, proj_loss: -0.310516, loss_mean_cls: 0.087197, grad_norm: 0.984962 +Steps: 0%| | 926/1000000 [03:53<68:39:22, 4.04it/s, grad_norm=0.985, loss_final=0.657, loss_mean=0.88, loss_mean_cls=0.0872, proj_loss=-0.311][2026-03-26 12:34:56] Step: 926, Training Logs: loss_final: 0.668690, loss_mean: 0.896672, proj_loss: -0.311895, loss_mean_cls: 0.083913, grad_norm: 0.560029 +Steps: 0%| | 927/1000000 [03:53<68:45:25, 4.04it/s, grad_norm=0.56, loss_final=0.669, loss_mean=0.897, loss_mean_cls=0.0839, proj_loss=-0.312][2026-03-26 12:34:56] Step: 927, Training Logs: loss_final: 0.661619, loss_mean: 0.882684, proj_loss: -0.308504, loss_mean_cls: 0.087439, grad_norm: 0.768070 +Steps: 0%| | 928/1000000 [03:53<68:43:46, 4.04it/s, grad_norm=0.768, loss_final=0.662, loss_mean=0.883, loss_mean_cls=0.0874, proj_loss=-0.309][2026-03-26 12:34:56] Step: 928, Training Logs: loss_final: 0.680070, loss_mean: 0.901608, proj_loss: -0.307719, loss_mean_cls: 0.086181, grad_norm: 0.840740 +Steps: 0%| | 929/1000000 [03:53<68:44:25, 4.04it/s, grad_norm=0.841, loss_final=0.68, loss_mean=0.902, loss_mean_cls=0.0862, proj_loss=-0.308][2026-03-26 12:34:57] Step: 929, Training Logs: loss_final: 0.640149, loss_mean: 0.872607, proj_loss: -0.318385, loss_mean_cls: 0.085927, grad_norm: 0.710321 +Steps: 0%| | 930/1000000 [03:54<68:42:33, 4.04it/s, grad_norm=0.71, loss_final=0.64, loss_mean=0.873, loss_mean_cls=0.0859, proj_loss=-0.318][2026-03-26 12:34:57] Step: 930, Training Logs: loss_final: 0.644477, loss_mean: 0.856758, proj_loss: -0.301012, loss_mean_cls: 0.088731, grad_norm: 0.582026 +Steps: 0%| | 931/1000000 [03:54<68:45:05, 4.04it/s, grad_norm=0.582, loss_final=0.644, loss_mean=0.857, loss_mean_cls=0.0887, proj_loss=-0.301][2026-03-26 12:34:57] Step: 931, Training Logs: loss_final: 0.648475, loss_mean: 0.874582, proj_loss: -0.312076, loss_mean_cls: 0.085970, grad_norm: 0.708700 +Steps: 0%| | 932/1000000 [03:54<68:43:27, 4.04it/s, grad_norm=0.709, loss_final=0.648, loss_mean=0.875, loss_mean_cls=0.086, proj_loss=-0.312][2026-03-26 12:34:57] Step: 932, Training Logs: loss_final: 0.637835, loss_mean: 0.861540, proj_loss: -0.311922, loss_mean_cls: 0.088216, grad_norm: 0.592059 +Steps: 0%| | 933/1000000 [03:54<68:41:48, 4.04it/s, grad_norm=0.592, loss_final=0.638, loss_mean=0.862, loss_mean_cls=0.0882, proj_loss=-0.312][2026-03-26 12:34:58] Step: 933, Training Logs: loss_final: 0.664222, loss_mean: 0.904575, proj_loss: -0.324834, loss_mean_cls: 0.084481, grad_norm: 0.770826 +Steps: 0%| | 934/1000000 [03:55<68:40:35, 4.04it/s, grad_norm=0.771, loss_final=0.664, loss_mean=0.905, loss_mean_cls=0.0845, proj_loss=-0.325][2026-03-26 12:34:58] Step: 934, Training Logs: loss_final: 0.631599, loss_mean: 0.850890, proj_loss: -0.307756, loss_mean_cls: 0.088465, grad_norm: 0.765354 +Steps: 0%| | 935/1000000 [03:55<68:42:02, 4.04it/s, grad_norm=0.765, loss_final=0.632, loss_mean=0.851, loss_mean_cls=0.0885, proj_loss=-0.308][2026-03-26 12:34:58] Step: 935, Training Logs: loss_final: 0.652676, loss_mean: 0.874739, proj_loss: -0.308529, loss_mean_cls: 0.086465, grad_norm: 0.442210 +Steps: 0%| | 936/1000000 [03:55<68:40:28, 4.04it/s, grad_norm=0.442, loss_final=0.653, loss_mean=0.875, loss_mean_cls=0.0865, proj_loss=-0.309][2026-03-26 12:34:58] Step: 936, Training Logs: loss_final: 0.639943, loss_mean: 0.876778, proj_loss: -0.321144, loss_mean_cls: 0.084309, grad_norm: 0.471354 +Steps: 0%| | 937/1000000 [03:55<68:40:52, 4.04it/s, grad_norm=0.471, loss_final=0.64, loss_mean=0.877, loss_mean_cls=0.0843, proj_loss=-0.321][2026-03-26 12:34:59] Step: 937, Training Logs: loss_final: 0.651049, loss_mean: 0.886001, proj_loss: -0.320778, loss_mean_cls: 0.085825, grad_norm: 0.599747 +Steps: 0%| | 938/1000000 [03:56<68:40:19, 4.04it/s, grad_norm=0.6, loss_final=0.651, loss_mean=0.886, loss_mean_cls=0.0858, proj_loss=-0.321][2026-03-26 12:34:59] Step: 938, Training Logs: loss_final: 0.652338, loss_mean: 0.889941, proj_loss: -0.322355, loss_mean_cls: 0.084753, grad_norm: 0.410043 +Steps: 0%| | 939/1000000 [03:56<68:42:46, 4.04it/s, grad_norm=0.41, loss_final=0.652, loss_mean=0.89, loss_mean_cls=0.0848, proj_loss=-0.322][2026-03-26 12:34:59] Step: 939, Training Logs: loss_final: 0.645694, loss_mean: 0.860887, proj_loss: -0.304292, loss_mean_cls: 0.089099, grad_norm: 0.541342 +Steps: 0%| | 940/1000000 [03:56<68:42:02, 4.04it/s, grad_norm=0.541, loss_final=0.646, loss_mean=0.861, loss_mean_cls=0.0891, proj_loss=-0.304][2026-03-26 12:34:59] Step: 940, Training Logs: loss_final: 0.658079, loss_mean: 0.875857, proj_loss: -0.304958, loss_mean_cls: 0.087179, grad_norm: 0.609464 +Steps: 0%| | 941/1000000 [03:56<68:41:24, 4.04it/s, grad_norm=0.609, loss_final=0.658, loss_mean=0.876, loss_mean_cls=0.0872, proj_loss=-0.305][2026-03-26 12:35:00] Step: 941, Training Logs: loss_final: 0.647134, loss_mean: 0.870943, proj_loss: -0.310592, loss_mean_cls: 0.086783, grad_norm: 0.404927 +Steps: 0%| | 942/1000000 [03:57<68:44:05, 4.04it/s, grad_norm=0.405, loss_final=0.647, loss_mean=0.871, loss_mean_cls=0.0868, proj_loss=-0.311][2026-03-26 12:35:00] Step: 942, Training Logs: loss_final: 0.653095, loss_mean: 0.873775, proj_loss: -0.307426, loss_mean_cls: 0.086746, grad_norm: 0.751986 +Steps: 0%| | 943/1000000 [03:57<68:42:27, 4.04it/s, grad_norm=0.752, loss_final=0.653, loss_mean=0.874, loss_mean_cls=0.0867, proj_loss=-0.307][2026-03-26 12:35:00] Step: 943, Training Logs: loss_final: 0.650516, loss_mean: 0.885287, proj_loss: -0.319453, loss_mean_cls: 0.084682, grad_norm: 0.493611 +Steps: 0%| | 944/1000000 [03:57<68:42:42, 4.04it/s, grad_norm=0.494, loss_final=0.651, loss_mean=0.885, loss_mean_cls=0.0847, proj_loss=-0.319][2026-03-26 12:35:00] Step: 944, Training Logs: loss_final: 0.658398, loss_mean: 0.889497, proj_loss: -0.316360, loss_mean_cls: 0.085260, grad_norm: 0.768878 +Steps: 0%| | 945/1000000 [03:57<68:40:55, 4.04it/s, grad_norm=0.769, loss_final=0.658, loss_mean=0.889, loss_mean_cls=0.0853, proj_loss=-0.316][2026-03-26 12:35:01] Step: 945, Training Logs: loss_final: 0.641795, loss_mean: 0.856186, proj_loss: -0.303996, loss_mean_cls: 0.089604, grad_norm: 0.563951 +Steps: 0%| | 946/1000000 [03:58<68:43:17, 4.04it/s, grad_norm=0.564, loss_final=0.642, loss_mean=0.856, loss_mean_cls=0.0896, proj_loss=-0.304][2026-03-26 12:35:01] Step: 946, Training Logs: loss_final: 0.646020, loss_mean: 0.877886, proj_loss: -0.317560, loss_mean_cls: 0.085694, grad_norm: 0.838644 +Steps: 0%| | 947/1000000 [03:58<68:42:36, 4.04it/s, grad_norm=0.839, loss_final=0.646, loss_mean=0.878, loss_mean_cls=0.0857, proj_loss=-0.318][2026-03-26 12:35:01] Step: 947, Training Logs: loss_final: 0.646030, loss_mean: 0.875764, proj_loss: -0.316618, loss_mean_cls: 0.086885, grad_norm: 0.547574 +Steps: 0%| | 948/1000000 [03:58<68:42:36, 4.04it/s, grad_norm=0.548, loss_final=0.646, loss_mean=0.876, loss_mean_cls=0.0869, proj_loss=-0.317][2026-03-26 12:35:01] Step: 948, Training Logs: loss_final: 0.667673, loss_mean: 0.901851, proj_loss: -0.320038, loss_mean_cls: 0.085859, grad_norm: 0.641582 +Steps: 0%| | 949/1000000 [03:58<68:42:08, 4.04it/s, grad_norm=0.642, loss_final=0.668, loss_mean=0.902, loss_mean_cls=0.0859, proj_loss=-0.32][2026-03-26 12:35:02] Step: 949, Training Logs: loss_final: 0.654145, loss_mean: 0.900216, proj_loss: -0.329831, loss_mean_cls: 0.083760, grad_norm: 0.744890 +Steps: 0%| | 950/1000000 [03:59<68:41:29, 4.04it/s, grad_norm=0.745, loss_final=0.654, loss_mean=0.9, loss_mean_cls=0.0838, proj_loss=-0.33][2026-03-26 12:35:02] Step: 950, Training Logs: loss_final: 0.653324, loss_mean: 0.890455, proj_loss: -0.321049, loss_mean_cls: 0.083918, grad_norm: 0.962176 +Steps: 0%| | 951/1000000 [03:59<68:39:07, 4.04it/s, grad_norm=0.962, loss_final=0.653, loss_mean=0.89, loss_mean_cls=0.0839, proj_loss=-0.321][2026-03-26 12:35:02] Step: 951, Training Logs: loss_final: 0.645509, loss_mean: 0.894814, proj_loss: -0.332809, loss_mean_cls: 0.083505, grad_norm: 0.614801 +Steps: 0%| | 952/1000000 [03:59<68:40:07, 4.04it/s, grad_norm=0.615, loss_final=0.646, loss_mean=0.895, loss_mean_cls=0.0835, proj_loss=-0.333][2026-03-26 12:35:02] Step: 952, Training Logs: loss_final: 0.646816, loss_mean: 0.867830, proj_loss: -0.309191, loss_mean_cls: 0.088177, grad_norm: 0.801632 +Steps: 0%| | 953/1000000 [03:59<68:36:16, 4.05it/s, grad_norm=0.802, loss_final=0.647, loss_mean=0.868, loss_mean_cls=0.0882, proj_loss=-0.309][2026-03-26 12:35:03] Step: 953, Training Logs: loss_final: 0.662997, loss_mean: 0.890422, proj_loss: -0.313718, loss_mean_cls: 0.086294, grad_norm: 1.024537 +Steps: 0%| | 954/1000000 [04:00<68:38:38, 4.04it/s, grad_norm=1.02, loss_final=0.663, loss_mean=0.89, loss_mean_cls=0.0863, proj_loss=-0.314][2026-03-26 12:35:03] Step: 954, Training Logs: loss_final: 0.661235, loss_mean: 0.877200, proj_loss: -0.304420, loss_mean_cls: 0.088455, grad_norm: 0.927826 +Steps: 0%| | 955/1000000 [04:00<68:37:31, 4.04it/s, grad_norm=0.928, loss_final=0.661, loss_mean=0.877, loss_mean_cls=0.0885, proj_loss=-0.304][2026-03-26 12:35:03] Step: 955, Training Logs: loss_final: 0.660837, loss_mean: 0.906284, proj_loss: -0.328773, loss_mean_cls: 0.083326, grad_norm: 0.792805 +Steps: 0%| | 956/1000000 [04:00<68:37:43, 4.04it/s, grad_norm=0.793, loss_final=0.661, loss_mean=0.906, loss_mean_cls=0.0833, proj_loss=-0.329][2026-03-26 12:35:03] Step: 956, Training Logs: loss_final: 0.653082, loss_mean: 0.883405, proj_loss: -0.316411, loss_mean_cls: 0.086088, grad_norm: 0.806134 +Steps: 0%| | 957/1000000 [04:00<68:36:56, 4.04it/s, grad_norm=0.806, loss_final=0.653, loss_mean=0.883, loss_mean_cls=0.0861, proj_loss=-0.316][2026-03-26 12:35:04] Step: 957, Training Logs: loss_final: 0.653095, loss_mean: 0.892701, proj_loss: -0.324295, loss_mean_cls: 0.084689, grad_norm: 1.043105 +Steps: 0%| | 958/1000000 [04:01<68:38:24, 4.04it/s, grad_norm=1.04, loss_final=0.653, loss_mean=0.893, loss_mean_cls=0.0847, proj_loss=-0.324][2026-03-26 12:35:04] Step: 958, Training Logs: loss_final: 0.638044, loss_mean: 0.865098, proj_loss: -0.314954, loss_mean_cls: 0.087900, grad_norm: 0.789672 +Steps: 0%| | 959/1000000 [04:01<68:39:24, 4.04it/s, grad_norm=0.79, loss_final=0.638, loss_mean=0.865, loss_mean_cls=0.0879, proj_loss=-0.315][2026-03-26 12:35:04] Step: 959, Training Logs: loss_final: 0.643814, loss_mean: 0.883659, proj_loss: -0.324361, loss_mean_cls: 0.084516, grad_norm: 0.654319 +Steps: 0%| | 960/1000000 [04:01<68:40:15, 4.04it/s, grad_norm=0.654, loss_final=0.644, loss_mean=0.884, loss_mean_cls=0.0845, proj_loss=-0.324][2026-03-26 12:35:04] Step: 960, Training Logs: loss_final: 0.641496, loss_mean: 0.884395, proj_loss: -0.326361, loss_mean_cls: 0.083462, grad_norm: 0.622635 +Steps: 0%| | 961/1000000 [04:01<68:39:28, 4.04it/s, grad_norm=0.623, loss_final=0.641, loss_mean=0.884, loss_mean_cls=0.0835, proj_loss=-0.326][2026-03-26 12:35:05] Step: 961, Training Logs: loss_final: 0.656399, loss_mean: 0.884992, proj_loss: -0.314383, loss_mean_cls: 0.085790, grad_norm: 0.562788 +Steps: 0%| | 962/1000000 [04:02<68:39:34, 4.04it/s, grad_norm=0.563, loss_final=0.656, loss_mean=0.885, loss_mean_cls=0.0858, proj_loss=-0.314][2026-03-26 12:35:05] Step: 962, Training Logs: loss_final: 0.636943, loss_mean: 0.884673, proj_loss: -0.331285, loss_mean_cls: 0.083556, grad_norm: 0.593439 +Steps: 0%| | 963/1000000 [04:02<68:40:00, 4.04it/s, grad_norm=0.593, loss_final=0.637, loss_mean=0.885, loss_mean_cls=0.0836, proj_loss=-0.331][2026-03-26 12:35:05] Step: 963, Training Logs: loss_final: 0.656615, loss_mean: 0.878198, proj_loss: -0.308860, loss_mean_cls: 0.087278, grad_norm: 0.789012 +Steps: 0%| | 964/1000000 [04:02<68:39:42, 4.04it/s, grad_norm=0.789, loss_final=0.657, loss_mean=0.878, loss_mean_cls=0.0873, proj_loss=-0.309][2026-03-26 12:35:05] Step: 964, Training Logs: loss_final: 0.646274, loss_mean: 0.885368, proj_loss: -0.323329, loss_mean_cls: 0.084235, grad_norm: 0.692760 +Steps: 0%| | 965/1000000 [04:02<68:39:36, 4.04it/s, grad_norm=0.693, loss_final=0.646, loss_mean=0.885, loss_mean_cls=0.0842, proj_loss=-0.323][2026-03-26 12:35:06] Step: 965, Training Logs: loss_final: 0.650703, loss_mean: 0.891951, proj_loss: -0.325784, loss_mean_cls: 0.084535, grad_norm: 0.876671 +Steps: 0%| | 966/1000000 [04:02<68:37:06, 4.04it/s, grad_norm=0.877, loss_final=0.651, loss_mean=0.892, loss_mean_cls=0.0845, proj_loss=-0.326][2026-03-26 12:35:06] Step: 966, Training Logs: loss_final: 0.657743, loss_mean: 0.881346, proj_loss: -0.309637, loss_mean_cls: 0.086034, grad_norm: 0.983802 +Steps: 0%| | 967/1000000 [04:03<68:38:04, 4.04it/s, grad_norm=0.984, loss_final=0.658, loss_mean=0.881, loss_mean_cls=0.086, proj_loss=-0.31][2026-03-26 12:35:06] Step: 967, Training Logs: loss_final: 0.658547, loss_mean: 0.896115, proj_loss: -0.322250, loss_mean_cls: 0.084681, grad_norm: 0.982856 +Steps: 0%| | 968/1000000 [04:03<68:36:21, 4.04it/s, grad_norm=0.983, loss_final=0.659, loss_mean=0.896, loss_mean_cls=0.0847, proj_loss=-0.322][2026-03-26 12:35:06] Step: 968, Training Logs: loss_final: 0.642158, loss_mean: 0.872357, proj_loss: -0.316255, loss_mean_cls: 0.086055, grad_norm: 0.449199 +Steps: 0%| | 969/1000000 [04:03<68:35:11, 4.05it/s, grad_norm=0.449, loss_final=0.642, loss_mean=0.872, loss_mean_cls=0.0861, proj_loss=-0.316][2026-03-26 12:35:07] Step: 969, Training Logs: loss_final: 0.660067, loss_mean: 0.894645, proj_loss: -0.319604, loss_mean_cls: 0.085026, grad_norm: 0.932147 +Steps: 0%| | 970/1000000 [04:03<68:37:57, 4.04it/s, grad_norm=0.932, loss_final=0.66, loss_mean=0.895, loss_mean_cls=0.085, proj_loss=-0.32][2026-03-26 12:35:07] Step: 970, Training Logs: loss_final: 0.658120, loss_mean: 0.890039, proj_loss: -0.317638, loss_mean_cls: 0.085719, grad_norm: 0.863120 +Steps: 0%| | 971/1000000 [04:04<68:36:48, 4.04it/s, grad_norm=0.863, loss_final=0.658, loss_mean=0.89, loss_mean_cls=0.0857, proj_loss=-0.318][2026-03-26 12:35:07] Step: 971, Training Logs: loss_final: 0.667753, loss_mean: 0.897131, proj_loss: -0.314539, loss_mean_cls: 0.085161, grad_norm: 0.894288 +Steps: 0%| | 972/1000000 [04:04<68:35:16, 4.05it/s, grad_norm=0.894, loss_final=0.668, loss_mean=0.897, loss_mean_cls=0.0852, proj_loss=-0.315][2026-03-26 12:35:07] Step: 972, Training Logs: loss_final: 0.645888, loss_mean: 0.873721, proj_loss: -0.314807, loss_mean_cls: 0.086974, grad_norm: 0.907949 +Steps: 0%| | 973/1000000 [04:04<68:35:55, 4.05it/s, grad_norm=0.908, loss_final=0.646, loss_mean=0.874, loss_mean_cls=0.087, proj_loss=-0.315][2026-03-26 12:35:08] Step: 973, Training Logs: loss_final: 0.649924, loss_mean: 0.889914, proj_loss: -0.323881, loss_mean_cls: 0.083891, grad_norm: 0.958614 +Steps: 0%| | 974/1000000 [04:04<68:35:30, 4.05it/s, grad_norm=0.959, loss_final=0.65, loss_mean=0.89, loss_mean_cls=0.0839, proj_loss=-0.324][2026-03-26 12:35:08] Step: 974, Training Logs: loss_final: 0.639818, loss_mean: 0.878410, proj_loss: -0.322839, loss_mean_cls: 0.084247, grad_norm: 0.861247 +Steps: 0%| | 975/1000000 [04:05<68:34:05, 4.05it/s, grad_norm=0.861, loss_final=0.64, loss_mean=0.878, loss_mean_cls=0.0842, proj_loss=-0.323][2026-03-26 12:35:08] Step: 975, Training Logs: loss_final: 0.646236, loss_mean: 0.851736, proj_loss: -0.296411, loss_mean_cls: 0.090911, grad_norm: 0.516389 +Steps: 0%| | 976/1000000 [04:05<68:38:01, 4.04it/s, grad_norm=0.516, loss_final=0.646, loss_mean=0.852, loss_mean_cls=0.0909, proj_loss=-0.296][2026-03-26 12:35:08] Step: 976, Training Logs: loss_final: 0.665623, loss_mean: 0.905927, proj_loss: -0.324011, loss_mean_cls: 0.083707, grad_norm: 0.600206 +Steps: 0%| | 977/1000000 [04:05<68:39:14, 4.04it/s, grad_norm=0.6, loss_final=0.666, loss_mean=0.906, loss_mean_cls=0.0837, proj_loss=-0.324][2026-03-26 12:35:09] Step: 977, Training Logs: loss_final: 0.657526, loss_mean: 0.894180, proj_loss: -0.321218, loss_mean_cls: 0.084563, grad_norm: 0.614058 +Steps: 0%| | 978/1000000 [04:05<68:39:20, 4.04it/s, grad_norm=0.614, loss_final=0.658, loss_mean=0.894, loss_mean_cls=0.0846, proj_loss=-0.321][2026-03-26 12:35:09] Step: 978, Training Logs: loss_final: 0.660830, loss_mean: 0.882051, proj_loss: -0.309053, loss_mean_cls: 0.087831, grad_norm: 0.710801 +Steps: 0%| | 979/1000000 [04:06<68:37:50, 4.04it/s, grad_norm=0.711, loss_final=0.661, loss_mean=0.882, loss_mean_cls=0.0878, proj_loss=-0.309][2026-03-26 12:35:09] Step: 979, Training Logs: loss_final: 0.646785, loss_mean: 0.891496, proj_loss: -0.328637, loss_mean_cls: 0.083926, grad_norm: 0.483550 +Steps: 0%| | 980/1000000 [04:06<68:40:34, 4.04it/s, grad_norm=0.484, loss_final=0.647, loss_mean=0.891, loss_mean_cls=0.0839, proj_loss=-0.329][2026-03-26 12:35:09] Step: 980, Training Logs: loss_final: 0.634192, loss_mean: 0.876802, proj_loss: -0.326799, loss_mean_cls: 0.084189, grad_norm: 0.512478 +Steps: 0%| | 981/1000000 [04:06<68:39:48, 4.04it/s, grad_norm=0.512, loss_final=0.634, loss_mean=0.877, loss_mean_cls=0.0842, proj_loss=-0.327][2026-03-26 12:35:10] Step: 981, Training Logs: loss_final: 0.644770, loss_mean: 0.887399, proj_loss: -0.326233, loss_mean_cls: 0.083604, grad_norm: 0.710640 +Steps: 0%| | 982/1000000 [04:06<68:41:48, 4.04it/s, grad_norm=0.711, loss_final=0.645, loss_mean=0.887, loss_mean_cls=0.0836, proj_loss=-0.326][2026-03-26 12:35:10] Step: 982, Training Logs: loss_final: 0.640644, loss_mean: 0.875582, proj_loss: -0.320672, loss_mean_cls: 0.085735, grad_norm: 0.437943 +Steps: 0%| | 983/1000000 [04:07<68:40:15, 4.04it/s, grad_norm=0.438, loss_final=0.641, loss_mean=0.876, loss_mean_cls=0.0857, proj_loss=-0.321][2026-03-26 12:35:10] Step: 983, Training Logs: loss_final: 0.644837, loss_mean: 0.874697, proj_loss: -0.316770, loss_mean_cls: 0.086911, grad_norm: 0.705386 +Steps: 0%| | 984/1000000 [04:07<68:42:05, 4.04it/s, grad_norm=0.705, loss_final=0.645, loss_mean=0.875, loss_mean_cls=0.0869, proj_loss=-0.317][2026-03-26 12:35:10] Step: 984, Training Logs: loss_final: 0.651311, loss_mean: 0.893664, proj_loss: -0.326589, loss_mean_cls: 0.084236, grad_norm: 0.420128 +Steps: 0%| | 985/1000000 [04:07<68:41:29, 4.04it/s, grad_norm=0.42, loss_final=0.651, loss_mean=0.894, loss_mean_cls=0.0842, proj_loss=-0.327][2026-03-26 12:35:11] Step: 985, Training Logs: loss_final: 0.651379, loss_mean: 0.893852, proj_loss: -0.326242, loss_mean_cls: 0.083769, grad_norm: 0.649048 +Steps: 0%| | 986/1000000 [04:07<68:39:44, 4.04it/s, grad_norm=0.649, loss_final=0.651, loss_mean=0.894, loss_mean_cls=0.0838, proj_loss=-0.326][2026-03-26 12:35:11] Step: 986, Training Logs: loss_final: 0.641603, loss_mean: 0.867066, proj_loss: -0.313119, loss_mean_cls: 0.087656, grad_norm: 0.780439 +Steps: 0%| | 987/1000000 [04:08<68:37:47, 4.04it/s, grad_norm=0.78, loss_final=0.642, loss_mean=0.867, loss_mean_cls=0.0877, proj_loss=-0.313][2026-03-26 12:35:11] Step: 987, Training Logs: loss_final: 0.650168, loss_mean: 0.889877, proj_loss: -0.324162, loss_mean_cls: 0.084453, grad_norm: 0.699616 +Steps: 0%| | 988/1000000 [04:08<68:37:38, 4.04it/s, grad_norm=0.7, loss_final=0.65, loss_mean=0.89, loss_mean_cls=0.0845, proj_loss=-0.324][2026-03-26 12:35:11] Step: 988, Training Logs: loss_final: 0.641915, loss_mean: 0.866745, proj_loss: -0.311972, loss_mean_cls: 0.087142, grad_norm: 1.208940 +Steps: 0%| | 989/1000000 [04:08<68:40:09, 4.04it/s, grad_norm=1.21, loss_final=0.642, loss_mean=0.867, loss_mean_cls=0.0871, proj_loss=-0.312][2026-03-26 12:35:12] Step: 989, Training Logs: loss_final: 0.637992, loss_mean: 0.873664, proj_loss: -0.320797, loss_mean_cls: 0.085125, grad_norm: 0.687915 +Steps: 0%| | 990/1000000 [04:08<68:39:07, 4.04it/s, grad_norm=0.688, loss_final=0.638, loss_mean=0.874, loss_mean_cls=0.0851, proj_loss=-0.321][2026-03-26 12:35:12] Step: 990, Training Logs: loss_final: 0.646140, loss_mean: 0.882266, proj_loss: -0.322354, loss_mean_cls: 0.086228, grad_norm: 0.887912 +Steps: 0%| | 991/1000000 [04:09<68:37:18, 4.04it/s, grad_norm=0.888, loss_final=0.646, loss_mean=0.882, loss_mean_cls=0.0862, proj_loss=-0.322][2026-03-26 12:35:12] Step: 991, Training Logs: loss_final: 0.635312, loss_mean: 0.858616, proj_loss: -0.310586, loss_mean_cls: 0.087282, grad_norm: 1.060831 +Steps: 0%| | 992/1000000 [04:09<68:40:03, 4.04it/s, grad_norm=1.06, loss_final=0.635, loss_mean=0.859, loss_mean_cls=0.0873, proj_loss=-0.311][2026-03-26 12:35:12] Step: 992, Training Logs: loss_final: 0.634202, loss_mean: 0.868510, proj_loss: -0.319583, loss_mean_cls: 0.085275, grad_norm: 0.545445 +Steps: 0%| | 993/1000000 [04:09<68:37:16, 4.04it/s, grad_norm=0.545, loss_final=0.634, loss_mean=0.869, loss_mean_cls=0.0853, proj_loss=-0.32][2026-03-26 12:35:13] Step: 993, Training Logs: loss_final: 0.641519, loss_mean: 0.889071, proj_loss: -0.330611, loss_mean_cls: 0.083059, grad_norm: 1.233173 +Steps: 0%| | 994/1000000 [04:09<68:35:21, 4.05it/s, grad_norm=1.23, loss_final=0.642, loss_mean=0.889, loss_mean_cls=0.0831, proj_loss=-0.331][2026-03-26 12:35:13] Step: 994, Training Logs: loss_final: 0.650681, loss_mean: 0.894927, proj_loss: -0.327820, loss_mean_cls: 0.083574, grad_norm: 0.986620 +Steps: 0%| | 995/1000000 [04:10<68:36:20, 4.04it/s, grad_norm=0.987, loss_final=0.651, loss_mean=0.895, loss_mean_cls=0.0836, proj_loss=-0.328][2026-03-26 12:35:13] Step: 995, Training Logs: loss_final: 0.635572, loss_mean: 0.882184, proj_loss: -0.329693, loss_mean_cls: 0.083081, grad_norm: 0.715541 +Steps: 0%| | 996/1000000 [04:10<68:37:32, 4.04it/s, grad_norm=0.716, loss_final=0.636, loss_mean=0.882, loss_mean_cls=0.0831, proj_loss=-0.33][2026-03-26 12:35:13] Step: 996, Training Logs: loss_final: 0.636518, loss_mean: 0.866150, proj_loss: -0.316270, loss_mean_cls: 0.086639, grad_norm: 0.768587 +Steps: 0%| | 997/1000000 [04:10<68:38:15, 4.04it/s, grad_norm=0.769, loss_final=0.637, loss_mean=0.866, loss_mean_cls=0.0866, proj_loss=-0.316][2026-03-26 12:35:14] Step: 997, Training Logs: loss_final: 0.637028, loss_mean: 0.866542, proj_loss: -0.316084, loss_mean_cls: 0.086569, grad_norm: 0.542799 +Steps: 0%| | 998/1000000 [04:10<68:38:01, 4.04it/s, grad_norm=0.543, loss_final=0.637, loss_mean=0.867, loss_mean_cls=0.0866, proj_loss=-0.316][2026-03-26 12:35:14] Step: 998, Training Logs: loss_final: 0.644504, loss_mean: 0.891957, proj_loss: -0.330938, loss_mean_cls: 0.083484, grad_norm: 0.778015 +Steps: 0%| | 999/1000000 [04:11<68:41:26, 4.04it/s, grad_norm=0.778, loss_final=0.645, loss_mean=0.892, loss_mean_cls=0.0835, proj_loss=-0.331][2026-03-26 12:35:14] Step: 999, Training Logs: loss_final: 0.645315, loss_mean: 0.884692, proj_loss: -0.324381, loss_mean_cls: 0.085004, grad_norm: 0.409642 +Steps: 0%| | 1000/1000000 [04:11<68:41:41, 4.04it/s, grad_norm=0.41, loss_final=0.645, loss_mean=0.885, loss_mean_cls=0.085, proj_loss=-0.324][2026-03-26 12:35:14] Step: 1000, Training Logs: loss_final: 0.645505, loss_mean: 0.889657, proj_loss: -0.328311, loss_mean_cls: 0.084159, grad_norm: 0.702422 +Steps: 0%| | 1001/1000000 [04:11<68:41:52, 4.04it/s, grad_norm=0.702, loss_final=0.646, loss_mean=0.89, loss_mean_cls=0.0842, proj_loss=-0.328][2026-03-26 12:35:15] Step: 1001, Training Logs: loss_final: 0.639093, loss_mean: 0.867874, proj_loss: -0.316696, loss_mean_cls: 0.087916, grad_norm: 0.608811 +Steps: 0%| | 1002/1000000 [04:11<68:39:37, 4.04it/s, grad_norm=0.609, loss_final=0.639, loss_mean=0.868, loss_mean_cls=0.0879, proj_loss=-0.317][2026-03-26 12:35:15] Step: 1002, Training Logs: loss_final: 0.619053, loss_mean: 0.858763, proj_loss: -0.325415, loss_mean_cls: 0.085706, grad_norm: 0.408551 +Steps: 0%| | 1003/1000000 [04:12<68:39:57, 4.04it/s, grad_norm=0.409, loss_final=0.619, loss_mean=0.859, loss_mean_cls=0.0857, proj_loss=-0.325][2026-03-26 12:35:15] Step: 1003, Training Logs: loss_final: 0.652473, loss_mean: 0.871659, proj_loss: -0.306432, loss_mean_cls: 0.087246, grad_norm: 0.779753 +Steps: 0%| | 1004/1000000 [04:12<68:43:39, 4.04it/s, grad_norm=0.78, loss_final=0.652, loss_mean=0.872, loss_mean_cls=0.0872, proj_loss=-0.306][2026-03-26 12:35:15] Step: 1004, Training Logs: loss_final: 0.642654, loss_mean: 0.902393, proj_loss: -0.341588, loss_mean_cls: 0.081849, grad_norm: 0.597855 +Steps: 0%| | 1005/1000000 [04:12<68:43:40, 4.04it/s, grad_norm=0.598, loss_final=0.643, loss_mean=0.902, loss_mean_cls=0.0818, proj_loss=-0.342][2026-03-26 12:35:16] Step: 1005, Training Logs: loss_final: 0.636124, loss_mean: 0.869663, proj_loss: -0.318932, loss_mean_cls: 0.085394, grad_norm: 0.597382 +Steps: 0%| | 1006/1000000 [04:12<68:41:42, 4.04it/s, grad_norm=0.597, loss_final=0.636, loss_mean=0.87, loss_mean_cls=0.0854, proj_loss=-0.319][2026-03-26 12:35:16] Step: 1006, Training Logs: loss_final: 0.643011, loss_mean: 0.871904, proj_loss: -0.315916, loss_mean_cls: 0.087023, grad_norm: 0.850822 +Steps: 0%| | 1007/1000000 [04:13<68:40:46, 4.04it/s, grad_norm=0.851, loss_final=0.643, loss_mean=0.872, loss_mean_cls=0.087, proj_loss=-0.316][2026-03-26 12:35:16] Step: 1007, Training Logs: loss_final: 0.633514, loss_mean: 0.875830, proj_loss: -0.327068, loss_mean_cls: 0.084752, grad_norm: 0.957104 +Steps: 0%| | 1008/1000000 [04:13<68:38:10, 4.04it/s, grad_norm=0.957, loss_final=0.634, loss_mean=0.876, loss_mean_cls=0.0848, proj_loss=-0.327][2026-03-26 12:35:16] Step: 1008, Training Logs: loss_final: 0.617028, loss_mean: 0.873256, proj_loss: -0.338714, loss_mean_cls: 0.082487, grad_norm: 0.722140 +Steps: 0%| | 1009/1000000 [04:13<68:39:36, 4.04it/s, grad_norm=0.722, loss_final=0.617, loss_mean=0.873, loss_mean_cls=0.0825, proj_loss=-0.339][2026-03-26 12:35:17] Step: 1009, Training Logs: loss_final: 0.640881, loss_mean: 0.878017, proj_loss: -0.323104, loss_mean_cls: 0.085967, grad_norm: 0.507103 +Steps: 0%| | 1010/1000000 [04:13<69:17:58, 4.00it/s, grad_norm=0.507, loss_final=0.641, loss_mean=0.878, loss_mean_cls=0.086, proj_loss=-0.323][2026-03-26 12:35:17] Step: 1010, Training Logs: loss_final: 0.623931, loss_mean: 0.873087, proj_loss: -0.331576, loss_mean_cls: 0.082420, grad_norm: 0.813961 +Steps: 0%| | 1011/1000000 [04:14<69:09:24, 4.01it/s, grad_norm=0.814, loss_final=0.624, loss_mean=0.873, loss_mean_cls=0.0824, proj_loss=-0.332][2026-03-26 12:35:17] Step: 1011, Training Logs: loss_final: 0.641977, loss_mean: 0.876068, proj_loss: -0.320963, loss_mean_cls: 0.086872, grad_norm: 0.965239 +Steps: 0%| | 1012/1000000 [04:14<69:00:38, 4.02it/s, grad_norm=0.965, loss_final=0.642, loss_mean=0.876, loss_mean_cls=0.0869, proj_loss=-0.321][2026-03-26 12:35:17] Step: 1012, Training Logs: loss_final: 0.643333, loss_mean: 0.888228, proj_loss: -0.328788, loss_mean_cls: 0.083894, grad_norm: 0.740913 +Steps: 0%| | 1013/1000000 [04:14<68:53:10, 4.03it/s, grad_norm=0.741, loss_final=0.643, loss_mean=0.888, loss_mean_cls=0.0839, proj_loss=-0.329][2026-03-26 12:35:18] Step: 1013, Training Logs: loss_final: 0.640151, loss_mean: 0.885487, proj_loss: -0.329320, loss_mean_cls: 0.083984, grad_norm: 0.703128 +Steps: 0%| | 1014/1000000 [04:14<69:00:28, 4.02it/s, grad_norm=0.703, loss_final=0.64, loss_mean=0.885, loss_mean_cls=0.084, proj_loss=-0.329][2026-03-26 12:35:18] Step: 1014, Training Logs: loss_final: 0.639778, loss_mean: 0.884401, proj_loss: -0.329232, loss_mean_cls: 0.084609, grad_norm: 0.901210 +Steps: 0%| | 1015/1000000 [04:15<71:58:22, 3.86it/s, grad_norm=0.901, loss_final=0.64, loss_mean=0.884, loss_mean_cls=0.0846, proj_loss=-0.329][2026-03-26 12:35:18] Step: 1015, Training Logs: loss_final: 0.647557, loss_mean: 0.892744, proj_loss: -0.328447, loss_mean_cls: 0.083260, grad_norm: 0.763102 +Steps: 0%| | 1016/1000000 [04:15<70:59:16, 3.91it/s, grad_norm=0.763, loss_final=0.648, loss_mean=0.893, loss_mean_cls=0.0833, proj_loss=-0.328][2026-03-26 12:35:18] Step: 1016, Training Logs: loss_final: 0.635096, loss_mean: 0.879310, proj_loss: -0.328383, loss_mean_cls: 0.084168, grad_norm: 0.479800 +Steps: 0%| | 1017/1000000 [04:15<70:16:44, 3.95it/s, grad_norm=0.48, loss_final=0.635, loss_mean=0.879, loss_mean_cls=0.0842, proj_loss=-0.328][2026-03-26 12:35:19] Step: 1017, Training Logs: loss_final: 0.630188, loss_mean: 0.852895, proj_loss: -0.309820, loss_mean_cls: 0.087113, grad_norm: 0.645474 +Steps: 0%| | 1018/1000000 [04:15<69:46:43, 3.98it/s, grad_norm=0.645, loss_final=0.63, loss_mean=0.853, loss_mean_cls=0.0871, proj_loss=-0.31][2026-03-26 12:35:19] Step: 1018, Training Logs: loss_final: 0.655722, loss_mean: 0.890897, proj_loss: -0.321057, loss_mean_cls: 0.085882, grad_norm: 0.686626 +Steps: 0%| | 1019/1000000 [04:16<69:27:09, 4.00it/s, grad_norm=0.687, loss_final=0.656, loss_mean=0.891, loss_mean_cls=0.0859, proj_loss=-0.321][2026-03-26 12:35:19] Step: 1019, Training Logs: loss_final: 0.630030, loss_mean: 0.857387, proj_loss: -0.314704, loss_mean_cls: 0.087348, grad_norm: 0.682789 +Steps: 0%| | 1020/1000000 [04:16<69:13:07, 4.01it/s, grad_norm=0.683, loss_final=0.63, loss_mean=0.857, loss_mean_cls=0.0873, proj_loss=-0.315][2026-03-26 12:35:19] Step: 1020, Training Logs: loss_final: 0.642728, loss_mean: 0.877381, proj_loss: -0.320054, loss_mean_cls: 0.085402, grad_norm: 0.831068 +Steps: 0%| | 1021/1000000 [04:16<69:02:58, 4.02it/s, grad_norm=0.831, loss_final=0.643, loss_mean=0.877, loss_mean_cls=0.0854, proj_loss=-0.32][2026-03-26 12:35:20] Step: 1021, Training Logs: loss_final: 0.636361, loss_mean: 0.865678, proj_loss: -0.317598, loss_mean_cls: 0.088281, grad_norm: 1.103792 +Steps: 0%| | 1022/1000000 [04:16<68:56:02, 4.03it/s, grad_norm=1.1, loss_final=0.636, loss_mean=0.866, loss_mean_cls=0.0883, proj_loss=-0.318][2026-03-26 12:35:20] Step: 1022, Training Logs: loss_final: 0.634531, loss_mean: 0.867538, proj_loss: -0.318377, loss_mean_cls: 0.085369, grad_norm: 0.436077 +Steps: 0%| | 1023/1000000 [04:17<68:52:48, 4.03it/s, grad_norm=0.436, loss_final=0.635, loss_mean=0.868, loss_mean_cls=0.0854, proj_loss=-0.318][2026-03-26 12:35:20] Step: 1023, Training Logs: loss_final: 0.651902, loss_mean: 0.882785, proj_loss: -0.316399, loss_mean_cls: 0.085516, grad_norm: 0.959542 +Steps: 0%| | 1024/1000000 [04:17<68:47:51, 4.03it/s, grad_norm=0.96, loss_final=0.652, loss_mean=0.883, loss_mean_cls=0.0855, proj_loss=-0.316][2026-03-26 12:35:20] Step: 1024, Training Logs: loss_final: 0.645028, loss_mean: 0.883953, proj_loss: -0.323585, loss_mean_cls: 0.084660, grad_norm: 0.572522 +Steps: 0%| | 1025/1000000 [04:17<68:49:37, 4.03it/s, grad_norm=0.573, loss_final=0.645, loss_mean=0.884, loss_mean_cls=0.0847, proj_loss=-0.324][2026-03-26 12:35:21] Step: 1025, Training Logs: loss_final: 0.636408, loss_mean: 0.876430, proj_loss: -0.325543, loss_mean_cls: 0.085521, grad_norm: 0.764086 +Steps: 0%| | 1026/1000000 [04:17<68:45:41, 4.04it/s, grad_norm=0.764, loss_final=0.636, loss_mean=0.876, loss_mean_cls=0.0855, proj_loss=-0.326][2026-03-26 12:35:21] Step: 1026, Training Logs: loss_final: 0.640206, loss_mean: 0.845415, proj_loss: -0.297324, loss_mean_cls: 0.092115, grad_norm: 0.491780 +Steps: 0%| | 1027/1000000 [04:18<68:46:43, 4.03it/s, grad_norm=0.492, loss_final=0.64, loss_mean=0.845, loss_mean_cls=0.0921, proj_loss=-0.297][2026-03-26 12:35:21] Step: 1027, Training Logs: loss_final: 0.625556, loss_mean: 0.860016, proj_loss: -0.320409, loss_mean_cls: 0.085949, grad_norm: 0.658471 +Steps: 0%| | 1028/1000000 [04:18<68:45:02, 4.04it/s, grad_norm=0.658, loss_final=0.626, loss_mean=0.86, loss_mean_cls=0.0859, proj_loss=-0.32][2026-03-26 12:35:21] Step: 1028, Training Logs: loss_final: 0.635625, loss_mean: 0.885855, proj_loss: -0.332446, loss_mean_cls: 0.082215, grad_norm: 0.554622 +Steps: 0%| | 1029/1000000 [04:18<68:43:07, 4.04it/s, grad_norm=0.555, loss_final=0.636, loss_mean=0.886, loss_mean_cls=0.0822, proj_loss=-0.332][2026-03-26 12:35:22] Step: 1029, Training Logs: loss_final: 0.647278, loss_mean: 0.877105, proj_loss: -0.316202, loss_mean_cls: 0.086376, grad_norm: 0.627150 +Steps: 0%| | 1030/1000000 [04:18<68:41:41, 4.04it/s, grad_norm=0.627, loss_final=0.647, loss_mean=0.877, loss_mean_cls=0.0864, proj_loss=-0.316][2026-03-26 12:35:22] Step: 1030, Training Logs: loss_final: 0.638290, loss_mean: 0.892178, proj_loss: -0.336932, loss_mean_cls: 0.083044, grad_norm: 0.761329 +Steps: 0%| | 1031/1000000 [04:19<68:43:03, 4.04it/s, grad_norm=0.761, loss_final=0.638, loss_mean=0.892, loss_mean_cls=0.083, proj_loss=-0.337][2026-03-26 12:35:22] Step: 1031, Training Logs: loss_final: 0.638115, loss_mean: 0.879266, proj_loss: -0.325660, loss_mean_cls: 0.084510, grad_norm: 0.689237 +Steps: 0%| | 1032/1000000 [04:19<68:46:55, 4.03it/s, grad_norm=0.689, loss_final=0.638, loss_mean=0.879, loss_mean_cls=0.0845, proj_loss=-0.326][2026-03-26 12:35:22] Step: 1032, Training Logs: loss_final: 0.634920, loss_mean: 0.887054, proj_loss: -0.335063, loss_mean_cls: 0.082929, grad_norm: 0.864254 +Steps: 0%| | 1033/1000000 [04:19<68:42:49, 4.04it/s, grad_norm=0.864, loss_final=0.635, loss_mean=0.887, loss_mean_cls=0.0829, proj_loss=-0.335][2026-03-26 12:35:23] Step: 1033, Training Logs: loss_final: 0.634912, loss_mean: 0.886671, proj_loss: -0.334693, loss_mean_cls: 0.082933, grad_norm: 0.531586 +Steps: 0%| | 1034/1000000 [04:19<68:43:17, 4.04it/s, grad_norm=0.532, loss_final=0.635, loss_mean=0.887, loss_mean_cls=0.0829, proj_loss=-0.335][2026-03-26 12:35:23] Step: 1034, Training Logs: loss_final: 0.636050, loss_mean: 0.876788, proj_loss: -0.325449, loss_mean_cls: 0.084711, grad_norm: 0.655029 +Steps: 0%| | 1035/1000000 [04:20<68:42:27, 4.04it/s, grad_norm=0.655, loss_final=0.636, loss_mean=0.877, loss_mean_cls=0.0847, proj_loss=-0.325][2026-03-26 12:35:23] Step: 1035, Training Logs: loss_final: 0.621130, loss_mean: 0.872215, proj_loss: -0.334178, loss_mean_cls: 0.083093, grad_norm: 0.554158 +Steps: 0%| | 1036/1000000 [04:20<68:44:52, 4.04it/s, grad_norm=0.554, loss_final=0.621, loss_mean=0.872, loss_mean_cls=0.0831, proj_loss=-0.334][2026-03-26 12:35:23] Step: 1036, Training Logs: loss_final: 0.654542, loss_mean: 0.894832, proj_loss: -0.324442, loss_mean_cls: 0.084153, grad_norm: 0.534695 +Steps: 0%| | 1037/1000000 [04:20<68:44:41, 4.04it/s, grad_norm=0.535, loss_final=0.655, loss_mean=0.895, loss_mean_cls=0.0842, proj_loss=-0.324][2026-03-26 12:35:24] Step: 1037, Training Logs: loss_final: 0.638036, loss_mean: 0.869340, proj_loss: -0.318802, loss_mean_cls: 0.087498, grad_norm: 0.539310 +Steps: 0%| | 1038/1000000 [04:20<68:47:18, 4.03it/s, grad_norm=0.539, loss_final=0.638, loss_mean=0.869, loss_mean_cls=0.0875, proj_loss=-0.319][2026-03-26 12:35:24] Step: 1038, Training Logs: loss_final: 0.616732, loss_mean: 0.853932, proj_loss: -0.323698, loss_mean_cls: 0.086498, grad_norm: 0.915693 +Steps: 0%| | 1039/1000000 [04:21<68:44:54, 4.04it/s, grad_norm=0.916, loss_final=0.617, loss_mean=0.854, loss_mean_cls=0.0865, proj_loss=-0.324][2026-03-26 12:35:24] Step: 1039, Training Logs: loss_final: 0.635105, loss_mean: 0.866635, proj_loss: -0.317701, loss_mean_cls: 0.086171, grad_norm: 0.479526 +Steps: 0%| | 1040/1000000 [04:21<68:46:09, 4.04it/s, grad_norm=0.48, loss_final=0.635, loss_mean=0.867, loss_mean_cls=0.0862, proj_loss=-0.318][2026-03-26 12:35:24] Step: 1040, Training Logs: loss_final: 0.625525, loss_mean: 0.874647, proj_loss: -0.333191, loss_mean_cls: 0.084069, grad_norm: 0.873146 +Steps: 0%| | 1041/1000000 [04:21<68:43:22, 4.04it/s, grad_norm=0.873, loss_final=0.626, loss_mean=0.875, loss_mean_cls=0.0841, proj_loss=-0.333][2026-03-26 12:35:24] Step: 1041, Training Logs: loss_final: 0.643365, loss_mean: 0.882972, proj_loss: -0.324624, loss_mean_cls: 0.085018, grad_norm: 0.596577 +Steps: 0%| | 1042/1000000 [04:21<68:41:24, 4.04it/s, grad_norm=0.597, loss_final=0.643, loss_mean=0.883, loss_mean_cls=0.085, proj_loss=-0.325][2026-03-26 12:35:25] Step: 1042, Training Logs: loss_final: 0.659670, loss_mean: 0.900001, proj_loss: -0.325333, loss_mean_cls: 0.085001, grad_norm: 0.796175 +Steps: 0%| | 1043/1000000 [04:22<68:42:05, 4.04it/s, grad_norm=0.796, loss_final=0.66, loss_mean=0.9, loss_mean_cls=0.085, proj_loss=-0.325][2026-03-26 12:35:25] Step: 1043, Training Logs: loss_final: 0.656960, loss_mean: 0.905342, proj_loss: -0.331638, loss_mean_cls: 0.083255, grad_norm: 0.767512 +Steps: 0%| | 1044/1000000 [04:22<68:44:35, 4.04it/s, grad_norm=0.768, loss_final=0.657, loss_mean=0.905, loss_mean_cls=0.0833, proj_loss=-0.332][2026-03-26 12:35:25] Step: 1044, Training Logs: loss_final: 0.630391, loss_mean: 0.859830, proj_loss: -0.316954, loss_mean_cls: 0.087515, grad_norm: 0.803635 +Steps: 0%| | 1045/1000000 [04:22<68:43:33, 4.04it/s, grad_norm=0.804, loss_final=0.63, loss_mean=0.86, loss_mean_cls=0.0875, proj_loss=-0.317][2026-03-26 12:35:25] Step: 1045, Training Logs: loss_final: 0.646929, loss_mean: 0.885119, proj_loss: -0.323008, loss_mean_cls: 0.084818, grad_norm: 0.606360 +Steps: 0%| | 1046/1000000 [04:22<69:40:50, 3.98it/s, grad_norm=0.606, loss_final=0.647, loss_mean=0.885, loss_mean_cls=0.0848, proj_loss=-0.323][2026-03-26 12:35:26] Step: 1046, Training Logs: loss_final: 0.641009, loss_mean: 0.892524, proj_loss: -0.334405, loss_mean_cls: 0.082890, grad_norm: 0.803921 +Steps: 0%| | 1047/1000000 [04:23<69:20:57, 4.00it/s, grad_norm=0.804, loss_final=0.641, loss_mean=0.893, loss_mean_cls=0.0829, proj_loss=-0.334][2026-03-26 12:35:26] Step: 1047, Training Logs: loss_final: 0.636406, loss_mean: 0.881756, proj_loss: -0.328776, loss_mean_cls: 0.083427, grad_norm: 1.085047 +Steps: 0%| | 1048/1000000 [04:23<69:13:15, 4.01it/s, grad_norm=1.09, loss_final=0.636, loss_mean=0.882, loss_mean_cls=0.0834, proj_loss=-0.329][2026-03-26 12:35:26] Step: 1048, Training Logs: loss_final: 0.627276, loss_mean: 0.878142, proj_loss: -0.335020, loss_mean_cls: 0.084154, grad_norm: 0.589065 +Steps: 0%| | 1049/1000000 [04:23<69:01:55, 4.02it/s, grad_norm=0.589, loss_final=0.627, loss_mean=0.878, loss_mean_cls=0.0842, proj_loss=-0.335][2026-03-26 12:35:26] Step: 1049, Training Logs: loss_final: 0.647257, loss_mean: 0.894695, proj_loss: -0.330592, loss_mean_cls: 0.083155, grad_norm: 0.628758 +Steps: 0%| | 1050/1000000 [04:23<68:59:45, 4.02it/s, grad_norm=0.629, loss_final=0.647, loss_mean=0.895, loss_mean_cls=0.0832, proj_loss=-0.331][2026-03-26 12:35:27] Step: 1050, Training Logs: loss_final: 0.632909, loss_mean: 0.864434, proj_loss: -0.317981, loss_mean_cls: 0.086456, grad_norm: 0.841391 +Steps: 0%| | 1051/1000000 [04:24<68:51:54, 4.03it/s, grad_norm=0.841, loss_final=0.633, loss_mean=0.864, loss_mean_cls=0.0865, proj_loss=-0.318][2026-03-26 12:35:27] Step: 1051, Training Logs: loss_final: 0.631618, loss_mean: 0.868019, proj_loss: -0.322831, loss_mean_cls: 0.086429, grad_norm: 0.483317 +Steps: 0%| | 1052/1000000 [04:24<68:50:17, 4.03it/s, grad_norm=0.483, loss_final=0.632, loss_mean=0.868, loss_mean_cls=0.0864, proj_loss=-0.323][2026-03-26 12:35:27] Step: 1052, Training Logs: loss_final: 0.628390, loss_mean: 0.873269, proj_loss: -0.329360, loss_mean_cls: 0.084481, grad_norm: 0.790342 +Steps: 0%| | 1053/1000000 [04:24<68:46:04, 4.04it/s, grad_norm=0.79, loss_final=0.628, loss_mean=0.873, loss_mean_cls=0.0845, proj_loss=-0.329][2026-03-26 12:35:27] Step: 1053, Training Logs: loss_final: 0.624525, loss_mean: 0.871773, proj_loss: -0.332862, loss_mean_cls: 0.085614, grad_norm: 0.493315 +Steps: 0%| | 1054/1000000 [04:24<68:45:44, 4.04it/s, grad_norm=0.493, loss_final=0.625, loss_mean=0.872, loss_mean_cls=0.0856, proj_loss=-0.333][2026-03-26 12:35:28] Step: 1054, Training Logs: loss_final: 0.622402, loss_mean: 0.868446, proj_loss: -0.330246, loss_mean_cls: 0.084202, grad_norm: 0.668282 +Steps: 0%| | 1055/1000000 [04:25<68:44:36, 4.04it/s, grad_norm=0.668, loss_final=0.622, loss_mean=0.868, loss_mean_cls=0.0842, proj_loss=-0.33][2026-03-26 12:35:28] Step: 1055, Training Logs: loss_final: 0.648620, loss_mean: 0.896286, proj_loss: -0.331820, loss_mean_cls: 0.084154, grad_norm: 0.619068 +Steps: 0%| | 1056/1000000 [04:25<68:46:11, 4.03it/s, grad_norm=0.619, loss_final=0.649, loss_mean=0.896, loss_mean_cls=0.0842, proj_loss=-0.332][2026-03-26 12:35:28] Step: 1056, Training Logs: loss_final: 0.631439, loss_mean: 0.867726, proj_loss: -0.322623, loss_mean_cls: 0.086336, grad_norm: 0.831770 +Steps: 0%| | 1057/1000000 [04:25<68:42:42, 4.04it/s, grad_norm=0.832, loss_final=0.631, loss_mean=0.868, loss_mean_cls=0.0863, proj_loss=-0.323][2026-03-26 12:35:28] Step: 1057, Training Logs: loss_final: 0.640224, loss_mean: 0.886693, proj_loss: -0.330323, loss_mean_cls: 0.083854, grad_norm: 0.529727 +Steps: 0%| | 1058/1000000 [04:25<68:46:16, 4.03it/s, grad_norm=0.53, loss_final=0.64, loss_mean=0.887, loss_mean_cls=0.0839, proj_loss=-0.33][2026-03-26 12:35:29] Step: 1058, Training Logs: loss_final: 0.621598, loss_mean: 0.871954, proj_loss: -0.334800, loss_mean_cls: 0.084443, grad_norm: 0.788665 +Steps: 0%| | 1059/1000000 [04:26<68:44:46, 4.04it/s, grad_norm=0.789, loss_final=0.622, loss_mean=0.872, loss_mean_cls=0.0844, proj_loss=-0.335][2026-03-26 12:35:29] Step: 1059, Training Logs: loss_final: 0.628682, loss_mean: 0.877623, proj_loss: -0.332969, loss_mean_cls: 0.084028, grad_norm: 0.762917 +Steps: 0%| | 1060/1000000 [04:26<68:42:58, 4.04it/s, grad_norm=0.763, loss_final=0.629, loss_mean=0.878, loss_mean_cls=0.084, proj_loss=-0.333][2026-03-26 12:35:29] Step: 1060, Training Logs: loss_final: 0.645306, loss_mean: 0.890538, proj_loss: -0.329435, loss_mean_cls: 0.084204, grad_norm: 0.649657 +Steps: 0%| | 1061/1000000 [04:26<68:42:30, 4.04it/s, grad_norm=0.65, loss_final=0.645, loss_mean=0.891, loss_mean_cls=0.0842, proj_loss=-0.329][2026-03-26 12:35:29] Step: 1061, Training Logs: loss_final: 0.633886, loss_mean: 0.861294, proj_loss: -0.315883, loss_mean_cls: 0.088475, grad_norm: 0.672725 +Steps: 0%| | 1062/1000000 [04:26<68:52:37, 4.03it/s, grad_norm=0.673, loss_final=0.634, loss_mean=0.861, loss_mean_cls=0.0885, proj_loss=-0.316][2026-03-26 12:35:30] Step: 1062, Training Logs: loss_final: 0.641928, loss_mean: 0.894160, proj_loss: -0.334864, loss_mean_cls: 0.082632, grad_norm: 0.542701 +Steps: 0%| | 1063/1000000 [04:27<68:48:44, 4.03it/s, grad_norm=0.543, loss_final=0.642, loss_mean=0.894, loss_mean_cls=0.0826, proj_loss=-0.335][2026-03-26 12:35:30] Step: 1063, Training Logs: loss_final: 0.622071, loss_mean: 0.866152, proj_loss: -0.328513, loss_mean_cls: 0.084432, grad_norm: 0.617517 +Steps: 0%| | 1064/1000000 [04:27<68:49:16, 4.03it/s, grad_norm=0.618, loss_final=0.622, loss_mean=0.866, loss_mean_cls=0.0844, proj_loss=-0.329][2026-03-26 12:35:30] Step: 1064, Training Logs: loss_final: 0.635041, loss_mean: 0.876337, proj_loss: -0.327180, loss_mean_cls: 0.085885, grad_norm: 0.518704 +Steps: 0%| | 1065/1000000 [04:27<68:46:51, 4.03it/s, grad_norm=0.519, loss_final=0.635, loss_mean=0.876, loss_mean_cls=0.0859, proj_loss=-0.327][2026-03-26 12:35:30] Step: 1065, Training Logs: loss_final: 0.622767, loss_mean: 0.868806, proj_loss: -0.331390, loss_mean_cls: 0.085351, grad_norm: 0.505156 +Steps: 0%| | 1066/1000000 [04:27<68:45:51, 4.04it/s, grad_norm=0.505, loss_final=0.623, loss_mean=0.869, loss_mean_cls=0.0854, proj_loss=-0.331][2026-03-26 12:35:31] Step: 1066, Training Logs: loss_final: 0.628233, loss_mean: 0.873289, proj_loss: -0.329890, loss_mean_cls: 0.084834, grad_norm: 0.640124 +Steps: 0%| | 1067/1000000 [04:28<68:45:34, 4.04it/s, grad_norm=0.64, loss_final=0.628, loss_mean=0.873, loss_mean_cls=0.0848, proj_loss=-0.33][2026-03-26 12:35:31] Step: 1067, Training Logs: loss_final: 0.626150, loss_mean: 0.871416, proj_loss: -0.330425, loss_mean_cls: 0.085159, grad_norm: 0.666240 +Steps: 0%| | 1068/1000000 [04:28<68:48:14, 4.03it/s, grad_norm=0.666, loss_final=0.626, loss_mean=0.871, loss_mean_cls=0.0852, proj_loss=-0.33][2026-03-26 12:35:31] Step: 1068, Training Logs: loss_final: 0.646854, loss_mean: 0.883886, proj_loss: -0.322405, loss_mean_cls: 0.085374, grad_norm: 0.894127 +Steps: 0%| | 1069/1000000 [04:28<68:49:09, 4.03it/s, grad_norm=0.894, loss_final=0.647, loss_mean=0.884, loss_mean_cls=0.0854, proj_loss=-0.322][2026-03-26 12:35:31] Step: 1069, Training Logs: loss_final: 0.626132, loss_mean: 0.878120, proj_loss: -0.335761, loss_mean_cls: 0.083773, grad_norm: 0.695475 +Steps: 0%| | 1070/1000000 [04:28<68:42:50, 4.04it/s, grad_norm=0.695, loss_final=0.626, loss_mean=0.878, loss_mean_cls=0.0838, proj_loss=-0.336][2026-03-26 12:35:32] Step: 1070, Training Logs: loss_final: 0.639274, loss_mean: 0.901738, proj_loss: -0.344085, loss_mean_cls: 0.081621, grad_norm: 0.835418 +Steps: 0%| | 1071/1000000 [04:29<68:41:52, 4.04it/s, grad_norm=0.835, loss_final=0.639, loss_mean=0.902, loss_mean_cls=0.0816, proj_loss=-0.344][2026-03-26 12:35:32] Step: 1071, Training Logs: loss_final: 0.621691, loss_mean: 0.851393, proj_loss: -0.315800, loss_mean_cls: 0.086098, grad_norm: 0.543115 +Steps: 0%| | 1072/1000000 [04:29<68:43:47, 4.04it/s, grad_norm=0.543, loss_final=0.622, loss_mean=0.851, loss_mean_cls=0.0861, proj_loss=-0.316][2026-03-26 12:35:32] Step: 1072, Training Logs: loss_final: 0.629942, loss_mean: 0.868210, proj_loss: -0.323250, loss_mean_cls: 0.084982, grad_norm: 0.806524 +Steps: 0%| | 1073/1000000 [04:29<68:44:46, 4.04it/s, grad_norm=0.807, loss_final=0.63, loss_mean=0.868, loss_mean_cls=0.085, proj_loss=-0.323][2026-03-26 12:35:32] Step: 1073, Training Logs: loss_final: 0.644832, loss_mean: 0.900941, proj_loss: -0.338035, loss_mean_cls: 0.081926, grad_norm: 0.737669 +Steps: 0%| | 1074/1000000 [04:29<68:42:39, 4.04it/s, grad_norm=0.738, loss_final=0.645, loss_mean=0.901, loss_mean_cls=0.0819, proj_loss=-0.338][2026-03-26 12:35:33] Step: 1074, Training Logs: loss_final: 0.630000, loss_mean: 0.878190, proj_loss: -0.331305, loss_mean_cls: 0.083114, grad_norm: 0.762077 +Steps: 0%| | 1075/1000000 [04:30<68:44:10, 4.04it/s, grad_norm=0.762, loss_final=0.63, loss_mean=0.878, loss_mean_cls=0.0831, proj_loss=-0.331][2026-03-26 12:35:33] Step: 1075, Training Logs: loss_final: 0.656760, loss_mean: 0.888689, proj_loss: -0.317603, loss_mean_cls: 0.085675, grad_norm: 0.734202 +Steps: 0%| | 1076/1000000 [04:30<68:42:14, 4.04it/s, grad_norm=0.734, loss_final=0.657, loss_mean=0.889, loss_mean_cls=0.0857, proj_loss=-0.318][2026-03-26 12:35:33] Step: 1076, Training Logs: loss_final: 0.624343, loss_mean: 0.870984, proj_loss: -0.330364, loss_mean_cls: 0.083722, grad_norm: 0.654100 +Steps: 0%| | 1077/1000000 [04:30<68:42:32, 4.04it/s, grad_norm=0.654, loss_final=0.624, loss_mean=0.871, loss_mean_cls=0.0837, proj_loss=-0.33][2026-03-26 12:35:33] Step: 1077, Training Logs: loss_final: 0.639933, loss_mean: 0.873851, proj_loss: -0.320543, loss_mean_cls: 0.086625, grad_norm: 0.767462 +Steps: 0%| | 1078/1000000 [04:30<68:40:46, 4.04it/s, grad_norm=0.767, loss_final=0.64, loss_mean=0.874, loss_mean_cls=0.0866, proj_loss=-0.321][2026-03-26 12:35:34] Step: 1078, Training Logs: loss_final: 0.624890, loss_mean: 0.868287, proj_loss: -0.328915, loss_mean_cls: 0.085517, grad_norm: 0.987382 +Steps: 0%| | 1079/1000000 [04:31<68:40:57, 4.04it/s, grad_norm=0.987, loss_final=0.625, loss_mean=0.868, loss_mean_cls=0.0855, proj_loss=-0.329][2026-03-26 12:35:34] Step: 1079, Training Logs: loss_final: 0.650587, loss_mean: 0.904256, proj_loss: -0.337062, loss_mean_cls: 0.083393, grad_norm: 0.524891 +Steps: 0%| | 1080/1000000 [04:31<68:42:54, 4.04it/s, grad_norm=0.525, loss_final=0.651, loss_mean=0.904, loss_mean_cls=0.0834, proj_loss=-0.337][2026-03-26 12:35:34] Step: 1080, Training Logs: loss_final: 0.641295, loss_mean: 0.898245, proj_loss: -0.340791, loss_mean_cls: 0.083841, grad_norm: 0.756158 +Steps: 0%| | 1081/1000000 [04:31<68:42:00, 4.04it/s, grad_norm=0.756, loss_final=0.641, loss_mean=0.898, loss_mean_cls=0.0838, proj_loss=-0.341][2026-03-26 12:35:34] Step: 1081, Training Logs: loss_final: 0.634814, loss_mean: 0.873788, proj_loss: -0.324831, loss_mean_cls: 0.085856, grad_norm: 0.670858 +Steps: 0%| | 1082/1000000 [04:31<68:41:57, 4.04it/s, grad_norm=0.671, loss_final=0.635, loss_mean=0.874, loss_mean_cls=0.0859, proj_loss=-0.325][2026-03-26 12:35:35] Step: 1082, Training Logs: loss_final: 0.634932, loss_mean: 0.873314, proj_loss: -0.326383, loss_mean_cls: 0.088001, grad_norm: 0.536538 +Steps: 0%| | 1083/1000000 [04:32<68:43:41, 4.04it/s, grad_norm=0.537, loss_final=0.635, loss_mean=0.873, loss_mean_cls=0.088, proj_loss=-0.326][2026-03-26 12:35:35] Step: 1083, Training Logs: loss_final: 0.627750, loss_mean: 0.871034, proj_loss: -0.327625, loss_mean_cls: 0.084341, grad_norm: 0.695763 +Steps: 0%| | 1084/1000000 [04:32<68:45:50, 4.04it/s, grad_norm=0.696, loss_final=0.628, loss_mean=0.871, loss_mean_cls=0.0843, proj_loss=-0.328][2026-03-26 12:35:35] Step: 1084, Training Logs: loss_final: 0.630057, loss_mean: 0.884776, proj_loss: -0.336984, loss_mean_cls: 0.082265, grad_norm: 0.724742 +Steps: 0%| | 1085/1000000 [04:32<68:44:41, 4.04it/s, grad_norm=0.725, loss_final=0.63, loss_mean=0.885, loss_mean_cls=0.0823, proj_loss=-0.337][2026-03-26 12:35:35] Step: 1085, Training Logs: loss_final: 0.628126, loss_mean: 0.863115, proj_loss: -0.321536, loss_mean_cls: 0.086547, grad_norm: 0.704739 +Steps: 0%| | 1086/1000000 [04:32<68:43:22, 4.04it/s, grad_norm=0.705, loss_final=0.628, loss_mean=0.863, loss_mean_cls=0.0865, proj_loss=-0.322][2026-03-26 12:35:36] Step: 1086, Training Logs: loss_final: 0.627488, loss_mean: 0.875140, proj_loss: -0.332367, loss_mean_cls: 0.084715, grad_norm: 0.568618 +Steps: 0%| | 1087/1000000 [04:33<68:42:45, 4.04it/s, grad_norm=0.569, loss_final=0.627, loss_mean=0.875, loss_mean_cls=0.0847, proj_loss=-0.332][2026-03-26 12:35:36] Step: 1087, Training Logs: loss_final: 0.627623, loss_mean: 0.865828, proj_loss: -0.325875, loss_mean_cls: 0.087671, grad_norm: 0.758316 +Steps: 0%| | 1088/1000000 [04:33<68:47:21, 4.03it/s, grad_norm=0.758, loss_final=0.628, loss_mean=0.866, loss_mean_cls=0.0877, proj_loss=-0.326][2026-03-26 12:35:36] Step: 1088, Training Logs: loss_final: 0.607152, loss_mean: 0.853230, proj_loss: -0.331722, loss_mean_cls: 0.085643, grad_norm: 0.531287 +Steps: 0%| | 1089/1000000 [04:33<68:45:33, 4.04it/s, grad_norm=0.531, loss_final=0.607, loss_mean=0.853, loss_mean_cls=0.0856, proj_loss=-0.332][2026-03-26 12:35:36] Step: 1089, Training Logs: loss_final: 0.620601, loss_mean: 0.853731, proj_loss: -0.319793, loss_mean_cls: 0.086662, grad_norm: 0.637669 +Steps: 0%| | 1090/1000000 [04:33<68:45:39, 4.04it/s, grad_norm=0.638, loss_final=0.621, loss_mean=0.854, loss_mean_cls=0.0867, proj_loss=-0.32][2026-03-26 12:35:37] Step: 1090, Training Logs: loss_final: 0.630491, loss_mean: 0.875757, proj_loss: -0.331214, loss_mean_cls: 0.085949, grad_norm: 0.617720 +Steps: 0%| | 1091/1000000 [04:33<68:40:45, 4.04it/s, grad_norm=0.618, loss_final=0.63, loss_mean=0.876, loss_mean_cls=0.0859, proj_loss=-0.331][2026-03-26 12:35:37] Step: 1091, Training Logs: loss_final: 0.620886, loss_mean: 0.879247, proj_loss: -0.341041, loss_mean_cls: 0.082680, grad_norm: 0.544795 +Steps: 0%| | 1092/1000000 [04:34<68:45:42, 4.04it/s, grad_norm=0.545, loss_final=0.621, loss_mean=0.879, loss_mean_cls=0.0827, proj_loss=-0.341][2026-03-26 12:35:37] Step: 1092, Training Logs: loss_final: 0.626673, loss_mean: 0.857159, proj_loss: -0.318263, loss_mean_cls: 0.087778, grad_norm: 0.557179 +Steps: 0%| | 1093/1000000 [04:34<68:41:34, 4.04it/s, grad_norm=0.557, loss_final=0.627, loss_mean=0.857, loss_mean_cls=0.0878, proj_loss=-0.318][2026-03-26 12:35:37] Step: 1093, Training Logs: loss_final: 0.639956, loss_mean: 0.873290, proj_loss: -0.319626, loss_mean_cls: 0.086292, grad_norm: 0.495139 +Steps: 0%| | 1094/1000000 [04:34<68:42:14, 4.04it/s, grad_norm=0.495, loss_final=0.64, loss_mean=0.873, loss_mean_cls=0.0863, proj_loss=-0.32][2026-03-26 12:35:38] Step: 1094, Training Logs: loss_final: 0.628332, loss_mean: 0.865453, proj_loss: -0.322875, loss_mean_cls: 0.085754, grad_norm: 0.369358 +Steps: 0%| | 1095/1000000 [04:34<68:40:35, 4.04it/s, grad_norm=0.369, loss_final=0.628, loss_mean=0.865, loss_mean_cls=0.0858, proj_loss=-0.323][2026-03-26 12:35:38] Step: 1095, Training Logs: loss_final: 0.634389, loss_mean: 0.880059, proj_loss: -0.329854, loss_mean_cls: 0.084183, grad_norm: 0.459964 +Steps: 0%| | 1096/1000000 [04:35<68:57:39, 4.02it/s, grad_norm=0.46, loss_final=0.634, loss_mean=0.88, loss_mean_cls=0.0842, proj_loss=-0.33][2026-03-26 12:35:38] Step: 1096, Training Logs: loss_final: 0.602451, loss_mean: 0.836432, proj_loss: -0.322355, loss_mean_cls: 0.088375, grad_norm: 0.370325 +Steps: 0%| | 1097/1000000 [04:35<68:50:27, 4.03it/s, grad_norm=0.37, loss_final=0.602, loss_mean=0.836, loss_mean_cls=0.0884, proj_loss=-0.322][2026-03-26 12:35:38] Step: 1097, Training Logs: loss_final: 0.625168, loss_mean: 0.866611, proj_loss: -0.327289, loss_mean_cls: 0.085846, grad_norm: 0.400652 +Steps: 0%| | 1098/1000000 [04:35<68:50:03, 4.03it/s, grad_norm=0.401, loss_final=0.625, loss_mean=0.867, loss_mean_cls=0.0858, proj_loss=-0.327][2026-03-26 12:35:39] Step: 1098, Training Logs: loss_final: 0.634921, loss_mean: 0.882715, proj_loss: -0.331964, loss_mean_cls: 0.084169, grad_norm: 0.493984 +Steps: 0%| | 1099/1000000 [04:35<68:45:12, 4.04it/s, grad_norm=0.494, loss_final=0.635, loss_mean=0.883, loss_mean_cls=0.0842, proj_loss=-0.332][2026-03-26 12:35:39] Step: 1099, Training Logs: loss_final: 0.616534, loss_mean: 0.861658, proj_loss: -0.330343, loss_mean_cls: 0.085219, grad_norm: 0.423465 +Steps: 0%| | 1100/1000000 [04:36<68:46:17, 4.03it/s, grad_norm=0.423, loss_final=0.617, loss_mean=0.862, loss_mean_cls=0.0852, proj_loss=-0.33][2026-03-26 12:35:39] Step: 1100, Training Logs: loss_final: 0.628090, loss_mean: 0.887502, proj_loss: -0.342010, loss_mean_cls: 0.082599, grad_norm: 0.444021 +Steps: 0%| | 1101/1000000 [04:36<68:42:45, 4.04it/s, grad_norm=0.444, loss_final=0.628, loss_mean=0.888, loss_mean_cls=0.0826, proj_loss=-0.342][2026-03-26 12:35:39] Step: 1101, Training Logs: loss_final: 0.611008, loss_mean: 0.862907, proj_loss: -0.335889, loss_mean_cls: 0.083990, grad_norm: 0.540821 +Steps: 0%| | 1102/1000000 [04:36<68:41:48, 4.04it/s, grad_norm=0.541, loss_final=0.611, loss_mean=0.863, loss_mean_cls=0.084, proj_loss=-0.336][2026-03-26 12:35:40] Step: 1102, Training Logs: loss_final: 0.628886, loss_mean: 0.863923, proj_loss: -0.320720, loss_mean_cls: 0.085682, grad_norm: 0.536975 +Steps: 0%| | 1103/1000000 [04:36<68:41:04, 4.04it/s, grad_norm=0.537, loss_final=0.629, loss_mean=0.864, loss_mean_cls=0.0857, proj_loss=-0.321][2026-03-26 12:35:40] Step: 1103, Training Logs: loss_final: 0.616841, loss_mean: 0.879773, proj_loss: -0.345539, loss_mean_cls: 0.082608, grad_norm: 0.392104 +Steps: 0%| | 1104/1000000 [04:37<68:40:43, 4.04it/s, grad_norm=0.392, loss_final=0.617, loss_mean=0.88, loss_mean_cls=0.0826, proj_loss=-0.346][2026-03-26 12:35:40] Step: 1104, Training Logs: loss_final: 0.642687, loss_mean: 0.876849, proj_loss: -0.320846, loss_mean_cls: 0.086683, grad_norm: 0.478455 +Steps: 0%| | 1105/1000000 [04:37<68:40:01, 4.04it/s, grad_norm=0.478, loss_final=0.643, loss_mean=0.877, loss_mean_cls=0.0867, proj_loss=-0.321][2026-03-26 12:35:40] Step: 1105, Training Logs: loss_final: 0.634217, loss_mean: 0.885514, proj_loss: -0.335235, loss_mean_cls: 0.083938, grad_norm: 0.480205 +Steps: 0%| | 1106/1000000 [04:37<68:38:25, 4.04it/s, grad_norm=0.48, loss_final=0.634, loss_mean=0.886, loss_mean_cls=0.0839, proj_loss=-0.335][2026-03-26 12:35:41] Step: 1106, Training Logs: loss_final: 0.620291, loss_mean: 0.866000, proj_loss: -0.330500, loss_mean_cls: 0.084791, grad_norm: 0.448817 +Steps: 0%| | 1107/1000000 [04:37<68:39:21, 4.04it/s, grad_norm=0.449, loss_final=0.62, loss_mean=0.866, loss_mean_cls=0.0848, proj_loss=-0.331][2026-03-26 12:35:41] Step: 1107, Training Logs: loss_final: 0.623990, loss_mean: 0.882450, proj_loss: -0.341456, loss_mean_cls: 0.082995, grad_norm: 0.396632 +Steps: 0%| | 1108/1000000 [04:38<68:41:43, 4.04it/s, grad_norm=0.397, loss_final=0.624, loss_mean=0.882, loss_mean_cls=0.083, proj_loss=-0.341][2026-03-26 12:35:41] Step: 1108, Training Logs: loss_final: 0.623346, loss_mean: 0.861933, proj_loss: -0.324161, loss_mean_cls: 0.085573, grad_norm: 0.669825 +Steps: 0%| | 1109/1000000 [04:38<68:43:47, 4.04it/s, grad_norm=0.67, loss_final=0.623, loss_mean=0.862, loss_mean_cls=0.0856, proj_loss=-0.324][2026-03-26 12:35:41] Step: 1109, Training Logs: loss_final: 0.626690, loss_mean: 0.875093, proj_loss: -0.333028, loss_mean_cls: 0.084625, grad_norm: 0.335340 +Steps: 0%| | 1110/1000000 [04:38<69:31:33, 3.99it/s, grad_norm=0.335, loss_final=0.627, loss_mean=0.875, loss_mean_cls=0.0846, proj_loss=-0.333][2026-03-26 12:35:42] Step: 1110, Training Logs: loss_final: 0.620250, loss_mean: 0.861325, proj_loss: -0.327522, loss_mean_cls: 0.086447, grad_norm: 0.744144 +Steps: 0%| | 1111/1000000 [04:38<69:45:41, 3.98it/s, grad_norm=0.744, loss_final=0.62, loss_mean=0.861, loss_mean_cls=0.0864, proj_loss=-0.328][2026-03-26 12:35:42] Step: 1111, Training Logs: loss_final: 0.641141, loss_mean: 0.869836, proj_loss: -0.316252, loss_mean_cls: 0.087557, grad_norm: 0.535399 +Steps: 0%| | 1112/1000000 [04:39<69:28:01, 3.99it/s, grad_norm=0.535, loss_final=0.641, loss_mean=0.87, loss_mean_cls=0.0876, proj_loss=-0.316][2026-03-26 12:35:42] Step: 1112, Training Logs: loss_final: 0.615361, loss_mean: 0.863700, proj_loss: -0.333290, loss_mean_cls: 0.084952, grad_norm: 0.735617 +Steps: 0%| | 1113/1000000 [04:39<69:16:01, 4.01it/s, grad_norm=0.736, loss_final=0.615, loss_mean=0.864, loss_mean_cls=0.085, proj_loss=-0.333][2026-03-26 12:35:42] Step: 1113, Training Logs: loss_final: 0.633082, loss_mean: 0.886260, proj_loss: -0.337732, loss_mean_cls: 0.084554, grad_norm: 0.520069 +Steps: 0%| | 1114/1000000 [04:39<69:07:49, 4.01it/s, grad_norm=0.52, loss_final=0.633, loss_mean=0.886, loss_mean_cls=0.0846, proj_loss=-0.338][2026-03-26 12:35:43] Step: 1114, Training Logs: loss_final: 0.615952, loss_mean: 0.860294, proj_loss: -0.329822, loss_mean_cls: 0.085480, grad_norm: 0.570281 +Steps: 0%| | 1115/1000000 [04:39<68:58:11, 4.02it/s, grad_norm=0.57, loss_final=0.616, loss_mean=0.86, loss_mean_cls=0.0855, proj_loss=-0.33][2026-03-26 12:35:43] Step: 1115, Training Logs: loss_final: 0.611761, loss_mean: 0.867100, proj_loss: -0.339827, loss_mean_cls: 0.084488, grad_norm: 0.680259 +Steps: 0%| | 1116/1000000 [04:40<68:52:58, 4.03it/s, grad_norm=0.68, loss_final=0.612, loss_mean=0.867, loss_mean_cls=0.0845, proj_loss=-0.34][2026-03-26 12:35:43] Step: 1116, Training Logs: loss_final: 0.618203, loss_mean: 0.873380, proj_loss: -0.338330, loss_mean_cls: 0.083153, grad_norm: 0.510702 +Steps: 0%| | 1117/1000000 [04:40<68:47:36, 4.03it/s, grad_norm=0.511, loss_final=0.618, loss_mean=0.873, loss_mean_cls=0.0832, proj_loss=-0.338][2026-03-26 12:35:43] Step: 1117, Training Logs: loss_final: 0.625927, loss_mean: 0.867392, proj_loss: -0.326472, loss_mean_cls: 0.085007, grad_norm: 0.575279 +Steps: 0%| | 1118/1000000 [04:40<68:43:20, 4.04it/s, grad_norm=0.575, loss_final=0.626, loss_mean=0.867, loss_mean_cls=0.085, proj_loss=-0.326][2026-03-26 12:35:44] Step: 1118, Training Logs: loss_final: 0.624167, loss_mean: 0.860017, proj_loss: -0.322287, loss_mean_cls: 0.086437, grad_norm: 0.470913 +Steps: 0%| | 1119/1000000 [04:40<68:41:50, 4.04it/s, grad_norm=0.471, loss_final=0.624, loss_mean=0.86, loss_mean_cls=0.0864, proj_loss=-0.322][2026-03-26 12:35:44] Step: 1119, Training Logs: loss_final: 0.620149, loss_mean: 0.881464, proj_loss: -0.343862, loss_mean_cls: 0.082547, grad_norm: 0.576050 +Steps: 0%| | 1120/1000000 [04:41<68:40:26, 4.04it/s, grad_norm=0.576, loss_final=0.62, loss_mean=0.881, loss_mean_cls=0.0825, proj_loss=-0.344][2026-03-26 12:35:44] Step: 1120, Training Logs: loss_final: 0.623135, loss_mean: 0.883763, proj_loss: -0.343429, loss_mean_cls: 0.082801, grad_norm: 0.506541 +Steps: 0%| | 1121/1000000 [04:41<68:42:04, 4.04it/s, grad_norm=0.507, loss_final=0.623, loss_mean=0.884, loss_mean_cls=0.0828, proj_loss=-0.343][2026-03-26 12:35:44] Step: 1121, Training Logs: loss_final: 0.631920, loss_mean: 0.872444, proj_loss: -0.326239, loss_mean_cls: 0.085714, grad_norm: 0.563257 +Steps: 0%| | 1122/1000000 [04:41<68:42:23, 4.04it/s, grad_norm=0.563, loss_final=0.632, loss_mean=0.872, loss_mean_cls=0.0857, proj_loss=-0.326][2026-03-26 12:35:45] Step: 1122, Training Logs: loss_final: 0.636081, loss_mean: 0.880707, proj_loss: -0.329527, loss_mean_cls: 0.084900, grad_norm: 0.471701 +Steps: 0%| | 1123/1000000 [04:41<68:43:20, 4.04it/s, grad_norm=0.472, loss_final=0.636, loss_mean=0.881, loss_mean_cls=0.0849, proj_loss=-0.33][2026-03-26 12:35:45] Step: 1123, Training Logs: loss_final: 0.634727, loss_mean: 0.886963, proj_loss: -0.336173, loss_mean_cls: 0.083936, grad_norm: 0.636536 +Steps: 0%| | 1124/1000000 [04:42<68:40:42, 4.04it/s, grad_norm=0.637, loss_final=0.635, loss_mean=0.887, loss_mean_cls=0.0839, proj_loss=-0.336][2026-03-26 12:35:45] Step: 1124, Training Logs: loss_final: 0.619721, loss_mean: 0.878910, proj_loss: -0.342752, loss_mean_cls: 0.083563, grad_norm: 0.387301 +Steps: 0%| | 1125/1000000 [04:42<68:43:07, 4.04it/s, grad_norm=0.387, loss_final=0.62, loss_mean=0.879, loss_mean_cls=0.0836, proj_loss=-0.343][2026-03-26 12:35:45] Step: 1125, Training Logs: loss_final: 0.617988, loss_mean: 0.865081, proj_loss: -0.332590, loss_mean_cls: 0.085497, grad_norm: 0.718451 +Steps: 0%| | 1126/1000000 [04:42<68:39:49, 4.04it/s, grad_norm=0.718, loss_final=0.618, loss_mean=0.865, loss_mean_cls=0.0855, proj_loss=-0.333][2026-03-26 12:35:46] Step: 1126, Training Logs: loss_final: 0.606442, loss_mean: 0.865382, proj_loss: -0.342196, loss_mean_cls: 0.083257, grad_norm: 0.438530 +Steps: 0%| | 1127/1000000 [04:42<68:36:42, 4.04it/s, grad_norm=0.439, loss_final=0.606, loss_mean=0.865, loss_mean_cls=0.0833, proj_loss=-0.342][2026-03-26 12:35:46] Step: 1127, Training Logs: loss_final: 0.628207, loss_mean: 0.872513, proj_loss: -0.329561, loss_mean_cls: 0.085255, grad_norm: 0.425318 +Steps: 0%| | 1128/1000000 [04:43<68:38:13, 4.04it/s, grad_norm=0.425, loss_final=0.628, loss_mean=0.873, loss_mean_cls=0.0853, proj_loss=-0.33][2026-03-26 12:35:46] Step: 1128, Training Logs: loss_final: 0.623332, loss_mean: 0.868664, proj_loss: -0.329704, loss_mean_cls: 0.084372, grad_norm: 0.502693 +Steps: 0%| | 1129/1000000 [04:43<68:40:00, 4.04it/s, grad_norm=0.503, loss_final=0.623, loss_mean=0.869, loss_mean_cls=0.0844, proj_loss=-0.33][2026-03-26 12:35:46] Step: 1129, Training Logs: loss_final: 0.619848, loss_mean: 0.861214, proj_loss: -0.327638, loss_mean_cls: 0.086272, grad_norm: 0.558311 +Steps: 0%| | 1130/1000000 [04:43<68:38:51, 4.04it/s, grad_norm=0.558, loss_final=0.62, loss_mean=0.861, loss_mean_cls=0.0863, proj_loss=-0.328][2026-03-26 12:35:47] Step: 1130, Training Logs: loss_final: 0.634306, loss_mean: 0.888779, proj_loss: -0.336966, loss_mean_cls: 0.082493, grad_norm: 0.494949 +Steps: 0%| | 1131/1000000 [04:43<68:38:51, 4.04it/s, grad_norm=0.495, loss_final=0.634, loss_mean=0.889, loss_mean_cls=0.0825, proj_loss=-0.337][2026-03-26 12:35:47] Step: 1131, Training Logs: loss_final: 0.612039, loss_mean: 0.860834, proj_loss: -0.333467, loss_mean_cls: 0.084673, grad_norm: 0.386787 +Steps: 0%| | 1132/1000000 [04:44<68:42:46, 4.04it/s, grad_norm=0.387, loss_final=0.612, loss_mean=0.861, loss_mean_cls=0.0847, proj_loss=-0.333][2026-03-26 12:35:47] Step: 1132, Training Logs: loss_final: 0.616035, loss_mean: 0.863427, proj_loss: -0.334371, loss_mean_cls: 0.086979, grad_norm: 0.487008 +Steps: 0%| | 1133/1000000 [04:44<68:41:34, 4.04it/s, grad_norm=0.487, loss_final=0.616, loss_mean=0.863, loss_mean_cls=0.087, proj_loss=-0.334][2026-03-26 12:35:47] Step: 1133, Training Logs: loss_final: 0.610633, loss_mean: 0.851933, proj_loss: -0.326925, loss_mean_cls: 0.085625, grad_norm: 0.524411 +Steps: 0%| | 1134/1000000 [04:44<68:41:41, 4.04it/s, grad_norm=0.524, loss_final=0.611, loss_mean=0.852, loss_mean_cls=0.0856, proj_loss=-0.327][2026-03-26 12:35:48] Step: 1134, Training Logs: loss_final: 0.625942, loss_mean: 0.878743, proj_loss: -0.335928, loss_mean_cls: 0.083128, grad_norm: 0.466330 +Steps: 0%| | 1135/1000000 [04:44<68:39:24, 4.04it/s, grad_norm=0.466, loss_final=0.626, loss_mean=0.879, loss_mean_cls=0.0831, proj_loss=-0.336][2026-03-26 12:35:48] Step: 1135, Training Logs: loss_final: 0.617926, loss_mean: 0.873435, proj_loss: -0.339159, loss_mean_cls: 0.083650, grad_norm: 0.390003 +Steps: 0%| | 1136/1000000 [04:45<68:42:19, 4.04it/s, grad_norm=0.39, loss_final=0.618, loss_mean=0.873, loss_mean_cls=0.0837, proj_loss=-0.339][2026-03-26 12:35:48] Step: 1136, Training Logs: loss_final: 0.615706, loss_mean: 0.870981, proj_loss: -0.338643, loss_mean_cls: 0.083368, grad_norm: 0.400937 +Steps: 0%| | 1137/1000000 [04:45<68:43:07, 4.04it/s, grad_norm=0.401, loss_final=0.616, loss_mean=0.871, loss_mean_cls=0.0834, proj_loss=-0.339][2026-03-26 12:35:48] Step: 1137, Training Logs: loss_final: 0.622472, loss_mean: 0.885525, proj_loss: -0.345067, loss_mean_cls: 0.082014, grad_norm: 0.366711 +Steps: 0%| | 1138/1000000 [04:45<68:42:35, 4.04it/s, grad_norm=0.367, loss_final=0.622, loss_mean=0.886, loss_mean_cls=0.082, proj_loss=-0.345][2026-03-26 12:35:49] Step: 1138, Training Logs: loss_final: 0.627970, loss_mean: 0.885235, proj_loss: -0.340779, loss_mean_cls: 0.083514, grad_norm: 0.533395 +Steps: 0%| | 1139/1000000 [04:45<68:39:03, 4.04it/s, grad_norm=0.533, loss_final=0.628, loss_mean=0.885, loss_mean_cls=0.0835, proj_loss=-0.341][2026-03-26 12:35:49] Step: 1139, Training Logs: loss_final: 0.628714, loss_mean: 0.863976, proj_loss: -0.321995, loss_mean_cls: 0.086732, grad_norm: 0.665151 +Steps: 0%| | 1140/1000000 [04:46<68:41:05, 4.04it/s, grad_norm=0.665, loss_final=0.629, loss_mean=0.864, loss_mean_cls=0.0867, proj_loss=-0.322][2026-03-26 12:35:49] Step: 1140, Training Logs: loss_final: 0.621723, loss_mean: 0.874606, proj_loss: -0.336356, loss_mean_cls: 0.083473, grad_norm: 0.536857 +Steps: 0%| | 1141/1000000 [04:46<68:37:54, 4.04it/s, grad_norm=0.537, loss_final=0.622, loss_mean=0.875, loss_mean_cls=0.0835, proj_loss=-0.336][2026-03-26 12:35:49] Step: 1141, Training Logs: loss_final: 0.620581, loss_mean: 0.876445, proj_loss: -0.340184, loss_mean_cls: 0.084319, grad_norm: 0.735103 +Steps: 0%| | 1142/1000000 [04:46<68:44:47, 4.04it/s, grad_norm=0.735, loss_final=0.621, loss_mean=0.876, loss_mean_cls=0.0843, proj_loss=-0.34][2026-03-26 12:35:50] Step: 1142, Training Logs: loss_final: 0.620843, loss_mean: 0.860657, proj_loss: -0.326662, loss_mean_cls: 0.086848, grad_norm: 0.752508 +Steps: 0%| | 1143/1000000 [04:46<68:42:51, 4.04it/s, grad_norm=0.753, loss_final=0.621, loss_mean=0.861, loss_mean_cls=0.0868, proj_loss=-0.327][2026-03-26 12:35:50] Step: 1143, Training Logs: loss_final: 0.614387, loss_mean: 0.873236, proj_loss: -0.342399, loss_mean_cls: 0.083550, grad_norm: 0.877988 +Steps: 0%| | 1144/1000000 [04:47<68:44:37, 4.04it/s, grad_norm=0.878, loss_final=0.614, loss_mean=0.873, loss_mean_cls=0.0835, proj_loss=-0.342][2026-03-26 12:35:50] Step: 1144, Training Logs: loss_final: 0.628887, loss_mean: 0.877175, proj_loss: -0.333267, loss_mean_cls: 0.084979, grad_norm: 0.833602 +Steps: 0%| | 1145/1000000 [04:47<68:41:35, 4.04it/s, grad_norm=0.834, loss_final=0.629, loss_mean=0.877, loss_mean_cls=0.085, proj_loss=-0.333][2026-03-26 12:35:50] Step: 1145, Training Logs: loss_final: 0.613039, loss_mean: 0.866646, proj_loss: -0.338394, loss_mean_cls: 0.084787, grad_norm: 0.518422 +Steps: 0%| | 1146/1000000 [04:47<68:41:45, 4.04it/s, grad_norm=0.518, loss_final=0.613, loss_mean=0.867, loss_mean_cls=0.0848, proj_loss=-0.338][2026-03-26 12:35:51] Step: 1146, Training Logs: loss_final: 0.612775, loss_mean: 0.877977, proj_loss: -0.347801, loss_mean_cls: 0.082598, grad_norm: 0.580239 +Steps: 0%| | 1147/1000000 [04:47<68:42:31, 4.04it/s, grad_norm=0.58, loss_final=0.613, loss_mean=0.878, loss_mean_cls=0.0826, proj_loss=-0.348][2026-03-26 12:35:51] Step: 1147, Training Logs: loss_final: 0.612159, loss_mean: 0.852774, proj_loss: -0.327228, loss_mean_cls: 0.086613, grad_norm: 0.464988 +Steps: 0%| | 1148/1000000 [04:48<68:40:12, 4.04it/s, grad_norm=0.465, loss_final=0.612, loss_mean=0.853, loss_mean_cls=0.0866, proj_loss=-0.327][2026-03-26 12:35:51] Step: 1148, Training Logs: loss_final: 0.616539, loss_mean: 0.869816, proj_loss: -0.336807, loss_mean_cls: 0.083531, grad_norm: 0.508180 +Steps: 0%| | 1149/1000000 [04:48<68:48:12, 4.03it/s, grad_norm=0.508, loss_final=0.617, loss_mean=0.87, loss_mean_cls=0.0835, proj_loss=-0.337][2026-03-26 12:35:51] Step: 1149, Training Logs: loss_final: 0.628022, loss_mean: 0.888072, proj_loss: -0.342259, loss_mean_cls: 0.082209, grad_norm: 0.487345 +Steps: 0%| | 1150/1000000 [04:48<68:48:37, 4.03it/s, grad_norm=0.487, loss_final=0.628, loss_mean=0.888, loss_mean_cls=0.0822, proj_loss=-0.342][2026-03-26 12:35:52] Step: 1150, Training Logs: loss_final: 0.635249, loss_mean: 0.885065, proj_loss: -0.333512, loss_mean_cls: 0.083695, grad_norm: 0.460157 +Steps: 0%| | 1151/1000000 [04:48<68:44:21, 4.04it/s, grad_norm=0.46, loss_final=0.635, loss_mean=0.885, loss_mean_cls=0.0837, proj_loss=-0.334][2026-03-26 12:35:52] Step: 1151, Training Logs: loss_final: 0.638023, loss_mean: 0.880968, proj_loss: -0.328806, loss_mean_cls: 0.085860, grad_norm: 0.420489 +Steps: 0%| | 1152/1000000 [04:49<68:42:43, 4.04it/s, grad_norm=0.42, loss_final=0.638, loss_mean=0.881, loss_mean_cls=0.0859, proj_loss=-0.329][2026-03-26 12:35:52] Step: 1152, Training Logs: loss_final: 0.608963, loss_mean: 0.870178, proj_loss: -0.344346, loss_mean_cls: 0.083132, grad_norm: 0.383540 +Steps: 0%| | 1153/1000000 [04:49<68:43:51, 4.04it/s, grad_norm=0.384, loss_final=0.609, loss_mean=0.87, loss_mean_cls=0.0831, proj_loss=-0.344][2026-03-26 12:35:52] Step: 1153, Training Logs: loss_final: 0.634167, loss_mean: 0.887214, proj_loss: -0.336992, loss_mean_cls: 0.083945, grad_norm: 0.491092 +Steps: 0%| | 1154/1000000 [04:49<68:42:13, 4.04it/s, grad_norm=0.491, loss_final=0.634, loss_mean=0.887, loss_mean_cls=0.0839, proj_loss=-0.337][2026-03-26 12:35:53] Step: 1154, Training Logs: loss_final: 0.631029, loss_mean: 0.891942, proj_loss: -0.343587, loss_mean_cls: 0.082674, grad_norm: 0.441443 +Steps: 0%| | 1155/1000000 [04:49<68:40:03, 4.04it/s, grad_norm=0.441, loss_final=0.631, loss_mean=0.892, loss_mean_cls=0.0827, proj_loss=-0.344][2026-03-26 12:35:53] Step: 1155, Training Logs: loss_final: 0.638010, loss_mean: 0.899436, proj_loss: -0.343470, loss_mean_cls: 0.082044, grad_norm: 0.327281 +Steps: 0%| | 1156/1000000 [04:50<68:40:52, 4.04it/s, grad_norm=0.327, loss_final=0.638, loss_mean=0.899, loss_mean_cls=0.082, proj_loss=-0.343][2026-03-26 12:35:53] Step: 1156, Training Logs: loss_final: 0.620416, loss_mean: 0.865782, proj_loss: -0.330835, loss_mean_cls: 0.085468, grad_norm: 0.629481 +Steps: 0%| | 1157/1000000 [04:50<68:42:47, 4.04it/s, grad_norm=0.629, loss_final=0.62, loss_mean=0.866, loss_mean_cls=0.0855, proj_loss=-0.331][2026-03-26 12:35:53] Step: 1157, Training Logs: loss_final: 0.617988, loss_mean: 0.862723, proj_loss: -0.329458, loss_mean_cls: 0.084723, grad_norm: 0.460117 +Steps: 0%| | 1158/1000000 [04:50<68:46:50, 4.03it/s, grad_norm=0.46, loss_final=0.618, loss_mean=0.863, loss_mean_cls=0.0847, proj_loss=-0.329][2026-03-26 12:35:53] Step: 1158, Training Logs: loss_final: 0.620570, loss_mean: 0.873469, proj_loss: -0.336707, loss_mean_cls: 0.083808, grad_norm: 0.756851 +Steps: 0%| | 1159/1000000 [04:50<68:49:29, 4.03it/s, grad_norm=0.757, loss_final=0.621, loss_mean=0.873, loss_mean_cls=0.0838, proj_loss=-0.337][2026-03-26 12:35:54] Step: 1159, Training Logs: loss_final: 0.609347, loss_mean: 0.884511, proj_loss: -0.355616, loss_mean_cls: 0.080452, grad_norm: 0.576268 +Steps: 0%| | 1160/1000000 [04:51<68:47:45, 4.03it/s, grad_norm=0.576, loss_final=0.609, loss_mean=0.885, loss_mean_cls=0.0805, proj_loss=-0.356][2026-03-26 12:35:54] Step: 1160, Training Logs: loss_final: 0.619801, loss_mean: 0.858571, proj_loss: -0.324334, loss_mean_cls: 0.085564, grad_norm: 0.541412 +Steps: 0%| | 1161/1000000 [04:51<68:45:54, 4.03it/s, grad_norm=0.541, loss_final=0.62, loss_mean=0.859, loss_mean_cls=0.0856, proj_loss=-0.324][2026-03-26 12:35:54] Step: 1161, Training Logs: loss_final: 0.633852, loss_mean: 0.904169, proj_loss: -0.350530, loss_mean_cls: 0.080213, grad_norm: 0.453064 +Steps: 0%| | 1162/1000000 [04:51<68:51:02, 4.03it/s, grad_norm=0.453, loss_final=0.634, loss_mean=0.904, loss_mean_cls=0.0802, proj_loss=-0.351][2026-03-26 12:35:54] Step: 1162, Training Logs: loss_final: 0.627892, loss_mean: 0.876217, proj_loss: -0.331820, loss_mean_cls: 0.083494, grad_norm: 0.482154 +Steps: 0%| | 1163/1000000 [04:51<69:04:22, 4.02it/s, grad_norm=0.482, loss_final=0.628, loss_mean=0.876, loss_mean_cls=0.0835, proj_loss=-0.332][2026-03-26 12:35:55] Step: 1163, Training Logs: loss_final: 0.614527, loss_mean: 0.879625, proj_loss: -0.347772, loss_mean_cls: 0.082674, grad_norm: 0.413148 +Steps: 0%| | 1164/1000000 [04:52<68:55:17, 4.03it/s, grad_norm=0.413, loss_final=0.615, loss_mean=0.88, loss_mean_cls=0.0827, proj_loss=-0.348][2026-03-26 12:35:55] Step: 1164, Training Logs: loss_final: 0.627956, loss_mean: 0.858120, proj_loss: -0.318664, loss_mean_cls: 0.088499, grad_norm: 0.413093 +Steps: 0%| | 1165/1000000 [04:52<68:50:58, 4.03it/s, grad_norm=0.413, loss_final=0.628, loss_mean=0.858, loss_mean_cls=0.0885, proj_loss=-0.319][2026-03-26 12:35:55] Step: 1165, Training Logs: loss_final: 0.622554, loss_mean: 0.890794, proj_loss: -0.349521, loss_mean_cls: 0.081280, grad_norm: 0.434631 +Steps: 0%| | 1166/1000000 [04:52<68:47:16, 4.03it/s, grad_norm=0.435, loss_final=0.623, loss_mean=0.891, loss_mean_cls=0.0813, proj_loss=-0.35][2026-03-26 12:35:55] Step: 1166, Training Logs: loss_final: 0.612638, loss_mean: 0.855671, proj_loss: -0.329686, loss_mean_cls: 0.086653, grad_norm: 0.538305 +Steps: 0%| | 1167/1000000 [04:52<68:48:28, 4.03it/s, grad_norm=0.538, loss_final=0.613, loss_mean=0.856, loss_mean_cls=0.0867, proj_loss=-0.33][2026-03-26 12:35:56] Step: 1167, Training Logs: loss_final: 0.617014, loss_mean: 0.864333, proj_loss: -0.332894, loss_mean_cls: 0.085576, grad_norm: 0.390013 +Steps: 0%| | 1168/1000000 [04:53<68:47:30, 4.03it/s, grad_norm=0.39, loss_final=0.617, loss_mean=0.864, loss_mean_cls=0.0856, proj_loss=-0.333][2026-03-26 12:35:56] Step: 1168, Training Logs: loss_final: 0.612167, loss_mean: 0.854257, proj_loss: -0.328251, loss_mean_cls: 0.086161, grad_norm: 0.551041 +Steps: 0%| | 1169/1000000 [04:53<68:44:11, 4.04it/s, grad_norm=0.551, loss_final=0.612, loss_mean=0.854, loss_mean_cls=0.0862, proj_loss=-0.328][2026-03-26 12:35:56] Step: 1169, Training Logs: loss_final: 0.613834, loss_mean: 0.858239, proj_loss: -0.329943, loss_mean_cls: 0.085538, grad_norm: 0.487411 +Steps: 0%| | 1170/1000000 [04:53<68:43:56, 4.04it/s, grad_norm=0.487, loss_final=0.614, loss_mean=0.858, loss_mean_cls=0.0855, proj_loss=-0.33][2026-03-26 12:35:56] Step: 1170, Training Logs: loss_final: 0.614293, loss_mean: 0.862504, proj_loss: -0.334025, loss_mean_cls: 0.085813, grad_norm: 0.660718 +Steps: 0%| | 1171/1000000 [04:53<68:42:34, 4.04it/s, grad_norm=0.661, loss_final=0.614, loss_mean=0.863, loss_mean_cls=0.0858, proj_loss=-0.334][2026-03-26 12:35:57] Step: 1171, Training Logs: loss_final: 0.608794, loss_mean: 0.885764, proj_loss: -0.357665, loss_mean_cls: 0.080696, grad_norm: 0.661494 +Steps: 0%| | 1172/1000000 [04:54<68:41:55, 4.04it/s, grad_norm=0.661, loss_final=0.609, loss_mean=0.886, loss_mean_cls=0.0807, proj_loss=-0.358][2026-03-26 12:35:57] Step: 1172, Training Logs: loss_final: 0.622584, loss_mean: 0.882786, proj_loss: -0.343827, loss_mean_cls: 0.083625, grad_norm: 0.850397 +Steps: 0%| | 1173/1000000 [04:54<68:43:21, 4.04it/s, grad_norm=0.85, loss_final=0.623, loss_mean=0.883, loss_mean_cls=0.0836, proj_loss=-0.344][2026-03-26 12:35:57] Step: 1173, Training Logs: loss_final: 0.623489, loss_mean: 0.867379, proj_loss: -0.329657, loss_mean_cls: 0.085767, grad_norm: 0.487572 +Steps: 0%| | 1174/1000000 [04:54<68:42:20, 4.04it/s, grad_norm=0.488, loss_final=0.623, loss_mean=0.867, loss_mean_cls=0.0858, proj_loss=-0.33][2026-03-26 12:35:57] Step: 1174, Training Logs: loss_final: 0.618723, loss_mean: 0.876269, proj_loss: -0.340988, loss_mean_cls: 0.083443, grad_norm: 0.622365 +Steps: 0%| | 1175/1000000 [04:54<68:43:03, 4.04it/s, grad_norm=0.622, loss_final=0.619, loss_mean=0.876, loss_mean_cls=0.0834, proj_loss=-0.341][2026-03-26 12:35:58] Step: 1175, Training Logs: loss_final: 0.626340, loss_mean: 0.876101, proj_loss: -0.335552, loss_mean_cls: 0.085791, grad_norm: 0.606091 +Steps: 0%| | 1176/1000000 [04:55<68:41:33, 4.04it/s, grad_norm=0.606, loss_final=0.626, loss_mean=0.876, loss_mean_cls=0.0858, proj_loss=-0.336][2026-03-26 12:35:58] Step: 1176, Training Logs: loss_final: 0.617859, loss_mean: 0.864309, proj_loss: -0.332156, loss_mean_cls: 0.085706, grad_norm: 0.862847 +Steps: 0%| | 1177/1000000 [04:55<68:41:16, 4.04it/s, grad_norm=0.863, loss_final=0.618, loss_mean=0.864, loss_mean_cls=0.0857, proj_loss=-0.332][2026-03-26 12:35:58] Step: 1177, Training Logs: loss_final: 0.619386, loss_mean: 0.864299, proj_loss: -0.330151, loss_mean_cls: 0.085237, grad_norm: 0.650158 +Steps: 0%| | 1178/1000000 [04:55<68:39:26, 4.04it/s, grad_norm=0.65, loss_final=0.619, loss_mean=0.864, loss_mean_cls=0.0852, proj_loss=-0.33][2026-03-26 12:35:58] Step: 1178, Training Logs: loss_final: 0.626619, loss_mean: 0.873738, proj_loss: -0.332362, loss_mean_cls: 0.085243, grad_norm: 0.637982 +Steps: 0%| | 1179/1000000 [04:55<68:42:26, 4.04it/s, grad_norm=0.638, loss_final=0.627, loss_mean=0.874, loss_mean_cls=0.0852, proj_loss=-0.332][2026-03-26 12:35:59] Step: 1179, Training Logs: loss_final: 0.629422, loss_mean: 0.881083, proj_loss: -0.334934, loss_mean_cls: 0.083273, grad_norm: 0.544834 +Steps: 0%| | 1180/1000000 [04:56<68:40:36, 4.04it/s, grad_norm=0.545, loss_final=0.629, loss_mean=0.881, loss_mean_cls=0.0833, proj_loss=-0.335][2026-03-26 12:35:59] Step: 1180, Training Logs: loss_final: 0.617209, loss_mean: 0.880277, proj_loss: -0.346564, loss_mean_cls: 0.083496, grad_norm: 0.602399 +Steps: 0%| | 1181/1000000 [04:56<68:39:01, 4.04it/s, grad_norm=0.602, loss_final=0.617, loss_mean=0.88, loss_mean_cls=0.0835, proj_loss=-0.347][2026-03-26 12:35:59] Step: 1181, Training Logs: loss_final: 0.637231, loss_mean: 0.903843, proj_loss: -0.348277, loss_mean_cls: 0.081665, grad_norm: 0.558539 +Steps: 0%| | 1182/1000000 [04:56<68:41:37, 4.04it/s, grad_norm=0.559, loss_final=0.637, loss_mean=0.904, loss_mean_cls=0.0817, proj_loss=-0.348][2026-03-26 12:35:59] Step: 1182, Training Logs: loss_final: 0.605183, loss_mean: 0.849076, proj_loss: -0.330354, loss_mean_cls: 0.086461, grad_norm: 0.517014 +Steps: 0%| | 1183/1000000 [04:56<68:37:57, 4.04it/s, grad_norm=0.517, loss_final=0.605, loss_mean=0.849, loss_mean_cls=0.0865, proj_loss=-0.33][2026-03-26 12:36:00] Step: 1183, Training Logs: loss_final: 0.633551, loss_mean: 0.886954, proj_loss: -0.336777, loss_mean_cls: 0.083374, grad_norm: 0.646356 +Steps: 0%| | 1184/1000000 [04:57<68:41:24, 4.04it/s, grad_norm=0.646, loss_final=0.634, loss_mean=0.887, loss_mean_cls=0.0834, proj_loss=-0.337][2026-03-26 12:36:00] Step: 1184, Training Logs: loss_final: 0.624965, loss_mean: 0.884046, proj_loss: -0.342775, loss_mean_cls: 0.083695, grad_norm: 0.590943 +Steps: 0%| | 1185/1000000 [04:57<68:40:35, 4.04it/s, grad_norm=0.591, loss_final=0.625, loss_mean=0.884, loss_mean_cls=0.0837, proj_loss=-0.343][2026-03-26 12:36:00] Step: 1185, Training Logs: loss_final: 0.613366, loss_mean: 0.869189, proj_loss: -0.340029, loss_mean_cls: 0.084206, grad_norm: 0.553084 +Steps: 0%| | 1186/1000000 [04:57<68:39:12, 4.04it/s, grad_norm=0.553, loss_final=0.613, loss_mean=0.869, loss_mean_cls=0.0842, proj_loss=-0.34][2026-03-26 12:36:00] Step: 1186, Training Logs: loss_final: 0.628533, loss_mean: 0.871821, proj_loss: -0.329857, loss_mean_cls: 0.086569, grad_norm: 0.472617 +Steps: 0%| | 1187/1000000 [04:57<68:38:44, 4.04it/s, grad_norm=0.473, loss_final=0.629, loss_mean=0.872, loss_mean_cls=0.0866, proj_loss=-0.33][2026-03-26 12:36:01] Step: 1187, Training Logs: loss_final: 0.603182, loss_mean: 0.836633, proj_loss: -0.321770, loss_mean_cls: 0.088319, grad_norm: 0.652483 +Steps: 0%| | 1188/1000000 [04:58<68:39:43, 4.04it/s, grad_norm=0.652, loss_final=0.603, loss_mean=0.837, loss_mean_cls=0.0883, proj_loss=-0.322][2026-03-26 12:36:01] Step: 1188, Training Logs: loss_final: 0.621008, loss_mean: 0.878617, proj_loss: -0.341341, loss_mean_cls: 0.083732, grad_norm: 0.587346 +Steps: 0%| | 1189/1000000 [04:58<68:38:08, 4.04it/s, grad_norm=0.587, loss_final=0.621, loss_mean=0.879, loss_mean_cls=0.0837, proj_loss=-0.341][2026-03-26 12:36:01] Step: 1189, Training Logs: loss_final: 0.619291, loss_mean: 0.873810, proj_loss: -0.337566, loss_mean_cls: 0.083047, grad_norm: 0.432318 +Steps: 0%| | 1190/1000000 [04:58<68:40:25, 4.04it/s, grad_norm=0.432, loss_final=0.619, loss_mean=0.874, loss_mean_cls=0.083, proj_loss=-0.338][2026-03-26 12:36:01] Step: 1190, Training Logs: loss_final: 0.628528, loss_mean: 0.878033, proj_loss: -0.334575, loss_mean_cls: 0.085069, grad_norm: 0.707575 +Steps: 0%| | 1191/1000000 [04:58<68:42:52, 4.04it/s, grad_norm=0.708, loss_final=0.629, loss_mean=0.878, loss_mean_cls=0.0851, proj_loss=-0.335][2026-03-26 12:36:02] Step: 1191, Training Logs: loss_final: 0.618629, loss_mean: 0.873938, proj_loss: -0.339349, loss_mean_cls: 0.084039, grad_norm: 0.464918 +Steps: 0%| | 1192/1000000 [04:59<68:43:07, 4.04it/s, grad_norm=0.465, loss_final=0.619, loss_mean=0.874, loss_mean_cls=0.084, proj_loss=-0.339][2026-03-26 12:36:02] Step: 1192, Training Logs: loss_final: 0.628933, loss_mean: 0.864455, proj_loss: -0.322789, loss_mean_cls: 0.087268, grad_norm: 0.756648 +Steps: 0%| | 1193/1000000 [04:59<68:40:27, 4.04it/s, grad_norm=0.757, loss_final=0.629, loss_mean=0.864, loss_mean_cls=0.0873, proj_loss=-0.323][2026-03-26 12:36:02] Step: 1193, Training Logs: loss_final: 0.625691, loss_mean: 0.882453, proj_loss: -0.340523, loss_mean_cls: 0.083762, grad_norm: 0.419855 +Steps: 0%| | 1194/1000000 [04:59<68:40:52, 4.04it/s, grad_norm=0.42, loss_final=0.626, loss_mean=0.882, loss_mean_cls=0.0838, proj_loss=-0.341][2026-03-26 12:36:02] Step: 1194, Training Logs: loss_final: 0.604775, loss_mean: 0.844194, proj_loss: -0.326293, loss_mean_cls: 0.086874, grad_norm: 0.906096 +Steps: 0%| | 1195/1000000 [04:59<68:39:52, 4.04it/s, grad_norm=0.906, loss_final=0.605, loss_mean=0.844, loss_mean_cls=0.0869, proj_loss=-0.326][2026-03-26 12:36:03] Step: 1195, Training Logs: loss_final: 0.608766, loss_mean: 0.878734, proj_loss: -0.352075, loss_mean_cls: 0.082107, grad_norm: 0.732931 +Steps: 0%| | 1196/1000000 [05:00<68:43:08, 4.04it/s, grad_norm=0.733, loss_final=0.609, loss_mean=0.879, loss_mean_cls=0.0821, proj_loss=-0.352][2026-03-26 12:36:03] Step: 1196, Training Logs: loss_final: 0.611586, loss_mean: 0.854833, proj_loss: -0.329401, loss_mean_cls: 0.086154, grad_norm: 0.562538 +Steps: 0%| | 1197/1000000 [05:00<68:41:07, 4.04it/s, grad_norm=0.563, loss_final=0.612, loss_mean=0.855, loss_mean_cls=0.0862, proj_loss=-0.329][2026-03-26 12:36:03] Step: 1197, Training Logs: loss_final: 0.632542, loss_mean: 0.896074, proj_loss: -0.345371, loss_mean_cls: 0.081840, grad_norm: 0.564963 +Steps: 0%| | 1198/1000000 [05:00<68:40:59, 4.04it/s, grad_norm=0.565, loss_final=0.633, loss_mean=0.896, loss_mean_cls=0.0818, proj_loss=-0.345][2026-03-26 12:36:03] Step: 1198, Training Logs: loss_final: 0.628960, loss_mean: 0.867159, proj_loss: -0.325433, loss_mean_cls: 0.087234, grad_norm: 0.820857 +Steps: 0%| | 1199/1000000 [05:00<68:41:18, 4.04it/s, grad_norm=0.821, loss_final=0.629, loss_mean=0.867, loss_mean_cls=0.0872, proj_loss=-0.325][2026-03-26 12:36:04] Step: 1199, Training Logs: loss_final: 0.633565, loss_mean: 0.894218, proj_loss: -0.344529, loss_mean_cls: 0.083877, grad_norm: 0.632806 +Steps: 0%| | 1200/1000000 [05:01<68:42:12, 4.04it/s, grad_norm=0.633, loss_final=0.634, loss_mean=0.894, loss_mean_cls=0.0839, proj_loss=-0.345][2026-03-26 12:36:04] Step: 1200, Training Logs: loss_final: 0.618491, loss_mean: 0.887401, proj_loss: -0.350423, loss_mean_cls: 0.081513, grad_norm: 0.496704 +Steps: 0%| | 1201/1000000 [05:01<68:44:44, 4.04it/s, grad_norm=0.497, loss_final=0.618, loss_mean=0.887, loss_mean_cls=0.0815, proj_loss=-0.35][2026-03-26 12:36:04] Step: 1201, Training Logs: loss_final: 0.609618, loss_mean: 0.875077, proj_loss: -0.348572, loss_mean_cls: 0.083114, grad_norm: 0.716606 +Steps: 0%| | 1202/1000000 [05:01<68:46:44, 4.03it/s, grad_norm=0.717, loss_final=0.61, loss_mean=0.875, loss_mean_cls=0.0831, proj_loss=-0.349][2026-03-26 12:36:04] Step: 1202, Training Logs: loss_final: 0.608659, loss_mean: 0.866584, proj_loss: -0.341802, loss_mean_cls: 0.083878, grad_norm: 0.669392 +Steps: 0%| | 1203/1000000 [05:01<68:48:04, 4.03it/s, grad_norm=0.669, loss_final=0.609, loss_mean=0.867, loss_mean_cls=0.0839, proj_loss=-0.342][2026-03-26 12:36:05] Step: 1203, Training Logs: loss_final: 0.620091, loss_mean: 0.866752, proj_loss: -0.333533, loss_mean_cls: 0.086872, grad_norm: 0.387514 +Steps: 0%| | 1204/1000000 [05:01<68:48:48, 4.03it/s, grad_norm=0.388, loss_final=0.62, loss_mean=0.867, loss_mean_cls=0.0869, proj_loss=-0.334][2026-03-26 12:36:05] Step: 1204, Training Logs: loss_final: 0.616193, loss_mean: 0.872535, proj_loss: -0.339883, loss_mean_cls: 0.083540, grad_norm: 0.589300 +Steps: 0%| | 1205/1000000 [05:02<68:45:58, 4.03it/s, grad_norm=0.589, loss_final=0.616, loss_mean=0.873, loss_mean_cls=0.0835, proj_loss=-0.34][2026-03-26 12:36:05] Step: 1205, Training Logs: loss_final: 0.606655, loss_mean: 0.864044, proj_loss: -0.341561, loss_mean_cls: 0.084172, grad_norm: 0.423558 +Steps: 0%| | 1206/1000000 [05:02<68:44:47, 4.04it/s, grad_norm=0.424, loss_final=0.607, loss_mean=0.864, loss_mean_cls=0.0842, proj_loss=-0.342][2026-03-26 12:36:05] Step: 1206, Training Logs: loss_final: 0.616705, loss_mean: 0.879788, proj_loss: -0.344512, loss_mean_cls: 0.081429, grad_norm: 0.394665 +Steps: 0%| | 1207/1000000 [05:02<68:47:05, 4.03it/s, grad_norm=0.395, loss_final=0.617, loss_mean=0.88, loss_mean_cls=0.0814, proj_loss=-0.345][2026-03-26 12:36:06] Step: 1207, Training Logs: loss_final: 0.603543, loss_mean: 0.873073, proj_loss: -0.350475, loss_mean_cls: 0.080945, grad_norm: 0.507236 +Steps: 0%| | 1208/1000000 [05:02<68:45:27, 4.04it/s, grad_norm=0.507, loss_final=0.604, loss_mean=0.873, loss_mean_cls=0.0809, proj_loss=-0.35][2026-03-26 12:36:06] Step: 1208, Training Logs: loss_final: 0.632411, loss_mean: 0.891063, proj_loss: -0.340949, loss_mean_cls: 0.082297, grad_norm: 0.716431 +Steps: 0%| | 1209/1000000 [05:03<68:42:05, 4.04it/s, grad_norm=0.716, loss_final=0.632, loss_mean=0.891, loss_mean_cls=0.0823, proj_loss=-0.341][2026-03-26 12:36:06] Step: 1209, Training Logs: loss_final: 0.635019, loss_mean: 0.877079, proj_loss: -0.327755, loss_mean_cls: 0.085694, grad_norm: 0.599530 +Steps: 0%| | 1210/1000000 [05:03<68:42:16, 4.04it/s, grad_norm=0.6, loss_final=0.635, loss_mean=0.877, loss_mean_cls=0.0857, proj_loss=-0.328][2026-03-26 12:36:06] Step: 1210, Training Logs: loss_final: 0.615425, loss_mean: 0.884499, proj_loss: -0.351134, loss_mean_cls: 0.082059, grad_norm: 0.628385 +Steps: 0%| | 1211/1000000 [05:03<68:43:42, 4.04it/s, grad_norm=0.628, loss_final=0.615, loss_mean=0.884, loss_mean_cls=0.0821, proj_loss=-0.351][2026-03-26 12:36:07] Step: 1211, Training Logs: loss_final: 0.618323, loss_mean: 0.887966, proj_loss: -0.350157, loss_mean_cls: 0.080514, grad_norm: 0.861106 +Steps: 0%| | 1212/1000000 [05:03<68:44:47, 4.04it/s, grad_norm=0.861, loss_final=0.618, loss_mean=0.888, loss_mean_cls=0.0805, proj_loss=-0.35][2026-03-26 12:36:07] Step: 1212, Training Logs: loss_final: 0.614651, loss_mean: 0.847934, proj_loss: -0.320881, loss_mean_cls: 0.087598, grad_norm: 0.474547 +Steps: 0%| | 1213/1000000 [05:04<68:42:55, 4.04it/s, grad_norm=0.475, loss_final=0.615, loss_mean=0.848, loss_mean_cls=0.0876, proj_loss=-0.321][2026-03-26 12:36:07] Step: 1213, Training Logs: loss_final: 0.634172, loss_mean: 0.881104, proj_loss: -0.330546, loss_mean_cls: 0.083614, grad_norm: 0.718086 +Steps: 0%| | 1214/1000000 [05:04<68:39:40, 4.04it/s, grad_norm=0.718, loss_final=0.634, loss_mean=0.881, loss_mean_cls=0.0836, proj_loss=-0.331][2026-03-26 12:36:07] Step: 1214, Training Logs: loss_final: 0.617380, loss_mean: 0.896913, proj_loss: -0.359596, loss_mean_cls: 0.080063, grad_norm: 0.510646 +Steps: 0%| | 1215/1000000 [05:04<68:40:45, 4.04it/s, grad_norm=0.511, loss_final=0.617, loss_mean=0.897, loss_mean_cls=0.0801, proj_loss=-0.36][2026-03-26 12:36:08] Step: 1215, Training Logs: loss_final: 0.618762, loss_mean: 0.858564, proj_loss: -0.325858, loss_mean_cls: 0.086056, grad_norm: 0.684971 +Steps: 0%| | 1216/1000000 [05:04<68:44:08, 4.04it/s, grad_norm=0.685, loss_final=0.619, loss_mean=0.859, loss_mean_cls=0.0861, proj_loss=-0.326][2026-03-26 12:36:08] Step: 1216, Training Logs: loss_final: 0.595738, loss_mean: 0.866198, proj_loss: -0.352841, loss_mean_cls: 0.082381, grad_norm: 0.582236 +Steps: 0%| | 1217/1000000 [05:05<69:37:31, 3.98it/s, grad_norm=0.582, loss_final=0.596, loss_mean=0.866, loss_mean_cls=0.0824, proj_loss=-0.353][2026-03-26 12:36:08] Step: 1217, Training Logs: loss_final: 0.622522, loss_mean: 0.885846, proj_loss: -0.345222, loss_mean_cls: 0.081898, grad_norm: 0.883975 +Steps: 0%| | 1218/1000000 [05:05<69:20:47, 4.00it/s, grad_norm=0.884, loss_final=0.623, loss_mean=0.886, loss_mean_cls=0.0819, proj_loss=-0.345][2026-03-26 12:36:08] Step: 1218, Training Logs: loss_final: 0.623448, loss_mean: 0.876936, proj_loss: -0.337861, loss_mean_cls: 0.084373, grad_norm: 0.691186 +Steps: 0%| | 1219/1000000 [05:05<69:08:05, 4.01it/s, grad_norm=0.691, loss_final=0.623, loss_mean=0.877, loss_mean_cls=0.0844, proj_loss=-0.338][2026-03-26 12:36:09] Step: 1219, Training Logs: loss_final: 0.625261, loss_mean: 0.864425, proj_loss: -0.327025, loss_mean_cls: 0.087861, grad_norm: 0.719100 +Steps: 0%| | 1220/1000000 [05:05<69:02:42, 4.02it/s, grad_norm=0.719, loss_final=0.625, loss_mean=0.864, loss_mean_cls=0.0879, proj_loss=-0.327][2026-03-26 12:36:09] Step: 1220, Training Logs: loss_final: 0.613742, loss_mean: 0.847381, proj_loss: -0.321155, loss_mean_cls: 0.087516, grad_norm: 0.565727 +Steps: 0%| | 1221/1000000 [05:06<68:55:43, 4.02it/s, grad_norm=0.566, loss_final=0.614, loss_mean=0.847, loss_mean_cls=0.0875, proj_loss=-0.321][2026-03-26 12:36:09] Step: 1221, Training Logs: loss_final: 0.614545, loss_mean: 0.860743, proj_loss: -0.332081, loss_mean_cls: 0.085883, grad_norm: 0.721867 +Steps: 0%| | 1222/1000000 [05:06<68:51:56, 4.03it/s, grad_norm=0.722, loss_final=0.615, loss_mean=0.861, loss_mean_cls=0.0859, proj_loss=-0.332][2026-03-26 12:36:09] Step: 1222, Training Logs: loss_final: 0.622682, loss_mean: 0.883716, proj_loss: -0.343626, loss_mean_cls: 0.082592, grad_norm: 0.704980 +Steps: 0%| | 1223/1000000 [05:06<68:49:37, 4.03it/s, grad_norm=0.705, loss_final=0.623, loss_mean=0.884, loss_mean_cls=0.0826, proj_loss=-0.344][2026-03-26 12:36:10] Step: 1223, Training Logs: loss_final: 0.624911, loss_mean: 0.874887, proj_loss: -0.334621, loss_mean_cls: 0.084645, grad_norm: 0.897032 +Steps: 0%| | 1224/1000000 [05:06<68:47:33, 4.03it/s, grad_norm=0.897, loss_final=0.625, loss_mean=0.875, loss_mean_cls=0.0846, proj_loss=-0.335][2026-03-26 12:36:10] Step: 1224, Training Logs: loss_final: 0.614746, loss_mean: 0.867371, proj_loss: -0.337166, loss_mean_cls: 0.084541, grad_norm: 0.551745 +Steps: 0%| | 1225/1000000 [05:07<68:44:10, 4.04it/s, grad_norm=0.552, loss_final=0.615, loss_mean=0.867, loss_mean_cls=0.0845, proj_loss=-0.337][2026-03-26 12:36:10] Step: 1225, Training Logs: loss_final: 0.618225, loss_mean: 0.873652, proj_loss: -0.339550, loss_mean_cls: 0.084123, grad_norm: 0.575791 +Steps: 0%| | 1226/1000000 [05:07<68:43:01, 4.04it/s, grad_norm=0.576, loss_final=0.618, loss_mean=0.874, loss_mean_cls=0.0841, proj_loss=-0.34][2026-03-26 12:36:10] Step: 1226, Training Logs: loss_final: 0.613770, loss_mean: 0.862950, proj_loss: -0.334389, loss_mean_cls: 0.085208, grad_norm: 0.693857 +Steps: 0%| | 1227/1000000 [05:07<68:41:15, 4.04it/s, grad_norm=0.694, loss_final=0.614, loss_mean=0.863, loss_mean_cls=0.0852, proj_loss=-0.334][2026-03-26 12:36:11] Step: 1227, Training Logs: loss_final: 0.638662, loss_mean: 0.881194, proj_loss: -0.327570, loss_mean_cls: 0.085038, grad_norm: 0.604077 +Steps: 0%| | 1228/1000000 [05:07<68:42:32, 4.04it/s, grad_norm=0.604, loss_final=0.639, loss_mean=0.881, loss_mean_cls=0.085, proj_loss=-0.328][2026-03-26 12:36:11] Step: 1228, Training Logs: loss_final: 0.613768, loss_mean: 0.869994, proj_loss: -0.340445, loss_mean_cls: 0.084219, grad_norm: 0.787373 +Steps: 0%| | 1229/1000000 [05:08<68:39:30, 4.04it/s, grad_norm=0.787, loss_final=0.614, loss_mean=0.87, loss_mean_cls=0.0842, proj_loss=-0.34][2026-03-26 12:36:11] Step: 1229, Training Logs: loss_final: 0.627297, loss_mean: 0.883144, proj_loss: -0.340652, loss_mean_cls: 0.084804, grad_norm: 0.959569 +Steps: 0%| | 1230/1000000 [05:08<68:39:58, 4.04it/s, grad_norm=0.96, loss_final=0.627, loss_mean=0.883, loss_mean_cls=0.0848, proj_loss=-0.341][2026-03-26 12:36:11] Step: 1230, Training Logs: loss_final: 0.627139, loss_mean: 0.885406, proj_loss: -0.341635, loss_mean_cls: 0.083368, grad_norm: 0.567037 +Steps: 0%| | 1231/1000000 [05:08<68:36:43, 4.04it/s, grad_norm=0.567, loss_final=0.627, loss_mean=0.885, loss_mean_cls=0.0834, proj_loss=-0.342][2026-03-26 12:36:12] Step: 1231, Training Logs: loss_final: 0.620364, loss_mean: 0.877218, proj_loss: -0.340061, loss_mean_cls: 0.083206, grad_norm: 0.598139 +Steps: 0%| | 1232/1000000 [05:08<68:42:07, 4.04it/s, grad_norm=0.598, loss_final=0.62, loss_mean=0.877, loss_mean_cls=0.0832, proj_loss=-0.34][2026-03-26 12:36:12] Step: 1232, Training Logs: loss_final: 0.625605, loss_mean: 0.878305, proj_loss: -0.336725, loss_mean_cls: 0.084026, grad_norm: 0.778688 +Steps: 0%| | 1233/1000000 [05:09<68:41:17, 4.04it/s, grad_norm=0.779, loss_final=0.626, loss_mean=0.878, loss_mean_cls=0.084, proj_loss=-0.337][2026-03-26 12:36:12] Step: 1233, Training Logs: loss_final: 0.630692, loss_mean: 0.882381, proj_loss: -0.334313, loss_mean_cls: 0.082624, grad_norm: 0.475945 +Steps: 0%| | 1234/1000000 [05:09<68:41:46, 4.04it/s, grad_norm=0.476, loss_final=0.631, loss_mean=0.882, loss_mean_cls=0.0826, proj_loss=-0.334][2026-03-26 12:36:12] Step: 1234, Training Logs: loss_final: 0.618028, loss_mean: 0.859923, proj_loss: -0.328930, loss_mean_cls: 0.087035, grad_norm: 0.602018 +Steps: 0%| | 1235/1000000 [05:09<68:38:59, 4.04it/s, grad_norm=0.602, loss_final=0.618, loss_mean=0.86, loss_mean_cls=0.087, proj_loss=-0.329][2026-03-26 12:36:13] Step: 1235, Training Logs: loss_final: 0.619094, loss_mean: 0.878217, proj_loss: -0.342786, loss_mean_cls: 0.083662, grad_norm: 0.714292 +Steps: 0%| | 1236/1000000 [05:09<68:40:19, 4.04it/s, grad_norm=0.714, loss_final=0.619, loss_mean=0.878, loss_mean_cls=0.0837, proj_loss=-0.343][2026-03-26 12:36:13] Step: 1236, Training Logs: loss_final: 0.618466, loss_mean: 0.875177, proj_loss: -0.340115, loss_mean_cls: 0.083404, grad_norm: 0.677405 +Steps: 0%| | 1237/1000000 [05:10<68:39:20, 4.04it/s, grad_norm=0.677, loss_final=0.618, loss_mean=0.875, loss_mean_cls=0.0834, proj_loss=-0.34][2026-03-26 12:36:13] Step: 1237, Training Logs: loss_final: 0.614667, loss_mean: 0.880717, proj_loss: -0.348518, loss_mean_cls: 0.082469, grad_norm: 0.504857 +Steps: 0%| | 1238/1000000 [05:10<68:39:58, 4.04it/s, grad_norm=0.505, loss_final=0.615, loss_mean=0.881, loss_mean_cls=0.0825, proj_loss=-0.349][2026-03-26 12:36:13] Step: 1238, Training Logs: loss_final: 0.629481, loss_mean: 0.884061, proj_loss: -0.337143, loss_mean_cls: 0.082563, grad_norm: 0.992857 +Steps: 0%| | 1239/1000000 [05:10<68:41:39, 4.04it/s, grad_norm=0.993, loss_final=0.629, loss_mean=0.884, loss_mean_cls=0.0826, proj_loss=-0.337][2026-03-26 12:36:14] Step: 1239, Training Logs: loss_final: 0.624753, loss_mean: 0.871855, proj_loss: -0.332643, loss_mean_cls: 0.085541, grad_norm: 0.851219 +Steps: 0%| | 1240/1000000 [05:10<68:42:45, 4.04it/s, grad_norm=0.851, loss_final=0.625, loss_mean=0.872, loss_mean_cls=0.0855, proj_loss=-0.333][2026-03-26 12:36:14] Step: 1240, Training Logs: loss_final: 0.629518, loss_mean: 0.885145, proj_loss: -0.338816, loss_mean_cls: 0.083189, grad_norm: 0.478427 +Steps: 0%| | 1241/1000000 [05:11<68:41:41, 4.04it/s, grad_norm=0.478, loss_final=0.63, loss_mean=0.885, loss_mean_cls=0.0832, proj_loss=-0.339][2026-03-26 12:36:14] Step: 1241, Training Logs: loss_final: 0.627109, loss_mean: 0.892086, proj_loss: -0.347247, loss_mean_cls: 0.082271, grad_norm: 1.075294 +Steps: 0%| | 1242/1000000 [05:11<68:43:47, 4.04it/s, grad_norm=1.08, loss_final=0.627, loss_mean=0.892, loss_mean_cls=0.0823, proj_loss=-0.347][2026-03-26 12:36:14] Step: 1242, Training Logs: loss_final: 0.612299, loss_mean: 0.853395, proj_loss: -0.326680, loss_mean_cls: 0.085584, grad_norm: 0.540709 +Steps: 0%| | 1243/1000000 [05:11<68:45:05, 4.04it/s, grad_norm=0.541, loss_final=0.612, loss_mean=0.853, loss_mean_cls=0.0856, proj_loss=-0.327][2026-03-26 12:36:15] Step: 1243, Training Logs: loss_final: 0.624866, loss_mean: 0.877176, proj_loss: -0.335591, loss_mean_cls: 0.083281, grad_norm: 1.009623 +Steps: 0%| | 1244/1000000 [05:11<68:47:15, 4.03it/s, grad_norm=1.01, loss_final=0.625, loss_mean=0.877, loss_mean_cls=0.0833, proj_loss=-0.336][2026-03-26 12:36:15] Step: 1244, Training Logs: loss_final: 0.615645, loss_mean: 0.872815, proj_loss: -0.341246, loss_mean_cls: 0.084076, grad_norm: 0.935730 +Steps: 0%| | 1245/1000000 [05:12<68:46:34, 4.03it/s, grad_norm=0.936, loss_final=0.616, loss_mean=0.873, loss_mean_cls=0.0841, proj_loss=-0.341][2026-03-26 12:36:15] Step: 1245, Training Logs: loss_final: 0.614246, loss_mean: 0.872729, proj_loss: -0.342487, loss_mean_cls: 0.084004, grad_norm: 0.715331 +Steps: 0%| | 1246/1000000 [05:12<68:45:01, 4.04it/s, grad_norm=0.715, loss_final=0.614, loss_mean=0.873, loss_mean_cls=0.084, proj_loss=-0.342][2026-03-26 12:36:15] Step: 1246, Training Logs: loss_final: 0.611953, loss_mean: 0.861448, proj_loss: -0.335032, loss_mean_cls: 0.085537, grad_norm: 0.573458 +Steps: 0%| | 1247/1000000 [05:12<68:45:20, 4.04it/s, grad_norm=0.573, loss_final=0.612, loss_mean=0.861, loss_mean_cls=0.0855, proj_loss=-0.335][2026-03-26 12:36:16] Step: 1247, Training Logs: loss_final: 0.614874, loss_mean: 0.885466, proj_loss: -0.351742, loss_mean_cls: 0.081150, grad_norm: 0.691885 +Steps: 0%| | 1248/1000000 [05:12<68:44:09, 4.04it/s, grad_norm=0.692, loss_final=0.615, loss_mean=0.885, loss_mean_cls=0.0812, proj_loss=-0.352][2026-03-26 12:36:16] Step: 1248, Training Logs: loss_final: 0.617597, loss_mean: 0.869443, proj_loss: -0.336105, loss_mean_cls: 0.084258, grad_norm: 0.406873 +Steps: 0%| | 1249/1000000 [05:13<68:44:22, 4.04it/s, grad_norm=0.407, loss_final=0.618, loss_mean=0.869, loss_mean_cls=0.0843, proj_loss=-0.336][2026-03-26 12:36:16] Step: 1249, Training Logs: loss_final: 0.622074, loss_mean: 0.890153, proj_loss: -0.349951, loss_mean_cls: 0.081872, grad_norm: 0.689834 +Steps: 0%| | 1250/1000000 [05:13<68:49:12, 4.03it/s, grad_norm=0.69, loss_final=0.622, loss_mean=0.89, loss_mean_cls=0.0819, proj_loss=-0.35][2026-03-26 12:36:16] Step: 1250, Training Logs: loss_final: 0.605741, loss_mean: 0.844803, proj_loss: -0.326885, loss_mean_cls: 0.087823, grad_norm: 0.543763 +Steps: 0%| | 1251/1000000 [05:13<68:47:24, 4.03it/s, grad_norm=0.544, loss_final=0.606, loss_mean=0.845, loss_mean_cls=0.0878, proj_loss=-0.327][2026-03-26 12:36:17] Step: 1251, Training Logs: loss_final: 0.627347, loss_mean: 0.891273, proj_loss: -0.345207, loss_mean_cls: 0.081281, grad_norm: 0.519453 +Steps: 0%| | 1252/1000000 [05:13<68:43:12, 4.04it/s, grad_norm=0.519, loss_final=0.627, loss_mean=0.891, loss_mean_cls=0.0813, proj_loss=-0.345][2026-03-26 12:36:17] Step: 1252, Training Logs: loss_final: 0.605632, loss_mean: 0.860362, proj_loss: -0.339087, loss_mean_cls: 0.084358, grad_norm: 0.687092 +Steps: 0%| | 1253/1000000 [05:14<68:43:35, 4.04it/s, grad_norm=0.687, loss_final=0.606, loss_mean=0.86, loss_mean_cls=0.0844, proj_loss=-0.339][2026-03-26 12:36:17] Step: 1253, Training Logs: loss_final: 0.611760, loss_mean: 0.871659, proj_loss: -0.344088, loss_mean_cls: 0.084189, grad_norm: 0.461224 +Steps: 0%| | 1254/1000000 [05:14<68:43:24, 4.04it/s, grad_norm=0.461, loss_final=0.612, loss_mean=0.872, loss_mean_cls=0.0842, proj_loss=-0.344][2026-03-26 12:36:17] Step: 1254, Training Logs: loss_final: 0.615045, loss_mean: 0.873530, proj_loss: -0.342017, loss_mean_cls: 0.083532, grad_norm: 0.640066 +Steps: 0%| | 1255/1000000 [05:14<69:26:51, 3.99it/s, grad_norm=0.64, loss_final=0.615, loss_mean=0.874, loss_mean_cls=0.0835, proj_loss=-0.342][2026-03-26 12:36:18] Step: 1255, Training Logs: loss_final: 0.619057, loss_mean: 0.881208, proj_loss: -0.343343, loss_mean_cls: 0.081192, grad_norm: 0.439881 +Steps: 0%| | 1256/1000000 [05:14<69:10:54, 4.01it/s, grad_norm=0.44, loss_final=0.619, loss_mean=0.881, loss_mean_cls=0.0812, proj_loss=-0.343][2026-03-26 12:36:18] Step: 1256, Training Logs: loss_final: 0.620827, loss_mean: 0.871212, proj_loss: -0.335310, loss_mean_cls: 0.084925, grad_norm: 0.696670 +Steps: 0%| | 1257/1000000 [05:15<69:04:50, 4.02it/s, grad_norm=0.697, loss_final=0.621, loss_mean=0.871, loss_mean_cls=0.0849, proj_loss=-0.335][2026-03-26 12:36:18] Step: 1257, Training Logs: loss_final: 0.604635, loss_mean: 0.868085, proj_loss: -0.345130, loss_mean_cls: 0.081681, grad_norm: 0.535472 +Steps: 0%| | 1258/1000000 [05:15<68:56:28, 4.02it/s, grad_norm=0.535, loss_final=0.605, loss_mean=0.868, loss_mean_cls=0.0817, proj_loss=-0.345][2026-03-26 12:36:18] Step: 1258, Training Logs: loss_final: 0.601631, loss_mean: 0.856805, proj_loss: -0.339293, loss_mean_cls: 0.084119, grad_norm: 0.445808 +Steps: 0%| | 1259/1000000 [05:15<68:48:58, 4.03it/s, grad_norm=0.446, loss_final=0.602, loss_mean=0.857, loss_mean_cls=0.0841, proj_loss=-0.339][2026-03-26 12:36:19] Step: 1259, Training Logs: loss_final: 0.612244, loss_mean: 0.871356, proj_loss: -0.341804, loss_mean_cls: 0.082693, grad_norm: 0.527094 +Steps: 0%| | 1260/1000000 [05:15<68:43:32, 4.04it/s, grad_norm=0.527, loss_final=0.612, loss_mean=0.871, loss_mean_cls=0.0827, proj_loss=-0.342][2026-03-26 12:36:19] Step: 1260, Training Logs: loss_final: 0.628524, loss_mean: 0.891462, proj_loss: -0.344410, loss_mean_cls: 0.081472, grad_norm: 0.435697 +Steps: 0%| | 1261/1000000 [05:16<68:43:03, 4.04it/s, grad_norm=0.436, loss_final=0.629, loss_mean=0.891, loss_mean_cls=0.0815, proj_loss=-0.344][2026-03-26 12:36:19] Step: 1261, Training Logs: loss_final: 0.604477, loss_mean: 0.874148, proj_loss: -0.351319, loss_mean_cls: 0.081648, grad_norm: 0.482770 +Steps: 0%| | 1262/1000000 [05:16<68:39:32, 4.04it/s, grad_norm=0.483, loss_final=0.604, loss_mean=0.874, loss_mean_cls=0.0816, proj_loss=-0.351][2026-03-26 12:36:19] Step: 1262, Training Logs: loss_final: 0.614478, loss_mean: 0.851602, proj_loss: -0.325697, loss_mean_cls: 0.088573, grad_norm: 0.483966 +Steps: 0%| | 1263/1000000 [05:16<68:41:05, 4.04it/s, grad_norm=0.484, loss_final=0.614, loss_mean=0.852, loss_mean_cls=0.0886, proj_loss=-0.326][2026-03-26 12:36:20] Step: 1263, Training Logs: loss_final: 0.620597, loss_mean: 0.871049, proj_loss: -0.336191, loss_mean_cls: 0.085739, grad_norm: 0.627640 +Steps: 0%| | 1264/1000000 [05:16<68:40:49, 4.04it/s, grad_norm=0.628, loss_final=0.621, loss_mean=0.871, loss_mean_cls=0.0857, proj_loss=-0.336][2026-03-26 12:36:20] Step: 1264, Training Logs: loss_final: 0.605150, loss_mean: 0.864043, proj_loss: -0.342224, loss_mean_cls: 0.083331, grad_norm: 0.384933 +Steps: 0%| | 1265/1000000 [05:17<71:23:37, 3.89it/s, grad_norm=0.385, loss_final=0.605, loss_mean=0.864, loss_mean_cls=0.0833, proj_loss=-0.342][2026-03-26 12:36:20] Step: 1265, Training Logs: loss_final: 0.608613, loss_mean: 0.870217, proj_loss: -0.344066, loss_mean_cls: 0.082462, grad_norm: 0.479246 +Steps: 0%| | 1266/1000000 [05:17<70:35:40, 3.93it/s, grad_norm=0.479, loss_final=0.609, loss_mean=0.87, loss_mean_cls=0.0825, proj_loss=-0.344][2026-03-26 12:36:20] Step: 1266, Training Logs: loss_final: 0.601715, loss_mean: 0.868237, proj_loss: -0.349265, loss_mean_cls: 0.082743, grad_norm: 0.449837 +Steps: 0%| | 1267/1000000 [05:17<69:59:11, 3.96it/s, grad_norm=0.45, loss_final=0.602, loss_mean=0.868, loss_mean_cls=0.0827, proj_loss=-0.349][2026-03-26 12:36:21] Step: 1267, Training Logs: loss_final: 0.609600, loss_mean: 0.858898, proj_loss: -0.334569, loss_mean_cls: 0.085271, grad_norm: 0.605642 +Steps: 0%| | 1268/1000000 [05:17<69:35:42, 3.99it/s, grad_norm=0.606, loss_final=0.61, loss_mean=0.859, loss_mean_cls=0.0853, proj_loss=-0.335][2026-03-26 12:36:21] Step: 1268, Training Logs: loss_final: 0.607008, loss_mean: 0.868912, proj_loss: -0.344085, loss_mean_cls: 0.082182, grad_norm: 0.583758 +Steps: 0%| | 1269/1000000 [05:18<69:19:21, 4.00it/s, grad_norm=0.584, loss_final=0.607, loss_mean=0.869, loss_mean_cls=0.0822, proj_loss=-0.344][2026-03-26 12:36:21] Step: 1269, Training Logs: loss_final: 0.619787, loss_mean: 0.881579, proj_loss: -0.344755, loss_mean_cls: 0.082963, grad_norm: 0.568007 +Steps: 0%| | 1270/1000000 [05:18<69:06:07, 4.01it/s, grad_norm=0.568, loss_final=0.62, loss_mean=0.882, loss_mean_cls=0.083, proj_loss=-0.345][2026-03-26 12:36:21] Step: 1270, Training Logs: loss_final: 0.628085, loss_mean: 0.888417, proj_loss: -0.343122, loss_mean_cls: 0.082789, grad_norm: 0.605930 +Steps: 0%| | 1271/1000000 [05:18<69:00:27, 4.02it/s, grad_norm=0.606, loss_final=0.628, loss_mean=0.888, loss_mean_cls=0.0828, proj_loss=-0.343][2026-03-26 12:36:22] Step: 1271, Training Logs: loss_final: 0.620823, loss_mean: 0.875137, proj_loss: -0.338697, loss_mean_cls: 0.084382, grad_norm: 0.627186 +Steps: 0%| | 1272/1000000 [05:18<68:55:09, 4.03it/s, grad_norm=0.627, loss_final=0.621, loss_mean=0.875, loss_mean_cls=0.0844, proj_loss=-0.339][2026-03-26 12:36:22] Step: 1272, Training Logs: loss_final: 0.621587, loss_mean: 0.874962, proj_loss: -0.338142, loss_mean_cls: 0.084767, grad_norm: 0.344280 +Steps: 0%| | 1273/1000000 [05:19<68:51:45, 4.03it/s, grad_norm=0.344, loss_final=0.622, loss_mean=0.875, loss_mean_cls=0.0848, proj_loss=-0.338][2026-03-26 12:36:22] Step: 1273, Training Logs: loss_final: 0.602542, loss_mean: 0.854631, proj_loss: -0.336762, loss_mean_cls: 0.084674, grad_norm: 0.589636 +Steps: 0%| | 1274/1000000 [05:19<68:49:15, 4.03it/s, grad_norm=0.59, loss_final=0.603, loss_mean=0.855, loss_mean_cls=0.0847, proj_loss=-0.337][2026-03-26 12:36:22] Step: 1274, Training Logs: loss_final: 0.606194, loss_mean: 0.865676, proj_loss: -0.342908, loss_mean_cls: 0.083425, grad_norm: 0.376352 +Steps: 0%| | 1275/1000000 [05:19<68:47:49, 4.03it/s, grad_norm=0.376, loss_final=0.606, loss_mean=0.866, loss_mean_cls=0.0834, proj_loss=-0.343][2026-03-26 12:36:23] Step: 1275, Training Logs: loss_final: 0.611738, loss_mean: 0.859857, proj_loss: -0.334277, loss_mean_cls: 0.086158, grad_norm: 0.375658 +Steps: 0%| | 1276/1000000 [05:19<68:44:24, 4.04it/s, grad_norm=0.376, loss_final=0.612, loss_mean=0.86, loss_mean_cls=0.0862, proj_loss=-0.334][2026-03-26 12:36:23] Step: 1276, Training Logs: loss_final: 0.598113, loss_mean: 0.839794, proj_loss: -0.329289, loss_mean_cls: 0.087607, grad_norm: 0.444121 +Steps: 0%| | 1277/1000000 [05:20<68:44:26, 4.04it/s, grad_norm=0.444, loss_final=0.598, loss_mean=0.84, loss_mean_cls=0.0876, proj_loss=-0.329][2026-03-26 12:36:23] Step: 1277, Training Logs: loss_final: 0.608368, loss_mean: 0.867760, proj_loss: -0.344054, loss_mean_cls: 0.084662, grad_norm: 0.392627 +Steps: 0%| | 1278/1000000 [05:20<68:42:04, 4.04it/s, grad_norm=0.393, loss_final=0.608, loss_mean=0.868, loss_mean_cls=0.0847, proj_loss=-0.344][2026-03-26 12:36:23] Step: 1278, Training Logs: loss_final: 0.620958, loss_mean: 0.876145, proj_loss: -0.339024, loss_mean_cls: 0.083837, grad_norm: 0.496307 +Steps: 0%| | 1279/1000000 [05:20<68:42:40, 4.04it/s, grad_norm=0.496, loss_final=0.621, loss_mean=0.876, loss_mean_cls=0.0838, proj_loss=-0.339][2026-03-26 12:36:24] Step: 1279, Training Logs: loss_final: 0.621772, loss_mean: 0.873240, proj_loss: -0.335974, loss_mean_cls: 0.084506, grad_norm: 0.483751 +Steps: 0%| | 1280/1000000 [05:20<68:42:03, 4.04it/s, grad_norm=0.484, loss_final=0.622, loss_mean=0.873, loss_mean_cls=0.0845, proj_loss=-0.336][2026-03-26 12:36:24] Step: 1280, Training Logs: loss_final: 0.628444, loss_mean: 0.883709, proj_loss: -0.340602, loss_mean_cls: 0.085338, grad_norm: 0.529198 +Steps: 0%| | 1281/1000000 [05:21<68:42:58, 4.04it/s, grad_norm=0.529, loss_final=0.628, loss_mean=0.884, loss_mean_cls=0.0853, proj_loss=-0.341][2026-03-26 12:36:24] Step: 1281, Training Logs: loss_final: 0.615011, loss_mean: 0.877244, proj_loss: -0.344321, loss_mean_cls: 0.082088, grad_norm: 0.521918 +Steps: 0%| | 1282/1000000 [05:21<68:41:53, 4.04it/s, grad_norm=0.522, loss_final=0.615, loss_mean=0.877, loss_mean_cls=0.0821, proj_loss=-0.344][2026-03-26 12:36:24] Step: 1282, Training Logs: loss_final: 0.613982, loss_mean: 0.870806, proj_loss: -0.340647, loss_mean_cls: 0.083824, grad_norm: 0.401370 +Steps: 0%| | 1283/1000000 [05:21<68:42:19, 4.04it/s, grad_norm=0.401, loss_final=0.614, loss_mean=0.871, loss_mean_cls=0.0838, proj_loss=-0.341][2026-03-26 12:36:25] Step: 1283, Training Logs: loss_final: 0.624070, loss_mean: 0.872338, proj_loss: -0.332665, loss_mean_cls: 0.084397, grad_norm: 0.384354 +Steps: 0%| | 1284/1000000 [05:21<68:40:46, 4.04it/s, grad_norm=0.384, loss_final=0.624, loss_mean=0.872, loss_mean_cls=0.0844, proj_loss=-0.333][2026-03-26 12:36:25] Step: 1284, Training Logs: loss_final: 0.606699, loss_mean: 0.865055, proj_loss: -0.343220, loss_mean_cls: 0.084864, grad_norm: 0.432017 +Steps: 0%| | 1285/1000000 [05:22<68:39:55, 4.04it/s, grad_norm=0.432, loss_final=0.607, loss_mean=0.865, loss_mean_cls=0.0849, proj_loss=-0.343][2026-03-26 12:36:25] Step: 1285, Training Logs: loss_final: 0.620784, loss_mean: 0.876757, proj_loss: -0.339935, loss_mean_cls: 0.083962, grad_norm: 0.618718 +Steps: 0%| | 1286/1000000 [05:22<68:40:17, 4.04it/s, grad_norm=0.619, loss_final=0.621, loss_mean=0.877, loss_mean_cls=0.084, proj_loss=-0.34][2026-03-26 12:36:25] Step: 1286, Training Logs: loss_final: 0.612452, loss_mean: 0.865473, proj_loss: -0.337247, loss_mean_cls: 0.084226, grad_norm: 0.634655 +Steps: 0%| | 1287/1000000 [05:22<68:41:12, 4.04it/s, grad_norm=0.635, loss_final=0.612, loss_mean=0.865, loss_mean_cls=0.0842, proj_loss=-0.337][2026-03-26 12:36:25] Step: 1287, Training Logs: loss_final: 0.610677, loss_mean: 0.864123, proj_loss: -0.338816, loss_mean_cls: 0.085370, grad_norm: 0.646733 +Steps: 0%| | 1288/1000000 [05:22<68:41:22, 4.04it/s, grad_norm=0.647, loss_final=0.611, loss_mean=0.864, loss_mean_cls=0.0854, proj_loss=-0.339][2026-03-26 12:36:26] Step: 1288, Training Logs: loss_final: 0.609805, loss_mean: 0.880179, proj_loss: -0.351623, loss_mean_cls: 0.081248, grad_norm: 0.459623 +Steps: 0%| | 1289/1000000 [05:23<68:38:34, 4.04it/s, grad_norm=0.46, loss_final=0.61, loss_mean=0.88, loss_mean_cls=0.0812, proj_loss=-0.352][2026-03-26 12:36:26] Step: 1289, Training Logs: loss_final: 0.612697, loss_mean: 0.887230, proj_loss: -0.354894, loss_mean_cls: 0.080361, grad_norm: 0.763949 +Steps: 0%| | 1290/1000000 [05:23<68:43:04, 4.04it/s, grad_norm=0.764, loss_final=0.613, loss_mean=0.887, loss_mean_cls=0.0804, proj_loss=-0.355][2026-03-26 12:36:26] Step: 1290, Training Logs: loss_final: 0.601087, loss_mean: 0.862168, proj_loss: -0.344660, loss_mean_cls: 0.083580, grad_norm: 0.376491 +Steps: 0%| | 1291/1000000 [05:23<68:41:02, 4.04it/s, grad_norm=0.376, loss_final=0.601, loss_mean=0.862, loss_mean_cls=0.0836, proj_loss=-0.345][2026-03-26 12:36:26] Step: 1291, Training Logs: loss_final: 0.613465, loss_mean: 0.877451, proj_loss: -0.346391, loss_mean_cls: 0.082404, grad_norm: 0.489877 +Steps: 0%| | 1292/1000000 [05:23<68:41:59, 4.04it/s, grad_norm=0.49, loss_final=0.613, loss_mean=0.877, loss_mean_cls=0.0824, proj_loss=-0.346][2026-03-26 12:36:27] Step: 1292, Training Logs: loss_final: 0.610210, loss_mean: 0.856015, proj_loss: -0.330456, loss_mean_cls: 0.084651, grad_norm: 0.646004 +Steps: 0%| | 1293/1000000 [05:24<68:41:49, 4.04it/s, grad_norm=0.646, loss_final=0.61, loss_mean=0.856, loss_mean_cls=0.0847, proj_loss=-0.33][2026-03-26 12:36:27] Step: 1293, Training Logs: loss_final: 0.611041, loss_mean: 0.863307, proj_loss: -0.335951, loss_mean_cls: 0.083686, grad_norm: 0.659107 +Steps: 0%| | 1294/1000000 [05:24<68:40:40, 4.04it/s, grad_norm=0.659, loss_final=0.611, loss_mean=0.863, loss_mean_cls=0.0837, proj_loss=-0.336][2026-03-26 12:36:27] Step: 1294, Training Logs: loss_final: 0.616652, loss_mean: 0.849244, proj_loss: -0.320279, loss_mean_cls: 0.087687, grad_norm: 0.747620 +Steps: 0%| | 1295/1000000 [05:24<68:39:19, 4.04it/s, grad_norm=0.748, loss_final=0.617, loss_mean=0.849, loss_mean_cls=0.0877, proj_loss=-0.32][2026-03-26 12:36:27] Step: 1295, Training Logs: loss_final: 0.615846, loss_mean: 0.882354, proj_loss: -0.347768, loss_mean_cls: 0.081260, grad_norm: 0.547603 +Steps: 0%| | 1296/1000000 [05:24<68:40:47, 4.04it/s, grad_norm=0.548, loss_final=0.616, loss_mean=0.882, loss_mean_cls=0.0813, proj_loss=-0.348][2026-03-26 12:36:28] Step: 1296, Training Logs: loss_final: 0.619132, loss_mean: 0.888498, proj_loss: -0.351756, loss_mean_cls: 0.082391, grad_norm: 0.502712 +Steps: 0%| | 1297/1000000 [05:25<69:28:05, 3.99it/s, grad_norm=0.503, loss_final=0.619, loss_mean=0.888, loss_mean_cls=0.0824, proj_loss=-0.352][2026-03-26 12:36:28] Step: 1297, Training Logs: loss_final: 0.611469, loss_mean: 0.866463, proj_loss: -0.340136, loss_mean_cls: 0.085142, grad_norm: 0.576179 +Steps: 0%| | 1298/1000000 [05:25<69:15:23, 4.01it/s, grad_norm=0.576, loss_final=0.611, loss_mean=0.866, loss_mean_cls=0.0851, proj_loss=-0.34][2026-03-26 12:36:28] Step: 1298, Training Logs: loss_final: 0.601786, loss_mean: 0.860057, proj_loss: -0.343666, loss_mean_cls: 0.085395, grad_norm: 0.851584 +Steps: 0%| | 1299/1000000 [05:25<69:01:33, 4.02it/s, grad_norm=0.852, loss_final=0.602, loss_mean=0.86, loss_mean_cls=0.0854, proj_loss=-0.344][2026-03-26 12:36:28] Step: 1299, Training Logs: loss_final: 0.610801, loss_mean: 0.858892, proj_loss: -0.332535, loss_mean_cls: 0.084444, grad_norm: 0.646770 +Steps: 0%| | 1300/1000000 [05:25<68:56:27, 4.02it/s, grad_norm=0.647, loss_final=0.611, loss_mean=0.859, loss_mean_cls=0.0844, proj_loss=-0.333][2026-03-26 12:36:29] Step: 1300, Training Logs: loss_final: 0.597995, loss_mean: 0.845973, proj_loss: -0.334407, loss_mean_cls: 0.086429, grad_norm: 0.658334 +Steps: 0%| | 1301/1000000 [05:26<68:48:21, 4.03it/s, grad_norm=0.658, loss_final=0.598, loss_mean=0.846, loss_mean_cls=0.0864, proj_loss=-0.334][2026-03-26 12:36:29] Step: 1301, Training Logs: loss_final: 0.604959, loss_mean: 0.853787, proj_loss: -0.334769, loss_mean_cls: 0.085941, grad_norm: 0.665360 +Steps: 0%| | 1302/1000000 [05:26<68:49:50, 4.03it/s, grad_norm=0.665, loss_final=0.605, loss_mean=0.854, loss_mean_cls=0.0859, proj_loss=-0.335][2026-03-26 12:36:29] Step: 1302, Training Logs: loss_final: 0.608544, loss_mean: 0.880515, proj_loss: -0.353563, loss_mean_cls: 0.081592, grad_norm: 0.875259 +Steps: 0%| | 1303/1000000 [05:26<68:44:32, 4.04it/s, grad_norm=0.875, loss_final=0.609, loss_mean=0.881, loss_mean_cls=0.0816, proj_loss=-0.354][2026-03-26 12:36:29] Step: 1303, Training Logs: loss_final: 0.608986, loss_mean: 0.870157, proj_loss: -0.344306, loss_mean_cls: 0.083134, grad_norm: 0.608591 +Steps: 0%| | 1304/1000000 [05:26<68:45:09, 4.03it/s, grad_norm=0.609, loss_final=0.609, loss_mean=0.87, loss_mean_cls=0.0831, proj_loss=-0.344][2026-03-26 12:36:30] Step: 1304, Training Logs: loss_final: 0.605606, loss_mean: 0.846110, proj_loss: -0.327543, loss_mean_cls: 0.087040, grad_norm: 0.612588 +Steps: 0%| | 1305/1000000 [05:27<68:43:30, 4.04it/s, grad_norm=0.613, loss_final=0.606, loss_mean=0.846, loss_mean_cls=0.087, proj_loss=-0.328][2026-03-26 12:36:30] Step: 1305, Training Logs: loss_final: 0.599536, loss_mean: 0.854039, proj_loss: -0.340714, loss_mean_cls: 0.086212, grad_norm: 0.609849 +Steps: 0%| | 1306/1000000 [05:27<68:42:24, 4.04it/s, grad_norm=0.61, loss_final=0.6, loss_mean=0.854, loss_mean_cls=0.0862, proj_loss=-0.341][2026-03-26 12:36:30] Step: 1306, Training Logs: loss_final: 0.616036, loss_mean: 0.869705, proj_loss: -0.337304, loss_mean_cls: 0.083635, grad_norm: 0.600719 +Steps: 0%| | 1307/1000000 [05:27<68:41:16, 4.04it/s, grad_norm=0.601, loss_final=0.616, loss_mean=0.87, loss_mean_cls=0.0836, proj_loss=-0.337][2026-03-26 12:36:30] Step: 1307, Training Logs: loss_final: 0.617364, loss_mean: 0.869846, proj_loss: -0.337603, loss_mean_cls: 0.085120, grad_norm: 0.526692 +Steps: 0%| | 1308/1000000 [05:27<68:40:32, 4.04it/s, grad_norm=0.527, loss_final=0.617, loss_mean=0.87, loss_mean_cls=0.0851, proj_loss=-0.338][2026-03-26 12:36:31] Step: 1308, Training Logs: loss_final: 0.598789, loss_mean: 0.863339, proj_loss: -0.347832, loss_mean_cls: 0.083282, grad_norm: 0.690316 +Steps: 0%| | 1309/1000000 [05:28<68:37:22, 4.04it/s, grad_norm=0.69, loss_final=0.599, loss_mean=0.863, loss_mean_cls=0.0833, proj_loss=-0.348][2026-03-26 12:36:31] Step: 1309, Training Logs: loss_final: 0.613279, loss_mean: 0.850375, proj_loss: -0.324424, loss_mean_cls: 0.087328, grad_norm: 0.516835 +Steps: 0%| | 1310/1000000 [05:28<68:37:14, 4.04it/s, grad_norm=0.517, loss_final=0.613, loss_mean=0.85, loss_mean_cls=0.0873, proj_loss=-0.324][2026-03-26 12:36:31] Step: 1310, Training Logs: loss_final: 0.603941, loss_mean: 0.852869, proj_loss: -0.333977, loss_mean_cls: 0.085049, grad_norm: 0.600309 +Steps: 0%| | 1311/1000000 [05:28<68:38:10, 4.04it/s, grad_norm=0.6, loss_final=0.604, loss_mean=0.853, loss_mean_cls=0.085, proj_loss=-0.334][2026-03-26 12:36:31] Step: 1311, Training Logs: loss_final: 0.611245, loss_mean: 0.862415, proj_loss: -0.337085, loss_mean_cls: 0.085915, grad_norm: 0.690744 +Steps: 0%| | 1312/1000000 [05:28<68:36:49, 4.04it/s, grad_norm=0.691, loss_final=0.611, loss_mean=0.862, loss_mean_cls=0.0859, proj_loss=-0.337][2026-03-26 12:36:32] Step: 1312, Training Logs: loss_final: 0.617415, loss_mean: 0.881532, proj_loss: -0.347131, loss_mean_cls: 0.083014, grad_norm: 0.530847 +Steps: 0%| | 1313/1000000 [05:29<68:37:24, 4.04it/s, grad_norm=0.531, loss_final=0.617, loss_mean=0.882, loss_mean_cls=0.083, proj_loss=-0.347][2026-03-26 12:36:32] Step: 1313, Training Logs: loss_final: 0.614918, loss_mean: 0.870590, proj_loss: -0.339267, loss_mean_cls: 0.083596, grad_norm: 0.586656 +Steps: 0%| | 1314/1000000 [05:29<68:36:30, 4.04it/s, grad_norm=0.587, loss_final=0.615, loss_mean=0.871, loss_mean_cls=0.0836, proj_loss=-0.339][2026-03-26 12:36:32] Step: 1314, Training Logs: loss_final: 0.608169, loss_mean: 0.865218, proj_loss: -0.341291, loss_mean_cls: 0.084242, grad_norm: 0.711037 +Steps: 0%| | 1315/1000000 [05:29<68:40:58, 4.04it/s, grad_norm=0.711, loss_final=0.608, loss_mean=0.865, loss_mean_cls=0.0842, proj_loss=-0.341][2026-03-26 12:36:32] Step: 1315, Training Logs: loss_final: 0.615347, loss_mean: 0.863994, proj_loss: -0.334239, loss_mean_cls: 0.085592, grad_norm: 0.612397 +Steps: 0%| | 1316/1000000 [05:29<68:41:11, 4.04it/s, grad_norm=0.612, loss_final=0.615, loss_mean=0.864, loss_mean_cls=0.0856, proj_loss=-0.334][2026-03-26 12:36:33] Step: 1316, Training Logs: loss_final: 0.619161, loss_mean: 0.856749, proj_loss: -0.324217, loss_mean_cls: 0.086629, grad_norm: 0.775053 +Steps: 0%| | 1317/1000000 [05:30<68:41:05, 4.04it/s, grad_norm=0.775, loss_final=0.619, loss_mean=0.857, loss_mean_cls=0.0866, proj_loss=-0.324][2026-03-26 12:36:33] Step: 1317, Training Logs: loss_final: 0.613099, loss_mean: 0.857366, proj_loss: -0.329111, loss_mean_cls: 0.084844, grad_norm: 0.692673 +Steps: 0%| | 1318/1000000 [05:30<68:39:11, 4.04it/s, grad_norm=0.693, loss_final=0.613, loss_mean=0.857, loss_mean_cls=0.0848, proj_loss=-0.329][2026-03-26 12:36:33] Step: 1318, Training Logs: loss_final: 0.608370, loss_mean: 0.855058, proj_loss: -0.333010, loss_mean_cls: 0.086321, grad_norm: 0.447567 +Steps: 0%| | 1319/1000000 [05:30<68:38:46, 4.04it/s, grad_norm=0.448, loss_final=0.608, loss_mean=0.855, loss_mean_cls=0.0863, proj_loss=-0.333][2026-03-26 12:36:33] Step: 1319, Training Logs: loss_final: 0.610681, loss_mean: 0.868420, proj_loss: -0.341510, loss_mean_cls: 0.083771, grad_norm: 0.408861 +Steps: 0%| | 1320/1000000 [05:30<68:36:16, 4.04it/s, grad_norm=0.409, loss_final=0.611, loss_mean=0.868, loss_mean_cls=0.0838, proj_loss=-0.342][2026-03-26 12:36:34] Step: 1320, Training Logs: loss_final: 0.619302, loss_mean: 0.878511, proj_loss: -0.342155, loss_mean_cls: 0.082946, grad_norm: 0.473648 +Steps: 0%| | 1321/1000000 [05:31<68:37:09, 4.04it/s, grad_norm=0.474, loss_final=0.619, loss_mean=0.879, loss_mean_cls=0.0829, proj_loss=-0.342][2026-03-26 12:36:34] Step: 1321, Training Logs: loss_final: 0.611448, loss_mean: 0.874104, proj_loss: -0.346700, loss_mean_cls: 0.084044, grad_norm: 0.440796 +Steps: 0%| | 1322/1000000 [05:31<68:36:14, 4.04it/s, grad_norm=0.441, loss_final=0.611, loss_mean=0.874, loss_mean_cls=0.084, proj_loss=-0.347][2026-03-26 12:36:34] Step: 1322, Training Logs: loss_final: 0.610829, loss_mean: 0.849917, proj_loss: -0.325752, loss_mean_cls: 0.086665, grad_norm: 0.508950 +Steps: 0%| | 1323/1000000 [05:31<68:39:24, 4.04it/s, grad_norm=0.509, loss_final=0.611, loss_mean=0.85, loss_mean_cls=0.0867, proj_loss=-0.326][2026-03-26 12:36:34] Step: 1323, Training Logs: loss_final: 0.600604, loss_mean: 0.847798, proj_loss: -0.331960, loss_mean_cls: 0.084766, grad_norm: 0.507002 +Steps: 0%| | 1324/1000000 [05:31<68:39:25, 4.04it/s, grad_norm=0.507, loss_final=0.601, loss_mean=0.848, loss_mean_cls=0.0848, proj_loss=-0.332][2026-03-26 12:36:35] Step: 1324, Training Logs: loss_final: 0.619410, loss_mean: 0.888751, proj_loss: -0.350468, loss_mean_cls: 0.081127, grad_norm: 0.384170 +Steps: 0%| | 1325/1000000 [05:32<68:40:58, 4.04it/s, grad_norm=0.384, loss_final=0.619, loss_mean=0.889, loss_mean_cls=0.0811, proj_loss=-0.35][2026-03-26 12:36:35] Step: 1325, Training Logs: loss_final: 0.625314, loss_mean: 0.890792, proj_loss: -0.346595, loss_mean_cls: 0.081117, grad_norm: 0.409707 +Steps: 0%| | 1326/1000000 [05:32<68:42:31, 4.04it/s, grad_norm=0.41, loss_final=0.625, loss_mean=0.891, loss_mean_cls=0.0811, proj_loss=-0.347][2026-03-26 12:36:35] Step: 1326, Training Logs: loss_final: 0.616100, loss_mean: 0.860692, proj_loss: -0.330875, loss_mean_cls: 0.086283, grad_norm: 0.390788 +Steps: 0%| | 1327/1000000 [05:32<68:39:05, 4.04it/s, grad_norm=0.391, loss_final=0.616, loss_mean=0.861, loss_mean_cls=0.0863, proj_loss=-0.331][2026-03-26 12:36:35] Step: 1327, Training Logs: loss_final: 0.616264, loss_mean: 0.875302, proj_loss: -0.341799, loss_mean_cls: 0.082761, grad_norm: 0.402311 +Steps: 0%| | 1328/1000000 [05:32<68:41:44, 4.04it/s, grad_norm=0.402, loss_final=0.616, loss_mean=0.875, loss_mean_cls=0.0828, proj_loss=-0.342][2026-03-26 12:36:36] Step: 1328, Training Logs: loss_final: 0.629573, loss_mean: 0.875934, proj_loss: -0.332057, loss_mean_cls: 0.085696, grad_norm: 0.453946 +Steps: 0%| | 1329/1000000 [05:33<68:42:13, 4.04it/s, grad_norm=0.454, loss_final=0.63, loss_mean=0.876, loss_mean_cls=0.0857, proj_loss=-0.332][2026-03-26 12:36:36] Step: 1329, Training Logs: loss_final: 0.623716, loss_mean: 0.887433, proj_loss: -0.346508, loss_mean_cls: 0.082791, grad_norm: 0.337914 +Steps: 0%| | 1330/1000000 [05:33<68:41:26, 4.04it/s, grad_norm=0.338, loss_final=0.624, loss_mean=0.887, loss_mean_cls=0.0828, proj_loss=-0.347][2026-03-26 12:36:36] Step: 1330, Training Logs: loss_final: 0.620890, loss_mean: 0.886128, proj_loss: -0.347539, loss_mean_cls: 0.082301, grad_norm: 0.638880 +Steps: 0%| | 1331/1000000 [05:33<68:41:21, 4.04it/s, grad_norm=0.639, loss_final=0.621, loss_mean=0.886, loss_mean_cls=0.0823, proj_loss=-0.348][2026-03-26 12:36:36] Step: 1331, Training Logs: loss_final: 0.621140, loss_mean: 0.889928, proj_loss: -0.350339, loss_mean_cls: 0.081552, grad_norm: 0.548028 +Steps: 0%| | 1332/1000000 [05:33<68:41:52, 4.04it/s, grad_norm=0.548, loss_final=0.621, loss_mean=0.89, loss_mean_cls=0.0816, proj_loss=-0.35][2026-03-26 12:36:37] Step: 1332, Training Logs: loss_final: 0.612839, loss_mean: 0.873454, proj_loss: -0.343046, loss_mean_cls: 0.082431, grad_norm: 0.660014 +Steps: 0%| | 1333/1000000 [05:33<68:39:12, 4.04it/s, grad_norm=0.66, loss_final=0.613, loss_mean=0.873, loss_mean_cls=0.0824, proj_loss=-0.343][2026-03-26 12:36:37] Step: 1333, Training Logs: loss_final: 0.605854, loss_mean: 0.862471, proj_loss: -0.340865, loss_mean_cls: 0.084248, grad_norm: 0.508658 +Steps: 0%| | 1334/1000000 [05:34<68:41:09, 4.04it/s, grad_norm=0.509, loss_final=0.606, loss_mean=0.862, loss_mean_cls=0.0842, proj_loss=-0.341][2026-03-26 12:36:37] Step: 1334, Training Logs: loss_final: 0.608654, loss_mean: 0.863986, proj_loss: -0.340016, loss_mean_cls: 0.084683, grad_norm: 0.670588 +Steps: 0%| | 1335/1000000 [05:34<68:40:26, 4.04it/s, grad_norm=0.671, loss_final=0.609, loss_mean=0.864, loss_mean_cls=0.0847, proj_loss=-0.34][2026-03-26 12:36:37] Step: 1335, Training Logs: loss_final: 0.620345, loss_mean: 0.864014, proj_loss: -0.330078, loss_mean_cls: 0.086410, grad_norm: 0.320587 +Steps: 0%| | 1336/1000000 [05:34<68:51:37, 4.03it/s, grad_norm=0.321, loss_final=0.62, loss_mean=0.864, loss_mean_cls=0.0864, proj_loss=-0.33][2026-03-26 12:36:38] Step: 1336, Training Logs: loss_final: 0.595741, loss_mean: 0.864541, proj_loss: -0.351275, loss_mean_cls: 0.082475, grad_norm: 0.648682 +Steps: 0%| | 1337/1000000 [05:34<68:47:54, 4.03it/s, grad_norm=0.649, loss_final=0.596, loss_mean=0.865, loss_mean_cls=0.0825, proj_loss=-0.351][2026-03-26 12:36:38] Step: 1337, Training Logs: loss_final: 0.621187, loss_mean: 0.884716, proj_loss: -0.345515, loss_mean_cls: 0.081987, grad_norm: 0.378218 +Steps: 0%| | 1338/1000000 [05:35<68:47:33, 4.03it/s, grad_norm=0.378, loss_final=0.621, loss_mean=0.885, loss_mean_cls=0.082, proj_loss=-0.346][2026-03-26 12:36:38] Step: 1338, Training Logs: loss_final: 0.613790, loss_mean: 0.855881, proj_loss: -0.329017, loss_mean_cls: 0.086926, grad_norm: 0.494173 +Steps: 0%| | 1339/1000000 [05:35<68:44:11, 4.04it/s, grad_norm=0.494, loss_final=0.614, loss_mean=0.856, loss_mean_cls=0.0869, proj_loss=-0.329][2026-03-26 12:36:38] Step: 1339, Training Logs: loss_final: 0.613078, loss_mean: 0.866247, proj_loss: -0.338054, loss_mean_cls: 0.084885, grad_norm: 0.490448 +Steps: 0%| | 1340/1000000 [05:35<68:43:38, 4.04it/s, grad_norm=0.49, loss_final=0.613, loss_mean=0.866, loss_mean_cls=0.0849, proj_loss=-0.338][2026-03-26 12:36:39] Step: 1340, Training Logs: loss_final: 0.604317, loss_mean: 0.872631, proj_loss: -0.351510, loss_mean_cls: 0.083196, grad_norm: 0.611964 +Steps: 0%| | 1341/1000000 [05:35<68:42:24, 4.04it/s, grad_norm=0.612, loss_final=0.604, loss_mean=0.873, loss_mean_cls=0.0832, proj_loss=-0.352][2026-03-26 12:36:39] Step: 1341, Training Logs: loss_final: 0.603098, loss_mean: 0.849731, proj_loss: -0.333369, loss_mean_cls: 0.086736, grad_norm: 0.626761 +Steps: 0%| | 1342/1000000 [05:36<68:41:29, 4.04it/s, grad_norm=0.627, loss_final=0.603, loss_mean=0.85, loss_mean_cls=0.0867, proj_loss=-0.333][2026-03-26 12:36:39] Step: 1342, Training Logs: loss_final: 0.615925, loss_mean: 0.876319, proj_loss: -0.344390, loss_mean_cls: 0.083996, grad_norm: 0.780825 +Steps: 0%| | 1343/1000000 [05:36<68:40:22, 4.04it/s, grad_norm=0.781, loss_final=0.616, loss_mean=0.876, loss_mean_cls=0.084, proj_loss=-0.344][2026-03-26 12:36:39] Step: 1343, Training Logs: loss_final: 0.598856, loss_mean: 0.862955, proj_loss: -0.346648, loss_mean_cls: 0.082549, grad_norm: 0.638331 +Steps: 0%| | 1344/1000000 [05:36<68:39:13, 4.04it/s, grad_norm=0.638, loss_final=0.599, loss_mean=0.863, loss_mean_cls=0.0825, proj_loss=-0.347][2026-03-26 12:36:40] Step: 1344, Training Logs: loss_final: 0.611004, loss_mean: 0.865609, proj_loss: -0.338830, loss_mean_cls: 0.084225, grad_norm: 0.569610 +Steps: 0%| | 1345/1000000 [05:36<68:40:56, 4.04it/s, grad_norm=0.57, loss_final=0.611, loss_mean=0.866, loss_mean_cls=0.0842, proj_loss=-0.339][2026-03-26 12:36:40] Step: 1345, Training Logs: loss_final: 0.608625, loss_mean: 0.881698, proj_loss: -0.354135, loss_mean_cls: 0.081061, grad_norm: 0.770071 +Steps: 0%| | 1346/1000000 [05:37<68:39:50, 4.04it/s, grad_norm=0.77, loss_final=0.609, loss_mean=0.882, loss_mean_cls=0.0811, proj_loss=-0.354][2026-03-26 12:36:40] Step: 1346, Training Logs: loss_final: 0.622765, loss_mean: 0.868909, proj_loss: -0.331898, loss_mean_cls: 0.085754, grad_norm: 0.825083 +Steps: 0%| | 1347/1000000 [05:37<68:38:26, 4.04it/s, grad_norm=0.825, loss_final=0.623, loss_mean=0.869, loss_mean_cls=0.0858, proj_loss=-0.332][2026-03-26 12:36:40] Step: 1347, Training Logs: loss_final: 0.605843, loss_mean: 0.873294, proj_loss: -0.349157, loss_mean_cls: 0.081706, grad_norm: 0.745728 +Steps: 0%| | 1348/1000000 [05:37<68:40:29, 4.04it/s, grad_norm=0.746, loss_final=0.606, loss_mean=0.873, loss_mean_cls=0.0817, proj_loss=-0.349][2026-03-26 12:36:41] Step: 1348, Training Logs: loss_final: 0.607464, loss_mean: 0.867309, proj_loss: -0.344165, loss_mean_cls: 0.084320, grad_norm: 0.643488 +Steps: 0%| | 1349/1000000 [05:37<68:40:40, 4.04it/s, grad_norm=0.643, loss_final=0.607, loss_mean=0.867, loss_mean_cls=0.0843, proj_loss=-0.344][2026-03-26 12:36:41] Step: 1349, Training Logs: loss_final: 0.591062, loss_mean: 0.835035, proj_loss: -0.330850, loss_mean_cls: 0.086876, grad_norm: 0.560750 +Steps: 0%| | 1350/1000000 [05:38<69:11:04, 4.01it/s, grad_norm=0.561, loss_final=0.591, loss_mean=0.835, loss_mean_cls=0.0869, proj_loss=-0.331][2026-03-26 12:36:41] Step: 1350, Training Logs: loss_final: 0.616268, loss_mean: 0.872233, proj_loss: -0.340556, loss_mean_cls: 0.084590, grad_norm: 0.658195 +Steps: 0%| | 1351/1000000 [05:38<69:04:24, 4.02it/s, grad_norm=0.658, loss_final=0.616, loss_mean=0.872, loss_mean_cls=0.0846, proj_loss=-0.341][2026-03-26 12:36:41] Step: 1351, Training Logs: loss_final: 0.617692, loss_mean: 0.889614, proj_loss: -0.354247, loss_mean_cls: 0.082325, grad_norm: 0.657129 +Steps: 0%| | 1352/1000000 [05:38<69:01:34, 4.02it/s, grad_norm=0.657, loss_final=0.618, loss_mean=0.89, loss_mean_cls=0.0823, proj_loss=-0.354][2026-03-26 12:36:42] Step: 1352, Training Logs: loss_final: 0.599037, loss_mean: 0.849792, proj_loss: -0.335555, loss_mean_cls: 0.084800, grad_norm: 0.827470 +Steps: 0%| | 1353/1000000 [05:38<69:07:47, 4.01it/s, grad_norm=0.827, loss_final=0.599, loss_mean=0.85, loss_mean_cls=0.0848, proj_loss=-0.336][2026-03-26 12:36:42] Step: 1353, Training Logs: loss_final: 0.613257, loss_mean: 0.866039, proj_loss: -0.338359, loss_mean_cls: 0.085577, grad_norm: 0.467871 +Steps: 0%| | 1354/1000000 [05:39<68:59:50, 4.02it/s, grad_norm=0.468, loss_final=0.613, loss_mean=0.866, loss_mean_cls=0.0856, proj_loss=-0.338][2026-03-26 12:36:42] Step: 1354, Training Logs: loss_final: 0.617947, loss_mean: 0.894286, proj_loss: -0.357638, loss_mean_cls: 0.081300, grad_norm: 0.642245 +Steps: 0%| | 1355/1000000 [05:39<68:52:59, 4.03it/s, grad_norm=0.642, loss_final=0.618, loss_mean=0.894, loss_mean_cls=0.0813, proj_loss=-0.358][2026-03-26 12:36:42] Step: 1355, Training Logs: loss_final: 0.612694, loss_mean: 0.851232, proj_loss: -0.324185, loss_mean_cls: 0.085647, grad_norm: 0.527559 +Steps: 0%| | 1356/1000000 [05:39<68:48:57, 4.03it/s, grad_norm=0.528, loss_final=0.613, loss_mean=0.851, loss_mean_cls=0.0856, proj_loss=-0.324][2026-03-26 12:36:43] Step: 1356, Training Logs: loss_final: 0.613216, loss_mean: 0.891581, proj_loss: -0.358711, loss_mean_cls: 0.080345, grad_norm: 0.407601 +Steps: 0%| | 1357/1000000 [05:39<68:44:56, 4.03it/s, grad_norm=0.408, loss_final=0.613, loss_mean=0.892, loss_mean_cls=0.0803, proj_loss=-0.359][2026-03-26 12:36:43] Step: 1357, Training Logs: loss_final: 0.597268, loss_mean: 0.854801, proj_loss: -0.341550, loss_mean_cls: 0.084017, grad_norm: 0.574650 +Steps: 0%| | 1358/1000000 [05:40<68:44:15, 4.04it/s, grad_norm=0.575, loss_final=0.597, loss_mean=0.855, loss_mean_cls=0.084, proj_loss=-0.342][2026-03-26 12:36:43] Step: 1358, Training Logs: loss_final: 0.594549, loss_mean: 0.850338, proj_loss: -0.342533, loss_mean_cls: 0.086744, grad_norm: 0.531220 +Steps: 0%| | 1359/1000000 [05:40<68:42:53, 4.04it/s, grad_norm=0.531, loss_final=0.595, loss_mean=0.85, loss_mean_cls=0.0867, proj_loss=-0.343][2026-03-26 12:36:43] Step: 1359, Training Logs: loss_final: 0.604329, loss_mean: 0.863702, proj_loss: -0.341989, loss_mean_cls: 0.082615, grad_norm: 0.572408 +Steps: 0%| | 1360/1000000 [05:40<68:39:46, 4.04it/s, grad_norm=0.572, loss_final=0.604, loss_mean=0.864, loss_mean_cls=0.0826, proj_loss=-0.342][2026-03-26 12:36:44] Step: 1360, Training Logs: loss_final: 0.613517, loss_mean: 0.873922, proj_loss: -0.343471, loss_mean_cls: 0.083066, grad_norm: 0.510006 +Steps: 0%| | 1361/1000000 [05:40<68:40:03, 4.04it/s, grad_norm=0.51, loss_final=0.614, loss_mean=0.874, loss_mean_cls=0.0831, proj_loss=-0.343][2026-03-26 12:36:44] Step: 1361, Training Logs: loss_final: 0.600642, loss_mean: 0.850599, proj_loss: -0.335893, loss_mean_cls: 0.085936, grad_norm: 0.634274 +Steps: 0%| | 1362/1000000 [05:41<68:39:20, 4.04it/s, grad_norm=0.634, loss_final=0.601, loss_mean=0.851, loss_mean_cls=0.0859, proj_loss=-0.336][2026-03-26 12:36:44] Step: 1362, Training Logs: loss_final: 0.610130, loss_mean: 0.863649, proj_loss: -0.337305, loss_mean_cls: 0.083786, grad_norm: 0.453937 +Steps: 0%| | 1363/1000000 [05:41<68:43:08, 4.04it/s, grad_norm=0.454, loss_final=0.61, loss_mean=0.864, loss_mean_cls=0.0838, proj_loss=-0.337][2026-03-26 12:36:44] Step: 1363, Training Logs: loss_final: 0.603440, loss_mean: 0.871933, proj_loss: -0.349625, loss_mean_cls: 0.081132, grad_norm: 0.486836 +Steps: 0%| | 1364/1000000 [05:41<68:42:15, 4.04it/s, grad_norm=0.487, loss_final=0.603, loss_mean=0.872, loss_mean_cls=0.0811, proj_loss=-0.35][2026-03-26 12:36:45] Step: 1364, Training Logs: loss_final: 0.618612, loss_mean: 0.874127, proj_loss: -0.339656, loss_mean_cls: 0.084141, grad_norm: 0.694603 +Steps: 0%| | 1365/1000000 [05:41<68:43:15, 4.04it/s, grad_norm=0.695, loss_final=0.619, loss_mean=0.874, loss_mean_cls=0.0841, proj_loss=-0.34][2026-03-26 12:36:45] Step: 1365, Training Logs: loss_final: 0.607560, loss_mean: 0.878567, proj_loss: -0.352874, loss_mean_cls: 0.081867, grad_norm: 0.384706 +Steps: 0%| | 1366/1000000 [05:42<68:41:18, 4.04it/s, grad_norm=0.385, loss_final=0.608, loss_mean=0.879, loss_mean_cls=0.0819, proj_loss=-0.353][2026-03-26 12:36:45] Step: 1366, Training Logs: loss_final: 0.616912, loss_mean: 0.868185, proj_loss: -0.337221, loss_mean_cls: 0.085948, grad_norm: 0.869208 +Steps: 0%| | 1367/1000000 [05:42<68:43:26, 4.04it/s, grad_norm=0.869, loss_final=0.617, loss_mean=0.868, loss_mean_cls=0.0859, proj_loss=-0.337][2026-03-26 12:36:45] Step: 1367, Training Logs: loss_final: 0.610021, loss_mean: 0.879562, proj_loss: -0.351020, loss_mean_cls: 0.081480, grad_norm: 0.417903 +Steps: 0%| | 1368/1000000 [05:42<68:42:58, 4.04it/s, grad_norm=0.418, loss_final=0.61, loss_mean=0.88, loss_mean_cls=0.0815, proj_loss=-0.351][2026-03-26 12:36:46] Step: 1368, Training Logs: loss_final: 0.617231, loss_mean: 0.879539, proj_loss: -0.344786, loss_mean_cls: 0.082479, grad_norm: 0.846736 +Steps: 0%| | 1369/1000000 [05:42<68:45:01, 4.03it/s, grad_norm=0.847, loss_final=0.617, loss_mean=0.88, loss_mean_cls=0.0825, proj_loss=-0.345][2026-03-26 12:36:46] Step: 1369, Training Logs: loss_final: 0.609338, loss_mean: 0.874817, proj_loss: -0.348595, loss_mean_cls: 0.083116, grad_norm: 0.519522 +Steps: 0%| | 1370/1000000 [05:43<68:43:16, 4.04it/s, grad_norm=0.52, loss_final=0.609, loss_mean=0.875, loss_mean_cls=0.0831, proj_loss=-0.349][2026-03-26 12:36:46] Step: 1370, Training Logs: loss_final: 0.593450, loss_mean: 0.850851, proj_loss: -0.342893, loss_mean_cls: 0.085491, grad_norm: 0.544421 +Steps: 0%| | 1371/1000000 [05:43<68:41:47, 4.04it/s, grad_norm=0.544, loss_final=0.593, loss_mean=0.851, loss_mean_cls=0.0855, proj_loss=-0.343][2026-03-26 12:36:46] Step: 1371, Training Logs: loss_final: 0.619983, loss_mean: 0.889808, proj_loss: -0.351855, loss_mean_cls: 0.082030, grad_norm: 0.661019 +Steps: 0%| | 1372/1000000 [05:43<68:42:58, 4.04it/s, grad_norm=0.661, loss_final=0.62, loss_mean=0.89, loss_mean_cls=0.082, proj_loss=-0.352][2026-03-26 12:36:47] Step: 1372, Training Logs: loss_final: 0.615251, loss_mean: 0.878832, proj_loss: -0.345480, loss_mean_cls: 0.081899, grad_norm: 0.457743 +Steps: 0%| | 1373/1000000 [05:43<68:43:11, 4.04it/s, grad_norm=0.458, loss_final=0.615, loss_mean=0.879, loss_mean_cls=0.0819, proj_loss=-0.345][2026-03-26 12:36:47] Step: 1373, Training Logs: loss_final: 0.589954, loss_mean: 0.834472, proj_loss: -0.330971, loss_mean_cls: 0.086453, grad_norm: 0.473442 +Steps: 0%| | 1374/1000000 [05:44<68:47:14, 4.03it/s, grad_norm=0.473, loss_final=0.59, loss_mean=0.834, loss_mean_cls=0.0865, proj_loss=-0.331][2026-03-26 12:36:47] Step: 1374, Training Logs: loss_final: 0.606181, loss_mean: 0.862203, proj_loss: -0.341397, loss_mean_cls: 0.085375, grad_norm: 0.447233 +Steps: 0%| | 1375/1000000 [05:44<68:45:59, 4.03it/s, grad_norm=0.447, loss_final=0.606, loss_mean=0.862, loss_mean_cls=0.0854, proj_loss=-0.341][2026-03-26 12:36:47] Step: 1375, Training Logs: loss_final: 0.619899, loss_mean: 0.861564, proj_loss: -0.328212, loss_mean_cls: 0.086547, grad_norm: 0.687790 +Steps: 0%| | 1376/1000000 [05:44<68:43:17, 4.04it/s, grad_norm=0.688, loss_final=0.62, loss_mean=0.862, loss_mean_cls=0.0865, proj_loss=-0.328][2026-03-26 12:36:48] Step: 1376, Training Logs: loss_final: 0.602240, loss_mean: 0.871428, proj_loss: -0.350691, loss_mean_cls: 0.081504, grad_norm: 0.437255 +Steps: 0%| | 1377/1000000 [05:44<68:41:48, 4.04it/s, grad_norm=0.437, loss_final=0.602, loss_mean=0.871, loss_mean_cls=0.0815, proj_loss=-0.351][2026-03-26 12:36:48] Step: 1377, Training Logs: loss_final: 0.617444, loss_mean: 0.875696, proj_loss: -0.342009, loss_mean_cls: 0.083757, grad_norm: 0.539332 +Steps: 0%| | 1378/1000000 [05:45<68:42:11, 4.04it/s, grad_norm=0.539, loss_final=0.617, loss_mean=0.876, loss_mean_cls=0.0838, proj_loss=-0.342][2026-03-26 12:36:48] Step: 1378, Training Logs: loss_final: 0.605848, loss_mean: 0.861406, proj_loss: -0.338398, loss_mean_cls: 0.082840, grad_norm: 0.532531 +Steps: 0%| | 1379/1000000 [05:45<68:40:46, 4.04it/s, grad_norm=0.533, loss_final=0.606, loss_mean=0.861, loss_mean_cls=0.0828, proj_loss=-0.338][2026-03-26 12:36:48] Step: 1379, Training Logs: loss_final: 0.616479, loss_mean: 0.864803, proj_loss: -0.333107, loss_mean_cls: 0.084783, grad_norm: 0.536482 +Steps: 0%| | 1380/1000000 [05:45<68:41:26, 4.04it/s, grad_norm=0.536, loss_final=0.616, loss_mean=0.865, loss_mean_cls=0.0848, proj_loss=-0.333][2026-03-26 12:36:49] Step: 1380, Training Logs: loss_final: 0.602079, loss_mean: 0.858276, proj_loss: -0.341707, loss_mean_cls: 0.085510, grad_norm: 0.480226 +Steps: 0%| | 1381/1000000 [05:45<68:41:28, 4.04it/s, grad_norm=0.48, loss_final=0.602, loss_mean=0.858, loss_mean_cls=0.0855, proj_loss=-0.342][2026-03-26 12:36:49] Step: 1381, Training Logs: loss_final: 0.598618, loss_mean: 0.845427, proj_loss: -0.333114, loss_mean_cls: 0.086304, grad_norm: 0.513925 +Steps: 0%| | 1382/1000000 [05:46<68:43:46, 4.04it/s, grad_norm=0.514, loss_final=0.599, loss_mean=0.845, loss_mean_cls=0.0863, proj_loss=-0.333][2026-03-26 12:36:49] Step: 1382, Training Logs: loss_final: 0.604723, loss_mean: 0.869929, proj_loss: -0.348634, loss_mean_cls: 0.083428, grad_norm: 0.425273 +Steps: 0%| | 1383/1000000 [05:46<68:41:18, 4.04it/s, grad_norm=0.425, loss_final=0.605, loss_mean=0.87, loss_mean_cls=0.0834, proj_loss=-0.349][2026-03-26 12:36:49] Step: 1383, Training Logs: loss_final: 0.604004, loss_mean: 0.872124, proj_loss: -0.351584, loss_mean_cls: 0.083464, grad_norm: 0.565045 +Steps: 0%| | 1384/1000000 [05:46<68:40:22, 4.04it/s, grad_norm=0.565, loss_final=0.604, loss_mean=0.872, loss_mean_cls=0.0835, proj_loss=-0.352][2026-03-26 12:36:50] Step: 1384, Training Logs: loss_final: 0.625384, loss_mean: 0.894662, proj_loss: -0.350641, loss_mean_cls: 0.081364, grad_norm: 0.456131 +Steps: 0%| | 1385/1000000 [05:46<68:37:43, 4.04it/s, grad_norm=0.456, loss_final=0.625, loss_mean=0.895, loss_mean_cls=0.0814, proj_loss=-0.351][2026-03-26 12:36:50] Step: 1385, Training Logs: loss_final: 0.611643, loss_mean: 0.874435, proj_loss: -0.346319, loss_mean_cls: 0.083526, grad_norm: 0.601464 +Steps: 0%| | 1386/1000000 [05:47<68:40:45, 4.04it/s, grad_norm=0.601, loss_final=0.612, loss_mean=0.874, loss_mean_cls=0.0835, proj_loss=-0.346][2026-03-26 12:36:50] Step: 1386, Training Logs: loss_final: 0.607089, loss_mean: 0.864524, proj_loss: -0.341275, loss_mean_cls: 0.083840, grad_norm: 0.568720 +Steps: 0%| | 1387/1000000 [05:47<68:38:13, 4.04it/s, grad_norm=0.569, loss_final=0.607, loss_mean=0.865, loss_mean_cls=0.0838, proj_loss=-0.341][2026-03-26 12:36:50] Step: 1387, Training Logs: loss_final: 0.599148, loss_mean: 0.858533, proj_loss: -0.342458, loss_mean_cls: 0.083072, grad_norm: 0.427418 +Steps: 0%| | 1388/1000000 [05:47<68:38:16, 4.04it/s, grad_norm=0.427, loss_final=0.599, loss_mean=0.859, loss_mean_cls=0.0831, proj_loss=-0.342][2026-03-26 12:36:51] Step: 1388, Training Logs: loss_final: 0.605904, loss_mean: 0.870399, proj_loss: -0.348108, loss_mean_cls: 0.083612, grad_norm: 0.615118 +Steps: 0%| | 1389/1000000 [05:47<68:40:07, 4.04it/s, grad_norm=0.615, loss_final=0.606, loss_mean=0.87, loss_mean_cls=0.0836, proj_loss=-0.348][2026-03-26 12:36:51] Step: 1389, Training Logs: loss_final: 0.608254, loss_mean: 0.859291, proj_loss: -0.336640, loss_mean_cls: 0.085603, grad_norm: 0.478843 +Steps: 0%| | 1390/1000000 [05:48<68:41:49, 4.04it/s, grad_norm=0.479, loss_final=0.608, loss_mean=0.859, loss_mean_cls=0.0856, proj_loss=-0.337][2026-03-26 12:36:51] Step: 1390, Training Logs: loss_final: 0.605960, loss_mean: 0.858991, proj_loss: -0.337343, loss_mean_cls: 0.084312, grad_norm: 0.629243 +Steps: 0%| | 1391/1000000 [05:48<68:42:12, 4.04it/s, grad_norm=0.629, loss_final=0.606, loss_mean=0.859, loss_mean_cls=0.0843, proj_loss=-0.337][2026-03-26 12:36:51] Step: 1391, Training Logs: loss_final: 0.608541, loss_mean: 0.867638, proj_loss: -0.341287, loss_mean_cls: 0.082190, grad_norm: 0.427088 +Steps: 0%| | 1392/1000000 [05:48<68:42:33, 4.04it/s, grad_norm=0.427, loss_final=0.609, loss_mean=0.868, loss_mean_cls=0.0822, proj_loss=-0.341][2026-03-26 12:36:52] Step: 1392, Training Logs: loss_final: 0.595577, loss_mean: 0.847420, proj_loss: -0.336566, loss_mean_cls: 0.084724, grad_norm: 0.523205 +Steps: 0%| | 1393/1000000 [05:48<68:42:04, 4.04it/s, grad_norm=0.523, loss_final=0.596, loss_mean=0.847, loss_mean_cls=0.0847, proj_loss=-0.337][2026-03-26 12:36:52] Step: 1393, Training Logs: loss_final: 0.618988, loss_mean: 0.883467, proj_loss: -0.347547, loss_mean_cls: 0.083068, grad_norm: 0.507238 +Steps: 0%| | 1394/1000000 [05:49<68:42:54, 4.04it/s, grad_norm=0.507, loss_final=0.619, loss_mean=0.883, loss_mean_cls=0.0831, proj_loss=-0.348][2026-03-26 12:36:52] Step: 1394, Training Logs: loss_final: 0.605527, loss_mean: 0.872225, proj_loss: -0.350360, loss_mean_cls: 0.083663, grad_norm: 0.491672 +Steps: 0%| | 1395/1000000 [05:49<68:44:08, 4.04it/s, grad_norm=0.492, loss_final=0.606, loss_mean=0.872, loss_mean_cls=0.0837, proj_loss=-0.35][2026-03-26 12:36:52] Step: 1395, Training Logs: loss_final: 0.593780, loss_mean: 0.856869, proj_loss: -0.346655, loss_mean_cls: 0.083566, grad_norm: 0.488597 +Steps: 0%| | 1396/1000000 [05:49<68:42:05, 4.04it/s, grad_norm=0.489, loss_final=0.594, loss_mean=0.857, loss_mean_cls=0.0836, proj_loss=-0.347][2026-03-26 12:36:53] Step: 1396, Training Logs: loss_final: 0.605549, loss_mean: 0.875228, proj_loss: -0.351064, loss_mean_cls: 0.081384, grad_norm: 0.587910 +Steps: 0%| | 1397/1000000 [05:49<68:41:21, 4.04it/s, grad_norm=0.588, loss_final=0.606, loss_mean=0.875, loss_mean_cls=0.0814, proj_loss=-0.351][2026-03-26 12:36:53] Step: 1397, Training Logs: loss_final: 0.615233, loss_mean: 0.883922, proj_loss: -0.351225, loss_mean_cls: 0.082537, grad_norm: 0.424241 +Steps: 0%| | 1398/1000000 [05:50<68:40:22, 4.04it/s, grad_norm=0.424, loss_final=0.615, loss_mean=0.884, loss_mean_cls=0.0825, proj_loss=-0.351][2026-03-26 12:36:53] Step: 1398, Training Logs: loss_final: 0.592063, loss_mean: 0.823606, proj_loss: -0.319807, loss_mean_cls: 0.088264, grad_norm: 0.471653 +Steps: 0%| | 1399/1000000 [05:50<68:41:09, 4.04it/s, grad_norm=0.472, loss_final=0.592, loss_mean=0.824, loss_mean_cls=0.0883, proj_loss=-0.32][2026-03-26 12:36:53] Step: 1399, Training Logs: loss_final: 0.589415, loss_mean: 0.836056, proj_loss: -0.332711, loss_mean_cls: 0.086070, grad_norm: 0.732752 +Steps: 0%| | 1400/1000000 [05:50<68:41:50, 4.04it/s, grad_norm=0.733, loss_final=0.589, loss_mean=0.836, loss_mean_cls=0.0861, proj_loss=-0.333][2026-03-26 12:36:53] Step: 1400, Training Logs: loss_final: 0.592437, loss_mean: 0.862261, proj_loss: -0.351600, loss_mean_cls: 0.081775, grad_norm: 0.544130 +Steps: 0%| | 1401/1000000 [05:50<68:40:08, 4.04it/s, grad_norm=0.544, loss_final=0.592, loss_mean=0.862, loss_mean_cls=0.0818, proj_loss=-0.352][2026-03-26 12:36:54] Step: 1401, Training Logs: loss_final: 0.591905, loss_mean: 0.837771, proj_loss: -0.331972, loss_mean_cls: 0.086105, grad_norm: 0.683899 +Steps: 0%| | 1402/1000000 [05:51<68:39:54, 4.04it/s, grad_norm=0.684, loss_final=0.592, loss_mean=0.838, loss_mean_cls=0.0861, proj_loss=-0.332][2026-03-26 12:36:54] Step: 1402, Training Logs: loss_final: 0.606007, loss_mean: 0.867842, proj_loss: -0.345433, loss_mean_cls: 0.083598, grad_norm: 0.404789 +Steps: 0%| | 1403/1000000 [05:51<68:36:56, 4.04it/s, grad_norm=0.405, loss_final=0.606, loss_mean=0.868, loss_mean_cls=0.0836, proj_loss=-0.345][2026-03-26 12:36:54] Step: 1403, Training Logs: loss_final: 0.610472, loss_mean: 0.867352, proj_loss: -0.341933, loss_mean_cls: 0.085053, grad_norm: 0.686802 +Steps: 0%| | 1404/1000000 [05:51<68:40:38, 4.04it/s, grad_norm=0.687, loss_final=0.61, loss_mean=0.867, loss_mean_cls=0.0851, proj_loss=-0.342][2026-03-26 12:36:54] Step: 1404, Training Logs: loss_final: 0.604368, loss_mean: 0.873193, proj_loss: -0.351964, loss_mean_cls: 0.083138, grad_norm: 0.455840 +Steps: 0%| | 1405/1000000 [05:51<68:38:12, 4.04it/s, grad_norm=0.456, loss_final=0.604, loss_mean=0.873, loss_mean_cls=0.0831, proj_loss=-0.352][2026-03-26 12:36:55] Step: 1405, Training Logs: loss_final: 0.613252, loss_mean: 0.857566, proj_loss: -0.331489, loss_mean_cls: 0.087176, grad_norm: 0.521834 +Steps: 0%| | 1406/1000000 [05:52<68:39:17, 4.04it/s, grad_norm=0.522, loss_final=0.613, loss_mean=0.858, loss_mean_cls=0.0872, proj_loss=-0.331][2026-03-26 12:36:55] Step: 1406, Training Logs: loss_final: 0.611461, loss_mean: 0.874658, proj_loss: -0.345179, loss_mean_cls: 0.081981, grad_norm: 0.583697 +Steps: 0%| | 1407/1000000 [05:52<68:41:20, 4.04it/s, grad_norm=0.584, loss_final=0.611, loss_mean=0.875, loss_mean_cls=0.082, proj_loss=-0.345][2026-03-26 12:36:55] Step: 1407, Training Logs: loss_final: 0.596660, loss_mean: 0.866184, proj_loss: -0.352259, loss_mean_cls: 0.082734, grad_norm: 0.591270 +Steps: 0%| | 1408/1000000 [05:52<68:40:34, 4.04it/s, grad_norm=0.591, loss_final=0.597, loss_mean=0.866, loss_mean_cls=0.0827, proj_loss=-0.352][2026-03-26 12:36:55] Step: 1408, Training Logs: loss_final: 0.606814, loss_mean: 0.864213, proj_loss: -0.340953, loss_mean_cls: 0.083554, grad_norm: 0.578657 +Steps: 0%| | 1409/1000000 [05:52<68:37:56, 4.04it/s, grad_norm=0.579, loss_final=0.607, loss_mean=0.864, loss_mean_cls=0.0836, proj_loss=-0.341][2026-03-26 12:36:56] Step: 1409, Training Logs: loss_final: 0.603288, loss_mean: 0.857342, proj_loss: -0.338909, loss_mean_cls: 0.084854, grad_norm: 0.500837 +Steps: 0%| | 1410/1000000 [05:53<68:38:51, 4.04it/s, grad_norm=0.501, loss_final=0.603, loss_mean=0.857, loss_mean_cls=0.0849, proj_loss=-0.339][2026-03-26 12:36:56] Step: 1410, Training Logs: loss_final: 0.612112, loss_mean: 0.882048, proj_loss: -0.351809, loss_mean_cls: 0.081872, grad_norm: 0.648577 +Steps: 0%| | 1411/1000000 [05:53<68:41:59, 4.04it/s, grad_norm=0.649, loss_final=0.612, loss_mean=0.882, loss_mean_cls=0.0819, proj_loss=-0.352][2026-03-26 12:36:56] Step: 1411, Training Logs: loss_final: 0.612381, loss_mean: 0.867824, proj_loss: -0.341788, loss_mean_cls: 0.086345, grad_norm: 0.631483 +Steps: 0%| | 1412/1000000 [05:53<68:42:13, 4.04it/s, grad_norm=0.631, loss_final=0.612, loss_mean=0.868, loss_mean_cls=0.0863, proj_loss=-0.342][2026-03-26 12:36:56] Step: 1412, Training Logs: loss_final: 0.601486, loss_mean: 0.861037, proj_loss: -0.343300, loss_mean_cls: 0.083749, grad_norm: 0.793462 +Steps: 0%| | 1413/1000000 [05:53<68:41:33, 4.04it/s, grad_norm=0.793, loss_final=0.601, loss_mean=0.861, loss_mean_cls=0.0837, proj_loss=-0.343][2026-03-26 12:36:57] Step: 1413, Training Logs: loss_final: 0.592875, loss_mean: 0.864312, proj_loss: -0.353235, loss_mean_cls: 0.081798, grad_norm: 0.726246 +Steps: 0%| | 1414/1000000 [05:54<70:42:43, 3.92it/s, grad_norm=0.726, loss_final=0.593, loss_mean=0.864, loss_mean_cls=0.0818, proj_loss=-0.353][2026-03-26 12:36:57] Step: 1414, Training Logs: loss_final: 0.614193, loss_mean: 0.874082, proj_loss: -0.342922, loss_mean_cls: 0.083033, grad_norm: 0.406875 +Steps: 0%| | 1415/1000000 [05:54<71:02:07, 3.90it/s, grad_norm=0.407, loss_final=0.614, loss_mean=0.874, loss_mean_cls=0.083, proj_loss=-0.343][2026-03-26 12:36:57] Step: 1415, Training Logs: loss_final: 0.609886, loss_mean: 0.872762, proj_loss: -0.344878, loss_mean_cls: 0.082002, grad_norm: 0.635435 +Steps: 0%| | 1416/1000000 [05:54<70:18:11, 3.95it/s, grad_norm=0.635, loss_final=0.61, loss_mean=0.873, loss_mean_cls=0.082, proj_loss=-0.345][2026-03-26 12:36:57] Step: 1416, Training Logs: loss_final: 0.590497, loss_mean: 0.871378, proj_loss: -0.361595, loss_mean_cls: 0.080713, grad_norm: 0.887107 +Steps: 0%| | 1417/1000000 [05:54<69:51:26, 3.97it/s, grad_norm=0.887, loss_final=0.59, loss_mean=0.871, loss_mean_cls=0.0807, proj_loss=-0.362][2026-03-26 12:36:58] Step: 1417, Training Logs: loss_final: 0.617695, loss_mean: 0.900197, proj_loss: -0.362322, loss_mean_cls: 0.079820, grad_norm: 0.884221 +Steps: 0%| | 1418/1000000 [05:55<69:27:58, 3.99it/s, grad_norm=0.884, loss_final=0.618, loss_mean=0.9, loss_mean_cls=0.0798, proj_loss=-0.362][2026-03-26 12:36:58] Step: 1418, Training Logs: loss_final: 0.598391, loss_mean: 0.869087, proj_loss: -0.353174, loss_mean_cls: 0.082478, grad_norm: 0.432184 +Steps: 0%| | 1419/1000000 [05:55<69:10:54, 4.01it/s, grad_norm=0.432, loss_final=0.598, loss_mean=0.869, loss_mean_cls=0.0825, proj_loss=-0.353][2026-03-26 12:36:58] Step: 1419, Training Logs: loss_final: 0.616437, loss_mean: 0.871236, proj_loss: -0.339851, loss_mean_cls: 0.085052, grad_norm: 1.000776 +Steps: 0%| | 1420/1000000 [05:55<69:00:07, 4.02it/s, grad_norm=1, loss_final=0.616, loss_mean=0.871, loss_mean_cls=0.0851, proj_loss=-0.34][2026-03-26 12:36:58] Step: 1420, Training Logs: loss_final: 0.607141, loss_mean: 0.883505, proj_loss: -0.356440, loss_mean_cls: 0.080075, grad_norm: 1.111762 +Steps: 0%| | 1421/1000000 [05:55<68:53:11, 4.03it/s, grad_norm=1.11, loss_final=0.607, loss_mean=0.884, loss_mean_cls=0.0801, proj_loss=-0.356][2026-03-26 12:36:59] Step: 1421, Training Logs: loss_final: 0.610191, loss_mean: 0.856679, proj_loss: -0.333095, loss_mean_cls: 0.086607, grad_norm: 0.624541 +Steps: 0%| | 1422/1000000 [05:56<68:47:13, 4.03it/s, grad_norm=0.625, loss_final=0.61, loss_mean=0.857, loss_mean_cls=0.0866, proj_loss=-0.333][2026-03-26 12:36:59] Step: 1422, Training Logs: loss_final: 0.608782, loss_mean: 0.870756, proj_loss: -0.345537, loss_mean_cls: 0.083563, grad_norm: 0.557108 +Steps: 0%| | 1423/1000000 [05:56<68:44:49, 4.03it/s, grad_norm=0.557, loss_final=0.609, loss_mean=0.871, loss_mean_cls=0.0836, proj_loss=-0.346][2026-03-26 12:36:59] Step: 1423, Training Logs: loss_final: 0.607638, loss_mean: 0.876142, proj_loss: -0.350931, loss_mean_cls: 0.082427, grad_norm: 0.859382 +Steps: 0%| | 1424/1000000 [05:56<68:41:49, 4.04it/s, grad_norm=0.859, loss_final=0.608, loss_mean=0.876, loss_mean_cls=0.0824, proj_loss=-0.351][2026-03-26 12:36:59] Step: 1424, Training Logs: loss_final: 0.598820, loss_mean: 0.853690, proj_loss: -0.340494, loss_mean_cls: 0.085624, grad_norm: 0.445106 +Steps: 0%| | 1425/1000000 [05:56<68:43:22, 4.04it/s, grad_norm=0.445, loss_final=0.599, loss_mean=0.854, loss_mean_cls=0.0856, proj_loss=-0.34][2026-03-26 12:37:00] Step: 1425, Training Logs: loss_final: 0.604859, loss_mean: 0.865591, proj_loss: -0.345016, loss_mean_cls: 0.084283, grad_norm: 0.654810 +Steps: 0%| | 1426/1000000 [05:57<68:44:03, 4.04it/s, grad_norm=0.655, loss_final=0.605, loss_mean=0.866, loss_mean_cls=0.0843, proj_loss=-0.345][2026-03-26 12:37:00] Step: 1426, Training Logs: loss_final: 0.601984, loss_mean: 0.862633, proj_loss: -0.345561, loss_mean_cls: 0.084912, grad_norm: 0.780032 +Steps: 0%| | 1427/1000000 [05:57<68:42:21, 4.04it/s, grad_norm=0.78, loss_final=0.602, loss_mean=0.863, loss_mean_cls=0.0849, proj_loss=-0.346][2026-03-26 12:37:00] Step: 1427, Training Logs: loss_final: 0.605380, loss_mean: 0.870423, proj_loss: -0.347181, loss_mean_cls: 0.082137, grad_norm: 0.533609 +Steps: 0%| | 1428/1000000 [05:57<68:42:16, 4.04it/s, grad_norm=0.534, loss_final=0.605, loss_mean=0.87, loss_mean_cls=0.0821, proj_loss=-0.347][2026-03-26 12:37:00] Step: 1428, Training Logs: loss_final: 0.612673, loss_mean: 0.889306, proj_loss: -0.357530, loss_mean_cls: 0.080897, grad_norm: 0.556594 +Steps: 0%| | 1429/1000000 [05:57<68:40:38, 4.04it/s, grad_norm=0.557, loss_final=0.613, loss_mean=0.889, loss_mean_cls=0.0809, proj_loss=-0.358][2026-03-26 12:37:01] Step: 1429, Training Logs: loss_final: 0.616945, loss_mean: 0.874966, proj_loss: -0.342297, loss_mean_cls: 0.084276, grad_norm: 0.673459 +Steps: 0%| | 1430/1000000 [05:58<68:40:24, 4.04it/s, grad_norm=0.673, loss_final=0.617, loss_mean=0.875, loss_mean_cls=0.0843, proj_loss=-0.342][2026-03-26 12:37:01] Step: 1430, Training Logs: loss_final: 0.605337, loss_mean: 0.855613, proj_loss: -0.335828, loss_mean_cls: 0.085553, grad_norm: 0.626803 +Steps: 0%| | 1431/1000000 [05:58<68:38:57, 4.04it/s, grad_norm=0.627, loss_final=0.605, loss_mean=0.856, loss_mean_cls=0.0856, proj_loss=-0.336][2026-03-26 12:37:01] Step: 1431, Training Logs: loss_final: 0.627787, loss_mean: 0.903174, proj_loss: -0.355696, loss_mean_cls: 0.080310, grad_norm: 0.442121 +Steps: 0%| | 1432/1000000 [05:58<68:42:16, 4.04it/s, grad_norm=0.442, loss_final=0.628, loss_mean=0.903, loss_mean_cls=0.0803, proj_loss=-0.356][2026-03-26 12:37:01] Step: 1432, Training Logs: loss_final: 0.608337, loss_mean: 0.887087, proj_loss: -0.359441, loss_mean_cls: 0.080691, grad_norm: 0.782745 +Steps: 0%| | 1433/1000000 [05:58<68:41:51, 4.04it/s, grad_norm=0.783, loss_final=0.608, loss_mean=0.887, loss_mean_cls=0.0807, proj_loss=-0.359][2026-03-26 12:37:02] Step: 1433, Training Logs: loss_final: 0.589302, loss_mean: 0.836991, proj_loss: -0.335494, loss_mean_cls: 0.087806, grad_norm: 0.409164 +Steps: 0%| | 1434/1000000 [05:59<68:41:29, 4.04it/s, grad_norm=0.409, loss_final=0.589, loss_mean=0.837, loss_mean_cls=0.0878, proj_loss=-0.335][2026-03-26 12:37:02] Step: 1434, Training Logs: loss_final: 0.607983, loss_mean: 0.864068, proj_loss: -0.341067, loss_mean_cls: 0.084982, grad_norm: 0.715925 +Steps: 0%| | 1435/1000000 [05:59<68:38:28, 4.04it/s, grad_norm=0.716, loss_final=0.608, loss_mean=0.864, loss_mean_cls=0.085, proj_loss=-0.341][2026-03-26 12:37:02] Step: 1435, Training Logs: loss_final: 0.595031, loss_mean: 0.857448, proj_loss: -0.346259, loss_mean_cls: 0.083842, grad_norm: 0.447056 +Steps: 0%| | 1436/1000000 [05:59<68:38:24, 4.04it/s, grad_norm=0.447, loss_final=0.595, loss_mean=0.857, loss_mean_cls=0.0838, proj_loss=-0.346][2026-03-26 12:37:02] Step: 1436, Training Logs: loss_final: 0.610491, loss_mean: 0.864043, proj_loss: -0.338645, loss_mean_cls: 0.085092, grad_norm: 0.653896 +Steps: 0%| | 1437/1000000 [05:59<68:37:28, 4.04it/s, grad_norm=0.654, loss_final=0.61, loss_mean=0.864, loss_mean_cls=0.0851, proj_loss=-0.339][2026-03-26 12:37:03] Step: 1437, Training Logs: loss_final: 0.604656, loss_mean: 0.873970, proj_loss: -0.352542, loss_mean_cls: 0.083228, grad_norm: 0.886509 +Steps: 0%| | 1438/1000000 [06:00<68:39:29, 4.04it/s, grad_norm=0.887, loss_final=0.605, loss_mean=0.874, loss_mean_cls=0.0832, proj_loss=-0.353][2026-03-26 12:37:03] Step: 1438, Training Logs: loss_final: 0.608293, loss_mean: 0.859793, proj_loss: -0.336746, loss_mean_cls: 0.085246, grad_norm: 1.117077 +Steps: 0%| | 1439/1000000 [06:00<68:37:13, 4.04it/s, grad_norm=1.12, loss_final=0.608, loss_mean=0.86, loss_mean_cls=0.0852, proj_loss=-0.337][2026-03-26 12:37:03] Step: 1439, Training Logs: loss_final: 0.592004, loss_mean: 0.853865, proj_loss: -0.345985, loss_mean_cls: 0.084124, grad_norm: 0.771564 +Steps: 0%| | 1440/1000000 [06:00<68:41:14, 4.04it/s, grad_norm=0.772, loss_final=0.592, loss_mean=0.854, loss_mean_cls=0.0841, proj_loss=-0.346][2026-03-26 12:37:03] Step: 1440, Training Logs: loss_final: 0.602567, loss_mean: 0.879939, proj_loss: -0.359224, loss_mean_cls: 0.081852, grad_norm: 1.041979 +Steps: 0%| | 1441/1000000 [06:00<68:39:02, 4.04it/s, grad_norm=1.04, loss_final=0.603, loss_mean=0.88, loss_mean_cls=0.0819, proj_loss=-0.359][2026-03-26 12:37:04] Step: 1441, Training Logs: loss_final: 0.597880, loss_mean: 0.869730, proj_loss: -0.354378, loss_mean_cls: 0.082528, grad_norm: 0.824856 +Steps: 0%| | 1442/1000000 [06:01<68:39:33, 4.04it/s, grad_norm=0.825, loss_final=0.598, loss_mean=0.87, loss_mean_cls=0.0825, proj_loss=-0.354][2026-03-26 12:37:04] Step: 1442, Training Logs: loss_final: 0.613177, loss_mean: 0.878779, proj_loss: -0.348101, loss_mean_cls: 0.082499, grad_norm: 0.502856 +Steps: 0%| | 1443/1000000 [06:01<68:39:08, 4.04it/s, grad_norm=0.503, loss_final=0.613, loss_mean=0.879, loss_mean_cls=0.0825, proj_loss=-0.348][2026-03-26 12:37:04] Step: 1443, Training Logs: loss_final: 0.603896, loss_mean: 0.862160, proj_loss: -0.342248, loss_mean_cls: 0.083984, grad_norm: 0.651760 +Steps: 0%| | 1444/1000000 [06:01<68:47:29, 4.03it/s, grad_norm=0.652, loss_final=0.604, loss_mean=0.862, loss_mean_cls=0.084, proj_loss=-0.342][2026-03-26 12:37:04] Step: 1444, Training Logs: loss_final: 0.616715, loss_mean: 0.880602, proj_loss: -0.347179, loss_mean_cls: 0.083292, grad_norm: 0.595200 +Steps: 0%| | 1445/1000000 [06:01<68:43:21, 4.04it/s, grad_norm=0.595, loss_final=0.617, loss_mean=0.881, loss_mean_cls=0.0833, proj_loss=-0.347][2026-03-26 12:37:05] Step: 1445, Training Logs: loss_final: 0.592707, loss_mean: 0.863350, proj_loss: -0.351578, loss_mean_cls: 0.080935, grad_norm: 0.605978 +Steps: 0%| | 1446/1000000 [06:02<68:42:42, 4.04it/s, grad_norm=0.606, loss_final=0.593, loss_mean=0.863, loss_mean_cls=0.0809, proj_loss=-0.352][2026-03-26 12:37:05] Step: 1446, Training Logs: loss_final: 0.604609, loss_mean: 0.874771, proj_loss: -0.352392, loss_mean_cls: 0.082230, grad_norm: 0.550313 +Steps: 0%| | 1447/1000000 [06:02<68:40:35, 4.04it/s, grad_norm=0.55, loss_final=0.605, loss_mean=0.875, loss_mean_cls=0.0822, proj_loss=-0.352][2026-03-26 12:37:05] Step: 1447, Training Logs: loss_final: 0.606759, loss_mean: 0.869907, proj_loss: -0.346238, loss_mean_cls: 0.083090, grad_norm: 0.551408 +Steps: 0%| | 1448/1000000 [06:02<68:42:35, 4.04it/s, grad_norm=0.551, loss_final=0.607, loss_mean=0.87, loss_mean_cls=0.0831, proj_loss=-0.346][2026-03-26 12:37:05] Step: 1448, Training Logs: loss_final: 0.607526, loss_mean: 0.857506, proj_loss: -0.336188, loss_mean_cls: 0.086208, grad_norm: 0.665012 +Steps: 0%| | 1449/1000000 [06:02<68:40:44, 4.04it/s, grad_norm=0.665, loss_final=0.608, loss_mean=0.858, loss_mean_cls=0.0862, proj_loss=-0.336][2026-03-26 12:37:06] Step: 1449, Training Logs: loss_final: 0.603475, loss_mean: 0.875264, proj_loss: -0.354816, loss_mean_cls: 0.083027, grad_norm: 0.427816 +Steps: 0%| | 1450/1000000 [06:03<68:40:05, 4.04it/s, grad_norm=0.428, loss_final=0.603, loss_mean=0.875, loss_mean_cls=0.083, proj_loss=-0.355][2026-03-26 12:37:06] Step: 1450, Training Logs: loss_final: 0.613814, loss_mean: 0.875714, proj_loss: -0.345511, loss_mean_cls: 0.083611, grad_norm: 0.564322 +Steps: 0%| | 1451/1000000 [06:03<68:40:26, 4.04it/s, grad_norm=0.564, loss_final=0.614, loss_mean=0.876, loss_mean_cls=0.0836, proj_loss=-0.346][2026-03-26 12:37:06] Step: 1451, Training Logs: loss_final: 0.610419, loss_mean: 0.874502, proj_loss: -0.347390, loss_mean_cls: 0.083307, grad_norm: 0.383917 +Steps: 0%| | 1452/1000000 [06:03<68:42:40, 4.04it/s, grad_norm=0.384, loss_final=0.61, loss_mean=0.875, loss_mean_cls=0.0833, proj_loss=-0.347][2026-03-26 12:37:06] Step: 1452, Training Logs: loss_final: 0.608748, loss_mean: 0.878646, proj_loss: -0.352342, loss_mean_cls: 0.082444, grad_norm: 0.580287 +Steps: 0%| | 1453/1000000 [06:03<68:41:18, 4.04it/s, grad_norm=0.58, loss_final=0.609, loss_mean=0.879, loss_mean_cls=0.0824, proj_loss=-0.352][2026-03-26 12:37:07] Step: 1453, Training Logs: loss_final: 0.612183, loss_mean: 0.890967, proj_loss: -0.359434, loss_mean_cls: 0.080650, grad_norm: 0.513324 +Steps: 0%| | 1454/1000000 [06:04<68:41:28, 4.04it/s, grad_norm=0.513, loss_final=0.612, loss_mean=0.891, loss_mean_cls=0.0806, proj_loss=-0.359][2026-03-26 12:37:07] Step: 1454, Training Logs: loss_final: 0.624078, loss_mean: 0.874247, proj_loss: -0.335730, loss_mean_cls: 0.085561, grad_norm: 0.503307 +Steps: 0%| | 1455/1000000 [06:04<68:41:00, 4.04it/s, grad_norm=0.503, loss_final=0.624, loss_mean=0.874, loss_mean_cls=0.0856, proj_loss=-0.336][2026-03-26 12:37:07] Step: 1455, Training Logs: loss_final: 0.616766, loss_mean: 0.891004, proj_loss: -0.355277, loss_mean_cls: 0.081039, grad_norm: 0.501974 +Steps: 0%| | 1456/1000000 [06:04<68:42:48, 4.04it/s, grad_norm=0.502, loss_final=0.617, loss_mean=0.891, loss_mean_cls=0.081, proj_loss=-0.355][2026-03-26 12:37:07] Step: 1456, Training Logs: loss_final: 0.608888, loss_mean: 0.859192, proj_loss: -0.335372, loss_mean_cls: 0.085068, grad_norm: 0.457858 +Steps: 0%| | 1457/1000000 [06:04<68:42:49, 4.04it/s, grad_norm=0.458, loss_final=0.609, loss_mean=0.859, loss_mean_cls=0.0851, proj_loss=-0.335][2026-03-26 12:37:08] Step: 1457, Training Logs: loss_final: 0.589671, loss_mean: 0.847648, proj_loss: -0.342395, loss_mean_cls: 0.084418, grad_norm: 0.549029 +Steps: 0%| | 1458/1000000 [06:04<68:42:44, 4.04it/s, grad_norm=0.549, loss_final=0.59, loss_mean=0.848, loss_mean_cls=0.0844, proj_loss=-0.342][2026-03-26 12:37:08] Step: 1458, Training Logs: loss_final: 0.594795, loss_mean: 0.852547, proj_loss: -0.342586, loss_mean_cls: 0.084834, grad_norm: 0.429392 +Steps: 0%| | 1459/1000000 [06:05<68:40:43, 4.04it/s, grad_norm=0.429, loss_final=0.595, loss_mean=0.853, loss_mean_cls=0.0848, proj_loss=-0.343][2026-03-26 12:37:08] Step: 1459, Training Logs: loss_final: 0.607233, loss_mean: 0.860487, proj_loss: -0.337357, loss_mean_cls: 0.084103, grad_norm: 0.385327 +Steps: 0%| | 1460/1000000 [06:05<68:40:13, 4.04it/s, grad_norm=0.385, loss_final=0.607, loss_mean=0.86, loss_mean_cls=0.0841, proj_loss=-0.337][2026-03-26 12:37:08] Step: 1460, Training Logs: loss_final: 0.589888, loss_mean: 0.853059, proj_loss: -0.347829, loss_mean_cls: 0.084657, grad_norm: 0.356503 +Steps: 0%| | 1461/1000000 [06:05<68:43:38, 4.04it/s, grad_norm=0.357, loss_final=0.59, loss_mean=0.853, loss_mean_cls=0.0847, proj_loss=-0.348][2026-03-26 12:37:09] Step: 1461, Training Logs: loss_final: 0.602386, loss_mean: 0.872624, proj_loss: -0.352929, loss_mean_cls: 0.082690, grad_norm: 0.356098 +Steps: 0%| | 1462/1000000 [06:05<68:41:31, 4.04it/s, grad_norm=0.356, loss_final=0.602, loss_mean=0.873, loss_mean_cls=0.0827, proj_loss=-0.353][2026-03-26 12:37:09] Step: 1462, Training Logs: loss_final: 0.599148, loss_mean: 0.852463, proj_loss: -0.338293, loss_mean_cls: 0.084979, grad_norm: 0.414538 +Steps: 0%| | 1463/1000000 [06:06<68:38:41, 4.04it/s, grad_norm=0.415, loss_final=0.599, loss_mean=0.852, loss_mean_cls=0.085, proj_loss=-0.338][2026-03-26 12:37:09] Step: 1463, Training Logs: loss_final: 0.601392, loss_mean: 0.864587, proj_loss: -0.346687, loss_mean_cls: 0.083492, grad_norm: 0.530387 +Steps: 0%| | 1464/1000000 [06:06<68:41:35, 4.04it/s, grad_norm=0.53, loss_final=0.601, loss_mean=0.865, loss_mean_cls=0.0835, proj_loss=-0.347][2026-03-26 12:37:09] Step: 1464, Training Logs: loss_final: 0.604596, loss_mean: 0.862709, proj_loss: -0.342469, loss_mean_cls: 0.084357, grad_norm: 0.633777 +Steps: 0%| | 1465/1000000 [06:06<68:43:07, 4.04it/s, grad_norm=0.634, loss_final=0.605, loss_mean=0.863, loss_mean_cls=0.0844, proj_loss=-0.342][2026-03-26 12:37:10] Step: 1465, Training Logs: loss_final: 0.595466, loss_mean: 0.845490, proj_loss: -0.334845, loss_mean_cls: 0.084821, grad_norm: 0.492193 +Steps: 0%| | 1466/1000000 [06:06<68:41:37, 4.04it/s, grad_norm=0.492, loss_final=0.595, loss_mean=0.845, loss_mean_cls=0.0848, proj_loss=-0.335][2026-03-26 12:37:10] Step: 1466, Training Logs: loss_final: 0.584625, loss_mean: 0.854318, proj_loss: -0.353299, loss_mean_cls: 0.083605, grad_norm: 0.476461 +Steps: 0%| | 1467/1000000 [06:07<71:08:57, 3.90it/s, grad_norm=0.476, loss_final=0.585, loss_mean=0.854, loss_mean_cls=0.0836, proj_loss=-0.353][2026-03-26 12:37:10] Step: 1467, Training Logs: loss_final: 0.603189, loss_mean: 0.857210, proj_loss: -0.339942, loss_mean_cls: 0.085921, grad_norm: 0.427643 +Steps: 0%| | 1468/1000000 [06:07<70:27:35, 3.94it/s, grad_norm=0.428, loss_final=0.603, loss_mean=0.857, loss_mean_cls=0.0859, proj_loss=-0.34][2026-03-26 12:37:10] Step: 1468, Training Logs: loss_final: 0.595453, loss_mean: 0.861984, proj_loss: -0.347906, loss_mean_cls: 0.081375, grad_norm: 0.489252 +Steps: 0%| | 1469/1000000 [06:07<69:56:50, 3.97it/s, grad_norm=0.489, loss_final=0.595, loss_mean=0.862, loss_mean_cls=0.0814, proj_loss=-0.348][2026-03-26 12:37:11] Step: 1469, Training Logs: loss_final: 0.610196, loss_mean: 0.857227, proj_loss: -0.333536, loss_mean_cls: 0.086504, grad_norm: 0.480260 +Steps: 0%| | 1470/1000000 [06:07<69:32:55, 3.99it/s, grad_norm=0.48, loss_final=0.61, loss_mean=0.857, loss_mean_cls=0.0865, proj_loss=-0.334][2026-03-26 12:37:11] Step: 1470, Training Logs: loss_final: 0.599035, loss_mean: 0.865244, proj_loss: -0.349655, loss_mean_cls: 0.083447, grad_norm: 0.716030 +Steps: 0%| | 1471/1000000 [06:08<69:16:46, 4.00it/s, grad_norm=0.716, loss_final=0.599, loss_mean=0.865, loss_mean_cls=0.0834, proj_loss=-0.35][2026-03-26 12:37:11] Step: 1471, Training Logs: loss_final: 0.609537, loss_mean: 0.867098, proj_loss: -0.342998, loss_mean_cls: 0.085437, grad_norm: 0.545370 +Steps: 0%| | 1472/1000000 [06:08<69:07:46, 4.01it/s, grad_norm=0.545, loss_final=0.61, loss_mean=0.867, loss_mean_cls=0.0854, proj_loss=-0.343][2026-03-26 12:37:11] Step: 1472, Training Logs: loss_final: 0.607913, loss_mean: 0.879087, proj_loss: -0.353981, loss_mean_cls: 0.082807, grad_norm: 0.603782 +Steps: 0%| | 1473/1000000 [06:08<69:03:52, 4.02it/s, grad_norm=0.604, loss_final=0.608, loss_mean=0.879, loss_mean_cls=0.0828, proj_loss=-0.354][2026-03-26 12:37:12] Step: 1473, Training Logs: loss_final: 0.608820, loss_mean: 0.875551, proj_loss: -0.348938, loss_mean_cls: 0.082207, grad_norm: 0.395784 +Steps: 0%| | 1474/1000000 [06:08<68:56:17, 4.02it/s, grad_norm=0.396, loss_final=0.609, loss_mean=0.876, loss_mean_cls=0.0822, proj_loss=-0.349][2026-03-26 12:37:12] Step: 1474, Training Logs: loss_final: 0.615213, loss_mean: 0.877059, proj_loss: -0.344809, loss_mean_cls: 0.082963, grad_norm: 0.731367 +Steps: 0%| | 1475/1000000 [06:09<68:49:09, 4.03it/s, grad_norm=0.731, loss_final=0.615, loss_mean=0.877, loss_mean_cls=0.083, proj_loss=-0.345][2026-03-26 12:37:12] Step: 1475, Training Logs: loss_final: 0.603342, loss_mean: 0.864808, proj_loss: -0.346343, loss_mean_cls: 0.084877, grad_norm: 0.521675 +Steps: 0%| | 1476/1000000 [06:09<68:49:49, 4.03it/s, grad_norm=0.522, loss_final=0.603, loss_mean=0.865, loss_mean_cls=0.0849, proj_loss=-0.346][2026-03-26 12:37:12] Step: 1476, Training Logs: loss_final: 0.613734, loss_mean: 0.881581, proj_loss: -0.350085, loss_mean_cls: 0.082238, grad_norm: 0.628101 +Steps: 0%| | 1477/1000000 [06:09<68:51:05, 4.03it/s, grad_norm=0.628, loss_final=0.614, loss_mean=0.882, loss_mean_cls=0.0822, proj_loss=-0.35][2026-03-26 12:37:13] Step: 1477, Training Logs: loss_final: 0.609121, loss_mean: 0.876663, proj_loss: -0.350682, loss_mean_cls: 0.083140, grad_norm: 0.531492 +Steps: 0%| | 1478/1000000 [06:09<68:46:23, 4.03it/s, grad_norm=0.531, loss_final=0.609, loss_mean=0.877, loss_mean_cls=0.0831, proj_loss=-0.351][2026-03-26 12:37:13] Step: 1478, Training Logs: loss_final: 0.600259, loss_mean: 0.868279, proj_loss: -0.351229, loss_mean_cls: 0.083209, grad_norm: 0.698413 +Steps: 0%| | 1479/1000000 [06:10<68:47:19, 4.03it/s, grad_norm=0.698, loss_final=0.6, loss_mean=0.868, loss_mean_cls=0.0832, proj_loss=-0.351][2026-03-26 12:37:13] Step: 1479, Training Logs: loss_final: 0.597691, loss_mean: 0.852059, proj_loss: -0.339362, loss_mean_cls: 0.084994, grad_norm: 0.462533 +Steps: 0%| | 1480/1000000 [06:10<68:44:37, 4.03it/s, grad_norm=0.463, loss_final=0.598, loss_mean=0.852, loss_mean_cls=0.085, proj_loss=-0.339][2026-03-26 12:37:13] Step: 1480, Training Logs: loss_final: 0.602476, loss_mean: 0.864301, proj_loss: -0.344468, loss_mean_cls: 0.082643, grad_norm: 0.733718 +Steps: 0%| | 1481/1000000 [06:10<68:46:47, 4.03it/s, grad_norm=0.734, loss_final=0.602, loss_mean=0.864, loss_mean_cls=0.0826, proj_loss=-0.344][2026-03-26 12:37:14] Step: 1481, Training Logs: loss_final: 0.608764, loss_mean: 0.859487, proj_loss: -0.336566, loss_mean_cls: 0.085844, grad_norm: 0.589703 +Steps: 0%| | 1482/1000000 [06:10<68:43:12, 4.04it/s, grad_norm=0.59, loss_final=0.609, loss_mean=0.859, loss_mean_cls=0.0858, proj_loss=-0.337][2026-03-26 12:37:14] Step: 1482, Training Logs: loss_final: 0.604574, loss_mean: 0.866910, proj_loss: -0.345902, loss_mean_cls: 0.083567, grad_norm: 0.722778 +Steps: 0%| | 1483/1000000 [06:11<68:41:28, 4.04it/s, grad_norm=0.723, loss_final=0.605, loss_mean=0.867, loss_mean_cls=0.0836, proj_loss=-0.346][2026-03-26 12:37:14] Step: 1483, Training Logs: loss_final: 0.596499, loss_mean: 0.853709, proj_loss: -0.341171, loss_mean_cls: 0.083961, grad_norm: 0.589813 +Steps: 0%| | 1484/1000000 [06:11<68:45:11, 4.03it/s, grad_norm=0.59, loss_final=0.596, loss_mean=0.854, loss_mean_cls=0.084, proj_loss=-0.341][2026-03-26 12:37:14] Step: 1484, Training Logs: loss_final: 0.609962, loss_mean: 0.863481, proj_loss: -0.338515, loss_mean_cls: 0.084996, grad_norm: 0.488095 +Steps: 0%| | 1485/1000000 [06:11<68:46:01, 4.03it/s, grad_norm=0.488, loss_final=0.61, loss_mean=0.863, loss_mean_cls=0.085, proj_loss=-0.339][2026-03-26 12:37:15] Step: 1485, Training Logs: loss_final: 0.603968, loss_mean: 0.849395, proj_loss: -0.331608, loss_mean_cls: 0.086181, grad_norm: 0.409931 +Steps: 0%| | 1486/1000000 [06:11<68:47:07, 4.03it/s, grad_norm=0.41, loss_final=0.604, loss_mean=0.849, loss_mean_cls=0.0862, proj_loss=-0.332][2026-03-26 12:37:15] Step: 1486, Training Logs: loss_final: 0.597771, loss_mean: 0.865898, proj_loss: -0.352501, loss_mean_cls: 0.084374, grad_norm: 0.543089 +Steps: 0%| | 1487/1000000 [06:12<68:44:29, 4.03it/s, grad_norm=0.543, loss_final=0.598, loss_mean=0.866, loss_mean_cls=0.0844, proj_loss=-0.353][2026-03-26 12:37:15] Step: 1487, Training Logs: loss_final: 0.599143, loss_mean: 0.869173, proj_loss: -0.353055, loss_mean_cls: 0.083025, grad_norm: 0.486771 +Steps: 0%| | 1488/1000000 [06:12<68:46:41, 4.03it/s, grad_norm=0.487, loss_final=0.599, loss_mean=0.869, loss_mean_cls=0.083, proj_loss=-0.353][2026-03-26 12:37:15] Step: 1488, Training Logs: loss_final: 0.584830, loss_mean: 0.846179, proj_loss: -0.345200, loss_mean_cls: 0.083851, grad_norm: 0.551605 +Steps: 0%| | 1489/1000000 [06:12<68:44:14, 4.04it/s, grad_norm=0.552, loss_final=0.585, loss_mean=0.846, loss_mean_cls=0.0839, proj_loss=-0.345][2026-03-26 12:37:16] Step: 1489, Training Logs: loss_final: 0.598938, loss_mean: 0.862273, proj_loss: -0.346201, loss_mean_cls: 0.082867, grad_norm: 0.412250 +Steps: 0%| | 1490/1000000 [06:12<68:43:23, 4.04it/s, grad_norm=0.412, loss_final=0.599, loss_mean=0.862, loss_mean_cls=0.0829, proj_loss=-0.346][2026-03-26 12:37:16] Step: 1490, Training Logs: loss_final: 0.600163, loss_mean: 0.884583, proj_loss: -0.364740, loss_mean_cls: 0.080320, grad_norm: 0.409885 +Steps: 0%| | 1491/1000000 [06:13<68:41:25, 4.04it/s, grad_norm=0.41, loss_final=0.6, loss_mean=0.885, loss_mean_cls=0.0803, proj_loss=-0.365][2026-03-26 12:37:16] Step: 1491, Training Logs: loss_final: 0.613672, loss_mean: 0.872230, proj_loss: -0.342606, loss_mean_cls: 0.084049, grad_norm: 0.553484 +Steps: 0%| | 1492/1000000 [06:13<68:44:18, 4.04it/s, grad_norm=0.553, loss_final=0.614, loss_mean=0.872, loss_mean_cls=0.084, proj_loss=-0.343][2026-03-26 12:37:16] Step: 1492, Training Logs: loss_final: 0.606157, loss_mean: 0.866963, proj_loss: -0.344510, loss_mean_cls: 0.083704, grad_norm: 0.460512 +Steps: 0%| | 1493/1000000 [06:13<68:41:13, 4.04it/s, grad_norm=0.461, loss_final=0.606, loss_mean=0.867, loss_mean_cls=0.0837, proj_loss=-0.345][2026-03-26 12:37:17] Step: 1493, Training Logs: loss_final: 0.588698, loss_mean: 0.855585, proj_loss: -0.349599, loss_mean_cls: 0.082712, grad_norm: 0.436609 +Steps: 0%| | 1494/1000000 [06:13<68:43:18, 4.04it/s, grad_norm=0.437, loss_final=0.589, loss_mean=0.856, loss_mean_cls=0.0827, proj_loss=-0.35][2026-03-26 12:37:17] Step: 1494, Training Logs: loss_final: 0.592914, loss_mean: 0.860380, proj_loss: -0.349564, loss_mean_cls: 0.082098, grad_norm: 0.482066 +Steps: 0%| | 1495/1000000 [06:14<68:41:25, 4.04it/s, grad_norm=0.482, loss_final=0.593, loss_mean=0.86, loss_mean_cls=0.0821, proj_loss=-0.35][2026-03-26 12:37:17] Step: 1495, Training Logs: loss_final: 0.606718, loss_mean: 0.874641, proj_loss: -0.351107, loss_mean_cls: 0.083184, grad_norm: 0.533984 +Steps: 0%| | 1496/1000000 [06:14<68:42:53, 4.04it/s, grad_norm=0.534, loss_final=0.607, loss_mean=0.875, loss_mean_cls=0.0832, proj_loss=-0.351][2026-03-26 12:37:17] Step: 1496, Training Logs: loss_final: 0.590998, loss_mean: 0.850419, proj_loss: -0.344373, loss_mean_cls: 0.084952, grad_norm: 0.630650 +Steps: 0%| | 1497/1000000 [06:14<68:40:51, 4.04it/s, grad_norm=0.631, loss_final=0.591, loss_mean=0.85, loss_mean_cls=0.085, proj_loss=-0.344][2026-03-26 12:37:18] Step: 1497, Training Logs: loss_final: 0.595781, loss_mean: 0.862868, proj_loss: -0.350259, loss_mean_cls: 0.083172, grad_norm: 0.399359 +Steps: 0%| | 1498/1000000 [06:14<68:39:39, 4.04it/s, grad_norm=0.399, loss_final=0.596, loss_mean=0.863, loss_mean_cls=0.0832, proj_loss=-0.35][2026-03-26 12:37:18] Step: 1498, Training Logs: loss_final: 0.579561, loss_mean: 0.842911, proj_loss: -0.347067, loss_mean_cls: 0.083717, grad_norm: 0.606149 +Steps: 0%| | 1499/1000000 [06:15<68:39:43, 4.04it/s, grad_norm=0.606, loss_final=0.58, loss_mean=0.843, loss_mean_cls=0.0837, proj_loss=-0.347][2026-03-26 12:37:18] Step: 1499, Training Logs: loss_final: 0.604115, loss_mean: 0.871033, proj_loss: -0.350408, loss_mean_cls: 0.083489, grad_norm: 0.495765 +Steps: 0%| | 1500/1000000 [06:15<69:15:29, 4.00it/s, grad_norm=0.496, loss_final=0.604, loss_mean=0.871, loss_mean_cls=0.0835, proj_loss=-0.35][2026-03-26 12:37:18] Step: 1500, Training Logs: loss_final: 0.596575, loss_mean: 0.854579, proj_loss: -0.343114, loss_mean_cls: 0.085109, grad_norm: 0.374631 +Steps: 0%| | 1501/1000000 [06:15<69:03:35, 4.02it/s, grad_norm=0.375, loss_final=0.597, loss_mean=0.855, loss_mean_cls=0.0851, proj_loss=-0.343][2026-03-26 12:37:19] Step: 1501, Training Logs: loss_final: 0.595700, loss_mean: 0.867005, proj_loss: -0.353929, loss_mean_cls: 0.082624, grad_norm: 0.434570 +Steps: 0%| | 1502/1000000 [06:15<68:57:23, 4.02it/s, grad_norm=0.435, loss_final=0.596, loss_mean=0.867, loss_mean_cls=0.0826, proj_loss=-0.354][2026-03-26 12:37:19] Step: 1502, Training Logs: loss_final: 0.596396, loss_mean: 0.868052, proj_loss: -0.354490, loss_mean_cls: 0.082834, grad_norm: 0.520688 +Steps: 0%| | 1503/1000000 [06:16<68:53:01, 4.03it/s, grad_norm=0.521, loss_final=0.596, loss_mean=0.868, loss_mean_cls=0.0828, proj_loss=-0.354][2026-03-26 12:37:19] Step: 1503, Training Logs: loss_final: 0.602995, loss_mean: 0.865039, proj_loss: -0.347579, loss_mean_cls: 0.085534, grad_norm: 0.345935 +Steps: 0%| | 1504/1000000 [06:16<68:50:34, 4.03it/s, grad_norm=0.346, loss_final=0.603, loss_mean=0.865, loss_mean_cls=0.0855, proj_loss=-0.348][2026-03-26 12:37:19] Step: 1504, Training Logs: loss_final: 0.584311, loss_mean: 0.863925, proj_loss: -0.361511, loss_mean_cls: 0.081897, grad_norm: 0.497337 +Steps: 0%| | 1505/1000000 [06:16<68:52:09, 4.03it/s, grad_norm=0.497, loss_final=0.584, loss_mean=0.864, loss_mean_cls=0.0819, proj_loss=-0.362][2026-03-26 12:37:20] Step: 1505, Training Logs: loss_final: 0.593236, loss_mean: 0.859802, proj_loss: -0.349270, loss_mean_cls: 0.082703, grad_norm: 0.546749 +Steps: 0%| | 1506/1000000 [06:16<69:03:25, 4.02it/s, grad_norm=0.547, loss_final=0.593, loss_mean=0.86, loss_mean_cls=0.0827, proj_loss=-0.349][2026-03-26 12:37:20] Step: 1506, Training Logs: loss_final: 0.602924, loss_mean: 0.861961, proj_loss: -0.344694, loss_mean_cls: 0.085658, grad_norm: 0.591596 +Steps: 0%| | 1507/1000000 [06:17<68:54:58, 4.02it/s, grad_norm=0.592, loss_final=0.603, loss_mean=0.862, loss_mean_cls=0.0857, proj_loss=-0.345][2026-03-26 12:37:20] Step: 1507, Training Logs: loss_final: 0.593454, loss_mean: 0.866226, proj_loss: -0.354449, loss_mean_cls: 0.081677, grad_norm: 0.552665 +Steps: 0%| | 1508/1000000 [06:17<68:52:23, 4.03it/s, grad_norm=0.553, loss_final=0.593, loss_mean=0.866, loss_mean_cls=0.0817, proj_loss=-0.354][2026-03-26 12:37:20] Step: 1508, Training Logs: loss_final: 0.590631, loss_mean: 0.876996, proj_loss: -0.367043, loss_mean_cls: 0.080679, grad_norm: 0.674759 +Steps: 0%| | 1509/1000000 [06:17<68:49:38, 4.03it/s, grad_norm=0.675, loss_final=0.591, loss_mean=0.877, loss_mean_cls=0.0807, proj_loss=-0.367][2026-03-26 12:37:21] Step: 1509, Training Logs: loss_final: 0.613068, loss_mean: 0.886781, proj_loss: -0.354193, loss_mean_cls: 0.080480, grad_norm: 0.970092 +Steps: 0%| | 1510/1000000 [06:17<68:44:18, 4.03it/s, grad_norm=0.97, loss_final=0.613, loss_mean=0.887, loss_mean_cls=0.0805, proj_loss=-0.354][2026-03-26 12:37:21] Step: 1510, Training Logs: loss_final: 0.604272, loss_mean: 0.871722, proj_loss: -0.349803, loss_mean_cls: 0.082353, grad_norm: 0.850378 +Steps: 0%| | 1511/1000000 [06:18<68:43:02, 4.04it/s, grad_norm=0.85, loss_final=0.604, loss_mean=0.872, loss_mean_cls=0.0824, proj_loss=-0.35][2026-03-26 12:37:21] Step: 1511, Training Logs: loss_final: 0.581053, loss_mean: 0.855481, proj_loss: -0.357489, loss_mean_cls: 0.083062, grad_norm: 0.467740 +Steps: 0%| | 1512/1000000 [06:18<68:45:01, 4.03it/s, grad_norm=0.468, loss_final=0.581, loss_mean=0.855, loss_mean_cls=0.0831, proj_loss=-0.357][2026-03-26 12:37:21] Step: 1512, Training Logs: loss_final: 0.606924, loss_mean: 0.855496, proj_loss: -0.334381, loss_mean_cls: 0.085809, grad_norm: 0.604759 +Steps: 0%| | 1513/1000000 [06:18<68:45:57, 4.03it/s, grad_norm=0.605, loss_final=0.607, loss_mean=0.855, loss_mean_cls=0.0858, proj_loss=-0.334][2026-03-26 12:37:22] Step: 1513, Training Logs: loss_final: 0.595961, loss_mean: 0.838338, proj_loss: -0.329756, loss_mean_cls: 0.087378, grad_norm: 0.672255 +Steps: 0%| | 1514/1000000 [06:18<68:43:28, 4.04it/s, grad_norm=0.672, loss_final=0.596, loss_mean=0.838, loss_mean_cls=0.0874, proj_loss=-0.33][2026-03-26 12:37:22] Step: 1514, Training Logs: loss_final: 0.575865, loss_mean: 0.853005, proj_loss: -0.359085, loss_mean_cls: 0.081944, grad_norm: 0.544047 +Steps: 0%| | 1515/1000000 [06:19<68:40:34, 4.04it/s, grad_norm=0.544, loss_final=0.576, loss_mean=0.853, loss_mean_cls=0.0819, proj_loss=-0.359][2026-03-26 12:37:22] Step: 1515, Training Logs: loss_final: 0.591913, loss_mean: 0.849200, proj_loss: -0.341815, loss_mean_cls: 0.084528, grad_norm: 0.632921 +Steps: 0%| | 1516/1000000 [06:19<68:44:16, 4.03it/s, grad_norm=0.633, loss_final=0.592, loss_mean=0.849, loss_mean_cls=0.0845, proj_loss=-0.342][2026-03-26 12:37:22] Step: 1516, Training Logs: loss_final: 0.587306, loss_mean: 0.856194, proj_loss: -0.351485, loss_mean_cls: 0.082597, grad_norm: 0.384241 +Steps: 0%| | 1517/1000000 [06:19<68:46:41, 4.03it/s, grad_norm=0.384, loss_final=0.587, loss_mean=0.856, loss_mean_cls=0.0826, proj_loss=-0.351][2026-03-26 12:37:23] Step: 1517, Training Logs: loss_final: 0.596889, loss_mean: 0.856343, proj_loss: -0.344318, loss_mean_cls: 0.084864, grad_norm: 0.516186 +Steps: 0%| | 1518/1000000 [06:19<68:44:00, 4.04it/s, grad_norm=0.516, loss_final=0.597, loss_mean=0.856, loss_mean_cls=0.0849, proj_loss=-0.344][2026-03-26 12:37:23] Step: 1518, Training Logs: loss_final: 0.604302, loss_mean: 0.863293, proj_loss: -0.343076, loss_mean_cls: 0.084085, grad_norm: 0.646422 +Steps: 0%| | 1519/1000000 [06:20<68:52:09, 4.03it/s, grad_norm=0.646, loss_final=0.604, loss_mean=0.863, loss_mean_cls=0.0841, proj_loss=-0.343][2026-03-26 12:37:23] Step: 1519, Training Logs: loss_final: 0.602014, loss_mean: 0.862623, proj_loss: -0.342975, loss_mean_cls: 0.082366, grad_norm: 0.648160 +Steps: 0%| | 1520/1000000 [06:20<71:25:03, 3.88it/s, grad_norm=0.648, loss_final=0.602, loss_mean=0.863, loss_mean_cls=0.0824, proj_loss=-0.343][2026-03-26 12:37:23] Step: 1520, Training Logs: loss_final: 0.590130, loss_mean: 0.849567, proj_loss: -0.344614, loss_mean_cls: 0.085176, grad_norm: 0.410080 +Steps: 0%| | 1521/1000000 [06:20<70:38:07, 3.93it/s, grad_norm=0.41, loss_final=0.59, loss_mean=0.85, loss_mean_cls=0.0852, proj_loss=-0.345][2026-03-26 12:37:24] Step: 1521, Training Logs: loss_final: 0.583741, loss_mean: 0.842043, proj_loss: -0.343245, loss_mean_cls: 0.084943, grad_norm: 0.471740 +Steps: 0%| | 1522/1000000 [06:20<70:06:22, 3.96it/s, grad_norm=0.472, loss_final=0.584, loss_mean=0.842, loss_mean_cls=0.0849, proj_loss=-0.343][2026-03-26 12:37:24] Step: 1522, Training Logs: loss_final: 0.597166, loss_mean: 0.842546, proj_loss: -0.332005, loss_mean_cls: 0.086625, grad_norm: 0.484481 +Steps: 0%| | 1523/1000000 [06:21<69:40:45, 3.98it/s, grad_norm=0.484, loss_final=0.597, loss_mean=0.843, loss_mean_cls=0.0866, proj_loss=-0.332][2026-03-26 12:37:24] Step: 1523, Training Logs: loss_final: 0.619568, loss_mean: 0.895429, proj_loss: -0.357160, loss_mean_cls: 0.081300, grad_norm: 0.494590 +Steps: 0%| | 1524/1000000 [06:21<69:23:29, 4.00it/s, grad_norm=0.495, loss_final=0.62, loss_mean=0.895, loss_mean_cls=0.0813, proj_loss=-0.357][2026-03-26 12:37:24] Step: 1524, Training Logs: loss_final: 0.592570, loss_mean: 0.849526, proj_loss: -0.343241, loss_mean_cls: 0.086285, grad_norm: 0.509505 +Steps: 0%| | 1525/1000000 [06:21<69:07:47, 4.01it/s, grad_norm=0.51, loss_final=0.593, loss_mean=0.85, loss_mean_cls=0.0863, proj_loss=-0.343][2026-03-26 12:37:25] Step: 1525, Training Logs: loss_final: 0.602763, loss_mean: 0.870654, proj_loss: -0.349101, loss_mean_cls: 0.081210, grad_norm: 0.442339 +Steps: 0%| | 1526/1000000 [06:21<69:04:22, 4.02it/s, grad_norm=0.442, loss_final=0.603, loss_mean=0.871, loss_mean_cls=0.0812, proj_loss=-0.349][2026-03-26 12:37:25] Step: 1526, Training Logs: loss_final: 0.581302, loss_mean: 0.855232, proj_loss: -0.355774, loss_mean_cls: 0.081844, grad_norm: 0.361251 +Steps: 0%| | 1527/1000000 [06:22<68:58:10, 4.02it/s, grad_norm=0.361, loss_final=0.581, loss_mean=0.855, loss_mean_cls=0.0818, proj_loss=-0.356][2026-03-26 12:37:25] Step: 1527, Training Logs: loss_final: 0.589933, loss_mean: 0.852760, proj_loss: -0.346498, loss_mean_cls: 0.083671, grad_norm: 0.316264 +Steps: 0%| | 1528/1000000 [06:22<68:52:16, 4.03it/s, grad_norm=0.316, loss_final=0.59, loss_mean=0.853, loss_mean_cls=0.0837, proj_loss=-0.346][2026-03-26 12:37:25] Step: 1528, Training Logs: loss_final: 0.575906, loss_mean: 0.865477, proj_loss: -0.369115, loss_mean_cls: 0.079544, grad_norm: 0.362903 +Steps: 0%| | 1529/1000000 [06:22<68:49:14, 4.03it/s, grad_norm=0.363, loss_final=0.576, loss_mean=0.865, loss_mean_cls=0.0795, proj_loss=-0.369][2026-03-26 12:37:26] Step: 1529, Training Logs: loss_final: 0.605733, loss_mean: 0.875880, proj_loss: -0.353445, loss_mean_cls: 0.083299, grad_norm: 0.392460 +Steps: 0%| | 1530/1000000 [06:22<68:46:45, 4.03it/s, grad_norm=0.392, loss_final=0.606, loss_mean=0.876, loss_mean_cls=0.0833, proj_loss=-0.353][2026-03-26 12:37:26] Step: 1530, Training Logs: loss_final: 0.587850, loss_mean: 0.853417, proj_loss: -0.349201, loss_mean_cls: 0.083635, grad_norm: 0.447516 +Steps: 0%| | 1531/1000000 [06:23<68:45:08, 4.03it/s, grad_norm=0.448, loss_final=0.588, loss_mean=0.853, loss_mean_cls=0.0836, proj_loss=-0.349][2026-03-26 12:37:26] Step: 1531, Training Logs: loss_final: 0.603313, loss_mean: 0.864410, proj_loss: -0.344524, loss_mean_cls: 0.083426, grad_norm: 0.415401 +Steps: 0%| | 1532/1000000 [06:23<68:42:59, 4.04it/s, grad_norm=0.415, loss_final=0.603, loss_mean=0.864, loss_mean_cls=0.0834, proj_loss=-0.345][2026-03-26 12:37:26] Step: 1532, Training Logs: loss_final: 0.584253, loss_mean: 0.856964, proj_loss: -0.355739, loss_mean_cls: 0.083028, grad_norm: 0.641485 +Steps: 0%| | 1533/1000000 [06:23<68:42:48, 4.04it/s, grad_norm=0.641, loss_final=0.584, loss_mean=0.857, loss_mean_cls=0.083, proj_loss=-0.356][2026-03-26 12:37:27] Step: 1533, Training Logs: loss_final: 0.598078, loss_mean: 0.849941, proj_loss: -0.337027, loss_mean_cls: 0.085163, grad_norm: 0.807033 +Steps: 0%| | 1534/1000000 [06:23<68:41:12, 4.04it/s, grad_norm=0.807, loss_final=0.598, loss_mean=0.85, loss_mean_cls=0.0852, proj_loss=-0.337][2026-03-26 12:37:27] Step: 1534, Training Logs: loss_final: 0.589850, loss_mean: 0.840981, proj_loss: -0.336296, loss_mean_cls: 0.085165, grad_norm: 0.510561 +Steps: 0%| | 1535/1000000 [06:24<68:45:08, 4.03it/s, grad_norm=0.511, loss_final=0.59, loss_mean=0.841, loss_mean_cls=0.0852, proj_loss=-0.336][2026-03-26 12:37:27] Step: 1535, Training Logs: loss_final: 0.586475, loss_mean: 0.830548, proj_loss: -0.331932, loss_mean_cls: 0.087858, grad_norm: 0.598248 +Steps: 0%| | 1536/1000000 [06:24<68:42:55, 4.04it/s, grad_norm=0.598, loss_final=0.586, loss_mean=0.831, loss_mean_cls=0.0879, proj_loss=-0.332][2026-03-26 12:37:27] Step: 1536, Training Logs: loss_final: 0.589303, loss_mean: 0.853122, proj_loss: -0.348146, loss_mean_cls: 0.084327, grad_norm: 0.771123 +Steps: 0%| | 1537/1000000 [06:24<68:43:14, 4.04it/s, grad_norm=0.771, loss_final=0.589, loss_mean=0.853, loss_mean_cls=0.0843, proj_loss=-0.348][2026-03-26 12:37:28] Step: 1537, Training Logs: loss_final: 0.588829, loss_mean: 0.867291, proj_loss: -0.360160, loss_mean_cls: 0.081698, grad_norm: 0.502459 +Steps: 0%| | 1538/1000000 [06:24<68:41:40, 4.04it/s, grad_norm=0.502, loss_final=0.589, loss_mean=0.867, loss_mean_cls=0.0817, proj_loss=-0.36][2026-03-26 12:37:28] Step: 1538, Training Logs: loss_final: 0.596492, loss_mean: 0.869045, proj_loss: -0.355409, loss_mean_cls: 0.082856, grad_norm: 0.537421 +Steps: 0%| | 1539/1000000 [06:25<68:38:44, 4.04it/s, grad_norm=0.537, loss_final=0.596, loss_mean=0.869, loss_mean_cls=0.0829, proj_loss=-0.355][2026-03-26 12:37:28] Step: 1539, Training Logs: loss_final: 0.584505, loss_mean: 0.847969, proj_loss: -0.348317, loss_mean_cls: 0.084854, grad_norm: 0.414102 +Steps: 0%| | 1540/1000000 [06:25<68:36:35, 4.04it/s, grad_norm=0.414, loss_final=0.585, loss_mean=0.848, loss_mean_cls=0.0849, proj_loss=-0.348][2026-03-26 12:37:28] Step: 1540, Training Logs: loss_final: 0.605367, loss_mean: 0.879375, proj_loss: -0.354767, loss_mean_cls: 0.080759, grad_norm: 0.924357 +Steps: 0%| | 1541/1000000 [06:25<68:36:56, 4.04it/s, grad_norm=0.924, loss_final=0.605, loss_mean=0.879, loss_mean_cls=0.0808, proj_loss=-0.355][2026-03-26 12:37:29] Step: 1541, Training Logs: loss_final: 0.609799, loss_mean: 0.883507, proj_loss: -0.354932, loss_mean_cls: 0.081225, grad_norm: 0.540262 +Steps: 0%| | 1542/1000000 [06:25<68:35:39, 4.04it/s, grad_norm=0.54, loss_final=0.61, loss_mean=0.884, loss_mean_cls=0.0812, proj_loss=-0.355][2026-03-26 12:37:29] Step: 1542, Training Logs: loss_final: 0.595806, loss_mean: 0.851144, proj_loss: -0.340609, loss_mean_cls: 0.085272, grad_norm: 0.992423 +Steps: 0%| | 1543/1000000 [06:26<68:37:39, 4.04it/s, grad_norm=0.992, loss_final=0.596, loss_mean=0.851, loss_mean_cls=0.0853, proj_loss=-0.341][2026-03-26 12:37:29] Step: 1543, Training Logs: loss_final: 0.603756, loss_mean: 0.873681, proj_loss: -0.353501, loss_mean_cls: 0.083576, grad_norm: 0.485043 +Steps: 0%| | 1544/1000000 [06:26<68:41:40, 4.04it/s, grad_norm=0.485, loss_final=0.604, loss_mean=0.874, loss_mean_cls=0.0836, proj_loss=-0.354][2026-03-26 12:37:29] Step: 1544, Training Logs: loss_final: 0.593565, loss_mean: 0.853183, proj_loss: -0.344059, loss_mean_cls: 0.084441, grad_norm: 0.645554 +Steps: 0%| | 1545/1000000 [06:26<68:38:44, 4.04it/s, grad_norm=0.646, loss_final=0.594, loss_mean=0.853, loss_mean_cls=0.0844, proj_loss=-0.344][2026-03-26 12:37:30] Step: 1545, Training Logs: loss_final: 0.603347, loss_mean: 0.869021, proj_loss: -0.348335, loss_mean_cls: 0.082661, grad_norm: 0.952404 +Steps: 0%| | 1546/1000000 [06:26<69:40:40, 3.98it/s, grad_norm=0.952, loss_final=0.603, loss_mean=0.869, loss_mean_cls=0.0827, proj_loss=-0.348][2026-03-26 12:37:30] Step: 1546, Training Logs: loss_final: 0.599382, loss_mean: 0.868533, proj_loss: -0.351505, loss_mean_cls: 0.082354, grad_norm: 0.712793 +Steps: 0%| | 1547/1000000 [06:27<69:19:24, 4.00it/s, grad_norm=0.713, loss_final=0.599, loss_mean=0.869, loss_mean_cls=0.0824, proj_loss=-0.352][2026-03-26 12:37:30] Step: 1547, Training Logs: loss_final: 0.598746, loss_mean: 0.854676, proj_loss: -0.340402, loss_mean_cls: 0.084472, grad_norm: 0.561949 +Steps: 0%| | 1548/1000000 [06:27<69:09:06, 4.01it/s, grad_norm=0.562, loss_final=0.599, loss_mean=0.855, loss_mean_cls=0.0845, proj_loss=-0.34][2026-03-26 12:37:30] Step: 1548, Training Logs: loss_final: 0.600505, loss_mean: 0.864245, proj_loss: -0.346469, loss_mean_cls: 0.082729, grad_norm: 0.591914 +Steps: 0%| | 1549/1000000 [06:27<68:58:17, 4.02it/s, grad_norm=0.592, loss_final=0.601, loss_mean=0.864, loss_mean_cls=0.0827, proj_loss=-0.346][2026-03-26 12:37:31] Step: 1549, Training Logs: loss_final: 0.578873, loss_mean: 0.836176, proj_loss: -0.343550, loss_mean_cls: 0.086247, grad_norm: 0.441614 +Steps: 0%| | 1550/1000000 [06:27<68:53:52, 4.03it/s, grad_norm=0.442, loss_final=0.579, loss_mean=0.836, loss_mean_cls=0.0862, proj_loss=-0.344][2026-03-26 12:37:31] Step: 1550, Training Logs: loss_final: 0.614489, loss_mean: 0.878744, proj_loss: -0.346814, loss_mean_cls: 0.082558, grad_norm: 0.614902 +Steps: 0%| | 1551/1000000 [06:28<68:54:41, 4.02it/s, grad_norm=0.615, loss_final=0.614, loss_mean=0.879, loss_mean_cls=0.0826, proj_loss=-0.347][2026-03-26 12:37:31] Step: 1551, Training Logs: loss_final: 0.620456, loss_mean: 0.872489, proj_loss: -0.336329, loss_mean_cls: 0.084296, grad_norm: 0.869025 +Steps: 0%| | 1552/1000000 [06:28<68:49:41, 4.03it/s, grad_norm=0.869, loss_final=0.62, loss_mean=0.872, loss_mean_cls=0.0843, proj_loss=-0.336][2026-03-26 12:37:31] Step: 1552, Training Logs: loss_final: 0.597004, loss_mean: 0.855947, proj_loss: -0.343943, loss_mean_cls: 0.084999, grad_norm: 0.536745 +Steps: 0%| | 1553/1000000 [06:28<68:53:09, 4.03it/s, grad_norm=0.537, loss_final=0.597, loss_mean=0.856, loss_mean_cls=0.085, proj_loss=-0.344][2026-03-26 12:37:32] Step: 1553, Training Logs: loss_final: 0.592502, loss_mean: 0.869879, proj_loss: -0.358647, loss_mean_cls: 0.081270, grad_norm: 0.744709 +Steps: 0%| | 1554/1000000 [06:28<68:44:46, 4.03it/s, grad_norm=0.745, loss_final=0.593, loss_mean=0.87, loss_mean_cls=0.0813, proj_loss=-0.359][2026-03-26 12:37:32] Step: 1554, Training Logs: loss_final: 0.593438, loss_mean: 0.862378, proj_loss: -0.352036, loss_mean_cls: 0.083095, grad_norm: 0.787013 +Steps: 0%| | 1555/1000000 [06:29<68:44:11, 4.03it/s, grad_norm=0.787, loss_final=0.593, loss_mean=0.862, loss_mean_cls=0.0831, proj_loss=-0.352][2026-03-26 12:37:32] Step: 1555, Training Logs: loss_final: 0.599922, loss_mean: 0.860471, proj_loss: -0.345460, loss_mean_cls: 0.084912, grad_norm: 0.455573 +Steps: 0%| | 1556/1000000 [06:29<68:46:59, 4.03it/s, grad_norm=0.456, loss_final=0.6, loss_mean=0.86, loss_mean_cls=0.0849, proj_loss=-0.345][2026-03-26 12:37:32] Step: 1556, Training Logs: loss_final: 0.608025, loss_mean: 0.888389, proj_loss: -0.359569, loss_mean_cls: 0.079206, grad_norm: 0.634972 +Steps: 0%| | 1557/1000000 [06:29<68:44:58, 4.03it/s, grad_norm=0.635, loss_final=0.608, loss_mean=0.888, loss_mean_cls=0.0792, proj_loss=-0.36][2026-03-26 12:37:32] Step: 1557, Training Logs: loss_final: 0.603482, loss_mean: 0.858333, proj_loss: -0.338748, loss_mean_cls: 0.083896, grad_norm: 0.880808 +Steps: 0%| | 1558/1000000 [06:29<68:44:24, 4.03it/s, grad_norm=0.881, loss_final=0.603, loss_mean=0.858, loss_mean_cls=0.0839, proj_loss=-0.339][2026-03-26 12:37:33] Step: 1558, Training Logs: loss_final: 0.603083, loss_mean: 0.866231, proj_loss: -0.346779, loss_mean_cls: 0.083631, grad_norm: 0.381530 +Steps: 0%| | 1559/1000000 [06:30<68:40:14, 4.04it/s, grad_norm=0.382, loss_final=0.603, loss_mean=0.866, loss_mean_cls=0.0836, proj_loss=-0.347][2026-03-26 12:37:33] Step: 1559, Training Logs: loss_final: 0.601054, loss_mean: 0.857351, proj_loss: -0.341360, loss_mean_cls: 0.085063, grad_norm: 0.856395 +Steps: 0%| | 1560/1000000 [06:30<68:42:16, 4.04it/s, grad_norm=0.856, loss_final=0.601, loss_mean=0.857, loss_mean_cls=0.0851, proj_loss=-0.341][2026-03-26 12:37:33] Step: 1560, Training Logs: loss_final: 0.598676, loss_mean: 0.862386, proj_loss: -0.348289, loss_mean_cls: 0.084579, grad_norm: 0.762277 +Steps: 0%| | 1561/1000000 [06:30<68:44:52, 4.03it/s, grad_norm=0.762, loss_final=0.599, loss_mean=0.862, loss_mean_cls=0.0846, proj_loss=-0.348][2026-03-26 12:37:33] Step: 1561, Training Logs: loss_final: 0.594480, loss_mean: 0.857944, proj_loss: -0.346954, loss_mean_cls: 0.083490, grad_norm: 0.619587 +Steps: 0%| | 1562/1000000 [06:30<68:44:30, 4.03it/s, grad_norm=0.62, loss_final=0.594, loss_mean=0.858, loss_mean_cls=0.0835, proj_loss=-0.347][2026-03-26 12:37:34] Step: 1562, Training Logs: loss_final: 0.606114, loss_mean: 0.891903, proj_loss: -0.365824, loss_mean_cls: 0.080035, grad_norm: 0.864933 +Steps: 0%| | 1563/1000000 [06:31<68:41:02, 4.04it/s, grad_norm=0.865, loss_final=0.606, loss_mean=0.892, loss_mean_cls=0.08, proj_loss=-0.366][2026-03-26 12:37:34] Step: 1563, Training Logs: loss_final: 0.599545, loss_mean: 0.865570, proj_loss: -0.349022, loss_mean_cls: 0.082997, grad_norm: 0.897992 +Steps: 0%| | 1564/1000000 [06:31<68:40:53, 4.04it/s, grad_norm=0.898, loss_final=0.6, loss_mean=0.866, loss_mean_cls=0.083, proj_loss=-0.349][2026-03-26 12:37:34] Step: 1564, Training Logs: loss_final: 0.585683, loss_mean: 0.846005, proj_loss: -0.344620, loss_mean_cls: 0.084298, grad_norm: 0.453597 +Steps: 0%| | 1565/1000000 [06:31<68:41:32, 4.04it/s, grad_norm=0.454, loss_final=0.586, loss_mean=0.846, loss_mean_cls=0.0843, proj_loss=-0.345][2026-03-26 12:37:34] Step: 1565, Training Logs: loss_final: 0.607528, loss_mean: 0.879724, proj_loss: -0.354026, loss_mean_cls: 0.081829, grad_norm: 0.716081 +Steps: 0%| | 1566/1000000 [06:31<68:43:43, 4.04it/s, grad_norm=0.716, loss_final=0.608, loss_mean=0.88, loss_mean_cls=0.0818, proj_loss=-0.354][2026-03-26 12:37:35] Step: 1566, Training Logs: loss_final: 0.609236, loss_mean: 0.876684, proj_loss: -0.350111, loss_mean_cls: 0.082662, grad_norm: 0.513455 +Steps: 0%| | 1567/1000000 [06:32<68:46:25, 4.03it/s, grad_norm=0.513, loss_final=0.609, loss_mean=0.877, loss_mean_cls=0.0827, proj_loss=-0.35][2026-03-26 12:37:35] Step: 1567, Training Logs: loss_final: 0.572828, loss_mean: 0.845825, proj_loss: -0.354340, loss_mean_cls: 0.081342, grad_norm: 0.623992 +Steps: 0%| | 1568/1000000 [06:32<68:47:57, 4.03it/s, grad_norm=0.624, loss_final=0.573, loss_mean=0.846, loss_mean_cls=0.0813, proj_loss=-0.354][2026-03-26 12:37:35] Step: 1568, Training Logs: loss_final: 0.609677, loss_mean: 0.881588, proj_loss: -0.353682, loss_mean_cls: 0.081771, grad_norm: 0.949404 +Steps: 0%| | 1569/1000000 [06:32<68:46:38, 4.03it/s, grad_norm=0.949, loss_final=0.61, loss_mean=0.882, loss_mean_cls=0.0818, proj_loss=-0.354][2026-03-26 12:37:35] Step: 1569, Training Logs: loss_final: 0.598962, loss_mean: 0.878673, proj_loss: -0.360306, loss_mean_cls: 0.080595, grad_norm: 0.478674 +Steps: 0%| | 1570/1000000 [06:32<68:48:46, 4.03it/s, grad_norm=0.479, loss_final=0.599, loss_mean=0.879, loss_mean_cls=0.0806, proj_loss=-0.36][2026-03-26 12:37:36] Step: 1570, Training Logs: loss_final: 0.590446, loss_mean: 0.864612, proj_loss: -0.356886, loss_mean_cls: 0.082720, grad_norm: 0.706343 +Steps: 0%| | 1571/1000000 [06:33<68:44:55, 4.03it/s, grad_norm=0.706, loss_final=0.59, loss_mean=0.865, loss_mean_cls=0.0827, proj_loss=-0.357][2026-03-26 12:37:36] Step: 1571, Training Logs: loss_final: 0.595601, loss_mean: 0.873553, proj_loss: -0.358129, loss_mean_cls: 0.080177, grad_norm: 0.395614 +Steps: 0%| | 1572/1000000 [06:33<68:47:14, 4.03it/s, grad_norm=0.396, loss_final=0.596, loss_mean=0.874, loss_mean_cls=0.0802, proj_loss=-0.358][2026-03-26 12:37:36] Step: 1572, Training Logs: loss_final: 0.602116, loss_mean: 0.850100, proj_loss: -0.334851, loss_mean_cls: 0.086867, grad_norm: 0.585154 +Steps: 0%| | 1573/1000000 [06:33<68:47:48, 4.03it/s, grad_norm=0.585, loss_final=0.602, loss_mean=0.85, loss_mean_cls=0.0869, proj_loss=-0.335][2026-03-26 12:37:36] Step: 1573, Training Logs: loss_final: 0.593982, loss_mean: 0.855530, proj_loss: -0.345446, loss_mean_cls: 0.083898, grad_norm: 0.800711 +Steps: 0%| | 1574/1000000 [06:33<68:45:15, 4.03it/s, grad_norm=0.801, loss_final=0.594, loss_mean=0.856, loss_mean_cls=0.0839, proj_loss=-0.345][2026-03-26 12:37:37] Step: 1574, Training Logs: loss_final: 0.598598, loss_mean: 0.872668, proj_loss: -0.354937, loss_mean_cls: 0.080867, grad_norm: 0.587712 +Steps: 0%| | 1575/1000000 [06:34<68:47:50, 4.03it/s, grad_norm=0.588, loss_final=0.599, loss_mean=0.873, loss_mean_cls=0.0809, proj_loss=-0.355][2026-03-26 12:37:37] Step: 1575, Training Logs: loss_final: 0.606877, loss_mean: 0.863258, proj_loss: -0.341172, loss_mean_cls: 0.084791, grad_norm: 0.611826 +Steps: 0%| | 1576/1000000 [06:34<68:47:26, 4.03it/s, grad_norm=0.612, loss_final=0.607, loss_mean=0.863, loss_mean_cls=0.0848, proj_loss=-0.341][2026-03-26 12:37:37] Step: 1576, Training Logs: loss_final: 0.582992, loss_mean: 0.861097, proj_loss: -0.360564, loss_mean_cls: 0.082459, grad_norm: 0.957700 +Steps: 0%| | 1577/1000000 [06:34<68:47:20, 4.03it/s, grad_norm=0.958, loss_final=0.583, loss_mean=0.861, loss_mean_cls=0.0825, proj_loss=-0.361][2026-03-26 12:37:37] Step: 1577, Training Logs: loss_final: 0.614281, loss_mean: 0.876574, proj_loss: -0.345243, loss_mean_cls: 0.082950, grad_norm: 0.872394 +Steps: 0%| | 1578/1000000 [06:34<68:48:07, 4.03it/s, grad_norm=0.872, loss_final=0.614, loss_mean=0.877, loss_mean_cls=0.083, proj_loss=-0.345][2026-03-26 12:37:38] Step: 1578, Training Logs: loss_final: 0.597134, loss_mean: 0.855457, proj_loss: -0.342788, loss_mean_cls: 0.084466, grad_norm: 0.546715 +Steps: 0%| | 1579/1000000 [06:35<68:43:44, 4.04it/s, grad_norm=0.547, loss_final=0.597, loss_mean=0.855, loss_mean_cls=0.0845, proj_loss=-0.343][2026-03-26 12:37:38] Step: 1579, Training Logs: loss_final: 0.593353, loss_mean: 0.857989, proj_loss: -0.349208, loss_mean_cls: 0.084571, grad_norm: 1.040653 +Steps: 0%| | 1580/1000000 [06:35<68:44:55, 4.03it/s, grad_norm=1.04, loss_final=0.593, loss_mean=0.858, loss_mean_cls=0.0846, proj_loss=-0.349][2026-03-26 12:37:38] Step: 1580, Training Logs: loss_final: 0.589583, loss_mean: 0.866708, proj_loss: -0.358214, loss_mean_cls: 0.081089, grad_norm: 0.837652 +Steps: 0%| | 1581/1000000 [06:35<68:46:42, 4.03it/s, grad_norm=0.838, loss_final=0.59, loss_mean=0.867, loss_mean_cls=0.0811, proj_loss=-0.358][2026-03-26 12:37:38] Step: 1581, Training Logs: loss_final: 0.595079, loss_mean: 0.853640, proj_loss: -0.344472, loss_mean_cls: 0.085912, grad_norm: 0.504574 +Steps: 0%| | 1582/1000000 [06:35<68:46:06, 4.03it/s, grad_norm=0.505, loss_final=0.595, loss_mean=0.854, loss_mean_cls=0.0859, proj_loss=-0.344][2026-03-26 12:37:39] Step: 1582, Training Logs: loss_final: 0.596863, loss_mean: 0.879281, proj_loss: -0.362389, loss_mean_cls: 0.079971, grad_norm: 0.619369 +Steps: 0%| | 1583/1000000 [06:36<68:45:46, 4.03it/s, grad_norm=0.619, loss_final=0.597, loss_mean=0.879, loss_mean_cls=0.08, proj_loss=-0.362][2026-03-26 12:37:39] Step: 1583, Training Logs: loss_final: 0.588632, loss_mean: 0.865571, proj_loss: -0.357947, loss_mean_cls: 0.081009, grad_norm: 0.442263 +Steps: 0%| | 1584/1000000 [06:36<68:44:32, 4.03it/s, grad_norm=0.442, loss_final=0.589, loss_mean=0.866, loss_mean_cls=0.081, proj_loss=-0.358][2026-03-26 12:37:39] Step: 1584, Training Logs: loss_final: 0.602090, loss_mean: 0.847819, proj_loss: -0.331770, loss_mean_cls: 0.086041, grad_norm: 0.723137 +Steps: 0%| | 1585/1000000 [06:36<68:45:44, 4.03it/s, grad_norm=0.723, loss_final=0.602, loss_mean=0.848, loss_mean_cls=0.086, proj_loss=-0.332][2026-03-26 12:37:39] Step: 1585, Training Logs: loss_final: 0.583897, loss_mean: 0.864086, proj_loss: -0.362405, loss_mean_cls: 0.082215, grad_norm: 0.541667 +Steps: 0%| | 1586/1000000 [06:36<68:41:48, 4.04it/s, grad_norm=0.542, loss_final=0.584, loss_mean=0.864, loss_mean_cls=0.0822, proj_loss=-0.362][2026-03-26 12:37:40] Step: 1586, Training Logs: loss_final: 0.597844, loss_mean: 0.878780, proj_loss: -0.361008, loss_mean_cls: 0.080073, grad_norm: 0.446809 +Steps: 0%| | 1587/1000000 [06:37<68:42:45, 4.04it/s, grad_norm=0.447, loss_final=0.598, loss_mean=0.879, loss_mean_cls=0.0801, proj_loss=-0.361][2026-03-26 12:37:40] Step: 1587, Training Logs: loss_final: 0.588439, loss_mean: 0.842493, proj_loss: -0.338820, loss_mean_cls: 0.084766, grad_norm: 0.627940 +Steps: 0%| | 1588/1000000 [06:37<68:42:42, 4.04it/s, grad_norm=0.628, loss_final=0.588, loss_mean=0.842, loss_mean_cls=0.0848, proj_loss=-0.339][2026-03-26 12:37:40] Step: 1588, Training Logs: loss_final: 0.609758, loss_mean: 0.882790, proj_loss: -0.355795, loss_mean_cls: 0.082763, grad_norm: 0.508703 +Steps: 0%| | 1589/1000000 [06:37<68:45:04, 4.03it/s, grad_norm=0.509, loss_final=0.61, loss_mean=0.883, loss_mean_cls=0.0828, proj_loss=-0.356][2026-03-26 12:37:40] Step: 1589, Training Logs: loss_final: 0.599332, loss_mean: 0.876588, proj_loss: -0.358604, loss_mean_cls: 0.081348, grad_norm: 0.567102 +Steps: 0%| | 1590/1000000 [06:37<68:43:30, 4.04it/s, grad_norm=0.567, loss_final=0.599, loss_mean=0.877, loss_mean_cls=0.0813, proj_loss=-0.359][2026-03-26 12:37:41] Step: 1590, Training Logs: loss_final: 0.596065, loss_mean: 0.857173, proj_loss: -0.346163, loss_mean_cls: 0.085054, grad_norm: 0.804655 +Steps: 0%| | 1591/1000000 [06:38<68:44:45, 4.03it/s, grad_norm=0.805, loss_final=0.596, loss_mean=0.857, loss_mean_cls=0.0851, proj_loss=-0.346][2026-03-26 12:37:41] Step: 1591, Training Logs: loss_final: 0.583174, loss_mean: 0.840975, proj_loss: -0.343081, loss_mean_cls: 0.085280, grad_norm: 0.436278 +Steps: 0%| | 1592/1000000 [06:38<68:43:57, 4.03it/s, grad_norm=0.436, loss_final=0.583, loss_mean=0.841, loss_mean_cls=0.0853, proj_loss=-0.343][2026-03-26 12:37:41] Step: 1592, Training Logs: loss_final: 0.594492, loss_mean: 0.860981, proj_loss: -0.348919, loss_mean_cls: 0.082430, grad_norm: 0.413784 +Steps: 0%| | 1593/1000000 [06:38<68:42:47, 4.04it/s, grad_norm=0.414, loss_final=0.594, loss_mean=0.861, loss_mean_cls=0.0824, proj_loss=-0.349][2026-03-26 12:37:41] Step: 1593, Training Logs: loss_final: 0.612364, loss_mean: 0.871557, proj_loss: -0.344213, loss_mean_cls: 0.085020, grad_norm: 0.574771 +Steps: 0%| | 1594/1000000 [06:38<68:44:18, 4.03it/s, grad_norm=0.575, loss_final=0.612, loss_mean=0.872, loss_mean_cls=0.085, proj_loss=-0.344][2026-03-26 12:37:42] Step: 1594, Training Logs: loss_final: 0.593569, loss_mean: 0.859172, proj_loss: -0.349537, loss_mean_cls: 0.083934, grad_norm: 0.508339 +Steps: 0%| | 1595/1000000 [06:39<68:44:22, 4.03it/s, grad_norm=0.508, loss_final=0.594, loss_mean=0.859, loss_mean_cls=0.0839, proj_loss=-0.35][2026-03-26 12:37:42] Step: 1595, Training Logs: loss_final: 0.608309, loss_mean: 0.877858, proj_loss: -0.351895, loss_mean_cls: 0.082346, grad_norm: 0.512603 +Steps: 0%| | 1596/1000000 [06:39<69:09:51, 4.01it/s, grad_norm=0.513, loss_final=0.608, loss_mean=0.878, loss_mean_cls=0.0823, proj_loss=-0.352][2026-03-26 12:37:42] Step: 1596, Training Logs: loss_final: 0.587030, loss_mean: 0.858439, proj_loss: -0.354208, loss_mean_cls: 0.082799, grad_norm: 0.608253 +Steps: 0%| | 1597/1000000 [06:39<69:01:49, 4.02it/s, grad_norm=0.608, loss_final=0.587, loss_mean=0.858, loss_mean_cls=0.0828, proj_loss=-0.354][2026-03-26 12:37:42] Step: 1597, Training Logs: loss_final: 0.603019, loss_mean: 0.868822, proj_loss: -0.349602, loss_mean_cls: 0.083799, grad_norm: 0.581635 +Steps: 0%| | 1598/1000000 [06:39<69:01:34, 4.02it/s, grad_norm=0.582, loss_final=0.603, loss_mean=0.869, loss_mean_cls=0.0838, proj_loss=-0.35][2026-03-26 12:37:43] Step: 1598, Training Logs: loss_final: 0.596243, loss_mean: 0.861260, proj_loss: -0.349073, loss_mean_cls: 0.084056, grad_norm: 0.470966 +Steps: 0%| | 1599/1000000 [06:40<68:53:35, 4.03it/s, grad_norm=0.471, loss_final=0.596, loss_mean=0.861, loss_mean_cls=0.0841, proj_loss=-0.349][2026-03-26 12:37:43] Step: 1599, Training Logs: loss_final: 0.595942, loss_mean: 0.867377, proj_loss: -0.353683, loss_mean_cls: 0.082247, grad_norm: 0.483338 +Steps: 0%| | 1600/1000000 [06:40<68:52:19, 4.03it/s, grad_norm=0.483, loss_final=0.596, loss_mean=0.867, loss_mean_cls=0.0822, proj_loss=-0.354][2026-03-26 12:37:43] Step: 1600, Training Logs: loss_final: 0.578516, loss_mean: 0.847097, proj_loss: -0.352376, loss_mean_cls: 0.083795, grad_norm: 0.496581 +Steps: 0%| | 1601/1000000 [06:40<68:51:28, 4.03it/s, grad_norm=0.497, loss_final=0.579, loss_mean=0.847, loss_mean_cls=0.0838, proj_loss=-0.352][2026-03-26 12:37:43] Step: 1601, Training Logs: loss_final: 0.578904, loss_mean: 0.855162, proj_loss: -0.359498, loss_mean_cls: 0.083240, grad_norm: 0.630129 +Steps: 0%| | 1602/1000000 [06:40<68:47:28, 4.03it/s, grad_norm=0.63, loss_final=0.579, loss_mean=0.855, loss_mean_cls=0.0832, proj_loss=-0.359][2026-03-26 12:37:44] Step: 1602, Training Logs: loss_final: 0.597148, loss_mean: 0.868021, proj_loss: -0.351719, loss_mean_cls: 0.080845, grad_norm: 0.352065 +Steps: 0%| | 1603/1000000 [06:41<68:44:21, 4.03it/s, grad_norm=0.352, loss_final=0.597, loss_mean=0.868, loss_mean_cls=0.0808, proj_loss=-0.352][2026-03-26 12:37:44] Step: 1603, Training Logs: loss_final: 0.591540, loss_mean: 0.852023, proj_loss: -0.346194, loss_mean_cls: 0.085711, grad_norm: 0.474017 +Steps: 0%| | 1604/1000000 [06:41<68:44:54, 4.03it/s, grad_norm=0.474, loss_final=0.592, loss_mean=0.852, loss_mean_cls=0.0857, proj_loss=-0.346][2026-03-26 12:37:44] Step: 1604, Training Logs: loss_final: 0.599014, loss_mean: 0.875053, proj_loss: -0.356724, loss_mean_cls: 0.080686, grad_norm: 0.670126 +Steps: 0%| | 1605/1000000 [06:41<68:45:09, 4.03it/s, grad_norm=0.67, loss_final=0.599, loss_mean=0.875, loss_mean_cls=0.0807, proj_loss=-0.357][2026-03-26 12:37:44] Step: 1605, Training Logs: loss_final: 0.601518, loss_mean: 0.875352, proj_loss: -0.355882, loss_mean_cls: 0.082049, grad_norm: 0.589878 +Steps: 0%| | 1606/1000000 [06:41<68:48:20, 4.03it/s, grad_norm=0.59, loss_final=0.602, loss_mean=0.875, loss_mean_cls=0.082, proj_loss=-0.356][2026-03-26 12:37:45] Step: 1606, Training Logs: loss_final: 0.604556, loss_mean: 0.886832, proj_loss: -0.362489, loss_mean_cls: 0.080214, grad_norm: 0.423548 +Steps: 0%| | 1607/1000000 [06:42<68:44:22, 4.03it/s, grad_norm=0.424, loss_final=0.605, loss_mean=0.887, loss_mean_cls=0.0802, proj_loss=-0.362][2026-03-26 12:37:45] Step: 1607, Training Logs: loss_final: 0.588982, loss_mean: 0.877741, proj_loss: -0.367162, loss_mean_cls: 0.078404, grad_norm: 0.462092 +Steps: 0%| | 1608/1000000 [06:42<68:44:16, 4.03it/s, grad_norm=0.462, loss_final=0.589, loss_mean=0.878, loss_mean_cls=0.0784, proj_loss=-0.367][2026-03-26 12:37:45] Step: 1608, Training Logs: loss_final: 0.589164, loss_mean: 0.866968, proj_loss: -0.359639, loss_mean_cls: 0.081835, grad_norm: 0.502293 +Steps: 0%| | 1609/1000000 [06:42<68:42:52, 4.04it/s, grad_norm=0.502, loss_final=0.589, loss_mean=0.867, loss_mean_cls=0.0818, proj_loss=-0.36][2026-03-26 12:37:45] Step: 1609, Training Logs: loss_final: 0.590434, loss_mean: 0.856882, proj_loss: -0.351249, loss_mean_cls: 0.084802, grad_norm: 0.393784 +Steps: 0%| | 1610/1000000 [06:42<68:43:03, 4.04it/s, grad_norm=0.394, loss_final=0.59, loss_mean=0.857, loss_mean_cls=0.0848, proj_loss=-0.351][2026-03-26 12:37:46] Step: 1610, Training Logs: loss_final: 0.584441, loss_mean: 0.849032, proj_loss: -0.348200, loss_mean_cls: 0.083609, grad_norm: 0.592998 +Steps: 0%| | 1611/1000000 [06:42<68:39:45, 4.04it/s, grad_norm=0.593, loss_final=0.584, loss_mean=0.849, loss_mean_cls=0.0836, proj_loss=-0.348][2026-03-26 12:37:46] Step: 1611, Training Logs: loss_final: 0.589870, loss_mean: 0.867563, proj_loss: -0.360378, loss_mean_cls: 0.082685, grad_norm: 0.578429 +Steps: 0%| | 1612/1000000 [06:43<68:41:09, 4.04it/s, grad_norm=0.578, loss_final=0.59, loss_mean=0.868, loss_mean_cls=0.0827, proj_loss=-0.36][2026-03-26 12:37:46] Step: 1612, Training Logs: loss_final: 0.574190, loss_mean: 0.835909, proj_loss: -0.345963, loss_mean_cls: 0.084244, grad_norm: 0.369920 +Steps: 0%| | 1613/1000000 [06:43<68:37:16, 4.04it/s, grad_norm=0.37, loss_final=0.574, loss_mean=0.836, loss_mean_cls=0.0842, proj_loss=-0.346][2026-03-26 12:37:46] Step: 1613, Training Logs: loss_final: 0.593287, loss_mean: 0.859394, proj_loss: -0.350505, loss_mean_cls: 0.084398, grad_norm: 0.426337 +Steps: 0%| | 1614/1000000 [06:43<68:38:14, 4.04it/s, grad_norm=0.426, loss_final=0.593, loss_mean=0.859, loss_mean_cls=0.0844, proj_loss=-0.351][2026-03-26 12:37:47] Step: 1614, Training Logs: loss_final: 0.605819, loss_mean: 0.880020, proj_loss: -0.355962, loss_mean_cls: 0.081761, grad_norm: 0.495235 +Steps: 0%| | 1615/1000000 [06:43<68:36:19, 4.04it/s, grad_norm=0.495, loss_final=0.606, loss_mean=0.88, loss_mean_cls=0.0818, proj_loss=-0.356][2026-03-26 12:37:47] Step: 1615, Training Logs: loss_final: 0.588846, loss_mean: 0.873887, proj_loss: -0.364868, loss_mean_cls: 0.079826, grad_norm: 0.398514 +Steps: 0%| | 1616/1000000 [06:44<68:38:52, 4.04it/s, grad_norm=0.399, loss_final=0.589, loss_mean=0.874, loss_mean_cls=0.0798, proj_loss=-0.365][2026-03-26 12:37:47] Step: 1616, Training Logs: loss_final: 0.590628, loss_mean: 0.853909, proj_loss: -0.347056, loss_mean_cls: 0.083776, grad_norm: 0.525901 +Steps: 0%| | 1617/1000000 [06:44<68:39:28, 4.04it/s, grad_norm=0.526, loss_final=0.591, loss_mean=0.854, loss_mean_cls=0.0838, proj_loss=-0.347][2026-03-26 12:37:47] Step: 1617, Training Logs: loss_final: 0.590288, loss_mean: 0.854311, proj_loss: -0.347321, loss_mean_cls: 0.083298, grad_norm: 0.666548 +Steps: 0%| | 1618/1000000 [06:44<68:41:02, 4.04it/s, grad_norm=0.667, loss_final=0.59, loss_mean=0.854, loss_mean_cls=0.0833, proj_loss=-0.347][2026-03-26 12:37:48] Step: 1618, Training Logs: loss_final: 0.591624, loss_mean: 0.870146, proj_loss: -0.359203, loss_mean_cls: 0.080681, grad_norm: 0.533651 +Steps: 0%| | 1619/1000000 [06:44<68:38:33, 4.04it/s, grad_norm=0.534, loss_final=0.592, loss_mean=0.87, loss_mean_cls=0.0807, proj_loss=-0.359][2026-03-26 12:37:48] Step: 1619, Training Logs: loss_final: 0.586333, loss_mean: 0.853667, proj_loss: -0.350829, loss_mean_cls: 0.083495, grad_norm: 0.635968 +Steps: 0%| | 1620/1000000 [06:45<68:46:46, 4.03it/s, grad_norm=0.636, loss_final=0.586, loss_mean=0.854, loss_mean_cls=0.0835, proj_loss=-0.351][2026-03-26 12:37:48] Step: 1620, Training Logs: loss_final: 0.591937, loss_mean: 0.844325, proj_loss: -0.338937, loss_mean_cls: 0.086549, grad_norm: 0.489211 +Steps: 0%| | 1621/1000000 [06:45<68:45:07, 4.03it/s, grad_norm=0.489, loss_final=0.592, loss_mean=0.844, loss_mean_cls=0.0865, proj_loss=-0.339][2026-03-26 12:37:48] Step: 1621, Training Logs: loss_final: 0.589305, loss_mean: 0.849892, proj_loss: -0.344117, loss_mean_cls: 0.083529, grad_norm: 0.688505 +Steps: 0%| | 1622/1000000 [06:45<68:43:14, 4.04it/s, grad_norm=0.689, loss_final=0.589, loss_mean=0.85, loss_mean_cls=0.0835, proj_loss=-0.344][2026-03-26 12:37:49] Step: 1622, Training Logs: loss_final: 0.576669, loss_mean: 0.850823, proj_loss: -0.357610, loss_mean_cls: 0.083456, grad_norm: 0.619058 +Steps: 0%| | 1623/1000000 [06:45<68:41:26, 4.04it/s, grad_norm=0.619, loss_final=0.577, loss_mean=0.851, loss_mean_cls=0.0835, proj_loss=-0.358][2026-03-26 12:37:49] Step: 1623, Training Logs: loss_final: 0.596766, loss_mean: 0.857872, proj_loss: -0.345182, loss_mean_cls: 0.084076, grad_norm: 0.529222 +Steps: 0%| | 1624/1000000 [06:46<68:41:33, 4.04it/s, grad_norm=0.529, loss_final=0.597, loss_mean=0.858, loss_mean_cls=0.0841, proj_loss=-0.345][2026-03-26 12:37:49] Step: 1624, Training Logs: loss_final: 0.608205, loss_mean: 0.884666, proj_loss: -0.358400, loss_mean_cls: 0.081939, grad_norm: 0.811426 +Steps: 0%| | 1625/1000000 [06:46<68:39:32, 4.04it/s, grad_norm=0.811, loss_final=0.608, loss_mean=0.885, loss_mean_cls=0.0819, proj_loss=-0.358][2026-03-26 12:37:49] Step: 1625, Training Logs: loss_final: 0.598804, loss_mean: 0.871826, proj_loss: -0.355511, loss_mean_cls: 0.082490, grad_norm: 0.518072 +Steps: 0%| | 1626/1000000 [06:46<68:40:16, 4.04it/s, grad_norm=0.518, loss_final=0.599, loss_mean=0.872, loss_mean_cls=0.0825, proj_loss=-0.356][2026-03-26 12:37:50] Step: 1626, Training Logs: loss_final: 0.597613, loss_mean: 0.852659, proj_loss: -0.339937, loss_mean_cls: 0.084892, grad_norm: 0.536563 +Steps: 0%| | 1627/1000000 [06:46<68:41:46, 4.04it/s, grad_norm=0.537, loss_final=0.598, loss_mean=0.853, loss_mean_cls=0.0849, proj_loss=-0.34][2026-03-26 12:37:50] Step: 1627, Training Logs: loss_final: 0.590872, loss_mean: 0.833788, proj_loss: -0.330411, loss_mean_cls: 0.087494, grad_norm: 0.796424 +Steps: 0%| | 1628/1000000 [06:47<68:45:24, 4.03it/s, grad_norm=0.796, loss_final=0.591, loss_mean=0.834, loss_mean_cls=0.0875, proj_loss=-0.33][2026-03-26 12:37:50] Step: 1628, Training Logs: loss_final: 0.597724, loss_mean: 0.861012, proj_loss: -0.347707, loss_mean_cls: 0.084419, grad_norm: 0.518404 +Steps: 0%| | 1629/1000000 [06:47<68:46:42, 4.03it/s, grad_norm=0.518, loss_final=0.598, loss_mean=0.861, loss_mean_cls=0.0844, proj_loss=-0.348][2026-03-26 12:37:50] Step: 1629, Training Logs: loss_final: 0.591773, loss_mean: 0.842866, proj_loss: -0.337473, loss_mean_cls: 0.086379, grad_norm: 0.779221 +Steps: 0%| | 1630/1000000 [06:47<68:42:35, 4.04it/s, grad_norm=0.779, loss_final=0.592, loss_mean=0.843, loss_mean_cls=0.0864, proj_loss=-0.337][2026-03-26 12:37:51] Step: 1630, Training Logs: loss_final: 0.590703, loss_mean: 0.859015, proj_loss: -0.351926, loss_mean_cls: 0.083614, grad_norm: 0.540169 +Steps: 0%| | 1631/1000000 [06:47<68:41:20, 4.04it/s, grad_norm=0.54, loss_final=0.591, loss_mean=0.859, loss_mean_cls=0.0836, proj_loss=-0.352][2026-03-26 12:37:51] Step: 1631, Training Logs: loss_final: 0.616444, loss_mean: 0.909585, proj_loss: -0.371708, loss_mean_cls: 0.078567, grad_norm: 0.440225 +Steps: 0%| | 1632/1000000 [06:48<68:41:04, 4.04it/s, grad_norm=0.44, loss_final=0.616, loss_mean=0.91, loss_mean_cls=0.0786, proj_loss=-0.372][2026-03-26 12:37:51] Step: 1632, Training Logs: loss_final: 0.597938, loss_mean: 0.875648, proj_loss: -0.358305, loss_mean_cls: 0.080595, grad_norm: 0.631411 +Steps: 0%| | 1633/1000000 [06:48<68:40:55, 4.04it/s, grad_norm=0.631, loss_final=0.598, loss_mean=0.876, loss_mean_cls=0.0806, proj_loss=-0.358][2026-03-26 12:37:51] Step: 1633, Training Logs: loss_final: 0.607464, loss_mean: 0.879013, proj_loss: -0.353745, loss_mean_cls: 0.082196, grad_norm: 0.352363 +Steps: 0%| | 1634/1000000 [06:48<68:40:48, 4.04it/s, grad_norm=0.352, loss_final=0.607, loss_mean=0.879, loss_mean_cls=0.0822, proj_loss=-0.354][2026-03-26 12:37:52] Step: 1634, Training Logs: loss_final: 0.601842, loss_mean: 0.883563, proj_loss: -0.361974, loss_mean_cls: 0.080253, grad_norm: 0.552051 +Steps: 0%| | 1635/1000000 [06:48<68:44:42, 4.03it/s, grad_norm=0.552, loss_final=0.602, loss_mean=0.884, loss_mean_cls=0.0803, proj_loss=-0.362][2026-03-26 12:37:52] Step: 1635, Training Logs: loss_final: 0.597671, loss_mean: 0.871645, proj_loss: -0.355734, loss_mean_cls: 0.081761, grad_norm: 0.466346 +Steps: 0%| | 1636/1000000 [06:49<68:39:04, 4.04it/s, grad_norm=0.466, loss_final=0.598, loss_mean=0.872, loss_mean_cls=0.0818, proj_loss=-0.356][2026-03-26 12:37:52] Step: 1636, Training Logs: loss_final: 0.585388, loss_mean: 0.845777, proj_loss: -0.345596, loss_mean_cls: 0.085207, grad_norm: 0.416422 +Steps: 0%| | 1637/1000000 [06:49<68:41:14, 4.04it/s, grad_norm=0.416, loss_final=0.585, loss_mean=0.846, loss_mean_cls=0.0852, proj_loss=-0.346][2026-03-26 12:37:52] Step: 1637, Training Logs: loss_final: 0.587145, loss_mean: 0.864046, proj_loss: -0.358342, loss_mean_cls: 0.081441, grad_norm: 0.549152 +Steps: 0%| | 1638/1000000 [06:49<68:39:58, 4.04it/s, grad_norm=0.549, loss_final=0.587, loss_mean=0.864, loss_mean_cls=0.0814, proj_loss=-0.358][2026-03-26 12:37:53] Step: 1638, Training Logs: loss_final: 0.582186, loss_mean: 0.839478, proj_loss: -0.341956, loss_mean_cls: 0.084664, grad_norm: 0.562367 +Steps: 0%| | 1639/1000000 [06:49<68:37:16, 4.04it/s, grad_norm=0.562, loss_final=0.582, loss_mean=0.839, loss_mean_cls=0.0847, proj_loss=-0.342][2026-03-26 12:37:53] Step: 1639, Training Logs: loss_final: 0.581694, loss_mean: 0.850899, proj_loss: -0.351887, loss_mean_cls: 0.082681, grad_norm: 0.666643 +Steps: 0%| | 1640/1000000 [06:50<68:42:54, 4.04it/s, grad_norm=0.667, loss_final=0.582, loss_mean=0.851, loss_mean_cls=0.0827, proj_loss=-0.352][2026-03-26 12:37:53] Step: 1640, Training Logs: loss_final: 0.579042, loss_mean: 0.846590, proj_loss: -0.351378, loss_mean_cls: 0.083831, grad_norm: 0.353443 +Steps: 0%| | 1641/1000000 [06:50<68:41:38, 4.04it/s, grad_norm=0.353, loss_final=0.579, loss_mean=0.847, loss_mean_cls=0.0838, proj_loss=-0.351][2026-03-26 12:37:53] Step: 1641, Training Logs: loss_final: 0.592582, loss_mean: 0.868512, proj_loss: -0.357646, loss_mean_cls: 0.081716, grad_norm: 0.511178 +Steps: 0%| | 1642/1000000 [06:50<68:42:36, 4.04it/s, grad_norm=0.511, loss_final=0.593, loss_mean=0.869, loss_mean_cls=0.0817, proj_loss=-0.358][2026-03-26 12:37:54] Step: 1642, Training Logs: loss_final: 0.581915, loss_mean: 0.849999, proj_loss: -0.351776, loss_mean_cls: 0.083692, grad_norm: 0.611467 +Steps: 0%| | 1643/1000000 [06:50<68:41:56, 4.04it/s, grad_norm=0.611, loss_final=0.582, loss_mean=0.85, loss_mean_cls=0.0837, proj_loss=-0.352][2026-03-26 12:37:54] Step: 1643, Training Logs: loss_final: 0.597338, loss_mean: 0.870147, proj_loss: -0.354128, loss_mean_cls: 0.081319, grad_norm: 0.388673 +Steps: 0%| | 1644/1000000 [06:51<68:42:13, 4.04it/s, grad_norm=0.389, loss_final=0.597, loss_mean=0.87, loss_mean_cls=0.0813, proj_loss=-0.354][2026-03-26 12:37:54] Step: 1644, Training Logs: loss_final: 0.588111, loss_mean: 0.867572, proj_loss: -0.360893, loss_mean_cls: 0.081433, grad_norm: 0.529821 +Steps: 0%| | 1645/1000000 [06:51<68:42:22, 4.04it/s, grad_norm=0.53, loss_final=0.588, loss_mean=0.868, loss_mean_cls=0.0814, proj_loss=-0.361][2026-03-26 12:37:54] Step: 1645, Training Logs: loss_final: 0.598297, loss_mean: 0.847023, proj_loss: -0.335665, loss_mean_cls: 0.086940, grad_norm: 0.959569 +Steps: 0%| | 1646/1000000 [06:51<68:43:21, 4.04it/s, grad_norm=0.96, loss_final=0.598, loss_mean=0.847, loss_mean_cls=0.0869, proj_loss=-0.336][2026-03-26 12:37:55] Step: 1646, Training Logs: loss_final: 0.593746, loss_mean: 0.886441, proj_loss: -0.372550, loss_mean_cls: 0.079855, grad_norm: 0.736647 +Steps: 0%| | 1647/1000000 [06:51<68:45:32, 4.03it/s, grad_norm=0.737, loss_final=0.594, loss_mean=0.886, loss_mean_cls=0.0799, proj_loss=-0.373][2026-03-26 12:37:55] Step: 1647, Training Logs: loss_final: 0.586568, loss_mean: 0.845116, proj_loss: -0.343402, loss_mean_cls: 0.084854, grad_norm: 0.647815 +Steps: 0%| | 1648/1000000 [06:52<68:48:16, 4.03it/s, grad_norm=0.648, loss_final=0.587, loss_mean=0.845, loss_mean_cls=0.0849, proj_loss=-0.343][2026-03-26 12:37:55] Step: 1648, Training Logs: loss_final: 0.587810, loss_mean: 0.846912, proj_loss: -0.343942, loss_mean_cls: 0.084840, grad_norm: 0.603763 +Steps: 0%| | 1649/1000000 [06:52<68:47:04, 4.03it/s, grad_norm=0.604, loss_final=0.588, loss_mean=0.847, loss_mean_cls=0.0848, proj_loss=-0.344][2026-03-26 12:37:55] Step: 1649, Training Logs: loss_final: 0.588134, loss_mean: 0.875712, proj_loss: -0.368327, loss_mean_cls: 0.080749, grad_norm: 0.788809 +Steps: 0%| | 1650/1000000 [06:52<68:46:58, 4.03it/s, grad_norm=0.789, loss_final=0.588, loss_mean=0.876, loss_mean_cls=0.0807, proj_loss=-0.368][2026-03-26 12:37:56] Step: 1650, Training Logs: loss_final: 0.582158, loss_mean: 0.850330, proj_loss: -0.351537, loss_mean_cls: 0.083365, grad_norm: 0.331393 +Steps: 0%| | 1651/1000000 [06:52<68:45:01, 4.03it/s, grad_norm=0.331, loss_final=0.582, loss_mean=0.85, loss_mean_cls=0.0834, proj_loss=-0.352][2026-03-26 12:37:56] Step: 1651, Training Logs: loss_final: 0.600039, loss_mean: 0.874092, proj_loss: -0.355540, loss_mean_cls: 0.081487, grad_norm: 0.742243 +Steps: 0%| | 1652/1000000 [06:53<68:44:58, 4.03it/s, grad_norm=0.742, loss_final=0.6, loss_mean=0.874, loss_mean_cls=0.0815, proj_loss=-0.356][2026-03-26 12:37:56] Step: 1652, Training Logs: loss_final: 0.590085, loss_mean: 0.845128, proj_loss: -0.341181, loss_mean_cls: 0.086138, grad_norm: 0.538975 +Steps: 0%| | 1653/1000000 [06:53<68:41:56, 4.04it/s, grad_norm=0.539, loss_final=0.59, loss_mean=0.845, loss_mean_cls=0.0861, proj_loss=-0.341][2026-03-26 12:37:56] Step: 1653, Training Logs: loss_final: 0.580380, loss_mean: 0.854407, proj_loss: -0.355766, loss_mean_cls: 0.081739, grad_norm: 0.427341 +Steps: 0%| | 1654/1000000 [06:53<68:42:30, 4.04it/s, grad_norm=0.427, loss_final=0.58, loss_mean=0.854, loss_mean_cls=0.0817, proj_loss=-0.356][2026-03-26 12:37:57] Step: 1654, Training Logs: loss_final: 0.594034, loss_mean: 0.865657, proj_loss: -0.354506, loss_mean_cls: 0.082883, grad_norm: 0.728035 +Steps: 0%| | 1655/1000000 [06:53<68:41:16, 4.04it/s, grad_norm=0.728, loss_final=0.594, loss_mean=0.866, loss_mean_cls=0.0829, proj_loss=-0.355][2026-03-26 12:37:57] Step: 1655, Training Logs: loss_final: 0.584008, loss_mean: 0.846565, proj_loss: -0.346599, loss_mean_cls: 0.084042, grad_norm: 0.592159 +Steps: 0%| | 1656/1000000 [06:54<68:45:54, 4.03it/s, grad_norm=0.592, loss_final=0.584, loss_mean=0.847, loss_mean_cls=0.084, proj_loss=-0.347][2026-03-26 12:37:57] Step: 1656, Training Logs: loss_final: 0.571042, loss_mean: 0.847789, proj_loss: -0.360813, loss_mean_cls: 0.084066, grad_norm: 0.685829 +Steps: 0%| | 1657/1000000 [06:54<68:41:50, 4.04it/s, grad_norm=0.686, loss_final=0.571, loss_mean=0.848, loss_mean_cls=0.0841, proj_loss=-0.361][2026-03-26 12:37:57] Step: 1657, Training Logs: loss_final: 0.585870, loss_mean: 0.838048, proj_loss: -0.339160, loss_mean_cls: 0.086983, grad_norm: 0.614078 +Steps: 0%| | 1658/1000000 [06:54<68:40:55, 4.04it/s, grad_norm=0.614, loss_final=0.586, loss_mean=0.838, loss_mean_cls=0.087, proj_loss=-0.339][2026-03-26 12:37:58] Step: 1658, Training Logs: loss_final: 0.597786, loss_mean: 0.870070, proj_loss: -0.353919, loss_mean_cls: 0.081635, grad_norm: 0.456322 +Steps: 0%| | 1659/1000000 [06:54<68:40:01, 4.04it/s, grad_norm=0.456, loss_final=0.598, loss_mean=0.87, loss_mean_cls=0.0816, proj_loss=-0.354][2026-03-26 12:37:58] Step: 1659, Training Logs: loss_final: 0.599839, loss_mean: 0.868664, proj_loss: -0.351998, loss_mean_cls: 0.083173, grad_norm: 1.015422 +Steps: 0%| | 1660/1000000 [06:55<68:41:31, 4.04it/s, grad_norm=1.02, loss_final=0.6, loss_mean=0.869, loss_mean_cls=0.0832, proj_loss=-0.352][2026-03-26 12:37:58] Step: 1660, Training Logs: loss_final: 0.580764, loss_mean: 0.846877, proj_loss: -0.350293, loss_mean_cls: 0.084180, grad_norm: 0.792975 +Steps: 0%| | 1661/1000000 [06:55<68:38:07, 4.04it/s, grad_norm=0.793, loss_final=0.581, loss_mean=0.847, loss_mean_cls=0.0842, proj_loss=-0.35][2026-03-26 12:37:58] Step: 1661, Training Logs: loss_final: 0.603712, loss_mean: 0.896004, proj_loss: -0.370527, loss_mean_cls: 0.078235, grad_norm: 0.587733 +Steps: 0%| | 1662/1000000 [06:55<68:39:21, 4.04it/s, grad_norm=0.588, loss_final=0.604, loss_mean=0.896, loss_mean_cls=0.0782, proj_loss=-0.371][2026-03-26 12:37:59] Step: 1662, Training Logs: loss_final: 0.590263, loss_mean: 0.862249, proj_loss: -0.354211, loss_mean_cls: 0.082226, grad_norm: 0.853680 +Steps: 0%| | 1663/1000000 [06:55<68:38:40, 4.04it/s, grad_norm=0.854, loss_final=0.59, loss_mean=0.862, loss_mean_cls=0.0822, proj_loss=-0.354][2026-03-26 12:37:59] Step: 1663, Training Logs: loss_final: 0.604393, loss_mean: 0.878214, proj_loss: -0.356191, loss_mean_cls: 0.082370, grad_norm: 0.551749 +Steps: 0%| | 1664/1000000 [06:56<68:42:04, 4.04it/s, grad_norm=0.552, loss_final=0.604, loss_mean=0.878, loss_mean_cls=0.0824, proj_loss=-0.356][2026-03-26 12:37:59] Step: 1664, Training Logs: loss_final: 0.593699, loss_mean: 0.870155, proj_loss: -0.358031, loss_mean_cls: 0.081575, grad_norm: 0.450970 +Steps: 0%| | 1665/1000000 [06:56<68:39:21, 4.04it/s, grad_norm=0.451, loss_final=0.594, loss_mean=0.87, loss_mean_cls=0.0816, proj_loss=-0.358][2026-03-26 12:37:59] Step: 1665, Training Logs: loss_final: 0.581052, loss_mean: 0.854883, proj_loss: -0.356167, loss_mean_cls: 0.082337, grad_norm: 0.535671 +Steps: 0%| | 1666/1000000 [06:56<68:38:51, 4.04it/s, grad_norm=0.536, loss_final=0.581, loss_mean=0.855, loss_mean_cls=0.0823, proj_loss=-0.356][2026-03-26 12:38:00] Step: 1666, Training Logs: loss_final: 0.575865, loss_mean: 0.852296, proj_loss: -0.358654, loss_mean_cls: 0.082222, grad_norm: 0.393043 +Steps: 0%| | 1667/1000000 [06:56<68:38:41, 4.04it/s, grad_norm=0.393, loss_final=0.576, loss_mean=0.852, loss_mean_cls=0.0822, proj_loss=-0.359][2026-03-26 12:38:00] Step: 1667, Training Logs: loss_final: 0.599643, loss_mean: 0.855330, proj_loss: -0.340685, loss_mean_cls: 0.084998, grad_norm: 0.583622 +Steps: 0%| | 1668/1000000 [06:57<68:41:40, 4.04it/s, grad_norm=0.584, loss_final=0.6, loss_mean=0.855, loss_mean_cls=0.085, proj_loss=-0.341][2026-03-26 12:38:00] Step: 1668, Training Logs: loss_final: 0.587532, loss_mean: 0.854092, proj_loss: -0.350662, loss_mean_cls: 0.084101, grad_norm: 0.565243 +Steps: 0%| | 1669/1000000 [06:57<68:41:42, 4.04it/s, grad_norm=0.565, loss_final=0.588, loss_mean=0.854, loss_mean_cls=0.0841, proj_loss=-0.351][2026-03-26 12:38:00] Step: 1669, Training Logs: loss_final: 0.593033, loss_mean: 0.872795, proj_loss: -0.361366, loss_mean_cls: 0.081604, grad_norm: 0.453949 +Steps: 0%| | 1670/1000000 [06:57<68:42:20, 4.04it/s, grad_norm=0.454, loss_final=0.593, loss_mean=0.873, loss_mean_cls=0.0816, proj_loss=-0.361][2026-03-26 12:38:01] Step: 1670, Training Logs: loss_final: 0.595970, loss_mean: 0.857178, proj_loss: -0.346609, loss_mean_cls: 0.085400, grad_norm: 0.843164 +Steps: 0%| | 1671/1000000 [06:57<68:40:54, 4.04it/s, grad_norm=0.843, loss_final=0.596, loss_mean=0.857, loss_mean_cls=0.0854, proj_loss=-0.347][2026-03-26 12:38:01] Step: 1671, Training Logs: loss_final: 0.578142, loss_mean: 0.837727, proj_loss: -0.344706, loss_mean_cls: 0.085121, grad_norm: 0.514767 +Steps: 0%| | 1672/1000000 [06:58<68:39:46, 4.04it/s, grad_norm=0.515, loss_final=0.578, loss_mean=0.838, loss_mean_cls=0.0851, proj_loss=-0.345][2026-03-26 12:38:01] Step: 1672, Training Logs: loss_final: 0.586530, loss_mean: 0.863768, proj_loss: -0.358525, loss_mean_cls: 0.081287, grad_norm: 0.659446 +Steps: 0%| | 1673/1000000 [06:58<68:39:33, 4.04it/s, grad_norm=0.659, loss_final=0.587, loss_mean=0.864, loss_mean_cls=0.0813, proj_loss=-0.359][2026-03-26 12:38:01] Step: 1673, Training Logs: loss_final: 0.602045, loss_mean: 0.877369, proj_loss: -0.356799, loss_mean_cls: 0.081475, grad_norm: 0.578711 +Steps: 0%| | 1674/1000000 [06:58<68:41:47, 4.04it/s, grad_norm=0.579, loss_final=0.602, loss_mean=0.877, loss_mean_cls=0.0815, proj_loss=-0.357][2026-03-26 12:38:01] Step: 1674, Training Logs: loss_final: 0.590638, loss_mean: 0.863735, proj_loss: -0.356294, loss_mean_cls: 0.083196, grad_norm: 0.521642 +Steps: 0%| | 1675/1000000 [06:58<68:56:51, 4.02it/s, grad_norm=0.522, loss_final=0.591, loss_mean=0.864, loss_mean_cls=0.0832, proj_loss=-0.356][2026-03-26 12:38:02] Step: 1675, Training Logs: loss_final: 0.596084, loss_mean: 0.867234, proj_loss: -0.353371, loss_mean_cls: 0.082221, grad_norm: 0.757960 +Steps: 0%| | 1676/1000000 [06:59<68:35:56, 4.04it/s, grad_norm=0.758, loss_final=0.596, loss_mean=0.867, loss_mean_cls=0.0822, proj_loss=-0.353][2026-03-26 12:38:02] Step: 1676, Training Logs: loss_final: 0.594149, loss_mean: 0.879083, proj_loss: -0.364661, loss_mean_cls: 0.079727, grad_norm: 0.672025 +Steps: 0%| | 1677/1000000 [06:59<68:41:03, 4.04it/s, grad_norm=0.672, loss_final=0.594, loss_mean=0.879, loss_mean_cls=0.0797, proj_loss=-0.365][2026-03-26 12:38:02] Step: 1677, Training Logs: loss_final: 0.587498, loss_mean: 0.840927, proj_loss: -0.339915, loss_mean_cls: 0.086486, grad_norm: 0.584725 +Steps: 0%| | 1678/1000000 [06:59<68:37:11, 4.04it/s, grad_norm=0.585, loss_final=0.587, loss_mean=0.841, loss_mean_cls=0.0865, proj_loss=-0.34][2026-03-26 12:38:02] Step: 1678, Training Logs: loss_final: 0.571595, loss_mean: 0.850573, proj_loss: -0.360596, loss_mean_cls: 0.081618, grad_norm: 0.655148 +Steps: 0%| | 1679/1000000 [06:59<68:38:49, 4.04it/s, grad_norm=0.655, loss_final=0.572, loss_mean=0.851, loss_mean_cls=0.0816, proj_loss=-0.361][2026-03-26 12:38:03] Step: 1679, Training Logs: loss_final: 0.579233, loss_mean: 0.856064, proj_loss: -0.360399, loss_mean_cls: 0.083568, grad_norm: 0.602019 +Steps: 0%| | 1680/1000000 [07:00<68:37:10, 4.04it/s, grad_norm=0.602, loss_final=0.579, loss_mean=0.856, loss_mean_cls=0.0836, proj_loss=-0.36][2026-03-26 12:38:03] Step: 1680, Training Logs: loss_final: 0.588619, loss_mean: 0.849084, proj_loss: -0.345158, loss_mean_cls: 0.084693, grad_norm: 0.717382 +Steps: 0%| | 1681/1000000 [07:00<68:36:10, 4.04it/s, grad_norm=0.717, loss_final=0.589, loss_mean=0.849, loss_mean_cls=0.0847, proj_loss=-0.345][2026-03-26 12:38:03] Step: 1681, Training Logs: loss_final: 0.586929, loss_mean: 0.862122, proj_loss: -0.358665, loss_mean_cls: 0.083472, grad_norm: 0.652555 +Steps: 0%| | 1682/1000000 [07:00<68:36:05, 4.04it/s, grad_norm=0.653, loss_final=0.587, loss_mean=0.862, loss_mean_cls=0.0835, proj_loss=-0.359][2026-03-26 12:38:03] Step: 1682, Training Logs: loss_final: 0.594114, loss_mean: 0.882872, proj_loss: -0.368926, loss_mean_cls: 0.080168, grad_norm: 0.584232 +Steps: 0%| | 1683/1000000 [07:00<68:38:02, 4.04it/s, grad_norm=0.584, loss_final=0.594, loss_mean=0.883, loss_mean_cls=0.0802, proj_loss=-0.369][2026-03-26 12:38:04] Step: 1683, Training Logs: loss_final: 0.593134, loss_mean: 0.856314, proj_loss: -0.347591, loss_mean_cls: 0.084411, grad_norm: 0.486632 +Steps: 0%| | 1684/1000000 [07:01<68:35:48, 4.04it/s, grad_norm=0.487, loss_final=0.593, loss_mean=0.856, loss_mean_cls=0.0844, proj_loss=-0.348][2026-03-26 12:38:04] Step: 1684, Training Logs: loss_final: 0.581470, loss_mean: 0.843995, proj_loss: -0.346791, loss_mean_cls: 0.084267, grad_norm: 0.717867 +Steps: 0%| | 1685/1000000 [07:01<68:40:50, 4.04it/s, grad_norm=0.718, loss_final=0.581, loss_mean=0.844, loss_mean_cls=0.0843, proj_loss=-0.347][2026-03-26 12:38:04] Step: 1685, Training Logs: loss_final: 0.588180, loss_mean: 0.858414, proj_loss: -0.353564, loss_mean_cls: 0.083331, grad_norm: 0.445237 +Steps: 0%| | 1686/1000000 [07:01<68:39:26, 4.04it/s, grad_norm=0.445, loss_final=0.588, loss_mean=0.858, loss_mean_cls=0.0833, proj_loss=-0.354][2026-03-26 12:38:04] Step: 1686, Training Logs: loss_final: 0.587825, loss_mean: 0.869776, proj_loss: -0.362856, loss_mean_cls: 0.080905, grad_norm: 0.524509 +Steps: 0%| | 1687/1000000 [07:01<68:38:33, 4.04it/s, grad_norm=0.525, loss_final=0.588, loss_mean=0.87, loss_mean_cls=0.0809, proj_loss=-0.363][2026-03-26 12:38:05] Step: 1687, Training Logs: loss_final: 0.583082, loss_mean: 0.854687, proj_loss: -0.356005, loss_mean_cls: 0.084400, grad_norm: 0.690808 +Steps: 0%| | 1688/1000000 [07:02<68:38:58, 4.04it/s, grad_norm=0.691, loss_final=0.583, loss_mean=0.855, loss_mean_cls=0.0844, proj_loss=-0.356][2026-03-26 12:38:05] Step: 1688, Training Logs: loss_final: 0.558234, loss_mean: 0.841530, proj_loss: -0.364469, loss_mean_cls: 0.081172, grad_norm: 0.661790 +Steps: 0%| | 1689/1000000 [07:02<68:39:56, 4.04it/s, grad_norm=0.662, loss_final=0.558, loss_mean=0.842, loss_mean_cls=0.0812, proj_loss=-0.364][2026-03-26 12:38:05] Step: 1689, Training Logs: loss_final: 0.573456, loss_mean: 0.837077, proj_loss: -0.348053, loss_mean_cls: 0.084431, grad_norm: 0.518090 +Steps: 0%| | 1690/1000000 [07:02<68:40:37, 4.04it/s, grad_norm=0.518, loss_final=0.573, loss_mean=0.837, loss_mean_cls=0.0844, proj_loss=-0.348][2026-03-26 12:38:05] Step: 1690, Training Logs: loss_final: 0.583210, loss_mean: 0.837693, proj_loss: -0.342042, loss_mean_cls: 0.087559, grad_norm: 0.828326 +Steps: 0%| | 1691/1000000 [07:02<68:39:13, 4.04it/s, grad_norm=0.828, loss_final=0.583, loss_mean=0.838, loss_mean_cls=0.0876, proj_loss=-0.342][2026-03-26 12:38:06] Step: 1691, Training Logs: loss_final: 0.606458, loss_mean: 0.879105, proj_loss: -0.356014, loss_mean_cls: 0.083367, grad_norm: 0.552635 +Steps: 0%| | 1692/1000000 [07:03<68:41:44, 4.04it/s, grad_norm=0.553, loss_final=0.606, loss_mean=0.879, loss_mean_cls=0.0834, proj_loss=-0.356][2026-03-26 12:38:06] Step: 1692, Training Logs: loss_final: 0.593906, loss_mean: 0.860929, proj_loss: -0.350388, loss_mean_cls: 0.083364, grad_norm: 0.637848 +Steps: 0%| | 1693/1000000 [07:03<68:40:53, 4.04it/s, grad_norm=0.638, loss_final=0.594, loss_mean=0.861, loss_mean_cls=0.0834, proj_loss=-0.35][2026-03-26 12:38:06] Step: 1693, Training Logs: loss_final: 0.582323, loss_mean: 0.865599, proj_loss: -0.363694, loss_mean_cls: 0.080418, grad_norm: 0.582677 +Steps: 0%| | 1694/1000000 [07:03<68:42:14, 4.04it/s, grad_norm=0.583, loss_final=0.582, loss_mean=0.866, loss_mean_cls=0.0804, proj_loss=-0.364][2026-03-26 12:38:06] Step: 1694, Training Logs: loss_final: 0.597485, loss_mean: 0.873070, proj_loss: -0.358184, loss_mean_cls: 0.082599, grad_norm: 0.580397 +Steps: 0%| | 1695/1000000 [07:03<68:39:34, 4.04it/s, grad_norm=0.58, loss_final=0.597, loss_mean=0.873, loss_mean_cls=0.0826, proj_loss=-0.358][2026-03-26 12:38:07] Step: 1695, Training Logs: loss_final: 0.583752, loss_mean: 0.874878, proj_loss: -0.370383, loss_mean_cls: 0.079257, grad_norm: 0.445255 +Steps: 0%| | 1696/1000000 [07:04<68:41:25, 4.04it/s, grad_norm=0.445, loss_final=0.584, loss_mean=0.875, loss_mean_cls=0.0793, proj_loss=-0.37][2026-03-26 12:38:07] Step: 1696, Training Logs: loss_final: 0.595468, loss_mean: 0.877060, proj_loss: -0.363237, loss_mean_cls: 0.081645, grad_norm: 0.449488 +Steps: 0%| | 1697/1000000 [07:04<68:50:18, 4.03it/s, grad_norm=0.449, loss_final=0.595, loss_mean=0.877, loss_mean_cls=0.0816, proj_loss=-0.363][2026-03-26 12:38:07] Step: 1697, Training Logs: loss_final: 0.595278, loss_mean: 0.862944, proj_loss: -0.352100, loss_mean_cls: 0.084433, grad_norm: 0.333147 +Steps: 0%| | 1698/1000000 [07:04<69:39:31, 3.98it/s, grad_norm=0.333, loss_final=0.595, loss_mean=0.863, loss_mean_cls=0.0844, proj_loss=-0.352][2026-03-26 12:38:07] Step: 1698, Training Logs: loss_final: 0.601825, loss_mean: 0.879374, proj_loss: -0.359527, loss_mean_cls: 0.081978, grad_norm: 0.461351 +Steps: 0%| | 1699/1000000 [07:04<69:22:04, 4.00it/s, grad_norm=0.461, loss_final=0.602, loss_mean=0.879, loss_mean_cls=0.082, proj_loss=-0.36][2026-03-26 12:38:08] Step: 1699, Training Logs: loss_final: 0.590857, loss_mean: 0.860263, proj_loss: -0.353048, loss_mean_cls: 0.083642, grad_norm: 0.466090 +Steps: 0%| | 1700/1000000 [07:05<69:13:09, 4.01it/s, grad_norm=0.466, loss_final=0.591, loss_mean=0.86, loss_mean_cls=0.0836, proj_loss=-0.353][2026-03-26 12:38:08] Step: 1700, Training Logs: loss_final: 0.587253, loss_mean: 0.852744, proj_loss: -0.350449, loss_mean_cls: 0.084958, grad_norm: 0.480217 +Steps: 0%| | 1701/1000000 [07:05<69:00:14, 4.02it/s, grad_norm=0.48, loss_final=0.587, loss_mean=0.853, loss_mean_cls=0.085, proj_loss=-0.35][2026-03-26 12:38:08] Step: 1701, Training Logs: loss_final: 0.578855, loss_mean: 0.837566, proj_loss: -0.344271, loss_mean_cls: 0.085559, grad_norm: 0.629080 +Steps: 0%| | 1702/1000000 [07:05<68:58:00, 4.02it/s, grad_norm=0.629, loss_final=0.579, loss_mean=0.838, loss_mean_cls=0.0856, proj_loss=-0.344][2026-03-26 12:38:08] Step: 1702, Training Logs: loss_final: 0.581404, loss_mean: 0.855674, proj_loss: -0.357622, loss_mean_cls: 0.083352, grad_norm: 0.584316 +Steps: 0%| | 1703/1000000 [07:05<68:50:54, 4.03it/s, grad_norm=0.584, loss_final=0.581, loss_mean=0.856, loss_mean_cls=0.0834, proj_loss=-0.358][2026-03-26 12:38:09] Step: 1703, Training Logs: loss_final: 0.590065, loss_mean: 0.880524, proj_loss: -0.369131, loss_mean_cls: 0.078672, grad_norm: 0.404495 +Steps: 0%| | 1704/1000000 [07:06<68:47:18, 4.03it/s, grad_norm=0.404, loss_final=0.59, loss_mean=0.881, loss_mean_cls=0.0787, proj_loss=-0.369][2026-03-26 12:38:09] Step: 1704, Training Logs: loss_final: 0.583521, loss_mean: 0.858743, proj_loss: -0.357001, loss_mean_cls: 0.081779, grad_norm: 0.593791 +Steps: 0%| | 1705/1000000 [07:06<68:42:14, 4.04it/s, grad_norm=0.594, loss_final=0.584, loss_mean=0.859, loss_mean_cls=0.0818, proj_loss=-0.357][2026-03-26 12:38:09] Step: 1705, Training Logs: loss_final: 0.568610, loss_mean: 0.852707, proj_loss: -0.365480, loss_mean_cls: 0.081383, grad_norm: 0.635073 +Steps: 0%| | 1706/1000000 [07:06<68:41:19, 4.04it/s, grad_norm=0.635, loss_final=0.569, loss_mean=0.853, loss_mean_cls=0.0814, proj_loss=-0.365][2026-03-26 12:38:09] Step: 1706, Training Logs: loss_final: 0.577586, loss_mean: 0.853787, proj_loss: -0.358326, loss_mean_cls: 0.082125, grad_norm: 0.454411 +Steps: 0%| | 1707/1000000 [07:06<68:40:36, 4.04it/s, grad_norm=0.454, loss_final=0.578, loss_mean=0.854, loss_mean_cls=0.0821, proj_loss=-0.358][2026-03-26 12:38:10] Step: 1707, Training Logs: loss_final: 0.586604, loss_mean: 0.840065, proj_loss: -0.340211, loss_mean_cls: 0.086751, grad_norm: 0.687161 +Steps: 0%| | 1708/1000000 [07:07<68:39:00, 4.04it/s, grad_norm=0.687, loss_final=0.587, loss_mean=0.84, loss_mean_cls=0.0868, proj_loss=-0.34][2026-03-26 12:38:10] Step: 1708, Training Logs: loss_final: 0.585711, loss_mean: 0.872389, proj_loss: -0.366805, loss_mean_cls: 0.080126, grad_norm: 0.654276 +Steps: 0%| | 1709/1000000 [07:07<68:37:51, 4.04it/s, grad_norm=0.654, loss_final=0.586, loss_mean=0.872, loss_mean_cls=0.0801, proj_loss=-0.367][2026-03-26 12:38:10] Step: 1709, Training Logs: loss_final: 0.588084, loss_mean: 0.856091, proj_loss: -0.351103, loss_mean_cls: 0.083095, grad_norm: 0.361701 +Steps: 0%| | 1710/1000000 [07:07<68:39:23, 4.04it/s, grad_norm=0.362, loss_final=0.588, loss_mean=0.856, loss_mean_cls=0.0831, proj_loss=-0.351][2026-03-26 12:38:10] Step: 1710, Training Logs: loss_final: 0.599200, loss_mean: 0.875292, proj_loss: -0.356984, loss_mean_cls: 0.080893, grad_norm: 0.502555 +Steps: 0%| | 1711/1000000 [07:07<68:41:00, 4.04it/s, grad_norm=0.503, loss_final=0.599, loss_mean=0.875, loss_mean_cls=0.0809, proj_loss=-0.357][2026-03-26 12:38:11] Step: 1711, Training Logs: loss_final: 0.596675, loss_mean: 0.870699, proj_loss: -0.356414, loss_mean_cls: 0.082391, grad_norm: 0.696119 +Steps: 0%| | 1712/1000000 [07:08<68:41:13, 4.04it/s, grad_norm=0.696, loss_final=0.597, loss_mean=0.871, loss_mean_cls=0.0824, proj_loss=-0.356][2026-03-26 12:38:11] Step: 1712, Training Logs: loss_final: 0.582288, loss_mean: 0.861008, proj_loss: -0.360879, loss_mean_cls: 0.082159, grad_norm: 0.490031 +Steps: 0%| | 1713/1000000 [07:08<68:38:05, 4.04it/s, grad_norm=0.49, loss_final=0.582, loss_mean=0.861, loss_mean_cls=0.0822, proj_loss=-0.361][2026-03-26 12:38:11] Step: 1713, Training Logs: loss_final: 0.598922, loss_mean: 0.861574, proj_loss: -0.346450, loss_mean_cls: 0.083798, grad_norm: 0.779622 +Steps: 0%| | 1714/1000000 [07:08<68:42:00, 4.04it/s, grad_norm=0.78, loss_final=0.599, loss_mean=0.862, loss_mean_cls=0.0838, proj_loss=-0.346][2026-03-26 12:38:11] Step: 1714, Training Logs: loss_final: 0.591911, loss_mean: 0.847285, proj_loss: -0.341584, loss_mean_cls: 0.086210, grad_norm: 0.716694 +Steps: 0%| | 1715/1000000 [07:08<68:41:08, 4.04it/s, grad_norm=0.717, loss_final=0.592, loss_mean=0.847, loss_mean_cls=0.0862, proj_loss=-0.342][2026-03-26 12:38:12] Step: 1715, Training Logs: loss_final: 0.585062, loss_mean: 0.859766, proj_loss: -0.357064, loss_mean_cls: 0.082361, grad_norm: 0.489387 +Steps: 0%| | 1716/1000000 [07:09<68:42:29, 4.04it/s, grad_norm=0.489, loss_final=0.585, loss_mean=0.86, loss_mean_cls=0.0824, proj_loss=-0.357][2026-03-26 12:38:12] Step: 1716, Training Logs: loss_final: 0.572551, loss_mean: 0.840108, proj_loss: -0.351410, loss_mean_cls: 0.083853, grad_norm: 0.783284 +Steps: 0%| | 1717/1000000 [07:09<68:40:33, 4.04it/s, grad_norm=0.783, loss_final=0.573, loss_mean=0.84, loss_mean_cls=0.0839, proj_loss=-0.351][2026-03-26 12:38:12] Step: 1717, Training Logs: loss_final: 0.592785, loss_mean: 0.880836, proj_loss: -0.366445, loss_mean_cls: 0.078395, grad_norm: 0.425338 +Steps: 0%| | 1718/1000000 [07:09<68:39:47, 4.04it/s, grad_norm=0.425, loss_final=0.593, loss_mean=0.881, loss_mean_cls=0.0784, proj_loss=-0.366][2026-03-26 12:38:12] Step: 1718, Training Logs: loss_final: 0.585868, loss_mean: 0.858827, proj_loss: -0.355252, loss_mean_cls: 0.082293, grad_norm: 0.406619 +Steps: 0%| | 1719/1000000 [07:09<68:40:23, 4.04it/s, grad_norm=0.407, loss_final=0.586, loss_mean=0.859, loss_mean_cls=0.0823, proj_loss=-0.355][2026-03-26 12:38:13] Step: 1719, Training Logs: loss_final: 0.593124, loss_mean: 0.874513, proj_loss: -0.361535, loss_mean_cls: 0.080146, grad_norm: 0.588913 +Steps: 0%| | 1720/1000000 [07:10<68:43:18, 4.04it/s, grad_norm=0.589, loss_final=0.593, loss_mean=0.875, loss_mean_cls=0.0801, proj_loss=-0.362][2026-03-26 12:38:13] Step: 1720, Training Logs: loss_final: 0.582530, loss_mean: 0.860223, proj_loss: -0.360174, loss_mean_cls: 0.082481, grad_norm: 0.476405 +Steps: 0%| | 1721/1000000 [07:10<68:38:41, 4.04it/s, grad_norm=0.476, loss_final=0.583, loss_mean=0.86, loss_mean_cls=0.0825, proj_loss=-0.36][2026-03-26 12:38:13] Step: 1721, Training Logs: loss_final: 0.587276, loss_mean: 0.857515, proj_loss: -0.354149, loss_mean_cls: 0.083910, grad_norm: 0.528571 +Steps: 0%| | 1722/1000000 [07:10<68:36:37, 4.04it/s, grad_norm=0.529, loss_final=0.587, loss_mean=0.858, loss_mean_cls=0.0839, proj_loss=-0.354][2026-03-26 12:38:13] Step: 1722, Training Logs: loss_final: 0.575003, loss_mean: 0.845158, proj_loss: -0.354892, loss_mean_cls: 0.084736, grad_norm: 0.429765 +Steps: 0%| | 1723/1000000 [07:10<68:37:52, 4.04it/s, grad_norm=0.43, loss_final=0.575, loss_mean=0.845, loss_mean_cls=0.0847, proj_loss=-0.355][2026-03-26 12:38:14] Step: 1723, Training Logs: loss_final: 0.582042, loss_mean: 0.840922, proj_loss: -0.344367, loss_mean_cls: 0.085487, grad_norm: 0.608229 +Steps: 0%| | 1724/1000000 [07:10<68:35:24, 4.04it/s, grad_norm=0.608, loss_final=0.582, loss_mean=0.841, loss_mean_cls=0.0855, proj_loss=-0.344][2026-03-26 12:38:14] Step: 1724, Training Logs: loss_final: 0.587481, loss_mean: 0.858341, proj_loss: -0.355341, loss_mean_cls: 0.084481, grad_norm: 0.560834 +Steps: 0%| | 1725/1000000 [07:11<68:36:31, 4.04it/s, grad_norm=0.561, loss_final=0.587, loss_mean=0.858, loss_mean_cls=0.0845, proj_loss=-0.355][2026-03-26 12:38:14] Step: 1725, Training Logs: loss_final: 0.590033, loss_mean: 0.851276, proj_loss: -0.346306, loss_mean_cls: 0.085064, grad_norm: 0.425122 +Steps: 0%| | 1726/1000000 [07:11<68:36:52, 4.04it/s, grad_norm=0.425, loss_final=0.59, loss_mean=0.851, loss_mean_cls=0.0851, proj_loss=-0.346][2026-03-26 12:38:14] Step: 1726, Training Logs: loss_final: 0.582392, loss_mean: 0.850489, proj_loss: -0.353016, loss_mean_cls: 0.084919, grad_norm: 0.608118 +Steps: 0%| | 1727/1000000 [07:11<68:37:35, 4.04it/s, grad_norm=0.608, loss_final=0.582, loss_mean=0.85, loss_mean_cls=0.0849, proj_loss=-0.353][2026-03-26 12:38:15] Step: 1727, Training Logs: loss_final: 0.595628, loss_mean: 0.879856, proj_loss: -0.364958, loss_mean_cls: 0.080730, grad_norm: 0.808054 +Steps: 0%| | 1728/1000000 [07:11<68:36:22, 4.04it/s, grad_norm=0.808, loss_final=0.596, loss_mean=0.88, loss_mean_cls=0.0807, proj_loss=-0.365][2026-03-26 12:38:15] Step: 1728, Training Logs: loss_final: 0.559925, loss_mean: 0.825760, proj_loss: -0.350817, loss_mean_cls: 0.084982, grad_norm: 0.655469 +Steps: 0%| | 1729/1000000 [07:12<68:35:54, 4.04it/s, grad_norm=0.655, loss_final=0.56, loss_mean=0.826, loss_mean_cls=0.085, proj_loss=-0.351][2026-03-26 12:38:15] Step: 1729, Training Logs: loss_final: 0.576317, loss_mean: 0.822930, proj_loss: -0.334691, loss_mean_cls: 0.088077, grad_norm: 0.522324 +Steps: 0%| | 1730/1000000 [07:12<68:42:38, 4.04it/s, grad_norm=0.522, loss_final=0.576, loss_mean=0.823, loss_mean_cls=0.0881, proj_loss=-0.335][2026-03-26 12:38:15] Step: 1730, Training Logs: loss_final: 0.578761, loss_mean: 0.853339, proj_loss: -0.357187, loss_mean_cls: 0.082610, grad_norm: 0.865536 +Steps: 0%| | 1731/1000000 [07:12<68:40:04, 4.04it/s, grad_norm=0.866, loss_final=0.579, loss_mean=0.853, loss_mean_cls=0.0826, proj_loss=-0.357][2026-03-26 12:38:16] Step: 1731, Training Logs: loss_final: 0.585616, loss_mean: 0.864414, proj_loss: -0.360818, loss_mean_cls: 0.082020, grad_norm: 0.534359 +Steps: 0%| | 1732/1000000 [07:12<68:39:17, 4.04it/s, grad_norm=0.534, loss_final=0.586, loss_mean=0.864, loss_mean_cls=0.082, proj_loss=-0.361][2026-03-26 12:38:16] Step: 1732, Training Logs: loss_final: 0.570154, loss_mean: 0.859741, proj_loss: -0.370610, loss_mean_cls: 0.081022, grad_norm: 0.718092 +Steps: 0%| | 1733/1000000 [07:13<68:39:41, 4.04it/s, grad_norm=0.718, loss_final=0.57, loss_mean=0.86, loss_mean_cls=0.081, proj_loss=-0.371][2026-03-26 12:38:16] Step: 1733, Training Logs: loss_final: 0.590195, loss_mean: 0.865768, proj_loss: -0.357943, loss_mean_cls: 0.082369, grad_norm: 0.715604 +Steps: 0%| | 1734/1000000 [07:13<68:39:45, 4.04it/s, grad_norm=0.716, loss_final=0.59, loss_mean=0.866, loss_mean_cls=0.0824, proj_loss=-0.358][2026-03-26 12:38:16] Step: 1734, Training Logs: loss_final: 0.574212, loss_mean: 0.837238, proj_loss: -0.349028, loss_mean_cls: 0.086002, grad_norm: 0.398073 +Steps: 0%| | 1735/1000000 [07:13<68:37:02, 4.04it/s, grad_norm=0.398, loss_final=0.574, loss_mean=0.837, loss_mean_cls=0.086, proj_loss=-0.349][2026-03-26 12:38:17] Step: 1735, Training Logs: loss_final: 0.592500, loss_mean: 0.870739, proj_loss: -0.358887, loss_mean_cls: 0.080648, grad_norm: 0.616603 +Steps: 0%| | 1736/1000000 [07:13<68:38:07, 4.04it/s, grad_norm=0.617, loss_final=0.593, loss_mean=0.871, loss_mean_cls=0.0806, proj_loss=-0.359][2026-03-26 12:38:17] Step: 1736, Training Logs: loss_final: 0.583747, loss_mean: 0.844794, proj_loss: -0.346035, loss_mean_cls: 0.084988, grad_norm: 0.783320 +Steps: 0%| | 1737/1000000 [07:14<68:38:46, 4.04it/s, grad_norm=0.783, loss_final=0.584, loss_mean=0.845, loss_mean_cls=0.085, proj_loss=-0.346][2026-03-26 12:38:17] Step: 1737, Training Logs: loss_final: 0.562615, loss_mean: 0.828767, proj_loss: -0.350851, loss_mean_cls: 0.084698, grad_norm: 0.377667 +Steps: 0%| | 1738/1000000 [07:14<68:38:44, 4.04it/s, grad_norm=0.378, loss_final=0.563, loss_mean=0.829, loss_mean_cls=0.0847, proj_loss=-0.351][2026-03-26 12:38:17] Step: 1738, Training Logs: loss_final: 0.591776, loss_mean: 0.868681, proj_loss: -0.359092, loss_mean_cls: 0.082186, grad_norm: 1.086943 +Steps: 0%| | 1739/1000000 [07:14<68:37:25, 4.04it/s, grad_norm=1.09, loss_final=0.592, loss_mean=0.869, loss_mean_cls=0.0822, proj_loss=-0.359][2026-03-26 12:38:18] Step: 1739, Training Logs: loss_final: 0.576837, loss_mean: 0.866565, proj_loss: -0.370325, loss_mean_cls: 0.080597, grad_norm: 0.523699 +Steps: 0%| | 1740/1000000 [07:14<68:38:00, 4.04it/s, grad_norm=0.524, loss_final=0.577, loss_mean=0.867, loss_mean_cls=0.0806, proj_loss=-0.37][2026-03-26 12:38:18] Step: 1740, Training Logs: loss_final: 0.590544, loss_mean: 0.867133, proj_loss: -0.357806, loss_mean_cls: 0.081217, grad_norm: 0.597871 +Steps: 0%| | 1741/1000000 [07:15<68:35:04, 4.04it/s, grad_norm=0.598, loss_final=0.591, loss_mean=0.867, loss_mean_cls=0.0812, proj_loss=-0.358][2026-03-26 12:38:18] Step: 1741, Training Logs: loss_final: 0.603502, loss_mean: 0.876153, proj_loss: -0.355531, loss_mean_cls: 0.082880, grad_norm: 0.749856 +Steps: 0%| | 1742/1000000 [07:15<68:35:55, 4.04it/s, grad_norm=0.75, loss_final=0.604, loss_mean=0.876, loss_mean_cls=0.0829, proj_loss=-0.356][2026-03-26 12:38:18] Step: 1742, Training Logs: loss_final: 0.570825, loss_mean: 0.845276, proj_loss: -0.357710, loss_mean_cls: 0.083259, grad_norm: 0.416492 +Steps: 0%| | 1743/1000000 [07:15<68:32:46, 4.05it/s, grad_norm=0.416, loss_final=0.571, loss_mean=0.845, loss_mean_cls=0.0833, proj_loss=-0.358][2026-03-26 12:38:19] Step: 1743, Training Logs: loss_final: 0.590866, loss_mean: 0.869816, proj_loss: -0.360314, loss_mean_cls: 0.081364, grad_norm: 0.836691 +Steps: 0%| | 1744/1000000 [07:15<68:37:25, 4.04it/s, grad_norm=0.837, loss_final=0.591, loss_mean=0.87, loss_mean_cls=0.0814, proj_loss=-0.36][2026-03-26 12:38:19] Step: 1744, Training Logs: loss_final: 0.590431, loss_mean: 0.845990, proj_loss: -0.340466, loss_mean_cls: 0.084907, grad_norm: 0.648301 +Steps: 0%| | 1745/1000000 [07:16<68:38:34, 4.04it/s, grad_norm=0.648, loss_final=0.59, loss_mean=0.846, loss_mean_cls=0.0849, proj_loss=-0.34][2026-03-26 12:38:19] Step: 1745, Training Logs: loss_final: 0.579731, loss_mean: 0.859248, proj_loss: -0.361552, loss_mean_cls: 0.082035, grad_norm: 0.676793 +Steps: 0%| | 1746/1000000 [07:16<69:16:00, 4.00it/s, grad_norm=0.677, loss_final=0.58, loss_mean=0.859, loss_mean_cls=0.082, proj_loss=-0.362][2026-03-26 12:38:19] Step: 1746, Training Logs: loss_final: 0.596874, loss_mean: 0.870646, proj_loss: -0.355904, loss_mean_cls: 0.082132, grad_norm: 0.738639 +Steps: 0%| | 1747/1000000 [07:16<69:03:34, 4.02it/s, grad_norm=0.739, loss_final=0.597, loss_mean=0.871, loss_mean_cls=0.0821, proj_loss=-0.356][2026-03-26 12:38:20] Step: 1747, Training Logs: loss_final: 0.598816, loss_mean: 0.874241, proj_loss: -0.357564, loss_mean_cls: 0.082139, grad_norm: 0.456216 +Steps: 0%| | 1748/1000000 [07:16<69:28:11, 3.99it/s, grad_norm=0.456, loss_final=0.599, loss_mean=0.874, loss_mean_cls=0.0821, proj_loss=-0.358][2026-03-26 12:38:20] Step: 1748, Training Logs: loss_final: 0.593932, loss_mean: 0.870716, proj_loss: -0.359042, loss_mean_cls: 0.082259, grad_norm: 0.576815 +Steps: 0%| | 1749/1000000 [07:17<69:12:14, 4.01it/s, grad_norm=0.577, loss_final=0.594, loss_mean=0.871, loss_mean_cls=0.0823, proj_loss=-0.359][2026-03-26 12:38:20] Step: 1749, Training Logs: loss_final: 0.591705, loss_mean: 0.861993, proj_loss: -0.354075, loss_mean_cls: 0.083787, grad_norm: 0.629950 +Steps: 0%| | 1750/1000000 [07:17<69:02:48, 4.02it/s, grad_norm=0.63, loss_final=0.592, loss_mean=0.862, loss_mean_cls=0.0838, proj_loss=-0.354][2026-03-26 12:38:20] Step: 1750, Training Logs: loss_final: 0.576564, loss_mean: 0.860149, proj_loss: -0.364601, loss_mean_cls: 0.081016, grad_norm: 0.425139 +Steps: 0%| | 1751/1000000 [07:17<68:53:12, 4.03it/s, grad_norm=0.425, loss_final=0.577, loss_mean=0.86, loss_mean_cls=0.081, proj_loss=-0.365][2026-03-26 12:38:21] Step: 1751, Training Logs: loss_final: 0.591242, loss_mean: 0.874315, proj_loss: -0.363989, loss_mean_cls: 0.080916, grad_norm: 0.629554 +Steps: 0%| | 1752/1000000 [07:17<68:48:44, 4.03it/s, grad_norm=0.63, loss_final=0.591, loss_mean=0.874, loss_mean_cls=0.0809, proj_loss=-0.364][2026-03-26 12:38:21] Step: 1752, Training Logs: loss_final: 0.584967, loss_mean: 0.835502, proj_loss: -0.336730, loss_mean_cls: 0.086196, grad_norm: 0.559564 +Steps: 0%| | 1753/1000000 [07:18<68:44:24, 4.03it/s, grad_norm=0.56, loss_final=0.585, loss_mean=0.836, loss_mean_cls=0.0862, proj_loss=-0.337][2026-03-26 12:38:21] Step: 1753, Training Logs: loss_final: 0.584995, loss_mean: 0.873680, proj_loss: -0.369268, loss_mean_cls: 0.080583, grad_norm: 0.386078 +Steps: 0%| | 1754/1000000 [07:18<68:43:03, 4.04it/s, grad_norm=0.386, loss_final=0.585, loss_mean=0.874, loss_mean_cls=0.0806, proj_loss=-0.369][2026-03-26 12:38:21] Step: 1754, Training Logs: loss_final: 0.589980, loss_mean: 0.869229, proj_loss: -0.360478, loss_mean_cls: 0.081229, grad_norm: 0.464172 +Steps: 0%| | 1755/1000000 [07:18<68:40:19, 4.04it/s, grad_norm=0.464, loss_final=0.59, loss_mean=0.869, loss_mean_cls=0.0812, proj_loss=-0.36][2026-03-26 12:38:22] Step: 1755, Training Logs: loss_final: 0.573798, loss_mean: 0.839173, proj_loss: -0.349103, loss_mean_cls: 0.083728, grad_norm: 0.447231 +Steps: 0%| | 1756/1000000 [07:18<68:41:07, 4.04it/s, grad_norm=0.447, loss_final=0.574, loss_mean=0.839, loss_mean_cls=0.0837, proj_loss=-0.349][2026-03-26 12:38:22] Step: 1756, Training Logs: loss_final: 0.564942, loss_mean: 0.839000, proj_loss: -0.357794, loss_mean_cls: 0.083736, grad_norm: 0.385845 +Steps: 0%| | 1757/1000000 [07:19<68:38:24, 4.04it/s, grad_norm=0.386, loss_final=0.565, loss_mean=0.839, loss_mean_cls=0.0837, proj_loss=-0.358][2026-03-26 12:38:22] Step: 1757, Training Logs: loss_final: 0.583114, loss_mean: 0.865406, proj_loss: -0.364308, loss_mean_cls: 0.082016, grad_norm: 0.428772 +Steps: 0%| | 1758/1000000 [07:19<68:38:46, 4.04it/s, grad_norm=0.429, loss_final=0.583, loss_mean=0.865, loss_mean_cls=0.082, proj_loss=-0.364][2026-03-26 12:38:22] Step: 1758, Training Logs: loss_final: 0.585587, loss_mean: 0.871794, proj_loss: -0.366870, loss_mean_cls: 0.080663, grad_norm: 0.309950 +Steps: 0%| | 1759/1000000 [07:19<68:38:17, 4.04it/s, grad_norm=0.31, loss_final=0.586, loss_mean=0.872, loss_mean_cls=0.0807, proj_loss=-0.367][2026-03-26 12:38:23] Step: 1759, Training Logs: loss_final: 0.579823, loss_mean: 0.849588, proj_loss: -0.354543, loss_mean_cls: 0.084778, grad_norm: 0.507051 +Steps: 0%| | 1760/1000000 [07:19<68:39:14, 4.04it/s, grad_norm=0.507, loss_final=0.58, loss_mean=0.85, loss_mean_cls=0.0848, proj_loss=-0.355][2026-03-26 12:38:23] Step: 1760, Training Logs: loss_final: 0.596607, loss_mean: 0.850764, proj_loss: -0.340092, loss_mean_cls: 0.085935, grad_norm: 0.436107 +Steps: 0%| | 1761/1000000 [07:20<68:43:57, 4.03it/s, grad_norm=0.436, loss_final=0.597, loss_mean=0.851, loss_mean_cls=0.0859, proj_loss=-0.34][2026-03-26 12:38:23] Step: 1761, Training Logs: loss_final: 0.591086, loss_mean: 0.850279, proj_loss: -0.345607, loss_mean_cls: 0.086414, grad_norm: 0.401876 +Steps: 0%| | 1762/1000000 [07:20<68:42:16, 4.04it/s, grad_norm=0.402, loss_final=0.591, loss_mean=0.85, loss_mean_cls=0.0864, proj_loss=-0.346][2026-03-26 12:38:23] Step: 1762, Training Logs: loss_final: 0.584335, loss_mean: 0.857218, proj_loss: -0.355933, loss_mean_cls: 0.083050, grad_norm: 0.406398 +Steps: 0%| | 1763/1000000 [07:20<68:41:20, 4.04it/s, grad_norm=0.406, loss_final=0.584, loss_mean=0.857, loss_mean_cls=0.0831, proj_loss=-0.356][2026-03-26 12:38:24] Step: 1763, Training Logs: loss_final: 0.577661, loss_mean: 0.870443, proj_loss: -0.372028, loss_mean_cls: 0.079247, grad_norm: 0.489515 +Steps: 0%| | 1764/1000000 [07:20<68:40:44, 4.04it/s, grad_norm=0.49, loss_final=0.578, loss_mean=0.87, loss_mean_cls=0.0792, proj_loss=-0.372][2026-03-26 12:38:24] Step: 1764, Training Logs: loss_final: 0.577257, loss_mean: 0.841611, proj_loss: -0.350264, loss_mean_cls: 0.085910, grad_norm: 0.441611 +Steps: 0%| | 1765/1000000 [07:21<68:51:34, 4.03it/s, grad_norm=0.442, loss_final=0.577, loss_mean=0.842, loss_mean_cls=0.0859, proj_loss=-0.35][2026-03-26 12:38:24] Step: 1765, Training Logs: loss_final: 0.581195, loss_mean: 0.844362, proj_loss: -0.348788, loss_mean_cls: 0.085621, grad_norm: 0.570365 +Steps: 0%| | 1766/1000000 [07:21<68:47:27, 4.03it/s, grad_norm=0.57, loss_final=0.581, loss_mean=0.844, loss_mean_cls=0.0856, proj_loss=-0.349][2026-03-26 12:38:24] Step: 1766, Training Logs: loss_final: 0.585916, loss_mean: 0.859540, proj_loss: -0.355741, loss_mean_cls: 0.082117, grad_norm: 0.627781 +Steps: 0%| | 1767/1000000 [07:21<68:45:42, 4.03it/s, grad_norm=0.628, loss_final=0.586, loss_mean=0.86, loss_mean_cls=0.0821, proj_loss=-0.356][2026-03-26 12:38:25] Step: 1767, Training Logs: loss_final: 0.572669, loss_mean: 0.863127, proj_loss: -0.371056, loss_mean_cls: 0.080598, grad_norm: 0.551412 +Steps: 0%| | 1768/1000000 [07:21<68:42:22, 4.04it/s, grad_norm=0.551, loss_final=0.573, loss_mean=0.863, loss_mean_cls=0.0806, proj_loss=-0.371][2026-03-26 12:38:25] Step: 1768, Training Logs: loss_final: 0.576381, loss_mean: 0.849711, proj_loss: -0.357006, loss_mean_cls: 0.083676, grad_norm: 0.950241 +Steps: 0%| | 1769/1000000 [07:22<68:43:23, 4.03it/s, grad_norm=0.95, loss_final=0.576, loss_mean=0.85, loss_mean_cls=0.0837, proj_loss=-0.357][2026-03-26 12:38:25] Step: 1769, Training Logs: loss_final: 0.589444, loss_mean: 0.860263, proj_loss: -0.354653, loss_mean_cls: 0.083834, grad_norm: 0.795202 +Steps: 0%| | 1770/1000000 [07:22<68:39:54, 4.04it/s, grad_norm=0.795, loss_final=0.589, loss_mean=0.86, loss_mean_cls=0.0838, proj_loss=-0.355][2026-03-26 12:38:25] Step: 1770, Training Logs: loss_final: 0.582929, loss_mean: 0.851741, proj_loss: -0.353633, loss_mean_cls: 0.084821, grad_norm: 0.536862 +Steps: 0%| | 1771/1000000 [07:22<68:41:48, 4.04it/s, grad_norm=0.537, loss_final=0.583, loss_mean=0.852, loss_mean_cls=0.0848, proj_loss=-0.354][2026-03-26 12:38:26] Step: 1771, Training Logs: loss_final: 0.578684, loss_mean: 0.851492, proj_loss: -0.355005, loss_mean_cls: 0.082198, grad_norm: 0.689853 +Steps: 0%| | 1772/1000000 [07:22<68:38:48, 4.04it/s, grad_norm=0.69, loss_final=0.579, loss_mean=0.851, loss_mean_cls=0.0822, proj_loss=-0.355][2026-03-26 12:38:26] Step: 1772, Training Logs: loss_final: 0.581108, loss_mean: 0.855645, proj_loss: -0.357581, loss_mean_cls: 0.083044, grad_norm: 0.525829 +Steps: 0%| | 1773/1000000 [07:23<68:38:52, 4.04it/s, grad_norm=0.526, loss_final=0.581, loss_mean=0.856, loss_mean_cls=0.083, proj_loss=-0.358][2026-03-26 12:38:26] Step: 1773, Training Logs: loss_final: 0.569164, loss_mean: 0.837649, proj_loss: -0.353078, loss_mean_cls: 0.084594, grad_norm: 0.544930 +Steps: 0%| | 1774/1000000 [07:23<72:34:26, 3.82it/s, grad_norm=0.545, loss_final=0.569, loss_mean=0.838, loss_mean_cls=0.0846, proj_loss=-0.353][2026-03-26 12:38:26] Step: 1774, Training Logs: loss_final: 0.595900, loss_mean: 0.852039, proj_loss: -0.341374, loss_mean_cls: 0.085235, grad_norm: 0.498662 +Steps: 0%| | 1775/1000000 [07:23<71:31:36, 3.88it/s, grad_norm=0.499, loss_final=0.596, loss_mean=0.852, loss_mean_cls=0.0852, proj_loss=-0.341][2026-03-26 12:38:27] Step: 1775, Training Logs: loss_final: 0.573966, loss_mean: 0.854910, proj_loss: -0.362382, loss_mean_cls: 0.081439, grad_norm: 0.752468 +Steps: 0%| | 1776/1000000 [07:23<70:40:32, 3.92it/s, grad_norm=0.752, loss_final=0.574, loss_mean=0.855, loss_mean_cls=0.0814, proj_loss=-0.362][2026-03-26 12:38:27] Step: 1776, Training Logs: loss_final: 0.581329, loss_mean: 0.846419, proj_loss: -0.349790, loss_mean_cls: 0.084700, grad_norm: 0.841046 +Steps: 0%| | 1777/1000000 [07:24<70:03:57, 3.96it/s, grad_norm=0.841, loss_final=0.581, loss_mean=0.846, loss_mean_cls=0.0847, proj_loss=-0.35][2026-03-26 12:38:27] Step: 1777, Training Logs: loss_final: 0.587786, loss_mean: 0.860951, proj_loss: -0.355342, loss_mean_cls: 0.082177, grad_norm: 0.600989 +Steps: 0%| | 1778/1000000 [07:24<69:39:31, 3.98it/s, grad_norm=0.601, loss_final=0.588, loss_mean=0.861, loss_mean_cls=0.0822, proj_loss=-0.355][2026-03-26 12:38:27] Step: 1778, Training Logs: loss_final: 0.594789, loss_mean: 0.859052, proj_loss: -0.348825, loss_mean_cls: 0.084562, grad_norm: 0.922863 +Steps: 0%| | 1779/1000000 [07:24<69:19:25, 4.00it/s, grad_norm=0.923, loss_final=0.595, loss_mean=0.859, loss_mean_cls=0.0846, proj_loss=-0.349][2026-03-26 12:38:28] Step: 1779, Training Logs: loss_final: 0.578842, loss_mean: 0.849623, proj_loss: -0.355236, loss_mean_cls: 0.084455, grad_norm: 0.844088 +Steps: 0%| | 1780/1000000 [07:24<69:09:08, 4.01it/s, grad_norm=0.844, loss_final=0.579, loss_mean=0.85, loss_mean_cls=0.0845, proj_loss=-0.355][2026-03-26 12:38:28] Step: 1780, Training Logs: loss_final: 0.590181, loss_mean: 0.863085, proj_loss: -0.355196, loss_mean_cls: 0.082292, grad_norm: 0.590356 +Steps: 0%| | 1781/1000000 [07:25<69:00:26, 4.02it/s, grad_norm=0.59, loss_final=0.59, loss_mean=0.863, loss_mean_cls=0.0823, proj_loss=-0.355][2026-03-26 12:38:28] Step: 1781, Training Logs: loss_final: 0.574754, loss_mean: 0.849481, proj_loss: -0.357919, loss_mean_cls: 0.083192, grad_norm: 0.959541 +Steps: 0%| | 1782/1000000 [07:25<68:53:02, 4.03it/s, grad_norm=0.96, loss_final=0.575, loss_mean=0.849, loss_mean_cls=0.0832, proj_loss=-0.358][2026-03-26 12:38:28] Step: 1782, Training Logs: loss_final: 0.594657, loss_mean: 0.876054, proj_loss: -0.362934, loss_mean_cls: 0.081538, grad_norm: 0.546855 +Steps: 0%| | 1783/1000000 [07:25<68:48:34, 4.03it/s, grad_norm=0.547, loss_final=0.595, loss_mean=0.876, loss_mean_cls=0.0815, proj_loss=-0.363][2026-03-26 12:38:29] Step: 1783, Training Logs: loss_final: 0.591695, loss_mean: 0.854730, proj_loss: -0.347386, loss_mean_cls: 0.084351, grad_norm: 0.848989 +Steps: 0%| | 1784/1000000 [07:25<68:47:09, 4.03it/s, grad_norm=0.849, loss_final=0.592, loss_mean=0.855, loss_mean_cls=0.0844, proj_loss=-0.347][2026-03-26 12:38:29] Step: 1784, Training Logs: loss_final: 0.576139, loss_mean: 0.848843, proj_loss: -0.356785, loss_mean_cls: 0.084081, grad_norm: 0.920913 +Steps: 0%| | 1785/1000000 [07:26<68:44:07, 4.03it/s, grad_norm=0.921, loss_final=0.576, loss_mean=0.849, loss_mean_cls=0.0841, proj_loss=-0.357][2026-03-26 12:38:29] Step: 1785, Training Logs: loss_final: 0.601292, loss_mean: 0.875234, proj_loss: -0.356618, loss_mean_cls: 0.082676, grad_norm: 0.762400 +Steps: 0%| | 1786/1000000 [07:26<68:41:44, 4.04it/s, grad_norm=0.762, loss_final=0.601, loss_mean=0.875, loss_mean_cls=0.0827, proj_loss=-0.357][2026-03-26 12:38:29] Step: 1786, Training Logs: loss_final: 0.576256, loss_mean: 0.840406, proj_loss: -0.349956, loss_mean_cls: 0.085806, grad_norm: 0.615777 +Steps: 0%| | 1787/1000000 [07:26<68:41:32, 4.04it/s, grad_norm=0.616, loss_final=0.576, loss_mean=0.84, loss_mean_cls=0.0858, proj_loss=-0.35][2026-03-26 12:38:30] Step: 1787, Training Logs: loss_final: 0.586820, loss_mean: 0.861532, proj_loss: -0.356910, loss_mean_cls: 0.082197, grad_norm: 0.784000 +Steps: 0%| | 1788/1000000 [07:26<68:41:03, 4.04it/s, grad_norm=0.784, loss_final=0.587, loss_mean=0.862, loss_mean_cls=0.0822, proj_loss=-0.357][2026-03-26 12:38:30] Step: 1788, Training Logs: loss_final: 0.572160, loss_mean: 0.840888, proj_loss: -0.353838, loss_mean_cls: 0.085110, grad_norm: 0.512877 +Steps: 0%| | 1789/1000000 [07:27<68:40:11, 4.04it/s, grad_norm=0.513, loss_final=0.572, loss_mean=0.841, loss_mean_cls=0.0851, proj_loss=-0.354][2026-03-26 12:38:30] Step: 1789, Training Logs: loss_final: 0.594037, loss_mean: 0.859858, proj_loss: -0.350950, loss_mean_cls: 0.085129, grad_norm: 0.787023 +Steps: 0%| | 1790/1000000 [07:27<68:40:21, 4.04it/s, grad_norm=0.787, loss_final=0.594, loss_mean=0.86, loss_mean_cls=0.0851, proj_loss=-0.351][2026-03-26 12:38:30] Step: 1790, Training Logs: loss_final: 0.569995, loss_mean: 0.846818, proj_loss: -0.359739, loss_mean_cls: 0.082916, grad_norm: 0.669056 +Steps: 0%| | 1791/1000000 [07:27<68:41:29, 4.04it/s, grad_norm=0.669, loss_final=0.57, loss_mean=0.847, loss_mean_cls=0.0829, proj_loss=-0.36][2026-03-26 12:38:31] Step: 1791, Training Logs: loss_final: 0.578486, loss_mean: 0.862646, proj_loss: -0.364982, loss_mean_cls: 0.080823, grad_norm: 0.581313 +Steps: 0%| | 1792/1000000 [07:27<68:41:15, 4.04it/s, grad_norm=0.581, loss_final=0.578, loss_mean=0.863, loss_mean_cls=0.0808, proj_loss=-0.365][2026-03-26 12:38:31] Step: 1792, Training Logs: loss_final: 0.581415, loss_mean: 0.847767, proj_loss: -0.352010, loss_mean_cls: 0.085658, grad_norm: 0.662865 +Steps: 0%| | 1793/1000000 [07:28<68:39:46, 4.04it/s, grad_norm=0.663, loss_final=0.581, loss_mean=0.848, loss_mean_cls=0.0857, proj_loss=-0.352][2026-03-26 12:38:31] Step: 1793, Training Logs: loss_final: 0.590630, loss_mean: 0.871793, proj_loss: -0.363138, loss_mean_cls: 0.081974, grad_norm: 0.551739 +Steps: 0%| | 1794/1000000 [07:28<68:41:20, 4.04it/s, grad_norm=0.552, loss_final=0.591, loss_mean=0.872, loss_mean_cls=0.082, proj_loss=-0.363][2026-03-26 12:38:31] Step: 1794, Training Logs: loss_final: 0.603807, loss_mean: 0.876556, proj_loss: -0.355251, loss_mean_cls: 0.082502, grad_norm: 1.097038 +Steps: 0%| | 1795/1000000 [07:28<68:39:25, 4.04it/s, grad_norm=1.1, loss_final=0.604, loss_mean=0.877, loss_mean_cls=0.0825, proj_loss=-0.355][2026-03-26 12:38:32] Step: 1795, Training Logs: loss_final: 0.591010, loss_mean: 0.855026, proj_loss: -0.347680, loss_mean_cls: 0.083664, grad_norm: 0.502994 +Steps: 0%| | 1796/1000000 [07:28<70:26:13, 3.94it/s, grad_norm=0.503, loss_final=0.591, loss_mean=0.855, loss_mean_cls=0.0837, proj_loss=-0.348][2026-03-26 12:38:32] Step: 1796, Training Logs: loss_final: 0.575935, loss_mean: 0.848158, proj_loss: -0.356180, loss_mean_cls: 0.083957, grad_norm: 0.724226 +Steps: 0%| | 1797/1000000 [07:29<69:59:36, 3.96it/s, grad_norm=0.724, loss_final=0.576, loss_mean=0.848, loss_mean_cls=0.084, proj_loss=-0.356][2026-03-26 12:38:32] Step: 1797, Training Logs: loss_final: 0.578987, loss_mean: 0.862659, proj_loss: -0.364861, loss_mean_cls: 0.081189, grad_norm: 0.584387 +Steps: 0%| | 1798/1000000 [07:29<70:30:12, 3.93it/s, grad_norm=0.584, loss_final=0.579, loss_mean=0.863, loss_mean_cls=0.0812, proj_loss=-0.365][2026-03-26 12:38:32] Step: 1798, Training Logs: loss_final: 0.595980, loss_mean: 0.876810, proj_loss: -0.362053, loss_mean_cls: 0.081224, grad_norm: 0.438060 +Steps: 0%| | 1799/1000000 [07:29<69:55:35, 3.97it/s, grad_norm=0.438, loss_final=0.596, loss_mean=0.877, loss_mean_cls=0.0812, proj_loss=-0.362][2026-03-26 12:38:33] Step: 1799, Training Logs: loss_final: 0.586342, loss_mean: 0.851188, proj_loss: -0.349411, loss_mean_cls: 0.084565, grad_norm: 0.558338 +Steps: 0%| | 1800/1000000 [07:29<69:32:42, 3.99it/s, grad_norm=0.558, loss_final=0.586, loss_mean=0.851, loss_mean_cls=0.0846, proj_loss=-0.349][2026-03-26 12:38:33] Step: 1800, Training Logs: loss_final: 0.604225, loss_mean: 0.883397, proj_loss: -0.360879, loss_mean_cls: 0.081707, grad_norm: 0.442446 +Steps: 0%| | 1801/1000000 [07:30<69:16:42, 4.00it/s, grad_norm=0.442, loss_final=0.604, loss_mean=0.883, loss_mean_cls=0.0817, proj_loss=-0.361][2026-03-26 12:38:33] Step: 1801, Training Logs: loss_final: 0.583376, loss_mean: 0.883448, proj_loss: -0.377908, loss_mean_cls: 0.077837, grad_norm: 0.757405 +Steps: 0%| | 1802/1000000 [07:30<69:05:23, 4.01it/s, grad_norm=0.757, loss_final=0.583, loss_mean=0.883, loss_mean_cls=0.0778, proj_loss=-0.378][2026-03-26 12:38:33] Step: 1802, Training Logs: loss_final: 0.570328, loss_mean: 0.851157, proj_loss: -0.362647, loss_mean_cls: 0.081817, grad_norm: 0.589134 +Steps: 0%| | 1803/1000000 [07:30<68:57:33, 4.02it/s, grad_norm=0.589, loss_final=0.57, loss_mean=0.851, loss_mean_cls=0.0818, proj_loss=-0.363][2026-03-26 12:38:34] Step: 1803, Training Logs: loss_final: 0.584632, loss_mean: 0.861154, proj_loss: -0.358716, loss_mean_cls: 0.082194, grad_norm: 0.474806 +Steps: 0%| | 1804/1000000 [07:30<68:51:13, 4.03it/s, grad_norm=0.475, loss_final=0.585, loss_mean=0.861, loss_mean_cls=0.0822, proj_loss=-0.359][2026-03-26 12:38:34] Step: 1804, Training Logs: loss_final: 0.573439, loss_mean: 0.836770, proj_loss: -0.348131, loss_mean_cls: 0.084800, grad_norm: 0.858730 +Steps: 0%| | 1805/1000000 [07:31<68:47:36, 4.03it/s, grad_norm=0.859, loss_final=0.573, loss_mean=0.837, loss_mean_cls=0.0848, proj_loss=-0.348][2026-03-26 12:38:34] Step: 1805, Training Logs: loss_final: 0.575625, loss_mean: 0.845280, proj_loss: -0.354521, loss_mean_cls: 0.084866, grad_norm: 0.374063 +Steps: 0%| | 1806/1000000 [07:31<68:43:45, 4.03it/s, grad_norm=0.374, loss_final=0.576, loss_mean=0.845, loss_mean_cls=0.0849, proj_loss=-0.355][2026-03-26 12:38:34] Step: 1806, Training Logs: loss_final: 0.581330, loss_mean: 0.852641, proj_loss: -0.356415, loss_mean_cls: 0.085104, grad_norm: 0.623491 +Steps: 0%| | 1807/1000000 [07:31<68:40:45, 4.04it/s, grad_norm=0.623, loss_final=0.581, loss_mean=0.853, loss_mean_cls=0.0851, proj_loss=-0.356][2026-03-26 12:38:35] Step: 1807, Training Logs: loss_final: 0.582924, loss_mean: 0.858506, proj_loss: -0.357752, loss_mean_cls: 0.082169, grad_norm: 0.418400 +Steps: 0%| | 1808/1000000 [07:31<68:39:58, 4.04it/s, grad_norm=0.418, loss_final=0.583, loss_mean=0.859, loss_mean_cls=0.0822, proj_loss=-0.358][2026-03-26 12:38:35] Step: 1808, Training Logs: loss_final: 0.598419, loss_mean: 0.859026, proj_loss: -0.346887, loss_mean_cls: 0.086280, grad_norm: 0.655981 +Steps: 0%| | 1809/1000000 [07:32<68:39:02, 4.04it/s, grad_norm=0.656, loss_final=0.598, loss_mean=0.859, loss_mean_cls=0.0863, proj_loss=-0.347][2026-03-26 12:38:35] Step: 1809, Training Logs: loss_final: 0.570964, loss_mean: 0.856487, proj_loss: -0.365779, loss_mean_cls: 0.080256, grad_norm: 0.473815 +Steps: 0%| | 1810/1000000 [07:32<68:40:05, 4.04it/s, grad_norm=0.474, loss_final=0.571, loss_mean=0.856, loss_mean_cls=0.0803, proj_loss=-0.366][2026-03-26 12:38:35] Step: 1810, Training Logs: loss_final: 0.580650, loss_mean: 0.856015, proj_loss: -0.357975, loss_mean_cls: 0.082610, grad_norm: 0.305684 +Steps: 0%| | 1811/1000000 [07:32<68:38:43, 4.04it/s, grad_norm=0.306, loss_final=0.581, loss_mean=0.856, loss_mean_cls=0.0826, proj_loss=-0.358][2026-03-26 12:38:36] Step: 1811, Training Logs: loss_final: 0.569738, loss_mean: 0.843958, proj_loss: -0.357932, loss_mean_cls: 0.083713, grad_norm: 0.697104 +Steps: 0%| | 1812/1000000 [07:32<68:40:14, 4.04it/s, grad_norm=0.697, loss_final=0.57, loss_mean=0.844, loss_mean_cls=0.0837, proj_loss=-0.358][2026-03-26 12:38:36] Step: 1812, Training Logs: loss_final: 0.581394, loss_mean: 0.855252, proj_loss: -0.357900, loss_mean_cls: 0.084042, grad_norm: 0.429414 +Steps: 0%| | 1813/1000000 [07:33<68:41:13, 4.04it/s, grad_norm=0.429, loss_final=0.581, loss_mean=0.855, loss_mean_cls=0.084, proj_loss=-0.358][2026-03-26 12:38:36] Step: 1813, Training Logs: loss_final: 0.571641, loss_mean: 0.859295, proj_loss: -0.368622, loss_mean_cls: 0.080968, grad_norm: 0.496096 +Steps: 0%| | 1814/1000000 [07:33<68:41:29, 4.04it/s, grad_norm=0.496, loss_final=0.572, loss_mean=0.859, loss_mean_cls=0.081, proj_loss=-0.369][2026-03-26 12:38:36] Step: 1814, Training Logs: loss_final: 0.576566, loss_mean: 0.857034, proj_loss: -0.362382, loss_mean_cls: 0.081914, grad_norm: 0.460126 +Steps: 0%| | 1815/1000000 [07:33<68:38:19, 4.04it/s, grad_norm=0.46, loss_final=0.577, loss_mean=0.857, loss_mean_cls=0.0819, proj_loss=-0.362][2026-03-26 12:38:37] Step: 1815, Training Logs: loss_final: 0.575068, loss_mean: 0.860304, proj_loss: -0.367421, loss_mean_cls: 0.082185, grad_norm: 0.468644 +Steps: 0%| | 1816/1000000 [07:33<68:38:34, 4.04it/s, grad_norm=0.469, loss_final=0.575, loss_mean=0.86, loss_mean_cls=0.0822, proj_loss=-0.367][2026-03-26 12:38:37] Step: 1816, Training Logs: loss_final: 0.572629, loss_mean: 0.835751, proj_loss: -0.348807, loss_mean_cls: 0.085685, grad_norm: 0.619069 +Steps: 0%| | 1817/1000000 [07:34<68:36:33, 4.04it/s, grad_norm=0.619, loss_final=0.573, loss_mean=0.836, loss_mean_cls=0.0857, proj_loss=-0.349][2026-03-26 12:38:37] Step: 1817, Training Logs: loss_final: 0.589992, loss_mean: 0.862700, proj_loss: -0.356243, loss_mean_cls: 0.083535, grad_norm: 0.548715 +Steps: 0%| | 1818/1000000 [07:34<68:38:11, 4.04it/s, grad_norm=0.549, loss_final=0.59, loss_mean=0.863, loss_mean_cls=0.0835, proj_loss=-0.356][2026-03-26 12:38:37] Step: 1818, Training Logs: loss_final: 0.565948, loss_mean: 0.847971, proj_loss: -0.364274, loss_mean_cls: 0.082250, grad_norm: 0.647637 +Steps: 0%| | 1819/1000000 [07:34<68:38:41, 4.04it/s, grad_norm=0.648, loss_final=0.566, loss_mean=0.848, loss_mean_cls=0.0823, proj_loss=-0.364][2026-03-26 12:38:37] Step: 1819, Training Logs: loss_final: 0.575984, loss_mean: 0.853121, proj_loss: -0.359499, loss_mean_cls: 0.082361, grad_norm: 0.488441 +Steps: 0%| | 1820/1000000 [07:34<68:43:00, 4.03it/s, grad_norm=0.488, loss_final=0.576, loss_mean=0.853, loss_mean_cls=0.0824, proj_loss=-0.359][2026-03-26 12:38:38] Step: 1820, Training Logs: loss_final: 0.583181, loss_mean: 0.859212, proj_loss: -0.359645, loss_mean_cls: 0.083614, grad_norm: 0.473263 +Steps: 0%| | 1821/1000000 [07:35<68:57:44, 4.02it/s, grad_norm=0.473, loss_final=0.583, loss_mean=0.859, loss_mean_cls=0.0836, proj_loss=-0.36][2026-03-26 12:38:38] Step: 1821, Training Logs: loss_final: 0.585284, loss_mean: 0.864423, proj_loss: -0.361308, loss_mean_cls: 0.082170, grad_norm: 0.414458 +Steps: 0%| | 1822/1000000 [07:35<68:54:45, 4.02it/s, grad_norm=0.414, loss_final=0.585, loss_mean=0.864, loss_mean_cls=0.0822, proj_loss=-0.361][2026-03-26 12:38:38] Step: 1822, Training Logs: loss_final: 0.578559, loss_mean: 0.843178, proj_loss: -0.349575, loss_mean_cls: 0.084956, grad_norm: 0.414759 +Steps: 0%| | 1823/1000000 [07:35<68:51:34, 4.03it/s, grad_norm=0.415, loss_final=0.579, loss_mean=0.843, loss_mean_cls=0.085, proj_loss=-0.35][2026-03-26 12:38:38] Step: 1823, Training Logs: loss_final: 0.593194, loss_mean: 0.874013, proj_loss: -0.363925, loss_mean_cls: 0.083106, grad_norm: 0.471716 +Steps: 0%| | 1824/1000000 [07:35<68:48:38, 4.03it/s, grad_norm=0.472, loss_final=0.593, loss_mean=0.874, loss_mean_cls=0.0831, proj_loss=-0.364][2026-03-26 12:38:39] Step: 1824, Training Logs: loss_final: 0.579441, loss_mean: 0.843127, proj_loss: -0.348245, loss_mean_cls: 0.084559, grad_norm: 0.380356 +Steps: 0%| | 1825/1000000 [07:36<68:44:46, 4.03it/s, grad_norm=0.38, loss_final=0.579, loss_mean=0.843, loss_mean_cls=0.0846, proj_loss=-0.348][2026-03-26 12:38:39] Step: 1825, Training Logs: loss_final: 0.588806, loss_mean: 0.866989, proj_loss: -0.360310, loss_mean_cls: 0.082127, grad_norm: 0.544421 +Steps: 0%| | 1826/1000000 [07:36<68:43:00, 4.03it/s, grad_norm=0.544, loss_final=0.589, loss_mean=0.867, loss_mean_cls=0.0821, proj_loss=-0.36][2026-03-26 12:38:39] Step: 1826, Training Logs: loss_final: 0.581146, loss_mean: 0.864298, proj_loss: -0.363578, loss_mean_cls: 0.080426, grad_norm: 0.575386 +Steps: 0%| | 1827/1000000 [07:36<68:39:48, 4.04it/s, grad_norm=0.575, loss_final=0.581, loss_mean=0.864, loss_mean_cls=0.0804, proj_loss=-0.364][2026-03-26 12:38:39] Step: 1827, Training Logs: loss_final: 0.580011, loss_mean: 0.866001, proj_loss: -0.367028, loss_mean_cls: 0.081038, grad_norm: 0.477554 +Steps: 0%| | 1828/1000000 [07:36<68:40:08, 4.04it/s, grad_norm=0.478, loss_final=0.58, loss_mean=0.866, loss_mean_cls=0.081, proj_loss=-0.367][2026-03-26 12:38:40] Step: 1828, Training Logs: loss_final: 0.586942, loss_mean: 0.856332, proj_loss: -0.354226, loss_mean_cls: 0.084836, grad_norm: 0.533627 +Steps: 0%| | 1829/1000000 [07:37<68:36:49, 4.04it/s, grad_norm=0.534, loss_final=0.587, loss_mean=0.856, loss_mean_cls=0.0848, proj_loss=-0.354][2026-03-26 12:38:40] Step: 1829, Training Logs: loss_final: 0.581209, loss_mean: 0.866191, proj_loss: -0.365697, loss_mean_cls: 0.080716, grad_norm: 0.392841 +Steps: 0%| | 1830/1000000 [07:37<68:37:12, 4.04it/s, grad_norm=0.393, loss_final=0.581, loss_mean=0.866, loss_mean_cls=0.0807, proj_loss=-0.366][2026-03-26 12:38:40] Step: 1830, Training Logs: loss_final: 0.589734, loss_mean: 0.875717, proj_loss: -0.366100, loss_mean_cls: 0.080117, grad_norm: 0.507280 +Steps: 0%| | 1831/1000000 [07:37<68:36:23, 4.04it/s, grad_norm=0.507, loss_final=0.59, loss_mean=0.876, loss_mean_cls=0.0801, proj_loss=-0.366][2026-03-26 12:38:40] Step: 1831, Training Logs: loss_final: 0.594649, loss_mean: 0.858640, proj_loss: -0.349067, loss_mean_cls: 0.085075, grad_norm: 0.538948 +Steps: 0%| | 1832/1000000 [07:37<68:37:26, 4.04it/s, grad_norm=0.539, loss_final=0.595, loss_mean=0.859, loss_mean_cls=0.0851, proj_loss=-0.349][2026-03-26 12:38:41] Step: 1832, Training Logs: loss_final: 0.575927, loss_mean: 0.849866, proj_loss: -0.356523, loss_mean_cls: 0.082585, grad_norm: 0.607829 +Steps: 0%| | 1833/1000000 [07:38<68:35:48, 4.04it/s, grad_norm=0.608, loss_final=0.576, loss_mean=0.85, loss_mean_cls=0.0826, proj_loss=-0.357][2026-03-26 12:38:41] Step: 1833, Training Logs: loss_final: 0.578091, loss_mean: 0.848121, proj_loss: -0.354393, loss_mean_cls: 0.084363, grad_norm: 0.427820 +Steps: 0%| | 1834/1000000 [07:38<68:37:49, 4.04it/s, grad_norm=0.428, loss_final=0.578, loss_mean=0.848, loss_mean_cls=0.0844, proj_loss=-0.354][2026-03-26 12:38:41] Step: 1834, Training Logs: loss_final: 0.571490, loss_mean: 0.837872, proj_loss: -0.350906, loss_mean_cls: 0.084524, grad_norm: 0.490600 +Steps: 0%| | 1835/1000000 [07:38<68:37:49, 4.04it/s, grad_norm=0.491, loss_final=0.571, loss_mean=0.838, loss_mean_cls=0.0845, proj_loss=-0.351][2026-03-26 12:38:41] Step: 1835, Training Logs: loss_final: 0.583543, loss_mean: 0.857675, proj_loss: -0.357602, loss_mean_cls: 0.083470, grad_norm: 0.362645 +Steps: 0%| | 1836/1000000 [07:38<68:39:35, 4.04it/s, grad_norm=0.363, loss_final=0.584, loss_mean=0.858, loss_mean_cls=0.0835, proj_loss=-0.358][2026-03-26 12:38:42] Step: 1836, Training Logs: loss_final: 0.602223, loss_mean: 0.891233, proj_loss: -0.368520, loss_mean_cls: 0.079510, grad_norm: 0.523390 +Steps: 0%| | 1837/1000000 [07:39<68:38:07, 4.04it/s, grad_norm=0.523, loss_final=0.602, loss_mean=0.891, loss_mean_cls=0.0795, proj_loss=-0.369][2026-03-26 12:38:42] Step: 1837, Training Logs: loss_final: 0.577288, loss_mean: 0.857545, proj_loss: -0.362903, loss_mean_cls: 0.082645, grad_norm: 0.467920 +Steps: 0%| | 1838/1000000 [07:39<68:38:37, 4.04it/s, grad_norm=0.468, loss_final=0.577, loss_mean=0.858, loss_mean_cls=0.0826, proj_loss=-0.363][2026-03-26 12:38:42] Step: 1838, Training Logs: loss_final: 0.578337, loss_mean: 0.859494, proj_loss: -0.362847, loss_mean_cls: 0.081691, grad_norm: 0.405950 +Steps: 0%| | 1839/1000000 [07:39<68:40:45, 4.04it/s, grad_norm=0.406, loss_final=0.578, loss_mean=0.859, loss_mean_cls=0.0817, proj_loss=-0.363][2026-03-26 12:38:42] Step: 1839, Training Logs: loss_final: 0.569275, loss_mean: 0.842293, proj_loss: -0.356276, loss_mean_cls: 0.083257, grad_norm: 0.575616 +Steps: 0%| | 1840/1000000 [07:39<68:42:33, 4.04it/s, grad_norm=0.576, loss_final=0.569, loss_mean=0.842, loss_mean_cls=0.0833, proj_loss=-0.356][2026-03-26 12:38:43] Step: 1840, Training Logs: loss_final: 0.582680, loss_mean: 0.858731, proj_loss: -0.359637, loss_mean_cls: 0.083587, grad_norm: 0.501760 +Steps: 0%| | 1841/1000000 [07:40<68:41:34, 4.04it/s, grad_norm=0.502, loss_final=0.583, loss_mean=0.859, loss_mean_cls=0.0836, proj_loss=-0.36][2026-03-26 12:38:43] Step: 1841, Training Logs: loss_final: 0.570756, loss_mean: 0.851727, proj_loss: -0.363533, loss_mean_cls: 0.082563, grad_norm: 0.592511 +Steps: 0%| | 1842/1000000 [07:40<68:41:05, 4.04it/s, grad_norm=0.593, loss_final=0.571, loss_mean=0.852, loss_mean_cls=0.0826, proj_loss=-0.364][2026-03-26 12:38:43] Step: 1842, Training Logs: loss_final: 0.589703, loss_mean: 0.877344, proj_loss: -0.369008, loss_mean_cls: 0.081367, grad_norm: 0.476338 +Steps: 0%| | 1843/1000000 [07:40<68:38:36, 4.04it/s, grad_norm=0.476, loss_final=0.59, loss_mean=0.877, loss_mean_cls=0.0814, proj_loss=-0.369][2026-03-26 12:38:43] Step: 1843, Training Logs: loss_final: 0.582786, loss_mean: 0.855534, proj_loss: -0.356444, loss_mean_cls: 0.083696, grad_norm: 0.639710 +Steps: 0%| | 1844/1000000 [07:40<68:38:14, 4.04it/s, grad_norm=0.64, loss_final=0.583, loss_mean=0.856, loss_mean_cls=0.0837, proj_loss=-0.356][2026-03-26 12:38:44] Step: 1844, Training Logs: loss_final: 0.574279, loss_mean: 0.844028, proj_loss: -0.354421, loss_mean_cls: 0.084671, grad_norm: 0.527502 +Steps: 0%| | 1845/1000000 [07:41<68:36:33, 4.04it/s, grad_norm=0.528, loss_final=0.574, loss_mean=0.844, loss_mean_cls=0.0847, proj_loss=-0.354][2026-03-26 12:38:44] Step: 1845, Training Logs: loss_final: 0.578336, loss_mean: 0.851448, proj_loss: -0.357149, loss_mean_cls: 0.084037, grad_norm: 0.481774 +Steps: 0%| | 1846/1000000 [07:41<68:36:04, 4.04it/s, grad_norm=0.482, loss_final=0.578, loss_mean=0.851, loss_mean_cls=0.084, proj_loss=-0.357][2026-03-26 12:38:44] Step: 1846, Training Logs: loss_final: 0.576896, loss_mean: 0.863523, proj_loss: -0.368687, loss_mean_cls: 0.082060, grad_norm: 0.540372 +Steps: 0%| | 1847/1000000 [07:41<68:55:10, 4.02it/s, grad_norm=0.54, loss_final=0.577, loss_mean=0.864, loss_mean_cls=0.0821, proj_loss=-0.369][2026-03-26 12:38:44] Step: 1847, Training Logs: loss_final: 0.578330, loss_mean: 0.841338, proj_loss: -0.348444, loss_mean_cls: 0.085437, grad_norm: 0.427541 +Steps: 0%| | 1848/1000000 [07:41<68:50:48, 4.03it/s, grad_norm=0.428, loss_final=0.578, loss_mean=0.841, loss_mean_cls=0.0854, proj_loss=-0.348][2026-03-26 12:38:45] Step: 1848, Training Logs: loss_final: 0.578747, loss_mean: 0.861863, proj_loss: -0.365089, loss_mean_cls: 0.081973, grad_norm: 0.413046 +Steps: 0%| | 1849/1000000 [07:42<68:46:44, 4.03it/s, grad_norm=0.413, loss_final=0.579, loss_mean=0.862, loss_mean_cls=0.082, proj_loss=-0.365][2026-03-26 12:38:45] Step: 1849, Training Logs: loss_final: 0.568753, loss_mean: 0.855842, proj_loss: -0.369548, loss_mean_cls: 0.082459, grad_norm: 0.524891 +Steps: 0%| | 1850/1000000 [07:42<68:41:50, 4.04it/s, grad_norm=0.525, loss_final=0.569, loss_mean=0.856, loss_mean_cls=0.0825, proj_loss=-0.37][2026-03-26 12:38:45] Step: 1850, Training Logs: loss_final: 0.585933, loss_mean: 0.868609, proj_loss: -0.365362, loss_mean_cls: 0.082686, grad_norm: 0.402255 +Steps: 0%| | 1851/1000000 [07:42<68:41:12, 4.04it/s, grad_norm=0.402, loss_final=0.586, loss_mean=0.869, loss_mean_cls=0.0827, proj_loss=-0.365][2026-03-26 12:38:45] Step: 1851, Training Logs: loss_final: 0.567337, loss_mean: 0.832348, proj_loss: -0.350377, loss_mean_cls: 0.085366, grad_norm: 0.624081 +Steps: 0%| | 1852/1000000 [07:42<68:37:18, 4.04it/s, grad_norm=0.624, loss_final=0.567, loss_mean=0.832, loss_mean_cls=0.0854, proj_loss=-0.35][2026-03-26 12:38:46] Step: 1852, Training Logs: loss_final: 0.581302, loss_mean: 0.853609, proj_loss: -0.356288, loss_mean_cls: 0.083980, grad_norm: 0.517375 +Steps: 0%| | 1853/1000000 [07:43<68:37:10, 4.04it/s, grad_norm=0.517, loss_final=0.581, loss_mean=0.854, loss_mean_cls=0.084, proj_loss=-0.356][2026-03-26 12:38:46] Step: 1853, Training Logs: loss_final: 0.570854, loss_mean: 0.857468, proj_loss: -0.367296, loss_mean_cls: 0.080682, grad_norm: 0.606730 +Steps: 0%| | 1854/1000000 [07:43<68:36:00, 4.04it/s, grad_norm=0.607, loss_final=0.571, loss_mean=0.857, loss_mean_cls=0.0807, proj_loss=-0.367][2026-03-26 12:38:46] Step: 1854, Training Logs: loss_final: 0.587593, loss_mean: 0.848141, proj_loss: -0.345942, loss_mean_cls: 0.085394, grad_norm: 0.400885 +Steps: 0%| | 1855/1000000 [07:43<68:35:33, 4.04it/s, grad_norm=0.401, loss_final=0.588, loss_mean=0.848, loss_mean_cls=0.0854, proj_loss=-0.346][2026-03-26 12:38:46] Step: 1855, Training Logs: loss_final: 0.582152, loss_mean: 0.860213, proj_loss: -0.360150, loss_mean_cls: 0.082090, grad_norm: 0.416737 +Steps: 0%| | 1856/1000000 [07:43<68:37:07, 4.04it/s, grad_norm=0.417, loss_final=0.582, loss_mean=0.86, loss_mean_cls=0.0821, proj_loss=-0.36][2026-03-26 12:38:47] Step: 1856, Training Logs: loss_final: 0.577451, loss_mean: 0.856321, proj_loss: -0.360760, loss_mean_cls: 0.081890, grad_norm: 0.430064 +Steps: 0%| | 1857/1000000 [07:44<68:37:36, 4.04it/s, grad_norm=0.43, loss_final=0.577, loss_mean=0.856, loss_mean_cls=0.0819, proj_loss=-0.361][2026-03-26 12:38:47] Step: 1857, Training Logs: loss_final: 0.582177, loss_mean: 0.855872, proj_loss: -0.356314, loss_mean_cls: 0.082619, grad_norm: 0.583636 +Steps: 0%| | 1858/1000000 [07:44<68:38:40, 4.04it/s, grad_norm=0.584, loss_final=0.582, loss_mean=0.856, loss_mean_cls=0.0826, proj_loss=-0.356][2026-03-26 12:38:47] Step: 1858, Training Logs: loss_final: 0.572606, loss_mean: 0.838969, proj_loss: -0.350830, loss_mean_cls: 0.084467, grad_norm: 0.401483 +Steps: 0%| | 1859/1000000 [07:44<68:40:08, 4.04it/s, grad_norm=0.401, loss_final=0.573, loss_mean=0.839, loss_mean_cls=0.0845, proj_loss=-0.351][2026-03-26 12:38:47] Step: 1859, Training Logs: loss_final: 0.584573, loss_mean: 0.866974, proj_loss: -0.362859, loss_mean_cls: 0.080457, grad_norm: 0.605967 +Steps: 0%| | 1860/1000000 [07:44<68:40:21, 4.04it/s, grad_norm=0.606, loss_final=0.585, loss_mean=0.867, loss_mean_cls=0.0805, proj_loss=-0.363][2026-03-26 12:38:48] Step: 1860, Training Logs: loss_final: 0.583825, loss_mean: 0.868508, proj_loss: -0.365982, loss_mean_cls: 0.081299, grad_norm: 0.580197 +Steps: 0%| | 1861/1000000 [07:45<68:39:19, 4.04it/s, grad_norm=0.58, loss_final=0.584, loss_mean=0.869, loss_mean_cls=0.0813, proj_loss=-0.366][2026-03-26 12:38:48] Step: 1861, Training Logs: loss_final: 0.583153, loss_mean: 0.871523, proj_loss: -0.368517, loss_mean_cls: 0.080146, grad_norm: 0.528065 +Steps: 0%| | 1862/1000000 [07:45<68:38:46, 4.04it/s, grad_norm=0.528, loss_final=0.583, loss_mean=0.872, loss_mean_cls=0.0801, proj_loss=-0.369][2026-03-26 12:38:48] Step: 1862, Training Logs: loss_final: 0.564095, loss_mean: 0.832066, proj_loss: -0.351916, loss_mean_cls: 0.083945, grad_norm: 0.691989 +Steps: 0%| | 1863/1000000 [07:45<68:39:08, 4.04it/s, grad_norm=0.692, loss_final=0.564, loss_mean=0.832, loss_mean_cls=0.0839, proj_loss=-0.352][2026-03-26 12:38:48] Step: 1863, Training Logs: loss_final: 0.568270, loss_mean: 0.857961, proj_loss: -0.370997, loss_mean_cls: 0.081307, grad_norm: 0.355872 +Steps: 0%| | 1864/1000000 [07:45<68:42:12, 4.04it/s, grad_norm=0.356, loss_final=0.568, loss_mean=0.858, loss_mean_cls=0.0813, proj_loss=-0.371][2026-03-26 12:38:49] Step: 1864, Training Logs: loss_final: 0.580519, loss_mean: 0.854758, proj_loss: -0.357924, loss_mean_cls: 0.083685, grad_norm: 0.608448 +Steps: 0%| | 1865/1000000 [07:46<68:41:05, 4.04it/s, grad_norm=0.608, loss_final=0.581, loss_mean=0.855, loss_mean_cls=0.0837, proj_loss=-0.358][2026-03-26 12:38:49] Step: 1865, Training Logs: loss_final: 0.587293, loss_mean: 0.865218, proj_loss: -0.361288, loss_mean_cls: 0.083363, grad_norm: 0.415520 +Steps: 0%| | 1866/1000000 [07:46<68:39:59, 4.04it/s, grad_norm=0.416, loss_final=0.587, loss_mean=0.865, loss_mean_cls=0.0834, proj_loss=-0.361][2026-03-26 12:38:49] Step: 1866, Training Logs: loss_final: 0.581208, loss_mean: 0.860915, proj_loss: -0.361757, loss_mean_cls: 0.082050, grad_norm: 0.577302 +Steps: 0%| | 1867/1000000 [07:46<68:38:01, 4.04it/s, grad_norm=0.577, loss_final=0.581, loss_mean=0.861, loss_mean_cls=0.0821, proj_loss=-0.362][2026-03-26 12:38:49] Step: 1867, Training Logs: loss_final: 0.581870, loss_mean: 0.844104, proj_loss: -0.347843, loss_mean_cls: 0.085609, grad_norm: 0.427197 +Steps: 0%| | 1868/1000000 [07:46<68:40:30, 4.04it/s, grad_norm=0.427, loss_final=0.582, loss_mean=0.844, loss_mean_cls=0.0856, proj_loss=-0.348][2026-03-26 12:38:50] Step: 1868, Training Logs: loss_final: 0.592528, loss_mean: 0.870317, proj_loss: -0.361494, loss_mean_cls: 0.083705, grad_norm: 0.551375 +Steps: 0%| | 1869/1000000 [07:46<68:38:52, 4.04it/s, grad_norm=0.551, loss_final=0.593, loss_mean=0.87, loss_mean_cls=0.0837, proj_loss=-0.361][2026-03-26 12:38:50] Step: 1869, Training Logs: loss_final: 0.582222, loss_mean: 0.874203, proj_loss: -0.371960, loss_mean_cls: 0.079980, grad_norm: 0.412210 +Steps: 0%| | 1870/1000000 [07:47<68:43:47, 4.03it/s, grad_norm=0.412, loss_final=0.582, loss_mean=0.874, loss_mean_cls=0.08, proj_loss=-0.372][2026-03-26 12:38:50] Step: 1870, Training Logs: loss_final: 0.555652, loss_mean: 0.839990, proj_loss: -0.367503, loss_mean_cls: 0.083165, grad_norm: 0.601584 +Steps: 0%| | 1871/1000000 [07:47<68:38:37, 4.04it/s, grad_norm=0.602, loss_final=0.556, loss_mean=0.84, loss_mean_cls=0.0832, proj_loss=-0.368][2026-03-26 12:38:50] Step: 1871, Training Logs: loss_final: 0.574957, loss_mean: 0.859390, proj_loss: -0.366637, loss_mean_cls: 0.082204, grad_norm: 0.692490 +Steps: 0%| | 1872/1000000 [07:47<68:41:04, 4.04it/s, grad_norm=0.692, loss_final=0.575, loss_mean=0.859, loss_mean_cls=0.0822, proj_loss=-0.367][2026-03-26 12:38:51] Step: 1872, Training Logs: loss_final: 0.576036, loss_mean: 0.858923, proj_loss: -0.364052, loss_mean_cls: 0.081165, grad_norm: 0.384323 +Steps: 0%| | 1873/1000000 [07:47<68:39:46, 4.04it/s, grad_norm=0.384, loss_final=0.576, loss_mean=0.859, loss_mean_cls=0.0812, proj_loss=-0.364][2026-03-26 12:38:51] Step: 1873, Training Logs: loss_final: 0.578503, loss_mean: 0.845640, proj_loss: -0.353331, loss_mean_cls: 0.086195, grad_norm: 0.578715 +Steps: 0%| | 1874/1000000 [07:48<68:40:15, 4.04it/s, grad_norm=0.579, loss_final=0.579, loss_mean=0.846, loss_mean_cls=0.0862, proj_loss=-0.353][2026-03-26 12:38:51] Step: 1874, Training Logs: loss_final: 0.579428, loss_mean: 0.849629, proj_loss: -0.354972, loss_mean_cls: 0.084771, grad_norm: 0.476636 +Steps: 0%| | 1875/1000000 [07:48<68:38:24, 4.04it/s, grad_norm=0.477, loss_final=0.579, loss_mean=0.85, loss_mean_cls=0.0848, proj_loss=-0.355][2026-03-26 12:38:51] Step: 1875, Training Logs: loss_final: 0.593170, loss_mean: 0.872085, proj_loss: -0.360138, loss_mean_cls: 0.081223, grad_norm: 0.590142 +Steps: 0%| | 1876/1000000 [07:48<68:37:45, 4.04it/s, grad_norm=0.59, loss_final=0.593, loss_mean=0.872, loss_mean_cls=0.0812, proj_loss=-0.36][2026-03-26 12:38:52] Step: 1876, Training Logs: loss_final: 0.591999, loss_mean: 0.867889, proj_loss: -0.358577, loss_mean_cls: 0.082687, grad_norm: 0.513656 +Steps: 0%| | 1877/1000000 [07:48<68:38:08, 4.04it/s, grad_norm=0.514, loss_final=0.592, loss_mean=0.868, loss_mean_cls=0.0827, proj_loss=-0.359][2026-03-26 12:38:52] Step: 1877, Training Logs: loss_final: 0.575925, loss_mean: 0.855221, proj_loss: -0.362828, loss_mean_cls: 0.083532, grad_norm: 0.499101 +Steps: 0%| | 1878/1000000 [07:49<68:39:42, 4.04it/s, grad_norm=0.499, loss_final=0.576, loss_mean=0.855, loss_mean_cls=0.0835, proj_loss=-0.363][2026-03-26 12:38:52] Step: 1878, Training Logs: loss_final: 0.566650, loss_mean: 0.854620, proj_loss: -0.369601, loss_mean_cls: 0.081631, grad_norm: 0.541187 +Steps: 0%| | 1879/1000000 [07:49<68:39:48, 4.04it/s, grad_norm=0.541, loss_final=0.567, loss_mean=0.855, loss_mean_cls=0.0816, proj_loss=-0.37][2026-03-26 12:38:52] Step: 1879, Training Logs: loss_final: 0.567756, loss_mean: 0.849183, proj_loss: -0.363524, loss_mean_cls: 0.082097, grad_norm: 0.509473 +Steps: 0%| | 1880/1000000 [07:49<68:39:07, 4.04it/s, grad_norm=0.509, loss_final=0.568, loss_mean=0.849, loss_mean_cls=0.0821, proj_loss=-0.364][2026-03-26 12:38:53] Step: 1880, Training Logs: loss_final: 0.574385, loss_mean: 0.845203, proj_loss: -0.354866, loss_mean_cls: 0.084047, grad_norm: 0.575782 +Steps: 0%| | 1881/1000000 [07:49<68:38:45, 4.04it/s, grad_norm=0.576, loss_final=0.574, loss_mean=0.845, loss_mean_cls=0.084, proj_loss=-0.355][2026-03-26 12:38:53] Step: 1881, Training Logs: loss_final: 0.582891, loss_mean: 0.876129, proj_loss: -0.373458, loss_mean_cls: 0.080220, grad_norm: 0.634291 +Steps: 0%| | 1882/1000000 [07:50<68:37:29, 4.04it/s, grad_norm=0.634, loss_final=0.583, loss_mean=0.876, loss_mean_cls=0.0802, proj_loss=-0.373][2026-03-26 12:38:53] Step: 1882, Training Logs: loss_final: 0.581242, loss_mean: 0.875496, proj_loss: -0.374429, loss_mean_cls: 0.080175, grad_norm: 0.455574 +Steps: 0%| | 1883/1000000 [07:50<68:39:50, 4.04it/s, grad_norm=0.456, loss_final=0.581, loss_mean=0.875, loss_mean_cls=0.0802, proj_loss=-0.374][2026-03-26 12:38:53] Step: 1883, Training Logs: loss_final: 0.578015, loss_mean: 0.848054, proj_loss: -0.353573, loss_mean_cls: 0.083535, grad_norm: 0.627949 +Steps: 0%| | 1884/1000000 [07:50<68:39:57, 4.04it/s, grad_norm=0.628, loss_final=0.578, loss_mean=0.848, loss_mean_cls=0.0835, proj_loss=-0.354][2026-03-26 12:38:54] Step: 1884, Training Logs: loss_final: 0.573014, loss_mean: 0.853811, proj_loss: -0.363403, loss_mean_cls: 0.082606, grad_norm: 0.599046 +Steps: 0%| | 1885/1000000 [07:50<68:43:46, 4.03it/s, grad_norm=0.599, loss_final=0.573, loss_mean=0.854, loss_mean_cls=0.0826, proj_loss=-0.363][2026-03-26 12:38:54] Step: 1885, Training Logs: loss_final: 0.576097, loss_mean: 0.855763, proj_loss: -0.363261, loss_mean_cls: 0.083594, grad_norm: 0.377972 +Steps: 0%| | 1886/1000000 [07:51<68:42:49, 4.03it/s, grad_norm=0.378, loss_final=0.576, loss_mean=0.856, loss_mean_cls=0.0836, proj_loss=-0.363][2026-03-26 12:38:54] Step: 1886, Training Logs: loss_final: 0.582523, loss_mean: 0.852691, proj_loss: -0.354280, loss_mean_cls: 0.084113, grad_norm: 0.507680 +Steps: 0%| | 1887/1000000 [07:51<68:43:40, 4.03it/s, grad_norm=0.508, loss_final=0.583, loss_mean=0.853, loss_mean_cls=0.0841, proj_loss=-0.354][2026-03-26 12:38:54] Step: 1887, Training Logs: loss_final: 0.561538, loss_mean: 0.850321, proj_loss: -0.370357, loss_mean_cls: 0.081574, grad_norm: 0.490142 +Steps: 0%| | 1888/1000000 [07:51<68:41:12, 4.04it/s, grad_norm=0.49, loss_final=0.562, loss_mean=0.85, loss_mean_cls=0.0816, proj_loss=-0.37][2026-03-26 12:38:55] Step: 1888, Training Logs: loss_final: 0.570025, loss_mean: 0.836947, proj_loss: -0.353649, loss_mean_cls: 0.086727, grad_norm: 0.442602 +Steps: 0%| | 1889/1000000 [07:51<68:43:26, 4.03it/s, grad_norm=0.443, loss_final=0.57, loss_mean=0.837, loss_mean_cls=0.0867, proj_loss=-0.354][2026-03-26 12:38:55] Step: 1889, Training Logs: loss_final: 0.573393, loss_mean: 0.843844, proj_loss: -0.355374, loss_mean_cls: 0.084923, grad_norm: 0.460525 +Steps: 0%| | 1890/1000000 [07:52<68:39:45, 4.04it/s, grad_norm=0.461, loss_final=0.573, loss_mean=0.844, loss_mean_cls=0.0849, proj_loss=-0.355][2026-03-26 12:38:55] Step: 1890, Training Logs: loss_final: 0.583964, loss_mean: 0.867645, proj_loss: -0.366301, loss_mean_cls: 0.082620, grad_norm: 0.358659 +Steps: 0%| | 1891/1000000 [07:52<68:41:14, 4.04it/s, grad_norm=0.359, loss_final=0.584, loss_mean=0.868, loss_mean_cls=0.0826, proj_loss=-0.366][2026-03-26 12:38:55] Step: 1891, Training Logs: loss_final: 0.569724, loss_mean: 0.861106, proj_loss: -0.371978, loss_mean_cls: 0.080596, grad_norm: 0.584031 +Steps: 0%| | 1892/1000000 [07:52<68:39:41, 4.04it/s, grad_norm=0.584, loss_final=0.57, loss_mean=0.861, loss_mean_cls=0.0806, proj_loss=-0.372][2026-03-26 12:38:56] Step: 1892, Training Logs: loss_final: 0.581500, loss_mean: 0.864453, proj_loss: -0.366013, loss_mean_cls: 0.083059, grad_norm: 0.637432 +Steps: 0%| | 1893/1000000 [07:52<68:40:22, 4.04it/s, grad_norm=0.637, loss_final=0.582, loss_mean=0.864, loss_mean_cls=0.0831, proj_loss=-0.366][2026-03-26 12:38:56] Step: 1893, Training Logs: loss_final: 0.565708, loss_mean: 0.874925, proj_loss: -0.386653, loss_mean_cls: 0.077436, grad_norm: 0.518797 +Steps: 0%| | 1894/1000000 [07:53<68:38:29, 4.04it/s, grad_norm=0.519, loss_final=0.566, loss_mean=0.875, loss_mean_cls=0.0774, proj_loss=-0.387][2026-03-26 12:38:56] Step: 1894, Training Logs: loss_final: 0.568270, loss_mean: 0.851732, proj_loss: -0.365135, loss_mean_cls: 0.081673, grad_norm: 0.500940 +Steps: 0%| | 1895/1000000 [07:53<68:41:00, 4.04it/s, grad_norm=0.501, loss_final=0.568, loss_mean=0.852, loss_mean_cls=0.0817, proj_loss=-0.365][2026-03-26 12:38:56] Step: 1895, Training Logs: loss_final: 0.578172, loss_mean: 0.864356, proj_loss: -0.367632, loss_mean_cls: 0.081449, grad_norm: 0.387241 +Steps: 0%| | 1896/1000000 [07:53<68:40:53, 4.04it/s, grad_norm=0.387, loss_final=0.578, loss_mean=0.864, loss_mean_cls=0.0814, proj_loss=-0.368][2026-03-26 12:38:57] Step: 1896, Training Logs: loss_final: 0.567843, loss_mean: 0.846952, proj_loss: -0.362489, loss_mean_cls: 0.083380, grad_norm: 0.497951 +Steps: 0%| | 1897/1000000 [07:53<68:38:07, 4.04it/s, grad_norm=0.498, loss_final=0.568, loss_mean=0.847, loss_mean_cls=0.0834, proj_loss=-0.362][2026-03-26 12:38:57] Step: 1897, Training Logs: loss_final: 0.564011, loss_mean: 0.850677, proj_loss: -0.369038, loss_mean_cls: 0.082372, grad_norm: 0.346225 +Steps: 0%| | 1898/1000000 [07:54<68:37:52, 4.04it/s, grad_norm=0.346, loss_final=0.564, loss_mean=0.851, loss_mean_cls=0.0824, proj_loss=-0.369][2026-03-26 12:38:57] Step: 1898, Training Logs: loss_final: 0.570941, loss_mean: 0.850087, proj_loss: -0.362778, loss_mean_cls: 0.083632, grad_norm: 0.448455 +Steps: 0%| | 1899/1000000 [07:54<68:38:37, 4.04it/s, grad_norm=0.448, loss_final=0.571, loss_mean=0.85, loss_mean_cls=0.0836, proj_loss=-0.363][2026-03-26 12:38:57] Step: 1899, Training Logs: loss_final: 0.579104, loss_mean: 0.864926, proj_loss: -0.367475, loss_mean_cls: 0.081653, grad_norm: 0.442193 +Steps: 0%| | 1900/1000000 [07:54<68:38:46, 4.04it/s, grad_norm=0.442, loss_final=0.579, loss_mean=0.865, loss_mean_cls=0.0817, proj_loss=-0.367][2026-03-26 12:38:58] Step: 1900, Training Logs: loss_final: 0.576613, loss_mean: 0.863287, proj_loss: -0.367764, loss_mean_cls: 0.081090, grad_norm: 0.364181 +Steps: 0%| | 1901/1000000 [07:54<68:40:26, 4.04it/s, grad_norm=0.364, loss_final=0.577, loss_mean=0.863, loss_mean_cls=0.0811, proj_loss=-0.368][2026-03-26 12:38:58] Step: 1901, Training Logs: loss_final: 0.572874, loss_mean: 0.844282, proj_loss: -0.357246, loss_mean_cls: 0.085838, grad_norm: 0.409986 +Steps: 0%| | 1902/1000000 [07:55<68:41:09, 4.04it/s, grad_norm=0.41, loss_final=0.573, loss_mean=0.844, loss_mean_cls=0.0858, proj_loss=-0.357][2026-03-26 12:38:58] Step: 1902, Training Logs: loss_final: 0.557568, loss_mean: 0.832371, proj_loss: -0.358876, loss_mean_cls: 0.084074, grad_norm: 0.488750 +Steps: 0%| | 1903/1000000 [07:55<68:38:31, 4.04it/s, grad_norm=0.489, loss_final=0.558, loss_mean=0.832, loss_mean_cls=0.0841, proj_loss=-0.359][2026-03-26 12:38:58] Step: 1903, Training Logs: loss_final: 0.565361, loss_mean: 0.845711, proj_loss: -0.364161, loss_mean_cls: 0.083811, grad_norm: 0.396198 +Steps: 0%| | 1904/1000000 [07:55<68:40:41, 4.04it/s, grad_norm=0.396, loss_final=0.565, loss_mean=0.846, loss_mean_cls=0.0838, proj_loss=-0.364][2026-03-26 12:38:59] Step: 1904, Training Logs: loss_final: 0.587782, loss_mean: 0.861521, proj_loss: -0.357855, loss_mean_cls: 0.084116, grad_norm: 0.492589 +Steps: 0%| | 1905/1000000 [07:55<68:40:37, 4.04it/s, grad_norm=0.493, loss_final=0.588, loss_mean=0.862, loss_mean_cls=0.0841, proj_loss=-0.358][2026-03-26 12:38:59] Step: 1905, Training Logs: loss_final: 0.563761, loss_mean: 0.850202, proj_loss: -0.367850, loss_mean_cls: 0.081409, grad_norm: 0.709466 +Steps: 0%| | 1906/1000000 [07:56<68:43:32, 4.03it/s, grad_norm=0.709, loss_final=0.564, loss_mean=0.85, loss_mean_cls=0.0814, proj_loss=-0.368][2026-03-26 12:38:59] Step: 1906, Training Logs: loss_final: 0.581840, loss_mean: 0.863757, proj_loss: -0.363501, loss_mean_cls: 0.081584, grad_norm: 0.810309 +Steps: 0%| | 1907/1000000 [07:56<68:40:26, 4.04it/s, grad_norm=0.81, loss_final=0.582, loss_mean=0.864, loss_mean_cls=0.0816, proj_loss=-0.364][2026-03-26 12:38:59] Step: 1907, Training Logs: loss_final: 0.566280, loss_mean: 0.871105, proj_loss: -0.382861, loss_mean_cls: 0.078037, grad_norm: 0.425915 +Steps: 0%| | 1908/1000000 [07:56<68:42:41, 4.03it/s, grad_norm=0.426, loss_final=0.566, loss_mean=0.871, loss_mean_cls=0.078, proj_loss=-0.383][2026-03-26 12:39:00] Step: 1908, Training Logs: loss_final: 0.585517, loss_mean: 0.865925, proj_loss: -0.362549, loss_mean_cls: 0.082141, grad_norm: 0.358721 +Steps: 0%| | 1909/1000000 [07:56<68:41:54, 4.04it/s, grad_norm=0.359, loss_final=0.586, loss_mean=0.866, loss_mean_cls=0.0821, proj_loss=-0.363][2026-03-26 12:39:00] Step: 1909, Training Logs: loss_final: 0.565593, loss_mean: 0.839453, proj_loss: -0.358282, loss_mean_cls: 0.084422, grad_norm: 0.450355 +Steps: 0%| | 1910/1000000 [07:57<68:40:11, 4.04it/s, grad_norm=0.45, loss_final=0.566, loss_mean=0.839, loss_mean_cls=0.0844, proj_loss=-0.358][2026-03-26 12:39:00] Step: 1910, Training Logs: loss_final: 0.564374, loss_mean: 0.852392, proj_loss: -0.370579, loss_mean_cls: 0.082562, grad_norm: 0.403424 +Steps: 0%| | 1911/1000000 [07:57<68:37:18, 4.04it/s, grad_norm=0.403, loss_final=0.564, loss_mean=0.852, loss_mean_cls=0.0826, proj_loss=-0.371][2026-03-26 12:39:00] Step: 1911, Training Logs: loss_final: 0.571041, loss_mean: 0.841236, proj_loss: -0.354969, loss_mean_cls: 0.084774, grad_norm: 0.411674 +Steps: 0%| | 1912/1000000 [07:57<68:41:13, 4.04it/s, grad_norm=0.412, loss_final=0.571, loss_mean=0.841, loss_mean_cls=0.0848, proj_loss=-0.355][2026-03-26 12:39:01] Step: 1912, Training Logs: loss_final: 0.562985, loss_mean: 0.857820, proj_loss: -0.376065, loss_mean_cls: 0.081230, grad_norm: 0.410504 +Steps: 0%| | 1913/1000000 [07:57<68:40:05, 4.04it/s, grad_norm=0.411, loss_final=0.563, loss_mean=0.858, loss_mean_cls=0.0812, proj_loss=-0.376][2026-03-26 12:39:01] Step: 1913, Training Logs: loss_final: 0.565754, loss_mean: 0.875596, proj_loss: -0.388086, loss_mean_cls: 0.078244, grad_norm: 0.415660 +Steps: 0%| | 1914/1000000 [07:58<68:39:53, 4.04it/s, grad_norm=0.416, loss_final=0.566, loss_mean=0.876, loss_mean_cls=0.0782, proj_loss=-0.388][2026-03-26 12:39:01] Step: 1914, Training Logs: loss_final: 0.570929, loss_mean: 0.851992, proj_loss: -0.363315, loss_mean_cls: 0.082252, grad_norm: 0.415302 +Steps: 0%| | 1915/1000000 [07:58<68:39:04, 4.04it/s, grad_norm=0.415, loss_final=0.571, loss_mean=0.852, loss_mean_cls=0.0823, proj_loss=-0.363][2026-03-26 12:39:01] Step: 1915, Training Logs: loss_final: 0.578365, loss_mean: 0.872385, proj_loss: -0.374273, loss_mean_cls: 0.080252, grad_norm: 0.355639 +Steps: 0%| | 1916/1000000 [07:58<68:39:08, 4.04it/s, grad_norm=0.356, loss_final=0.578, loss_mean=0.872, loss_mean_cls=0.0803, proj_loss=-0.374][2026-03-26 12:39:02] Step: 1916, Training Logs: loss_final: 0.565827, loss_mean: 0.850634, proj_loss: -0.366771, loss_mean_cls: 0.081963, grad_norm: 0.418700 +Steps: 0%| | 1917/1000000 [07:58<68:38:07, 4.04it/s, grad_norm=0.419, loss_final=0.566, loss_mean=0.851, loss_mean_cls=0.082, proj_loss=-0.367][2026-03-26 12:39:02] Step: 1917, Training Logs: loss_final: 0.574871, loss_mean: 0.864732, proj_loss: -0.371550, loss_mean_cls: 0.081689, grad_norm: 0.459947 +Steps: 0%| | 1918/1000000 [07:59<68:38:20, 4.04it/s, grad_norm=0.46, loss_final=0.575, loss_mean=0.865, loss_mean_cls=0.0817, proj_loss=-0.372][2026-03-26 12:39:02] Step: 1918, Training Logs: loss_final: 0.574689, loss_mean: 0.863509, proj_loss: -0.370517, loss_mean_cls: 0.081697, grad_norm: 0.447872 +Steps: 0%| | 1919/1000000 [07:59<68:37:54, 4.04it/s, grad_norm=0.448, loss_final=0.575, loss_mean=0.864, loss_mean_cls=0.0817, proj_loss=-0.371][2026-03-26 12:39:02] Step: 1919, Training Logs: loss_final: 0.569433, loss_mean: 0.845870, proj_loss: -0.360460, loss_mean_cls: 0.084022, grad_norm: 0.637805 +Steps: 0%| | 1920/1000000 [07:59<68:37:28, 4.04it/s, grad_norm=0.638, loss_final=0.569, loss_mean=0.846, loss_mean_cls=0.084, proj_loss=-0.36][2026-03-26 12:39:03] Step: 1920, Training Logs: loss_final: 0.582887, loss_mean: 0.858726, proj_loss: -0.360615, loss_mean_cls: 0.084777, grad_norm: 0.467352 +Steps: 0%| | 1921/1000000 [07:59<68:39:43, 4.04it/s, grad_norm=0.467, loss_final=0.583, loss_mean=0.859, loss_mean_cls=0.0848, proj_loss=-0.361][2026-03-26 12:39:03] Step: 1921, Training Logs: loss_final: 0.591813, loss_mean: 0.896290, proj_loss: -0.381260, loss_mean_cls: 0.076784, grad_norm: 0.556187 +Steps: 0%| | 1922/1000000 [08:00<68:39:40, 4.04it/s, grad_norm=0.556, loss_final=0.592, loss_mean=0.896, loss_mean_cls=0.0768, proj_loss=-0.381][2026-03-26 12:39:03] Step: 1922, Training Logs: loss_final: 0.579358, loss_mean: 0.872754, proj_loss: -0.373821, loss_mean_cls: 0.080425, grad_norm: 0.801386 +Steps: 0%| | 1923/1000000 [08:00<68:38:26, 4.04it/s, grad_norm=0.801, loss_final=0.579, loss_mean=0.873, loss_mean_cls=0.0804, proj_loss=-0.374][2026-03-26 12:39:03] Step: 1923, Training Logs: loss_final: 0.569570, loss_mean: 0.843750, proj_loss: -0.357876, loss_mean_cls: 0.083697, grad_norm: 0.599063 +Steps: 0%| | 1924/1000000 [08:00<68:37:25, 4.04it/s, grad_norm=0.599, loss_final=0.57, loss_mean=0.844, loss_mean_cls=0.0837, proj_loss=-0.358][2026-03-26 12:39:04] Step: 1924, Training Logs: loss_final: 0.572761, loss_mean: 0.854165, proj_loss: -0.363441, loss_mean_cls: 0.082036, grad_norm: 0.446547 +Steps: 0%| | 1925/1000000 [08:00<68:36:04, 4.04it/s, grad_norm=0.447, loss_final=0.573, loss_mean=0.854, loss_mean_cls=0.082, proj_loss=-0.363][2026-03-26 12:39:04] Step: 1925, Training Logs: loss_final: 0.568386, loss_mean: 0.869321, proj_loss: -0.380211, loss_mean_cls: 0.079276, grad_norm: 0.613290 +Steps: 0%| | 1926/1000000 [08:01<96:18:33, 2.88it/s, grad_norm=0.613, loss_final=0.568, loss_mean=0.869, loss_mean_cls=0.0793, proj_loss=-0.38][2026-03-26 12:39:04] Step: 1926, Training Logs: loss_final: 0.582405, loss_mean: 0.851538, proj_loss: -0.353806, loss_mean_cls: 0.084672, grad_norm: 0.486133 +Steps: 0%| | 1927/1000000 [08:01<87:58:24, 3.15it/s, grad_norm=0.486, loss_final=0.582, loss_mean=0.852, loss_mean_cls=0.0847, proj_loss=-0.354][2026-03-26 12:39:05] Step: 1927, Training Logs: loss_final: 0.583119, loss_mean: 0.870120, proj_loss: -0.368181, loss_mean_cls: 0.081180, grad_norm: 0.624738 +Steps: 0%| | 1928/1000000 [08:01<82:08:52, 3.37it/s, grad_norm=0.625, loss_final=0.583, loss_mean=0.87, loss_mean_cls=0.0812, proj_loss=-0.368][2026-03-26 12:39:05] Step: 1928, Training Logs: loss_final: 0.585531, loss_mean: 0.864324, proj_loss: -0.362758, loss_mean_cls: 0.083964, grad_norm: 0.781246 +Steps: 0%| | 1929/1000000 [08:02<78:03:10, 3.55it/s, grad_norm=0.781, loss_final=0.586, loss_mean=0.864, loss_mean_cls=0.084, proj_loss=-0.363][2026-03-26 12:39:05] Step: 1929, Training Logs: loss_final: 0.582163, loss_mean: 0.864504, proj_loss: -0.364268, loss_mean_cls: 0.081926, grad_norm: 0.518337 +Steps: 0%| | 1930/1000000 [08:02<75:13:31, 3.69it/s, grad_norm=0.518, loss_final=0.582, loss_mean=0.865, loss_mean_cls=0.0819, proj_loss=-0.364][2026-03-26 12:39:05] Step: 1930, Training Logs: loss_final: 0.572510, loss_mean: 0.844132, proj_loss: -0.358398, loss_mean_cls: 0.086775, grad_norm: 0.485615 +Steps: 0%| | 1931/1000000 [08:02<73:12:14, 3.79it/s, grad_norm=0.486, loss_final=0.573, loss_mean=0.844, loss_mean_cls=0.0868, proj_loss=-0.358][2026-03-26 12:39:06] Step: 1931, Training Logs: loss_final: 0.574103, loss_mean: 0.858025, proj_loss: -0.366771, loss_mean_cls: 0.082849, grad_norm: 0.600702 +Steps: 0%| | 1932/1000000 [08:02<71:49:30, 3.86it/s, grad_norm=0.601, loss_final=0.574, loss_mean=0.858, loss_mean_cls=0.0828, proj_loss=-0.367][2026-03-26 12:39:06] Step: 1932, Training Logs: loss_final: 0.559536, loss_mean: 0.827754, proj_loss: -0.354923, loss_mean_cls: 0.086705, grad_norm: 0.478621 +Steps: 0%| | 1933/1000000 [08:03<70:53:24, 3.91it/s, grad_norm=0.479, loss_final=0.56, loss_mean=0.828, loss_mean_cls=0.0867, proj_loss=-0.355][2026-03-26 12:39:06] Step: 1933, Training Logs: loss_final: 0.566318, loss_mean: 0.846349, proj_loss: -0.364382, loss_mean_cls: 0.084351, grad_norm: 0.451289 +Steps: 0%| | 1934/1000000 [08:03<70:12:41, 3.95it/s, grad_norm=0.451, loss_final=0.566, loss_mean=0.846, loss_mean_cls=0.0844, proj_loss=-0.364][2026-03-26 12:39:06] Step: 1934, Training Logs: loss_final: 0.574776, loss_mean: 0.856315, proj_loss: -0.364370, loss_mean_cls: 0.082831, grad_norm: 0.364257 +Steps: 0%| | 1935/1000000 [08:03<69:43:47, 3.98it/s, grad_norm=0.364, loss_final=0.575, loss_mean=0.856, loss_mean_cls=0.0828, proj_loss=-0.364][2026-03-26 12:39:07] Step: 1935, Training Logs: loss_final: 0.564803, loss_mean: 0.862745, proj_loss: -0.376966, loss_mean_cls: 0.079024, grad_norm: 0.483041 +Steps: 0%| | 1936/1000000 [08:03<69:22:16, 4.00it/s, grad_norm=0.483, loss_final=0.565, loss_mean=0.863, loss_mean_cls=0.079, proj_loss=-0.377][2026-03-26 12:39:07] Step: 1936, Training Logs: loss_final: 0.564878, loss_mean: 0.849477, proj_loss: -0.367984, loss_mean_cls: 0.083385, grad_norm: 0.496163 +Steps: 0%| | 1937/1000000 [08:04<69:52:34, 3.97it/s, grad_norm=0.496, loss_final=0.565, loss_mean=0.849, loss_mean_cls=0.0834, proj_loss=-0.368][2026-03-26 12:39:07] Step: 1937, Training Logs: loss_final: 0.562369, loss_mean: 0.826746, proj_loss: -0.351042, loss_mean_cls: 0.086664, grad_norm: 0.420049 +Steps: 0%| | 1938/1000000 [08:04<69:32:09, 3.99it/s, grad_norm=0.42, loss_final=0.562, loss_mean=0.827, loss_mean_cls=0.0867, proj_loss=-0.351][2026-03-26 12:39:07] Step: 1938, Training Logs: loss_final: 0.560635, loss_mean: 0.862059, proj_loss: -0.380709, loss_mean_cls: 0.079285, grad_norm: 0.697148 +Steps: 0%| | 1939/1000000 [08:04<69:13:48, 4.00it/s, grad_norm=0.697, loss_final=0.561, loss_mean=0.862, loss_mean_cls=0.0793, proj_loss=-0.381][2026-03-26 12:39:08] Step: 1939, Training Logs: loss_final: 0.570607, loss_mean: 0.870455, proj_loss: -0.378453, loss_mean_cls: 0.078605, grad_norm: 0.782318 +Steps: 0%| | 1940/1000000 [08:04<69:03:13, 4.01it/s, grad_norm=0.782, loss_final=0.571, loss_mean=0.87, loss_mean_cls=0.0786, proj_loss=-0.378][2026-03-26 12:39:08] Step: 1940, Training Logs: loss_final: 0.563898, loss_mean: 0.852278, proj_loss: -0.370697, loss_mean_cls: 0.082318, grad_norm: 0.527229 +Steps: 0%| | 1941/1000000 [08:05<68:55:49, 4.02it/s, grad_norm=0.527, loss_final=0.564, loss_mean=0.852, loss_mean_cls=0.0823, proj_loss=-0.371][2026-03-26 12:39:08] Step: 1941, Training Logs: loss_final: 0.581009, loss_mean: 0.872442, proj_loss: -0.372044, loss_mean_cls: 0.080610, grad_norm: 0.539068 +Steps: 0%| | 1942/1000000 [08:05<68:49:06, 4.03it/s, grad_norm=0.539, loss_final=0.581, loss_mean=0.872, loss_mean_cls=0.0806, proj_loss=-0.372][2026-03-26 12:39:08] Step: 1942, Training Logs: loss_final: 0.554792, loss_mean: 0.845454, proj_loss: -0.374065, loss_mean_cls: 0.083402, grad_norm: 0.650047 +Steps: 0%| | 1943/1000000 [08:05<68:46:54, 4.03it/s, grad_norm=0.65, loss_final=0.555, loss_mean=0.845, loss_mean_cls=0.0834, proj_loss=-0.374][2026-03-26 12:39:09] Step: 1943, Training Logs: loss_final: 0.569093, loss_mean: 0.856253, proj_loss: -0.367953, loss_mean_cls: 0.080793, grad_norm: 0.439904 +Steps: 0%| | 1944/1000000 [08:05<68:43:06, 4.03it/s, grad_norm=0.44, loss_final=0.569, loss_mean=0.856, loss_mean_cls=0.0808, proj_loss=-0.368][2026-03-26 12:39:09] Step: 1944, Training Logs: loss_final: 0.561905, loss_mean: 0.841076, proj_loss: -0.363416, loss_mean_cls: 0.084245, grad_norm: 0.602077 +Steps: 0%| | 1945/1000000 [08:06<68:42:33, 4.03it/s, grad_norm=0.602, loss_final=0.562, loss_mean=0.841, loss_mean_cls=0.0842, proj_loss=-0.363][2026-03-26 12:39:09] Step: 1945, Training Logs: loss_final: 0.586525, loss_mean: 0.871891, proj_loss: -0.366948, loss_mean_cls: 0.081583, grad_norm: 0.492539 +Steps: 0%| | 1946/1000000 [08:06<68:41:28, 4.04it/s, grad_norm=0.493, loss_final=0.587, loss_mean=0.872, loss_mean_cls=0.0816, proj_loss=-0.367][2026-03-26 12:39:09] Step: 1946, Training Logs: loss_final: 0.571811, loss_mean: 0.870541, proj_loss: -0.377592, loss_mean_cls: 0.078861, grad_norm: 0.461280 +Steps: 0%| | 1947/1000000 [08:06<68:39:46, 4.04it/s, grad_norm=0.461, loss_final=0.572, loss_mean=0.871, loss_mean_cls=0.0789, proj_loss=-0.378][2026-03-26 12:39:10] Step: 1947, Training Logs: loss_final: 0.549480, loss_mean: 0.826344, proj_loss: -0.361422, loss_mean_cls: 0.084559, grad_norm: 0.396881 +Steps: 0%| | 1948/1000000 [08:06<68:37:45, 4.04it/s, grad_norm=0.397, loss_final=0.549, loss_mean=0.826, loss_mean_cls=0.0846, proj_loss=-0.361][2026-03-26 12:39:10] Step: 1948, Training Logs: loss_final: 0.560765, loss_mean: 0.842770, proj_loss: -0.365888, loss_mean_cls: 0.083883, grad_norm: 0.372762 +Steps: 0%| | 1949/1000000 [08:07<68:40:21, 4.04it/s, grad_norm=0.373, loss_final=0.561, loss_mean=0.843, loss_mean_cls=0.0839, proj_loss=-0.366][2026-03-26 12:39:10] Step: 1949, Training Logs: loss_final: 0.563210, loss_mean: 0.835639, proj_loss: -0.356882, loss_mean_cls: 0.084453, grad_norm: 0.502711 +Steps: 0%| | 1950/1000000 [08:07<68:39:19, 4.04it/s, grad_norm=0.503, loss_final=0.563, loss_mean=0.836, loss_mean_cls=0.0845, proj_loss=-0.357][2026-03-26 12:39:10] Step: 1950, Training Logs: loss_final: 0.561651, loss_mean: 0.830235, proj_loss: -0.354248, loss_mean_cls: 0.085664, grad_norm: 0.353740 +Steps: 0%| | 1951/1000000 [08:07<68:41:23, 4.04it/s, grad_norm=0.354, loss_final=0.562, loss_mean=0.83, loss_mean_cls=0.0857, proj_loss=-0.354][2026-03-26 12:39:11] Step: 1951, Training Logs: loss_final: 0.588535, loss_mean: 0.878662, proj_loss: -0.372026, loss_mean_cls: 0.081899, grad_norm: 0.345761 +Steps: 0%| | 1952/1000000 [08:07<68:40:47, 4.04it/s, grad_norm=0.346, loss_final=0.589, loss_mean=0.879, loss_mean_cls=0.0819, proj_loss=-0.372][2026-03-26 12:39:11] Step: 1952, Training Logs: loss_final: 0.570283, loss_mean: 0.837620, proj_loss: -0.353188, loss_mean_cls: 0.085851, grad_norm: 0.591819 +Steps: 0%| | 1953/1000000 [08:08<68:39:07, 4.04it/s, grad_norm=0.592, loss_final=0.57, loss_mean=0.838, loss_mean_cls=0.0859, proj_loss=-0.353][2026-03-26 12:39:11] Step: 1953, Training Logs: loss_final: 0.571833, loss_mean: 0.844444, proj_loss: -0.357682, loss_mean_cls: 0.085070, grad_norm: 0.560144 +Steps: 0%| | 1954/1000000 [08:08<68:37:31, 4.04it/s, grad_norm=0.56, loss_final=0.572, loss_mean=0.844, loss_mean_cls=0.0851, proj_loss=-0.358][2026-03-26 12:39:11] Step: 1954, Training Logs: loss_final: 0.558718, loss_mean: 0.848452, proj_loss: -0.371706, loss_mean_cls: 0.081972, grad_norm: 0.337337 +Steps: 0%| | 1955/1000000 [08:08<68:36:27, 4.04it/s, grad_norm=0.337, loss_final=0.559, loss_mean=0.848, loss_mean_cls=0.082, proj_loss=-0.372][2026-03-26 12:39:12] Step: 1955, Training Logs: loss_final: 0.561142, loss_mean: 0.850895, proj_loss: -0.371331, loss_mean_cls: 0.081578, grad_norm: 0.372603 +Steps: 0%| | 1956/1000000 [08:08<68:35:12, 4.04it/s, grad_norm=0.373, loss_final=0.561, loss_mean=0.851, loss_mean_cls=0.0816, proj_loss=-0.371][2026-03-26 12:39:12] Step: 1956, Training Logs: loss_final: 0.564821, loss_mean: 0.835914, proj_loss: -0.356404, loss_mean_cls: 0.085311, grad_norm: 0.579141 +Steps: 0%| | 1957/1000000 [08:09<69:03:27, 4.01it/s, grad_norm=0.579, loss_final=0.565, loss_mean=0.836, loss_mean_cls=0.0853, proj_loss=-0.356][2026-03-26 12:39:12] Step: 1957, Training Logs: loss_final: 0.576884, loss_mean: 0.860776, proj_loss: -0.366172, loss_mean_cls: 0.082280, grad_norm: 0.470795 +Steps: 0%| | 1958/1000000 [08:09<68:54:14, 4.02it/s, grad_norm=0.471, loss_final=0.577, loss_mean=0.861, loss_mean_cls=0.0823, proj_loss=-0.366][2026-03-26 12:39:12] Step: 1958, Training Logs: loss_final: 0.554082, loss_mean: 0.826465, proj_loss: -0.357371, loss_mean_cls: 0.084988, grad_norm: 0.462141 +Steps: 0%| | 1959/1000000 [08:09<68:49:48, 4.03it/s, grad_norm=0.462, loss_final=0.554, loss_mean=0.826, loss_mean_cls=0.085, proj_loss=-0.357][2026-03-26 12:39:13] Step: 1959, Training Logs: loss_final: 0.554593, loss_mean: 0.839373, proj_loss: -0.368030, loss_mean_cls: 0.083249, grad_norm: 0.436647 +Steps: 0%| | 1960/1000000 [08:09<68:47:07, 4.03it/s, grad_norm=0.437, loss_final=0.555, loss_mean=0.839, loss_mean_cls=0.0832, proj_loss=-0.368][2026-03-26 12:39:13] Step: 1960, Training Logs: loss_final: 0.571854, loss_mean: 0.868327, proj_loss: -0.376340, loss_mean_cls: 0.079867, grad_norm: 0.549399 +Steps: 0%| | 1961/1000000 [08:10<68:43:35, 4.03it/s, grad_norm=0.549, loss_final=0.572, loss_mean=0.868, loss_mean_cls=0.0799, proj_loss=-0.376][2026-03-26 12:39:13] Step: 1961, Training Logs: loss_final: 0.568253, loss_mean: 0.849175, proj_loss: -0.364791, loss_mean_cls: 0.083869, grad_norm: 0.446062 +Steps: 0%| | 1962/1000000 [08:10<68:42:24, 4.04it/s, grad_norm=0.446, loss_final=0.568, loss_mean=0.849, loss_mean_cls=0.0839, proj_loss=-0.365][2026-03-26 12:39:13] Step: 1962, Training Logs: loss_final: 0.568422, loss_mean: 0.858113, proj_loss: -0.371590, loss_mean_cls: 0.081900, grad_norm: 0.365032 +Steps: 0%| | 1963/1000000 [08:10<68:42:10, 4.04it/s, grad_norm=0.365, loss_final=0.568, loss_mean=0.858, loss_mean_cls=0.0819, proj_loss=-0.372][2026-03-26 12:39:14] Step: 1963, Training Logs: loss_final: 0.578477, loss_mean: 0.852627, proj_loss: -0.359339, loss_mean_cls: 0.085189, grad_norm: 0.591511 +Steps: 0%| | 1964/1000000 [08:10<68:44:27, 4.03it/s, grad_norm=0.592, loss_final=0.578, loss_mean=0.853, loss_mean_cls=0.0852, proj_loss=-0.359][2026-03-26 12:39:14] Step: 1964, Training Logs: loss_final: 0.574149, loss_mean: 0.870317, proj_loss: -0.377603, loss_mean_cls: 0.081435, grad_norm: 0.370038 +Steps: 0%| | 1965/1000000 [08:11<68:41:30, 4.04it/s, grad_norm=0.37, loss_final=0.574, loss_mean=0.87, loss_mean_cls=0.0814, proj_loss=-0.378][2026-03-26 12:39:14] Step: 1965, Training Logs: loss_final: 0.563441, loss_mean: 0.841986, proj_loss: -0.363233, loss_mean_cls: 0.084688, grad_norm: 0.651581 +Steps: 0%| | 1966/1000000 [08:11<68:43:38, 4.03it/s, grad_norm=0.652, loss_final=0.563, loss_mean=0.842, loss_mean_cls=0.0847, proj_loss=-0.363][2026-03-26 12:39:14] Step: 1966, Training Logs: loss_final: 0.577298, loss_mean: 0.859849, proj_loss: -0.366715, loss_mean_cls: 0.084164, grad_norm: 0.439199 +Steps: 0%| | 1967/1000000 [08:11<68:40:46, 4.04it/s, grad_norm=0.439, loss_final=0.577, loss_mean=0.86, loss_mean_cls=0.0842, proj_loss=-0.367][2026-03-26 12:39:14] Step: 1967, Training Logs: loss_final: 0.568665, loss_mean: 0.847476, proj_loss: -0.362569, loss_mean_cls: 0.083757, grad_norm: 0.474437 +Steps: 0%| | 1968/1000000 [08:11<68:42:49, 4.03it/s, grad_norm=0.474, loss_final=0.569, loss_mean=0.847, loss_mean_cls=0.0838, proj_loss=-0.363][2026-03-26 12:39:15] Step: 1968, Training Logs: loss_final: 0.578196, loss_mean: 0.865761, proj_loss: -0.368990, loss_mean_cls: 0.081425, grad_norm: 0.474857 +Steps: 0%| | 1969/1000000 [08:12<68:41:12, 4.04it/s, grad_norm=0.475, loss_final=0.578, loss_mean=0.866, loss_mean_cls=0.0814, proj_loss=-0.369][2026-03-26 12:39:15] Step: 1969, Training Logs: loss_final: 0.560249, loss_mean: 0.845943, proj_loss: -0.368136, loss_mean_cls: 0.082442, grad_norm: 0.515100 +Steps: 0%| | 1970/1000000 [08:12<68:42:17, 4.04it/s, grad_norm=0.515, loss_final=0.56, loss_mean=0.846, loss_mean_cls=0.0824, proj_loss=-0.368][2026-03-26 12:39:15] Step: 1970, Training Logs: loss_final: 0.546478, loss_mean: 0.839726, proj_loss: -0.374729, loss_mean_cls: 0.081481, grad_norm: 0.450462 +Steps: 0%| | 1971/1000000 [08:12<68:41:58, 4.04it/s, grad_norm=0.45, loss_final=0.546, loss_mean=0.84, loss_mean_cls=0.0815, proj_loss=-0.375][2026-03-26 12:39:15] Step: 1971, Training Logs: loss_final: 0.579712, loss_mean: 0.864533, proj_loss: -0.367716, loss_mean_cls: 0.082894, grad_norm: 0.383640 +Steps: 0%| | 1972/1000000 [08:12<68:39:54, 4.04it/s, grad_norm=0.384, loss_final=0.58, loss_mean=0.865, loss_mean_cls=0.0829, proj_loss=-0.368][2026-03-26 12:39:16] Step: 1972, Training Logs: loss_final: 0.575440, loss_mean: 0.862008, proj_loss: -0.368679, loss_mean_cls: 0.082111, grad_norm: 0.496208 +Steps: 0%| | 1973/1000000 [08:13<68:37:18, 4.04it/s, grad_norm=0.496, loss_final=0.575, loss_mean=0.862, loss_mean_cls=0.0821, proj_loss=-0.369][2026-03-26 12:39:16] Step: 1973, Training Logs: loss_final: 0.562846, loss_mean: 0.850188, proj_loss: -0.369867, loss_mean_cls: 0.082525, grad_norm: 0.354546 +Steps: 0%| | 1974/1000000 [08:13<68:45:15, 4.03it/s, grad_norm=0.355, loss_final=0.563, loss_mean=0.85, loss_mean_cls=0.0825, proj_loss=-0.37][2026-03-26 12:39:16] Step: 1974, Training Logs: loss_final: 0.584891, loss_mean: 0.873614, proj_loss: -0.371586, loss_mean_cls: 0.082862, grad_norm: 0.460013 +Steps: 0%| | 1975/1000000 [08:13<68:40:26, 4.04it/s, grad_norm=0.46, loss_final=0.585, loss_mean=0.874, loss_mean_cls=0.0829, proj_loss=-0.372][2026-03-26 12:39:16] Step: 1975, Training Logs: loss_final: 0.578941, loss_mean: 0.875344, proj_loss: -0.377527, loss_mean_cls: 0.081125, grad_norm: 0.436725 +Steps: 0%| | 1976/1000000 [08:13<68:42:02, 4.04it/s, grad_norm=0.437, loss_final=0.579, loss_mean=0.875, loss_mean_cls=0.0811, proj_loss=-0.378][2026-03-26 12:39:17] Step: 1976, Training Logs: loss_final: 0.563753, loss_mean: 0.848445, proj_loss: -0.368204, loss_mean_cls: 0.083512, grad_norm: 0.492850 +Steps: 0%| | 1977/1000000 [08:14<68:38:49, 4.04it/s, grad_norm=0.493, loss_final=0.564, loss_mean=0.848, loss_mean_cls=0.0835, proj_loss=-0.368][2026-03-26 12:39:17] Step: 1977, Training Logs: loss_final: 0.560673, loss_mean: 0.844399, proj_loss: -0.368070, loss_mean_cls: 0.084345, grad_norm: 0.424055 +Steps: 0%| | 1978/1000000 [08:14<68:41:32, 4.04it/s, grad_norm=0.424, loss_final=0.561, loss_mean=0.844, loss_mean_cls=0.0843, proj_loss=-0.368][2026-03-26 12:39:17] Step: 1978, Training Logs: loss_final: 0.573244, loss_mean: 0.847463, proj_loss: -0.360157, loss_mean_cls: 0.085938, grad_norm: 0.458569 +Steps: 0%| | 1979/1000000 [08:14<68:38:21, 4.04it/s, grad_norm=0.459, loss_final=0.573, loss_mean=0.847, loss_mean_cls=0.0859, proj_loss=-0.36][2026-03-26 12:39:17] Step: 1979, Training Logs: loss_final: 0.554683, loss_mean: 0.842701, proj_loss: -0.370959, loss_mean_cls: 0.082941, grad_norm: 0.370802 +Steps: 0%| | 1980/1000000 [08:14<68:37:53, 4.04it/s, grad_norm=0.371, loss_final=0.555, loss_mean=0.843, loss_mean_cls=0.0829, proj_loss=-0.371][2026-03-26 12:39:18] Step: 1980, Training Logs: loss_final: 0.566072, loss_mean: 0.843459, proj_loss: -0.361912, loss_mean_cls: 0.084526, grad_norm: 0.457071 +Steps: 0%| | 1981/1000000 [08:15<68:37:55, 4.04it/s, grad_norm=0.457, loss_final=0.566, loss_mean=0.843, loss_mean_cls=0.0845, proj_loss=-0.362][2026-03-26 12:39:18] Step: 1981, Training Logs: loss_final: 0.568179, loss_mean: 0.846770, proj_loss: -0.363852, loss_mean_cls: 0.085261, grad_norm: 0.389031 +Steps: 0%| | 1982/1000000 [08:15<68:37:37, 4.04it/s, grad_norm=0.389, loss_final=0.568, loss_mean=0.847, loss_mean_cls=0.0853, proj_loss=-0.364][2026-03-26 12:39:18] Step: 1982, Training Logs: loss_final: 0.564653, loss_mean: 0.848793, proj_loss: -0.367776, loss_mean_cls: 0.083636, grad_norm: 0.552638 +Steps: 0%| | 1983/1000000 [08:15<68:36:30, 4.04it/s, grad_norm=0.553, loss_final=0.565, loss_mean=0.849, loss_mean_cls=0.0836, proj_loss=-0.368][2026-03-26 12:39:18] Step: 1983, Training Logs: loss_final: 0.581364, loss_mean: 0.872664, proj_loss: -0.373103, loss_mean_cls: 0.081804, grad_norm: 0.636463 +Steps: 0%| | 1984/1000000 [08:15<68:41:48, 4.04it/s, grad_norm=0.636, loss_final=0.581, loss_mean=0.873, loss_mean_cls=0.0818, proj_loss=-0.373][2026-03-26 12:39:19] Step: 1984, Training Logs: loss_final: 0.561449, loss_mean: 0.858190, proj_loss: -0.377898, loss_mean_cls: 0.081157, grad_norm: 0.568227 +Steps: 0%| | 1985/1000000 [08:16<68:41:50, 4.04it/s, grad_norm=0.568, loss_final=0.561, loss_mean=0.858, loss_mean_cls=0.0812, proj_loss=-0.378][2026-03-26 12:39:19] Step: 1985, Training Logs: loss_final: 0.562555, loss_mean: 0.831654, proj_loss: -0.355213, loss_mean_cls: 0.086114, grad_norm: 0.448801 +Steps: 0%| | 1986/1000000 [08:16<68:40:21, 4.04it/s, grad_norm=0.449, loss_final=0.563, loss_mean=0.832, loss_mean_cls=0.0861, proj_loss=-0.355][2026-03-26 12:39:19] Step: 1986, Training Logs: loss_final: 0.561890, loss_mean: 0.839686, proj_loss: -0.363732, loss_mean_cls: 0.085937, grad_norm: 0.631301 +Steps: 0%| | 1987/1000000 [08:16<68:41:45, 4.04it/s, grad_norm=0.631, loss_final=0.562, loss_mean=0.84, loss_mean_cls=0.0859, proj_loss=-0.364][2026-03-26 12:39:19] Step: 1987, Training Logs: loss_final: 0.573644, loss_mean: 0.867082, proj_loss: -0.375218, loss_mean_cls: 0.081780, grad_norm: 0.570668 +Steps: 0%| | 1988/1000000 [08:16<68:41:23, 4.04it/s, grad_norm=0.571, loss_final=0.574, loss_mean=0.867, loss_mean_cls=0.0818, proj_loss=-0.375][2026-03-26 12:39:20] Step: 1988, Training Logs: loss_final: 0.555399, loss_mean: 0.844178, proj_loss: -0.371178, loss_mean_cls: 0.082400, grad_norm: 0.521155 +Steps: 0%| | 1989/1000000 [08:17<68:41:19, 4.04it/s, grad_norm=0.521, loss_final=0.555, loss_mean=0.844, loss_mean_cls=0.0824, proj_loss=-0.371][2026-03-26 12:39:20] Step: 1989, Training Logs: loss_final: 0.564530, loss_mean: 0.859883, proj_loss: -0.375204, loss_mean_cls: 0.079851, grad_norm: 0.472840 +Steps: 0%| | 1990/1000000 [08:17<69:29:21, 3.99it/s, grad_norm=0.473, loss_final=0.565, loss_mean=0.86, loss_mean_cls=0.0799, proj_loss=-0.375][2026-03-26 12:39:20] Step: 1990, Training Logs: loss_final: 0.566601, loss_mean: 0.853770, proj_loss: -0.370321, loss_mean_cls: 0.083152, grad_norm: 0.491784 +Steps: 0%| | 1991/1000000 [08:17<69:14:19, 4.00it/s, grad_norm=0.492, loss_final=0.567, loss_mean=0.854, loss_mean_cls=0.0832, proj_loss=-0.37][2026-03-26 12:39:20] Step: 1991, Training Logs: loss_final: 0.564991, loss_mean: 0.846477, proj_loss: -0.365479, loss_mean_cls: 0.083993, grad_norm: 0.469916 +Steps: 0%| | 1992/1000000 [08:17<69:01:48, 4.02it/s, grad_norm=0.47, loss_final=0.565, loss_mean=0.846, loss_mean_cls=0.084, proj_loss=-0.365][2026-03-26 12:39:21] Step: 1992, Training Logs: loss_final: 0.561737, loss_mean: 0.851932, proj_loss: -0.372456, loss_mean_cls: 0.082262, grad_norm: 0.470302 +Steps: 0%| | 1993/1000000 [08:18<68:59:56, 4.02it/s, grad_norm=0.47, loss_final=0.562, loss_mean=0.852, loss_mean_cls=0.0823, proj_loss=-0.372][2026-03-26 12:39:21] Step: 1993, Training Logs: loss_final: 0.575296, loss_mean: 0.864025, proj_loss: -0.370024, loss_mean_cls: 0.081294, grad_norm: 0.461436 +Steps: 0%| | 1994/1000000 [08:18<68:56:04, 4.02it/s, grad_norm=0.461, loss_final=0.575, loss_mean=0.864, loss_mean_cls=0.0813, proj_loss=-0.37][2026-03-26 12:39:21] Step: 1994, Training Logs: loss_final: 0.565759, loss_mean: 0.858840, proj_loss: -0.374923, loss_mean_cls: 0.081842, grad_norm: 0.448403 +Steps: 0%| | 1995/1000000 [08:18<69:01:00, 4.02it/s, grad_norm=0.448, loss_final=0.566, loss_mean=0.859, loss_mean_cls=0.0818, proj_loss=-0.375][2026-03-26 12:39:21] Step: 1995, Training Logs: loss_final: 0.570103, loss_mean: 0.857012, proj_loss: -0.369231, loss_mean_cls: 0.082321, grad_norm: 0.708960 +Steps: 0%| | 1996/1000000 [08:18<68:55:35, 4.02it/s, grad_norm=0.709, loss_final=0.57, loss_mean=0.857, loss_mean_cls=0.0823, proj_loss=-0.369][2026-03-26 12:39:22] Step: 1996, Training Logs: loss_final: 0.563127, loss_mean: 0.854625, proj_loss: -0.374477, loss_mean_cls: 0.082978, grad_norm: 0.446784 +Steps: 0%| | 1997/1000000 [08:19<68:51:41, 4.03it/s, grad_norm=0.447, loss_final=0.563, loss_mean=0.855, loss_mean_cls=0.083, proj_loss=-0.374][2026-03-26 12:39:22] Step: 1997, Training Logs: loss_final: 0.561740, loss_mean: 0.853605, proj_loss: -0.372244, loss_mean_cls: 0.080380, grad_norm: 0.502007 +Steps: 0%| | 1998/1000000 [08:19<68:45:30, 4.03it/s, grad_norm=0.502, loss_final=0.562, loss_mean=0.854, loss_mean_cls=0.0804, proj_loss=-0.372][2026-03-26 12:39:22] Step: 1998, Training Logs: loss_final: 0.558658, loss_mean: 0.840586, proj_loss: -0.366341, loss_mean_cls: 0.084412, grad_norm: 0.371268 +Steps: 0%| | 1999/1000000 [08:19<68:43:42, 4.03it/s, grad_norm=0.371, loss_final=0.559, loss_mean=0.841, loss_mean_cls=0.0844, proj_loss=-0.366][2026-03-26 12:39:22] Step: 1999, Training Logs: loss_final: 0.566955, loss_mean: 0.865365, proj_loss: -0.379522, loss_mean_cls: 0.081112, grad_norm: 0.418650 +Steps: 0%| | 2000/1000000 [08:19<68:41:15, 4.04it/s, grad_norm=0.419, loss_final=0.567, loss_mean=0.865, loss_mean_cls=0.0811, proj_loss=-0.38][2026-03-26 12:39:23] Step: 2000, Training Logs: loss_final: 0.578232, loss_mean: 0.873327, proj_loss: -0.375998, loss_mean_cls: 0.080903, grad_norm: 0.399886 +Steps: 0%| | 2001/1000000 [08:20<68:40:46, 4.04it/s, grad_norm=0.4, loss_final=0.578, loss_mean=0.873, loss_mean_cls=0.0809, proj_loss=-0.376][2026-03-26 12:39:23] Step: 2001, Training Logs: loss_final: 0.572368, loss_mean: 0.866353, proj_loss: -0.375292, loss_mean_cls: 0.081306, grad_norm: 0.425930 +Steps: 0%| | 2002/1000000 [08:20<68:38:48, 4.04it/s, grad_norm=0.426, loss_final=0.572, loss_mean=0.866, loss_mean_cls=0.0813, proj_loss=-0.375][2026-03-26 12:39:23] Step: 2002, Training Logs: loss_final: 0.556757, loss_mean: 0.835569, proj_loss: -0.362729, loss_mean_cls: 0.083917, grad_norm: 0.542523 +Steps: 0%| | 2003/1000000 [08:20<68:37:25, 4.04it/s, grad_norm=0.543, loss_final=0.557, loss_mean=0.836, loss_mean_cls=0.0839, proj_loss=-0.363][2026-03-26 12:39:23] Step: 2003, Training Logs: loss_final: 0.566101, loss_mean: 0.861210, proj_loss: -0.376121, loss_mean_cls: 0.081012, grad_norm: 0.695613 +Steps: 0%| | 2004/1000000 [08:20<68:37:42, 4.04it/s, grad_norm=0.696, loss_final=0.566, loss_mean=0.861, loss_mean_cls=0.081, proj_loss=-0.376][2026-03-26 12:39:24] Step: 2004, Training Logs: loss_final: 0.569596, loss_mean: 0.858755, proj_loss: -0.372118, loss_mean_cls: 0.082960, grad_norm: 0.434280 +Steps: 0%| | 2005/1000000 [08:21<68:37:13, 4.04it/s, grad_norm=0.434, loss_final=0.57, loss_mean=0.859, loss_mean_cls=0.083, proj_loss=-0.372][2026-03-26 12:39:24] Step: 2005, Training Logs: loss_final: 0.551335, loss_mean: 0.844352, proj_loss: -0.375728, loss_mean_cls: 0.082711, grad_norm: 0.420749 +Steps: 0%| | 2006/1000000 [08:21<68:36:21, 4.04it/s, grad_norm=0.421, loss_final=0.551, loss_mean=0.844, loss_mean_cls=0.0827, proj_loss=-0.376][2026-03-26 12:39:24] Step: 2006, Training Logs: loss_final: 0.583765, loss_mean: 0.878378, proj_loss: -0.375643, loss_mean_cls: 0.081030, grad_norm: 0.444436 +Steps: 0%| | 2007/1000000 [08:21<68:36:40, 4.04it/s, grad_norm=0.444, loss_final=0.584, loss_mean=0.878, loss_mean_cls=0.081, proj_loss=-0.376][2026-03-26 12:39:24] Step: 2007, Training Logs: loss_final: 0.544415, loss_mean: 0.838710, proj_loss: -0.376815, loss_mean_cls: 0.082520, grad_norm: 0.462792 +Steps: 0%| | 2008/1000000 [08:21<68:37:16, 4.04it/s, grad_norm=0.463, loss_final=0.544, loss_mean=0.839, loss_mean_cls=0.0825, proj_loss=-0.377][2026-03-26 12:39:25] Step: 2008, Training Logs: loss_final: 0.551633, loss_mean: 0.831737, proj_loss: -0.364620, loss_mean_cls: 0.084516, grad_norm: 0.405474 +Steps: 0%| | 2009/1000000 [08:22<68:36:51, 4.04it/s, grad_norm=0.405, loss_final=0.552, loss_mean=0.832, loss_mean_cls=0.0845, proj_loss=-0.365][2026-03-26 12:39:25] Step: 2009, Training Logs: loss_final: 0.574004, loss_mean: 0.871956, proj_loss: -0.377900, loss_mean_cls: 0.079948, grad_norm: 0.406608 +Steps: 0%| | 2010/1000000 [08:22<68:44:49, 4.03it/s, grad_norm=0.407, loss_final=0.574, loss_mean=0.872, loss_mean_cls=0.0799, proj_loss=-0.378][2026-03-26 12:39:25] Step: 2010, Training Logs: loss_final: 0.566725, loss_mean: 0.838057, proj_loss: -0.357610, loss_mean_cls: 0.086278, grad_norm: 0.429462 +Steps: 0%| | 2011/1000000 [08:22<68:43:01, 4.03it/s, grad_norm=0.429, loss_final=0.567, loss_mean=0.838, loss_mean_cls=0.0863, proj_loss=-0.358][2026-03-26 12:39:25] Step: 2011, Training Logs: loss_final: 0.566830, loss_mean: 0.852873, proj_loss: -0.368921, loss_mean_cls: 0.082879, grad_norm: 0.444288 +Steps: 0%| | 2012/1000000 [08:22<68:45:22, 4.03it/s, grad_norm=0.444, loss_final=0.567, loss_mean=0.853, loss_mean_cls=0.0829, proj_loss=-0.369][2026-03-26 12:39:26] Step: 2012, Training Logs: loss_final: 0.557734, loss_mean: 0.823386, proj_loss: -0.353773, loss_mean_cls: 0.088121, grad_norm: 0.459970 +Steps: 0%| | 2013/1000000 [08:23<68:43:54, 4.03it/s, grad_norm=0.46, loss_final=0.558, loss_mean=0.823, loss_mean_cls=0.0881, proj_loss=-0.354][2026-03-26 12:39:26] Step: 2013, Training Logs: loss_final: 0.560350, loss_mean: 0.844515, proj_loss: -0.367634, loss_mean_cls: 0.083469, grad_norm: 0.457281 +Steps: 0%| | 2014/1000000 [08:23<68:43:21, 4.03it/s, grad_norm=0.457, loss_final=0.56, loss_mean=0.845, loss_mean_cls=0.0835, proj_loss=-0.368][2026-03-26 12:39:26] Step: 2014, Training Logs: loss_final: 0.571875, loss_mean: 0.859071, proj_loss: -0.370494, loss_mean_cls: 0.083298, grad_norm: 0.590092 +Steps: 0%| | 2015/1000000 [08:23<68:42:15, 4.03it/s, grad_norm=0.59, loss_final=0.572, loss_mean=0.859, loss_mean_cls=0.0833, proj_loss=-0.37][2026-03-26 12:39:26] Step: 2015, Training Logs: loss_final: 0.571962, loss_mean: 0.880183, proj_loss: -0.386963, loss_mean_cls: 0.078743, grad_norm: 0.485193 +Steps: 0%| | 2016/1000000 [08:23<68:40:02, 4.04it/s, grad_norm=0.485, loss_final=0.572, loss_mean=0.88, loss_mean_cls=0.0787, proj_loss=-0.387][2026-03-26 12:39:27] Step: 2016, Training Logs: loss_final: 0.568077, loss_mean: 0.852046, proj_loss: -0.367059, loss_mean_cls: 0.083090, grad_norm: 0.524904 +Steps: 0%| | 2017/1000000 [08:24<68:39:49, 4.04it/s, grad_norm=0.525, loss_final=0.568, loss_mean=0.852, loss_mean_cls=0.0831, proj_loss=-0.367][2026-03-26 12:39:27] Step: 2017, Training Logs: loss_final: 0.576083, loss_mean: 0.865394, proj_loss: -0.371850, loss_mean_cls: 0.082539, grad_norm: 0.478835 +Steps: 0%| | 2018/1000000 [08:24<68:39:09, 4.04it/s, grad_norm=0.479, loss_final=0.576, loss_mean=0.865, loss_mean_cls=0.0825, proj_loss=-0.372][2026-03-26 12:39:27] Step: 2018, Training Logs: loss_final: 0.559553, loss_mean: 0.838264, proj_loss: -0.363017, loss_mean_cls: 0.084306, grad_norm: 0.593362 +Steps: 0%| | 2019/1000000 [08:24<68:36:10, 4.04it/s, grad_norm=0.593, loss_final=0.56, loss_mean=0.838, loss_mean_cls=0.0843, proj_loss=-0.363][2026-03-26 12:39:27] Step: 2019, Training Logs: loss_final: 0.561520, loss_mean: 0.862724, proj_loss: -0.380123, loss_mean_cls: 0.078919, grad_norm: 0.664650 +Steps: 0%| | 2020/1000000 [08:24<68:37:41, 4.04it/s, grad_norm=0.665, loss_final=0.562, loss_mean=0.863, loss_mean_cls=0.0789, proj_loss=-0.38][2026-03-26 12:39:28] Step: 2020, Training Logs: loss_final: 0.570328, loss_mean: 0.851627, proj_loss: -0.364969, loss_mean_cls: 0.083670, grad_norm: 0.555371 +Steps: 0%| | 2021/1000000 [08:24<68:36:37, 4.04it/s, grad_norm=0.555, loss_final=0.57, loss_mean=0.852, loss_mean_cls=0.0837, proj_loss=-0.365][2026-03-26 12:39:28] Step: 2021, Training Logs: loss_final: 0.564435, loss_mean: 0.853975, proj_loss: -0.372204, loss_mean_cls: 0.082664, grad_norm: 0.443536 +Steps: 0%| | 2022/1000000 [08:25<68:38:05, 4.04it/s, grad_norm=0.444, loss_final=0.564, loss_mean=0.854, loss_mean_cls=0.0827, proj_loss=-0.372][2026-03-26 12:39:28] Step: 2022, Training Logs: loss_final: 0.569830, loss_mean: 0.845156, proj_loss: -0.360383, loss_mean_cls: 0.085057, grad_norm: 0.694447 +Steps: 0%| | 2023/1000000 [08:25<68:34:02, 4.04it/s, grad_norm=0.694, loss_final=0.57, loss_mean=0.845, loss_mean_cls=0.0851, proj_loss=-0.36][2026-03-26 12:39:28] Step: 2023, Training Logs: loss_final: 0.568308, loss_mean: 0.868713, proj_loss: -0.380757, loss_mean_cls: 0.080351, grad_norm: 0.571837 +Steps: 0%| | 2024/1000000 [08:25<68:37:21, 4.04it/s, grad_norm=0.572, loss_final=0.568, loss_mean=0.869, loss_mean_cls=0.0804, proj_loss=-0.381][2026-03-26 12:39:29] Step: 2024, Training Logs: loss_final: 0.584924, loss_mean: 0.864279, proj_loss: -0.361218, loss_mean_cls: 0.081863, grad_norm: 0.824011 +Steps: 0%| | 2025/1000000 [08:25<68:38:18, 4.04it/s, grad_norm=0.824, loss_final=0.585, loss_mean=0.864, loss_mean_cls=0.0819, proj_loss=-0.361][2026-03-26 12:39:29] Step: 2025, Training Logs: loss_final: 0.565045, loss_mean: 0.866236, proj_loss: -0.381673, loss_mean_cls: 0.080482, grad_norm: 0.417830 +Steps: 0%| | 2026/1000000 [08:26<70:57:04, 3.91it/s, grad_norm=0.418, loss_final=0.565, loss_mean=0.866, loss_mean_cls=0.0805, proj_loss=-0.382][2026-03-26 12:39:29] Step: 2026, Training Logs: loss_final: 0.551370, loss_mean: 0.836345, proj_loss: -0.370187, loss_mean_cls: 0.085211, grad_norm: 0.508687 +Steps: 0%| | 2027/1000000 [08:26<72:01:11, 3.85it/s, grad_norm=0.509, loss_final=0.551, loss_mean=0.836, loss_mean_cls=0.0852, proj_loss=-0.37][2026-03-26 12:39:29] Step: 2027, Training Logs: loss_final: 0.567082, loss_mean: 0.855899, proj_loss: -0.370893, loss_mean_cls: 0.082077, grad_norm: 0.582944 +Steps: 0%| | 2028/1000000 [08:26<70:57:52, 3.91it/s, grad_norm=0.583, loss_final=0.567, loss_mean=0.856, loss_mean_cls=0.0821, proj_loss=-0.371][2026-03-26 12:39:30] Step: 2028, Training Logs: loss_final: 0.564905, loss_mean: 0.826225, proj_loss: -0.348575, loss_mean_cls: 0.087256, grad_norm: 0.406455 +Steps: 0%| | 2029/1000000 [08:27<70:17:37, 3.94it/s, grad_norm=0.406, loss_final=0.565, loss_mean=0.826, loss_mean_cls=0.0873, proj_loss=-0.349][2026-03-26 12:39:30] Step: 2029, Training Logs: loss_final: 0.578685, loss_mean: 0.866928, proj_loss: -0.370490, loss_mean_cls: 0.082248, grad_norm: 0.497329 +Steps: 0%| | 2030/1000000 [08:27<69:47:43, 3.97it/s, grad_norm=0.497, loss_final=0.579, loss_mean=0.867, loss_mean_cls=0.0822, proj_loss=-0.37][2026-03-26 12:39:30] Step: 2030, Training Logs: loss_final: 0.556449, loss_mean: 0.852328, proj_loss: -0.378093, loss_mean_cls: 0.082215, grad_norm: 0.563723 +Steps: 0%| | 2031/1000000 [08:27<69:26:03, 3.99it/s, grad_norm=0.564, loss_final=0.556, loss_mean=0.852, loss_mean_cls=0.0822, proj_loss=-0.378][2026-03-26 12:39:30] Step: 2031, Training Logs: loss_final: 0.565335, loss_mean: 0.857779, proj_loss: -0.373725, loss_mean_cls: 0.081282, grad_norm: 0.431567 +Steps: 0%| | 2032/1000000 [08:27<69:11:19, 4.01it/s, grad_norm=0.432, loss_final=0.565, loss_mean=0.858, loss_mean_cls=0.0813, proj_loss=-0.374][2026-03-26 12:39:31] Step: 2032, Training Logs: loss_final: 0.565008, loss_mean: 0.869178, proj_loss: -0.384435, loss_mean_cls: 0.080266, grad_norm: 0.504594 +Steps: 0%| | 2033/1000000 [08:28<71:06:51, 3.90it/s, grad_norm=0.505, loss_final=0.565, loss_mean=0.869, loss_mean_cls=0.0803, proj_loss=-0.384][2026-03-26 12:39:31] Step: 2033, Training Logs: loss_final: 0.562555, loss_mean: 0.852098, proj_loss: -0.372435, loss_mean_cls: 0.082892, grad_norm: 0.467962 +Steps: 0%| | 2034/1000000 [08:28<69:48:46, 3.97it/s, grad_norm=0.468, loss_final=0.563, loss_mean=0.852, loss_mean_cls=0.0829, proj_loss=-0.372][2026-03-26 12:39:31] Step: 2034, Training Logs: loss_final: 0.575355, loss_mean: 0.862997, proj_loss: -0.370270, loss_mean_cls: 0.082629, grad_norm: 0.389943 +Steps: 0%| | 2035/1000000 [08:28<69:27:41, 3.99it/s, grad_norm=0.39, loss_final=0.575, loss_mean=0.863, loss_mean_cls=0.0826, proj_loss=-0.37][2026-03-26 12:39:31] Step: 2035, Training Logs: loss_final: 0.566067, loss_mean: 0.855755, proj_loss: -0.371775, loss_mean_cls: 0.082087, grad_norm: 0.644871 +Steps: 0%| | 2036/1000000 [08:28<69:12:41, 4.01it/s, grad_norm=0.645, loss_final=0.566, loss_mean=0.856, loss_mean_cls=0.0821, proj_loss=-0.372][2026-03-26 12:39:32] Step: 2036, Training Logs: loss_final: 0.573330, loss_mean: 0.863336, proj_loss: -0.372645, loss_mean_cls: 0.082638, grad_norm: 0.433626 +Steps: 0%| | 2037/1000000 [08:29<69:04:53, 4.01it/s, grad_norm=0.434, loss_final=0.573, loss_mean=0.863, loss_mean_cls=0.0826, proj_loss=-0.373][2026-03-26 12:39:32] Step: 2037, Training Logs: loss_final: 0.569234, loss_mean: 0.852075, proj_loss: -0.365826, loss_mean_cls: 0.082985, grad_norm: 0.658070 +Steps: 0%| | 2038/1000000 [08:29<69:00:21, 4.02it/s, grad_norm=0.658, loss_final=0.569, loss_mean=0.852, loss_mean_cls=0.083, proj_loss=-0.366][2026-03-26 12:39:32] Step: 2038, Training Logs: loss_final: 0.563229, loss_mean: 0.857578, proj_loss: -0.375713, loss_mean_cls: 0.081365, grad_norm: 0.488247 +Steps: 0%| | 2039/1000000 [08:29<68:54:14, 4.02it/s, grad_norm=0.488, loss_final=0.563, loss_mean=0.858, loss_mean_cls=0.0814, proj_loss=-0.376][2026-03-26 12:39:32] Step: 2039, Training Logs: loss_final: 0.569046, loss_mean: 0.860697, proj_loss: -0.373724, loss_mean_cls: 0.082073, grad_norm: 0.396079 +Steps: 0%| | 2040/1000000 [08:29<68:48:04, 4.03it/s, grad_norm=0.396, loss_final=0.569, loss_mean=0.861, loss_mean_cls=0.0821, proj_loss=-0.374][2026-03-26 12:39:33] Step: 2040, Training Logs: loss_final: 0.557946, loss_mean: 0.852417, proj_loss: -0.376738, loss_mean_cls: 0.082266, grad_norm: 0.399322 +Steps: 0%| | 2041/1000000 [08:30<68:46:21, 4.03it/s, grad_norm=0.399, loss_final=0.558, loss_mean=0.852, loss_mean_cls=0.0823, proj_loss=-0.377][2026-03-26 12:39:33] Step: 2041, Training Logs: loss_final: 0.550654, loss_mean: 0.842473, proj_loss: -0.375460, loss_mean_cls: 0.083641, grad_norm: 0.477591 +Steps: 0%| | 2042/1000000 [08:30<68:44:45, 4.03it/s, grad_norm=0.478, loss_final=0.551, loss_mean=0.842, loss_mean_cls=0.0836, proj_loss=-0.375][2026-03-26 12:39:33] Step: 2042, Training Logs: loss_final: 0.547232, loss_mean: 0.822578, proj_loss: -0.362920, loss_mean_cls: 0.087574, grad_norm: 0.383200 +Steps: 0%| | 2043/1000000 [08:30<68:43:23, 4.03it/s, grad_norm=0.383, loss_final=0.547, loss_mean=0.823, loss_mean_cls=0.0876, proj_loss=-0.363][2026-03-26 12:39:33] Step: 2043, Training Logs: loss_final: 0.557075, loss_mean: 0.864333, proj_loss: -0.386528, loss_mean_cls: 0.079271, grad_norm: 0.451967 +Steps: 0%| | 2044/1000000 [08:30<68:46:06, 4.03it/s, grad_norm=0.452, loss_final=0.557, loss_mean=0.864, loss_mean_cls=0.0793, proj_loss=-0.387][2026-03-26 12:39:34] Step: 2044, Training Logs: loss_final: 0.568727, loss_mean: 0.853338, proj_loss: -0.368826, loss_mean_cls: 0.084215, grad_norm: 0.346988 +Steps: 0%| | 2045/1000000 [08:31<69:43:04, 3.98it/s, grad_norm=0.347, loss_final=0.569, loss_mean=0.853, loss_mean_cls=0.0842, proj_loss=-0.369][2026-03-26 12:39:34] Step: 2045, Training Logs: loss_final: 0.558135, loss_mean: 0.846455, proj_loss: -0.371272, loss_mean_cls: 0.082952, grad_norm: 0.376954 +Steps: 0%| | 2046/1000000 [08:31<69:25:44, 3.99it/s, grad_norm=0.377, loss_final=0.558, loss_mean=0.846, loss_mean_cls=0.083, proj_loss=-0.371][2026-03-26 12:39:34] Step: 2046, Training Logs: loss_final: 0.575167, loss_mean: 0.871668, proj_loss: -0.378120, loss_mean_cls: 0.081619, grad_norm: 0.413385 +Steps: 0%| | 2047/1000000 [08:31<69:11:32, 4.01it/s, grad_norm=0.413, loss_final=0.575, loss_mean=0.872, loss_mean_cls=0.0816, proj_loss=-0.378][2026-03-26 12:39:34] Step: 2047, Training Logs: loss_final: 0.566668, loss_mean: 0.858620, proj_loss: -0.374212, loss_mean_cls: 0.082260, grad_norm: 0.342514 +Steps: 0%| | 2048/1000000 [08:31<69:01:44, 4.02it/s, grad_norm=0.343, loss_final=0.567, loss_mean=0.859, loss_mean_cls=0.0823, proj_loss=-0.374][2026-03-26 12:39:35] Step: 2048, Training Logs: loss_final: 0.548154, loss_mean: 0.842009, proj_loss: -0.375781, loss_mean_cls: 0.081926, grad_norm: 0.396062 +Steps: 0%| | 2049/1000000 [08:32<68:54:40, 4.02it/s, grad_norm=0.396, loss_final=0.548, loss_mean=0.842, loss_mean_cls=0.0819, proj_loss=-0.376][2026-03-26 12:39:35] Step: 2049, Training Logs: loss_final: 0.560567, loss_mean: 0.858300, proj_loss: -0.378693, loss_mean_cls: 0.080960, grad_norm: 0.631903 +Steps: 0%| | 2050/1000000 [08:32<68:50:34, 4.03it/s, grad_norm=0.632, loss_final=0.561, loss_mean=0.858, loss_mean_cls=0.081, proj_loss=-0.379][2026-03-26 12:39:35] Step: 2050, Training Logs: loss_final: 0.544363, loss_mean: 0.833642, proj_loss: -0.371898, loss_mean_cls: 0.082619, grad_norm: 0.512213 +Steps: 0%| | 2051/1000000 [08:32<68:51:07, 4.03it/s, grad_norm=0.512, loss_final=0.544, loss_mean=0.834, loss_mean_cls=0.0826, proj_loss=-0.372][2026-03-26 12:39:35] Step: 2051, Training Logs: loss_final: 0.555383, loss_mean: 0.846869, proj_loss: -0.374223, loss_mean_cls: 0.082737, grad_norm: 0.430827 +Steps: 0%| | 2052/1000000 [08:32<68:50:39, 4.03it/s, grad_norm=0.431, loss_final=0.555, loss_mean=0.847, loss_mean_cls=0.0827, proj_loss=-0.374][2026-03-26 12:39:36] Step: 2052, Training Logs: loss_final: 0.565403, loss_mean: 0.851260, proj_loss: -0.370129, loss_mean_cls: 0.084272, grad_norm: 0.453529 +Steps: 0%| | 2053/1000000 [08:33<68:51:04, 4.03it/s, grad_norm=0.454, loss_final=0.565, loss_mean=0.851, loss_mean_cls=0.0843, proj_loss=-0.37][2026-03-26 12:39:36] Step: 2053, Training Logs: loss_final: 0.566996, loss_mean: 0.853357, proj_loss: -0.369585, loss_mean_cls: 0.083223, grad_norm: 0.667638 +Steps: 0%| | 2054/1000000 [08:33<68:47:27, 4.03it/s, grad_norm=0.668, loss_final=0.567, loss_mean=0.853, loss_mean_cls=0.0832, proj_loss=-0.37][2026-03-26 12:39:36] Step: 2054, Training Logs: loss_final: 0.561298, loss_mean: 0.837211, proj_loss: -0.361427, loss_mean_cls: 0.085513, grad_norm: 0.615892 +Steps: 0%| | 2055/1000000 [08:33<68:45:46, 4.03it/s, grad_norm=0.616, loss_final=0.561, loss_mean=0.837, loss_mean_cls=0.0855, proj_loss=-0.361][2026-03-26 12:39:36] Step: 2055, Training Logs: loss_final: 0.565428, loss_mean: 0.861262, proj_loss: -0.377594, loss_mean_cls: 0.081759, grad_norm: 0.531483 +Steps: 0%| | 2056/1000000 [08:33<68:46:03, 4.03it/s, grad_norm=0.531, loss_final=0.565, loss_mean=0.861, loss_mean_cls=0.0818, proj_loss=-0.378][2026-03-26 12:39:37] Step: 2056, Training Logs: loss_final: 0.574359, loss_mean: 0.858295, proj_loss: -0.368315, loss_mean_cls: 0.084380, grad_norm: 0.822477 +Steps: 0%| | 2057/1000000 [08:33<68:42:12, 4.03it/s, grad_norm=0.822, loss_final=0.574, loss_mean=0.858, loss_mean_cls=0.0844, proj_loss=-0.368][2026-03-26 12:39:37] Step: 2057, Training Logs: loss_final: 0.575576, loss_mean: 0.872016, proj_loss: -0.377486, loss_mean_cls: 0.081046, grad_norm: 0.491816 +Steps: 0%| | 2058/1000000 [08:34<68:41:35, 4.04it/s, grad_norm=0.492, loss_final=0.576, loss_mean=0.872, loss_mean_cls=0.081, proj_loss=-0.377][2026-03-26 12:39:37] Step: 2058, Training Logs: loss_final: 0.553979, loss_mean: 0.845728, proj_loss: -0.374096, loss_mean_cls: 0.082347, grad_norm: 0.444600 +Steps: 0%| | 2059/1000000 [08:34<68:39:08, 4.04it/s, grad_norm=0.445, loss_final=0.554, loss_mean=0.846, loss_mean_cls=0.0823, proj_loss=-0.374][2026-03-26 12:39:37] Step: 2059, Training Logs: loss_final: 0.561537, loss_mean: 0.858147, proj_loss: -0.378490, loss_mean_cls: 0.081880, grad_norm: 0.459837 +Steps: 0%| | 2060/1000000 [08:34<68:38:49, 4.04it/s, grad_norm=0.46, loss_final=0.562, loss_mean=0.858, loss_mean_cls=0.0819, proj_loss=-0.378][2026-03-26 12:39:38] Step: 2060, Training Logs: loss_final: 0.556150, loss_mean: 0.859086, proj_loss: -0.382970, loss_mean_cls: 0.080035, grad_norm: 0.563207 +Steps: 0%| | 2061/1000000 [08:34<68:37:26, 4.04it/s, grad_norm=0.563, loss_final=0.556, loss_mean=0.859, loss_mean_cls=0.08, proj_loss=-0.383][2026-03-26 12:39:38] Step: 2061, Training Logs: loss_final: 0.547911, loss_mean: 0.839889, proj_loss: -0.374926, loss_mean_cls: 0.082948, grad_norm: 0.395809 +Steps: 0%| | 2062/1000000 [08:35<83:57:52, 3.30it/s, grad_norm=0.396, loss_final=0.548, loss_mean=0.84, loss_mean_cls=0.0829, proj_loss=-0.375][2026-03-26 12:39:38] Step: 2062, Training Logs: loss_final: 0.567474, loss_mean: 0.849907, proj_loss: -0.367070, loss_mean_cls: 0.084637, grad_norm: 0.684430 +Steps: 0%| | 2063/1000000 [08:35<79:23:56, 3.49it/s, grad_norm=0.684, loss_final=0.567, loss_mean=0.85, loss_mean_cls=0.0846, proj_loss=-0.367][2026-03-26 12:39:39] Step: 2063, Training Logs: loss_final: 0.559511, loss_mean: 0.854434, proj_loss: -0.375976, loss_mean_cls: 0.081053, grad_norm: 0.403366 +Steps: 0%| | 2064/1000000 [08:35<76:08:39, 3.64it/s, grad_norm=0.403, loss_final=0.56, loss_mean=0.854, loss_mean_cls=0.0811, proj_loss=-0.376][2026-03-26 12:39:39] Step: 2064, Training Logs: loss_final: 0.550597, loss_mean: 0.847554, proj_loss: -0.377627, loss_mean_cls: 0.080669, grad_norm: 0.446226 +Steps: 0%| | 2065/1000000 [08:36<73:54:28, 3.75it/s, grad_norm=0.446, loss_final=0.551, loss_mean=0.848, loss_mean_cls=0.0807, proj_loss=-0.378][2026-03-26 12:39:39] Step: 2065, Training Logs: loss_final: 0.563433, loss_mean: 0.843738, proj_loss: -0.365380, loss_mean_cls: 0.085075, grad_norm: 0.509305 +Steps: 0%| | 2066/1000000 [08:36<72:19:02, 3.83it/s, grad_norm=0.509, loss_final=0.563, loss_mean=0.844, loss_mean_cls=0.0851, proj_loss=-0.365][2026-03-26 12:39:39] Step: 2066, Training Logs: loss_final: 0.550270, loss_mean: 0.853014, proj_loss: -0.383152, loss_mean_cls: 0.080408, grad_norm: 0.467092 +Steps: 0%| | 2067/1000000 [08:36<71:13:03, 3.89it/s, grad_norm=0.467, loss_final=0.55, loss_mean=0.853, loss_mean_cls=0.0804, proj_loss=-0.383][2026-03-26 12:39:40] Step: 2067, Training Logs: loss_final: 0.546543, loss_mean: 0.836815, proj_loss: -0.375363, loss_mean_cls: 0.085091, grad_norm: 0.405935 +Steps: 0%| | 2068/1000000 [08:36<70:23:53, 3.94it/s, grad_norm=0.406, loss_final=0.547, loss_mean=0.837, loss_mean_cls=0.0851, proj_loss=-0.375][2026-03-26 12:39:40] Step: 2068, Training Logs: loss_final: 0.559751, loss_mean: 0.865542, proj_loss: -0.386125, loss_mean_cls: 0.080334, grad_norm: 0.368467 +Steps: 0%| | 2069/1000000 [08:37<69:51:12, 3.97it/s, grad_norm=0.368, loss_final=0.56, loss_mean=0.866, loss_mean_cls=0.0803, proj_loss=-0.386][2026-03-26 12:39:40] Step: 2069, Training Logs: loss_final: 0.567251, loss_mean: 0.852153, proj_loss: -0.369211, loss_mean_cls: 0.084309, grad_norm: 0.476185 +Steps: 0%| | 2070/1000000 [08:37<69:30:21, 3.99it/s, grad_norm=0.476, loss_final=0.567, loss_mean=0.852, loss_mean_cls=0.0843, proj_loss=-0.369][2026-03-26 12:39:40] Step: 2070, Training Logs: loss_final: 0.551372, loss_mean: 0.831425, proj_loss: -0.364806, loss_mean_cls: 0.084753, grad_norm: 0.415390 +Steps: 0%| | 2071/1000000 [08:37<69:15:52, 4.00it/s, grad_norm=0.415, loss_final=0.551, loss_mean=0.831, loss_mean_cls=0.0848, proj_loss=-0.365][2026-03-26 12:39:41] Step: 2071, Training Logs: loss_final: 0.559762, loss_mean: 0.857890, proj_loss: -0.379765, loss_mean_cls: 0.081638, grad_norm: 0.531426 +Steps: 0%| | 2072/1000000 [08:37<69:02:46, 4.01it/s, grad_norm=0.531, loss_final=0.56, loss_mean=0.858, loss_mean_cls=0.0816, proj_loss=-0.38][2026-03-26 12:39:41] Step: 2072, Training Logs: loss_final: 0.555826, loss_mean: 0.848221, proj_loss: -0.376223, loss_mean_cls: 0.083828, grad_norm: 0.575543 +Steps: 0%| | 2073/1000000 [08:38<68:57:48, 4.02it/s, grad_norm=0.576, loss_final=0.556, loss_mean=0.848, loss_mean_cls=0.0838, proj_loss=-0.376][2026-03-26 12:39:41] Step: 2073, Training Logs: loss_final: 0.561594, loss_mean: 0.862797, proj_loss: -0.382359, loss_mean_cls: 0.081157, grad_norm: 0.597612 +Steps: 0%| | 2074/1000000 [08:38<68:50:55, 4.03it/s, grad_norm=0.598, loss_final=0.562, loss_mean=0.863, loss_mean_cls=0.0812, proj_loss=-0.382][2026-03-26 12:39:41] Step: 2074, Training Logs: loss_final: 0.556281, loss_mean: 0.850174, proj_loss: -0.376514, loss_mean_cls: 0.082621, grad_norm: 0.331661 +Steps: 0%| | 2075/1000000 [08:38<84:51:08, 3.27it/s, grad_norm=0.332, loss_final=0.556, loss_mean=0.85, loss_mean_cls=0.0826, proj_loss=-0.377][2026-03-26 12:39:42] Step: 2075, Training Logs: loss_final: 0.572268, loss_mean: 0.870234, proj_loss: -0.378679, loss_mean_cls: 0.080713, grad_norm: 0.536580 +Steps: 0%| | 2076/1000000 [08:39<79:55:05, 3.47it/s, grad_norm=0.537, loss_final=0.572, loss_mean=0.87, loss_mean_cls=0.0807, proj_loss=-0.379][2026-03-26 12:39:42] Step: 2076, Training Logs: loss_final: 0.557943, loss_mean: 0.859958, proj_loss: -0.381739, loss_mean_cls: 0.079724, grad_norm: 0.441089 +Steps: 0%| | 2077/1000000 [08:39<76:31:59, 3.62it/s, grad_norm=0.441, loss_final=0.558, loss_mean=0.86, loss_mean_cls=0.0797, proj_loss=-0.382][2026-03-26 12:39:42] Step: 2077, Training Logs: loss_final: 0.552865, loss_mean: 0.832454, proj_loss: -0.364137, loss_mean_cls: 0.084547, grad_norm: 0.405949 +Steps: 0%| | 2078/1000000 [08:39<74:09:47, 3.74it/s, grad_norm=0.406, loss_final=0.553, loss_mean=0.832, loss_mean_cls=0.0845, proj_loss=-0.364][2026-03-26 12:39:42] Step: 2078, Training Logs: loss_final: 0.569942, loss_mean: 0.853253, proj_loss: -0.367580, loss_mean_cls: 0.084269, grad_norm: 0.516095 +Steps: 0%| | 2079/1000000 [08:39<72:30:36, 3.82it/s, grad_norm=0.516, loss_final=0.57, loss_mean=0.853, loss_mean_cls=0.0843, proj_loss=-0.368][2026-03-26 12:39:43] Step: 2079, Training Logs: loss_final: 0.554047, loss_mean: 0.850654, proj_loss: -0.378206, loss_mean_cls: 0.081600, grad_norm: 0.726194 +Steps: 0%| | 2080/1000000 [08:40<71:20:00, 3.89it/s, grad_norm=0.726, loss_final=0.554, loss_mean=0.851, loss_mean_cls=0.0816, proj_loss=-0.378][2026-03-26 12:39:43] Step: 2080, Training Logs: loss_final: 0.552497, loss_mean: 0.845116, proj_loss: -0.374658, loss_mean_cls: 0.082038, grad_norm: 0.676407 +Steps: 0%| | 2081/1000000 [08:40<70:28:41, 3.93it/s, grad_norm=0.676, loss_final=0.552, loss_mean=0.845, loss_mean_cls=0.082, proj_loss=-0.375][2026-03-26 12:39:43] Step: 2081, Training Logs: loss_final: 0.548481, loss_mean: 0.814478, proj_loss: -0.355335, loss_mean_cls: 0.089339, grad_norm: 0.628662 +Steps: 0%| | 2082/1000000 [08:40<69:58:24, 3.96it/s, grad_norm=0.629, loss_final=0.548, loss_mean=0.814, loss_mean_cls=0.0893, proj_loss=-0.355][2026-03-26 12:39:43] Step: 2082, Training Logs: loss_final: 0.553423, loss_mean: 0.853212, proj_loss: -0.381807, loss_mean_cls: 0.082018, grad_norm: 0.464001 +Steps: 0%| | 2083/1000000 [08:40<69:36:00, 3.98it/s, grad_norm=0.464, loss_final=0.553, loss_mean=0.853, loss_mean_cls=0.082, proj_loss=-0.382][2026-03-26 12:39:44] Step: 2083, Training Logs: loss_final: 0.561521, loss_mean: 0.846633, proj_loss: -0.369020, loss_mean_cls: 0.083908, grad_norm: 1.066763 +Steps: 0%| | 2084/1000000 [08:41<69:20:02, 4.00it/s, grad_norm=1.07, loss_final=0.562, loss_mean=0.847, loss_mean_cls=0.0839, proj_loss=-0.369][2026-03-26 12:39:44] Step: 2084, Training Logs: loss_final: 0.559716, loss_mean: 0.862732, proj_loss: -0.383696, loss_mean_cls: 0.080680, grad_norm: 0.773900 +Steps: 0%| | 2085/1000000 [08:41<69:07:55, 4.01it/s, grad_norm=0.774, loss_final=0.56, loss_mean=0.863, loss_mean_cls=0.0807, proj_loss=-0.384][2026-03-26 12:39:44] Step: 2085, Training Logs: loss_final: 0.564234, loss_mean: 0.863264, proj_loss: -0.379465, loss_mean_cls: 0.080436, grad_norm: 0.492503 +Steps: 0%| | 2086/1000000 [08:41<68:58:59, 4.02it/s, grad_norm=0.493, loss_final=0.564, loss_mean=0.863, loss_mean_cls=0.0804, proj_loss=-0.379][2026-03-26 12:39:44] Step: 2086, Training Logs: loss_final: 0.570972, loss_mean: 0.868630, proj_loss: -0.379455, loss_mean_cls: 0.081797, grad_norm: 0.434963 +Steps: 0%| | 2087/1000000 [08:41<84:49:29, 3.27it/s, grad_norm=0.435, loss_final=0.571, loss_mean=0.869, loss_mean_cls=0.0818, proj_loss=-0.379][2026-03-26 12:39:45] Step: 2087, Training Logs: loss_final: 0.565813, loss_mean: 0.863997, proj_loss: -0.380636, loss_mean_cls: 0.082452, grad_norm: 0.423847 +Steps: 0%| | 2088/1000000 [08:42<79:55:53, 3.47it/s, grad_norm=0.424, loss_final=0.566, loss_mean=0.864, loss_mean_cls=0.0825, proj_loss=-0.381][2026-03-26 12:39:45] Step: 2088, Training Logs: loss_final: 0.560125, loss_mean: 0.866728, proj_loss: -0.386616, loss_mean_cls: 0.080012, grad_norm: 0.443824 +Steps: 0%| | 2089/1000000 [08:42<76:28:43, 3.62it/s, grad_norm=0.444, loss_final=0.56, loss_mean=0.867, loss_mean_cls=0.08, proj_loss=-0.387][2026-03-26 12:39:45] Step: 2089, Training Logs: loss_final: 0.553382, loss_mean: 0.839602, proj_loss: -0.369371, loss_mean_cls: 0.083151, grad_norm: 0.522180 +Steps: 0%| | 2090/1000000 [08:42<74:09:33, 3.74it/s, grad_norm=0.522, loss_final=0.553, loss_mean=0.84, loss_mean_cls=0.0832, proj_loss=-0.369][2026-03-26 12:39:46] Step: 2090, Training Logs: loss_final: 0.552159, loss_mean: 0.849697, proj_loss: -0.380395, loss_mean_cls: 0.082857, grad_norm: 0.471917 +Steps: 0%| | 2091/1000000 [08:42<72:28:05, 3.83it/s, grad_norm=0.472, loss_final=0.552, loss_mean=0.85, loss_mean_cls=0.0829, proj_loss=-0.38][2026-03-26 12:39:46] Step: 2091, Training Logs: loss_final: 0.551668, loss_mean: 0.844886, proj_loss: -0.376466, loss_mean_cls: 0.083248, grad_norm: 0.422653 +Steps: 0%| | 2092/1000000 [08:43<71:24:51, 3.88it/s, grad_norm=0.423, loss_final=0.552, loss_mean=0.845, loss_mean_cls=0.0832, proj_loss=-0.376][2026-03-26 12:39:46] Step: 2092, Training Logs: loss_final: 0.574891, loss_mean: 0.872684, proj_loss: -0.380068, loss_mean_cls: 0.082274, grad_norm: 0.555476 +Steps: 0%| | 2093/1000000 [08:43<70:41:59, 3.92it/s, grad_norm=0.555, loss_final=0.575, loss_mean=0.873, loss_mean_cls=0.0823, proj_loss=-0.38][2026-03-26 12:39:46] Step: 2093, Training Logs: loss_final: 0.558271, loss_mean: 0.854953, proj_loss: -0.379110, loss_mean_cls: 0.082429, grad_norm: 0.407906 +Steps: 0%| | 2094/1000000 [08:43<70:08:31, 3.95it/s, grad_norm=0.408, loss_final=0.558, loss_mean=0.855, loss_mean_cls=0.0824, proj_loss=-0.379][2026-03-26 12:39:47] Step: 2094, Training Logs: loss_final: 0.549825, loss_mean: 0.852744, proj_loss: -0.383571, loss_mean_cls: 0.080651, grad_norm: 0.639780 +Steps: 0%| | 2095/1000000 [08:43<69:55:40, 3.96it/s, grad_norm=0.64, loss_final=0.55, loss_mean=0.853, loss_mean_cls=0.0807, proj_loss=-0.384][2026-03-26 12:39:47] Step: 2095, Training Logs: loss_final: 0.552675, loss_mean: 0.850159, proj_loss: -0.378756, loss_mean_cls: 0.081272, grad_norm: 0.764975 +Steps: 0%| | 2096/1000000 [08:44<69:32:25, 3.99it/s, grad_norm=0.765, loss_final=0.553, loss_mean=0.85, loss_mean_cls=0.0813, proj_loss=-0.379][2026-03-26 12:39:47] Step: 2096, Training Logs: loss_final: 0.563347, loss_mean: 0.857232, proj_loss: -0.375894, loss_mean_cls: 0.082008, grad_norm: 0.496243 +Steps: 0%| | 2097/1000000 [08:44<69:13:50, 4.00it/s, grad_norm=0.496, loss_final=0.563, loss_mean=0.857, loss_mean_cls=0.082, proj_loss=-0.376][2026-03-26 12:39:47] Step: 2097, Training Logs: loss_final: 0.558889, loss_mean: 0.852680, proj_loss: -0.376055, loss_mean_cls: 0.082264, grad_norm: 0.469473 +Steps: 0%| | 2098/1000000 [08:44<69:04:40, 4.01it/s, grad_norm=0.469, loss_final=0.559, loss_mean=0.853, loss_mean_cls=0.0823, proj_loss=-0.376][2026-03-26 12:39:48] Step: 2098, Training Logs: loss_final: 0.559177, loss_mean: 0.857320, proj_loss: -0.380361, loss_mean_cls: 0.082218, grad_norm: 0.434151 +Steps: 0%| | 2099/1000000 [08:44<68:55:38, 4.02it/s, grad_norm=0.434, loss_final=0.559, loss_mean=0.857, loss_mean_cls=0.0822, proj_loss=-0.38][2026-03-26 12:39:48] Step: 2099, Training Logs: loss_final: 0.551736, loss_mean: 0.848326, proj_loss: -0.378660, loss_mean_cls: 0.082070, grad_norm: 0.586865 +Steps: 0%| | 2100/1000000 [08:45<68:49:57, 4.03it/s, grad_norm=0.587, loss_final=0.552, loss_mean=0.848, loss_mean_cls=0.0821, proj_loss=-0.379][2026-03-26 12:39:48] Step: 2100, Training Logs: loss_final: 0.541747, loss_mean: 0.836777, proj_loss: -0.377722, loss_mean_cls: 0.082693, grad_norm: 0.463014 +Steps: 0%| | 2101/1000000 [08:45<68:44:55, 4.03it/s, grad_norm=0.463, loss_final=0.542, loss_mean=0.837, loss_mean_cls=0.0827, proj_loss=-0.378][2026-03-26 12:39:48] Step: 2101, Training Logs: loss_final: 0.539627, loss_mean: 0.828542, proj_loss: -0.372739, loss_mean_cls: 0.083824, grad_norm: 0.359859 +Steps: 0%| | 2102/1000000 [08:45<68:45:17, 4.03it/s, grad_norm=0.36, loss_final=0.54, loss_mean=0.829, loss_mean_cls=0.0838, proj_loss=-0.373][2026-03-26 12:39:49] Step: 2102, Training Logs: loss_final: 0.536938, loss_mean: 0.827120, proj_loss: -0.374542, loss_mean_cls: 0.084360, grad_norm: 0.463823 +Steps: 0%| | 2103/1000000 [08:45<68:39:47, 4.04it/s, grad_norm=0.464, loss_final=0.537, loss_mean=0.827, loss_mean_cls=0.0844, proj_loss=-0.375][2026-03-26 12:39:49] Step: 2103, Training Logs: loss_final: 0.546954, loss_mean: 0.832432, proj_loss: -0.370437, loss_mean_cls: 0.084959, grad_norm: 0.454804 +Steps: 0%| | 2104/1000000 [08:46<68:38:35, 4.04it/s, grad_norm=0.455, loss_final=0.547, loss_mean=0.832, loss_mean_cls=0.085, proj_loss=-0.37][2026-03-26 12:39:49] Step: 2104, Training Logs: loss_final: 0.547076, loss_mean: 0.836648, proj_loss: -0.373016, loss_mean_cls: 0.083443, grad_norm: 0.429002 +Steps: 0%| | 2105/1000000 [08:46<68:37:51, 4.04it/s, grad_norm=0.429, loss_final=0.547, loss_mean=0.837, loss_mean_cls=0.0834, proj_loss=-0.373][2026-03-26 12:39:49] Step: 2105, Training Logs: loss_final: 0.560883, loss_mean: 0.864507, proj_loss: -0.383608, loss_mean_cls: 0.079985, grad_norm: 0.508034 +Steps: 0%| | 2106/1000000 [08:46<68:36:46, 4.04it/s, grad_norm=0.508, loss_final=0.561, loss_mean=0.865, loss_mean_cls=0.08, proj_loss=-0.384][2026-03-26 12:39:50] Step: 2106, Training Logs: loss_final: 0.553405, loss_mean: 0.847695, proj_loss: -0.377365, loss_mean_cls: 0.083075, grad_norm: 0.408217 +Steps: 0%| | 2107/1000000 [08:46<68:41:57, 4.03it/s, grad_norm=0.408, loss_final=0.553, loss_mean=0.848, loss_mean_cls=0.0831, proj_loss=-0.377][2026-03-26 12:39:50] Step: 2107, Training Logs: loss_final: 0.567954, loss_mean: 0.851877, proj_loss: -0.368630, loss_mean_cls: 0.084707, grad_norm: 0.446655 +Steps: 0%| | 2108/1000000 [08:47<68:39:30, 4.04it/s, grad_norm=0.447, loss_final=0.568, loss_mean=0.852, loss_mean_cls=0.0847, proj_loss=-0.369][2026-03-26 12:39:50] Step: 2108, Training Logs: loss_final: 0.568052, loss_mean: 0.860254, proj_loss: -0.376448, loss_mean_cls: 0.084246, grad_norm: 0.496218 +Steps: 0%| | 2109/1000000 [08:47<68:38:17, 4.04it/s, grad_norm=0.496, loss_final=0.568, loss_mean=0.86, loss_mean_cls=0.0842, proj_loss=-0.376][2026-03-26 12:39:50] Step: 2109, Training Logs: loss_final: 0.562902, loss_mean: 0.859196, proj_loss: -0.377120, loss_mean_cls: 0.080826, grad_norm: 0.557407 +Steps: 0%| | 2110/1000000 [08:47<68:34:36, 4.04it/s, grad_norm=0.557, loss_final=0.563, loss_mean=0.859, loss_mean_cls=0.0808, proj_loss=-0.377][2026-03-26 12:39:51] Step: 2110, Training Logs: loss_final: 0.557225, loss_mean: 0.841425, proj_loss: -0.369252, loss_mean_cls: 0.085053, grad_norm: 0.360192 +Steps: 0%| | 2111/1000000 [08:47<68:33:50, 4.04it/s, grad_norm=0.36, loss_final=0.557, loss_mean=0.841, loss_mean_cls=0.0851, proj_loss=-0.369][2026-03-26 12:39:51] Step: 2111, Training Logs: loss_final: 0.554845, loss_mean: 0.840997, proj_loss: -0.370875, loss_mean_cls: 0.084724, grad_norm: 0.600922 +Steps: 0%| | 2112/1000000 [08:48<68:34:52, 4.04it/s, grad_norm=0.601, loss_final=0.555, loss_mean=0.841, loss_mean_cls=0.0847, proj_loss=-0.371][2026-03-26 12:39:51] Step: 2112, Training Logs: loss_final: 0.555440, loss_mean: 0.850191, proj_loss: -0.376899, loss_mean_cls: 0.082147, grad_norm: 0.477994 +Steps: 0%| | 2113/1000000 [08:48<68:34:18, 4.04it/s, grad_norm=0.478, loss_final=0.555, loss_mean=0.85, loss_mean_cls=0.0821, proj_loss=-0.377][2026-03-26 12:39:51] Step: 2113, Training Logs: loss_final: 0.570909, loss_mean: 0.851296, proj_loss: -0.365699, loss_mean_cls: 0.085312, grad_norm: 0.483958 +Steps: 0%| | 2114/1000000 [08:48<68:36:48, 4.04it/s, grad_norm=0.484, loss_final=0.571, loss_mean=0.851, loss_mean_cls=0.0853, proj_loss=-0.366][2026-03-26 12:39:52] Step: 2114, Training Logs: loss_final: 0.564332, loss_mean: 0.854080, proj_loss: -0.373095, loss_mean_cls: 0.083348, grad_norm: 0.636277 +Steps: 0%| | 2115/1000000 [08:48<68:38:48, 4.04it/s, grad_norm=0.636, loss_final=0.564, loss_mean=0.854, loss_mean_cls=0.0833, proj_loss=-0.373][2026-03-26 12:39:52] Step: 2115, Training Logs: loss_final: 0.560553, loss_mean: 0.845180, proj_loss: -0.368634, loss_mean_cls: 0.084007, grad_norm: 0.604513 +Steps: 0%| | 2116/1000000 [08:49<68:38:21, 4.04it/s, grad_norm=0.605, loss_final=0.561, loss_mean=0.845, loss_mean_cls=0.084, proj_loss=-0.369][2026-03-26 12:39:52] Step: 2116, Training Logs: loss_final: 0.564302, loss_mean: 0.862028, proj_loss: -0.378458, loss_mean_cls: 0.080732, grad_norm: 0.329773 +Steps: 0%| | 2117/1000000 [08:49<68:38:18, 4.04it/s, grad_norm=0.33, loss_final=0.564, loss_mean=0.862, loss_mean_cls=0.0807, proj_loss=-0.378][2026-03-26 12:39:52] Step: 2117, Training Logs: loss_final: 0.561622, loss_mean: 0.860993, proj_loss: -0.381813, loss_mean_cls: 0.082442, grad_norm: 0.444693 +Steps: 0%| | 2118/1000000 [08:49<68:39:24, 4.04it/s, grad_norm=0.445, loss_final=0.562, loss_mean=0.861, loss_mean_cls=0.0824, proj_loss=-0.382][2026-03-26 12:39:53] Step: 2118, Training Logs: loss_final: 0.559198, loss_mean: 0.867616, proj_loss: -0.389135, loss_mean_cls: 0.080716, grad_norm: 0.357858 +Steps: 0%| | 2119/1000000 [08:49<68:37:09, 4.04it/s, grad_norm=0.358, loss_final=0.559, loss_mean=0.868, loss_mean_cls=0.0807, proj_loss=-0.389][2026-03-26 12:39:53] Step: 2119, Training Logs: loss_final: 0.549401, loss_mean: 0.821380, proj_loss: -0.358798, loss_mean_cls: 0.086820, grad_norm: 0.526680 +Steps: 0%| | 2120/1000000 [08:50<68:41:02, 4.04it/s, grad_norm=0.527, loss_final=0.549, loss_mean=0.821, loss_mean_cls=0.0868, proj_loss=-0.359][2026-03-26 12:39:53] Step: 2120, Training Logs: loss_final: 0.559369, loss_mean: 0.854169, proj_loss: -0.379043, loss_mean_cls: 0.084243, grad_norm: 0.534412 +Steps: 0%| | 2121/1000000 [08:50<68:39:30, 4.04it/s, grad_norm=0.534, loss_final=0.559, loss_mean=0.854, loss_mean_cls=0.0842, proj_loss=-0.379][2026-03-26 12:39:53] Step: 2121, Training Logs: loss_final: 0.557358, loss_mean: 0.852799, proj_loss: -0.378505, loss_mean_cls: 0.083064, grad_norm: 0.724754 +Steps: 0%| | 2122/1000000 [08:50<68:39:46, 4.04it/s, grad_norm=0.725, loss_final=0.557, loss_mean=0.853, loss_mean_cls=0.0831, proj_loss=-0.379][2026-03-26 12:39:54] Step: 2122, Training Logs: loss_final: 0.553819, loss_mean: 0.858533, proj_loss: -0.384101, loss_mean_cls: 0.079387, grad_norm: 0.409641 +Steps: 0%| | 2123/1000000 [08:50<68:40:31, 4.04it/s, grad_norm=0.41, loss_final=0.554, loss_mean=0.859, loss_mean_cls=0.0794, proj_loss=-0.384][2026-03-26 12:39:54] Step: 2123, Training Logs: loss_final: 0.556750, loss_mean: 0.848444, proj_loss: -0.374511, loss_mean_cls: 0.082818, grad_norm: 0.531197 +Steps: 0%| | 2124/1000000 [08:51<68:38:51, 4.04it/s, grad_norm=0.531, loss_final=0.557, loss_mean=0.848, loss_mean_cls=0.0828, proj_loss=-0.375][2026-03-26 12:39:54] Step: 2124, Training Logs: loss_final: 0.569426, loss_mean: 0.871154, proj_loss: -0.384301, loss_mean_cls: 0.082573, grad_norm: 0.447887 +Steps: 0%| | 2125/1000000 [08:51<68:37:34, 4.04it/s, grad_norm=0.448, loss_final=0.569, loss_mean=0.871, loss_mean_cls=0.0826, proj_loss=-0.384][2026-03-26 12:39:54] Step: 2125, Training Logs: loss_final: 0.555238, loss_mean: 0.847389, proj_loss: -0.375278, loss_mean_cls: 0.083126, grad_norm: 0.525532 +Steps: 0%| | 2126/1000000 [08:51<68:38:32, 4.04it/s, grad_norm=0.526, loss_final=0.555, loss_mean=0.847, loss_mean_cls=0.0831, proj_loss=-0.375][2026-03-26 12:39:55] Step: 2126, Training Logs: loss_final: 0.557773, loss_mean: 0.854427, proj_loss: -0.379048, loss_mean_cls: 0.082394, grad_norm: 0.493444 +Steps: 0%| | 2127/1000000 [08:51<68:36:37, 4.04it/s, grad_norm=0.493, loss_final=0.558, loss_mean=0.854, loss_mean_cls=0.0824, proj_loss=-0.379][2026-03-26 12:39:55] Step: 2127, Training Logs: loss_final: 0.547708, loss_mean: 0.842579, proj_loss: -0.378342, loss_mean_cls: 0.083471, grad_norm: 0.365515 +Steps: 0%| | 2128/1000000 [08:52<68:40:42, 4.04it/s, grad_norm=0.366, loss_final=0.548, loss_mean=0.843, loss_mean_cls=0.0835, proj_loss=-0.378][2026-03-26 12:39:55] Step: 2128, Training Logs: loss_final: 0.543398, loss_mean: 0.836839, proj_loss: -0.376695, loss_mean_cls: 0.083254, grad_norm: 0.499246 +Steps: 0%| | 2129/1000000 [08:52<68:37:35, 4.04it/s, grad_norm=0.499, loss_final=0.543, loss_mean=0.837, loss_mean_cls=0.0833, proj_loss=-0.377][2026-03-26 12:39:55] Step: 2129, Training Logs: loss_final: 0.543861, loss_mean: 0.836774, proj_loss: -0.377355, loss_mean_cls: 0.084442, grad_norm: 0.482603 +Steps: 0%| | 2130/1000000 [08:52<68:42:47, 4.03it/s, grad_norm=0.483, loss_final=0.544, loss_mean=0.837, loss_mean_cls=0.0844, proj_loss=-0.377][2026-03-26 12:39:56] Step: 2130, Training Logs: loss_final: 0.554765, loss_mean: 0.836371, proj_loss: -0.367555, loss_mean_cls: 0.085949, grad_norm: 0.476172 +Steps: 0%| | 2131/1000000 [08:52<68:38:35, 4.04it/s, grad_norm=0.476, loss_final=0.555, loss_mean=0.836, loss_mean_cls=0.0859, proj_loss=-0.368][2026-03-26 12:39:56] Step: 2131, Training Logs: loss_final: 0.538885, loss_mean: 0.840130, proj_loss: -0.383186, loss_mean_cls: 0.081940, grad_norm: 0.529672 +Steps: 0%| | 2132/1000000 [08:53<68:37:52, 4.04it/s, grad_norm=0.53, loss_final=0.539, loss_mean=0.84, loss_mean_cls=0.0819, proj_loss=-0.383][2026-03-26 12:39:56] Step: 2132, Training Logs: loss_final: 0.554164, loss_mean: 0.832985, proj_loss: -0.364492, loss_mean_cls: 0.085672, grad_norm: 0.385489 +Steps: 0%| | 2133/1000000 [08:53<68:37:29, 4.04it/s, grad_norm=0.385, loss_final=0.554, loss_mean=0.833, loss_mean_cls=0.0857, proj_loss=-0.364][2026-03-26 12:39:56] Step: 2133, Training Logs: loss_final: 0.563470, loss_mean: 0.862568, proj_loss: -0.381677, loss_mean_cls: 0.082580, grad_norm: 0.436272 +Steps: 0%| | 2134/1000000 [08:53<68:37:18, 4.04it/s, grad_norm=0.436, loss_final=0.563, loss_mean=0.863, loss_mean_cls=0.0826, proj_loss=-0.382][2026-03-26 12:39:57] Step: 2134, Training Logs: loss_final: 0.574884, loss_mean: 0.876069, proj_loss: -0.381480, loss_mean_cls: 0.080294, grad_norm: 0.405696 +Steps: 0%| | 2135/1000000 [08:53<68:35:38, 4.04it/s, grad_norm=0.406, loss_final=0.575, loss_mean=0.876, loss_mean_cls=0.0803, proj_loss=-0.381][2026-03-26 12:39:57] Step: 2135, Training Logs: loss_final: 0.548454, loss_mean: 0.838562, proj_loss: -0.374035, loss_mean_cls: 0.083928, grad_norm: 0.514416 +Steps: 0%| | 2136/1000000 [08:54<68:42:28, 4.03it/s, grad_norm=0.514, loss_final=0.548, loss_mean=0.839, loss_mean_cls=0.0839, proj_loss=-0.374][2026-03-26 12:39:57] Step: 2136, Training Logs: loss_final: 0.549433, loss_mean: 0.844055, proj_loss: -0.378108, loss_mean_cls: 0.083485, grad_norm: 0.577382 +Steps: 0%| | 2137/1000000 [08:54<68:45:40, 4.03it/s, grad_norm=0.577, loss_final=0.549, loss_mean=0.844, loss_mean_cls=0.0835, proj_loss=-0.378][2026-03-26 12:39:57] Step: 2137, Training Logs: loss_final: 0.539926, loss_mean: 0.828599, proj_loss: -0.372839, loss_mean_cls: 0.084166, grad_norm: 0.473395 +Steps: 0%| | 2138/1000000 [08:54<68:47:17, 4.03it/s, grad_norm=0.473, loss_final=0.54, loss_mean=0.829, loss_mean_cls=0.0842, proj_loss=-0.373][2026-03-26 12:39:58] Step: 2138, Training Logs: loss_final: 0.555875, loss_mean: 0.854898, proj_loss: -0.381646, loss_mean_cls: 0.082623, grad_norm: 0.410280 +Steps: 0%| | 2139/1000000 [08:54<68:48:34, 4.03it/s, grad_norm=0.41, loss_final=0.556, loss_mean=0.855, loss_mean_cls=0.0826, proj_loss=-0.382][2026-03-26 12:39:58] Step: 2139, Training Logs: loss_final: 0.561920, loss_mean: 0.876188, proj_loss: -0.393010, loss_mean_cls: 0.078742, grad_norm: 0.578250 +Steps: 0%| | 2140/1000000 [08:55<68:50:39, 4.03it/s, grad_norm=0.578, loss_final=0.562, loss_mean=0.876, loss_mean_cls=0.0787, proj_loss=-0.393][2026-03-26 12:39:58] Step: 2140, Training Logs: loss_final: 0.553591, loss_mean: 0.846471, proj_loss: -0.377270, loss_mean_cls: 0.084389, grad_norm: 0.722137 +Steps: 0%| | 2141/1000000 [08:55<69:02:03, 4.02it/s, grad_norm=0.722, loss_final=0.554, loss_mean=0.846, loss_mean_cls=0.0844, proj_loss=-0.377][2026-03-26 12:39:58] Step: 2141, Training Logs: loss_final: 0.552959, loss_mean: 0.850809, proj_loss: -0.380947, loss_mean_cls: 0.083097, grad_norm: 0.497810 +Steps: 0%| | 2142/1000000 [08:55<68:54:15, 4.02it/s, grad_norm=0.498, loss_final=0.553, loss_mean=0.851, loss_mean_cls=0.0831, proj_loss=-0.381][2026-03-26 12:39:59] Step: 2142, Training Logs: loss_final: 0.560023, loss_mean: 0.851311, proj_loss: -0.374891, loss_mean_cls: 0.083603, grad_norm: 0.704994 +Steps: 0%| | 2143/1000000 [08:55<68:49:26, 4.03it/s, grad_norm=0.705, loss_final=0.56, loss_mean=0.851, loss_mean_cls=0.0836, proj_loss=-0.375][2026-03-26 12:39:59] Step: 2143, Training Logs: loss_final: 0.550745, loss_mean: 0.845189, proj_loss: -0.377153, loss_mean_cls: 0.082709, grad_norm: 0.696182 +Steps: 0%| | 2144/1000000 [08:56<68:46:33, 4.03it/s, grad_norm=0.696, loss_final=0.551, loss_mean=0.845, loss_mean_cls=0.0827, proj_loss=-0.377][2026-03-26 12:39:59] Step: 2144, Training Logs: loss_final: 0.562809, loss_mean: 0.852005, proj_loss: -0.373340, loss_mean_cls: 0.084143, grad_norm: 0.499896 +Steps: 0%| | 2145/1000000 [08:56<68:43:10, 4.03it/s, grad_norm=0.5, loss_final=0.563, loss_mean=0.852, loss_mean_cls=0.0841, proj_loss=-0.373][2026-03-26 12:39:59] Step: 2145, Training Logs: loss_final: 0.559308, loss_mean: 0.854710, proj_loss: -0.377529, loss_mean_cls: 0.082127, grad_norm: 0.507270 +Steps: 0%| | 2146/1000000 [08:56<68:41:19, 4.04it/s, grad_norm=0.507, loss_final=0.559, loss_mean=0.855, loss_mean_cls=0.0821, proj_loss=-0.378][2026-03-26 12:39:59] Step: 2146, Training Logs: loss_final: 0.565376, loss_mean: 0.842174, proj_loss: -0.364039, loss_mean_cls: 0.087242, grad_norm: 0.696147 +Steps: 0%| | 2147/1000000 [08:56<68:39:15, 4.04it/s, grad_norm=0.696, loss_final=0.565, loss_mean=0.842, loss_mean_cls=0.0872, proj_loss=-0.364][2026-03-26 12:40:00] Step: 2147, Training Logs: loss_final: 0.557446, loss_mean: 0.862036, proj_loss: -0.385785, loss_mean_cls: 0.081195, grad_norm: 0.507385 +Steps: 0%| | 2148/1000000 [08:57<68:41:22, 4.04it/s, grad_norm=0.507, loss_final=0.557, loss_mean=0.862, loss_mean_cls=0.0812, proj_loss=-0.386][2026-03-26 12:40:00] Step: 2148, Training Logs: loss_final: 0.565984, loss_mean: 0.880361, proj_loss: -0.392076, loss_mean_cls: 0.077699, grad_norm: 0.537092 +Steps: 0%| | 2149/1000000 [08:57<68:39:14, 4.04it/s, grad_norm=0.537, loss_final=0.566, loss_mean=0.88, loss_mean_cls=0.0777, proj_loss=-0.392][2026-03-26 12:40:00] Step: 2149, Training Logs: loss_final: 0.552629, loss_mean: 0.854038, proj_loss: -0.382858, loss_mean_cls: 0.081450, grad_norm: 0.848113 +Steps: 0%| | 2150/1000000 [08:57<68:36:45, 4.04it/s, grad_norm=0.848, loss_final=0.553, loss_mean=0.854, loss_mean_cls=0.0814, proj_loss=-0.383][2026-03-26 12:40:00] Step: 2150, Training Logs: loss_final: 0.544847, loss_mean: 0.840417, proj_loss: -0.378561, loss_mean_cls: 0.082992, grad_norm: 0.651430 +Steps: 0%| | 2151/1000000 [08:57<68:40:12, 4.04it/s, grad_norm=0.651, loss_final=0.545, loss_mean=0.84, loss_mean_cls=0.083, proj_loss=-0.379][2026-03-26 12:40:01] Step: 2151, Training Logs: loss_final: 0.547757, loss_mean: 0.850250, proj_loss: -0.383536, loss_mean_cls: 0.081043, grad_norm: 0.315504 +Steps: 0%| | 2152/1000000 [08:58<68:39:58, 4.04it/s, grad_norm=0.316, loss_final=0.548, loss_mean=0.85, loss_mean_cls=0.081, proj_loss=-0.384][2026-03-26 12:40:01] Step: 2152, Training Logs: loss_final: 0.547040, loss_mean: 0.838166, proj_loss: -0.374609, loss_mean_cls: 0.083483, grad_norm: 0.568666 +Steps: 0%| | 2153/1000000 [08:58<68:40:20, 4.04it/s, grad_norm=0.569, loss_final=0.547, loss_mean=0.838, loss_mean_cls=0.0835, proj_loss=-0.375][2026-03-26 12:40:01] Step: 2153, Training Logs: loss_final: 0.554204, loss_mean: 0.851052, proj_loss: -0.379405, loss_mean_cls: 0.082557, grad_norm: 0.752964 +Steps: 0%| | 2154/1000000 [08:58<68:40:49, 4.04it/s, grad_norm=0.753, loss_final=0.554, loss_mean=0.851, loss_mean_cls=0.0826, proj_loss=-0.379][2026-03-26 12:40:01] Step: 2154, Training Logs: loss_final: 0.557829, loss_mean: 0.846074, proj_loss: -0.373425, loss_mean_cls: 0.085181, grad_norm: 0.451067 +Steps: 0%| | 2155/1000000 [08:58<68:53:03, 4.02it/s, grad_norm=0.451, loss_final=0.558, loss_mean=0.846, loss_mean_cls=0.0852, proj_loss=-0.373][2026-03-26 12:40:02] Step: 2155, Training Logs: loss_final: 0.562135, loss_mean: 0.859621, proj_loss: -0.379665, loss_mean_cls: 0.082179, grad_norm: 0.599600 +Steps: 0%| | 2156/1000000 [08:59<68:48:19, 4.03it/s, grad_norm=0.6, loss_final=0.562, loss_mean=0.86, loss_mean_cls=0.0822, proj_loss=-0.38][2026-03-26 12:40:02] Step: 2156, Training Logs: loss_final: 0.567531, loss_mean: 0.870371, proj_loss: -0.384855, loss_mean_cls: 0.082014, grad_norm: 0.819418 +Steps: 0%| | 2157/1000000 [08:59<68:46:10, 4.03it/s, grad_norm=0.819, loss_final=0.568, loss_mean=0.87, loss_mean_cls=0.082, proj_loss=-0.385][2026-03-26 12:40:02] Step: 2157, Training Logs: loss_final: 0.553056, loss_mean: 0.857001, proj_loss: -0.384593, loss_mean_cls: 0.080648, grad_norm: 0.466457 +Steps: 0%| | 2158/1000000 [08:59<68:46:28, 4.03it/s, grad_norm=0.466, loss_final=0.553, loss_mean=0.857, loss_mean_cls=0.0806, proj_loss=-0.385][2026-03-26 12:40:02] Step: 2158, Training Logs: loss_final: 0.549917, loss_mean: 0.848494, proj_loss: -0.381210, loss_mean_cls: 0.082634, grad_norm: 0.396841 +Steps: 0%| | 2159/1000000 [08:59<68:41:27, 4.04it/s, grad_norm=0.397, loss_final=0.55, loss_mean=0.848, loss_mean_cls=0.0826, proj_loss=-0.381][2026-03-26 12:40:03] Step: 2159, Training Logs: loss_final: 0.561609, loss_mean: 0.859007, proj_loss: -0.379083, loss_mean_cls: 0.081685, grad_norm: 0.495713 +Steps: 0%| | 2160/1000000 [09:00<68:39:23, 4.04it/s, grad_norm=0.496, loss_final=0.562, loss_mean=0.859, loss_mean_cls=0.0817, proj_loss=-0.379][2026-03-26 12:40:03] Step: 2160, Training Logs: loss_final: 0.547786, loss_mean: 0.840916, proj_loss: -0.377624, loss_mean_cls: 0.084495, grad_norm: 0.405676 +Steps: 0%| | 2161/1000000 [09:00<68:37:46, 4.04it/s, grad_norm=0.406, loss_final=0.548, loss_mean=0.841, loss_mean_cls=0.0845, proj_loss=-0.378][2026-03-26 12:40:03] Step: 2161, Training Logs: loss_final: 0.542764, loss_mean: 0.827688, proj_loss: -0.370888, loss_mean_cls: 0.085964, grad_norm: 0.517001 +Steps: 0%| | 2162/1000000 [09:00<68:38:45, 4.04it/s, grad_norm=0.517, loss_final=0.543, loss_mean=0.828, loss_mean_cls=0.086, proj_loss=-0.371][2026-03-26 12:40:03] Step: 2162, Training Logs: loss_final: 0.536109, loss_mean: 0.825876, proj_loss: -0.375722, loss_mean_cls: 0.085955, grad_norm: 0.711201 +Steps: 0%| | 2163/1000000 [09:00<68:37:15, 4.04it/s, grad_norm=0.711, loss_final=0.536, loss_mean=0.826, loss_mean_cls=0.086, proj_loss=-0.376][2026-03-26 12:40:04] Step: 2163, Training Logs: loss_final: 0.553045, loss_mean: 0.843053, proj_loss: -0.373400, loss_mean_cls: 0.083391, grad_norm: 0.394707 +Steps: 0%| | 2164/1000000 [09:01<68:39:24, 4.04it/s, grad_norm=0.395, loss_final=0.553, loss_mean=0.843, loss_mean_cls=0.0834, proj_loss=-0.373][2026-03-26 12:40:04] Step: 2164, Training Logs: loss_final: 0.553136, loss_mean: 0.852206, proj_loss: -0.381934, loss_mean_cls: 0.082864, grad_norm: 0.482143 +Steps: 0%| | 2165/1000000 [09:01<68:38:40, 4.04it/s, grad_norm=0.482, loss_final=0.553, loss_mean=0.852, loss_mean_cls=0.0829, proj_loss=-0.382][2026-03-26 12:40:04] Step: 2165, Training Logs: loss_final: 0.560133, loss_mean: 0.860169, proj_loss: -0.381802, loss_mean_cls: 0.081766, grad_norm: 0.456895 +Steps: 0%| | 2166/1000000 [09:01<68:41:18, 4.04it/s, grad_norm=0.457, loss_final=0.56, loss_mean=0.86, loss_mean_cls=0.0818, proj_loss=-0.382][2026-03-26 12:40:04] Step: 2166, Training Logs: loss_final: 0.568052, loss_mean: 0.871271, proj_loss: -0.383514, loss_mean_cls: 0.080295, grad_norm: 0.439630 +Steps: 0%| | 2167/1000000 [09:01<68:39:43, 4.04it/s, grad_norm=0.44, loss_final=0.568, loss_mean=0.871, loss_mean_cls=0.0803, proj_loss=-0.384][2026-03-26 12:40:05] Step: 2167, Training Logs: loss_final: 0.548231, loss_mean: 0.845740, proj_loss: -0.380437, loss_mean_cls: 0.082928, grad_norm: 0.527887 +Steps: 0%| | 2168/1000000 [09:02<68:40:04, 4.04it/s, grad_norm=0.528, loss_final=0.548, loss_mean=0.846, loss_mean_cls=0.0829, proj_loss=-0.38][2026-03-26 12:40:05] Step: 2168, Training Logs: loss_final: 0.548435, loss_mean: 0.841617, proj_loss: -0.377105, loss_mean_cls: 0.083924, grad_norm: 0.454696 +Steps: 0%| | 2169/1000000 [09:02<68:51:11, 4.03it/s, grad_norm=0.455, loss_final=0.548, loss_mean=0.842, loss_mean_cls=0.0839, proj_loss=-0.377][2026-03-26 12:40:05] Step: 2169, Training Logs: loss_final: 0.557366, loss_mean: 0.851219, proj_loss: -0.378023, loss_mean_cls: 0.084170, grad_norm: 0.515829 +Steps: 0%| | 2170/1000000 [09:02<69:03:10, 4.01it/s, grad_norm=0.516, loss_final=0.557, loss_mean=0.851, loss_mean_cls=0.0842, proj_loss=-0.378][2026-03-26 12:40:05] Step: 2170, Training Logs: loss_final: 0.559580, loss_mean: 0.856026, proj_loss: -0.378922, loss_mean_cls: 0.082476, grad_norm: 0.444669 +Steps: 0%| | 2171/1000000 [09:02<68:54:52, 4.02it/s, grad_norm=0.445, loss_final=0.56, loss_mean=0.856, loss_mean_cls=0.0825, proj_loss=-0.379][2026-03-26 12:40:06] Step: 2171, Training Logs: loss_final: 0.542291, loss_mean: 0.838879, proj_loss: -0.379754, loss_mean_cls: 0.083165, grad_norm: 0.416628 +Steps: 0%| | 2172/1000000 [09:03<69:06:53, 4.01it/s, grad_norm=0.417, loss_final=0.542, loss_mean=0.839, loss_mean_cls=0.0832, proj_loss=-0.38][2026-03-26 12:40:06] Step: 2172, Training Logs: loss_final: 0.546034, loss_mean: 0.845081, proj_loss: -0.382665, loss_mean_cls: 0.083618, grad_norm: 0.529786 +Steps: 0%| | 2173/1000000 [09:03<69:13:40, 4.00it/s, grad_norm=0.53, loss_final=0.546, loss_mean=0.845, loss_mean_cls=0.0836, proj_loss=-0.383][2026-03-26 12:40:06] Step: 2173, Training Logs: loss_final: 0.548894, loss_mean: 0.837343, proj_loss: -0.372768, loss_mean_cls: 0.084319, grad_norm: 0.376778 +Steps: 0%| | 2174/1000000 [09:03<69:07:25, 4.01it/s, grad_norm=0.377, loss_final=0.549, loss_mean=0.837, loss_mean_cls=0.0843, proj_loss=-0.373][2026-03-26 12:40:06] Step: 2174, Training Logs: loss_final: 0.552918, loss_mean: 0.873225, proj_loss: -0.398203, loss_mean_cls: 0.077896, grad_norm: 0.441818 +Steps: 0%| | 2175/1000000 [09:03<69:06:11, 4.01it/s, grad_norm=0.442, loss_final=0.553, loss_mean=0.873, loss_mean_cls=0.0779, proj_loss=-0.398][2026-03-26 12:40:07] Step: 2175, Training Logs: loss_final: 0.555909, loss_mean: 0.835488, proj_loss: -0.366096, loss_mean_cls: 0.086517, grad_norm: 0.498610 +Steps: 0%| | 2176/1000000 [09:04<69:12:13, 4.01it/s, grad_norm=0.499, loss_final=0.556, loss_mean=0.835, loss_mean_cls=0.0865, proj_loss=-0.366][2026-03-26 12:40:07] Step: 2176, Training Logs: loss_final: 0.553089, loss_mean: 0.849139, proj_loss: -0.380001, loss_mean_cls: 0.083952, grad_norm: 0.601755 +Steps: 0%| | 2177/1000000 [09:04<70:32:25, 3.93it/s, grad_norm=0.602, loss_final=0.553, loss_mean=0.849, loss_mean_cls=0.084, proj_loss=-0.38][2026-03-26 12:40:07] Step: 2177, Training Logs: loss_final: 0.552249, loss_mean: 0.847380, proj_loss: -0.378375, loss_mean_cls: 0.083245, grad_norm: 0.556785 +Steps: 0%| | 2178/1000000 [09:04<70:16:20, 3.94it/s, grad_norm=0.557, loss_final=0.552, loss_mean=0.847, loss_mean_cls=0.0832, proj_loss=-0.378][2026-03-26 12:40:07] Step: 2178, Training Logs: loss_final: 0.553131, loss_mean: 0.853767, proj_loss: -0.382958, loss_mean_cls: 0.082323, grad_norm: 0.464471 +Steps: 0%| | 2179/1000000 [09:04<69:47:24, 3.97it/s, grad_norm=0.464, loss_final=0.553, loss_mean=0.854, loss_mean_cls=0.0823, proj_loss=-0.383][2026-03-26 12:40:08] Step: 2179, Training Logs: loss_final: 0.552884, loss_mean: 0.861205, proj_loss: -0.387440, loss_mean_cls: 0.079119, grad_norm: 0.568403 +Steps: 0%| | 2180/1000000 [09:05<69:41:45, 3.98it/s, grad_norm=0.568, loss_final=0.553, loss_mean=0.861, loss_mean_cls=0.0791, proj_loss=-0.387][2026-03-26 12:40:08] Step: 2180, Training Logs: loss_final: 0.554053, loss_mean: 0.862401, proj_loss: -0.388919, loss_mean_cls: 0.080571, grad_norm: 0.728435 +Steps: 0%| | 2181/1000000 [09:05<69:38:25, 3.98it/s, grad_norm=0.728, loss_final=0.554, loss_mean=0.862, loss_mean_cls=0.0806, proj_loss=-0.389][2026-03-26 12:40:08] Step: 2181, Training Logs: loss_final: 0.539230, loss_mean: 0.827035, proj_loss: -0.374291, loss_mean_cls: 0.086486, grad_norm: 0.481165 +Steps: 0%| | 2182/1000000 [09:05<69:21:35, 4.00it/s, grad_norm=0.481, loss_final=0.539, loss_mean=0.827, loss_mean_cls=0.0865, proj_loss=-0.374][2026-03-26 12:40:08] Step: 2182, Training Logs: loss_final: 0.555297, loss_mean: 0.851413, proj_loss: -0.379183, loss_mean_cls: 0.083067, grad_norm: 0.524866 +Steps: 0%| | 2183/1000000 [09:05<69:25:19, 3.99it/s, grad_norm=0.525, loss_final=0.555, loss_mean=0.851, loss_mean_cls=0.0831, proj_loss=-0.379][2026-03-26 12:40:09] Step: 2183, Training Logs: loss_final: 0.551052, loss_mean: 0.856734, proj_loss: -0.386334, loss_mean_cls: 0.080651, grad_norm: 0.769253 +Steps: 0%| | 2184/1000000 [09:06<69:26:57, 3.99it/s, grad_norm=0.769, loss_final=0.551, loss_mean=0.857, loss_mean_cls=0.0807, proj_loss=-0.386][2026-03-26 12:40:09] Step: 2184, Training Logs: loss_final: 0.544445, loss_mean: 0.841529, proj_loss: -0.379296, loss_mean_cls: 0.082212, grad_norm: 0.803565 +Steps: 0%| | 2185/1000000 [09:06<69:10:26, 4.01it/s, grad_norm=0.804, loss_final=0.544, loss_mean=0.842, loss_mean_cls=0.0822, proj_loss=-0.379][2026-03-26 12:40:09] Step: 2185, Training Logs: loss_final: 0.558576, loss_mean: 0.855215, proj_loss: -0.380297, loss_mean_cls: 0.083657, grad_norm: 0.517332 +Steps: 0%| | 2186/1000000 [09:06<69:20:28, 4.00it/s, grad_norm=0.517, loss_final=0.559, loss_mean=0.855, loss_mean_cls=0.0837, proj_loss=-0.38][2026-03-26 12:40:09] Step: 2186, Training Logs: loss_final: 0.556796, loss_mean: 0.866458, proj_loss: -0.390207, loss_mean_cls: 0.080544, grad_norm: 0.664812 +Steps: 0%| | 2187/1000000 [09:06<69:25:42, 3.99it/s, grad_norm=0.665, loss_final=0.557, loss_mean=0.866, loss_mean_cls=0.0805, proj_loss=-0.39][2026-03-26 12:40:10] Step: 2187, Training Logs: loss_final: 0.568977, loss_mean: 0.868272, proj_loss: -0.380742, loss_mean_cls: 0.081446, grad_norm: 0.866716 +Steps: 0%| | 2188/1000000 [09:07<69:11:48, 4.01it/s, grad_norm=0.867, loss_final=0.569, loss_mean=0.868, loss_mean_cls=0.0814, proj_loss=-0.381][2026-03-26 12:40:10] Step: 2188, Training Logs: loss_final: 0.557442, loss_mean: 0.862751, proj_loss: -0.385939, loss_mean_cls: 0.080630, grad_norm: 0.571675 +Steps: 0%| | 2189/1000000 [09:07<68:58:40, 4.02it/s, grad_norm=0.572, loss_final=0.557, loss_mean=0.863, loss_mean_cls=0.0806, proj_loss=-0.386][2026-03-26 12:40:10] Step: 2189, Training Logs: loss_final: 0.560761, loss_mean: 0.871221, proj_loss: -0.389484, loss_mean_cls: 0.079024, grad_norm: 0.646763 +Steps: 0%| | 2190/1000000 [09:07<68:53:29, 4.02it/s, grad_norm=0.647, loss_final=0.561, loss_mean=0.871, loss_mean_cls=0.079, proj_loss=-0.389][2026-03-26 12:40:10] Step: 2190, Training Logs: loss_final: 0.558572, loss_mean: 0.851576, proj_loss: -0.376630, loss_mean_cls: 0.083627, grad_norm: 0.893512 +Steps: 0%| | 2191/1000000 [09:07<68:47:35, 4.03it/s, grad_norm=0.894, loss_final=0.559, loss_mean=0.852, loss_mean_cls=0.0836, proj_loss=-0.377][2026-03-26 12:40:11] Step: 2191, Training Logs: loss_final: 0.553179, loss_mean: 0.853008, proj_loss: -0.381709, loss_mean_cls: 0.081880, grad_norm: 0.637223 +Steps: 0%| | 2192/1000000 [09:08<68:46:38, 4.03it/s, grad_norm=0.637, loss_final=0.553, loss_mean=0.853, loss_mean_cls=0.0819, proj_loss=-0.382][2026-03-26 12:40:11] Step: 2192, Training Logs: loss_final: 0.559814, loss_mean: 0.844985, proj_loss: -0.370712, loss_mean_cls: 0.085542, grad_norm: 0.499564 +Steps: 0%| | 2193/1000000 [09:08<68:44:01, 4.03it/s, grad_norm=0.5, loss_final=0.56, loss_mean=0.845, loss_mean_cls=0.0855, proj_loss=-0.371][2026-03-26 12:40:11] Step: 2193, Training Logs: loss_final: 0.576210, loss_mean: 0.878473, proj_loss: -0.382660, loss_mean_cls: 0.080397, grad_norm: 0.723425 +Steps: 0%| | 2194/1000000 [09:08<68:42:57, 4.03it/s, grad_norm=0.723, loss_final=0.576, loss_mean=0.878, loss_mean_cls=0.0804, proj_loss=-0.383][2026-03-26 12:40:11] Step: 2194, Training Logs: loss_final: 0.544696, loss_mean: 0.838915, proj_loss: -0.377466, loss_mean_cls: 0.083248, grad_norm: 0.511569 +Steps: 0%| | 2195/1000000 [09:08<68:42:48, 4.03it/s, grad_norm=0.512, loss_final=0.545, loss_mean=0.839, loss_mean_cls=0.0832, proj_loss=-0.377][2026-03-26 12:40:12] Step: 2195, Training Logs: loss_final: 0.555972, loss_mean: 0.844810, proj_loss: -0.372950, loss_mean_cls: 0.084113, grad_norm: 0.578811 +Steps: 0%| | 2196/1000000 [09:09<68:43:24, 4.03it/s, grad_norm=0.579, loss_final=0.556, loss_mean=0.845, loss_mean_cls=0.0841, proj_loss=-0.373][2026-03-26 12:40:12] Step: 2196, Training Logs: loss_final: 0.545824, loss_mean: 0.838674, proj_loss: -0.378315, loss_mean_cls: 0.085465, grad_norm: 0.664294 +Steps: 0%| | 2197/1000000 [09:09<68:40:10, 4.04it/s, grad_norm=0.664, loss_final=0.546, loss_mean=0.839, loss_mean_cls=0.0855, proj_loss=-0.378][2026-03-26 12:40:12] Step: 2197, Training Logs: loss_final: 0.543232, loss_mean: 0.845888, proj_loss: -0.384876, loss_mean_cls: 0.082220, grad_norm: 0.702343 +Steps: 0%| | 2198/1000000 [09:09<68:43:53, 4.03it/s, grad_norm=0.702, loss_final=0.543, loss_mean=0.846, loss_mean_cls=0.0822, proj_loss=-0.385][2026-03-26 12:40:12] Step: 2198, Training Logs: loss_final: 0.557159, loss_mean: 0.861473, proj_loss: -0.385846, loss_mean_cls: 0.081532, grad_norm: 0.579940 +Steps: 0%| | 2199/1000000 [09:09<68:40:00, 4.04it/s, grad_norm=0.58, loss_final=0.557, loss_mean=0.861, loss_mean_cls=0.0815, proj_loss=-0.386][2026-03-26 12:40:13] Step: 2199, Training Logs: loss_final: 0.554446, loss_mean: 0.849046, proj_loss: -0.378265, loss_mean_cls: 0.083665, grad_norm: 0.351268 +Steps: 0%| | 2200/1000000 [09:10<68:45:02, 4.03it/s, grad_norm=0.351, loss_final=0.554, loss_mean=0.849, loss_mean_cls=0.0837, proj_loss=-0.378][2026-03-26 12:40:13] Step: 2200, Training Logs: loss_final: 0.561387, loss_mean: 0.870383, proj_loss: -0.390223, loss_mean_cls: 0.081227, grad_norm: 0.633735 +Steps: 0%| | 2201/1000000 [09:10<68:42:47, 4.03it/s, grad_norm=0.634, loss_final=0.561, loss_mean=0.87, loss_mean_cls=0.0812, proj_loss=-0.39][2026-03-26 12:40:13] Step: 2201, Training Logs: loss_final: 0.546358, loss_mean: 0.848665, proj_loss: -0.384796, loss_mean_cls: 0.082489, grad_norm: 0.639277 +Steps: 0%| | 2202/1000000 [09:10<68:41:14, 4.04it/s, grad_norm=0.639, loss_final=0.546, loss_mean=0.849, loss_mean_cls=0.0825, proj_loss=-0.385][2026-03-26 12:40:13] Step: 2202, Training Logs: loss_final: 0.557625, loss_mean: 0.866397, proj_loss: -0.389472, loss_mean_cls: 0.080700, grad_norm: 0.520335 +Steps: 0%| | 2203/1000000 [09:10<68:37:40, 4.04it/s, grad_norm=0.52, loss_final=0.558, loss_mean=0.866, loss_mean_cls=0.0807, proj_loss=-0.389][2026-03-26 12:40:14] Step: 2203, Training Logs: loss_final: 0.542505, loss_mean: 0.826484, proj_loss: -0.371143, loss_mean_cls: 0.087165, grad_norm: 0.858463 +Steps: 0%| | 2204/1000000 [09:11<68:38:40, 4.04it/s, grad_norm=0.858, loss_final=0.543, loss_mean=0.826, loss_mean_cls=0.0872, proj_loss=-0.371][2026-03-26 12:40:14] Step: 2204, Training Logs: loss_final: 0.552752, loss_mean: 0.857210, proj_loss: -0.386220, loss_mean_cls: 0.081762, grad_norm: 0.501349 +Steps: 0%| | 2205/1000000 [09:11<68:36:32, 4.04it/s, grad_norm=0.501, loss_final=0.553, loss_mean=0.857, loss_mean_cls=0.0818, proj_loss=-0.386][2026-03-26 12:40:14] Step: 2205, Training Logs: loss_final: 0.561210, loss_mean: 0.869996, proj_loss: -0.388618, loss_mean_cls: 0.079831, grad_norm: 0.733774 +Steps: 0%| | 2206/1000000 [09:11<68:38:11, 4.04it/s, grad_norm=0.734, loss_final=0.561, loss_mean=0.87, loss_mean_cls=0.0798, proj_loss=-0.389][2026-03-26 12:40:14] Step: 2206, Training Logs: loss_final: 0.559422, loss_mean: 0.871735, proj_loss: -0.391040, loss_mean_cls: 0.078728, grad_norm: 0.574723 +Steps: 0%| | 2207/1000000 [09:11<68:35:55, 4.04it/s, grad_norm=0.575, loss_final=0.559, loss_mean=0.872, loss_mean_cls=0.0787, proj_loss=-0.391][2026-03-26 12:40:15] Step: 2207, Training Logs: loss_final: 0.563091, loss_mean: 0.858767, proj_loss: -0.378353, loss_mean_cls: 0.082677, grad_norm: 0.424063 +Steps: 0%| | 2208/1000000 [09:12<68:36:00, 4.04it/s, grad_norm=0.424, loss_final=0.563, loss_mean=0.859, loss_mean_cls=0.0827, proj_loss=-0.378][2026-03-26 12:40:15] Step: 2208, Training Logs: loss_final: 0.552747, loss_mean: 0.842961, proj_loss: -0.374615, loss_mean_cls: 0.084401, grad_norm: 0.565246 +Steps: 0%| | 2209/1000000 [09:12<68:35:43, 4.04it/s, grad_norm=0.565, loss_final=0.553, loss_mean=0.843, loss_mean_cls=0.0844, proj_loss=-0.375][2026-03-26 12:40:15] Step: 2209, Training Logs: loss_final: 0.555459, loss_mean: 0.862117, proj_loss: -0.388232, loss_mean_cls: 0.081573, grad_norm: 0.351666 +Steps: 0%| | 2210/1000000 [09:12<68:37:02, 4.04it/s, grad_norm=0.352, loss_final=0.555, loss_mean=0.862, loss_mean_cls=0.0816, proj_loss=-0.388][2026-03-26 12:40:15] Step: 2210, Training Logs: loss_final: 0.556460, loss_mean: 0.870709, proj_loss: -0.394044, loss_mean_cls: 0.079796, grad_norm: 0.363375 +Steps: 0%| | 2211/1000000 [09:12<68:37:17, 4.04it/s, grad_norm=0.363, loss_final=0.556, loss_mean=0.871, loss_mean_cls=0.0798, proj_loss=-0.394][2026-03-26 12:40:16] Step: 2211, Training Logs: loss_final: 0.558158, loss_mean: 0.845133, proj_loss: -0.371348, loss_mean_cls: 0.084373, grad_norm: 0.444786 +Steps: 0%| | 2212/1000000 [09:13<68:38:51, 4.04it/s, grad_norm=0.445, loss_final=0.558, loss_mean=0.845, loss_mean_cls=0.0844, proj_loss=-0.371][2026-03-26 12:40:16] Step: 2212, Training Logs: loss_final: 0.563182, loss_mean: 0.863480, proj_loss: -0.382062, loss_mean_cls: 0.081764, grad_norm: 0.394562 +Steps: 0%| | 2213/1000000 [09:13<68:39:25, 4.04it/s, grad_norm=0.395, loss_final=0.563, loss_mean=0.863, loss_mean_cls=0.0818, proj_loss=-0.382][2026-03-26 12:40:16] Step: 2213, Training Logs: loss_final: 0.574274, loss_mean: 0.881175, proj_loss: -0.388126, loss_mean_cls: 0.081225, grad_norm: 0.484201 +Steps: 0%| | 2214/1000000 [09:13<68:38:11, 4.04it/s, grad_norm=0.484, loss_final=0.574, loss_mean=0.881, loss_mean_cls=0.0812, proj_loss=-0.388][2026-03-26 12:40:16] Step: 2214, Training Logs: loss_final: 0.549586, loss_mean: 0.849319, proj_loss: -0.382303, loss_mean_cls: 0.082570, grad_norm: 0.596735 +Steps: 0%| | 2215/1000000 [09:13<68:38:45, 4.04it/s, grad_norm=0.597, loss_final=0.55, loss_mean=0.849, loss_mean_cls=0.0826, proj_loss=-0.382][2026-03-26 12:40:17] Step: 2215, Training Logs: loss_final: 0.549501, loss_mean: 0.839203, proj_loss: -0.375801, loss_mean_cls: 0.086099, grad_norm: 0.592708 +Steps: 0%| | 2216/1000000 [09:14<68:39:40, 4.04it/s, grad_norm=0.593, loss_final=0.55, loss_mean=0.839, loss_mean_cls=0.0861, proj_loss=-0.376][2026-03-26 12:40:17] Step: 2216, Training Logs: loss_final: 0.560318, loss_mean: 0.848583, proj_loss: -0.373205, loss_mean_cls: 0.084940, grad_norm: 0.626167 +Steps: 0%| | 2217/1000000 [09:14<68:38:06, 4.04it/s, grad_norm=0.626, loss_final=0.56, loss_mean=0.849, loss_mean_cls=0.0849, proj_loss=-0.373][2026-03-26 12:40:17] Step: 2217, Training Logs: loss_final: 0.551611, loss_mean: 0.851647, proj_loss: -0.381997, loss_mean_cls: 0.081960, grad_norm: 0.641268 +Steps: 0%| | 2218/1000000 [09:14<68:36:13, 4.04it/s, grad_norm=0.641, loss_final=0.552, loss_mean=0.852, loss_mean_cls=0.082, proj_loss=-0.382][2026-03-26 12:40:17] Step: 2218, Training Logs: loss_final: 0.561659, loss_mean: 0.874148, proj_loss: -0.392295, loss_mean_cls: 0.079806, grad_norm: 0.691909 +Steps: 0%| | 2219/1000000 [09:14<68:37:03, 4.04it/s, grad_norm=0.692, loss_final=0.562, loss_mean=0.874, loss_mean_cls=0.0798, proj_loss=-0.392][2026-03-26 12:40:18] Step: 2219, Training Logs: loss_final: 0.537749, loss_mean: 0.839158, proj_loss: -0.383204, loss_mean_cls: 0.081795, grad_norm: 0.406387 +Steps: 0%| | 2220/1000000 [09:14<68:38:48, 4.04it/s, grad_norm=0.406, loss_final=0.538, loss_mean=0.839, loss_mean_cls=0.0818, proj_loss=-0.383][2026-03-26 12:40:18] Step: 2220, Training Logs: loss_final: 0.543733, loss_mean: 0.831825, proj_loss: -0.372857, loss_mean_cls: 0.084764, grad_norm: 0.664217 +Steps: 0%| | 2221/1000000 [09:15<68:37:38, 4.04it/s, grad_norm=0.664, loss_final=0.544, loss_mean=0.832, loss_mean_cls=0.0848, proj_loss=-0.373][2026-03-26 12:40:18] Step: 2221, Training Logs: loss_final: 0.560199, loss_mean: 0.860336, proj_loss: -0.381527, loss_mean_cls: 0.081390, grad_norm: 0.434622 +Steps: 0%| | 2222/1000000 [09:15<68:36:12, 4.04it/s, grad_norm=0.435, loss_final=0.56, loss_mean=0.86, loss_mean_cls=0.0814, proj_loss=-0.382][2026-03-26 12:40:18] Step: 2222, Training Logs: loss_final: 0.544198, loss_mean: 0.843981, proj_loss: -0.381543, loss_mean_cls: 0.081761, grad_norm: 0.411286 +Steps: 0%| | 2223/1000000 [09:15<68:37:32, 4.04it/s, grad_norm=0.411, loss_final=0.544, loss_mean=0.844, loss_mean_cls=0.0818, proj_loss=-0.382][2026-03-26 12:40:19] Step: 2223, Training Logs: loss_final: 0.541278, loss_mean: 0.838403, proj_loss: -0.380537, loss_mean_cls: 0.083411, grad_norm: 0.579807 +Steps: 0%| | 2224/1000000 [09:15<68:36:14, 4.04it/s, grad_norm=0.58, loss_final=0.541, loss_mean=0.838, loss_mean_cls=0.0834, proj_loss=-0.381][2026-03-26 12:40:19] Step: 2224, Training Logs: loss_final: 0.554989, loss_mean: 0.839405, proj_loss: -0.370203, loss_mean_cls: 0.085787, grad_norm: 0.440268 +Steps: 0%| | 2225/1000000 [09:16<68:36:31, 4.04it/s, grad_norm=0.44, loss_final=0.555, loss_mean=0.839, loss_mean_cls=0.0858, proj_loss=-0.37][2026-03-26 12:40:19] Step: 2225, Training Logs: loss_final: 0.552824, loss_mean: 0.860059, proj_loss: -0.388116, loss_mean_cls: 0.080881, grad_norm: 0.571114 +Steps: 0%| | 2226/1000000 [09:16<68:36:51, 4.04it/s, grad_norm=0.571, loss_final=0.553, loss_mean=0.86, loss_mean_cls=0.0809, proj_loss=-0.388][2026-03-26 12:40:19] Step: 2226, Training Logs: loss_final: 0.542934, loss_mean: 0.849458, proj_loss: -0.388048, loss_mean_cls: 0.081523, grad_norm: 0.511576 +Steps: 0%| | 2227/1000000 [09:16<68:37:10, 4.04it/s, grad_norm=0.512, loss_final=0.543, loss_mean=0.849, loss_mean_cls=0.0815, proj_loss=-0.388][2026-03-26 12:40:20] Step: 2227, Training Logs: loss_final: 0.559446, loss_mean: 0.858191, proj_loss: -0.380699, loss_mean_cls: 0.081954, grad_norm: 0.396165 +Steps: 0%| | 2228/1000000 [09:16<68:59:46, 4.02it/s, grad_norm=0.396, loss_final=0.559, loss_mean=0.858, loss_mean_cls=0.082, proj_loss=-0.381][2026-03-26 12:40:20] Step: 2228, Training Logs: loss_final: 0.561423, loss_mean: 0.866720, proj_loss: -0.386358, loss_mean_cls: 0.081061, grad_norm: 0.505204 +Steps: 0%| | 2229/1000000 [09:17<68:38:05, 4.04it/s, grad_norm=0.505, loss_final=0.561, loss_mean=0.867, loss_mean_cls=0.0811, proj_loss=-0.386][2026-03-26 12:40:20] Step: 2229, Training Logs: loss_final: 0.545377, loss_mean: 0.832729, proj_loss: -0.373796, loss_mean_cls: 0.086443, grad_norm: 0.487966 +Steps: 0%| | 2230/1000000 [09:17<68:36:30, 4.04it/s, grad_norm=0.488, loss_final=0.545, loss_mean=0.833, loss_mean_cls=0.0864, proj_loss=-0.374][2026-03-26 12:40:20] Step: 2230, Training Logs: loss_final: 0.558024, loss_mean: 0.847975, proj_loss: -0.375028, loss_mean_cls: 0.085076, grad_norm: 0.418262 +Steps: 0%| | 2231/1000000 [09:17<68:34:16, 4.04it/s, grad_norm=0.418, loss_final=0.558, loss_mean=0.848, loss_mean_cls=0.0851, proj_loss=-0.375][2026-03-26 12:40:21] Step: 2231, Training Logs: loss_final: 0.550558, loss_mean: 0.853359, proj_loss: -0.384735, loss_mean_cls: 0.081934, grad_norm: 0.533118 +Steps: 0%| | 2232/1000000 [09:17<68:35:18, 4.04it/s, grad_norm=0.533, loss_final=0.551, loss_mean=0.853, loss_mean_cls=0.0819, proj_loss=-0.385][2026-03-26 12:40:21] Step: 2232, Training Logs: loss_final: 0.541339, loss_mean: 0.844040, proj_loss: -0.384360, loss_mean_cls: 0.081659, grad_norm: 0.459992 +Steps: 0%| | 2233/1000000 [09:18<68:32:47, 4.04it/s, grad_norm=0.46, loss_final=0.541, loss_mean=0.844, loss_mean_cls=0.0817, proj_loss=-0.384][2026-03-26 12:40:21] Step: 2233, Training Logs: loss_final: 0.544635, loss_mean: 0.852526, proj_loss: -0.389444, loss_mean_cls: 0.081553, grad_norm: 0.440614 +Steps: 0%| | 2234/1000000 [09:18<68:36:47, 4.04it/s, grad_norm=0.441, loss_final=0.545, loss_mean=0.853, loss_mean_cls=0.0816, proj_loss=-0.389][2026-03-26 12:40:21] Step: 2234, Training Logs: loss_final: 0.557522, loss_mean: 0.863642, proj_loss: -0.386616, loss_mean_cls: 0.080497, grad_norm: 0.668079 +Steps: 0%| | 2235/1000000 [09:18<68:34:58, 4.04it/s, grad_norm=0.668, loss_final=0.558, loss_mean=0.864, loss_mean_cls=0.0805, proj_loss=-0.387][2026-03-26 12:40:22] Step: 2235, Training Logs: loss_final: 0.553884, loss_mean: 0.854455, proj_loss: -0.383171, loss_mean_cls: 0.082600, grad_norm: 0.469329 +Steps: 0%| | 2236/1000000 [09:18<68:35:53, 4.04it/s, grad_norm=0.469, loss_final=0.554, loss_mean=0.854, loss_mean_cls=0.0826, proj_loss=-0.383][2026-03-26 12:40:22] Step: 2236, Training Logs: loss_final: 0.553358, loss_mean: 0.846182, proj_loss: -0.376943, loss_mean_cls: 0.084118, grad_norm: 0.449677 +Steps: 0%| | 2237/1000000 [09:19<68:36:30, 4.04it/s, grad_norm=0.45, loss_final=0.553, loss_mean=0.846, loss_mean_cls=0.0841, proj_loss=-0.377][2026-03-26 12:40:22] Step: 2237, Training Logs: loss_final: 0.559755, loss_mean: 0.868081, proj_loss: -0.389254, loss_mean_cls: 0.080928, grad_norm: 0.408141 +Steps: 0%| | 2238/1000000 [09:19<69:12:53, 4.00it/s, grad_norm=0.408, loss_final=0.56, loss_mean=0.868, loss_mean_cls=0.0809, proj_loss=-0.389][2026-03-26 12:40:22] Step: 2238, Training Logs: loss_final: 0.552253, loss_mean: 0.859959, proj_loss: -0.388596, loss_mean_cls: 0.080890, grad_norm: 0.375948 +Steps: 0%| | 2239/1000000 [09:19<69:01:28, 4.02it/s, grad_norm=0.376, loss_final=0.552, loss_mean=0.86, loss_mean_cls=0.0809, proj_loss=-0.389][2026-03-26 12:40:23] Step: 2239, Training Logs: loss_final: 0.559042, loss_mean: 0.859255, proj_loss: -0.381844, loss_mean_cls: 0.081630, grad_norm: 0.442074 +Steps: 0%| | 2240/1000000 [09:19<68:56:49, 4.02it/s, grad_norm=0.442, loss_final=0.559, loss_mean=0.859, loss_mean_cls=0.0816, proj_loss=-0.382][2026-03-26 12:40:23] Step: 2240, Training Logs: loss_final: 0.533731, loss_mean: 0.852057, proj_loss: -0.396595, loss_mean_cls: 0.078269, grad_norm: 0.523839 +Steps: 0%| | 2241/1000000 [09:20<68:48:27, 4.03it/s, grad_norm=0.524, loss_final=0.534, loss_mean=0.852, loss_mean_cls=0.0783, proj_loss=-0.397][2026-03-26 12:40:23] Step: 2241, Training Logs: loss_final: 0.555582, loss_mean: 0.852101, proj_loss: -0.380890, loss_mean_cls: 0.084371, grad_norm: 0.518561 +Steps: 0%| | 2242/1000000 [09:20<68:44:35, 4.03it/s, grad_norm=0.519, loss_final=0.556, loss_mean=0.852, loss_mean_cls=0.0844, proj_loss=-0.381][2026-03-26 12:40:23] Step: 2242, Training Logs: loss_final: 0.557786, loss_mean: 0.855661, proj_loss: -0.380476, loss_mean_cls: 0.082602, grad_norm: 0.478100 +Steps: 0%| | 2243/1000000 [09:20<68:40:44, 4.04it/s, grad_norm=0.478, loss_final=0.558, loss_mean=0.856, loss_mean_cls=0.0826, proj_loss=-0.38][2026-03-26 12:40:24] Step: 2243, Training Logs: loss_final: 0.551505, loss_mean: 0.843914, proj_loss: -0.375797, loss_mean_cls: 0.083389, grad_norm: 0.444338 +Steps: 0%| | 2244/1000000 [09:20<68:41:55, 4.03it/s, grad_norm=0.444, loss_final=0.552, loss_mean=0.844, loss_mean_cls=0.0834, proj_loss=-0.376][2026-03-26 12:40:24] Step: 2244, Training Logs: loss_final: 0.552066, loss_mean: 0.847364, proj_loss: -0.378128, loss_mean_cls: 0.082830, grad_norm: 0.461291 +Steps: 0%| | 2245/1000000 [09:21<68:38:56, 4.04it/s, grad_norm=0.461, loss_final=0.552, loss_mean=0.847, loss_mean_cls=0.0828, proj_loss=-0.378][2026-03-26 12:40:24] Step: 2245, Training Logs: loss_final: 0.539609, loss_mean: 0.837352, proj_loss: -0.381784, loss_mean_cls: 0.084041, grad_norm: 0.647153 +Steps: 0%| | 2246/1000000 [09:21<68:38:57, 4.04it/s, grad_norm=0.647, loss_final=0.54, loss_mean=0.837, loss_mean_cls=0.084, proj_loss=-0.382][2026-03-26 12:40:24] Step: 2246, Training Logs: loss_final: 0.532779, loss_mean: 0.825048, proj_loss: -0.375843, loss_mean_cls: 0.083574, grad_norm: 0.614011 +Steps: 0%| | 2247/1000000 [09:21<68:36:53, 4.04it/s, grad_norm=0.614, loss_final=0.533, loss_mean=0.825, loss_mean_cls=0.0836, proj_loss=-0.376][2026-03-26 12:40:25] Step: 2247, Training Logs: loss_final: 0.557843, loss_mean: 0.864663, proj_loss: -0.388776, loss_mean_cls: 0.081956, grad_norm: 0.626720 +Steps: 0%| | 2248/1000000 [09:21<68:40:09, 4.04it/s, grad_norm=0.627, loss_final=0.558, loss_mean=0.865, loss_mean_cls=0.082, proj_loss=-0.389][2026-03-26 12:40:25] Step: 2248, Training Logs: loss_final: 0.562251, loss_mean: 0.873929, proj_loss: -0.390828, loss_mean_cls: 0.079150, grad_norm: 0.456049 +Steps: 0%| | 2249/1000000 [09:22<68:36:16, 4.04it/s, grad_norm=0.456, loss_final=0.562, loss_mean=0.874, loss_mean_cls=0.0792, proj_loss=-0.391][2026-03-26 12:40:25] Step: 2249, Training Logs: loss_final: 0.546458, loss_mean: 0.848681, proj_loss: -0.385038, loss_mean_cls: 0.082815, grad_norm: 0.667467 +Steps: 0%| | 2250/1000000 [09:22<68:36:26, 4.04it/s, grad_norm=0.667, loss_final=0.546, loss_mean=0.849, loss_mean_cls=0.0828, proj_loss=-0.385][2026-03-26 12:40:25] Step: 2250, Training Logs: loss_final: 0.547359, loss_mean: 0.858387, proj_loss: -0.391484, loss_mean_cls: 0.080456, grad_norm: 0.521961 +Steps: 0%| | 2251/1000000 [09:22<68:34:21, 4.04it/s, grad_norm=0.522, loss_final=0.547, loss_mean=0.858, loss_mean_cls=0.0805, proj_loss=-0.391][2026-03-26 12:40:26] Step: 2251, Training Logs: loss_final: 0.546889, loss_mean: 0.842281, proj_loss: -0.378900, loss_mean_cls: 0.083508, grad_norm: 0.547818 +Steps: 0%| | 2252/1000000 [09:22<68:36:48, 4.04it/s, grad_norm=0.548, loss_final=0.547, loss_mean=0.842, loss_mean_cls=0.0835, proj_loss=-0.379][2026-03-26 12:40:26] Step: 2252, Training Logs: loss_final: 0.553678, loss_mean: 0.856951, proj_loss: -0.385640, loss_mean_cls: 0.082366, grad_norm: 0.502309 +Steps: 0%| | 2253/1000000 [09:23<68:35:50, 4.04it/s, grad_norm=0.502, loss_final=0.554, loss_mean=0.857, loss_mean_cls=0.0824, proj_loss=-0.386][2026-03-26 12:40:26] Step: 2253, Training Logs: loss_final: 0.558241, loss_mean: 0.863702, proj_loss: -0.387455, loss_mean_cls: 0.081994, grad_norm: 0.762693 +Steps: 0%| | 2254/1000000 [09:23<68:36:05, 4.04it/s, grad_norm=0.763, loss_final=0.558, loss_mean=0.864, loss_mean_cls=0.082, proj_loss=-0.387][2026-03-26 12:40:26] Step: 2254, Training Logs: loss_final: 0.541875, loss_mean: 0.850847, proj_loss: -0.389298, loss_mean_cls: 0.080326, grad_norm: 0.506931 +Steps: 0%| | 2255/1000000 [09:23<68:34:56, 4.04it/s, grad_norm=0.507, loss_final=0.542, loss_mean=0.851, loss_mean_cls=0.0803, proj_loss=-0.389][2026-03-26 12:40:27] Step: 2255, Training Logs: loss_final: 0.561125, loss_mean: 0.860525, proj_loss: -0.382048, loss_mean_cls: 0.082648, grad_norm: 0.603732 +Steps: 0%| | 2256/1000000 [09:23<68:35:20, 4.04it/s, grad_norm=0.604, loss_final=0.561, loss_mean=0.861, loss_mean_cls=0.0826, proj_loss=-0.382][2026-03-26 12:40:27] Step: 2256, Training Logs: loss_final: 0.539927, loss_mean: 0.853895, proj_loss: -0.394331, loss_mean_cls: 0.080363, grad_norm: 0.613554 +Steps: 0%| | 2257/1000000 [09:24<68:35:34, 4.04it/s, grad_norm=0.614, loss_final=0.54, loss_mean=0.854, loss_mean_cls=0.0804, proj_loss=-0.394][2026-03-26 12:40:27] Step: 2257, Training Logs: loss_final: 0.552066, loss_mean: 0.846145, proj_loss: -0.378811, loss_mean_cls: 0.084732, grad_norm: 0.601582 +Steps: 0%| | 2258/1000000 [09:24<68:35:55, 4.04it/s, grad_norm=0.602, loss_final=0.552, loss_mean=0.846, loss_mean_cls=0.0847, proj_loss=-0.379][2026-03-26 12:40:27] Step: 2258, Training Logs: loss_final: 0.552665, loss_mean: 0.844459, proj_loss: -0.376690, loss_mean_cls: 0.084897, grad_norm: 0.543103 +Steps: 0%| | 2259/1000000 [09:24<68:36:29, 4.04it/s, grad_norm=0.543, loss_final=0.553, loss_mean=0.844, loss_mean_cls=0.0849, proj_loss=-0.377][2026-03-26 12:40:28] Step: 2259, Training Logs: loss_final: 0.565814, loss_mean: 0.872021, proj_loss: -0.388609, loss_mean_cls: 0.082402, grad_norm: 0.612770 +Steps: 0%| | 2260/1000000 [09:24<68:36:01, 4.04it/s, grad_norm=0.613, loss_final=0.566, loss_mean=0.872, loss_mean_cls=0.0824, proj_loss=-0.389][2026-03-26 12:40:28] Step: 2260, Training Logs: loss_final: 0.552172, loss_mean: 0.835999, proj_loss: -0.368728, loss_mean_cls: 0.084900, grad_norm: 0.466882 +Steps: 0%| | 2261/1000000 [09:25<68:34:32, 4.04it/s, grad_norm=0.467, loss_final=0.552, loss_mean=0.836, loss_mean_cls=0.0849, proj_loss=-0.369][2026-03-26 12:40:28] Step: 2261, Training Logs: loss_final: 0.540371, loss_mean: 0.840108, proj_loss: -0.383783, loss_mean_cls: 0.084046, grad_norm: 0.466342 +Steps: 0%| | 2262/1000000 [09:25<68:34:56, 4.04it/s, grad_norm=0.466, loss_final=0.54, loss_mean=0.84, loss_mean_cls=0.084, proj_loss=-0.384][2026-03-26 12:40:28] Step: 2262, Training Logs: loss_final: 0.550669, loss_mean: 0.844383, proj_loss: -0.377564, loss_mean_cls: 0.083850, grad_norm: 0.464874 +Steps: 0%| | 2263/1000000 [09:25<68:34:24, 4.04it/s, grad_norm=0.465, loss_final=0.551, loss_mean=0.844, loss_mean_cls=0.0838, proj_loss=-0.378][2026-03-26 12:40:29] Step: 2263, Training Logs: loss_final: 0.546374, loss_mean: 0.838420, proj_loss: -0.376961, loss_mean_cls: 0.084915, grad_norm: 0.440689 +Steps: 0%| | 2264/1000000 [09:25<68:32:31, 4.04it/s, grad_norm=0.441, loss_final=0.546, loss_mean=0.838, loss_mean_cls=0.0849, proj_loss=-0.377][2026-03-26 12:40:29] Step: 2264, Training Logs: loss_final: 0.540415, loss_mean: 0.845822, proj_loss: -0.387409, loss_mean_cls: 0.082002, grad_norm: 0.428136 +Steps: 0%| | 2265/1000000 [09:26<68:34:56, 4.04it/s, grad_norm=0.428, loss_final=0.54, loss_mean=0.846, loss_mean_cls=0.082, proj_loss=-0.387][2026-03-26 12:40:29] Step: 2265, Training Logs: loss_final: 0.547679, loss_mean: 0.851409, proj_loss: -0.386678, loss_mean_cls: 0.082948, grad_norm: 0.518554 +Steps: 0%| | 2266/1000000 [09:26<68:32:08, 4.04it/s, grad_norm=0.519, loss_final=0.548, loss_mean=0.851, loss_mean_cls=0.0829, proj_loss=-0.387][2026-03-26 12:40:29] Step: 2266, Training Logs: loss_final: 0.546548, loss_mean: 0.849321, proj_loss: -0.384467, loss_mean_cls: 0.081694, grad_norm: 0.344418 +Steps: 0%| | 2267/1000000 [09:26<68:33:08, 4.04it/s, grad_norm=0.344, loss_final=0.547, loss_mean=0.849, loss_mean_cls=0.0817, proj_loss=-0.384][2026-03-26 12:40:30] Step: 2267, Training Logs: loss_final: 0.561614, loss_mean: 0.854573, proj_loss: -0.377787, loss_mean_cls: 0.084827, grad_norm: 0.516626 +Steps: 0%| | 2268/1000000 [09:26<68:32:52, 4.04it/s, grad_norm=0.517, loss_final=0.562, loss_mean=0.855, loss_mean_cls=0.0848, proj_loss=-0.378][2026-03-26 12:40:30] Step: 2268, Training Logs: loss_final: 0.555611, loss_mean: 0.850160, proj_loss: -0.378569, loss_mean_cls: 0.084021, grad_norm: 0.443292 +Steps: 0%| | 2269/1000000 [09:27<68:33:50, 4.04it/s, grad_norm=0.443, loss_final=0.556, loss_mean=0.85, loss_mean_cls=0.084, proj_loss=-0.379][2026-03-26 12:40:30] Step: 2269, Training Logs: loss_final: 0.543896, loss_mean: 0.851747, proj_loss: -0.389138, loss_mean_cls: 0.081287, grad_norm: 0.482973 +Steps: 0%| | 2270/1000000 [09:27<68:32:58, 4.04it/s, grad_norm=0.483, loss_final=0.544, loss_mean=0.852, loss_mean_cls=0.0813, proj_loss=-0.389][2026-03-26 12:40:30] Step: 2270, Training Logs: loss_final: 0.555321, loss_mean: 0.863259, proj_loss: -0.388597, loss_mean_cls: 0.080659, grad_norm: 0.424787 +Steps: 0%| | 2271/1000000 [09:27<68:33:10, 4.04it/s, grad_norm=0.425, loss_final=0.555, loss_mean=0.863, loss_mean_cls=0.0807, proj_loss=-0.389][2026-03-26 12:40:31] Step: 2271, Training Logs: loss_final: 0.554276, loss_mean: 0.873523, proj_loss: -0.397713, loss_mean_cls: 0.078466, grad_norm: 0.472665 +Steps: 0%| | 2272/1000000 [09:27<68:34:34, 4.04it/s, grad_norm=0.473, loss_final=0.554, loss_mean=0.874, loss_mean_cls=0.0785, proj_loss=-0.398][2026-03-26 12:40:31] Step: 2272, Training Logs: loss_final: 0.554193, loss_mean: 0.869423, proj_loss: -0.394889, loss_mean_cls: 0.079659, grad_norm: 0.506662 +Steps: 0%| | 2273/1000000 [09:28<68:35:36, 4.04it/s, grad_norm=0.507, loss_final=0.554, loss_mean=0.869, loss_mean_cls=0.0797, proj_loss=-0.395][2026-03-26 12:40:31] Step: 2273, Training Logs: loss_final: 0.566811, loss_mean: 0.871895, proj_loss: -0.386429, loss_mean_cls: 0.081345, grad_norm: 0.333570 +Steps: 0%| | 2274/1000000 [09:28<68:37:50, 4.04it/s, grad_norm=0.334, loss_final=0.567, loss_mean=0.872, loss_mean_cls=0.0813, proj_loss=-0.386][2026-03-26 12:40:31] Step: 2274, Training Logs: loss_final: 0.546926, loss_mean: 0.856216, proj_loss: -0.392278, loss_mean_cls: 0.082989, grad_norm: 0.467951 +Steps: 0%| | 2275/1000000 [09:28<68:37:47, 4.04it/s, grad_norm=0.468, loss_final=0.547, loss_mean=0.856, loss_mean_cls=0.083, proj_loss=-0.392][2026-03-26 12:40:32] Step: 2275, Training Logs: loss_final: 0.549371, loss_mean: 0.851602, proj_loss: -0.384786, loss_mean_cls: 0.082555, grad_norm: 0.574285 +Steps: 0%| | 2276/1000000 [09:28<68:37:53, 4.04it/s, grad_norm=0.574, loss_final=0.549, loss_mean=0.852, loss_mean_cls=0.0826, proj_loss=-0.385][2026-03-26 12:40:32] Step: 2276, Training Logs: loss_final: 0.551217, loss_mean: 0.857327, proj_loss: -0.388253, loss_mean_cls: 0.082143, grad_norm: 0.395457 +Steps: 0%| | 2277/1000000 [09:29<68:37:32, 4.04it/s, grad_norm=0.395, loss_final=0.551, loss_mean=0.857, loss_mean_cls=0.0821, proj_loss=-0.388][2026-03-26 12:40:32] Step: 2277, Training Logs: loss_final: 0.561246, loss_mean: 0.857640, proj_loss: -0.379422, loss_mean_cls: 0.083028, grad_norm: 0.421440 +Steps: 0%| | 2278/1000000 [09:29<68:39:49, 4.04it/s, grad_norm=0.421, loss_final=0.561, loss_mean=0.858, loss_mean_cls=0.083, proj_loss=-0.379][2026-03-26 12:40:32] Step: 2278, Training Logs: loss_final: 0.540820, loss_mean: 0.840237, proj_loss: -0.382814, loss_mean_cls: 0.083398, grad_norm: 0.361336 +Steps: 0%| | 2279/1000000 [09:29<70:08:15, 3.95it/s, grad_norm=0.361, loss_final=0.541, loss_mean=0.84, loss_mean_cls=0.0834, proj_loss=-0.383][2026-03-26 12:40:33] Step: 2279, Training Logs: loss_final: 0.539850, loss_mean: 0.852986, proj_loss: -0.393663, loss_mean_cls: 0.080528, grad_norm: 0.393891 +Steps: 0%| | 2280/1000000 [09:29<69:53:21, 3.97it/s, grad_norm=0.394, loss_final=0.54, loss_mean=0.853, loss_mean_cls=0.0805, proj_loss=-0.394][2026-03-26 12:40:33] Step: 2280, Training Logs: loss_final: 0.530692, loss_mean: 0.841317, proj_loss: -0.391812, loss_mean_cls: 0.081188, grad_norm: 0.358416 +Steps: 0%| | 2281/1000000 [09:30<69:32:29, 3.99it/s, grad_norm=0.358, loss_final=0.531, loss_mean=0.841, loss_mean_cls=0.0812, proj_loss=-0.392][2026-03-26 12:40:33] Step: 2281, Training Logs: loss_final: 0.542910, loss_mean: 0.858787, proj_loss: -0.395422, loss_mean_cls: 0.079545, grad_norm: 0.325124 +Steps: 0%| | 2282/1000000 [09:30<69:20:15, 4.00it/s, grad_norm=0.325, loss_final=0.543, loss_mean=0.859, loss_mean_cls=0.0795, proj_loss=-0.395][2026-03-26 12:40:33] Step: 2282, Training Logs: loss_final: 0.552208, loss_mean: 0.862292, proj_loss: -0.390827, loss_mean_cls: 0.080743, grad_norm: 0.472818 +Steps: 0%| | 2283/1000000 [09:30<69:06:34, 4.01it/s, grad_norm=0.473, loss_final=0.552, loss_mean=0.862, loss_mean_cls=0.0807, proj_loss=-0.391][2026-03-26 12:40:34] Step: 2283, Training Logs: loss_final: 0.535965, loss_mean: 0.847774, proj_loss: -0.392006, loss_mean_cls: 0.080196, grad_norm: 0.346886 +Steps: 0%| | 2284/1000000 [09:30<69:03:53, 4.01it/s, grad_norm=0.347, loss_final=0.536, loss_mean=0.848, loss_mean_cls=0.0802, proj_loss=-0.392][2026-03-26 12:40:34] Step: 2284, Training Logs: loss_final: 0.543335, loss_mean: 0.843753, proj_loss: -0.384361, loss_mean_cls: 0.083943, grad_norm: 0.361347 +Steps: 0%| | 2285/1000000 [09:31<68:54:13, 4.02it/s, grad_norm=0.361, loss_final=0.543, loss_mean=0.844, loss_mean_cls=0.0839, proj_loss=-0.384][2026-03-26 12:40:34] Step: 2285, Training Logs: loss_final: 0.546855, loss_mean: 0.855092, proj_loss: -0.389079, loss_mean_cls: 0.080842, grad_norm: 0.411701 +Steps: 0%| | 2286/1000000 [09:31<68:50:03, 4.03it/s, grad_norm=0.412, loss_final=0.547, loss_mean=0.855, loss_mean_cls=0.0808, proj_loss=-0.389][2026-03-26 12:40:34] Step: 2286, Training Logs: loss_final: 0.544474, loss_mean: 0.852281, proj_loss: -0.389999, loss_mean_cls: 0.082192, grad_norm: 0.389881 +Steps: 0%| | 2287/1000000 [09:31<68:46:28, 4.03it/s, grad_norm=0.39, loss_final=0.544, loss_mean=0.852, loss_mean_cls=0.0822, proj_loss=-0.39][2026-03-26 12:40:35] Step: 2287, Training Logs: loss_final: 0.536400, loss_mean: 0.840850, proj_loss: -0.385791, loss_mean_cls: 0.081341, grad_norm: 0.490041 +Steps: 0%| | 2288/1000000 [09:31<68:44:36, 4.03it/s, grad_norm=0.49, loss_final=0.536, loss_mean=0.841, loss_mean_cls=0.0813, proj_loss=-0.386][2026-03-26 12:40:35] Step: 2288, Training Logs: loss_final: 0.530205, loss_mean: 0.845143, proj_loss: -0.395053, loss_mean_cls: 0.080115, grad_norm: 0.398471 +Steps: 0%| | 2289/1000000 [09:32<68:45:30, 4.03it/s, grad_norm=0.398, loss_final=0.53, loss_mean=0.845, loss_mean_cls=0.0801, proj_loss=-0.395][2026-03-26 12:40:35] Step: 2289, Training Logs: loss_final: 0.553328, loss_mean: 0.850414, proj_loss: -0.381931, loss_mean_cls: 0.084845, grad_norm: 0.453362 +Steps: 0%| | 2290/1000000 [09:32<68:42:20, 4.03it/s, grad_norm=0.453, loss_final=0.553, loss_mean=0.85, loss_mean_cls=0.0848, proj_loss=-0.382][2026-03-26 12:40:35] Step: 2290, Training Logs: loss_final: 0.549501, loss_mean: 0.850545, proj_loss: -0.384303, loss_mean_cls: 0.083259, grad_norm: 0.637942 +Steps: 0%| | 2291/1000000 [09:32<68:37:42, 4.04it/s, grad_norm=0.638, loss_final=0.55, loss_mean=0.851, loss_mean_cls=0.0833, proj_loss=-0.384][2026-03-26 12:40:35] Step: 2291, Training Logs: loss_final: 0.545799, loss_mean: 0.849906, proj_loss: -0.386800, loss_mean_cls: 0.082693, grad_norm: 0.649063 +Steps: 0%| | 2292/1000000 [09:32<68:39:09, 4.04it/s, grad_norm=0.649, loss_final=0.546, loss_mean=0.85, loss_mean_cls=0.0827, proj_loss=-0.387][2026-03-26 12:40:36] Step: 2292, Training Logs: loss_final: 0.555220, loss_mean: 0.862577, proj_loss: -0.389294, loss_mean_cls: 0.081938, grad_norm: 0.711500 +Steps: 0%| | 2293/1000000 [09:33<68:36:09, 4.04it/s, grad_norm=0.712, loss_final=0.555, loss_mean=0.863, loss_mean_cls=0.0819, proj_loss=-0.389][2026-03-26 12:40:36] Step: 2293, Training Logs: loss_final: 0.525251, loss_mean: 0.823329, proj_loss: -0.383098, loss_mean_cls: 0.085020, grad_norm: 0.429775 +Steps: 0%| | 2294/1000000 [09:33<68:38:09, 4.04it/s, grad_norm=0.43, loss_final=0.525, loss_mean=0.823, loss_mean_cls=0.085, proj_loss=-0.383][2026-03-26 12:40:36] Step: 2294, Training Logs: loss_final: 0.537344, loss_mean: 0.851239, proj_loss: -0.393826, loss_mean_cls: 0.079931, grad_norm: 0.429437 +Steps: 0%| | 2295/1000000 [09:33<68:35:45, 4.04it/s, grad_norm=0.429, loss_final=0.537, loss_mean=0.851, loss_mean_cls=0.0799, proj_loss=-0.394][2026-03-26 12:40:36] Step: 2295, Training Logs: loss_final: 0.529724, loss_mean: 0.835172, proj_loss: -0.387710, loss_mean_cls: 0.082262, grad_norm: 0.547423 +Steps: 0%| | 2296/1000000 [09:33<68:36:08, 4.04it/s, grad_norm=0.547, loss_final=0.53, loss_mean=0.835, loss_mean_cls=0.0823, proj_loss=-0.388][2026-03-26 12:40:37] Step: 2296, Training Logs: loss_final: 0.548249, loss_mean: 0.854172, proj_loss: -0.387977, loss_mean_cls: 0.082054, grad_norm: 0.610567 +Steps: 0%| | 2297/1000000 [09:34<68:34:14, 4.04it/s, grad_norm=0.611, loss_final=0.548, loss_mean=0.854, loss_mean_cls=0.0821, proj_loss=-0.388][2026-03-26 12:40:37] Step: 2297, Training Logs: loss_final: 0.556535, loss_mean: 0.867427, proj_loss: -0.391882, loss_mean_cls: 0.080990, grad_norm: 0.495835 +Steps: 0%| | 2298/1000000 [09:34<68:35:33, 4.04it/s, grad_norm=0.496, loss_final=0.557, loss_mean=0.867, loss_mean_cls=0.081, proj_loss=-0.392][2026-03-26 12:40:37] Step: 2298, Training Logs: loss_final: 0.563391, loss_mean: 0.855284, proj_loss: -0.375862, loss_mean_cls: 0.083969, grad_norm: 0.424200 +Steps: 0%| | 2299/1000000 [09:34<69:41:57, 3.98it/s, grad_norm=0.424, loss_final=0.563, loss_mean=0.855, loss_mean_cls=0.084, proj_loss=-0.376][2026-03-26 12:40:37] Step: 2299, Training Logs: loss_final: 0.554657, loss_mean: 0.865068, proj_loss: -0.390804, loss_mean_cls: 0.080393, grad_norm: 0.563030 +Steps: 0%| | 2300/1000000 [09:34<69:26:22, 3.99it/s, grad_norm=0.563, loss_final=0.555, loss_mean=0.865, loss_mean_cls=0.0804, proj_loss=-0.391][2026-03-26 12:40:38] Step: 2300, Training Logs: loss_final: 0.542849, loss_mean: 0.839518, proj_loss: -0.379204, loss_mean_cls: 0.082535, grad_norm: 0.611259 +Steps: 0%| | 2301/1000000 [09:35<69:10:32, 4.01it/s, grad_norm=0.611, loss_final=0.543, loss_mean=0.84, loss_mean_cls=0.0825, proj_loss=-0.379][2026-03-26 12:40:38] Step: 2301, Training Logs: loss_final: 0.544494, loss_mean: 0.847616, proj_loss: -0.385681, loss_mean_cls: 0.082558, grad_norm: 0.491178 +Steps: 0%| | 2302/1000000 [09:35<68:59:09, 4.02it/s, grad_norm=0.491, loss_final=0.544, loss_mean=0.848, loss_mean_cls=0.0826, proj_loss=-0.386][2026-03-26 12:40:38] Step: 2302, Training Logs: loss_final: 0.544655, loss_mean: 0.850465, proj_loss: -0.387692, loss_mean_cls: 0.081882, grad_norm: 0.407172 +Steps: 0%| | 2303/1000000 [09:35<68:55:01, 4.02it/s, grad_norm=0.407, loss_final=0.545, loss_mean=0.85, loss_mean_cls=0.0819, proj_loss=-0.388][2026-03-26 12:40:38] Step: 2303, Training Logs: loss_final: 0.547162, loss_mean: 0.842016, proj_loss: -0.380251, loss_mean_cls: 0.085397, grad_norm: 0.689598 +Steps: 0%| | 2304/1000000 [09:35<68:50:13, 4.03it/s, grad_norm=0.69, loss_final=0.547, loss_mean=0.842, loss_mean_cls=0.0854, proj_loss=-0.38][2026-03-26 12:40:39] Step: 2304, Training Logs: loss_final: 0.522960, loss_mean: 0.828050, proj_loss: -0.387043, loss_mean_cls: 0.081953, grad_norm: 0.717010 +Steps: 0%| | 2305/1000000 [09:36<68:44:37, 4.03it/s, grad_norm=0.717, loss_final=0.523, loss_mean=0.828, loss_mean_cls=0.082, proj_loss=-0.387][2026-03-26 12:40:39] Step: 2305, Training Logs: loss_final: 0.549070, loss_mean: 0.854219, proj_loss: -0.387219, loss_mean_cls: 0.082070, grad_norm: 0.768286 +Steps: 0%| | 2306/1000000 [09:36<68:42:41, 4.03it/s, grad_norm=0.768, loss_final=0.549, loss_mean=0.854, loss_mean_cls=0.0821, proj_loss=-0.387][2026-03-26 12:40:39] Step: 2306, Training Logs: loss_final: 0.550658, loss_mean: 0.858924, proj_loss: -0.389320, loss_mean_cls: 0.081054, grad_norm: 0.608434 +Steps: 0%| | 2307/1000000 [09:36<68:40:17, 4.04it/s, grad_norm=0.608, loss_final=0.551, loss_mean=0.859, loss_mean_cls=0.0811, proj_loss=-0.389][2026-03-26 12:40:39] Step: 2307, Training Logs: loss_final: 0.532805, loss_mean: 0.830003, proj_loss: -0.381748, loss_mean_cls: 0.084550, grad_norm: 0.448745 +Steps: 0%| | 2308/1000000 [09:36<68:42:08, 4.03it/s, grad_norm=0.449, loss_final=0.533, loss_mean=0.83, loss_mean_cls=0.0845, proj_loss=-0.382][2026-03-26 12:40:40] Step: 2308, Training Logs: loss_final: 0.520976, loss_mean: 0.813964, proj_loss: -0.378159, loss_mean_cls: 0.085171, grad_norm: 0.584446 +Steps: 0%| | 2309/1000000 [09:37<68:39:07, 4.04it/s, grad_norm=0.584, loss_final=0.521, loss_mean=0.814, loss_mean_cls=0.0852, proj_loss=-0.378][2026-03-26 12:40:40] Step: 2309, Training Logs: loss_final: 0.534705, loss_mean: 0.829562, proj_loss: -0.381523, loss_mean_cls: 0.086665, grad_norm: 0.454788 +Steps: 0%| | 2310/1000000 [09:37<68:38:10, 4.04it/s, grad_norm=0.455, loss_final=0.535, loss_mean=0.83, loss_mean_cls=0.0867, proj_loss=-0.382][2026-03-26 12:40:40] Step: 2310, Training Logs: loss_final: 0.546692, loss_mean: 0.855061, proj_loss: -0.388993, loss_mean_cls: 0.080625, grad_norm: 0.448385 +Steps: 0%| | 2311/1000000 [09:37<68:36:13, 4.04it/s, grad_norm=0.448, loss_final=0.547, loss_mean=0.855, loss_mean_cls=0.0806, proj_loss=-0.389][2026-03-26 12:40:40] Step: 2311, Training Logs: loss_final: 0.537621, loss_mean: 0.832394, proj_loss: -0.379214, loss_mean_cls: 0.084441, grad_norm: 0.460536 +Steps: 0%| | 2312/1000000 [09:37<68:35:25, 4.04it/s, grad_norm=0.461, loss_final=0.538, loss_mean=0.832, loss_mean_cls=0.0844, proj_loss=-0.379][2026-03-26 12:40:41] Step: 2312, Training Logs: loss_final: 0.538307, loss_mean: 0.844168, proj_loss: -0.387644, loss_mean_cls: 0.081783, grad_norm: 0.555700 +Steps: 0%| | 2313/1000000 [09:38<68:36:16, 4.04it/s, grad_norm=0.556, loss_final=0.538, loss_mean=0.844, loss_mean_cls=0.0818, proj_loss=-0.388][2026-03-26 12:40:41] Step: 2313, Training Logs: loss_final: 0.538631, loss_mean: 0.837179, proj_loss: -0.382805, loss_mean_cls: 0.084257, grad_norm: 0.419918 +Steps: 0%| | 2314/1000000 [09:38<68:36:05, 4.04it/s, grad_norm=0.42, loss_final=0.539, loss_mean=0.837, loss_mean_cls=0.0843, proj_loss=-0.383][2026-03-26 12:40:41] Step: 2314, Training Logs: loss_final: 0.543728, loss_mean: 0.838906, proj_loss: -0.378841, loss_mean_cls: 0.083662, grad_norm: 0.476593 +Steps: 0%| | 2315/1000000 [09:38<68:36:42, 4.04it/s, grad_norm=0.477, loss_final=0.544, loss_mean=0.839, loss_mean_cls=0.0837, proj_loss=-0.379][2026-03-26 12:40:41] Step: 2315, Training Logs: loss_final: 0.551003, loss_mean: 0.862135, proj_loss: -0.392280, loss_mean_cls: 0.081148, grad_norm: 0.432579 +Steps: 0%| | 2316/1000000 [09:38<68:38:20, 4.04it/s, grad_norm=0.433, loss_final=0.551, loss_mean=0.862, loss_mean_cls=0.0811, proj_loss=-0.392][2026-03-26 12:40:42] Step: 2316, Training Logs: loss_final: 0.546267, loss_mean: 0.858299, proj_loss: -0.392563, loss_mean_cls: 0.080532, grad_norm: 0.354603 +Steps: 0%| | 2317/1000000 [09:39<68:38:39, 4.04it/s, grad_norm=0.355, loss_final=0.546, loss_mean=0.858, loss_mean_cls=0.0805, proj_loss=-0.393][2026-03-26 12:40:42] Step: 2317, Training Logs: loss_final: 0.548722, loss_mean: 0.848276, proj_loss: -0.383673, loss_mean_cls: 0.084119, grad_norm: 0.431819 +Steps: 0%| | 2318/1000000 [09:39<68:37:12, 4.04it/s, grad_norm=0.432, loss_final=0.549, loss_mean=0.848, loss_mean_cls=0.0841, proj_loss=-0.384][2026-03-26 12:40:42] Step: 2318, Training Logs: loss_final: 0.551064, loss_mean: 0.854750, proj_loss: -0.385873, loss_mean_cls: 0.082186, grad_norm: 0.460780 +Steps: 0%| | 2319/1000000 [09:39<68:37:59, 4.04it/s, grad_norm=0.461, loss_final=0.551, loss_mean=0.855, loss_mean_cls=0.0822, proj_loss=-0.386][2026-03-26 12:40:42] Step: 2319, Training Logs: loss_final: 0.546032, loss_mean: 0.850620, proj_loss: -0.386741, loss_mean_cls: 0.082153, grad_norm: 0.430940 +Steps: 0%| | 2320/1000000 [09:39<68:37:18, 4.04it/s, grad_norm=0.431, loss_final=0.546, loss_mean=0.851, loss_mean_cls=0.0822, proj_loss=-0.387][2026-03-26 12:40:43] Step: 2320, Training Logs: loss_final: 0.532848, loss_mean: 0.840962, proj_loss: -0.390767, loss_mean_cls: 0.082654, grad_norm: 0.477423 +Steps: 0%| | 2321/1000000 [09:40<70:11:52, 3.95it/s, grad_norm=0.477, loss_final=0.533, loss_mean=0.841, loss_mean_cls=0.0827, proj_loss=-0.391][2026-03-26 12:40:43] Step: 2321, Training Logs: loss_final: 0.538402, loss_mean: 0.827615, proj_loss: -0.375723, loss_mean_cls: 0.086509, grad_norm: 0.333382 +Steps: 0%| | 2322/1000000 [09:40<68:54:13, 4.02it/s, grad_norm=0.333, loss_final=0.538, loss_mean=0.828, loss_mean_cls=0.0865, proj_loss=-0.376][2026-03-26 12:40:43] Step: 2322, Training Logs: loss_final: 0.526271, loss_mean: 0.833098, proj_loss: -0.389093, loss_mean_cls: 0.082265, grad_norm: 0.392364 +Steps: 0%| | 2323/1000000 [09:40<68:49:49, 4.03it/s, grad_norm=0.392, loss_final=0.526, loss_mean=0.833, loss_mean_cls=0.0823, proj_loss=-0.389][2026-03-26 12:40:43] Step: 2323, Training Logs: loss_final: 0.541679, loss_mean: 0.839014, proj_loss: -0.381030, loss_mean_cls: 0.083694, grad_norm: 0.448577 +Steps: 0%| | 2324/1000000 [09:40<68:44:31, 4.03it/s, grad_norm=0.449, loss_final=0.542, loss_mean=0.839, loss_mean_cls=0.0837, proj_loss=-0.381][2026-03-26 12:40:44] Step: 2324, Training Logs: loss_final: 0.546865, loss_mean: 0.858490, proj_loss: -0.392394, loss_mean_cls: 0.080770, grad_norm: 0.388418 +Steps: 0%| | 2325/1000000 [09:41<68:47:19, 4.03it/s, grad_norm=0.388, loss_final=0.547, loss_mean=0.858, loss_mean_cls=0.0808, proj_loss=-0.392][2026-03-26 12:40:44] Step: 2325, Training Logs: loss_final: 0.546068, loss_mean: 0.859237, proj_loss: -0.393971, loss_mean_cls: 0.080802, grad_norm: 0.519814 +Steps: 0%| | 2326/1000000 [09:41<68:40:45, 4.04it/s, grad_norm=0.52, loss_final=0.546, loss_mean=0.859, loss_mean_cls=0.0808, proj_loss=-0.394][2026-03-26 12:40:44] Step: 2326, Training Logs: loss_final: 0.547875, loss_mean: 0.838136, proj_loss: -0.375917, loss_mean_cls: 0.085656, grad_norm: 0.545045 +Steps: 0%| | 2327/1000000 [09:41<68:41:47, 4.03it/s, grad_norm=0.545, loss_final=0.548, loss_mean=0.838, loss_mean_cls=0.0857, proj_loss=-0.376][2026-03-26 12:40:44] Step: 2327, Training Logs: loss_final: 0.537496, loss_mean: 0.845303, proj_loss: -0.390615, loss_mean_cls: 0.082807, grad_norm: 0.371322 +Steps: 0%| | 2328/1000000 [09:41<68:37:10, 4.04it/s, grad_norm=0.371, loss_final=0.537, loss_mean=0.845, loss_mean_cls=0.0828, proj_loss=-0.391][2026-03-26 12:40:45] Step: 2328, Training Logs: loss_final: 0.545623, loss_mean: 0.857729, proj_loss: -0.391389, loss_mean_cls: 0.079282, grad_norm: 0.543552 +Steps: 0%| | 2329/1000000 [09:42<68:37:27, 4.04it/s, grad_norm=0.544, loss_final=0.546, loss_mean=0.858, loss_mean_cls=0.0793, proj_loss=-0.391][2026-03-26 12:40:45] Step: 2329, Training Logs: loss_final: 0.554323, loss_mean: 0.864172, proj_loss: -0.391175, loss_mean_cls: 0.081326, grad_norm: 0.472241 +Steps: 0%| | 2330/1000000 [09:42<68:34:35, 4.04it/s, grad_norm=0.472, loss_final=0.554, loss_mean=0.864, loss_mean_cls=0.0813, proj_loss=-0.391][2026-03-26 12:40:45] Step: 2330, Training Logs: loss_final: 0.541582, loss_mean: 0.859501, proj_loss: -0.395932, loss_mean_cls: 0.078014, grad_norm: 0.346398 +Steps: 0%| | 2331/1000000 [09:42<68:35:18, 4.04it/s, grad_norm=0.346, loss_final=0.542, loss_mean=0.86, loss_mean_cls=0.078, proj_loss=-0.396][2026-03-26 12:40:45] Step: 2331, Training Logs: loss_final: 0.549142, loss_mean: 0.843574, proj_loss: -0.379085, loss_mean_cls: 0.084654, grad_norm: 0.546344 +Steps: 0%| | 2332/1000000 [09:42<68:36:43, 4.04it/s, grad_norm=0.546, loss_final=0.549, loss_mean=0.844, loss_mean_cls=0.0847, proj_loss=-0.379][2026-03-26 12:40:46] Step: 2332, Training Logs: loss_final: 0.538355, loss_mean: 0.837922, proj_loss: -0.383282, loss_mean_cls: 0.083715, grad_norm: 0.338586 +Steps: 0%| | 2333/1000000 [09:43<68:35:39, 4.04it/s, grad_norm=0.339, loss_final=0.538, loss_mean=0.838, loss_mean_cls=0.0837, proj_loss=-0.383][2026-03-26 12:40:46] Step: 2333, Training Logs: loss_final: 0.543262, loss_mean: 0.861480, proj_loss: -0.397848, loss_mean_cls: 0.079630, grad_norm: 0.617534 +Steps: 0%| | 2334/1000000 [09:43<68:35:41, 4.04it/s, grad_norm=0.618, loss_final=0.543, loss_mean=0.861, loss_mean_cls=0.0796, proj_loss=-0.398][2026-03-26 12:40:46] Step: 2334, Training Logs: loss_final: 0.553320, loss_mean: 0.838791, proj_loss: -0.371667, loss_mean_cls: 0.086196, grad_norm: 0.392768 +Steps: 0%| | 2335/1000000 [09:43<68:35:22, 4.04it/s, grad_norm=0.393, loss_final=0.553, loss_mean=0.839, loss_mean_cls=0.0862, proj_loss=-0.372][2026-03-26 12:40:46] Step: 2335, Training Logs: loss_final: 0.544384, loss_mean: 0.852716, proj_loss: -0.390326, loss_mean_cls: 0.081994, grad_norm: 0.443491 +Steps: 0%| | 2336/1000000 [09:43<68:38:51, 4.04it/s, grad_norm=0.443, loss_final=0.544, loss_mean=0.853, loss_mean_cls=0.082, proj_loss=-0.39][2026-03-26 12:40:47] Step: 2336, Training Logs: loss_final: 0.547318, loss_mean: 0.848673, proj_loss: -0.384466, loss_mean_cls: 0.083111, grad_norm: 0.509062 +Steps: 0%| | 2337/1000000 [09:44<68:41:42, 4.03it/s, grad_norm=0.509, loss_final=0.547, loss_mean=0.849, loss_mean_cls=0.0831, proj_loss=-0.384][2026-03-26 12:40:47] Step: 2337, Training Logs: loss_final: 0.530738, loss_mean: 0.838613, proj_loss: -0.389023, loss_mean_cls: 0.081149, grad_norm: 0.495077 +Steps: 0%| | 2338/1000000 [09:44<68:42:50, 4.03it/s, grad_norm=0.495, loss_final=0.531, loss_mean=0.839, loss_mean_cls=0.0811, proj_loss=-0.389][2026-03-26 12:40:47] Step: 2338, Training Logs: loss_final: 0.553020, loss_mean: 0.851230, proj_loss: -0.381306, loss_mean_cls: 0.083097, grad_norm: 0.468763 +Steps: 0%| | 2339/1000000 [09:44<68:39:58, 4.04it/s, grad_norm=0.469, loss_final=0.553, loss_mean=0.851, loss_mean_cls=0.0831, proj_loss=-0.381][2026-03-26 12:40:47] Step: 2339, Training Logs: loss_final: 0.555292, loss_mean: 0.847273, proj_loss: -0.377217, loss_mean_cls: 0.085236, grad_norm: 0.497873 +Steps: 0%| | 2340/1000000 [09:44<68:38:26, 4.04it/s, grad_norm=0.498, loss_final=0.555, loss_mean=0.847, loss_mean_cls=0.0852, proj_loss=-0.377][2026-03-26 12:40:48] Step: 2340, Training Logs: loss_final: 0.544439, loss_mean: 0.844005, proj_loss: -0.381852, loss_mean_cls: 0.082286, grad_norm: 0.503703 +Steps: 0%| | 2341/1000000 [09:45<68:37:54, 4.04it/s, grad_norm=0.504, loss_final=0.544, loss_mean=0.844, loss_mean_cls=0.0823, proj_loss=-0.382][2026-03-26 12:40:48] Step: 2341, Training Logs: loss_final: 0.542551, loss_mean: 0.855742, proj_loss: -0.393574, loss_mean_cls: 0.080383, grad_norm: 0.410102 +Steps: 0%| | 2342/1000000 [09:45<68:42:19, 4.03it/s, grad_norm=0.41, loss_final=0.543, loss_mean=0.856, loss_mean_cls=0.0804, proj_loss=-0.394][2026-03-26 12:40:48] Step: 2342, Training Logs: loss_final: 0.547314, loss_mean: 0.862467, proj_loss: -0.394462, loss_mean_cls: 0.079309, grad_norm: 0.452434 +Steps: 0%| | 2343/1000000 [09:45<68:34:38, 4.04it/s, grad_norm=0.452, loss_final=0.547, loss_mean=0.862, loss_mean_cls=0.0793, proj_loss=-0.394][2026-03-26 12:40:48] Step: 2343, Training Logs: loss_final: 0.555124, loss_mean: 0.853152, proj_loss: -0.381233, loss_mean_cls: 0.083205, grad_norm: 0.593198 +Steps: 0%| | 2344/1000000 [09:45<68:35:14, 4.04it/s, grad_norm=0.593, loss_final=0.555, loss_mean=0.853, loss_mean_cls=0.0832, proj_loss=-0.381][2026-03-26 12:40:49] Step: 2344, Training Logs: loss_final: 0.542734, loss_mean: 0.866789, proj_loss: -0.401209, loss_mean_cls: 0.077153, grad_norm: 0.439830 +Steps: 0%| | 2345/1000000 [09:45<68:35:44, 4.04it/s, grad_norm=0.44, loss_final=0.543, loss_mean=0.867, loss_mean_cls=0.0772, proj_loss=-0.401][2026-03-26 12:40:49] Step: 2345, Training Logs: loss_final: 0.541546, loss_mean: 0.850966, proj_loss: -0.390449, loss_mean_cls: 0.081029, grad_norm: 0.525448 +Steps: 0%| | 2346/1000000 [09:46<68:35:54, 4.04it/s, grad_norm=0.525, loss_final=0.542, loss_mean=0.851, loss_mean_cls=0.081, proj_loss=-0.39][2026-03-26 12:40:49] Step: 2346, Training Logs: loss_final: 0.548146, loss_mean: 0.851289, proj_loss: -0.386753, loss_mean_cls: 0.083610, grad_norm: 0.599697 +Steps: 0%| | 2347/1000000 [09:46<68:33:49, 4.04it/s, grad_norm=0.6, loss_final=0.548, loss_mean=0.851, loss_mean_cls=0.0836, proj_loss=-0.387][2026-03-26 12:40:49] Step: 2347, Training Logs: loss_final: 0.524122, loss_mean: 0.832619, proj_loss: -0.391033, loss_mean_cls: 0.082536, grad_norm: 0.612368 +Steps: 0%| | 2348/1000000 [09:46<68:53:27, 4.02it/s, grad_norm=0.612, loss_final=0.524, loss_mean=0.833, loss_mean_cls=0.0825, proj_loss=-0.391][2026-03-26 12:40:50] Step: 2348, Training Logs: loss_final: 0.543243, loss_mean: 0.849775, proj_loss: -0.389442, loss_mean_cls: 0.082910, grad_norm: 0.471510 +Steps: 0%| | 2349/1000000 [09:46<68:48:05, 4.03it/s, grad_norm=0.472, loss_final=0.543, loss_mean=0.85, loss_mean_cls=0.0829, proj_loss=-0.389][2026-03-26 12:40:50] Step: 2349, Training Logs: loss_final: 0.532110, loss_mean: 0.839936, proj_loss: -0.389189, loss_mean_cls: 0.081362, grad_norm: 0.438335 +Steps: 0%| | 2350/1000000 [09:47<68:47:41, 4.03it/s, grad_norm=0.438, loss_final=0.532, loss_mean=0.84, loss_mean_cls=0.0814, proj_loss=-0.389][2026-03-26 12:40:50] Step: 2350, Training Logs: loss_final: 0.547840, loss_mean: 0.859837, proj_loss: -0.393823, loss_mean_cls: 0.081826, grad_norm: 0.543775 +Steps: 0%| | 2351/1000000 [09:47<68:43:33, 4.03it/s, grad_norm=0.544, loss_final=0.548, loss_mean=0.86, loss_mean_cls=0.0818, proj_loss=-0.394][2026-03-26 12:40:50] Step: 2351, Training Logs: loss_final: 0.533180, loss_mean: 0.832005, proj_loss: -0.382488, loss_mean_cls: 0.083663, grad_norm: 0.408780 +Steps: 0%| | 2352/1000000 [09:47<68:42:08, 4.03it/s, grad_norm=0.409, loss_final=0.533, loss_mean=0.832, loss_mean_cls=0.0837, proj_loss=-0.382][2026-03-26 12:40:51] Step: 2352, Training Logs: loss_final: 0.553160, loss_mean: 0.858512, proj_loss: -0.387093, loss_mean_cls: 0.081740, grad_norm: 0.424980 +Steps: 0%| | 2353/1000000 [09:47<68:40:37, 4.04it/s, grad_norm=0.425, loss_final=0.553, loss_mean=0.859, loss_mean_cls=0.0817, proj_loss=-0.387][2026-03-26 12:40:51] Step: 2353, Training Logs: loss_final: 0.530913, loss_mean: 0.839332, proj_loss: -0.390363, loss_mean_cls: 0.081944, grad_norm: 0.472047 +Steps: 0%| | 2354/1000000 [09:48<68:41:06, 4.03it/s, grad_norm=0.472, loss_final=0.531, loss_mean=0.839, loss_mean_cls=0.0819, proj_loss=-0.39][2026-03-26 12:40:51] Step: 2354, Training Logs: loss_final: 0.555470, loss_mean: 0.852629, proj_loss: -0.381047, loss_mean_cls: 0.083888, grad_norm: 0.484916 +Steps: 0%| | 2355/1000000 [09:48<68:38:14, 4.04it/s, grad_norm=0.485, loss_final=0.555, loss_mean=0.853, loss_mean_cls=0.0839, proj_loss=-0.381][2026-03-26 12:40:51] Step: 2355, Training Logs: loss_final: 0.544628, loss_mean: 0.848803, proj_loss: -0.387267, loss_mean_cls: 0.083092, grad_norm: 0.397145 +Steps: 0%| | 2356/1000000 [09:48<68:44:39, 4.03it/s, grad_norm=0.397, loss_final=0.545, loss_mean=0.849, loss_mean_cls=0.0831, proj_loss=-0.387][2026-03-26 12:40:52] Step: 2356, Training Logs: loss_final: 0.548797, loss_mean: 0.865943, proj_loss: -0.396976, loss_mean_cls: 0.079830, grad_norm: 0.390494 +Steps: 0%| | 2357/1000000 [09:48<68:48:30, 4.03it/s, grad_norm=0.39, loss_final=0.549, loss_mean=0.866, loss_mean_cls=0.0798, proj_loss=-0.397][2026-03-26 12:40:52] Step: 2357, Training Logs: loss_final: 0.545520, loss_mean: 0.853227, proj_loss: -0.390591, loss_mean_cls: 0.082884, grad_norm: 0.360030 +Steps: 0%| | 2358/1000000 [09:49<68:44:04, 4.03it/s, grad_norm=0.36, loss_final=0.546, loss_mean=0.853, loss_mean_cls=0.0829, proj_loss=-0.391][2026-03-26 12:40:52] Step: 2358, Training Logs: loss_final: 0.551479, loss_mean: 0.866183, proj_loss: -0.394719, loss_mean_cls: 0.080016, grad_norm: 0.411070 +Steps: 0%| | 2359/1000000 [09:49<68:42:24, 4.03it/s, grad_norm=0.411, loss_final=0.551, loss_mean=0.866, loss_mean_cls=0.08, proj_loss=-0.395][2026-03-26 12:40:52] Step: 2359, Training Logs: loss_final: 0.543245, loss_mean: 0.846571, proj_loss: -0.386252, loss_mean_cls: 0.082926, grad_norm: 0.399572 +Steps: 0%| | 2360/1000000 [09:49<68:40:08, 4.04it/s, grad_norm=0.4, loss_final=0.543, loss_mean=0.847, loss_mean_cls=0.0829, proj_loss=-0.386][2026-03-26 12:40:53] Step: 2360, Training Logs: loss_final: 0.517503, loss_mean: 0.817132, proj_loss: -0.384112, loss_mean_cls: 0.084484, grad_norm: 0.429949 +Steps: 0%| | 2361/1000000 [09:49<68:36:58, 4.04it/s, grad_norm=0.43, loss_final=0.518, loss_mean=0.817, loss_mean_cls=0.0845, proj_loss=-0.384][2026-03-26 12:40:53] Step: 2361, Training Logs: loss_final: 0.534970, loss_mean: 0.848410, proj_loss: -0.393838, loss_mean_cls: 0.080398, grad_norm: 0.434382 +Steps: 0%| | 2362/1000000 [09:50<68:36:22, 4.04it/s, grad_norm=0.434, loss_final=0.535, loss_mean=0.848, loss_mean_cls=0.0804, proj_loss=-0.394][2026-03-26 12:40:53] Step: 2362, Training Logs: loss_final: 0.554040, loss_mean: 0.851164, proj_loss: -0.381135, loss_mean_cls: 0.084011, grad_norm: 0.386858 +Steps: 0%| | 2363/1000000 [09:50<68:36:06, 4.04it/s, grad_norm=0.387, loss_final=0.554, loss_mean=0.851, loss_mean_cls=0.084, proj_loss=-0.381][2026-03-26 12:40:53] Step: 2363, Training Logs: loss_final: 0.537255, loss_mean: 0.840891, proj_loss: -0.386349, loss_mean_cls: 0.082713, grad_norm: 0.553668 +Steps: 0%| | 2364/1000000 [09:50<68:42:38, 4.03it/s, grad_norm=0.554, loss_final=0.537, loss_mean=0.841, loss_mean_cls=0.0827, proj_loss=-0.386][2026-03-26 12:40:54] Step: 2364, Training Logs: loss_final: 0.544276, loss_mean: 0.853171, proj_loss: -0.390626, loss_mean_cls: 0.081732, grad_norm: 0.425018 +Steps: 0%| | 2365/1000000 [09:50<68:40:48, 4.03it/s, grad_norm=0.425, loss_final=0.544, loss_mean=0.853, loss_mean_cls=0.0817, proj_loss=-0.391][2026-03-26 12:40:54] Step: 2365, Training Logs: loss_final: 0.535680, loss_mean: 0.832334, proj_loss: -0.381305, loss_mean_cls: 0.084650, grad_norm: 0.421623 +Steps: 0%| | 2366/1000000 [09:51<68:39:01, 4.04it/s, grad_norm=0.422, loss_final=0.536, loss_mean=0.832, loss_mean_cls=0.0846, proj_loss=-0.381][2026-03-26 12:40:54] Step: 2366, Training Logs: loss_final: 0.549797, loss_mean: 0.857275, proj_loss: -0.390528, loss_mean_cls: 0.083049, grad_norm: 0.439282 +Steps: 0%| | 2367/1000000 [09:51<68:35:42, 4.04it/s, grad_norm=0.439, loss_final=0.55, loss_mean=0.857, loss_mean_cls=0.083, proj_loss=-0.391][2026-03-26 12:40:54] Step: 2367, Training Logs: loss_final: 0.552317, loss_mean: 0.850041, proj_loss: -0.381912, loss_mean_cls: 0.084188, grad_norm: 0.623601 +Steps: 0%| | 2368/1000000 [09:51<68:35:36, 4.04it/s, grad_norm=0.624, loss_final=0.552, loss_mean=0.85, loss_mean_cls=0.0842, proj_loss=-0.382][2026-03-26 12:40:55] Step: 2368, Training Logs: loss_final: 0.556066, loss_mean: 0.873616, proj_loss: -0.397903, loss_mean_cls: 0.080353, grad_norm: 0.457469 +Steps: 0%| | 2369/1000000 [09:51<68:39:29, 4.04it/s, grad_norm=0.457, loss_final=0.556, loss_mean=0.874, loss_mean_cls=0.0804, proj_loss=-0.398][2026-03-26 12:40:55] Step: 2369, Training Logs: loss_final: 0.547124, loss_mean: 0.846609, proj_loss: -0.382440, loss_mean_cls: 0.082955, grad_norm: 0.536281 +Steps: 0%| | 2370/1000000 [09:52<68:38:47, 4.04it/s, grad_norm=0.536, loss_final=0.547, loss_mean=0.847, loss_mean_cls=0.083, proj_loss=-0.382][2026-03-26 12:40:55] Step: 2370, Training Logs: loss_final: 0.540668, loss_mean: 0.845659, proj_loss: -0.386370, loss_mean_cls: 0.081380, grad_norm: 0.598361 +Steps: 0%| | 2371/1000000 [09:52<68:39:29, 4.04it/s, grad_norm=0.598, loss_final=0.541, loss_mean=0.846, loss_mean_cls=0.0814, proj_loss=-0.386][2026-03-26 12:40:55] Step: 2371, Training Logs: loss_final: 0.550030, loss_mean: 0.875272, proj_loss: -0.401811, loss_mean_cls: 0.076568, grad_norm: 0.674143 +Steps: 0%| | 2372/1000000 [09:52<68:37:07, 4.04it/s, grad_norm=0.674, loss_final=0.55, loss_mean=0.875, loss_mean_cls=0.0766, proj_loss=-0.402][2026-03-26 12:40:56] Step: 2372, Training Logs: loss_final: 0.534230, loss_mean: 0.846839, proj_loss: -0.393574, loss_mean_cls: 0.080965, grad_norm: 0.576352 +Steps: 0%| | 2373/1000000 [09:52<68:42:43, 4.03it/s, grad_norm=0.576, loss_final=0.534, loss_mean=0.847, loss_mean_cls=0.081, proj_loss=-0.394][2026-03-26 12:40:56] Step: 2373, Training Logs: loss_final: 0.545531, loss_mean: 0.853760, proj_loss: -0.391449, loss_mean_cls: 0.083220, grad_norm: 0.650477 +Steps: 0%| | 2374/1000000 [09:53<68:42:32, 4.03it/s, grad_norm=0.65, loss_final=0.546, loss_mean=0.854, loss_mean_cls=0.0832, proj_loss=-0.391][2026-03-26 12:40:56] Step: 2374, Training Logs: loss_final: 0.559050, loss_mean: 0.869066, proj_loss: -0.391500, loss_mean_cls: 0.081484, grad_norm: 0.565057 +Steps: 0%| | 2375/1000000 [09:53<68:39:44, 4.04it/s, grad_norm=0.565, loss_final=0.559, loss_mean=0.869, loss_mean_cls=0.0815, proj_loss=-0.391][2026-03-26 12:40:56] Step: 2375, Training Logs: loss_final: 0.532570, loss_mean: 0.842051, proj_loss: -0.390698, loss_mean_cls: 0.081218, grad_norm: 0.662434 +Steps: 0%| | 2376/1000000 [09:53<68:40:04, 4.04it/s, grad_norm=0.662, loss_final=0.533, loss_mean=0.842, loss_mean_cls=0.0812, proj_loss=-0.391][2026-03-26 12:40:57] Step: 2376, Training Logs: loss_final: 0.550880, loss_mean: 0.853404, proj_loss: -0.384918, loss_mean_cls: 0.082395, grad_norm: 0.771579 +Steps: 0%| | 2377/1000000 [09:53<68:37:02, 4.04it/s, grad_norm=0.772, loss_final=0.551, loss_mean=0.853, loss_mean_cls=0.0824, proj_loss=-0.385][2026-03-26 12:40:57] Step: 2377, Training Logs: loss_final: 0.546006, loss_mean: 0.856574, proj_loss: -0.391177, loss_mean_cls: 0.080610, grad_norm: 0.526486 +Steps: 0%| | 2378/1000000 [09:54<68:39:32, 4.04it/s, grad_norm=0.526, loss_final=0.546, loss_mean=0.857, loss_mean_cls=0.0806, proj_loss=-0.391][2026-03-26 12:40:57] Step: 2378, Training Logs: loss_final: 0.538391, loss_mean: 0.839766, proj_loss: -0.385193, loss_mean_cls: 0.083819, grad_norm: 0.683381 +Steps: 0%| | 2379/1000000 [09:54<68:38:37, 4.04it/s, grad_norm=0.683, loss_final=0.538, loss_mean=0.84, loss_mean_cls=0.0838, proj_loss=-0.385][2026-03-26 12:40:57] Step: 2379, Training Logs: loss_final: 0.530847, loss_mean: 0.836344, proj_loss: -0.387740, loss_mean_cls: 0.082243, grad_norm: 0.585175 +Steps: 0%| | 2380/1000000 [09:54<68:38:22, 4.04it/s, grad_norm=0.585, loss_final=0.531, loss_mean=0.836, loss_mean_cls=0.0822, proj_loss=-0.388][2026-03-26 12:40:58] Step: 2380, Training Logs: loss_final: 0.533898, loss_mean: 0.834088, proj_loss: -0.384894, loss_mean_cls: 0.084705, grad_norm: 0.566666 +Steps: 0%| | 2381/1000000 [09:54<68:41:02, 4.03it/s, grad_norm=0.567, loss_final=0.534, loss_mean=0.834, loss_mean_cls=0.0847, proj_loss=-0.385][2026-03-26 12:40:58] Step: 2381, Training Logs: loss_final: 0.544286, loss_mean: 0.842742, proj_loss: -0.383693, loss_mean_cls: 0.085236, grad_norm: 0.428047 +Steps: 0%| | 2382/1000000 [09:55<68:41:42, 4.03it/s, grad_norm=0.428, loss_final=0.544, loss_mean=0.843, loss_mean_cls=0.0852, proj_loss=-0.384][2026-03-26 12:40:58] Step: 2382, Training Logs: loss_final: 0.540905, loss_mean: 0.836239, proj_loss: -0.380064, loss_mean_cls: 0.084730, grad_norm: 0.651135 +Steps: 0%| | 2383/1000000 [09:55<68:41:13, 4.03it/s, grad_norm=0.651, loss_final=0.541, loss_mean=0.836, loss_mean_cls=0.0847, proj_loss=-0.38][2026-03-26 12:40:58] Step: 2383, Training Logs: loss_final: 0.551859, loss_mean: 0.854345, proj_loss: -0.385136, loss_mean_cls: 0.082650, grad_norm: 0.618376 +Steps: 0%| | 2384/1000000 [09:55<68:40:12, 4.04it/s, grad_norm=0.618, loss_final=0.552, loss_mean=0.854, loss_mean_cls=0.0827, proj_loss=-0.385][2026-03-26 12:40:59] Step: 2384, Training Logs: loss_final: 0.535731, loss_mean: 0.852923, proj_loss: -0.396682, loss_mean_cls: 0.079490, grad_norm: 0.305354 +Steps: 0%| | 2385/1000000 [09:55<68:38:48, 4.04it/s, grad_norm=0.305, loss_final=0.536, loss_mean=0.853, loss_mean_cls=0.0795, proj_loss=-0.397][2026-03-26 12:40:59] Step: 2385, Training Logs: loss_final: 0.542254, loss_mean: 0.841820, proj_loss: -0.383811, loss_mean_cls: 0.084245, grad_norm: 0.490346 +Steps: 0%| | 2386/1000000 [09:56<68:40:29, 4.04it/s, grad_norm=0.49, loss_final=0.542, loss_mean=0.842, loss_mean_cls=0.0842, proj_loss=-0.384][2026-03-26 12:40:59] Step: 2386, Training Logs: loss_final: 0.529937, loss_mean: 0.831054, proj_loss: -0.384111, loss_mean_cls: 0.082994, grad_norm: 0.385024 +Steps: 0%| | 2387/1000000 [09:56<68:38:29, 4.04it/s, grad_norm=0.385, loss_final=0.53, loss_mean=0.831, loss_mean_cls=0.083, proj_loss=-0.384][2026-03-26 12:40:59] Step: 2387, Training Logs: loss_final: 0.526979, loss_mean: 0.842414, proj_loss: -0.396377, loss_mean_cls: 0.080942, grad_norm: 0.339990 +Steps: 0%| | 2388/1000000 [09:56<68:41:48, 4.03it/s, grad_norm=0.34, loss_final=0.527, loss_mean=0.842, loss_mean_cls=0.0809, proj_loss=-0.396][2026-03-26 12:41:00] Step: 2388, Training Logs: loss_final: 0.540889, loss_mean: 0.843701, proj_loss: -0.386386, loss_mean_cls: 0.083574, grad_norm: 0.563672 +Steps: 0%| | 2389/1000000 [09:56<68:37:47, 4.04it/s, grad_norm=0.564, loss_final=0.541, loss_mean=0.844, loss_mean_cls=0.0836, proj_loss=-0.386][2026-03-26 12:41:00] Step: 2389, Training Logs: loss_final: 0.522237, loss_mean: 0.843316, proj_loss: -0.400302, loss_mean_cls: 0.079223, grad_norm: 0.342046 +Steps: 0%| | 2390/1000000 [09:57<68:40:47, 4.03it/s, grad_norm=0.342, loss_final=0.522, loss_mean=0.843, loss_mean_cls=0.0792, proj_loss=-0.4][2026-03-26 12:41:00] Step: 2390, Training Logs: loss_final: 0.523551, loss_mean: 0.822946, proj_loss: -0.384953, loss_mean_cls: 0.085557, grad_norm: 0.496053 +Steps: 0%| | 2391/1000000 [09:57<68:39:14, 4.04it/s, grad_norm=0.496, loss_final=0.524, loss_mean=0.823, loss_mean_cls=0.0856, proj_loss=-0.385][2026-03-26 12:41:00] Step: 2391, Training Logs: loss_final: 0.544458, loss_mean: 0.864118, proj_loss: -0.398246, loss_mean_cls: 0.078586, grad_norm: 0.386572 +Steps: 0%| | 2392/1000000 [09:57<68:40:16, 4.04it/s, grad_norm=0.387, loss_final=0.544, loss_mean=0.864, loss_mean_cls=0.0786, proj_loss=-0.398][2026-03-26 12:41:01] Step: 2392, Training Logs: loss_final: 0.540987, loss_mean: 0.849273, proj_loss: -0.390066, loss_mean_cls: 0.081781, grad_norm: 0.644277 +Steps: 0%| | 2393/1000000 [09:57<68:37:08, 4.04it/s, grad_norm=0.644, loss_final=0.541, loss_mean=0.849, loss_mean_cls=0.0818, proj_loss=-0.39][2026-03-26 12:41:01] Step: 2393, Training Logs: loss_final: 0.535344, loss_mean: 0.845236, proj_loss: -0.391388, loss_mean_cls: 0.081497, grad_norm: 0.499594 +Steps: 0%| | 2394/1000000 [09:58<68:37:31, 4.04it/s, grad_norm=0.5, loss_final=0.535, loss_mean=0.845, loss_mean_cls=0.0815, proj_loss=-0.391][2026-03-26 12:41:01] Step: 2394, Training Logs: loss_final: 0.542763, loss_mean: 0.862352, proj_loss: -0.398795, loss_mean_cls: 0.079206, grad_norm: 0.513977 +Steps: 0%| | 2395/1000000 [09:58<68:37:17, 4.04it/s, grad_norm=0.514, loss_final=0.543, loss_mean=0.862, loss_mean_cls=0.0792, proj_loss=-0.399][2026-03-26 12:41:01] Step: 2395, Training Logs: loss_final: 0.534385, loss_mean: 0.850003, proj_loss: -0.395981, loss_mean_cls: 0.080362, grad_norm: 0.420517 +Steps: 0%| | 2396/1000000 [09:58<68:40:02, 4.04it/s, grad_norm=0.421, loss_final=0.534, loss_mean=0.85, loss_mean_cls=0.0804, proj_loss=-0.396][2026-03-26 12:41:02] Step: 2396, Training Logs: loss_final: 0.539947, loss_mean: 0.846055, proj_loss: -0.387706, loss_mean_cls: 0.081599, grad_norm: 0.428847 +Steps: 0%| | 2397/1000000 [09:58<68:37:50, 4.04it/s, grad_norm=0.429, loss_final=0.54, loss_mean=0.846, loss_mean_cls=0.0816, proj_loss=-0.388][2026-03-26 12:41:02] Step: 2397, Training Logs: loss_final: 0.539015, loss_mean: 0.857528, proj_loss: -0.400016, loss_mean_cls: 0.081502, grad_norm: 0.345908 +Steps: 0%| | 2398/1000000 [09:59<68:43:44, 4.03it/s, grad_norm=0.346, loss_final=0.539, loss_mean=0.858, loss_mean_cls=0.0815, proj_loss=-0.4][2026-03-26 12:41:02] Step: 2398, Training Logs: loss_final: 0.526940, loss_mean: 0.839859, proj_loss: -0.394255, loss_mean_cls: 0.081335, grad_norm: 0.416840 +Steps: 0%| | 2399/1000000 [09:59<68:38:46, 4.04it/s, grad_norm=0.417, loss_final=0.527, loss_mean=0.84, loss_mean_cls=0.0813, proj_loss=-0.394][2026-03-26 12:41:02] Step: 2399, Training Logs: loss_final: 0.546285, loss_mean: 0.858962, proj_loss: -0.393198, loss_mean_cls: 0.080522, grad_norm: 0.370813 +Steps: 0%| | 2400/1000000 [09:59<68:40:42, 4.03it/s, grad_norm=0.371, loss_final=0.546, loss_mean=0.859, loss_mean_cls=0.0805, proj_loss=-0.393][2026-03-26 12:41:03] Step: 2400, Training Logs: loss_final: 0.545453, loss_mean: 0.852777, proj_loss: -0.391038, loss_mean_cls: 0.083715, grad_norm: 0.457551 +Steps: 0%| | 2401/1000000 [09:59<68:41:02, 4.03it/s, grad_norm=0.458, loss_final=0.545, loss_mean=0.853, loss_mean_cls=0.0837, proj_loss=-0.391][2026-03-26 12:41:03] Step: 2401, Training Logs: loss_final: 0.547782, loss_mean: 0.851498, proj_loss: -0.386953, loss_mean_cls: 0.083237, grad_norm: 0.376150 +Steps: 0%| | 2402/1000000 [10:00<68:39:30, 4.04it/s, grad_norm=0.376, loss_final=0.548, loss_mean=0.851, loss_mean_cls=0.0832, proj_loss=-0.387][2026-03-26 12:41:03] Step: 2402, Training Logs: loss_final: 0.532207, loss_mean: 0.839939, proj_loss: -0.390140, loss_mean_cls: 0.082408, grad_norm: 0.423609 +Steps: 0%| | 2403/1000000 [10:00<68:38:05, 4.04it/s, grad_norm=0.424, loss_final=0.532, loss_mean=0.84, loss_mean_cls=0.0824, proj_loss=-0.39][2026-03-26 12:41:03] Step: 2403, Training Logs: loss_final: 0.546309, loss_mean: 0.847648, proj_loss: -0.384284, loss_mean_cls: 0.082945, grad_norm: 0.361442 +Steps: 0%| | 2404/1000000 [10:00<68:38:27, 4.04it/s, grad_norm=0.361, loss_final=0.546, loss_mean=0.848, loss_mean_cls=0.0829, proj_loss=-0.384][2026-03-26 12:41:04] Step: 2404, Training Logs: loss_final: 0.542241, loss_mean: 0.843387, proj_loss: -0.384533, loss_mean_cls: 0.083387, grad_norm: 0.418502 +Steps: 0%| | 2405/1000000 [10:00<68:41:27, 4.03it/s, grad_norm=0.419, loss_final=0.542, loss_mean=0.843, loss_mean_cls=0.0834, proj_loss=-0.385][2026-03-26 12:41:04] Step: 2405, Training Logs: loss_final: 0.546635, loss_mean: 0.844624, proj_loss: -0.381850, loss_mean_cls: 0.083860, grad_norm: 0.532235 +Steps: 0%| | 2406/1000000 [10:01<68:39:13, 4.04it/s, grad_norm=0.532, loss_final=0.547, loss_mean=0.845, loss_mean_cls=0.0839, proj_loss=-0.382][2026-03-26 12:41:04] Step: 2406, Training Logs: loss_final: 0.535624, loss_mean: 0.847074, proj_loss: -0.392680, loss_mean_cls: 0.081230, grad_norm: 0.446092 +Steps: 0%| | 2407/1000000 [10:01<68:42:42, 4.03it/s, grad_norm=0.446, loss_final=0.536, loss_mean=0.847, loss_mean_cls=0.0812, proj_loss=-0.393][2026-03-26 12:41:04] Step: 2407, Training Logs: loss_final: 0.545328, loss_mean: 0.841670, proj_loss: -0.381060, loss_mean_cls: 0.084718, grad_norm: 0.472360 +Steps: 0%| | 2408/1000000 [10:01<68:39:44, 4.04it/s, grad_norm=0.472, loss_final=0.545, loss_mean=0.842, loss_mean_cls=0.0847, proj_loss=-0.381][2026-03-26 12:41:04] Step: 2408, Training Logs: loss_final: 0.540402, loss_mean: 0.853946, proj_loss: -0.392789, loss_mean_cls: 0.079245, grad_norm: 0.624887 +Steps: 0%| | 2409/1000000 [10:01<68:37:56, 4.04it/s, grad_norm=0.625, loss_final=0.54, loss_mean=0.854, loss_mean_cls=0.0792, proj_loss=-0.393][2026-03-26 12:41:05] Step: 2409, Training Logs: loss_final: 0.545396, loss_mean: 0.862800, proj_loss: -0.397528, loss_mean_cls: 0.080124, grad_norm: 0.481568 +Steps: 0%| | 2410/1000000 [10:02<68:38:29, 4.04it/s, grad_norm=0.482, loss_final=0.545, loss_mean=0.863, loss_mean_cls=0.0801, proj_loss=-0.398][2026-03-26 12:41:05] Step: 2410, Training Logs: loss_final: 0.531003, loss_mean: 0.848341, proj_loss: -0.397705, loss_mean_cls: 0.080367, grad_norm: 0.469463 +Steps: 0%| | 2411/1000000 [10:02<68:34:16, 4.04it/s, grad_norm=0.469, loss_final=0.531, loss_mean=0.848, loss_mean_cls=0.0804, proj_loss=-0.398][2026-03-26 12:41:05] Step: 2411, Training Logs: loss_final: 0.555797, loss_mean: 0.873146, proj_loss: -0.396380, loss_mean_cls: 0.079031, grad_norm: 0.552224 +Steps: 0%| | 2412/1000000 [10:02<68:34:13, 4.04it/s, grad_norm=0.552, loss_final=0.556, loss_mean=0.873, loss_mean_cls=0.079, proj_loss=-0.396][2026-03-26 12:41:05] Step: 2412, Training Logs: loss_final: 0.539170, loss_mean: 0.848347, proj_loss: -0.391357, loss_mean_cls: 0.082180, grad_norm: 0.476445 +Steps: 0%| | 2413/1000000 [10:02<68:36:04, 4.04it/s, grad_norm=0.476, loss_final=0.539, loss_mean=0.848, loss_mean_cls=0.0822, proj_loss=-0.391][2026-03-26 12:41:06] Step: 2413, Training Logs: loss_final: 0.527400, loss_mean: 0.821100, proj_loss: -0.380335, loss_mean_cls: 0.086635, grad_norm: 0.447648 +Steps: 0%| | 2414/1000000 [10:03<68:34:34, 4.04it/s, grad_norm=0.448, loss_final=0.527, loss_mean=0.821, loss_mean_cls=0.0866, proj_loss=-0.38][2026-03-26 12:41:06] Step: 2414, Training Logs: loss_final: 0.528731, loss_mean: 0.825356, proj_loss: -0.383413, loss_mean_cls: 0.086788, grad_norm: 0.547448 +Steps: 0%| | 2415/1000000 [10:03<68:33:38, 4.04it/s, grad_norm=0.547, loss_final=0.529, loss_mean=0.825, loss_mean_cls=0.0868, proj_loss=-0.383][2026-03-26 12:41:06] Step: 2415, Training Logs: loss_final: 0.542337, loss_mean: 0.844866, proj_loss: -0.385897, loss_mean_cls: 0.083368, grad_norm: 0.425545 +Steps: 0%| | 2416/1000000 [10:03<68:32:32, 4.04it/s, grad_norm=0.426, loss_final=0.542, loss_mean=0.845, loss_mean_cls=0.0834, proj_loss=-0.386][2026-03-26 12:41:06] Step: 2416, Training Logs: loss_final: 0.542689, loss_mean: 0.843057, proj_loss: -0.384811, loss_mean_cls: 0.084443, grad_norm: 0.544295 +Steps: 0%| | 2417/1000000 [10:03<68:34:34, 4.04it/s, grad_norm=0.544, loss_final=0.543, loss_mean=0.843, loss_mean_cls=0.0844, proj_loss=-0.385][2026-03-26 12:41:07] Step: 2417, Training Logs: loss_final: 0.546755, loss_mean: 0.850825, proj_loss: -0.386770, loss_mean_cls: 0.082700, grad_norm: 0.343982 +Steps: 0%| | 2418/1000000 [10:04<69:12:24, 4.00it/s, grad_norm=0.344, loss_final=0.547, loss_mean=0.851, loss_mean_cls=0.0827, proj_loss=-0.387][2026-03-26 12:41:07] Step: 2418, Training Logs: loss_final: 0.535308, loss_mean: 0.846641, proj_loss: -0.394240, loss_mean_cls: 0.082908, grad_norm: 0.643553 +Steps: 0%| | 2419/1000000 [10:04<70:36:42, 3.92it/s, grad_norm=0.644, loss_final=0.535, loss_mean=0.847, loss_mean_cls=0.0829, proj_loss=-0.394][2026-03-26 12:41:07] Step: 2419, Training Logs: loss_final: 0.531039, loss_mean: 0.857045, proj_loss: -0.403749, loss_mean_cls: 0.077742, grad_norm: 0.769306 +Steps: 0%| | 2420/1000000 [10:04<70:01:25, 3.96it/s, grad_norm=0.769, loss_final=0.531, loss_mean=0.857, loss_mean_cls=0.0777, proj_loss=-0.404][2026-03-26 12:41:07] Step: 2420, Training Logs: loss_final: 0.543493, loss_mean: 0.844615, proj_loss: -0.385122, loss_mean_cls: 0.084000, grad_norm: 0.713129 +Steps: 0%| | 2421/1000000 [10:04<69:36:16, 3.98it/s, grad_norm=0.713, loss_final=0.543, loss_mean=0.845, loss_mean_cls=0.084, proj_loss=-0.385][2026-03-26 12:41:08] Step: 2421, Training Logs: loss_final: 0.543880, loss_mean: 0.839573, proj_loss: -0.379329, loss_mean_cls: 0.083635, grad_norm: 0.505805 +Steps: 0%| | 2422/1000000 [10:05<69:15:49, 4.00it/s, grad_norm=0.506, loss_final=0.544, loss_mean=0.84, loss_mean_cls=0.0836, proj_loss=-0.379][2026-03-26 12:41:08] Step: 2422, Training Logs: loss_final: 0.540435, loss_mean: 0.857896, proj_loss: -0.398172, loss_mean_cls: 0.080711, grad_norm: 0.580430 +Steps: 0%| | 2423/1000000 [10:05<69:04:59, 4.01it/s, grad_norm=0.58, loss_final=0.54, loss_mean=0.858, loss_mean_cls=0.0807, proj_loss=-0.398][2026-03-26 12:41:08] Step: 2423, Training Logs: loss_final: 0.531475, loss_mean: 0.832444, proj_loss: -0.384578, loss_mean_cls: 0.083609, grad_norm: 0.657810 +Steps: 0%| | 2424/1000000 [10:05<68:54:18, 4.02it/s, grad_norm=0.658, loss_final=0.531, loss_mean=0.832, loss_mean_cls=0.0836, proj_loss=-0.385][2026-03-26 12:41:08] Step: 2424, Training Logs: loss_final: 0.545300, loss_mean: 0.858736, proj_loss: -0.393782, loss_mean_cls: 0.080346, grad_norm: 0.425652 +Steps: 0%| | 2425/1000000 [10:05<68:50:10, 4.03it/s, grad_norm=0.426, loss_final=0.545, loss_mean=0.859, loss_mean_cls=0.0803, proj_loss=-0.394][2026-03-26 12:41:09] Step: 2425, Training Logs: loss_final: 0.533891, loss_mean: 0.846977, proj_loss: -0.395277, loss_mean_cls: 0.082191, grad_norm: 0.690569 +Steps: 0%| | 2426/1000000 [10:06<68:47:13, 4.03it/s, grad_norm=0.691, loss_final=0.534, loss_mean=0.847, loss_mean_cls=0.0822, proj_loss=-0.395][2026-03-26 12:41:09] Step: 2426, Training Logs: loss_final: 0.551648, loss_mean: 0.855721, proj_loss: -0.385323, loss_mean_cls: 0.081250, grad_norm: 0.810485 +Steps: 0%| | 2427/1000000 [10:06<68:43:40, 4.03it/s, grad_norm=0.81, loss_final=0.552, loss_mean=0.856, loss_mean_cls=0.0813, proj_loss=-0.385][2026-03-26 12:41:09] Step: 2427, Training Logs: loss_final: 0.552392, loss_mean: 0.870911, proj_loss: -0.398393, loss_mean_cls: 0.079874, grad_norm: 0.406350 +Steps: 0%| | 2428/1000000 [10:06<68:39:59, 4.04it/s, grad_norm=0.406, loss_final=0.552, loss_mean=0.871, loss_mean_cls=0.0799, proj_loss=-0.398][2026-03-26 12:41:09] Step: 2428, Training Logs: loss_final: 0.514794, loss_mean: 0.817225, proj_loss: -0.387649, loss_mean_cls: 0.085218, grad_norm: 0.532264 +Steps: 0%| | 2429/1000000 [10:06<68:38:54, 4.04it/s, grad_norm=0.532, loss_final=0.515, loss_mean=0.817, loss_mean_cls=0.0852, proj_loss=-0.388][2026-03-26 12:41:10] Step: 2429, Training Logs: loss_final: 0.545504, loss_mean: 0.853895, proj_loss: -0.389964, loss_mean_cls: 0.081573, grad_norm: 0.737514 +Steps: 0%| | 2430/1000000 [10:07<68:38:05, 4.04it/s, grad_norm=0.738, loss_final=0.546, loss_mean=0.854, loss_mean_cls=0.0816, proj_loss=-0.39][2026-03-26 12:41:10] Step: 2430, Training Logs: loss_final: 0.536006, loss_mean: 0.842535, proj_loss: -0.387151, loss_mean_cls: 0.080623, grad_norm: 0.406726 +Steps: 0%| | 2431/1000000 [10:07<68:38:20, 4.04it/s, grad_norm=0.407, loss_final=0.536, loss_mean=0.843, loss_mean_cls=0.0806, proj_loss=-0.387][2026-03-26 12:41:10] Step: 2431, Training Logs: loss_final: 0.533237, loss_mean: 0.852084, proj_loss: -0.398584, loss_mean_cls: 0.079738, grad_norm: 0.706594 +Steps: 0%| | 2432/1000000 [10:07<68:39:19, 4.04it/s, grad_norm=0.707, loss_final=0.533, loss_mean=0.852, loss_mean_cls=0.0797, proj_loss=-0.399][2026-03-26 12:41:10] Step: 2432, Training Logs: loss_final: 0.523502, loss_mean: 0.829934, proj_loss: -0.390184, loss_mean_cls: 0.083752, grad_norm: 0.803372 +Steps: 0%| | 2433/1000000 [10:07<68:41:50, 4.03it/s, grad_norm=0.803, loss_final=0.524, loss_mean=0.83, loss_mean_cls=0.0838, proj_loss=-0.39][2026-03-26 12:41:11] Step: 2433, Training Logs: loss_final: 0.539187, loss_mean: 0.841401, proj_loss: -0.386359, loss_mean_cls: 0.084145, grad_norm: 0.362228 +Steps: 0%| | 2434/1000000 [10:08<68:39:08, 4.04it/s, grad_norm=0.362, loss_final=0.539, loss_mean=0.841, loss_mean_cls=0.0841, proj_loss=-0.386][2026-03-26 12:41:11] Step: 2434, Training Logs: loss_final: 0.559961, loss_mean: 0.882152, proj_loss: -0.400579, loss_mean_cls: 0.078387, grad_norm: 0.917272 +Steps: 0%| | 2435/1000000 [10:08<68:38:12, 4.04it/s, grad_norm=0.917, loss_final=0.56, loss_mean=0.882, loss_mean_cls=0.0784, proj_loss=-0.401][2026-03-26 12:41:11] Step: 2435, Training Logs: loss_final: 0.551593, loss_mean: 0.862179, proj_loss: -0.391642, loss_mean_cls: 0.081056, grad_norm: 0.950975 +Steps: 0%| | 2436/1000000 [10:08<68:40:34, 4.03it/s, grad_norm=0.951, loss_final=0.552, loss_mean=0.862, loss_mean_cls=0.0811, proj_loss=-0.392][2026-03-26 12:41:11] Step: 2436, Training Logs: loss_final: 0.532890, loss_mean: 0.848273, proj_loss: -0.395656, loss_mean_cls: 0.080273, grad_norm: 0.431115 +Steps: 0%| | 2437/1000000 [10:08<68:42:43, 4.03it/s, grad_norm=0.431, loss_final=0.533, loss_mean=0.848, loss_mean_cls=0.0803, proj_loss=-0.396][2026-03-26 12:41:12] Step: 2437, Training Logs: loss_final: 0.541059, loss_mean: 0.842585, proj_loss: -0.385669, loss_mean_cls: 0.084142, grad_norm: 0.708289 +Steps: 0%| | 2438/1000000 [10:09<68:43:24, 4.03it/s, grad_norm=0.708, loss_final=0.541, loss_mean=0.843, loss_mean_cls=0.0841, proj_loss=-0.386][2026-03-26 12:41:12] Step: 2438, Training Logs: loss_final: 0.524255, loss_mean: 0.827869, proj_loss: -0.386737, loss_mean_cls: 0.083122, grad_norm: 0.831789 +Steps: 0%| | 2439/1000000 [10:09<68:42:49, 4.03it/s, grad_norm=0.832, loss_final=0.524, loss_mean=0.828, loss_mean_cls=0.0831, proj_loss=-0.387][2026-03-26 12:41:12] Step: 2439, Training Logs: loss_final: 0.529691, loss_mean: 0.843470, proj_loss: -0.395626, loss_mean_cls: 0.081847, grad_norm: 0.496608 +Steps: 0%| | 2440/1000000 [10:09<68:41:28, 4.03it/s, grad_norm=0.497, loss_final=0.53, loss_mean=0.843, loss_mean_cls=0.0818, proj_loss=-0.396][2026-03-26 12:41:12] Step: 2440, Training Logs: loss_final: 0.549614, loss_mean: 0.863599, proj_loss: -0.393573, loss_mean_cls: 0.079587, grad_norm: 0.557821 +Steps: 0%| | 2441/1000000 [10:09<68:37:57, 4.04it/s, grad_norm=0.558, loss_final=0.55, loss_mean=0.864, loss_mean_cls=0.0796, proj_loss=-0.394][2026-03-26 12:41:13] Step: 2441, Training Logs: loss_final: 0.543286, loss_mean: 0.853529, proj_loss: -0.390539, loss_mean_cls: 0.080297, grad_norm: 0.563494 +Steps: 0%| | 2442/1000000 [10:10<68:37:55, 4.04it/s, grad_norm=0.563, loss_final=0.543, loss_mean=0.854, loss_mean_cls=0.0803, proj_loss=-0.391][2026-03-26 12:41:13] Step: 2442, Training Logs: loss_final: 0.530274, loss_mean: 0.832667, proj_loss: -0.387454, loss_mean_cls: 0.085062, grad_norm: 0.504427 +Steps: 0%| | 2443/1000000 [10:10<68:41:24, 4.03it/s, grad_norm=0.504, loss_final=0.53, loss_mean=0.833, loss_mean_cls=0.0851, proj_loss=-0.387][2026-03-26 12:41:13] Step: 2443, Training Logs: loss_final: 0.541258, loss_mean: 0.848697, proj_loss: -0.389626, loss_mean_cls: 0.082188, grad_norm: 0.602185 +Steps: 0%| | 2444/1000000 [10:10<68:40:45, 4.03it/s, grad_norm=0.602, loss_final=0.541, loss_mean=0.849, loss_mean_cls=0.0822, proj_loss=-0.39][2026-03-26 12:41:13] Step: 2444, Training Logs: loss_final: 0.548126, loss_mean: 0.851160, proj_loss: -0.386683, loss_mean_cls: 0.083649, grad_norm: 0.688138 +Steps: 0%| | 2445/1000000 [10:10<68:38:02, 4.04it/s, grad_norm=0.688, loss_final=0.548, loss_mean=0.851, loss_mean_cls=0.0836, proj_loss=-0.387][2026-03-26 12:41:14] Step: 2445, Training Logs: loss_final: 0.522063, loss_mean: 0.824721, proj_loss: -0.388340, loss_mean_cls: 0.085682, grad_norm: 0.409297 +Steps: 0%| | 2446/1000000 [10:11<68:40:57, 4.03it/s, grad_norm=0.409, loss_final=0.522, loss_mean=0.825, loss_mean_cls=0.0857, proj_loss=-0.388][2026-03-26 12:41:14] Step: 2446, Training Logs: loss_final: 0.518731, loss_mean: 0.827791, proj_loss: -0.392556, loss_mean_cls: 0.083497, grad_norm: 0.492711 +Steps: 0%| | 2447/1000000 [10:11<68:39:31, 4.04it/s, grad_norm=0.493, loss_final=0.519, loss_mean=0.828, loss_mean_cls=0.0835, proj_loss=-0.393][2026-03-26 12:41:14] Step: 2447, Training Logs: loss_final: 0.513983, loss_mean: 0.819054, proj_loss: -0.388924, loss_mean_cls: 0.083852, grad_norm: 0.513393 +Steps: 0%| | 2448/1000000 [10:11<68:38:02, 4.04it/s, grad_norm=0.513, loss_final=0.514, loss_mean=0.819, loss_mean_cls=0.0839, proj_loss=-0.389][2026-03-26 12:41:14] Step: 2448, Training Logs: loss_final: 0.541897, loss_mean: 0.864276, proj_loss: -0.402572, loss_mean_cls: 0.080194, grad_norm: 0.392599 +Steps: 0%| | 2449/1000000 [10:11<68:33:29, 4.04it/s, grad_norm=0.393, loss_final=0.542, loss_mean=0.864, loss_mean_cls=0.0802, proj_loss=-0.403][2026-03-26 12:41:15] Step: 2449, Training Logs: loss_final: 0.549205, loss_mean: 0.857959, proj_loss: -0.391128, loss_mean_cls: 0.082374, grad_norm: 0.383270 +Steps: 0%| | 2450/1000000 [10:12<68:35:10, 4.04it/s, grad_norm=0.383, loss_final=0.549, loss_mean=0.858, loss_mean_cls=0.0824, proj_loss=-0.391][2026-03-26 12:41:15] Step: 2450, Training Logs: loss_final: 0.539366, loss_mean: 0.833955, proj_loss: -0.380146, loss_mean_cls: 0.085556, grad_norm: 0.490695 +Steps: 0%| | 2451/1000000 [10:12<68:34:31, 4.04it/s, grad_norm=0.491, loss_final=0.539, loss_mean=0.834, loss_mean_cls=0.0856, proj_loss=-0.38][2026-03-26 12:41:15] Step: 2451, Training Logs: loss_final: 0.546893, loss_mean: 0.862138, proj_loss: -0.394853, loss_mean_cls: 0.079607, grad_norm: 0.370415 +Steps: 0%| | 2452/1000000 [10:12<68:35:11, 4.04it/s, grad_norm=0.37, loss_final=0.547, loss_mean=0.862, loss_mean_cls=0.0796, proj_loss=-0.395][2026-03-26 12:41:15] Step: 2452, Training Logs: loss_final: 0.552938, loss_mean: 0.866602, proj_loss: -0.392757, loss_mean_cls: 0.079092, grad_norm: 0.403667 +Steps: 0%| | 2453/1000000 [10:12<68:34:38, 4.04it/s, grad_norm=0.404, loss_final=0.553, loss_mean=0.867, loss_mean_cls=0.0791, proj_loss=-0.393][2026-03-26 12:41:16] Step: 2453, Training Logs: loss_final: 0.533298, loss_mean: 0.831962, proj_loss: -0.383274, loss_mean_cls: 0.084609, grad_norm: 0.452453 +Steps: 0%| | 2454/1000000 [10:13<68:38:36, 4.04it/s, grad_norm=0.452, loss_final=0.533, loss_mean=0.832, loss_mean_cls=0.0846, proj_loss=-0.383][2026-03-26 12:41:16] Step: 2454, Training Logs: loss_final: 0.540801, loss_mean: 0.850266, proj_loss: -0.392136, loss_mean_cls: 0.082671, grad_norm: 0.437592 +Steps: 0%| | 2455/1000000 [10:13<68:38:59, 4.04it/s, grad_norm=0.438, loss_final=0.541, loss_mean=0.85, loss_mean_cls=0.0827, proj_loss=-0.392][2026-03-26 12:41:16] Step: 2455, Training Logs: loss_final: 0.522665, loss_mean: 0.832326, proj_loss: -0.393295, loss_mean_cls: 0.083634, grad_norm: 0.496896 +Steps: 0%| | 2456/1000000 [10:13<68:38:35, 4.04it/s, grad_norm=0.497, loss_final=0.523, loss_mean=0.832, loss_mean_cls=0.0836, proj_loss=-0.393][2026-03-26 12:41:16] Step: 2456, Training Logs: loss_final: 0.534052, loss_mean: 0.843051, proj_loss: -0.391453, loss_mean_cls: 0.082454, grad_norm: 0.456411 +Steps: 0%| | 2457/1000000 [10:13<68:34:36, 4.04it/s, grad_norm=0.456, loss_final=0.534, loss_mean=0.843, loss_mean_cls=0.0825, proj_loss=-0.391][2026-03-26 12:41:17] Step: 2457, Training Logs: loss_final: 0.547048, loss_mean: 0.865787, proj_loss: -0.399380, loss_mean_cls: 0.080641, grad_norm: 0.460056 +Steps: 0%| | 2458/1000000 [10:14<68:33:41, 4.04it/s, grad_norm=0.46, loss_final=0.547, loss_mean=0.866, loss_mean_cls=0.0806, proj_loss=-0.399][2026-03-26 12:41:17] Step: 2458, Training Logs: loss_final: 0.550245, loss_mean: 0.861720, proj_loss: -0.392783, loss_mean_cls: 0.081308, grad_norm: 0.573273 +Steps: 0%| | 2459/1000000 [10:14<68:36:32, 4.04it/s, grad_norm=0.573, loss_final=0.55, loss_mean=0.862, loss_mean_cls=0.0813, proj_loss=-0.393][2026-03-26 12:41:17] Step: 2459, Training Logs: loss_final: 0.556517, loss_mean: 0.874695, proj_loss: -0.398108, loss_mean_cls: 0.079930, grad_norm: 0.446419 +Steps: 0%| | 2460/1000000 [10:14<68:36:49, 4.04it/s, grad_norm=0.446, loss_final=0.557, loss_mean=0.875, loss_mean_cls=0.0799, proj_loss=-0.398][2026-03-26 12:41:17] Step: 2460, Training Logs: loss_final: 0.548115, loss_mean: 0.852184, proj_loss: -0.386614, loss_mean_cls: 0.082545, grad_norm: 0.597321 +Steps: 0%| | 2461/1000000 [10:14<68:34:23, 4.04it/s, grad_norm=0.597, loss_final=0.548, loss_mean=0.852, loss_mean_cls=0.0825, proj_loss=-0.387][2026-03-26 12:41:18] Step: 2461, Training Logs: loss_final: 0.543965, loss_mean: 0.872073, proj_loss: -0.404861, loss_mean_cls: 0.076754, grad_norm: 0.571522 +Steps: 0%| | 2462/1000000 [10:15<68:43:10, 4.03it/s, grad_norm=0.572, loss_final=0.544, loss_mean=0.872, loss_mean_cls=0.0768, proj_loss=-0.405][2026-03-26 12:41:18] Step: 2462, Training Logs: loss_final: 0.536902, loss_mean: 0.846521, proj_loss: -0.390504, loss_mean_cls: 0.080884, grad_norm: 0.553902 +Steps: 0%| | 2463/1000000 [10:15<68:42:19, 4.03it/s, grad_norm=0.554, loss_final=0.537, loss_mean=0.847, loss_mean_cls=0.0809, proj_loss=-0.391][2026-03-26 12:41:18] Step: 2463, Training Logs: loss_final: 0.528799, loss_mean: 0.829631, proj_loss: -0.383820, loss_mean_cls: 0.082988, grad_norm: 0.378091 +Steps: 0%| | 2464/1000000 [10:15<68:42:04, 4.03it/s, grad_norm=0.378, loss_final=0.529, loss_mean=0.83, loss_mean_cls=0.083, proj_loss=-0.384][2026-03-26 12:41:18] Step: 2464, Training Logs: loss_final: 0.539985, loss_mean: 0.850076, proj_loss: -0.391925, loss_mean_cls: 0.081834, grad_norm: 0.445344 +Steps: 0%| | 2465/1000000 [10:15<68:41:23, 4.03it/s, grad_norm=0.445, loss_final=0.54, loss_mean=0.85, loss_mean_cls=0.0818, proj_loss=-0.392][2026-03-26 12:41:19] Step: 2465, Training Logs: loss_final: 0.526918, loss_mean: 0.835262, proj_loss: -0.391403, loss_mean_cls: 0.083058, grad_norm: 0.454317 +Steps: 0%| | 2466/1000000 [10:15<68:39:03, 4.04it/s, grad_norm=0.454, loss_final=0.527, loss_mean=0.835, loss_mean_cls=0.0831, proj_loss=-0.391][2026-03-26 12:41:19] Step: 2466, Training Logs: loss_final: 0.553765, loss_mean: 0.857691, proj_loss: -0.385822, loss_mean_cls: 0.081896, grad_norm: 0.326018 +Steps: 0%| | 2467/1000000 [10:16<68:37:41, 4.04it/s, grad_norm=0.326, loss_final=0.554, loss_mean=0.858, loss_mean_cls=0.0819, proj_loss=-0.386][2026-03-26 12:41:19] Step: 2467, Training Logs: loss_final: 0.532605, loss_mean: 0.841699, proj_loss: -0.391628, loss_mean_cls: 0.082534, grad_norm: 0.385788 +Steps: 0%| | 2468/1000000 [10:16<68:35:17, 4.04it/s, grad_norm=0.386, loss_final=0.533, loss_mean=0.842, loss_mean_cls=0.0825, proj_loss=-0.392][2026-03-26 12:41:19] Step: 2468, Training Logs: loss_final: 0.538737, loss_mean: 0.844323, proj_loss: -0.388440, loss_mean_cls: 0.082854, grad_norm: 0.354769 +Steps: 0%| | 2469/1000000 [10:16<68:35:43, 4.04it/s, grad_norm=0.355, loss_final=0.539, loss_mean=0.844, loss_mean_cls=0.0829, proj_loss=-0.388][2026-03-26 12:41:20] Step: 2469, Training Logs: loss_final: 0.543649, loss_mean: 0.860106, proj_loss: -0.396470, loss_mean_cls: 0.080012, grad_norm: 0.347491 +Steps: 0%| | 2470/1000000 [10:16<68:35:43, 4.04it/s, grad_norm=0.347, loss_final=0.544, loss_mean=0.86, loss_mean_cls=0.08, proj_loss=-0.396][2026-03-26 12:41:20] Step: 2470, Training Logs: loss_final: 0.542155, loss_mean: 0.851682, proj_loss: -0.391603, loss_mean_cls: 0.082076, grad_norm: 0.634934 +Steps: 0%| | 2471/1000000 [10:17<68:35:19, 4.04it/s, grad_norm=0.635, loss_final=0.542, loss_mean=0.852, loss_mean_cls=0.0821, proj_loss=-0.392][2026-03-26 12:41:20] Step: 2471, Training Logs: loss_final: 0.545561, loss_mean: 0.858392, proj_loss: -0.394152, loss_mean_cls: 0.081321, grad_norm: 0.388199 +Steps: 0%| | 2472/1000000 [10:17<68:36:07, 4.04it/s, grad_norm=0.388, loss_final=0.546, loss_mean=0.858, loss_mean_cls=0.0813, proj_loss=-0.394][2026-03-26 12:41:20] Step: 2472, Training Logs: loss_final: 0.530968, loss_mean: 0.840829, proj_loss: -0.392747, loss_mean_cls: 0.082886, grad_norm: 0.498088 +Steps: 0%| | 2473/1000000 [10:17<68:34:05, 4.04it/s, grad_norm=0.498, loss_final=0.531, loss_mean=0.841, loss_mean_cls=0.0829, proj_loss=-0.393][2026-03-26 12:41:21] Step: 2473, Training Logs: loss_final: 0.537800, loss_mean: 0.845483, proj_loss: -0.389712, loss_mean_cls: 0.082029, grad_norm: 0.529125 +Steps: 0%| | 2474/1000000 [10:17<68:35:46, 4.04it/s, grad_norm=0.529, loss_final=0.538, loss_mean=0.845, loss_mean_cls=0.082, proj_loss=-0.39][2026-03-26 12:41:21] Step: 2474, Training Logs: loss_final: 0.540151, loss_mean: 0.853837, proj_loss: -0.394536, loss_mean_cls: 0.080850, grad_norm: 0.495432 +Steps: 0%| | 2475/1000000 [10:18<68:34:09, 4.04it/s, grad_norm=0.495, loss_final=0.54, loss_mean=0.854, loss_mean_cls=0.0808, proj_loss=-0.395][2026-03-26 12:41:21] Step: 2475, Training Logs: loss_final: 0.542414, loss_mean: 0.849168, proj_loss: -0.388183, loss_mean_cls: 0.081429, grad_norm: 0.604373 +Steps: 0%| | 2476/1000000 [10:18<68:32:09, 4.04it/s, grad_norm=0.604, loss_final=0.542, loss_mean=0.849, loss_mean_cls=0.0814, proj_loss=-0.388][2026-03-26 12:41:21] Step: 2476, Training Logs: loss_final: 0.539163, loss_mean: 0.855653, proj_loss: -0.397351, loss_mean_cls: 0.080861, grad_norm: 0.472816 +Steps: 0%| | 2477/1000000 [10:18<68:39:17, 4.04it/s, grad_norm=0.473, loss_final=0.539, loss_mean=0.856, loss_mean_cls=0.0809, proj_loss=-0.397][2026-03-26 12:41:22] Step: 2477, Training Logs: loss_final: 0.548073, loss_mean: 0.851180, proj_loss: -0.387167, loss_mean_cls: 0.084061, grad_norm: 0.741807 +Steps: 0%| | 2478/1000000 [10:18<68:40:54, 4.03it/s, grad_norm=0.742, loss_final=0.548, loss_mean=0.851, loss_mean_cls=0.0841, proj_loss=-0.387][2026-03-26 12:41:22] Step: 2478, Training Logs: loss_final: 0.526119, loss_mean: 0.827252, proj_loss: -0.384634, loss_mean_cls: 0.083501, grad_norm: 0.367756 +Steps: 0%| | 2479/1000000 [10:19<68:45:15, 4.03it/s, grad_norm=0.368, loss_final=0.526, loss_mean=0.827, loss_mean_cls=0.0835, proj_loss=-0.385][2026-03-26 12:41:22] Step: 2479, Training Logs: loss_final: 0.534933, loss_mean: 0.847046, proj_loss: -0.394386, loss_mean_cls: 0.082273, grad_norm: 0.595787 +Steps: 0%| | 2480/1000000 [10:19<68:45:48, 4.03it/s, grad_norm=0.596, loss_final=0.535, loss_mean=0.847, loss_mean_cls=0.0823, proj_loss=-0.394][2026-03-26 12:41:22] Step: 2480, Training Logs: loss_final: 0.556306, loss_mean: 0.867389, proj_loss: -0.393329, loss_mean_cls: 0.082247, grad_norm: 0.670764 +Steps: 0%| | 2481/1000000 [10:19<68:42:56, 4.03it/s, grad_norm=0.671, loss_final=0.556, loss_mean=0.867, loss_mean_cls=0.0822, proj_loss=-0.393][2026-03-26 12:41:23] Step: 2481, Training Logs: loss_final: 0.535058, loss_mean: 0.842074, proj_loss: -0.389590, loss_mean_cls: 0.082574, grad_norm: 0.338420 +Steps: 0%| | 2482/1000000 [10:19<68:42:07, 4.03it/s, grad_norm=0.338, loss_final=0.535, loss_mean=0.842, loss_mean_cls=0.0826, proj_loss=-0.39][2026-03-26 12:41:23] Step: 2482, Training Logs: loss_final: 0.525365, loss_mean: 0.820481, proj_loss: -0.380285, loss_mean_cls: 0.085169, grad_norm: 0.733374 +Steps: 0%| | 2483/1000000 [10:20<68:39:10, 4.04it/s, grad_norm=0.733, loss_final=0.525, loss_mean=0.82, loss_mean_cls=0.0852, proj_loss=-0.38][2026-03-26 12:41:23] Step: 2483, Training Logs: loss_final: 0.546194, loss_mean: 0.867843, proj_loss: -0.401018, loss_mean_cls: 0.079368, grad_norm: 0.740346 +Steps: 0%| | 2484/1000000 [10:20<68:44:04, 4.03it/s, grad_norm=0.74, loss_final=0.546, loss_mean=0.868, loss_mean_cls=0.0794, proj_loss=-0.401][2026-03-26 12:41:23] Step: 2484, Training Logs: loss_final: 0.552659, loss_mean: 0.867147, proj_loss: -0.394908, loss_mean_cls: 0.080419, grad_norm: 0.490111 +Steps: 0%| | 2485/1000000 [10:20<69:20:40, 4.00it/s, grad_norm=0.49, loss_final=0.553, loss_mean=0.867, loss_mean_cls=0.0804, proj_loss=-0.395][2026-03-26 12:41:24] Step: 2485, Training Logs: loss_final: 0.536021, loss_mean: 0.851884, proj_loss: -0.395816, loss_mean_cls: 0.079953, grad_norm: 0.379746 +Steps: 0%| | 2486/1000000 [10:20<69:06:11, 4.01it/s, grad_norm=0.38, loss_final=0.536, loss_mean=0.852, loss_mean_cls=0.08, proj_loss=-0.396][2026-03-26 12:41:24] Step: 2486, Training Logs: loss_final: 0.541800, loss_mean: 0.849505, proj_loss: -0.390585, loss_mean_cls: 0.082881, grad_norm: 0.685948 +Steps: 0%| | 2487/1000000 [10:21<68:56:28, 4.02it/s, grad_norm=0.686, loss_final=0.542, loss_mean=0.85, loss_mean_cls=0.0829, proj_loss=-0.391][2026-03-26 12:41:24] Step: 2487, Training Logs: loss_final: 0.554464, loss_mean: 0.869131, proj_loss: -0.395740, loss_mean_cls: 0.081073, grad_norm: 0.473266 +Steps: 0%| | 2488/1000000 [10:21<68:52:07, 4.02it/s, grad_norm=0.473, loss_final=0.554, loss_mean=0.869, loss_mean_cls=0.0811, proj_loss=-0.396][2026-03-26 12:41:24] Step: 2488, Training Logs: loss_final: 0.520747, loss_mean: 0.825102, proj_loss: -0.388672, loss_mean_cls: 0.084317, grad_norm: 0.416573 +Steps: 0%| | 2489/1000000 [10:21<68:46:48, 4.03it/s, grad_norm=0.417, loss_final=0.521, loss_mean=0.825, loss_mean_cls=0.0843, proj_loss=-0.389][2026-03-26 12:41:25] Step: 2489, Training Logs: loss_final: 0.536707, loss_mean: 0.848697, proj_loss: -0.394669, loss_mean_cls: 0.082679, grad_norm: 0.529249 +Steps: 0%| | 2490/1000000 [10:21<68:43:47, 4.03it/s, grad_norm=0.529, loss_final=0.537, loss_mean=0.849, loss_mean_cls=0.0827, proj_loss=-0.395][2026-03-26 12:41:25] Step: 2490, Training Logs: loss_final: 0.535888, loss_mean: 0.837432, proj_loss: -0.385995, loss_mean_cls: 0.084451, grad_norm: 0.507915 +Steps: 0%| | 2491/1000000 [10:22<68:41:32, 4.03it/s, grad_norm=0.508, loss_final=0.536, loss_mean=0.837, loss_mean_cls=0.0845, proj_loss=-0.386][2026-03-26 12:41:25] Step: 2491, Training Logs: loss_final: 0.539290, loss_mean: 0.854856, proj_loss: -0.396165, loss_mean_cls: 0.080599, grad_norm: 0.370376 +Steps: 0%| | 2492/1000000 [10:22<68:42:03, 4.03it/s, grad_norm=0.37, loss_final=0.539, loss_mean=0.855, loss_mean_cls=0.0806, proj_loss=-0.396][2026-03-26 12:41:25] Step: 2492, Training Logs: loss_final: 0.551907, loss_mean: 0.869060, proj_loss: -0.397585, loss_mean_cls: 0.080432, grad_norm: 0.577201 +Steps: 0%| | 2493/1000000 [10:22<68:39:35, 4.04it/s, grad_norm=0.577, loss_final=0.552, loss_mean=0.869, loss_mean_cls=0.0804, proj_loss=-0.398][2026-03-26 12:41:26] Step: 2493, Training Logs: loss_final: 0.522658, loss_mean: 0.821436, proj_loss: -0.384230, loss_mean_cls: 0.085452, grad_norm: 0.387910 +Steps: 0%| | 2494/1000000 [10:22<68:39:16, 4.04it/s, grad_norm=0.388, loss_final=0.523, loss_mean=0.821, loss_mean_cls=0.0855, proj_loss=-0.384][2026-03-26 12:41:26] Step: 2494, Training Logs: loss_final: 0.539408, loss_mean: 0.840392, proj_loss: -0.384950, loss_mean_cls: 0.083965, grad_norm: 0.639216 +Steps: 0%| | 2495/1000000 [10:23<68:38:44, 4.04it/s, grad_norm=0.639, loss_final=0.539, loss_mean=0.84, loss_mean_cls=0.084, proj_loss=-0.385][2026-03-26 12:41:26] Step: 2495, Training Logs: loss_final: 0.531161, loss_mean: 0.839263, proj_loss: -0.392987, loss_mean_cls: 0.084885, grad_norm: 0.603279 +Steps: 0%| | 2496/1000000 [10:23<68:37:25, 4.04it/s, grad_norm=0.603, loss_final=0.531, loss_mean=0.839, loss_mean_cls=0.0849, proj_loss=-0.393][2026-03-26 12:41:26] Step: 2496, Training Logs: loss_final: 0.543072, loss_mean: 0.851853, proj_loss: -0.390810, loss_mean_cls: 0.082030, grad_norm: 0.300942 +Steps: 0%| | 2497/1000000 [10:23<68:35:29, 4.04it/s, grad_norm=0.301, loss_final=0.543, loss_mean=0.852, loss_mean_cls=0.082, proj_loss=-0.391][2026-03-26 12:41:27] Step: 2497, Training Logs: loss_final: 0.532872, loss_mean: 0.826409, proj_loss: -0.378177, loss_mean_cls: 0.084640, grad_norm: 0.649052 +Steps: 0%| | 2498/1000000 [10:23<68:40:45, 4.03it/s, grad_norm=0.649, loss_final=0.533, loss_mean=0.826, loss_mean_cls=0.0846, proj_loss=-0.378][2026-03-26 12:41:27] Step: 2498, Training Logs: loss_final: 0.520814, loss_mean: 0.823068, proj_loss: -0.388381, loss_mean_cls: 0.086127, grad_norm: 0.503848 +Steps: 0%| | 2499/1000000 [10:24<68:37:36, 4.04it/s, grad_norm=0.504, loss_final=0.521, loss_mean=0.823, loss_mean_cls=0.0861, proj_loss=-0.388][2026-03-26 12:41:27] Step: 2499, Training Logs: loss_final: 0.532446, loss_mean: 0.835467, proj_loss: -0.387220, loss_mean_cls: 0.084199, grad_norm: 0.440481 +Steps: 0%| | 2500/1000000 [10:24<68:37:51, 4.04it/s, grad_norm=0.44, loss_final=0.532, loss_mean=0.835, loss_mean_cls=0.0842, proj_loss=-0.387][2026-03-26 12:41:27] Step: 2500, Training Logs: loss_final: 0.538768, loss_mean: 0.856498, proj_loss: -0.398196, loss_mean_cls: 0.080465, grad_norm: 0.574866 +Steps: 0%| | 2501/1000000 [10:24<68:39:50, 4.04it/s, grad_norm=0.575, loss_final=0.539, loss_mean=0.856, loss_mean_cls=0.0805, proj_loss=-0.398][2026-03-26 12:41:28] Step: 2501, Training Logs: loss_final: 0.530195, loss_mean: 0.840914, proj_loss: -0.393275, loss_mean_cls: 0.082555, grad_norm: 0.613967 +Steps: 0%| | 2502/1000000 [10:24<68:41:30, 4.03it/s, grad_norm=0.614, loss_final=0.53, loss_mean=0.841, loss_mean_cls=0.0826, proj_loss=-0.393][2026-03-26 12:41:28] Step: 2502, Training Logs: loss_final: 0.546807, loss_mean: 0.841067, proj_loss: -0.379502, loss_mean_cls: 0.085242, grad_norm: 0.433349 +Steps: 0%| | 2503/1000000 [10:25<68:38:18, 4.04it/s, grad_norm=0.433, loss_final=0.547, loss_mean=0.841, loss_mean_cls=0.0852, proj_loss=-0.38][2026-03-26 12:41:28] Step: 2503, Training Logs: loss_final: 0.513483, loss_mean: 0.831806, proj_loss: -0.400482, loss_mean_cls: 0.082158, grad_norm: 0.431630 +Steps: 0%| | 2504/1000000 [10:25<68:37:41, 4.04it/s, grad_norm=0.432, loss_final=0.513, loss_mean=0.832, loss_mean_cls=0.0822, proj_loss=-0.4][2026-03-26 12:41:28] Step: 2504, Training Logs: loss_final: 0.519599, loss_mean: 0.836903, proj_loss: -0.398819, loss_mean_cls: 0.081515, grad_norm: 0.438576 +Steps: 0%| | 2505/1000000 [10:25<68:36:29, 4.04it/s, grad_norm=0.439, loss_final=0.52, loss_mean=0.837, loss_mean_cls=0.0815, proj_loss=-0.399][2026-03-26 12:41:29] Step: 2505, Training Logs: loss_final: 0.535631, loss_mean: 0.841955, proj_loss: -0.389088, loss_mean_cls: 0.082764, grad_norm: 0.368424 +Steps: 0%| | 2506/1000000 [10:25<68:37:36, 4.04it/s, grad_norm=0.368, loss_final=0.536, loss_mean=0.842, loss_mean_cls=0.0828, proj_loss=-0.389][2026-03-26 12:41:29] Step: 2506, Training Logs: loss_final: 0.534040, loss_mean: 0.847346, proj_loss: -0.393717, loss_mean_cls: 0.080410, grad_norm: 0.403347 +Steps: 0%| | 2507/1000000 [10:26<68:38:06, 4.04it/s, grad_norm=0.403, loss_final=0.534, loss_mean=0.847, loss_mean_cls=0.0804, proj_loss=-0.394][2026-03-26 12:41:29] Step: 2507, Training Logs: loss_final: 0.547118, loss_mean: 0.853570, proj_loss: -0.390097, loss_mean_cls: 0.083645, grad_norm: 0.557720 +Steps: 0%| | 2508/1000000 [10:26<68:37:48, 4.04it/s, grad_norm=0.558, loss_final=0.547, loss_mean=0.854, loss_mean_cls=0.0836, proj_loss=-0.39][2026-03-26 12:41:29] Step: 2508, Training Logs: loss_final: 0.532918, loss_mean: 0.843877, proj_loss: -0.394215, loss_mean_cls: 0.083256, grad_norm: 0.386315 +Steps: 0%| | 2509/1000000 [10:26<68:38:19, 4.04it/s, grad_norm=0.386, loss_final=0.533, loss_mean=0.844, loss_mean_cls=0.0833, proj_loss=-0.394][2026-03-26 12:41:30] Step: 2509, Training Logs: loss_final: 0.545934, loss_mean: 0.851795, proj_loss: -0.388225, loss_mean_cls: 0.082363, grad_norm: 0.479331 +Steps: 0%| | 2510/1000000 [10:26<68:38:38, 4.04it/s, grad_norm=0.479, loss_final=0.546, loss_mean=0.852, loss_mean_cls=0.0824, proj_loss=-0.388][2026-03-26 12:41:30] Step: 2510, Training Logs: loss_final: 0.564151, loss_mean: 0.879469, proj_loss: -0.395821, loss_mean_cls: 0.080503, grad_norm: 0.427866 +Steps: 0%| | 2511/1000000 [10:27<68:38:01, 4.04it/s, grad_norm=0.428, loss_final=0.564, loss_mean=0.879, loss_mean_cls=0.0805, proj_loss=-0.396][2026-03-26 12:41:30] Step: 2511, Training Logs: loss_final: 0.530918, loss_mean: 0.837604, proj_loss: -0.390182, loss_mean_cls: 0.083496, grad_norm: 0.471001 +Steps: 0%| | 2512/1000000 [10:27<68:35:48, 4.04it/s, grad_norm=0.471, loss_final=0.531, loss_mean=0.838, loss_mean_cls=0.0835, proj_loss=-0.39][2026-03-26 12:41:30] Step: 2512, Training Logs: loss_final: 0.539405, loss_mean: 0.857391, proj_loss: -0.399238, loss_mean_cls: 0.081251, grad_norm: 0.544568 +Steps: 0%| | 2513/1000000 [10:27<68:36:07, 4.04it/s, grad_norm=0.545, loss_final=0.539, loss_mean=0.857, loss_mean_cls=0.0813, proj_loss=-0.399][2026-03-26 12:41:31] Step: 2513, Training Logs: loss_final: 0.533666, loss_mean: 0.841777, proj_loss: -0.393275, loss_mean_cls: 0.085164, grad_norm: 0.452368 +Steps: 0%| | 2514/1000000 [10:27<68:36:25, 4.04it/s, grad_norm=0.452, loss_final=0.534, loss_mean=0.842, loss_mean_cls=0.0852, proj_loss=-0.393][2026-03-26 12:41:31] Step: 2514, Training Logs: loss_final: 0.522316, loss_mean: 0.839943, proj_loss: -0.399972, loss_mean_cls: 0.082345, grad_norm: 0.467789 +Steps: 0%| | 2515/1000000 [10:28<68:41:52, 4.03it/s, grad_norm=0.468, loss_final=0.522, loss_mean=0.84, loss_mean_cls=0.0823, proj_loss=-0.4][2026-03-26 12:41:31] Step: 2515, Training Logs: loss_final: 0.555630, loss_mean: 0.870090, proj_loss: -0.394901, loss_mean_cls: 0.080441, grad_norm: 0.456710 +Steps: 0%| | 2516/1000000 [10:28<68:40:44, 4.03it/s, grad_norm=0.457, loss_final=0.556, loss_mean=0.87, loss_mean_cls=0.0804, proj_loss=-0.395][2026-03-26 12:41:31] Step: 2516, Training Logs: loss_final: 0.537230, loss_mean: 0.862019, proj_loss: -0.404389, loss_mean_cls: 0.079599, grad_norm: 0.515317 +Steps: 0%| | 2517/1000000 [10:28<68:48:21, 4.03it/s, grad_norm=0.515, loss_final=0.537, loss_mean=0.862, loss_mean_cls=0.0796, proj_loss=-0.404][2026-03-26 12:41:32] Step: 2517, Training Logs: loss_final: 0.540380, loss_mean: 0.843862, proj_loss: -0.386658, loss_mean_cls: 0.083176, grad_norm: 0.516132 +Steps: 0%| | 2518/1000000 [10:28<68:41:14, 4.03it/s, grad_norm=0.516, loss_final=0.54, loss_mean=0.844, loss_mean_cls=0.0832, proj_loss=-0.387][2026-03-26 12:41:32] Step: 2518, Training Logs: loss_final: 0.536558, loss_mean: 0.839728, proj_loss: -0.387376, loss_mean_cls: 0.084205, grad_norm: 0.474858 +Steps: 0%| | 2519/1000000 [10:29<68:40:21, 4.03it/s, grad_norm=0.475, loss_final=0.537, loss_mean=0.84, loss_mean_cls=0.0842, proj_loss=-0.387][2026-03-26 12:41:32] Step: 2519, Training Logs: loss_final: 0.515997, loss_mean: 0.821177, proj_loss: -0.390143, loss_mean_cls: 0.084963, grad_norm: 0.487179 +Steps: 0%| | 2520/1000000 [10:29<68:42:45, 4.03it/s, grad_norm=0.487, loss_final=0.516, loss_mean=0.821, loss_mean_cls=0.085, proj_loss=-0.39][2026-03-26 12:41:32] Step: 2520, Training Logs: loss_final: 0.529239, loss_mean: 0.841826, proj_loss: -0.394214, loss_mean_cls: 0.081627, grad_norm: 0.461034 +Steps: 0%| | 2521/1000000 [10:29<68:38:55, 4.04it/s, grad_norm=0.461, loss_final=0.529, loss_mean=0.842, loss_mean_cls=0.0816, proj_loss=-0.394][2026-03-26 12:41:33] Step: 2521, Training Logs: loss_final: 0.533560, loss_mean: 0.842411, proj_loss: -0.391833, loss_mean_cls: 0.082982, grad_norm: 0.560054 +Steps: 0%| | 2522/1000000 [10:29<68:38:32, 4.04it/s, grad_norm=0.56, loss_final=0.534, loss_mean=0.842, loss_mean_cls=0.083, proj_loss=-0.392][2026-03-26 12:41:33] Step: 2522, Training Logs: loss_final: 0.534540, loss_mean: 0.849111, proj_loss: -0.397048, loss_mean_cls: 0.082477, grad_norm: 0.427075 +Steps: 0%| | 2523/1000000 [10:30<68:37:10, 4.04it/s, grad_norm=0.427, loss_final=0.535, loss_mean=0.849, loss_mean_cls=0.0825, proj_loss=-0.397][2026-03-26 12:41:33] Step: 2523, Training Logs: loss_final: 0.541943, loss_mean: 0.861378, proj_loss: -0.399345, loss_mean_cls: 0.079910, grad_norm: 0.562633 +Steps: 0%| | 2524/1000000 [10:30<68:37:04, 4.04it/s, grad_norm=0.563, loss_final=0.542, loss_mean=0.861, loss_mean_cls=0.0799, proj_loss=-0.399][2026-03-26 12:41:33] Step: 2524, Training Logs: loss_final: 0.515737, loss_mean: 0.821395, proj_loss: -0.389474, loss_mean_cls: 0.083816, grad_norm: 0.580501 +Steps: 0%| | 2525/1000000 [10:30<68:37:38, 4.04it/s, grad_norm=0.581, loss_final=0.516, loss_mean=0.821, loss_mean_cls=0.0838, proj_loss=-0.389][2026-03-26 12:41:34] Step: 2525, Training Logs: loss_final: 0.534501, loss_mean: 0.841478, proj_loss: -0.390074, loss_mean_cls: 0.083096, grad_norm: 0.446225 +Steps: 0%| | 2526/1000000 [10:30<68:39:04, 4.04it/s, grad_norm=0.446, loss_final=0.535, loss_mean=0.841, loss_mean_cls=0.0831, proj_loss=-0.39][2026-03-26 12:41:34] Step: 2526, Training Logs: loss_final: 0.537881, loss_mean: 0.849896, proj_loss: -0.392800, loss_mean_cls: 0.080784, grad_norm: 0.452323 +Steps: 0%| | 2527/1000000 [10:31<68:39:23, 4.04it/s, grad_norm=0.452, loss_final=0.538, loss_mean=0.85, loss_mean_cls=0.0808, proj_loss=-0.393][2026-03-26 12:41:34] Step: 2527, Training Logs: loss_final: 0.543059, loss_mean: 0.862522, proj_loss: -0.399227, loss_mean_cls: 0.079764, grad_norm: 0.354954 +Steps: 0%| | 2528/1000000 [10:31<68:38:55, 4.04it/s, grad_norm=0.355, loss_final=0.543, loss_mean=0.863, loss_mean_cls=0.0798, proj_loss=-0.399][2026-03-26 12:41:34] Step: 2528, Training Logs: loss_final: 0.540562, loss_mean: 0.860429, proj_loss: -0.400587, loss_mean_cls: 0.080720, grad_norm: 0.388055 +Steps: 0%| | 2529/1000000 [10:31<68:36:53, 4.04it/s, grad_norm=0.388, loss_final=0.541, loss_mean=0.86, loss_mean_cls=0.0807, proj_loss=-0.401][2026-03-26 12:41:35] Step: 2529, Training Logs: loss_final: 0.535594, loss_mean: 0.833212, proj_loss: -0.382933, loss_mean_cls: 0.085314, grad_norm: 0.512397 +Steps: 0%| | 2530/1000000 [10:31<68:36:30, 4.04it/s, grad_norm=0.512, loss_final=0.536, loss_mean=0.833, loss_mean_cls=0.0853, proj_loss=-0.383][2026-03-26 12:41:35] Step: 2530, Training Logs: loss_final: 0.532608, loss_mean: 0.844551, proj_loss: -0.393877, loss_mean_cls: 0.081933, grad_norm: 0.357000 +Steps: 0%| | 2531/1000000 [10:32<68:38:08, 4.04it/s, grad_norm=0.357, loss_final=0.533, loss_mean=0.845, loss_mean_cls=0.0819, proj_loss=-0.394][2026-03-26 12:41:35] Step: 2531, Training Logs: loss_final: 0.529236, loss_mean: 0.837711, proj_loss: -0.390840, loss_mean_cls: 0.082365, grad_norm: 0.378802 +Steps: 0%| | 2532/1000000 [10:32<70:57:48, 3.90it/s, grad_norm=0.379, loss_final=0.529, loss_mean=0.838, loss_mean_cls=0.0824, proj_loss=-0.391][2026-03-26 12:41:35] Step: 2532, Training Logs: loss_final: 0.545203, loss_mean: 0.866764, proj_loss: -0.401843, loss_mean_cls: 0.080282, grad_norm: 0.350110 +Steps: 0%| | 2533/1000000 [10:32<73:33:47, 3.77it/s, grad_norm=0.35, loss_final=0.545, loss_mean=0.867, loss_mean_cls=0.0803, proj_loss=-0.402][2026-03-26 12:41:36] Step: 2533, Training Logs: loss_final: 0.527084, loss_mean: 0.830351, proj_loss: -0.386672, loss_mean_cls: 0.083406, grad_norm: 0.468871 +Steps: 0%| | 2534/1000000 [10:32<72:06:00, 3.84it/s, grad_norm=0.469, loss_final=0.527, loss_mean=0.83, loss_mean_cls=0.0834, proj_loss=-0.387][2026-03-26 12:41:36] Step: 2534, Training Logs: loss_final: 0.539069, loss_mean: 0.859868, proj_loss: -0.401116, loss_mean_cls: 0.080317, grad_norm: 0.456678 +Steps: 0%| | 2535/1000000 [10:33<71:04:01, 3.90it/s, grad_norm=0.457, loss_final=0.539, loss_mean=0.86, loss_mean_cls=0.0803, proj_loss=-0.401][2026-03-26 12:41:36] Step: 2535, Training Logs: loss_final: 0.534936, loss_mean: 0.835350, proj_loss: -0.386445, loss_mean_cls: 0.086031, grad_norm: 0.411665 +Steps: 0%| | 2536/1000000 [10:33<70:21:08, 3.94it/s, grad_norm=0.412, loss_final=0.535, loss_mean=0.835, loss_mean_cls=0.086, proj_loss=-0.386][2026-03-26 12:41:36] Step: 2536, Training Logs: loss_final: 0.540148, loss_mean: 0.858794, proj_loss: -0.397967, loss_mean_cls: 0.079321, grad_norm: 0.383360 +Steps: 0%| | 2537/1000000 [10:33<69:49:05, 3.97it/s, grad_norm=0.383, loss_final=0.54, loss_mean=0.859, loss_mean_cls=0.0793, proj_loss=-0.398][2026-03-26 12:41:37] Step: 2537, Training Logs: loss_final: 0.534431, loss_mean: 0.845767, proj_loss: -0.393240, loss_mean_cls: 0.081904, grad_norm: 0.364118 +Steps: 0%| | 2538/1000000 [10:33<69:29:48, 3.99it/s, grad_norm=0.364, loss_final=0.534, loss_mean=0.846, loss_mean_cls=0.0819, proj_loss=-0.393][2026-03-26 12:41:37] Step: 2538, Training Logs: loss_final: 0.557660, loss_mean: 0.875553, proj_loss: -0.398018, loss_mean_cls: 0.080125, grad_norm: 0.395474 +Steps: 0%| | 2539/1000000 [10:34<69:18:19, 4.00it/s, grad_norm=0.395, loss_final=0.558, loss_mean=0.876, loss_mean_cls=0.0801, proj_loss=-0.398][2026-03-26 12:41:37] Step: 2539, Training Logs: loss_final: 0.530137, loss_mean: 0.861948, proj_loss: -0.410260, loss_mean_cls: 0.078449, grad_norm: 0.540127 +Steps: 0%| | 2540/1000000 [10:34<69:10:42, 4.01it/s, grad_norm=0.54, loss_final=0.53, loss_mean=0.862, loss_mean_cls=0.0784, proj_loss=-0.41][2026-03-26 12:41:37] Step: 2540, Training Logs: loss_final: 0.533760, loss_mean: 0.836329, proj_loss: -0.386534, loss_mean_cls: 0.083965, grad_norm: 0.800487 +Steps: 0%| | 2541/1000000 [10:34<68:59:21, 4.02it/s, grad_norm=0.8, loss_final=0.534, loss_mean=0.836, loss_mean_cls=0.084, proj_loss=-0.387][2026-03-26 12:41:38] Step: 2541, Training Logs: loss_final: 0.542208, loss_mean: 0.851770, proj_loss: -0.392847, loss_mean_cls: 0.083285, grad_norm: 0.525762 +Steps: 0%| | 2542/1000000 [10:34<69:00:02, 4.02it/s, grad_norm=0.526, loss_final=0.542, loss_mean=0.852, loss_mean_cls=0.0833, proj_loss=-0.393][2026-03-26 12:41:38] Step: 2542, Training Logs: loss_final: 0.527187, loss_mean: 0.839392, proj_loss: -0.394667, loss_mean_cls: 0.082463, grad_norm: 0.456467 +Steps: 0%| | 2543/1000000 [10:35<68:48:57, 4.03it/s, grad_norm=0.456, loss_final=0.527, loss_mean=0.839, loss_mean_cls=0.0825, proj_loss=-0.395][2026-03-26 12:41:38] Step: 2543, Training Logs: loss_final: 0.551967, loss_mean: 0.868711, proj_loss: -0.397898, loss_mean_cls: 0.081154, grad_norm: 0.637196 +Steps: 0%| | 2544/1000000 [10:35<68:44:09, 4.03it/s, grad_norm=0.637, loss_final=0.552, loss_mean=0.869, loss_mean_cls=0.0812, proj_loss=-0.398][2026-03-26 12:41:38] Step: 2544, Training Logs: loss_final: 0.519313, loss_mean: 0.827364, proj_loss: -0.390379, loss_mean_cls: 0.082328, grad_norm: 0.537246 +Steps: 0%| | 2545/1000000 [10:35<69:36:30, 3.98it/s, grad_norm=0.537, loss_final=0.519, loss_mean=0.827, loss_mean_cls=0.0823, proj_loss=-0.39][2026-03-26 12:41:39] Step: 2545, Training Logs: loss_final: 0.538745, loss_mean: 0.854840, proj_loss: -0.397352, loss_mean_cls: 0.081258, grad_norm: 0.317283 +Steps: 0%| | 2546/1000000 [10:35<69:24:25, 3.99it/s, grad_norm=0.317, loss_final=0.539, loss_mean=0.855, loss_mean_cls=0.0813, proj_loss=-0.397][2026-03-26 12:41:39] Step: 2546, Training Logs: loss_final: 0.533063, loss_mean: 0.840774, proj_loss: -0.391116, loss_mean_cls: 0.083405, grad_norm: 0.482120 +Steps: 0%| | 2547/1000000 [10:36<69:04:49, 4.01it/s, grad_norm=0.482, loss_final=0.533, loss_mean=0.841, loss_mean_cls=0.0834, proj_loss=-0.391][2026-03-26 12:41:39] Step: 2547, Training Logs: loss_final: 0.537204, loss_mean: 0.851465, proj_loss: -0.395859, loss_mean_cls: 0.081599, grad_norm: 0.393855 +Steps: 0%| | 2548/1000000 [10:36<68:57:01, 4.02it/s, grad_norm=0.394, loss_final=0.537, loss_mean=0.851, loss_mean_cls=0.0816, proj_loss=-0.396][2026-03-26 12:41:39] Step: 2548, Training Logs: loss_final: 0.537866, loss_mean: 0.840211, proj_loss: -0.387460, loss_mean_cls: 0.085115, grad_norm: 0.526441 +Steps: 0%| | 2549/1000000 [10:36<68:51:50, 4.02it/s, grad_norm=0.526, loss_final=0.538, loss_mean=0.84, loss_mean_cls=0.0851, proj_loss=-0.387][2026-03-26 12:41:40] Step: 2549, Training Logs: loss_final: 0.539660, loss_mean: 0.852921, proj_loss: -0.395618, loss_mean_cls: 0.082356, grad_norm: 0.576550 +Steps: 0%| | 2550/1000000 [10:36<68:49:27, 4.03it/s, grad_norm=0.577, loss_final=0.54, loss_mean=0.853, loss_mean_cls=0.0824, proj_loss=-0.396][2026-03-26 12:41:40] Step: 2550, Training Logs: loss_final: 0.535572, loss_mean: 0.854474, proj_loss: -0.399572, loss_mean_cls: 0.080670, grad_norm: 0.362113 +Steps: 0%| | 2551/1000000 [10:37<68:49:48, 4.03it/s, grad_norm=0.362, loss_final=0.536, loss_mean=0.854, loss_mean_cls=0.0807, proj_loss=-0.4][2026-03-26 12:41:40] Step: 2551, Training Logs: loss_final: 0.549085, loss_mean: 0.858945, proj_loss: -0.391585, loss_mean_cls: 0.081725, grad_norm: 0.501796 +Steps: 0%| | 2552/1000000 [10:37<68:43:53, 4.03it/s, grad_norm=0.502, loss_final=0.549, loss_mean=0.859, loss_mean_cls=0.0817, proj_loss=-0.392][2026-03-26 12:41:40] Step: 2552, Training Logs: loss_final: 0.540455, loss_mean: 0.853571, proj_loss: -0.394868, loss_mean_cls: 0.081752, grad_norm: 0.568275 +Steps: 0%| | 2553/1000000 [10:37<68:45:45, 4.03it/s, grad_norm=0.568, loss_final=0.54, loss_mean=0.854, loss_mean_cls=0.0818, proj_loss=-0.395][2026-03-26 12:41:41] Step: 2553, Training Logs: loss_final: 0.531362, loss_mean: 0.838204, proj_loss: -0.390916, loss_mean_cls: 0.084074, grad_norm: 0.352791 +Steps: 0%| | 2554/1000000 [10:37<68:43:03, 4.03it/s, grad_norm=0.353, loss_final=0.531, loss_mean=0.838, loss_mean_cls=0.0841, proj_loss=-0.391][2026-03-26 12:41:41] Step: 2554, Training Logs: loss_final: 0.534827, loss_mean: 0.848038, proj_loss: -0.394546, loss_mean_cls: 0.081334, grad_norm: 0.473955 +Steps: 0%| | 2555/1000000 [10:38<68:40:49, 4.03it/s, grad_norm=0.474, loss_final=0.535, loss_mean=0.848, loss_mean_cls=0.0813, proj_loss=-0.395][2026-03-26 12:41:41] Step: 2555, Training Logs: loss_final: 0.540235, loss_mean: 0.855665, proj_loss: -0.397001, loss_mean_cls: 0.081571, grad_norm: 0.540073 +Steps: 0%| | 2556/1000000 [10:38<68:40:23, 4.03it/s, grad_norm=0.54, loss_final=0.54, loss_mean=0.856, loss_mean_cls=0.0816, proj_loss=-0.397][2026-03-26 12:41:41] Step: 2556, Training Logs: loss_final: 0.531206, loss_mean: 0.844092, proj_loss: -0.394182, loss_mean_cls: 0.081296, grad_norm: 0.442155 +Steps: 0%| | 2557/1000000 [10:38<68:40:24, 4.03it/s, grad_norm=0.442, loss_final=0.531, loss_mean=0.844, loss_mean_cls=0.0813, proj_loss=-0.394][2026-03-26 12:41:42] Step: 2557, Training Logs: loss_final: 0.537366, loss_mean: 0.855410, proj_loss: -0.398776, loss_mean_cls: 0.080732, grad_norm: 0.391826 +Steps: 0%| | 2558/1000000 [10:38<68:40:00, 4.03it/s, grad_norm=0.392, loss_final=0.537, loss_mean=0.855, loss_mean_cls=0.0807, proj_loss=-0.399][2026-03-26 12:41:42] Step: 2558, Training Logs: loss_final: 0.555466, loss_mean: 0.874527, proj_loss: -0.399643, loss_mean_cls: 0.080582, grad_norm: 0.598595 +Steps: 0%| | 2559/1000000 [10:39<71:01:50, 3.90it/s, grad_norm=0.599, loss_final=0.555, loss_mean=0.875, loss_mean_cls=0.0806, proj_loss=-0.4][2026-03-26 12:41:42] Step: 2559, Training Logs: loss_final: 0.534826, loss_mean: 0.843787, proj_loss: -0.392051, loss_mean_cls: 0.083090, grad_norm: 0.482750 +Steps: 0%| | 2560/1000000 [10:39<70:17:22, 3.94it/s, grad_norm=0.483, loss_final=0.535, loss_mean=0.844, loss_mean_cls=0.0831, proj_loss=-0.392][2026-03-26 12:41:42] Step: 2560, Training Logs: loss_final: 0.530714, loss_mean: 0.829385, proj_loss: -0.384024, loss_mean_cls: 0.085353, grad_norm: 0.501174 +Steps: 0%| | 2561/1000000 [10:39<69:47:57, 3.97it/s, grad_norm=0.501, loss_final=0.531, loss_mean=0.829, loss_mean_cls=0.0854, proj_loss=-0.384][2026-03-26 12:41:43] Step: 2561, Training Logs: loss_final: 0.558008, loss_mean: 0.874665, proj_loss: -0.396823, loss_mean_cls: 0.080167, grad_norm: 0.843945 +Steps: 0%| | 2562/1000000 [10:39<69:26:20, 3.99it/s, grad_norm=0.844, loss_final=0.558, loss_mean=0.875, loss_mean_cls=0.0802, proj_loss=-0.397][2026-03-26 12:41:43] Step: 2562, Training Logs: loss_final: 0.544401, loss_mean: 0.849934, proj_loss: -0.389015, loss_mean_cls: 0.083482, grad_norm: 0.996795 +Steps: 0%| | 2563/1000000 [10:40<69:16:25, 4.00it/s, grad_norm=0.997, loss_final=0.544, loss_mean=0.85, loss_mean_cls=0.0835, proj_loss=-0.389][2026-03-26 12:41:43] Step: 2563, Training Logs: loss_final: 0.527564, loss_mean: 0.839619, proj_loss: -0.394212, loss_mean_cls: 0.082156, grad_norm: 0.433851 +Steps: 0%| | 2564/1000000 [10:40<69:04:23, 4.01it/s, grad_norm=0.434, loss_final=0.528, loss_mean=0.84, loss_mean_cls=0.0822, proj_loss=-0.394][2026-03-26 12:41:43] Step: 2564, Training Logs: loss_final: 0.544310, loss_mean: 0.856144, proj_loss: -0.393362, loss_mean_cls: 0.081528, grad_norm: 0.680933 +Steps: 0%| | 2565/1000000 [10:40<68:56:24, 4.02it/s, grad_norm=0.681, loss_final=0.544, loss_mean=0.856, loss_mean_cls=0.0815, proj_loss=-0.393][2026-03-26 12:41:44] Step: 2565, Training Logs: loss_final: 0.538723, loss_mean: 0.857821, proj_loss: -0.398933, loss_mean_cls: 0.079835, grad_norm: 0.619794 +Steps: 0%| | 2566/1000000 [10:40<68:48:44, 4.03it/s, grad_norm=0.62, loss_final=0.539, loss_mean=0.858, loss_mean_cls=0.0798, proj_loss=-0.399][2026-03-26 12:41:44] Step: 2566, Training Logs: loss_final: 0.545052, loss_mean: 0.851887, proj_loss: -0.390096, loss_mean_cls: 0.083261, grad_norm: 0.349748 +Steps: 0%| | 2567/1000000 [10:41<68:44:56, 4.03it/s, grad_norm=0.35, loss_final=0.545, loss_mean=0.852, loss_mean_cls=0.0833, proj_loss=-0.39][2026-03-26 12:41:44] Step: 2567, Training Logs: loss_final: 0.545139, loss_mean: 0.855537, proj_loss: -0.393591, loss_mean_cls: 0.083192, grad_norm: 0.445615 +Steps: 0%| | 2568/1000000 [10:41<68:42:42, 4.03it/s, grad_norm=0.446, loss_final=0.545, loss_mean=0.856, loss_mean_cls=0.0832, proj_loss=-0.394][2026-03-26 12:41:44] Step: 2568, Training Logs: loss_final: 0.527402, loss_mean: 0.827682, proj_loss: -0.383535, loss_mean_cls: 0.083254, grad_norm: 0.445208 +Steps: 0%| | 2569/1000000 [10:41<68:40:07, 4.03it/s, grad_norm=0.445, loss_final=0.527, loss_mean=0.828, loss_mean_cls=0.0833, proj_loss=-0.384][2026-03-26 12:41:45] Step: 2569, Training Logs: loss_final: 0.550506, loss_mean: 0.861300, proj_loss: -0.392408, loss_mean_cls: 0.081614, grad_norm: 0.380644 +Steps: 0%| | 2570/1000000 [10:41<68:37:58, 4.04it/s, grad_norm=0.381, loss_final=0.551, loss_mean=0.861, loss_mean_cls=0.0816, proj_loss=-0.392][2026-03-26 12:41:45] Step: 2570, Training Logs: loss_final: 0.530579, loss_mean: 0.838687, proj_loss: -0.391786, loss_mean_cls: 0.083678, grad_norm: 0.367806 +Steps: 0%| | 2571/1000000 [10:42<68:40:12, 4.03it/s, grad_norm=0.368, loss_final=0.531, loss_mean=0.839, loss_mean_cls=0.0837, proj_loss=-0.392][2026-03-26 12:41:45] Step: 2571, Training Logs: loss_final: 0.530436, loss_mean: 0.849391, proj_loss: -0.400405, loss_mean_cls: 0.081450, grad_norm: 0.537260 +Steps: 0%| | 2572/1000000 [10:42<68:35:11, 4.04it/s, grad_norm=0.537, loss_final=0.53, loss_mean=0.849, loss_mean_cls=0.0815, proj_loss=-0.4][2026-03-26 12:41:45] Step: 2572, Training Logs: loss_final: 0.517161, loss_mean: 0.831604, proj_loss: -0.396337, loss_mean_cls: 0.081894, grad_norm: 0.623421 +Steps: 0%| | 2573/1000000 [10:42<68:34:37, 4.04it/s, grad_norm=0.623, loss_final=0.517, loss_mean=0.832, loss_mean_cls=0.0819, proj_loss=-0.396][2026-03-26 12:41:46] Step: 2573, Training Logs: loss_final: 0.547158, loss_mean: 0.862874, proj_loss: -0.396614, loss_mean_cls: 0.080898, grad_norm: 0.744260 +Steps: 0%| | 2574/1000000 [10:42<68:36:07, 4.04it/s, grad_norm=0.744, loss_final=0.547, loss_mean=0.863, loss_mean_cls=0.0809, proj_loss=-0.397][2026-03-26 12:41:46] Step: 2574, Training Logs: loss_final: 0.531210, loss_mean: 0.845057, proj_loss: -0.395724, loss_mean_cls: 0.081876, grad_norm: 0.400017 +Steps: 0%| | 2575/1000000 [10:43<68:37:46, 4.04it/s, grad_norm=0.4, loss_final=0.531, loss_mean=0.845, loss_mean_cls=0.0819, proj_loss=-0.396][2026-03-26 12:41:46] Step: 2575, Training Logs: loss_final: 0.518635, loss_mean: 0.828724, proj_loss: -0.392973, loss_mean_cls: 0.082885, grad_norm: 0.478011 +Steps: 0%| | 2576/1000000 [10:43<68:37:41, 4.04it/s, grad_norm=0.478, loss_final=0.519, loss_mean=0.829, loss_mean_cls=0.0829, proj_loss=-0.393][2026-03-26 12:41:46] Step: 2576, Training Logs: loss_final: 0.528650, loss_mean: 0.832963, proj_loss: -0.388356, loss_mean_cls: 0.084043, grad_norm: 0.346304 +Steps: 0%| | 2577/1000000 [10:43<68:36:35, 4.04it/s, grad_norm=0.346, loss_final=0.529, loss_mean=0.833, loss_mean_cls=0.084, proj_loss=-0.388][2026-03-26 12:41:47] Step: 2577, Training Logs: loss_final: 0.524377, loss_mean: 0.832811, proj_loss: -0.391263, loss_mean_cls: 0.082829, grad_norm: 0.376404 +Steps: 0%| | 2578/1000000 [10:43<68:37:53, 4.04it/s, grad_norm=0.376, loss_final=0.524, loss_mean=0.833, loss_mean_cls=0.0828, proj_loss=-0.391][2026-03-26 12:41:47] Step: 2578, Training Logs: loss_final: 0.545391, loss_mean: 0.845862, proj_loss: -0.384226, loss_mean_cls: 0.083755, grad_norm: 0.484311 +Steps: 0%| | 2579/1000000 [10:44<68:37:40, 4.04it/s, grad_norm=0.484, loss_final=0.545, loss_mean=0.846, loss_mean_cls=0.0838, proj_loss=-0.384][2026-03-26 12:41:47] Step: 2579, Training Logs: loss_final: 0.515590, loss_mean: 0.833594, proj_loss: -0.398822, loss_mean_cls: 0.080817, grad_norm: 0.408802 +Steps: 0%| | 2580/1000000 [10:44<68:38:32, 4.04it/s, grad_norm=0.409, loss_final=0.516, loss_mean=0.834, loss_mean_cls=0.0808, proj_loss=-0.399][2026-03-26 12:41:47] Step: 2580, Training Logs: loss_final: 0.541286, loss_mean: 0.843755, proj_loss: -0.386225, loss_mean_cls: 0.083755, grad_norm: 0.680428 +Steps: 0%| | 2581/1000000 [10:44<68:37:21, 4.04it/s, grad_norm=0.68, loss_final=0.541, loss_mean=0.844, loss_mean_cls=0.0838, proj_loss=-0.386][2026-03-26 12:41:47] Step: 2581, Training Logs: loss_final: 0.560088, loss_mean: 0.878865, proj_loss: -0.398669, loss_mean_cls: 0.079892, grad_norm: 0.674297 +Steps: 0%| | 2582/1000000 [10:44<68:37:27, 4.04it/s, grad_norm=0.674, loss_final=0.56, loss_mean=0.879, loss_mean_cls=0.0799, proj_loss=-0.399][2026-03-26 12:41:48] Step: 2582, Training Logs: loss_final: 0.548346, loss_mean: 0.863535, proj_loss: -0.396380, loss_mean_cls: 0.081190, grad_norm: 0.371783 +Steps: 0%| | 2583/1000000 [10:45<68:38:31, 4.04it/s, grad_norm=0.372, loss_final=0.548, loss_mean=0.864, loss_mean_cls=0.0812, proj_loss=-0.396][2026-03-26 12:41:48] Step: 2583, Training Logs: loss_final: 0.531049, loss_mean: 0.835264, proj_loss: -0.389191, loss_mean_cls: 0.084976, grad_norm: 0.660593 +Steps: 0%| | 2584/1000000 [10:45<68:40:15, 4.03it/s, grad_norm=0.661, loss_final=0.531, loss_mean=0.835, loss_mean_cls=0.085, proj_loss=-0.389][2026-03-26 12:41:48] Step: 2584, Training Logs: loss_final: 0.530296, loss_mean: 0.842833, proj_loss: -0.395457, loss_mean_cls: 0.082920, grad_norm: 0.461040 +Steps: 0%| | 2585/1000000 [10:45<68:38:06, 4.04it/s, grad_norm=0.461, loss_final=0.53, loss_mean=0.843, loss_mean_cls=0.0829, proj_loss=-0.395][2026-03-26 12:41:48] Step: 2585, Training Logs: loss_final: 0.539850, loss_mean: 0.856857, proj_loss: -0.398398, loss_mean_cls: 0.081391, grad_norm: 0.414221 +Steps: 0%| | 2586/1000000 [10:45<68:40:34, 4.03it/s, grad_norm=0.414, loss_final=0.54, loss_mean=0.857, loss_mean_cls=0.0814, proj_loss=-0.398][2026-03-26 12:41:49] Step: 2586, Training Logs: loss_final: 0.538871, loss_mean: 0.863106, proj_loss: -0.402835, loss_mean_cls: 0.078601, grad_norm: 0.555233 +Steps: 0%| | 2587/1000000 [10:46<68:37:05, 4.04it/s, grad_norm=0.555, loss_final=0.539, loss_mean=0.863, loss_mean_cls=0.0786, proj_loss=-0.403][2026-03-26 12:41:49] Step: 2587, Training Logs: loss_final: 0.528657, loss_mean: 0.840889, proj_loss: -0.394880, loss_mean_cls: 0.082648, grad_norm: 0.568521 +Steps: 0%| | 2588/1000000 [10:46<68:41:40, 4.03it/s, grad_norm=0.569, loss_final=0.529, loss_mean=0.841, loss_mean_cls=0.0826, proj_loss=-0.395][2026-03-26 12:41:49] Step: 2588, Training Logs: loss_final: 0.539210, loss_mean: 0.855395, proj_loss: -0.398001, loss_mean_cls: 0.081816, grad_norm: 0.426456 +Steps: 0%| | 2589/1000000 [10:46<68:33:21, 4.04it/s, grad_norm=0.426, loss_final=0.539, loss_mean=0.855, loss_mean_cls=0.0818, proj_loss=-0.398][2026-03-26 12:41:49] Step: 2589, Training Logs: loss_final: 0.527743, loss_mean: 0.846944, proj_loss: -0.399686, loss_mean_cls: 0.080485, grad_norm: 0.493898 +Steps: 0%| | 2590/1000000 [10:46<68:37:23, 4.04it/s, grad_norm=0.494, loss_final=0.528, loss_mean=0.847, loss_mean_cls=0.0805, proj_loss=-0.4][2026-03-26 12:41:50] Step: 2590, Training Logs: loss_final: 0.529808, loss_mean: 0.847866, proj_loss: -0.400541, loss_mean_cls: 0.082483, grad_norm: 0.382471 +Steps: 0%| | 2591/1000000 [10:47<68:34:06, 4.04it/s, grad_norm=0.382, loss_final=0.53, loss_mean=0.848, loss_mean_cls=0.0825, proj_loss=-0.401][2026-03-26 12:41:50] Step: 2591, Training Logs: loss_final: 0.541720, loss_mean: 0.858341, proj_loss: -0.398028, loss_mean_cls: 0.081407, grad_norm: 0.572524 +Steps: 0%| | 2592/1000000 [10:47<68:38:00, 4.04it/s, grad_norm=0.573, loss_final=0.542, loss_mean=0.858, loss_mean_cls=0.0814, proj_loss=-0.398][2026-03-26 12:41:50] Step: 2592, Training Logs: loss_final: 0.536659, loss_mean: 0.844980, proj_loss: -0.391999, loss_mean_cls: 0.083677, grad_norm: 0.625622 +Steps: 0%| | 2593/1000000 [10:47<68:49:34, 4.03it/s, grad_norm=0.626, loss_final=0.537, loss_mean=0.845, loss_mean_cls=0.0837, proj_loss=-0.392][2026-03-26 12:41:50] Step: 2593, Training Logs: loss_final: 0.543620, loss_mean: 0.864990, proj_loss: -0.401815, loss_mean_cls: 0.080444, grad_norm: 0.606814 +Steps: 0%| | 2594/1000000 [10:47<68:48:48, 4.03it/s, grad_norm=0.607, loss_final=0.544, loss_mean=0.865, loss_mean_cls=0.0804, proj_loss=-0.402][2026-03-26 12:41:51] Step: 2594, Training Logs: loss_final: 0.534803, loss_mean: 0.840696, proj_loss: -0.388904, loss_mean_cls: 0.083011, grad_norm: 0.373391 +Steps: 0%| | 2595/1000000 [10:48<68:46:29, 4.03it/s, grad_norm=0.373, loss_final=0.535, loss_mean=0.841, loss_mean_cls=0.083, proj_loss=-0.389][2026-03-26 12:41:51] Step: 2595, Training Logs: loss_final: 0.541452, loss_mean: 0.855616, proj_loss: -0.396203, loss_mean_cls: 0.082039, grad_norm: 0.823154 +Steps: 0%| | 2596/1000000 [10:48<68:44:05, 4.03it/s, grad_norm=0.823, loss_final=0.541, loss_mean=0.856, loss_mean_cls=0.082, proj_loss=-0.396][2026-03-26 12:41:51] Step: 2596, Training Logs: loss_final: 0.541623, loss_mean: 0.847796, proj_loss: -0.389523, loss_mean_cls: 0.083350, grad_norm: 0.582302 +Steps: 0%| | 2597/1000000 [10:48<68:43:28, 4.03it/s, grad_norm=0.582, loss_final=0.542, loss_mean=0.848, loss_mean_cls=0.0834, proj_loss=-0.39][2026-03-26 12:41:51] Step: 2597, Training Logs: loss_final: 0.538037, loss_mean: 0.848608, proj_loss: -0.392847, loss_mean_cls: 0.082275, grad_norm: 0.401560 +Steps: 0%| | 2598/1000000 [10:48<68:41:51, 4.03it/s, grad_norm=0.402, loss_final=0.538, loss_mean=0.849, loss_mean_cls=0.0823, proj_loss=-0.393][2026-03-26 12:41:52] Step: 2598, Training Logs: loss_final: 0.531045, loss_mean: 0.847974, proj_loss: -0.397806, loss_mean_cls: 0.080877, grad_norm: 0.502066 +Steps: 0%| | 2599/1000000 [10:49<68:43:03, 4.03it/s, grad_norm=0.502, loss_final=0.531, loss_mean=0.848, loss_mean_cls=0.0809, proj_loss=-0.398][2026-03-26 12:41:52] Step: 2599, Training Logs: loss_final: 0.531264, loss_mean: 0.855946, proj_loss: -0.404366, loss_mean_cls: 0.079684, grad_norm: 0.685352 +Steps: 0%| | 2600/1000000 [10:49<68:43:33, 4.03it/s, grad_norm=0.685, loss_final=0.531, loss_mean=0.856, loss_mean_cls=0.0797, proj_loss=-0.404][2026-03-26 12:41:52] Step: 2600, Training Logs: loss_final: 0.541603, loss_mean: 0.854641, proj_loss: -0.394540, loss_mean_cls: 0.081501, grad_norm: 0.502642 +Steps: 0%| | 2601/1000000 [10:49<68:45:00, 4.03it/s, grad_norm=0.503, loss_final=0.542, loss_mean=0.855, loss_mean_cls=0.0815, proj_loss=-0.395][2026-03-26 12:41:52] Step: 2601, Training Logs: loss_final: 0.545412, loss_mean: 0.864557, proj_loss: -0.399805, loss_mean_cls: 0.080660, grad_norm: 0.513088 +Steps: 0%| | 2602/1000000 [10:49<68:44:47, 4.03it/s, grad_norm=0.513, loss_final=0.545, loss_mean=0.865, loss_mean_cls=0.0807, proj_loss=-0.4][2026-03-26 12:41:53] Step: 2602, Training Logs: loss_final: 0.551601, loss_mean: 0.862486, proj_loss: -0.393944, loss_mean_cls: 0.083059, grad_norm: 0.622144 +Steps: 0%| | 2603/1000000 [10:50<68:43:06, 4.03it/s, grad_norm=0.622, loss_final=0.552, loss_mean=0.862, loss_mean_cls=0.0831, proj_loss=-0.394][2026-03-26 12:41:53] Step: 2603, Training Logs: loss_final: 0.533526, loss_mean: 0.842118, proj_loss: -0.392903, loss_mean_cls: 0.084310, grad_norm: 0.641349 +Steps: 0%| | 2604/1000000 [10:50<68:42:22, 4.03it/s, grad_norm=0.641, loss_final=0.534, loss_mean=0.842, loss_mean_cls=0.0843, proj_loss=-0.393][2026-03-26 12:41:53] Step: 2604, Training Logs: loss_final: 0.540696, loss_mean: 0.856591, proj_loss: -0.396586, loss_mean_cls: 0.080690, grad_norm: 0.710582 +Steps: 0%| | 2605/1000000 [10:50<68:46:46, 4.03it/s, grad_norm=0.711, loss_final=0.541, loss_mean=0.857, loss_mean_cls=0.0807, proj_loss=-0.397][2026-03-26 12:41:53] Step: 2605, Training Logs: loss_final: 0.541445, loss_mean: 0.852315, proj_loss: -0.393472, loss_mean_cls: 0.082602, grad_norm: 0.388841 +Steps: 0%| | 2606/1000000 [10:50<68:42:56, 4.03it/s, grad_norm=0.389, loss_final=0.541, loss_mean=0.852, loss_mean_cls=0.0826, proj_loss=-0.393][2026-03-26 12:41:54] Step: 2606, Training Logs: loss_final: 0.519834, loss_mean: 0.831428, proj_loss: -0.395300, loss_mean_cls: 0.083706, grad_norm: 0.570586 +Steps: 0%| | 2607/1000000 [10:51<68:47:58, 4.03it/s, grad_norm=0.571, loss_final=0.52, loss_mean=0.831, loss_mean_cls=0.0837, proj_loss=-0.395][2026-03-26 12:41:54] Step: 2607, Training Logs: loss_final: 0.552046, loss_mean: 0.868067, proj_loss: -0.396243, loss_mean_cls: 0.080221, grad_norm: 0.887443 +Steps: 0%| | 2608/1000000 [10:51<68:40:23, 4.03it/s, grad_norm=0.887, loss_final=0.552, loss_mean=0.868, loss_mean_cls=0.0802, proj_loss=-0.396][2026-03-26 12:41:54] Step: 2608, Training Logs: loss_final: 0.521468, loss_mean: 0.825670, proj_loss: -0.388294, loss_mean_cls: 0.084092, grad_norm: 0.401012 +Steps: 0%| | 2609/1000000 [10:51<68:39:05, 4.04it/s, grad_norm=0.401, loss_final=0.521, loss_mean=0.826, loss_mean_cls=0.0841, proj_loss=-0.388][2026-03-26 12:41:54] Step: 2609, Training Logs: loss_final: 0.535227, loss_mean: 0.849984, proj_loss: -0.397236, loss_mean_cls: 0.082479, grad_norm: 0.394572 +Steps: 0%| | 2610/1000000 [10:51<68:38:38, 4.04it/s, grad_norm=0.395, loss_final=0.535, loss_mean=0.85, loss_mean_cls=0.0825, proj_loss=-0.397][2026-03-26 12:41:55] Step: 2610, Training Logs: loss_final: 0.525485, loss_mean: 0.832052, proj_loss: -0.390484, loss_mean_cls: 0.083916, grad_norm: 0.473531 +Steps: 0%| | 2611/1000000 [10:52<68:39:17, 4.04it/s, grad_norm=0.474, loss_final=0.525, loss_mean=0.832, loss_mean_cls=0.0839, proj_loss=-0.39][2026-03-26 12:41:55] Step: 2611, Training Logs: loss_final: 0.537300, loss_mean: 0.857155, proj_loss: -0.399885, loss_mean_cls: 0.080029, grad_norm: 0.412723 +Steps: 0%| | 2612/1000000 [10:52<68:38:44, 4.04it/s, grad_norm=0.413, loss_final=0.537, loss_mean=0.857, loss_mean_cls=0.08, proj_loss=-0.4][2026-03-26 12:41:55] Step: 2612, Training Logs: loss_final: 0.546679, loss_mean: 0.870087, proj_loss: -0.402474, loss_mean_cls: 0.079065, grad_norm: 0.334177 +Steps: 0%| | 2613/1000000 [10:52<68:36:01, 4.04it/s, grad_norm=0.334, loss_final=0.547, loss_mean=0.87, loss_mean_cls=0.0791, proj_loss=-0.402][2026-03-26 12:41:55] Step: 2613, Training Logs: loss_final: 0.527500, loss_mean: 0.847707, proj_loss: -0.400754, loss_mean_cls: 0.080547, grad_norm: 0.561593 +Steps: 0%| | 2614/1000000 [10:52<68:36:38, 4.04it/s, grad_norm=0.562, loss_final=0.527, loss_mean=0.848, loss_mean_cls=0.0805, proj_loss=-0.401][2026-03-26 12:41:56] Step: 2614, Training Logs: loss_final: 0.534198, loss_mean: 0.845895, proj_loss: -0.393981, loss_mean_cls: 0.082283, grad_norm: 0.450003 +Steps: 0%| | 2615/1000000 [10:53<68:37:59, 4.04it/s, grad_norm=0.45, loss_final=0.534, loss_mean=0.846, loss_mean_cls=0.0823, proj_loss=-0.394][2026-03-26 12:41:56] Step: 2615, Training Logs: loss_final: 0.524285, loss_mean: 0.827267, proj_loss: -0.387925, loss_mean_cls: 0.084943, grad_norm: 0.353151 +Steps: 0%| | 2616/1000000 [10:53<68:39:20, 4.04it/s, grad_norm=0.353, loss_final=0.524, loss_mean=0.827, loss_mean_cls=0.0849, proj_loss=-0.388][2026-03-26 12:41:56] Step: 2616, Training Logs: loss_final: 0.536785, loss_mean: 0.848100, proj_loss: -0.393981, loss_mean_cls: 0.082666, grad_norm: 0.571436 +Steps: 0%| | 2617/1000000 [10:53<68:39:36, 4.04it/s, grad_norm=0.571, loss_final=0.537, loss_mean=0.848, loss_mean_cls=0.0827, proj_loss=-0.394][2026-03-26 12:41:56] Step: 2617, Training Logs: loss_final: 0.535740, loss_mean: 0.850967, proj_loss: -0.397761, loss_mean_cls: 0.082534, grad_norm: 0.687966 +Steps: 0%| | 2618/1000000 [10:53<68:41:48, 4.03it/s, grad_norm=0.688, loss_final=0.536, loss_mean=0.851, loss_mean_cls=0.0825, proj_loss=-0.398][2026-03-26 12:41:57] Step: 2618, Training Logs: loss_final: 0.510888, loss_mean: 0.819665, proj_loss: -0.392972, loss_mean_cls: 0.084195, grad_norm: 0.567877 +Steps: 0%| | 2619/1000000 [10:54<68:41:53, 4.03it/s, grad_norm=0.568, loss_final=0.511, loss_mean=0.82, loss_mean_cls=0.0842, proj_loss=-0.393][2026-03-26 12:41:57] Step: 2619, Training Logs: loss_final: 0.544431, loss_mean: 0.857309, proj_loss: -0.394292, loss_mean_cls: 0.081415, grad_norm: 0.406584 +Steps: 0%| | 2620/1000000 [10:54<68:43:41, 4.03it/s, grad_norm=0.407, loss_final=0.544, loss_mean=0.857, loss_mean_cls=0.0814, proj_loss=-0.394][2026-03-26 12:41:57] Step: 2620, Training Logs: loss_final: 0.527537, loss_mean: 0.828495, proj_loss: -0.385542, loss_mean_cls: 0.084584, grad_norm: 0.619971 +Steps: 0%| | 2621/1000000 [10:54<68:41:04, 4.03it/s, grad_norm=0.62, loss_final=0.528, loss_mean=0.828, loss_mean_cls=0.0846, proj_loss=-0.386][2026-03-26 12:41:57] Step: 2621, Training Logs: loss_final: 0.534828, loss_mean: 0.854122, proj_loss: -0.400371, loss_mean_cls: 0.081078, grad_norm: 0.539521 +Steps: 0%| | 2622/1000000 [10:54<68:42:00, 4.03it/s, grad_norm=0.54, loss_final=0.535, loss_mean=0.854, loss_mean_cls=0.0811, proj_loss=-0.4][2026-03-26 12:41:58] Step: 2622, Training Logs: loss_final: 0.516223, loss_mean: 0.829065, proj_loss: -0.394679, loss_mean_cls: 0.081836, grad_norm: 0.482167 +Steps: 0%| | 2623/1000000 [10:55<68:43:02, 4.03it/s, grad_norm=0.482, loss_final=0.516, loss_mean=0.829, loss_mean_cls=0.0818, proj_loss=-0.395][2026-03-26 12:41:58] Step: 2623, Training Logs: loss_final: 0.524356, loss_mean: 0.825629, proj_loss: -0.386113, loss_mean_cls: 0.084840, grad_norm: 0.722894 +Steps: 0%| | 2624/1000000 [10:55<68:47:23, 4.03it/s, grad_norm=0.723, loss_final=0.524, loss_mean=0.826, loss_mean_cls=0.0848, proj_loss=-0.386][2026-03-26 12:41:58] Step: 2624, Training Logs: loss_final: 0.538210, loss_mean: 0.863760, proj_loss: -0.404924, loss_mean_cls: 0.079374, grad_norm: 0.527010 +Steps: 0%| | 2625/1000000 [10:55<68:42:15, 4.03it/s, grad_norm=0.527, loss_final=0.538, loss_mean=0.864, loss_mean_cls=0.0794, proj_loss=-0.405][2026-03-26 12:41:58] Step: 2625, Training Logs: loss_final: 0.519899, loss_mean: 0.832613, proj_loss: -0.394955, loss_mean_cls: 0.082240, grad_norm: 0.394813 +Steps: 0%| | 2626/1000000 [10:55<68:41:11, 4.03it/s, grad_norm=0.395, loss_final=0.52, loss_mean=0.833, loss_mean_cls=0.0822, proj_loss=-0.395][2026-03-26 12:41:59] Step: 2626, Training Logs: loss_final: 0.531302, loss_mean: 0.839734, proj_loss: -0.390320, loss_mean_cls: 0.081889, grad_norm: 0.705108 +Steps: 0%| | 2627/1000000 [10:56<68:42:47, 4.03it/s, grad_norm=0.705, loss_final=0.531, loss_mean=0.84, loss_mean_cls=0.0819, proj_loss=-0.39][2026-03-26 12:41:59] Step: 2627, Training Logs: loss_final: 0.533206, loss_mean: 0.849011, proj_loss: -0.397528, loss_mean_cls: 0.081723, grad_norm: 0.356879 +Steps: 0%| | 2628/1000000 [10:56<68:43:53, 4.03it/s, grad_norm=0.357, loss_final=0.533, loss_mean=0.849, loss_mean_cls=0.0817, proj_loss=-0.398][2026-03-26 12:41:59] Step: 2628, Training Logs: loss_final: 0.539502, loss_mean: 0.846689, proj_loss: -0.389503, loss_mean_cls: 0.082316, grad_norm: 0.681615 +Steps: 0%| | 2629/1000000 [10:56<68:43:19, 4.03it/s, grad_norm=0.682, loss_final=0.54, loss_mean=0.847, loss_mean_cls=0.0823, proj_loss=-0.39][2026-03-26 12:41:59] Step: 2629, Training Logs: loss_final: 0.536918, loss_mean: 0.853538, proj_loss: -0.398323, loss_mean_cls: 0.081703, grad_norm: 0.796747 +Steps: 0%| | 2630/1000000 [10:56<68:43:28, 4.03it/s, grad_norm=0.797, loss_final=0.537, loss_mean=0.854, loss_mean_cls=0.0817, proj_loss=-0.398][2026-03-26 12:42:00] Step: 2630, Training Logs: loss_final: 0.509613, loss_mean: 0.813662, proj_loss: -0.388290, loss_mean_cls: 0.084240, grad_norm: 0.458025 +Steps: 0%| | 2631/1000000 [10:57<68:41:17, 4.03it/s, grad_norm=0.458, loss_final=0.51, loss_mean=0.814, loss_mean_cls=0.0842, proj_loss=-0.388][2026-03-26 12:42:00] Step: 2631, Training Logs: loss_final: 0.508498, loss_mean: 0.822909, proj_loss: -0.397894, loss_mean_cls: 0.083483, grad_norm: 0.738269 +Steps: 0%| | 2632/1000000 [10:57<68:40:55, 4.03it/s, grad_norm=0.738, loss_final=0.508, loss_mean=0.823, loss_mean_cls=0.0835, proj_loss=-0.398][2026-03-26 12:42:00] Step: 2632, Training Logs: loss_final: 0.534912, loss_mean: 0.849465, proj_loss: -0.395705, loss_mean_cls: 0.081152, grad_norm: 0.429709 +Steps: 0%| | 2633/1000000 [10:57<68:42:15, 4.03it/s, grad_norm=0.43, loss_final=0.535, loss_mean=0.849, loss_mean_cls=0.0812, proj_loss=-0.396][2026-03-26 12:42:00] Step: 2633, Training Logs: loss_final: 0.534471, loss_mean: 0.846456, proj_loss: -0.393482, loss_mean_cls: 0.081497, grad_norm: 0.500522 +Steps: 0%| | 2634/1000000 [10:57<68:43:35, 4.03it/s, grad_norm=0.501, loss_final=0.534, loss_mean=0.846, loss_mean_cls=0.0815, proj_loss=-0.393][2026-03-26 12:42:01] Step: 2634, Training Logs: loss_final: 0.519518, loss_mean: 0.834585, proj_loss: -0.397660, loss_mean_cls: 0.082593, grad_norm: 0.451370 +Steps: 0%| | 2635/1000000 [10:57<68:40:31, 4.03it/s, grad_norm=0.451, loss_final=0.52, loss_mean=0.835, loss_mean_cls=0.0826, proj_loss=-0.398][2026-03-26 12:42:01] Step: 2635, Training Logs: loss_final: 0.542749, loss_mean: 0.869671, proj_loss: -0.404055, loss_mean_cls: 0.077133, grad_norm: 0.463494 +Steps: 0%| | 2636/1000000 [10:58<68:40:12, 4.03it/s, grad_norm=0.463, loss_final=0.543, loss_mean=0.87, loss_mean_cls=0.0771, proj_loss=-0.404][2026-03-26 12:42:01] Step: 2636, Training Logs: loss_final: 0.529464, loss_mean: 0.847088, proj_loss: -0.398134, loss_mean_cls: 0.080510, grad_norm: 0.526885 +Steps: 0%| | 2637/1000000 [10:58<68:40:26, 4.03it/s, grad_norm=0.527, loss_final=0.529, loss_mean=0.847, loss_mean_cls=0.0805, proj_loss=-0.398][2026-03-26 12:42:01] Step: 2637, Training Logs: loss_final: 0.529603, loss_mean: 0.845699, proj_loss: -0.397647, loss_mean_cls: 0.081550, grad_norm: 0.365903 +Steps: 0%| | 2638/1000000 [10:58<69:15:02, 4.00it/s, grad_norm=0.366, loss_final=0.53, loss_mean=0.846, loss_mean_cls=0.0816, proj_loss=-0.398][2026-03-26 12:42:02] Step: 2638, Training Logs: loss_final: 0.531357, loss_mean: 0.841713, proj_loss: -0.392510, loss_mean_cls: 0.082154, grad_norm: 0.478854 +Steps: 0%| | 2639/1000000 [10:58<69:34:37, 3.98it/s, grad_norm=0.479, loss_final=0.531, loss_mean=0.842, loss_mean_cls=0.0822, proj_loss=-0.393][2026-03-26 12:42:02] Step: 2639, Training Logs: loss_final: 0.538271, loss_mean: 0.852235, proj_loss: -0.396547, loss_mean_cls: 0.082584, grad_norm: 0.446622 +Steps: 0%| | 2640/1000000 [10:59<69:19:20, 4.00it/s, grad_norm=0.447, loss_final=0.538, loss_mean=0.852, loss_mean_cls=0.0826, proj_loss=-0.397][2026-03-26 12:42:02] Step: 2640, Training Logs: loss_final: 0.529086, loss_mean: 0.834956, proj_loss: -0.389382, loss_mean_cls: 0.083512, grad_norm: 0.375912 +Steps: 0%| | 2641/1000000 [10:59<69:10:33, 4.00it/s, grad_norm=0.376, loss_final=0.529, loss_mean=0.835, loss_mean_cls=0.0835, proj_loss=-0.389][2026-03-26 12:42:02] Step: 2641, Training Logs: loss_final: 0.530505, loss_mean: 0.855092, proj_loss: -0.403674, loss_mean_cls: 0.079087, grad_norm: 0.686250 +Steps: 0%| | 2642/1000000 [10:59<68:56:19, 4.02it/s, grad_norm=0.686, loss_final=0.531, loss_mean=0.855, loss_mean_cls=0.0791, proj_loss=-0.404][2026-03-26 12:42:03] Step: 2642, Training Logs: loss_final: 0.540376, loss_mean: 0.848423, proj_loss: -0.389507, loss_mean_cls: 0.081460, grad_norm: 0.799225 +Steps: 0%| | 2643/1000000 [10:59<68:55:07, 4.02it/s, grad_norm=0.799, loss_final=0.54, loss_mean=0.848, loss_mean_cls=0.0815, proj_loss=-0.39][2026-03-26 12:42:03] Step: 2643, Training Logs: loss_final: 0.527260, loss_mean: 0.833091, proj_loss: -0.388712, loss_mean_cls: 0.082881, grad_norm: 0.364175 +Steps: 0%| | 2644/1000000 [11:00<68:45:42, 4.03it/s, grad_norm=0.364, loss_final=0.527, loss_mean=0.833, loss_mean_cls=0.0829, proj_loss=-0.389][2026-03-26 12:42:03] Step: 2644, Training Logs: loss_final: 0.549046, loss_mean: 0.873525, proj_loss: -0.404988, loss_mean_cls: 0.080509, grad_norm: 0.722374 +Steps: 0%| | 2645/1000000 [11:00<68:43:24, 4.03it/s, grad_norm=0.722, loss_final=0.549, loss_mean=0.874, loss_mean_cls=0.0805, proj_loss=-0.405][2026-03-26 12:42:03] Step: 2645, Training Logs: loss_final: 0.539750, loss_mean: 0.847389, proj_loss: -0.390148, loss_mean_cls: 0.082509, grad_norm: 0.804762 +Steps: 0%| | 2646/1000000 [11:00<68:41:47, 4.03it/s, grad_norm=0.805, loss_final=0.54, loss_mean=0.847, loss_mean_cls=0.0825, proj_loss=-0.39][2026-03-26 12:42:04] Step: 2646, Training Logs: loss_final: 0.522797, loss_mean: 0.830537, proj_loss: -0.390131, loss_mean_cls: 0.082391, grad_norm: 0.424703 +Steps: 0%| | 2647/1000000 [11:00<68:41:32, 4.03it/s, grad_norm=0.425, loss_final=0.523, loss_mean=0.831, loss_mean_cls=0.0824, proj_loss=-0.39][2026-03-26 12:42:04] Step: 2647, Training Logs: loss_final: 0.521937, loss_mean: 0.840246, proj_loss: -0.400037, loss_mean_cls: 0.081728, grad_norm: 0.753641 +Steps: 0%| | 2648/1000000 [11:01<68:41:34, 4.03it/s, grad_norm=0.754, loss_final=0.522, loss_mean=0.84, loss_mean_cls=0.0817, proj_loss=-0.4][2026-03-26 12:42:04] Step: 2648, Training Logs: loss_final: 0.518689, loss_mean: 0.831880, proj_loss: -0.397026, loss_mean_cls: 0.083835, grad_norm: 0.473480 +Steps: 0%| | 2649/1000000 [11:01<68:37:58, 4.04it/s, grad_norm=0.473, loss_final=0.519, loss_mean=0.832, loss_mean_cls=0.0838, proj_loss=-0.397][2026-03-26 12:42:04] Step: 2649, Training Logs: loss_final: 0.537771, loss_mean: 0.842501, proj_loss: -0.387948, loss_mean_cls: 0.083218, grad_norm: 0.572615 +Steps: 0%| | 2650/1000000 [11:01<68:41:08, 4.03it/s, grad_norm=0.573, loss_final=0.538, loss_mean=0.843, loss_mean_cls=0.0832, proj_loss=-0.388][2026-03-26 12:42:05] Step: 2650, Training Logs: loss_final: 0.525317, loss_mean: 0.844532, proj_loss: -0.400255, loss_mean_cls: 0.081040, grad_norm: 0.735019 +Steps: 0%| | 2651/1000000 [11:01<68:39:56, 4.03it/s, grad_norm=0.735, loss_final=0.525, loss_mean=0.845, loss_mean_cls=0.081, proj_loss=-0.4][2026-03-26 12:42:05] Step: 2651, Training Logs: loss_final: 0.526618, loss_mean: 0.839885, proj_loss: -0.395791, loss_mean_cls: 0.082524, grad_norm: 0.486953 +Steps: 0%| | 2652/1000000 [11:02<68:40:03, 4.03it/s, grad_norm=0.487, loss_final=0.527, loss_mean=0.84, loss_mean_cls=0.0825, proj_loss=-0.396][2026-03-26 12:42:05] Step: 2652, Training Logs: loss_final: 0.541633, loss_mean: 0.860936, proj_loss: -0.400092, loss_mean_cls: 0.080789, grad_norm: 0.608036 +Steps: 0%| | 2653/1000000 [11:02<68:42:41, 4.03it/s, grad_norm=0.608, loss_final=0.542, loss_mean=0.861, loss_mean_cls=0.0808, proj_loss=-0.4][2026-03-26 12:42:05] Step: 2653, Training Logs: loss_final: 0.524337, loss_mean: 0.843571, proj_loss: -0.400318, loss_mean_cls: 0.081084, grad_norm: 0.612939 +Steps: 0%| | 2654/1000000 [11:02<68:39:00, 4.04it/s, grad_norm=0.613, loss_final=0.524, loss_mean=0.844, loss_mean_cls=0.0811, proj_loss=-0.4][2026-03-26 12:42:06] Step: 2654, Training Logs: loss_final: 0.538490, loss_mean: 0.847968, proj_loss: -0.392219, loss_mean_cls: 0.082741, grad_norm: 0.478763 +Steps: 0%| | 2655/1000000 [11:02<68:39:48, 4.03it/s, grad_norm=0.479, loss_final=0.538, loss_mean=0.848, loss_mean_cls=0.0827, proj_loss=-0.392][2026-03-26 12:42:06] Step: 2655, Training Logs: loss_final: 0.521211, loss_mean: 0.828993, proj_loss: -0.391240, loss_mean_cls: 0.083458, grad_norm: 0.857876 +Steps: 0%| | 2656/1000000 [11:03<68:44:33, 4.03it/s, grad_norm=0.858, loss_final=0.521, loss_mean=0.829, loss_mean_cls=0.0835, proj_loss=-0.391][2026-03-26 12:42:06] Step: 2656, Training Logs: loss_final: 0.538026, loss_mean: 0.844205, proj_loss: -0.388787, loss_mean_cls: 0.082609, grad_norm: 0.455187 +Steps: 0%| | 2657/1000000 [11:03<69:51:49, 3.97it/s, grad_norm=0.455, loss_final=0.538, loss_mean=0.844, loss_mean_cls=0.0826, proj_loss=-0.389][2026-03-26 12:42:06] Step: 2657, Training Logs: loss_final: 0.556642, loss_mean: 0.871463, proj_loss: -0.395667, loss_mean_cls: 0.080845, grad_norm: 0.677059 +Steps: 0%| | 2658/1000000 [11:03<69:30:54, 3.99it/s, grad_norm=0.677, loss_final=0.557, loss_mean=0.871, loss_mean_cls=0.0808, proj_loss=-0.396][2026-03-26 12:42:07] Step: 2658, Training Logs: loss_final: 0.525760, loss_mean: 0.840935, proj_loss: -0.398161, loss_mean_cls: 0.082985, grad_norm: 0.942033 +Steps: 0%| | 2659/1000000 [11:03<69:14:21, 4.00it/s, grad_norm=0.942, loss_final=0.526, loss_mean=0.841, loss_mean_cls=0.083, proj_loss=-0.398][2026-03-26 12:42:07] Step: 2659, Training Logs: loss_final: 0.540413, loss_mean: 0.863194, proj_loss: -0.402234, loss_mean_cls: 0.079453, grad_norm: 0.331380 +Steps: 0%| | 2660/1000000 [11:04<69:03:38, 4.01it/s, grad_norm=0.331, loss_final=0.54, loss_mean=0.863, loss_mean_cls=0.0795, proj_loss=-0.402][2026-03-26 12:42:07] Step: 2660, Training Logs: loss_final: 0.531337, loss_mean: 0.842583, proj_loss: -0.393537, loss_mean_cls: 0.082291, grad_norm: 0.813183 +Steps: 0%| | 2661/1000000 [11:04<68:56:51, 4.02it/s, grad_norm=0.813, loss_final=0.531, loss_mean=0.843, loss_mean_cls=0.0823, proj_loss=-0.394][2026-03-26 12:42:07] Step: 2661, Training Logs: loss_final: 0.531458, loss_mean: 0.841286, proj_loss: -0.393756, loss_mean_cls: 0.083928, grad_norm: 0.535826 +Steps: 0%| | 2662/1000000 [11:04<68:57:59, 4.02it/s, grad_norm=0.536, loss_final=0.531, loss_mean=0.841, loss_mean_cls=0.0839, proj_loss=-0.394][2026-03-26 12:42:08] Step: 2662, Training Logs: loss_final: 0.524633, loss_mean: 0.829034, proj_loss: -0.387944, loss_mean_cls: 0.083544, grad_norm: 0.748711 +Steps: 0%| | 2663/1000000 [11:04<68:49:55, 4.02it/s, grad_norm=0.749, loss_final=0.525, loss_mean=0.829, loss_mean_cls=0.0835, proj_loss=-0.388][2026-03-26 12:42:08] Step: 2663, Training Logs: loss_final: 0.544434, loss_mean: 0.862283, proj_loss: -0.399683, loss_mean_cls: 0.081834, grad_norm: 1.006295 +Steps: 0%| | 2664/1000000 [11:05<68:50:44, 4.02it/s, grad_norm=1.01, loss_final=0.544, loss_mean=0.862, loss_mean_cls=0.0818, proj_loss=-0.4][2026-03-26 12:42:08] Step: 2664, Training Logs: loss_final: 0.530182, loss_mean: 0.841187, proj_loss: -0.394173, loss_mean_cls: 0.083168, grad_norm: 0.368109 +Steps: 0%| | 2665/1000000 [11:05<68:43:36, 4.03it/s, grad_norm=0.368, loss_final=0.53, loss_mean=0.841, loss_mean_cls=0.0832, proj_loss=-0.394][2026-03-26 12:42:08] Step: 2665, Training Logs: loss_final: 0.531565, loss_mean: 0.838440, proj_loss: -0.391143, loss_mean_cls: 0.084267, grad_norm: 0.592148 +Steps: 0%| | 2666/1000000 [11:05<68:43:53, 4.03it/s, grad_norm=0.592, loss_final=0.532, loss_mean=0.838, loss_mean_cls=0.0843, proj_loss=-0.391][2026-03-26 12:42:09] Step: 2666, Training Logs: loss_final: 0.520935, loss_mean: 0.835246, proj_loss: -0.397159, loss_mean_cls: 0.082848, grad_norm: 0.538479 +Steps: 0%| | 2667/1000000 [11:05<68:43:38, 4.03it/s, grad_norm=0.538, loss_final=0.521, loss_mean=0.835, loss_mean_cls=0.0828, proj_loss=-0.397][2026-03-26 12:42:09] Step: 2667, Training Logs: loss_final: 0.535069, loss_mean: 0.858515, proj_loss: -0.401775, loss_mean_cls: 0.078329, grad_norm: 0.343568 +Steps: 0%| | 2668/1000000 [11:06<68:40:50, 4.03it/s, grad_norm=0.344, loss_final=0.535, loss_mean=0.859, loss_mean_cls=0.0783, proj_loss=-0.402][2026-03-26 12:42:09] Step: 2668, Training Logs: loss_final: 0.533656, loss_mean: 0.856175, proj_loss: -0.403818, loss_mean_cls: 0.081299, grad_norm: 0.529546 +Steps: 0%| | 2669/1000000 [11:06<68:40:40, 4.03it/s, grad_norm=0.53, loss_final=0.534, loss_mean=0.856, loss_mean_cls=0.0813, proj_loss=-0.404][2026-03-26 12:42:09] Step: 2669, Training Logs: loss_final: 0.516755, loss_mean: 0.829753, proj_loss: -0.395980, loss_mean_cls: 0.082982, grad_norm: 0.330502 +Steps: 0%| | 2670/1000000 [11:06<68:37:53, 4.04it/s, grad_norm=0.331, loss_final=0.517, loss_mean=0.83, loss_mean_cls=0.083, proj_loss=-0.396][2026-03-26 12:42:10] Step: 2670, Training Logs: loss_final: 0.542564, loss_mean: 0.852970, proj_loss: -0.392742, loss_mean_cls: 0.082336, grad_norm: 0.492822 +Steps: 0%| | 2671/1000000 [11:06<68:38:17, 4.04it/s, grad_norm=0.493, loss_final=0.543, loss_mean=0.853, loss_mean_cls=0.0823, proj_loss=-0.393][2026-03-26 12:42:10] Step: 2671, Training Logs: loss_final: 0.526063, loss_mean: 0.828617, proj_loss: -0.386824, loss_mean_cls: 0.084270, grad_norm: 0.441151 +Steps: 0%| | 2672/1000000 [11:07<68:37:34, 4.04it/s, grad_norm=0.441, loss_final=0.526, loss_mean=0.829, loss_mean_cls=0.0843, proj_loss=-0.387][2026-03-26 12:42:10] Step: 2672, Training Logs: loss_final: 0.534060, loss_mean: 0.852762, proj_loss: -0.399594, loss_mean_cls: 0.080893, grad_norm: 0.391765 +Steps: 0%| | 2673/1000000 [11:07<68:39:04, 4.04it/s, grad_norm=0.392, loss_final=0.534, loss_mean=0.853, loss_mean_cls=0.0809, proj_loss=-0.4][2026-03-26 12:42:10] Step: 2673, Training Logs: loss_final: 0.528548, loss_mean: 0.841495, proj_loss: -0.396202, loss_mean_cls: 0.083255, grad_norm: 0.434387 +Steps: 0%| | 2674/1000000 [11:07<68:36:09, 4.04it/s, grad_norm=0.434, loss_final=0.529, loss_mean=0.841, loss_mean_cls=0.0833, proj_loss=-0.396][2026-03-26 12:42:11] Step: 2674, Training Logs: loss_final: 0.526370, loss_mean: 0.852036, proj_loss: -0.405364, loss_mean_cls: 0.079698, grad_norm: 0.327531 +Steps: 0%| | 2675/1000000 [11:07<68:36:51, 4.04it/s, grad_norm=0.328, loss_final=0.526, loss_mean=0.852, loss_mean_cls=0.0797, proj_loss=-0.405][2026-03-26 12:42:11] Step: 2675, Training Logs: loss_final: 0.546434, loss_mean: 0.864437, proj_loss: -0.398646, loss_mean_cls: 0.080643, grad_norm: 0.452478 +Steps: 0%| | 2676/1000000 [11:08<68:36:34, 4.04it/s, grad_norm=0.452, loss_final=0.546, loss_mean=0.864, loss_mean_cls=0.0806, proj_loss=-0.399][2026-03-26 12:42:11] Step: 2676, Training Logs: loss_final: 0.536410, loss_mean: 0.839735, proj_loss: -0.387768, loss_mean_cls: 0.084443, grad_norm: 0.339759 +Steps: 0%| | 2677/1000000 [11:08<68:36:45, 4.04it/s, grad_norm=0.34, loss_final=0.536, loss_mean=0.84, loss_mean_cls=0.0844, proj_loss=-0.388][2026-03-26 12:42:11] Step: 2677, Training Logs: loss_final: 0.531687, loss_mean: 0.839209, proj_loss: -0.389627, loss_mean_cls: 0.082104, grad_norm: 0.393274 +Steps: 0%| | 2678/1000000 [11:08<68:41:09, 4.03it/s, grad_norm=0.393, loss_final=0.532, loss_mean=0.839, loss_mean_cls=0.0821, proj_loss=-0.39][2026-03-26 12:42:12] Step: 2678, Training Logs: loss_final: 0.524137, loss_mean: 0.839395, proj_loss: -0.396524, loss_mean_cls: 0.081266, grad_norm: 0.364570 +Steps: 0%| | 2679/1000000 [11:08<68:43:34, 4.03it/s, grad_norm=0.365, loss_final=0.524, loss_mean=0.839, loss_mean_cls=0.0813, proj_loss=-0.397][2026-03-26 12:42:12] Step: 2679, Training Logs: loss_final: 0.534986, loss_mean: 0.852311, proj_loss: -0.398276, loss_mean_cls: 0.080951, grad_norm: 0.658778 +Steps: 0%| | 2680/1000000 [11:09<68:41:37, 4.03it/s, grad_norm=0.659, loss_final=0.535, loss_mean=0.852, loss_mean_cls=0.081, proj_loss=-0.398][2026-03-26 12:42:12] Step: 2680, Training Logs: loss_final: 0.537070, loss_mean: 0.850288, proj_loss: -0.395396, loss_mean_cls: 0.082178, grad_norm: 0.596665 +Steps: 0%| | 2681/1000000 [11:09<68:39:53, 4.03it/s, grad_norm=0.597, loss_final=0.537, loss_mean=0.85, loss_mean_cls=0.0822, proj_loss=-0.395][2026-03-26 12:42:12] Step: 2681, Training Logs: loss_final: 0.527427, loss_mean: 0.840244, proj_loss: -0.395542, loss_mean_cls: 0.082725, grad_norm: 0.341907 +Steps: 0%| | 2682/1000000 [11:09<68:38:36, 4.04it/s, grad_norm=0.342, loss_final=0.527, loss_mean=0.84, loss_mean_cls=0.0827, proj_loss=-0.396][2026-03-26 12:42:13] Step: 2682, Training Logs: loss_final: 0.526401, loss_mean: 0.841003, proj_loss: -0.396187, loss_mean_cls: 0.081586, grad_norm: 0.473142 +Steps: 0%| | 2683/1000000 [11:09<68:36:06, 4.04it/s, grad_norm=0.473, loss_final=0.526, loss_mean=0.841, loss_mean_cls=0.0816, proj_loss=-0.396][2026-03-26 12:42:13] Step: 2683, Training Logs: loss_final: 0.525380, loss_mean: 0.838393, proj_loss: -0.395418, loss_mean_cls: 0.082406, grad_norm: 0.357192 +Steps: 0%| | 2684/1000000 [11:10<68:39:04, 4.04it/s, grad_norm=0.357, loss_final=0.525, loss_mean=0.838, loss_mean_cls=0.0824, proj_loss=-0.395][2026-03-26 12:42:13] Step: 2684, Training Logs: loss_final: 0.533953, loss_mean: 0.857072, proj_loss: -0.403537, loss_mean_cls: 0.080418, grad_norm: 0.315993 +Steps: 0%| | 2685/1000000 [11:10<68:35:28, 4.04it/s, grad_norm=0.316, loss_final=0.534, loss_mean=0.857, loss_mean_cls=0.0804, proj_loss=-0.404][2026-03-26 12:42:13] Step: 2685, Training Logs: loss_final: 0.536992, loss_mean: 0.864886, proj_loss: -0.407061, loss_mean_cls: 0.079167, grad_norm: 0.434975 +Steps: 0%| | 2686/1000000 [11:10<68:34:38, 4.04it/s, grad_norm=0.435, loss_final=0.537, loss_mean=0.865, loss_mean_cls=0.0792, proj_loss=-0.407][2026-03-26 12:42:14] Step: 2686, Training Logs: loss_final: 0.514028, loss_mean: 0.821871, proj_loss: -0.392270, loss_mean_cls: 0.084427, grad_norm: 0.385993 +Steps: 0%| | 2687/1000000 [11:10<68:36:13, 4.04it/s, grad_norm=0.386, loss_final=0.514, loss_mean=0.822, loss_mean_cls=0.0844, proj_loss=-0.392][2026-03-26 12:42:14] Step: 2687, Training Logs: loss_final: 0.548963, loss_mean: 0.869268, proj_loss: -0.400134, loss_mean_cls: 0.079829, grad_norm: 0.539705 +Steps: 0%| | 2688/1000000 [11:11<68:38:14, 4.04it/s, grad_norm=0.54, loss_final=0.549, loss_mean=0.869, loss_mean_cls=0.0798, proj_loss=-0.4][2026-03-26 12:42:14] Step: 2688, Training Logs: loss_final: 0.531601, loss_mean: 0.852159, proj_loss: -0.401093, loss_mean_cls: 0.080535, grad_norm: 0.472516 +Steps: 0%| | 2689/1000000 [11:11<68:36:48, 4.04it/s, grad_norm=0.473, loss_final=0.532, loss_mean=0.852, loss_mean_cls=0.0805, proj_loss=-0.401][2026-03-26 12:42:14] Step: 2689, Training Logs: loss_final: 0.523670, loss_mean: 0.836410, proj_loss: -0.396198, loss_mean_cls: 0.083458, grad_norm: 0.382997 +Steps: 0%| | 2690/1000000 [11:11<68:37:35, 4.04it/s, grad_norm=0.383, loss_final=0.524, loss_mean=0.836, loss_mean_cls=0.0835, proj_loss=-0.396][2026-03-26 12:42:15] Step: 2690, Training Logs: loss_final: 0.520644, loss_mean: 0.829385, proj_loss: -0.394093, loss_mean_cls: 0.085352, grad_norm: 0.476484 +Steps: 0%| | 2691/1000000 [11:11<68:35:47, 4.04it/s, grad_norm=0.476, loss_final=0.521, loss_mean=0.829, loss_mean_cls=0.0854, proj_loss=-0.394][2026-03-26 12:42:15] Step: 2691, Training Logs: loss_final: 0.530223, loss_mean: 0.840146, proj_loss: -0.393739, loss_mean_cls: 0.083815, grad_norm: 0.335030 +Steps: 0%| | 2692/1000000 [11:12<68:37:36, 4.04it/s, grad_norm=0.335, loss_final=0.53, loss_mean=0.84, loss_mean_cls=0.0838, proj_loss=-0.394][2026-03-26 12:42:15] Step: 2692, Training Logs: loss_final: 0.523123, loss_mean: 0.833588, proj_loss: -0.393992, loss_mean_cls: 0.083527, grad_norm: 0.399950 +Steps: 0%| | 2693/1000000 [11:12<68:38:53, 4.04it/s, grad_norm=0.4, loss_final=0.523, loss_mean=0.834, loss_mean_cls=0.0835, proj_loss=-0.394][2026-03-26 12:42:15] Step: 2693, Training Logs: loss_final: 0.528428, loss_mean: 0.846050, proj_loss: -0.399478, loss_mean_cls: 0.081856, grad_norm: 0.341836 +Steps: 0%| | 2694/1000000 [11:12<68:40:27, 4.03it/s, grad_norm=0.342, loss_final=0.528, loss_mean=0.846, loss_mean_cls=0.0819, proj_loss=-0.399][2026-03-26 12:42:16] Step: 2694, Training Logs: loss_final: 0.533481, loss_mean: 0.850105, proj_loss: -0.397840, loss_mean_cls: 0.081217, grad_norm: 0.425069 +Steps: 0%| | 2695/1000000 [11:12<68:42:22, 4.03it/s, grad_norm=0.425, loss_final=0.533, loss_mean=0.85, loss_mean_cls=0.0812, proj_loss=-0.398][2026-03-26 12:42:16] Step: 2695, Training Logs: loss_final: 0.529790, loss_mean: 0.854318, proj_loss: -0.404480, loss_mean_cls: 0.079953, grad_norm: 0.439435 +Steps: 0%| | 2696/1000000 [11:13<68:40:16, 4.03it/s, grad_norm=0.439, loss_final=0.53, loss_mean=0.854, loss_mean_cls=0.08, proj_loss=-0.404][2026-03-26 12:42:16] Step: 2696, Training Logs: loss_final: 0.525601, loss_mean: 0.844804, proj_loss: -0.400249, loss_mean_cls: 0.081046, grad_norm: 0.381779 +Steps: 0%| | 2697/1000000 [11:13<68:36:28, 4.04it/s, grad_norm=0.382, loss_final=0.526, loss_mean=0.845, loss_mean_cls=0.081, proj_loss=-0.4][2026-03-26 12:42:16] Step: 2697, Training Logs: loss_final: 0.529183, loss_mean: 0.836427, proj_loss: -0.392519, loss_mean_cls: 0.085275, grad_norm: 0.705183 +Steps: 0%| | 2698/1000000 [11:13<68:35:56, 4.04it/s, grad_norm=0.705, loss_final=0.529, loss_mean=0.836, loss_mean_cls=0.0853, proj_loss=-0.393][2026-03-26 12:42:17] Step: 2698, Training Logs: loss_final: 0.526330, loss_mean: 0.839599, proj_loss: -0.395388, loss_mean_cls: 0.082119, grad_norm: 0.364676 +Steps: 0%| | 2699/1000000 [11:13<68:35:58, 4.04it/s, grad_norm=0.365, loss_final=0.526, loss_mean=0.84, loss_mean_cls=0.0821, proj_loss=-0.395][2026-03-26 12:42:17] Step: 2699, Training Logs: loss_final: 0.527912, loss_mean: 0.834774, proj_loss: -0.389874, loss_mean_cls: 0.083012, grad_norm: 0.486871 +Steps: 0%| | 2700/1000000 [11:14<68:36:58, 4.04it/s, grad_norm=0.487, loss_final=0.528, loss_mean=0.835, loss_mean_cls=0.083, proj_loss=-0.39][2026-03-26 12:42:17] Step: 2700, Training Logs: loss_final: 0.532064, loss_mean: 0.859474, proj_loss: -0.406916, loss_mean_cls: 0.079507, grad_norm: 0.499088 +Steps: 0%| | 2701/1000000 [11:14<68:37:38, 4.04it/s, grad_norm=0.499, loss_final=0.532, loss_mean=0.859, loss_mean_cls=0.0795, proj_loss=-0.407][2026-03-26 12:42:17] Step: 2701, Training Logs: loss_final: 0.534863, loss_mean: 0.853778, proj_loss: -0.397564, loss_mean_cls: 0.078649, grad_norm: 0.334536 +Steps: 0%| | 2702/1000000 [11:14<68:38:46, 4.04it/s, grad_norm=0.335, loss_final=0.535, loss_mean=0.854, loss_mean_cls=0.0786, proj_loss=-0.398][2026-03-26 12:42:18] Step: 2702, Training Logs: loss_final: 0.537794, loss_mean: 0.844886, proj_loss: -0.389581, loss_mean_cls: 0.082489, grad_norm: 0.360217 +Steps: 0%| | 2703/1000000 [11:14<68:37:21, 4.04it/s, grad_norm=0.36, loss_final=0.538, loss_mean=0.845, loss_mean_cls=0.0825, proj_loss=-0.39][2026-03-26 12:42:18] Step: 2703, Training Logs: loss_final: 0.547111, loss_mean: 0.860701, proj_loss: -0.395666, loss_mean_cls: 0.082077, grad_norm: 0.343755 +Steps: 0%| | 2704/1000000 [11:15<68:35:34, 4.04it/s, grad_norm=0.344, loss_final=0.547, loss_mean=0.861, loss_mean_cls=0.0821, proj_loss=-0.396][2026-03-26 12:42:18] Step: 2704, Training Logs: loss_final: 0.529033, loss_mean: 0.843061, proj_loss: -0.398236, loss_mean_cls: 0.084208, grad_norm: 0.427139 +Steps: 0%| | 2705/1000000 [11:15<68:34:22, 4.04it/s, grad_norm=0.427, loss_final=0.529, loss_mean=0.843, loss_mean_cls=0.0842, proj_loss=-0.398][2026-03-26 12:42:18] Step: 2705, Training Logs: loss_final: 0.523335, loss_mean: 0.853169, proj_loss: -0.408895, loss_mean_cls: 0.079061, grad_norm: 0.316805 +Steps: 0%| | 2706/1000000 [11:15<68:34:51, 4.04it/s, grad_norm=0.317, loss_final=0.523, loss_mean=0.853, loss_mean_cls=0.0791, proj_loss=-0.409][2026-03-26 12:42:19] Step: 2706, Training Logs: loss_final: 0.532451, loss_mean: 0.852936, proj_loss: -0.401902, loss_mean_cls: 0.081416, grad_norm: 0.388626 +Steps: 0%| | 2707/1000000 [11:15<68:34:55, 4.04it/s, grad_norm=0.389, loss_final=0.532, loss_mean=0.853, loss_mean_cls=0.0814, proj_loss=-0.402][2026-03-26 12:42:19] Step: 2707, Training Logs: loss_final: 0.528196, loss_mean: 0.843667, proj_loss: -0.397869, loss_mean_cls: 0.082398, grad_norm: 0.326260 +Steps: 0%| | 2708/1000000 [11:16<68:35:29, 4.04it/s, grad_norm=0.326, loss_final=0.528, loss_mean=0.844, loss_mean_cls=0.0824, proj_loss=-0.398][2026-03-26 12:42:19] Step: 2708, Training Logs: loss_final: 0.537629, loss_mean: 0.861557, proj_loss: -0.404242, loss_mean_cls: 0.080314, grad_norm: 0.318805 +Steps: 0%| | 2709/1000000 [11:16<68:38:27, 4.04it/s, grad_norm=0.319, loss_final=0.538, loss_mean=0.862, loss_mean_cls=0.0803, proj_loss=-0.404][2026-03-26 12:42:19] Step: 2709, Training Logs: loss_final: 0.530995, loss_mean: 0.843052, proj_loss: -0.394127, loss_mean_cls: 0.082070, grad_norm: 0.403207 +Steps: 0%| | 2710/1000000 [11:16<68:44:37, 4.03it/s, grad_norm=0.403, loss_final=0.531, loss_mean=0.843, loss_mean_cls=0.0821, proj_loss=-0.394][2026-03-26 12:42:19] Step: 2710, Training Logs: loss_final: 0.522797, loss_mean: 0.841753, proj_loss: -0.400957, loss_mean_cls: 0.082001, grad_norm: 0.378469 +Steps: 0%| | 2711/1000000 [11:16<68:42:53, 4.03it/s, grad_norm=0.378, loss_final=0.523, loss_mean=0.842, loss_mean_cls=0.082, proj_loss=-0.401][2026-03-26 12:42:20] Step: 2711, Training Logs: loss_final: 0.544811, loss_mean: 0.864036, proj_loss: -0.400469, loss_mean_cls: 0.081244, grad_norm: 0.345203 +Steps: 0%| | 2712/1000000 [11:17<68:40:55, 4.03it/s, grad_norm=0.345, loss_final=0.545, loss_mean=0.864, loss_mean_cls=0.0812, proj_loss=-0.4][2026-03-26 12:42:20] Step: 2712, Training Logs: loss_final: 0.526170, loss_mean: 0.836390, proj_loss: -0.394336, loss_mean_cls: 0.084116, grad_norm: 0.319150 +Steps: 0%| | 2713/1000000 [11:17<68:40:20, 4.03it/s, grad_norm=0.319, loss_final=0.526, loss_mean=0.836, loss_mean_cls=0.0841, proj_loss=-0.394][2026-03-26 12:42:20] Step: 2713, Training Logs: loss_final: 0.520635, loss_mean: 0.841686, proj_loss: -0.400850, loss_mean_cls: 0.079799, grad_norm: 0.314414 +Steps: 0%| | 2714/1000000 [11:17<68:38:28, 4.04it/s, grad_norm=0.314, loss_final=0.521, loss_mean=0.842, loss_mean_cls=0.0798, proj_loss=-0.401][2026-03-26 12:42:20] Step: 2714, Training Logs: loss_final: 0.526102, loss_mean: 0.854679, proj_loss: -0.406844, loss_mean_cls: 0.078267, grad_norm: 0.448849 +Steps: 0%| | 2715/1000000 [11:17<68:38:06, 4.04it/s, grad_norm=0.449, loss_final=0.526, loss_mean=0.855, loss_mean_cls=0.0783, proj_loss=-0.407][2026-03-26 12:42:21] Step: 2715, Training Logs: loss_final: 0.526525, loss_mean: 0.840060, proj_loss: -0.395754, loss_mean_cls: 0.082219, grad_norm: 0.368186 +Steps: 0%| | 2716/1000000 [11:18<68:35:27, 4.04it/s, grad_norm=0.368, loss_final=0.527, loss_mean=0.84, loss_mean_cls=0.0822, proj_loss=-0.396][2026-03-26 12:42:21] Step: 2716, Training Logs: loss_final: 0.531225, loss_mean: 0.860559, proj_loss: -0.407031, loss_mean_cls: 0.077698, grad_norm: 0.402494 +Steps: 0%| | 2717/1000000 [11:18<68:35:08, 4.04it/s, grad_norm=0.402, loss_final=0.531, loss_mean=0.861, loss_mean_cls=0.0777, proj_loss=-0.407][2026-03-26 12:42:21] Step: 2717, Training Logs: loss_final: 0.525076, loss_mean: 0.837122, proj_loss: -0.395341, loss_mean_cls: 0.083294, grad_norm: 0.640375 +Steps: 0%| | 2718/1000000 [11:18<68:32:55, 4.04it/s, grad_norm=0.64, loss_final=0.525, loss_mean=0.837, loss_mean_cls=0.0833, proj_loss=-0.395][2026-03-26 12:42:21] Step: 2718, Training Logs: loss_final: 0.517354, loss_mean: 0.836293, proj_loss: -0.401097, loss_mean_cls: 0.082159, grad_norm: 0.331555 +Steps: 0%| | 2719/1000000 [11:18<68:36:01, 4.04it/s, grad_norm=0.332, loss_final=0.517, loss_mean=0.836, loss_mean_cls=0.0822, proj_loss=-0.401][2026-03-26 12:42:22] Step: 2719, Training Logs: loss_final: 0.537267, loss_mean: 0.853462, proj_loss: -0.397431, loss_mean_cls: 0.081236, grad_norm: 0.610321 +Steps: 0%| | 2720/1000000 [11:19<68:35:25, 4.04it/s, grad_norm=0.61, loss_final=0.537, loss_mean=0.853, loss_mean_cls=0.0812, proj_loss=-0.397][2026-03-26 12:42:22] Step: 2720, Training Logs: loss_final: 0.537456, loss_mean: 0.862482, proj_loss: -0.404067, loss_mean_cls: 0.079042, grad_norm: 0.539956 +Steps: 0%| | 2721/1000000 [11:19<68:35:29, 4.04it/s, grad_norm=0.54, loss_final=0.537, loss_mean=0.862, loss_mean_cls=0.079, proj_loss=-0.404][2026-03-26 12:42:22] Step: 2721, Training Logs: loss_final: 0.540737, loss_mean: 0.858338, proj_loss: -0.399536, loss_mean_cls: 0.081935, grad_norm: 0.463418 +Steps: 0%| | 2722/1000000 [11:19<68:34:13, 4.04it/s, grad_norm=0.463, loss_final=0.541, loss_mean=0.858, loss_mean_cls=0.0819, proj_loss=-0.4][2026-03-26 12:42:22] Step: 2722, Training Logs: loss_final: 0.526930, loss_mean: 0.840884, proj_loss: -0.395945, loss_mean_cls: 0.081992, grad_norm: 0.514718 +Steps: 0%| | 2723/1000000 [11:19<68:35:56, 4.04it/s, grad_norm=0.515, loss_final=0.527, loss_mean=0.841, loss_mean_cls=0.082, proj_loss=-0.396][2026-03-26 12:42:23] Step: 2723, Training Logs: loss_final: 0.524841, loss_mean: 0.835617, proj_loss: -0.393255, loss_mean_cls: 0.082479, grad_norm: 0.642873 +Steps: 0%| | 2724/1000000 [11:20<68:35:05, 4.04it/s, grad_norm=0.643, loss_final=0.525, loss_mean=0.836, loss_mean_cls=0.0825, proj_loss=-0.393][2026-03-26 12:42:23] Step: 2724, Training Logs: loss_final: 0.521581, loss_mean: 0.839101, proj_loss: -0.398863, loss_mean_cls: 0.081343, grad_norm: 0.442086 +Steps: 0%| | 2725/1000000 [11:20<68:35:28, 4.04it/s, grad_norm=0.442, loss_final=0.522, loss_mean=0.839, loss_mean_cls=0.0813, proj_loss=-0.399][2026-03-26 12:42:23] Step: 2725, Training Logs: loss_final: 0.530047, loss_mean: 0.851527, proj_loss: -0.400825, loss_mean_cls: 0.079345, grad_norm: 0.398297 +Steps: 0%| | 2726/1000000 [11:20<68:34:45, 4.04it/s, grad_norm=0.398, loss_final=0.53, loss_mean=0.852, loss_mean_cls=0.0793, proj_loss=-0.401][2026-03-26 12:42:23] Step: 2726, Training Logs: loss_final: 0.545983, loss_mean: 0.862810, proj_loss: -0.396299, loss_mean_cls: 0.079472, grad_norm: 0.555412 +Steps: 0%| | 2727/1000000 [11:20<68:39:07, 4.04it/s, grad_norm=0.555, loss_final=0.546, loss_mean=0.863, loss_mean_cls=0.0795, proj_loss=-0.396][2026-03-26 12:42:24] Step: 2727, Training Logs: loss_final: 0.523579, loss_mean: 0.820787, proj_loss: -0.382847, loss_mean_cls: 0.085639, grad_norm: 0.449196 +Steps: 0%| | 2728/1000000 [11:21<68:40:44, 4.03it/s, grad_norm=0.449, loss_final=0.524, loss_mean=0.821, loss_mean_cls=0.0856, proj_loss=-0.383][2026-03-26 12:42:24] Step: 2728, Training Logs: loss_final: 0.526055, loss_mean: 0.842152, proj_loss: -0.398215, loss_mean_cls: 0.082118, grad_norm: 0.434967 +Steps: 0%| | 2729/1000000 [11:21<68:40:14, 4.03it/s, grad_norm=0.435, loss_final=0.526, loss_mean=0.842, loss_mean_cls=0.0821, proj_loss=-0.398][2026-03-26 12:42:24] Step: 2729, Training Logs: loss_final: 0.524552, loss_mean: 0.849575, proj_loss: -0.404527, loss_mean_cls: 0.079504, grad_norm: 0.585142 +Steps: 0%| | 2730/1000000 [11:21<68:41:16, 4.03it/s, grad_norm=0.585, loss_final=0.525, loss_mean=0.85, loss_mean_cls=0.0795, proj_loss=-0.405][2026-03-26 12:42:24] Step: 2730, Training Logs: loss_final: 0.541449, loss_mean: 0.862693, proj_loss: -0.401307, loss_mean_cls: 0.080063, grad_norm: 0.420103 +Steps: 0%| | 2731/1000000 [11:21<69:14:28, 4.00it/s, grad_norm=0.42, loss_final=0.541, loss_mean=0.863, loss_mean_cls=0.0801, proj_loss=-0.401][2026-03-26 12:42:25] Step: 2731, Training Logs: loss_final: 0.534063, loss_mean: 0.858881, proj_loss: -0.405119, loss_mean_cls: 0.080301, grad_norm: 0.555073 +Steps: 0%| | 2732/1000000 [11:22<69:01:43, 4.01it/s, grad_norm=0.555, loss_final=0.534, loss_mean=0.859, loss_mean_cls=0.0803, proj_loss=-0.405][2026-03-26 12:42:25] Step: 2732, Training Logs: loss_final: 0.527661, loss_mean: 0.849277, proj_loss: -0.403359, loss_mean_cls: 0.081743, grad_norm: 0.474701 +Steps: 0%| | 2733/1000000 [11:22<68:55:46, 4.02it/s, grad_norm=0.475, loss_final=0.528, loss_mean=0.849, loss_mean_cls=0.0817, proj_loss=-0.403][2026-03-26 12:42:25] Step: 2733, Training Logs: loss_final: 0.527410, loss_mean: 0.837358, proj_loss: -0.393007, loss_mean_cls: 0.083059, grad_norm: 0.452860 +Steps: 0%| | 2734/1000000 [11:22<68:49:31, 4.02it/s, grad_norm=0.453, loss_final=0.527, loss_mean=0.837, loss_mean_cls=0.0831, proj_loss=-0.393][2026-03-26 12:42:25] Step: 2734, Training Logs: loss_final: 0.525676, loss_mean: 0.836822, proj_loss: -0.393006, loss_mean_cls: 0.081860, grad_norm: 0.538782 +Steps: 0%| | 2735/1000000 [11:22<68:44:42, 4.03it/s, grad_norm=0.539, loss_final=0.526, loss_mean=0.837, loss_mean_cls=0.0819, proj_loss=-0.393][2026-03-26 12:42:26] Step: 2735, Training Logs: loss_final: 0.523466, loss_mean: 0.834571, proj_loss: -0.393518, loss_mean_cls: 0.082413, grad_norm: 0.390497 +Steps: 0%| | 2736/1000000 [11:23<68:40:49, 4.03it/s, grad_norm=0.39, loss_final=0.523, loss_mean=0.835, loss_mean_cls=0.0824, proj_loss=-0.394][2026-03-26 12:42:26] Step: 2736, Training Logs: loss_final: 0.548901, loss_mean: 0.874941, proj_loss: -0.405176, loss_mean_cls: 0.079136, grad_norm: 0.541149 +Steps: 0%| | 2737/1000000 [11:23<68:35:55, 4.04it/s, grad_norm=0.541, loss_final=0.549, loss_mean=0.875, loss_mean_cls=0.0791, proj_loss=-0.405][2026-03-26 12:42:26] Step: 2737, Training Logs: loss_final: 0.519589, loss_mean: 0.826012, proj_loss: -0.391475, loss_mean_cls: 0.085052, grad_norm: 0.471069 +Steps: 0%| | 2738/1000000 [11:23<68:38:08, 4.04it/s, grad_norm=0.471, loss_final=0.52, loss_mean=0.826, loss_mean_cls=0.0851, proj_loss=-0.391][2026-03-26 12:42:26] Step: 2738, Training Logs: loss_final: 0.537418, loss_mean: 0.857033, proj_loss: -0.401224, loss_mean_cls: 0.081610, grad_norm: 0.517812 +Steps: 0%| | 2739/1000000 [11:23<68:36:10, 4.04it/s, grad_norm=0.518, loss_final=0.537, loss_mean=0.857, loss_mean_cls=0.0816, proj_loss=-0.401][2026-03-26 12:42:27] Step: 2739, Training Logs: loss_final: 0.522370, loss_mean: 0.837977, proj_loss: -0.398152, loss_mean_cls: 0.082545, grad_norm: 0.638656 +Steps: 0%| | 2740/1000000 [11:24<68:38:02, 4.04it/s, grad_norm=0.639, loss_final=0.522, loss_mean=0.838, loss_mean_cls=0.0825, proj_loss=-0.398][2026-03-26 12:42:27] Step: 2740, Training Logs: loss_final: 0.521308, loss_mean: 0.830042, proj_loss: -0.393314, loss_mean_cls: 0.084580, grad_norm: 0.418741 +Steps: 0%| | 2741/1000000 [11:24<68:35:34, 4.04it/s, grad_norm=0.419, loss_final=0.521, loss_mean=0.83, loss_mean_cls=0.0846, proj_loss=-0.393][2026-03-26 12:42:27] Step: 2741, Training Logs: loss_final: 0.524721, loss_mean: 0.837826, proj_loss: -0.396274, loss_mean_cls: 0.083168, grad_norm: 0.512185 +Steps: 0%| | 2742/1000000 [11:24<68:38:06, 4.04it/s, grad_norm=0.512, loss_final=0.525, loss_mean=0.838, loss_mean_cls=0.0832, proj_loss=-0.396][2026-03-26 12:42:27] Step: 2742, Training Logs: loss_final: 0.526222, loss_mean: 0.835757, proj_loss: -0.391594, loss_mean_cls: 0.082060, grad_norm: 0.556875 +Steps: 0%| | 2743/1000000 [11:24<68:37:12, 4.04it/s, grad_norm=0.557, loss_final=0.526, loss_mean=0.836, loss_mean_cls=0.0821, proj_loss=-0.392][2026-03-26 12:42:28] Step: 2743, Training Logs: loss_final: 0.520276, loss_mean: 0.832469, proj_loss: -0.395114, loss_mean_cls: 0.082920, grad_norm: 0.526202 +Steps: 0%| | 2744/1000000 [11:25<68:39:49, 4.03it/s, grad_norm=0.526, loss_final=0.52, loss_mean=0.832, loss_mean_cls=0.0829, proj_loss=-0.395][2026-03-26 12:42:28] Step: 2744, Training Logs: loss_final: 0.538953, loss_mean: 0.857124, proj_loss: -0.398765, loss_mean_cls: 0.080594, grad_norm: 0.699973 +Steps: 0%| | 2745/1000000 [11:25<68:39:16, 4.03it/s, grad_norm=0.7, loss_final=0.539, loss_mean=0.857, loss_mean_cls=0.0806, proj_loss=-0.399][2026-03-26 12:42:28] Step: 2745, Training Logs: loss_final: 0.545489, loss_mean: 0.877559, proj_loss: -0.410240, loss_mean_cls: 0.078170, grad_norm: 0.636780 +Steps: 0%| | 2746/1000000 [11:25<68:42:26, 4.03it/s, grad_norm=0.637, loss_final=0.545, loss_mean=0.878, loss_mean_cls=0.0782, proj_loss=-0.41][2026-03-26 12:42:28] Step: 2746, Training Logs: loss_final: 0.528451, loss_mean: 0.840212, proj_loss: -0.392665, loss_mean_cls: 0.080904, grad_norm: 0.434377 +Steps: 0%| | 2747/1000000 [11:25<68:41:06, 4.03it/s, grad_norm=0.434, loss_final=0.528, loss_mean=0.84, loss_mean_cls=0.0809, proj_loss=-0.393][2026-03-26 12:42:29] Step: 2747, Training Logs: loss_final: 0.529068, loss_mean: 0.844425, proj_loss: -0.397283, loss_mean_cls: 0.081926, grad_norm: 0.483990 +Steps: 0%| | 2748/1000000 [11:26<68:39:18, 4.03it/s, grad_norm=0.484, loss_final=0.529, loss_mean=0.844, loss_mean_cls=0.0819, proj_loss=-0.397][2026-03-26 12:42:29] Step: 2748, Training Logs: loss_final: 0.520737, loss_mean: 0.839470, proj_loss: -0.400660, loss_mean_cls: 0.081926, grad_norm: 0.598773 +Steps: 0%| | 2749/1000000 [11:26<68:37:07, 4.04it/s, grad_norm=0.599, loss_final=0.521, loss_mean=0.839, loss_mean_cls=0.0819, proj_loss=-0.401][2026-03-26 12:42:29] Step: 2749, Training Logs: loss_final: 0.544720, loss_mean: 0.846041, proj_loss: -0.384150, loss_mean_cls: 0.082828, grad_norm: 0.436331 +Steps: 0%| | 2750/1000000 [11:26<68:36:54, 4.04it/s, grad_norm=0.436, loss_final=0.545, loss_mean=0.846, loss_mean_cls=0.0828, proj_loss=-0.384][2026-03-26 12:42:29] Step: 2750, Training Logs: loss_final: 0.531322, loss_mean: 0.857792, proj_loss: -0.405951, loss_mean_cls: 0.079481, grad_norm: 0.486506 +Steps: 0%| | 2751/1000000 [11:26<68:35:52, 4.04it/s, grad_norm=0.487, loss_final=0.531, loss_mean=0.858, loss_mean_cls=0.0795, proj_loss=-0.406][2026-03-26 12:42:30] Step: 2751, Training Logs: loss_final: 0.523939, loss_mean: 0.834209, proj_loss: -0.393534, loss_mean_cls: 0.083264, grad_norm: 0.524459 +Steps: 0%| | 2752/1000000 [11:27<68:38:26, 4.04it/s, grad_norm=0.524, loss_final=0.524, loss_mean=0.834, loss_mean_cls=0.0833, proj_loss=-0.394][2026-03-26 12:42:30] Step: 2752, Training Logs: loss_final: 0.516479, loss_mean: 0.842601, proj_loss: -0.407129, loss_mean_cls: 0.081008, grad_norm: 0.499243 +Steps: 0%| | 2753/1000000 [11:27<68:38:29, 4.04it/s, grad_norm=0.499, loss_final=0.516, loss_mean=0.843, loss_mean_cls=0.081, proj_loss=-0.407][2026-03-26 12:42:30] Step: 2753, Training Logs: loss_final: 0.526815, loss_mean: 0.852018, proj_loss: -0.404804, loss_mean_cls: 0.079601, grad_norm: 0.708748 +Steps: 0%| | 2754/1000000 [11:27<68:38:20, 4.04it/s, grad_norm=0.709, loss_final=0.527, loss_mean=0.852, loss_mean_cls=0.0796, proj_loss=-0.405][2026-03-26 12:42:30] Step: 2754, Training Logs: loss_final: 0.536053, loss_mean: 0.842165, proj_loss: -0.390056, loss_mean_cls: 0.083944, grad_norm: 0.330923 +Steps: 0%| | 2755/1000000 [11:27<68:39:50, 4.03it/s, grad_norm=0.331, loss_final=0.536, loss_mean=0.842, loss_mean_cls=0.0839, proj_loss=-0.39][2026-03-26 12:42:31] Step: 2755, Training Logs: loss_final: 0.514745, loss_mean: 0.829592, proj_loss: -0.396688, loss_mean_cls: 0.081841, grad_norm: 0.564813 +Steps: 0%| | 2756/1000000 [11:28<68:37:16, 4.04it/s, grad_norm=0.565, loss_final=0.515, loss_mean=0.83, loss_mean_cls=0.0818, proj_loss=-0.397][2026-03-26 12:42:31] Step: 2756, Training Logs: loss_final: 0.513880, loss_mean: 0.832389, proj_loss: -0.399497, loss_mean_cls: 0.080988, grad_norm: 0.657708 +Steps: 0%| | 2757/1000000 [11:28<68:36:56, 4.04it/s, grad_norm=0.658, loss_final=0.514, loss_mean=0.832, loss_mean_cls=0.081, proj_loss=-0.399][2026-03-26 12:42:31] Step: 2757, Training Logs: loss_final: 0.535096, loss_mean: 0.853720, proj_loss: -0.398226, loss_mean_cls: 0.079603, grad_norm: 0.384596 +Steps: 0%| | 2758/1000000 [11:28<68:34:30, 4.04it/s, grad_norm=0.385, loss_final=0.535, loss_mean=0.854, loss_mean_cls=0.0796, proj_loss=-0.398][2026-03-26 12:42:31] Step: 2758, Training Logs: loss_final: 0.542452, loss_mean: 0.867565, proj_loss: -0.404696, loss_mean_cls: 0.079583, grad_norm: 0.464208 +Steps: 0%| | 2759/1000000 [11:28<68:34:42, 4.04it/s, grad_norm=0.464, loss_final=0.542, loss_mean=0.868, loss_mean_cls=0.0796, proj_loss=-0.405][2026-03-26 12:42:32] Step: 2759, Training Logs: loss_final: 0.529300, loss_mean: 0.846166, proj_loss: -0.397293, loss_mean_cls: 0.080427, grad_norm: 0.397602 +Steps: 0%| | 2760/1000000 [11:28<68:34:58, 4.04it/s, grad_norm=0.398, loss_final=0.529, loss_mean=0.846, loss_mean_cls=0.0804, proj_loss=-0.397][2026-03-26 12:42:32] Step: 2760, Training Logs: loss_final: 0.532226, loss_mean: 0.840916, proj_loss: -0.391785, loss_mean_cls: 0.083095, grad_norm: 0.301224 +Steps: 0%| | 2761/1000000 [11:29<68:38:08, 4.04it/s, grad_norm=0.301, loss_final=0.532, loss_mean=0.841, loss_mean_cls=0.0831, proj_loss=-0.392][2026-03-26 12:42:32] Step: 2761, Training Logs: loss_final: 0.520533, loss_mean: 0.831288, proj_loss: -0.393383, loss_mean_cls: 0.082628, grad_norm: 0.346655 +Steps: 0%| | 2762/1000000 [11:29<68:36:43, 4.04it/s, grad_norm=0.347, loss_final=0.521, loss_mean=0.831, loss_mean_cls=0.0826, proj_loss=-0.393][2026-03-26 12:42:32] Step: 2762, Training Logs: loss_final: 0.541219, loss_mean: 0.848310, proj_loss: -0.390828, loss_mean_cls: 0.083737, grad_norm: 0.478581 +Steps: 0%| | 2763/1000000 [11:29<68:42:52, 4.03it/s, grad_norm=0.479, loss_final=0.541, loss_mean=0.848, loss_mean_cls=0.0837, proj_loss=-0.391][2026-03-26 12:42:33] Step: 2763, Training Logs: loss_final: 0.539069, loss_mean: 0.861883, proj_loss: -0.403962, loss_mean_cls: 0.081148, grad_norm: 0.367743 +Steps: 0%| | 2764/1000000 [11:29<68:39:03, 4.04it/s, grad_norm=0.368, loss_final=0.539, loss_mean=0.862, loss_mean_cls=0.0811, proj_loss=-0.404][2026-03-26 12:42:33] Step: 2764, Training Logs: loss_final: 0.525609, loss_mean: 0.843774, proj_loss: -0.400255, loss_mean_cls: 0.082090, grad_norm: 0.424179 +Steps: 0%| | 2765/1000000 [11:30<68:39:38, 4.03it/s, grad_norm=0.424, loss_final=0.526, loss_mean=0.844, loss_mean_cls=0.0821, proj_loss=-0.4][2026-03-26 12:42:33] Step: 2765, Training Logs: loss_final: 0.545582, loss_mean: 0.861201, proj_loss: -0.398411, loss_mean_cls: 0.082793, grad_norm: 0.393663 +Steps: 0%| | 2766/1000000 [11:30<68:38:40, 4.04it/s, grad_norm=0.394, loss_final=0.546, loss_mean=0.861, loss_mean_cls=0.0828, proj_loss=-0.398][2026-03-26 12:42:33] Step: 2766, Training Logs: loss_final: 0.511661, loss_mean: 0.830911, proj_loss: -0.400542, loss_mean_cls: 0.081292, grad_norm: 0.651159 +Steps: 0%| | 2767/1000000 [11:30<68:36:39, 4.04it/s, grad_norm=0.651, loss_final=0.512, loss_mean=0.831, loss_mean_cls=0.0813, proj_loss=-0.401][2026-03-26 12:42:34] Step: 2767, Training Logs: loss_final: 0.528416, loss_mean: 0.844689, proj_loss: -0.398671, loss_mean_cls: 0.082397, grad_norm: 0.489575 +Steps: 0%| | 2768/1000000 [11:30<68:35:47, 4.04it/s, grad_norm=0.49, loss_final=0.528, loss_mean=0.845, loss_mean_cls=0.0824, proj_loss=-0.399][2026-03-26 12:42:34] Step: 2768, Training Logs: loss_final: 0.535558, loss_mean: 0.849918, proj_loss: -0.395922, loss_mean_cls: 0.081562, grad_norm: 0.396431 +Steps: 0%| | 2769/1000000 [11:31<68:36:55, 4.04it/s, grad_norm=0.396, loss_final=0.536, loss_mean=0.85, loss_mean_cls=0.0816, proj_loss=-0.396][2026-03-26 12:42:34] Step: 2769, Training Logs: loss_final: 0.541556, loss_mean: 0.863181, proj_loss: -0.400639, loss_mean_cls: 0.079014, grad_norm: 0.694814 +Steps: 0%| | 2770/1000000 [11:31<68:37:35, 4.04it/s, grad_norm=0.695, loss_final=0.542, loss_mean=0.863, loss_mean_cls=0.079, proj_loss=-0.401][2026-03-26 12:42:34] Step: 2770, Training Logs: loss_final: 0.517055, loss_mean: 0.838834, proj_loss: -0.403144, loss_mean_cls: 0.081365, grad_norm: 0.578267 +Steps: 0%| | 2771/1000000 [11:31<68:34:42, 4.04it/s, grad_norm=0.578, loss_final=0.517, loss_mean=0.839, loss_mean_cls=0.0814, proj_loss=-0.403][2026-03-26 12:42:35] Step: 2771, Training Logs: loss_final: 0.520391, loss_mean: 0.824929, proj_loss: -0.390301, loss_mean_cls: 0.085763, grad_norm: 0.370044 +Steps: 0%| | 2772/1000000 [11:31<68:35:26, 4.04it/s, grad_norm=0.37, loss_final=0.52, loss_mean=0.825, loss_mean_cls=0.0858, proj_loss=-0.39][2026-03-26 12:42:35] Step: 2772, Training Logs: loss_final: 0.509742, loss_mean: 0.824748, proj_loss: -0.398088, loss_mean_cls: 0.083082, grad_norm: 0.363453 +Steps: 0%| | 2773/1000000 [11:32<68:36:02, 4.04it/s, grad_norm=0.363, loss_final=0.51, loss_mean=0.825, loss_mean_cls=0.0831, proj_loss=-0.398][2026-03-26 12:42:35] Step: 2773, Training Logs: loss_final: 0.525185, loss_mean: 0.848506, proj_loss: -0.403669, loss_mean_cls: 0.080348, grad_norm: 0.394249 +Steps: 0%| | 2774/1000000 [11:32<68:37:28, 4.04it/s, grad_norm=0.394, loss_final=0.525, loss_mean=0.849, loss_mean_cls=0.0803, proj_loss=-0.404][2026-03-26 12:42:35] Step: 2774, Training Logs: loss_final: 0.536218, loss_mean: 0.852486, proj_loss: -0.398120, loss_mean_cls: 0.081851, grad_norm: 0.444257 +Steps: 0%| | 2775/1000000 [11:32<68:35:53, 4.04it/s, grad_norm=0.444, loss_final=0.536, loss_mean=0.852, loss_mean_cls=0.0819, proj_loss=-0.398][2026-03-26 12:42:36] Step: 2775, Training Logs: loss_final: 0.538917, loss_mean: 0.865147, proj_loss: -0.406192, loss_mean_cls: 0.079962, grad_norm: 0.572958 +Steps: 0%| | 2776/1000000 [11:32<68:35:24, 4.04it/s, grad_norm=0.573, loss_final=0.539, loss_mean=0.865, loss_mean_cls=0.08, proj_loss=-0.406][2026-03-26 12:42:36] Step: 2776, Training Logs: loss_final: 0.543037, loss_mean: 0.860149, proj_loss: -0.397770, loss_mean_cls: 0.080657, grad_norm: 0.630824 +Steps: 0%| | 2777/1000000 [11:33<69:46:54, 3.97it/s, grad_norm=0.631, loss_final=0.543, loss_mean=0.86, loss_mean_cls=0.0807, proj_loss=-0.398][2026-03-26 12:42:36] Step: 2777, Training Logs: loss_final: 0.534660, loss_mean: 0.845882, proj_loss: -0.393089, loss_mean_cls: 0.081866, grad_norm: 0.303408 +Steps: 0%| | 2778/1000000 [11:33<69:24:31, 3.99it/s, grad_norm=0.303, loss_final=0.535, loss_mean=0.846, loss_mean_cls=0.0819, proj_loss=-0.393][2026-03-26 12:42:36] Step: 2778, Training Logs: loss_final: 0.517851, loss_mean: 0.829448, proj_loss: -0.394652, loss_mean_cls: 0.083055, grad_norm: 0.539603 +Steps: 0%| | 2779/1000000 [11:33<69:09:12, 4.01it/s, grad_norm=0.54, loss_final=0.518, loss_mean=0.829, loss_mean_cls=0.0831, proj_loss=-0.395][2026-03-26 12:42:37] Step: 2779, Training Logs: loss_final: 0.541133, loss_mean: 0.869454, proj_loss: -0.406786, loss_mean_cls: 0.078465, grad_norm: 0.327988 +Steps: 0%| | 2780/1000000 [11:33<69:00:53, 4.01it/s, grad_norm=0.328, loss_final=0.541, loss_mean=0.869, loss_mean_cls=0.0785, proj_loss=-0.407][2026-03-26 12:42:37] Step: 2780, Training Logs: loss_final: 0.520568, loss_mean: 0.837723, proj_loss: -0.398885, loss_mean_cls: 0.081730, grad_norm: 0.364921 +Steps: 0%| | 2781/1000000 [11:34<68:53:23, 4.02it/s, grad_norm=0.365, loss_final=0.521, loss_mean=0.838, loss_mean_cls=0.0817, proj_loss=-0.399][2026-03-26 12:42:37] Step: 2781, Training Logs: loss_final: 0.520295, loss_mean: 0.838413, proj_loss: -0.400443, loss_mean_cls: 0.082324, grad_norm: 0.460120 +Steps: 0%| | 2782/1000000 [11:34<68:48:52, 4.03it/s, grad_norm=0.46, loss_final=0.52, loss_mean=0.838, loss_mean_cls=0.0823, proj_loss=-0.4][2026-03-26 12:42:37] Step: 2782, Training Logs: loss_final: 0.531708, loss_mean: 0.848874, proj_loss: -0.398554, loss_mean_cls: 0.081388, grad_norm: 0.365466 +Steps: 0%| | 2783/1000000 [11:34<68:43:56, 4.03it/s, grad_norm=0.365, loss_final=0.532, loss_mean=0.849, loss_mean_cls=0.0814, proj_loss=-0.399][2026-03-26 12:42:38] Step: 2783, Training Logs: loss_final: 0.527651, loss_mean: 0.844341, proj_loss: -0.398820, loss_mean_cls: 0.082130, grad_norm: 0.518993 +Steps: 0%| | 2784/1000000 [11:34<68:41:01, 4.03it/s, grad_norm=0.519, loss_final=0.528, loss_mean=0.844, loss_mean_cls=0.0821, proj_loss=-0.399][2026-03-26 12:42:38] Step: 2784, Training Logs: loss_final: 0.535261, loss_mean: 0.853461, proj_loss: -0.399194, loss_mean_cls: 0.080993, grad_norm: 0.454107 +Steps: 0%| | 2785/1000000 [11:35<68:39:48, 4.03it/s, grad_norm=0.454, loss_final=0.535, loss_mean=0.853, loss_mean_cls=0.081, proj_loss=-0.399][2026-03-26 12:42:38] Step: 2785, Training Logs: loss_final: 0.513618, loss_mean: 0.834466, proj_loss: -0.401646, loss_mean_cls: 0.080799, grad_norm: 0.501514 +Steps: 0%| | 2786/1000000 [11:35<68:46:03, 4.03it/s, grad_norm=0.502, loss_final=0.514, loss_mean=0.834, loss_mean_cls=0.0808, proj_loss=-0.402][2026-03-26 12:42:38] Step: 2786, Training Logs: loss_final: 0.529653, loss_mean: 0.840916, proj_loss: -0.394196, loss_mean_cls: 0.082933, grad_norm: 0.862202 +Steps: 0%| | 2787/1000000 [11:35<71:27:34, 3.88it/s, grad_norm=0.862, loss_final=0.53, loss_mean=0.841, loss_mean_cls=0.0829, proj_loss=-0.394][2026-03-26 12:42:39] Step: 2787, Training Logs: loss_final: 0.514681, loss_mean: 0.822690, proj_loss: -0.391764, loss_mean_cls: 0.083754, grad_norm: 0.478589 +Steps: 0%| | 2788/1000000 [11:35<70:34:18, 3.93it/s, grad_norm=0.479, loss_final=0.515, loss_mean=0.823, loss_mean_cls=0.0838, proj_loss=-0.392][2026-03-26 12:42:39] Step: 2788, Training Logs: loss_final: 0.507442, loss_mean: 0.829000, proj_loss: -0.403105, loss_mean_cls: 0.081546, grad_norm: 0.454483 +Steps: 0%| | 2789/1000000 [11:36<69:58:20, 3.96it/s, grad_norm=0.454, loss_final=0.507, loss_mean=0.829, loss_mean_cls=0.0815, proj_loss=-0.403][2026-03-26 12:42:39] Step: 2789, Training Logs: loss_final: 0.508721, loss_mean: 0.826625, proj_loss: -0.401306, loss_mean_cls: 0.083402, grad_norm: 0.449911 +Steps: 0%| | 2790/1000000 [11:36<69:32:23, 3.98it/s, grad_norm=0.45, loss_final=0.509, loss_mean=0.827, loss_mean_cls=0.0834, proj_loss=-0.401][2026-03-26 12:42:39] Step: 2790, Training Logs: loss_final: 0.537448, loss_mean: 0.856722, proj_loss: -0.399976, loss_mean_cls: 0.080702, grad_norm: 0.280484 +Steps: 0%| | 2791/1000000 [11:36<69:16:03, 4.00it/s, grad_norm=0.28, loss_final=0.537, loss_mean=0.857, loss_mean_cls=0.0807, proj_loss=-0.4][2026-03-26 12:42:40] Step: 2791, Training Logs: loss_final: 0.527313, loss_mean: 0.840197, proj_loss: -0.395993, loss_mean_cls: 0.083109, grad_norm: 0.471754 +Steps: 0%| | 2792/1000000 [11:36<69:02:13, 4.01it/s, grad_norm=0.472, loss_final=0.527, loss_mean=0.84, loss_mean_cls=0.0831, proj_loss=-0.396][2026-03-26 12:42:40] Step: 2792, Training Logs: loss_final: 0.515051, loss_mean: 0.826086, proj_loss: -0.394980, loss_mean_cls: 0.083946, grad_norm: 0.426019 +Steps: 0%| | 2793/1000000 [11:37<68:53:14, 4.02it/s, grad_norm=0.426, loss_final=0.515, loss_mean=0.826, loss_mean_cls=0.0839, proj_loss=-0.395][2026-03-26 12:42:40] Step: 2793, Training Logs: loss_final: 0.538088, loss_mean: 0.854998, proj_loss: -0.398785, loss_mean_cls: 0.081875, grad_norm: 0.539010 +Steps: 0%| | 2794/1000000 [11:37<68:47:20, 4.03it/s, grad_norm=0.539, loss_final=0.538, loss_mean=0.855, loss_mean_cls=0.0819, proj_loss=-0.399][2026-03-26 12:42:40] Step: 2794, Training Logs: loss_final: 0.540750, loss_mean: 0.852129, proj_loss: -0.394828, loss_mean_cls: 0.083448, grad_norm: 0.680704 +Steps: 0%| | 2795/1000000 [11:37<68:45:40, 4.03it/s, grad_norm=0.681, loss_final=0.541, loss_mean=0.852, loss_mean_cls=0.0834, proj_loss=-0.395][2026-03-26 12:42:41] Step: 2795, Training Logs: loss_final: 0.529274, loss_mean: 0.840354, proj_loss: -0.394653, loss_mean_cls: 0.083574, grad_norm: 0.408751 +Steps: 0%| | 2796/1000000 [11:37<69:38:37, 3.98it/s, grad_norm=0.409, loss_final=0.529, loss_mean=0.84, loss_mean_cls=0.0836, proj_loss=-0.395][2026-03-26 12:42:41] Step: 2796, Training Logs: loss_final: 0.522781, loss_mean: 0.847307, proj_loss: -0.404069, loss_mean_cls: 0.079543, grad_norm: 0.490863 +Steps: 0%| | 2797/1000000 [11:38<69:18:10, 4.00it/s, grad_norm=0.491, loss_final=0.523, loss_mean=0.847, loss_mean_cls=0.0795, proj_loss=-0.404][2026-03-26 12:42:41] Step: 2797, Training Logs: loss_final: 0.531896, loss_mean: 0.839782, proj_loss: -0.391897, loss_mean_cls: 0.084011, grad_norm: 0.650112 +Steps: 0%| | 2798/1000000 [11:38<69:05:03, 4.01it/s, grad_norm=0.65, loss_final=0.532, loss_mean=0.84, loss_mean_cls=0.084, proj_loss=-0.392][2026-03-26 12:42:41] Step: 2798, Training Logs: loss_final: 0.519204, loss_mean: 0.847247, proj_loss: -0.406395, loss_mean_cls: 0.078353, grad_norm: 0.302926 +Steps: 0%| | 2799/1000000 [11:38<68:55:34, 4.02it/s, grad_norm=0.303, loss_final=0.519, loss_mean=0.847, loss_mean_cls=0.0784, proj_loss=-0.406][2026-03-26 12:42:42] Step: 2799, Training Logs: loss_final: 0.527319, loss_mean: 0.832655, proj_loss: -0.388658, loss_mean_cls: 0.083322, grad_norm: 0.556613 +Steps: 0%| | 2800/1000000 [11:38<68:48:24, 4.03it/s, grad_norm=0.557, loss_final=0.527, loss_mean=0.833, loss_mean_cls=0.0833, proj_loss=-0.389][2026-03-26 12:42:42] Step: 2800, Training Logs: loss_final: 0.533913, loss_mean: 0.844771, proj_loss: -0.393094, loss_mean_cls: 0.082237, grad_norm: 0.371172 +Steps: 0%| | 2801/1000000 [11:39<68:43:49, 4.03it/s, grad_norm=0.371, loss_final=0.534, loss_mean=0.845, loss_mean_cls=0.0822, proj_loss=-0.393][2026-03-26 12:42:42] Step: 2801, Training Logs: loss_final: 0.523476, loss_mean: 0.837448, proj_loss: -0.398404, loss_mean_cls: 0.084432, grad_norm: 0.685072 +Steps: 0%| | 2802/1000000 [11:39<70:09:19, 3.95it/s, grad_norm=0.685, loss_final=0.523, loss_mean=0.837, loss_mean_cls=0.0844, proj_loss=-0.398][2026-03-26 12:42:42] Step: 2802, Training Logs: loss_final: 0.529986, loss_mean: 0.852440, proj_loss: -0.404059, loss_mean_cls: 0.081605, grad_norm: 0.686490 +Steps: 0%| | 2803/1000000 [11:39<69:48:50, 3.97it/s, grad_norm=0.686, loss_final=0.53, loss_mean=0.852, loss_mean_cls=0.0816, proj_loss=-0.404][2026-03-26 12:42:43] Step: 2803, Training Logs: loss_final: 0.524534, loss_mean: 0.841128, proj_loss: -0.398878, loss_mean_cls: 0.082284, grad_norm: 0.344776 +Steps: 0%| | 2804/1000000 [11:39<69:26:14, 3.99it/s, grad_norm=0.345, loss_final=0.525, loss_mean=0.841, loss_mean_cls=0.0823, proj_loss=-0.399][2026-03-26 12:42:43] Step: 2804, Training Logs: loss_final: 0.544635, loss_mean: 0.866159, proj_loss: -0.400340, loss_mean_cls: 0.078816, grad_norm: 0.489415 +Steps: 0%| | 2805/1000000 [11:40<69:09:55, 4.00it/s, grad_norm=0.489, loss_final=0.545, loss_mean=0.866, loss_mean_cls=0.0788, proj_loss=-0.4][2026-03-26 12:42:43] Step: 2805, Training Logs: loss_final: 0.531305, loss_mean: 0.849952, proj_loss: -0.399641, loss_mean_cls: 0.080993, grad_norm: 0.516691 +Steps: 0%| | 2806/1000000 [11:40<68:59:56, 4.01it/s, grad_norm=0.517, loss_final=0.531, loss_mean=0.85, loss_mean_cls=0.081, proj_loss=-0.4][2026-03-26 12:42:43] Step: 2806, Training Logs: loss_final: 0.517339, loss_mean: 0.835059, proj_loss: -0.401277, loss_mean_cls: 0.083557, grad_norm: 0.333924 +Steps: 0%| | 2807/1000000 [11:40<68:54:43, 4.02it/s, grad_norm=0.334, loss_final=0.517, loss_mean=0.835, loss_mean_cls=0.0836, proj_loss=-0.401][2026-03-26 12:42:44] Step: 2807, Training Logs: loss_final: 0.540344, loss_mean: 0.869888, proj_loss: -0.407889, loss_mean_cls: 0.078344, grad_norm: 0.383309 +Steps: 0%| | 2808/1000000 [11:40<68:47:45, 4.03it/s, grad_norm=0.383, loss_final=0.54, loss_mean=0.87, loss_mean_cls=0.0783, proj_loss=-0.408][2026-03-26 12:42:44] Step: 2808, Training Logs: loss_final: 0.525928, loss_mean: 0.849602, proj_loss: -0.404279, loss_mean_cls: 0.080605, grad_norm: 0.414474 +Steps: 0%| | 2809/1000000 [11:41<68:42:33, 4.03it/s, grad_norm=0.414, loss_final=0.526, loss_mean=0.85, loss_mean_cls=0.0806, proj_loss=-0.404][2026-03-26 12:42:44] Step: 2809, Training Logs: loss_final: 0.515776, loss_mean: 0.820132, proj_loss: -0.389485, loss_mean_cls: 0.085129, grad_norm: 0.483155 +Steps: 0%| | 2810/1000000 [11:41<68:41:19, 4.03it/s, grad_norm=0.483, loss_final=0.516, loss_mean=0.82, loss_mean_cls=0.0851, proj_loss=-0.389][2026-03-26 12:42:44] Step: 2810, Training Logs: loss_final: 0.541381, loss_mean: 0.842896, proj_loss: -0.385410, loss_mean_cls: 0.083895, grad_norm: 0.439476 +Steps: 0%| | 2811/1000000 [11:41<68:39:34, 4.03it/s, grad_norm=0.439, loss_final=0.541, loss_mean=0.843, loss_mean_cls=0.0839, proj_loss=-0.385][2026-03-26 12:42:45] Step: 2811, Training Logs: loss_final: 0.538736, loss_mean: 0.851497, proj_loss: -0.394587, loss_mean_cls: 0.081827, grad_norm: 0.380264 +Steps: 0%| | 2812/1000000 [11:41<68:38:25, 4.04it/s, grad_norm=0.38, loss_final=0.539, loss_mean=0.851, loss_mean_cls=0.0818, proj_loss=-0.395][2026-03-26 12:42:45] Step: 2812, Training Logs: loss_final: 0.516734, loss_mean: 0.840134, proj_loss: -0.405375, loss_mean_cls: 0.081975, grad_norm: 0.403905 +Steps: 0%| | 2813/1000000 [11:42<68:32:35, 4.04it/s, grad_norm=0.404, loss_final=0.517, loss_mean=0.84, loss_mean_cls=0.082, proj_loss=-0.405][2026-03-26 12:42:45] Step: 2813, Training Logs: loss_final: 0.512730, loss_mean: 0.818443, proj_loss: -0.389688, loss_mean_cls: 0.083975, grad_norm: 0.395489 +Steps: 0%| | 2814/1000000 [11:42<68:36:11, 4.04it/s, grad_norm=0.395, loss_final=0.513, loss_mean=0.818, loss_mean_cls=0.084, proj_loss=-0.39][2026-03-26 12:42:45] Step: 2814, Training Logs: loss_final: 0.513695, loss_mean: 0.829831, proj_loss: -0.398479, loss_mean_cls: 0.082343, grad_norm: 0.406315 +Steps: 0%| | 2815/1000000 [11:42<68:33:27, 4.04it/s, grad_norm=0.406, loss_final=0.514, loss_mean=0.83, loss_mean_cls=0.0823, proj_loss=-0.398][2026-03-26 12:42:46] Step: 2815, Training Logs: loss_final: 0.529598, loss_mean: 0.860815, proj_loss: -0.409636, loss_mean_cls: 0.078419, grad_norm: 0.461381 +Steps: 0%| | 2816/1000000 [11:42<68:33:09, 4.04it/s, grad_norm=0.461, loss_final=0.53, loss_mean=0.861, loss_mean_cls=0.0784, proj_loss=-0.41][2026-03-26 12:42:46] Step: 2816, Training Logs: loss_final: 0.532346, loss_mean: 0.846957, proj_loss: -0.397188, loss_mean_cls: 0.082577, grad_norm: 0.445181 +Steps: 0%| | 2817/1000000 [11:43<68:29:51, 4.04it/s, grad_norm=0.445, loss_final=0.532, loss_mean=0.847, loss_mean_cls=0.0826, proj_loss=-0.397][2026-03-26 12:42:46] Step: 2817, Training Logs: loss_final: 0.522532, loss_mean: 0.847292, proj_loss: -0.405469, loss_mean_cls: 0.080709, grad_norm: 0.435953 +Steps: 0%| | 2818/1000000 [11:43<68:31:34, 4.04it/s, grad_norm=0.436, loss_final=0.523, loss_mean=0.847, loss_mean_cls=0.0807, proj_loss=-0.405][2026-03-26 12:42:46] Step: 2818, Training Logs: loss_final: 0.528809, loss_mean: 0.851770, proj_loss: -0.403773, loss_mean_cls: 0.080811, grad_norm: 0.508567 +Steps: 0%| | 2819/1000000 [11:43<68:30:54, 4.04it/s, grad_norm=0.509, loss_final=0.529, loss_mean=0.852, loss_mean_cls=0.0808, proj_loss=-0.404][2026-03-26 12:42:47] Step: 2819, Training Logs: loss_final: 0.533036, loss_mean: 0.851539, proj_loss: -0.401754, loss_mean_cls: 0.083252, grad_norm: 0.290906 +Steps: 0%| | 2820/1000000 [11:43<68:30:34, 4.04it/s, grad_norm=0.291, loss_final=0.533, loss_mean=0.852, loss_mean_cls=0.0833, proj_loss=-0.402][2026-03-26 12:42:47] Step: 2820, Training Logs: loss_final: 0.523630, loss_mean: 0.830724, proj_loss: -0.392663, loss_mean_cls: 0.085569, grad_norm: 0.464530 +Steps: 0%| | 2821/1000000 [11:44<68:30:57, 4.04it/s, grad_norm=0.465, loss_final=0.524, loss_mean=0.831, loss_mean_cls=0.0856, proj_loss=-0.393][2026-03-26 12:42:47] Step: 2821, Training Logs: loss_final: 0.519426, loss_mean: 0.839738, proj_loss: -0.400548, loss_mean_cls: 0.080236, grad_norm: 0.474498 +Steps: 0%| | 2822/1000000 [11:44<68:33:57, 4.04it/s, grad_norm=0.474, loss_final=0.519, loss_mean=0.84, loss_mean_cls=0.0802, proj_loss=-0.401][2026-03-26 12:42:47] Step: 2822, Training Logs: loss_final: 0.526352, loss_mean: 0.842279, proj_loss: -0.397227, loss_mean_cls: 0.081300, grad_norm: 0.343492 +Steps: 0%| | 2823/1000000 [11:44<68:31:24, 4.04it/s, grad_norm=0.343, loss_final=0.526, loss_mean=0.842, loss_mean_cls=0.0813, proj_loss=-0.397][2026-03-26 12:42:48] Step: 2823, Training Logs: loss_final: 0.537195, loss_mean: 0.846410, proj_loss: -0.392411, loss_mean_cls: 0.083197, grad_norm: 0.450623 +Steps: 0%| | 2824/1000000 [11:44<68:30:11, 4.04it/s, grad_norm=0.451, loss_final=0.537, loss_mean=0.846, loss_mean_cls=0.0832, proj_loss=-0.392][2026-03-26 12:42:48] Step: 2824, Training Logs: loss_final: 0.512294, loss_mean: 0.819011, proj_loss: -0.390939, loss_mean_cls: 0.084222, grad_norm: 0.460685 +Steps: 0%| | 2825/1000000 [11:45<68:28:38, 4.05it/s, grad_norm=0.461, loss_final=0.512, loss_mean=0.819, loss_mean_cls=0.0842, proj_loss=-0.391][2026-03-26 12:42:48] Step: 2825, Training Logs: loss_final: 0.528361, loss_mean: 0.845501, proj_loss: -0.398972, loss_mean_cls: 0.081832, grad_norm: 0.358023 +Steps: 0%| | 2826/1000000 [11:45<68:29:36, 4.04it/s, grad_norm=0.358, loss_final=0.528, loss_mean=0.846, loss_mean_cls=0.0818, proj_loss=-0.399][2026-03-26 12:42:48] Step: 2826, Training Logs: loss_final: 0.535006, loss_mean: 0.845216, proj_loss: -0.392870, loss_mean_cls: 0.082660, grad_norm: 0.494882 +Steps: 0%| | 2827/1000000 [11:45<68:32:23, 4.04it/s, grad_norm=0.495, loss_final=0.535, loss_mean=0.845, loss_mean_cls=0.0827, proj_loss=-0.393][2026-03-26 12:42:49] Step: 2827, Training Logs: loss_final: 0.521092, loss_mean: 0.837367, proj_loss: -0.398351, loss_mean_cls: 0.082076, grad_norm: 0.332697 +Steps: 0%| | 2828/1000000 [11:45<68:33:19, 4.04it/s, grad_norm=0.333, loss_final=0.521, loss_mean=0.837, loss_mean_cls=0.0821, proj_loss=-0.398][2026-03-26 12:42:49] Step: 2828, Training Logs: loss_final: 0.513542, loss_mean: 0.819799, proj_loss: -0.391252, loss_mean_cls: 0.084995, grad_norm: 0.374762 +Steps: 0%| | 2829/1000000 [11:46<68:32:46, 4.04it/s, grad_norm=0.375, loss_final=0.514, loss_mean=0.82, loss_mean_cls=0.085, proj_loss=-0.391][2026-03-26 12:42:49] Step: 2829, Training Logs: loss_final: 0.519421, loss_mean: 0.833790, proj_loss: -0.396443, loss_mean_cls: 0.082074, grad_norm: 0.465053 +Steps: 0%| | 2830/1000000 [11:46<68:33:46, 4.04it/s, grad_norm=0.465, loss_final=0.519, loss_mean=0.834, loss_mean_cls=0.0821, proj_loss=-0.396][2026-03-26 12:42:49] Step: 2830, Training Logs: loss_final: 0.533474, loss_mean: 0.839454, proj_loss: -0.390024, loss_mean_cls: 0.084045, grad_norm: 0.505070 +Steps: 0%| | 2831/1000000 [11:46<68:34:18, 4.04it/s, grad_norm=0.505, loss_final=0.533, loss_mean=0.839, loss_mean_cls=0.084, proj_loss=-0.39][2026-03-26 12:42:50] Step: 2831, Training Logs: loss_final: 0.515982, loss_mean: 0.830166, proj_loss: -0.397376, loss_mean_cls: 0.083193, grad_norm: 0.505766 +Steps: 0%| | 2832/1000000 [11:46<68:33:18, 4.04it/s, grad_norm=0.506, loss_final=0.516, loss_mean=0.83, loss_mean_cls=0.0832, proj_loss=-0.397][2026-03-26 12:42:50] Step: 2832, Training Logs: loss_final: 0.529133, loss_mean: 0.853937, proj_loss: -0.404502, loss_mean_cls: 0.079699, grad_norm: 0.560772 +Steps: 0%| | 2833/1000000 [11:47<68:33:49, 4.04it/s, grad_norm=0.561, loss_final=0.529, loss_mean=0.854, loss_mean_cls=0.0797, proj_loss=-0.405][2026-03-26 12:42:50] Step: 2833, Training Logs: loss_final: 0.520375, loss_mean: 0.843040, proj_loss: -0.402258, loss_mean_cls: 0.079593, grad_norm: 0.440674 +Steps: 0%| | 2834/1000000 [11:47<68:32:44, 4.04it/s, grad_norm=0.441, loss_final=0.52, loss_mean=0.843, loss_mean_cls=0.0796, proj_loss=-0.402][2026-03-26 12:42:50] Step: 2834, Training Logs: loss_final: 0.512626, loss_mean: 0.828517, proj_loss: -0.398708, loss_mean_cls: 0.082817, grad_norm: 0.433162 +Steps: 0%| | 2835/1000000 [11:47<68:32:46, 4.04it/s, grad_norm=0.433, loss_final=0.513, loss_mean=0.829, loss_mean_cls=0.0828, proj_loss=-0.399][2026-03-26 12:42:51] Step: 2835, Training Logs: loss_final: 0.522236, loss_mean: 0.828390, proj_loss: -0.390628, loss_mean_cls: 0.084474, grad_norm: 0.636718 +Steps: 0%| | 2836/1000000 [11:47<68:32:25, 4.04it/s, grad_norm=0.637, loss_final=0.522, loss_mean=0.828, loss_mean_cls=0.0845, proj_loss=-0.391][2026-03-26 12:42:51] Step: 2836, Training Logs: loss_final: 0.516582, loss_mean: 0.837401, proj_loss: -0.402530, loss_mean_cls: 0.081710, grad_norm: 0.410208 +Steps: 0%| | 2837/1000000 [11:48<68:34:38, 4.04it/s, grad_norm=0.41, loss_final=0.517, loss_mean=0.837, loss_mean_cls=0.0817, proj_loss=-0.403][2026-03-26 12:42:51] Step: 2837, Training Logs: loss_final: 0.526432, loss_mean: 0.848940, proj_loss: -0.401942, loss_mean_cls: 0.079434, grad_norm: 0.672370 +Steps: 0%| | 2838/1000000 [11:48<68:33:08, 4.04it/s, grad_norm=0.672, loss_final=0.526, loss_mean=0.849, loss_mean_cls=0.0794, proj_loss=-0.402][2026-03-26 12:42:51] Step: 2838, Training Logs: loss_final: 0.526406, loss_mean: 0.843573, proj_loss: -0.399620, loss_mean_cls: 0.082452, grad_norm: 0.706284 +Steps: 0%| | 2839/1000000 [11:48<68:33:15, 4.04it/s, grad_norm=0.706, loss_final=0.526, loss_mean=0.844, loss_mean_cls=0.0825, proj_loss=-0.4][2026-03-26 12:42:52] Step: 2839, Training Logs: loss_final: 0.535650, loss_mean: 0.833876, proj_loss: -0.382612, loss_mean_cls: 0.084385, grad_norm: 0.388900 +Steps: 0%| | 2840/1000000 [11:48<68:31:09, 4.04it/s, grad_norm=0.389, loss_final=0.536, loss_mean=0.834, loss_mean_cls=0.0844, proj_loss=-0.383][2026-03-26 12:42:52] Step: 2840, Training Logs: loss_final: 0.523950, loss_mean: 0.828479, proj_loss: -0.391048, loss_mean_cls: 0.086519, grad_norm: 0.427293 +Steps: 0%| | 2841/1000000 [11:49<68:32:31, 4.04it/s, grad_norm=0.427, loss_final=0.524, loss_mean=0.828, loss_mean_cls=0.0865, proj_loss=-0.391][2026-03-26 12:42:52] Step: 2841, Training Logs: loss_final: 0.529839, loss_mean: 0.873371, proj_loss: -0.418985, loss_mean_cls: 0.075454, grad_norm: 0.367621 +Steps: 0%| | 2842/1000000 [11:49<68:32:33, 4.04it/s, grad_norm=0.368, loss_final=0.53, loss_mean=0.873, loss_mean_cls=0.0755, proj_loss=-0.419][2026-03-26 12:42:52] Step: 2842, Training Logs: loss_final: 0.522192, loss_mean: 0.843277, proj_loss: -0.401317, loss_mean_cls: 0.080232, grad_norm: 0.430590 +Steps: 0%| | 2843/1000000 [11:49<68:33:10, 4.04it/s, grad_norm=0.431, loss_final=0.522, loss_mean=0.843, loss_mean_cls=0.0802, proj_loss=-0.401][2026-03-26 12:42:53] Step: 2843, Training Logs: loss_final: 0.516826, loss_mean: 0.844658, proj_loss: -0.405858, loss_mean_cls: 0.078025, grad_norm: 0.354067 +Steps: 0%| | 2844/1000000 [11:49<68:53:17, 4.02it/s, grad_norm=0.354, loss_final=0.517, loss_mean=0.845, loss_mean_cls=0.078, proj_loss=-0.406][2026-03-26 12:42:53] Step: 2844, Training Logs: loss_final: 0.524015, loss_mean: 0.833186, proj_loss: -0.394330, loss_mean_cls: 0.085159, grad_norm: 0.536538 +Steps: 0%| | 2845/1000000 [11:50<68:47:44, 4.03it/s, grad_norm=0.537, loss_final=0.524, loss_mean=0.833, loss_mean_cls=0.0852, proj_loss=-0.394][2026-03-26 12:42:53] Step: 2845, Training Logs: loss_final: 0.520226, loss_mean: 0.837503, proj_loss: -0.398465, loss_mean_cls: 0.081188, grad_norm: 0.417243 +Steps: 0%| | 2846/1000000 [11:50<68:48:12, 4.03it/s, grad_norm=0.417, loss_final=0.52, loss_mean=0.838, loss_mean_cls=0.0812, proj_loss=-0.398][2026-03-26 12:42:53] Step: 2846, Training Logs: loss_final: 0.510632, loss_mean: 0.823796, proj_loss: -0.395145, loss_mean_cls: 0.081981, grad_norm: 0.358186 +Steps: 0%| | 2847/1000000 [11:50<68:43:09, 4.03it/s, grad_norm=0.358, loss_final=0.511, loss_mean=0.824, loss_mean_cls=0.082, proj_loss=-0.395][2026-03-26 12:42:54] Step: 2847, Training Logs: loss_final: 0.517625, loss_mean: 0.824674, proj_loss: -0.392970, loss_mean_cls: 0.085921, grad_norm: 0.312696 +Steps: 0%| | 2848/1000000 [11:50<68:40:44, 4.03it/s, grad_norm=0.313, loss_final=0.518, loss_mean=0.825, loss_mean_cls=0.0859, proj_loss=-0.393][2026-03-26 12:42:54] Step: 2848, Training Logs: loss_final: 0.527294, loss_mean: 0.835587, proj_loss: -0.391798, loss_mean_cls: 0.083505, grad_norm: 0.482141 +Steps: 0%| | 2849/1000000 [11:51<68:37:43, 4.04it/s, grad_norm=0.482, loss_final=0.527, loss_mean=0.836, loss_mean_cls=0.0835, proj_loss=-0.392][2026-03-26 12:42:54] Step: 2849, Training Logs: loss_final: 0.521937, loss_mean: 0.837239, proj_loss: -0.398345, loss_mean_cls: 0.083043, grad_norm: 0.390563 +Steps: 0%| | 2850/1000000 [11:51<68:36:53, 4.04it/s, grad_norm=0.391, loss_final=0.522, loss_mean=0.837, loss_mean_cls=0.083, proj_loss=-0.398][2026-03-26 12:42:54] Step: 2850, Training Logs: loss_final: 0.527152, loss_mean: 0.844402, proj_loss: -0.399253, loss_mean_cls: 0.082003, grad_norm: 0.571958 +Steps: 0%| | 2851/1000000 [11:51<68:39:41, 4.03it/s, grad_norm=0.572, loss_final=0.527, loss_mean=0.844, loss_mean_cls=0.082, proj_loss=-0.399][2026-03-26 12:42:54] Step: 2851, Training Logs: loss_final: 0.519986, loss_mean: 0.827582, proj_loss: -0.393239, loss_mean_cls: 0.085643, grad_norm: 0.378440 +Steps: 0%| | 2852/1000000 [11:51<68:38:58, 4.03it/s, grad_norm=0.378, loss_final=0.52, loss_mean=0.828, loss_mean_cls=0.0856, proj_loss=-0.393][2026-03-26 12:42:55] Step: 2852, Training Logs: loss_final: 0.518987, loss_mean: 0.842846, proj_loss: -0.403601, loss_mean_cls: 0.079742, grad_norm: 0.555726 +Steps: 0%| | 2853/1000000 [11:52<68:37:05, 4.04it/s, grad_norm=0.556, loss_final=0.519, loss_mean=0.843, loss_mean_cls=0.0797, proj_loss=-0.404][2026-03-26 12:42:55] Step: 2853, Training Logs: loss_final: 0.528301, loss_mean: 0.855535, proj_loss: -0.406953, loss_mean_cls: 0.079719, grad_norm: 0.449940 +Steps: 0%| | 2854/1000000 [11:52<68:37:53, 4.04it/s, grad_norm=0.45, loss_final=0.528, loss_mean=0.856, loss_mean_cls=0.0797, proj_loss=-0.407][2026-03-26 12:42:55] Step: 2854, Training Logs: loss_final: 0.529756, loss_mean: 0.852896, proj_loss: -0.402982, loss_mean_cls: 0.079842, grad_norm: 0.366818 +Steps: 0%| | 2855/1000000 [11:52<68:34:12, 4.04it/s, grad_norm=0.367, loss_final=0.53, loss_mean=0.853, loss_mean_cls=0.0798, proj_loss=-0.403][2026-03-26 12:42:55] Step: 2855, Training Logs: loss_final: 0.543381, loss_mean: 0.854135, proj_loss: -0.393010, loss_mean_cls: 0.082256, grad_norm: 0.538984 +Steps: 0%| | 2856/1000000 [11:52<68:37:25, 4.04it/s, grad_norm=0.539, loss_final=0.543, loss_mean=0.854, loss_mean_cls=0.0823, proj_loss=-0.393][2026-03-26 12:42:56] Step: 2856, Training Logs: loss_final: 0.530674, loss_mean: 0.845433, proj_loss: -0.397813, loss_mean_cls: 0.083054, grad_norm: 0.562518 +Steps: 0%| | 2857/1000000 [11:53<68:33:37, 4.04it/s, grad_norm=0.563, loss_final=0.531, loss_mean=0.845, loss_mean_cls=0.0831, proj_loss=-0.398][2026-03-26 12:42:56] Step: 2857, Training Logs: loss_final: 0.511290, loss_mean: 0.832273, proj_loss: -0.401901, loss_mean_cls: 0.080918, grad_norm: 0.503635 +Steps: 0%| | 2858/1000000 [11:53<68:34:30, 4.04it/s, grad_norm=0.504, loss_final=0.511, loss_mean=0.832, loss_mean_cls=0.0809, proj_loss=-0.402][2026-03-26 12:42:56] Step: 2858, Training Logs: loss_final: 0.528654, loss_mean: 0.857717, proj_loss: -0.408461, loss_mean_cls: 0.079398, grad_norm: 0.541414 +Steps: 0%| | 2859/1000000 [11:53<68:33:08, 4.04it/s, grad_norm=0.541, loss_final=0.529, loss_mean=0.858, loss_mean_cls=0.0794, proj_loss=-0.408][2026-03-26 12:42:56] Step: 2859, Training Logs: loss_final: 0.533946, loss_mean: 0.854754, proj_loss: -0.401238, loss_mean_cls: 0.080430, grad_norm: 0.364676 +Steps: 0%| | 2860/1000000 [11:53<68:33:50, 4.04it/s, grad_norm=0.365, loss_final=0.534, loss_mean=0.855, loss_mean_cls=0.0804, proj_loss=-0.401][2026-03-26 12:42:57] Step: 2860, Training Logs: loss_final: 0.526292, loss_mean: 0.845155, proj_loss: -0.400468, loss_mean_cls: 0.081605, grad_norm: 0.473215 +Steps: 0%| | 2861/1000000 [11:54<68:32:39, 4.04it/s, grad_norm=0.473, loss_final=0.526, loss_mean=0.845, loss_mean_cls=0.0816, proj_loss=-0.4][2026-03-26 12:42:57] Step: 2861, Training Logs: loss_final: 0.533143, loss_mean: 0.852546, proj_loss: -0.401313, loss_mean_cls: 0.081909, grad_norm: 0.405349 +Steps: 0%| | 2862/1000000 [11:54<68:33:28, 4.04it/s, grad_norm=0.405, loss_final=0.533, loss_mean=0.853, loss_mean_cls=0.0819, proj_loss=-0.401][2026-03-26 12:42:57] Step: 2862, Training Logs: loss_final: 0.512955, loss_mean: 0.836917, proj_loss: -0.404797, loss_mean_cls: 0.080835, grad_norm: 0.359299 +Steps: 0%| | 2863/1000000 [11:54<68:31:54, 4.04it/s, grad_norm=0.359, loss_final=0.513, loss_mean=0.837, loss_mean_cls=0.0808, proj_loss=-0.405][2026-03-26 12:42:57] Step: 2863, Training Logs: loss_final: 0.514089, loss_mean: 0.832794, proj_loss: -0.402056, loss_mean_cls: 0.083351, grad_norm: 0.448521 +Steps: 0%| | 2864/1000000 [11:54<68:33:26, 4.04it/s, grad_norm=0.449, loss_final=0.514, loss_mean=0.833, loss_mean_cls=0.0834, proj_loss=-0.402][2026-03-26 12:42:58] Step: 2864, Training Logs: loss_final: 0.500346, loss_mean: 0.809226, proj_loss: -0.393542, loss_mean_cls: 0.084662, grad_norm: 0.436281 +Steps: 0%| | 2865/1000000 [11:55<68:36:00, 4.04it/s, grad_norm=0.436, loss_final=0.5, loss_mean=0.809, loss_mean_cls=0.0847, proj_loss=-0.394][2026-03-26 12:42:58] Step: 2865, Training Logs: loss_final: 0.508602, loss_mean: 0.821802, proj_loss: -0.396256, loss_mean_cls: 0.083056, grad_norm: 0.332707 +Steps: 0%| | 2866/1000000 [11:55<68:35:12, 4.04it/s, grad_norm=0.333, loss_final=0.509, loss_mean=0.822, loss_mean_cls=0.0831, proj_loss=-0.396][2026-03-26 12:42:58] Step: 2866, Training Logs: loss_final: 0.528620, loss_mean: 0.857968, proj_loss: -0.408736, loss_mean_cls: 0.079388, grad_norm: 0.408504 +Steps: 0%| | 2867/1000000 [11:55<68:33:21, 4.04it/s, grad_norm=0.409, loss_final=0.529, loss_mean=0.858, loss_mean_cls=0.0794, proj_loss=-0.409][2026-03-26 12:42:58] Step: 2867, Training Logs: loss_final: 0.529091, loss_mean: 0.852980, proj_loss: -0.403939, loss_mean_cls: 0.080051, grad_norm: 0.327080 +Steps: 0%| | 2868/1000000 [11:55<68:37:58, 4.04it/s, grad_norm=0.327, loss_final=0.529, loss_mean=0.853, loss_mean_cls=0.0801, proj_loss=-0.404][2026-03-26 12:42:59] Step: 2868, Training Logs: loss_final: 0.515873, loss_mean: 0.834019, proj_loss: -0.400199, loss_mean_cls: 0.082053, grad_norm: 0.264719 +Steps: 0%| | 2869/1000000 [11:56<68:37:21, 4.04it/s, grad_norm=0.265, loss_final=0.516, loss_mean=0.834, loss_mean_cls=0.0821, proj_loss=-0.4][2026-03-26 12:42:59] Step: 2869, Training Logs: loss_final: 0.532933, loss_mean: 0.839076, proj_loss: -0.390446, loss_mean_cls: 0.084303, grad_norm: 0.408225 +Steps: 0%| | 2870/1000000 [11:56<68:37:20, 4.04it/s, grad_norm=0.408, loss_final=0.533, loss_mean=0.839, loss_mean_cls=0.0843, proj_loss=-0.39][2026-03-26 12:42:59] Step: 2870, Training Logs: loss_final: 0.514413, loss_mean: 0.824406, proj_loss: -0.393397, loss_mean_cls: 0.083404, grad_norm: 0.338733 +Steps: 0%| | 2871/1000000 [11:56<68:33:50, 4.04it/s, grad_norm=0.339, loss_final=0.514, loss_mean=0.824, loss_mean_cls=0.0834, proj_loss=-0.393][2026-03-26 12:42:59] Step: 2871, Training Logs: loss_final: 0.548470, loss_mean: 0.863969, proj_loss: -0.396313, loss_mean_cls: 0.080813, grad_norm: 0.582433 +Steps: 0%| | 2872/1000000 [11:56<68:32:25, 4.04it/s, grad_norm=0.582, loss_final=0.548, loss_mean=0.864, loss_mean_cls=0.0808, proj_loss=-0.396][2026-03-26 12:43:00] Step: 2872, Training Logs: loss_final: 0.525192, loss_mean: 0.849372, proj_loss: -0.405489, loss_mean_cls: 0.081309, grad_norm: 0.453910 +Steps: 0%| | 2873/1000000 [11:57<68:32:02, 4.04it/s, grad_norm=0.454, loss_final=0.525, loss_mean=0.849, loss_mean_cls=0.0813, proj_loss=-0.405][2026-03-26 12:43:00] Step: 2873, Training Logs: loss_final: 0.500449, loss_mean: 0.806636, proj_loss: -0.392148, loss_mean_cls: 0.085961, grad_norm: 0.696622 +Steps: 0%| | 2874/1000000 [11:57<68:32:25, 4.04it/s, grad_norm=0.697, loss_final=0.5, loss_mean=0.807, loss_mean_cls=0.086, proj_loss=-0.392][2026-03-26 12:43:00] Step: 2874, Training Logs: loss_final: 0.526428, loss_mean: 0.845659, proj_loss: -0.401588, loss_mean_cls: 0.082357, grad_norm: 0.362560 +Steps: 0%| | 2875/1000000 [11:57<68:32:15, 4.04it/s, grad_norm=0.363, loss_final=0.526, loss_mean=0.846, loss_mean_cls=0.0824, proj_loss=-0.402][2026-03-26 12:43:00] Step: 2875, Training Logs: loss_final: 0.510454, loss_mean: 0.836569, proj_loss: -0.406416, loss_mean_cls: 0.080301, grad_norm: 0.630087 +Steps: 0%| | 2876/1000000 [11:57<68:30:42, 4.04it/s, grad_norm=0.63, loss_final=0.51, loss_mean=0.837, loss_mean_cls=0.0803, proj_loss=-0.406][2026-03-26 12:43:01] Step: 2876, Training Logs: loss_final: 0.538291, loss_mean: 0.872615, proj_loss: -0.412821, loss_mean_cls: 0.078497, grad_norm: 0.835697 +Steps: 0%| | 2877/1000000 [11:58<68:33:02, 4.04it/s, grad_norm=0.836, loss_final=0.538, loss_mean=0.873, loss_mean_cls=0.0785, proj_loss=-0.413][2026-03-26 12:43:01] Step: 2877, Training Logs: loss_final: 0.531688, loss_mean: 0.849377, proj_loss: -0.400403, loss_mean_cls: 0.082715, grad_norm: 0.618852 +Steps: 0%| | 2878/1000000 [11:58<68:32:50, 4.04it/s, grad_norm=0.619, loss_final=0.532, loss_mean=0.849, loss_mean_cls=0.0827, proj_loss=-0.4][2026-03-26 12:43:01] Step: 2878, Training Logs: loss_final: 0.508413, loss_mean: 0.833283, proj_loss: -0.404841, loss_mean_cls: 0.079971, grad_norm: 0.477735 +Steps: 0%| | 2879/1000000 [11:58<68:34:27, 4.04it/s, grad_norm=0.478, loss_final=0.508, loss_mean=0.833, loss_mean_cls=0.08, proj_loss=-0.405][2026-03-26 12:43:01] Step: 2879, Training Logs: loss_final: 0.525780, loss_mean: 0.839405, proj_loss: -0.395782, loss_mean_cls: 0.082157, grad_norm: 0.985447 +Steps: 0%| | 2880/1000000 [11:58<68:33:06, 4.04it/s, grad_norm=0.985, loss_final=0.526, loss_mean=0.839, loss_mean_cls=0.0822, proj_loss=-0.396][2026-03-26 12:43:02] Step: 2880, Training Logs: loss_final: 0.523444, loss_mean: 0.835496, proj_loss: -0.395126, loss_mean_cls: 0.083074, grad_norm: 0.491500 +Steps: 0%| | 2881/1000000 [11:59<68:32:10, 4.04it/s, grad_norm=0.491, loss_final=0.523, loss_mean=0.835, loss_mean_cls=0.0831, proj_loss=-0.395][2026-03-26 12:43:02] Step: 2881, Training Logs: loss_final: 0.532375, loss_mean: 0.849817, proj_loss: -0.399123, loss_mean_cls: 0.081681, grad_norm: 0.705324 +Steps: 0%| | 2882/1000000 [11:59<68:32:35, 4.04it/s, grad_norm=0.705, loss_final=0.532, loss_mean=0.85, loss_mean_cls=0.0817, proj_loss=-0.399][2026-03-26 12:43:02] Step: 2882, Training Logs: loss_final: 0.524926, loss_mean: 0.844613, proj_loss: -0.402342, loss_mean_cls: 0.082655, grad_norm: 0.873473 +Steps: 0%| | 2883/1000000 [11:59<68:34:56, 4.04it/s, grad_norm=0.873, loss_final=0.525, loss_mean=0.845, loss_mean_cls=0.0827, proj_loss=-0.402][2026-03-26 12:43:02] Step: 2883, Training Logs: loss_final: 0.524504, loss_mean: 0.839435, proj_loss: -0.397943, loss_mean_cls: 0.083012, grad_norm: 0.395603 +Steps: 0%| | 2884/1000000 [11:59<68:33:08, 4.04it/s, grad_norm=0.396, loss_final=0.525, loss_mean=0.839, loss_mean_cls=0.083, proj_loss=-0.398][2026-03-26 12:43:03] Step: 2884, Training Logs: loss_final: 0.533273, loss_mean: 0.848562, proj_loss: -0.398856, loss_mean_cls: 0.083567, grad_norm: 0.868994 +Steps: 0%| | 2885/1000000 [12:00<68:33:44, 4.04it/s, grad_norm=0.869, loss_final=0.533, loss_mean=0.849, loss_mean_cls=0.0836, proj_loss=-0.399][2026-03-26 12:43:03] Step: 2885, Training Logs: loss_final: 0.532933, loss_mean: 0.851306, proj_loss: -0.401130, loss_mean_cls: 0.082757, grad_norm: 0.555139 +Steps: 0%| | 2886/1000000 [12:00<68:34:01, 4.04it/s, grad_norm=0.555, loss_final=0.533, loss_mean=0.851, loss_mean_cls=0.0828, proj_loss=-0.401][2026-03-26 12:43:03] Step: 2886, Training Logs: loss_final: 0.529912, loss_mean: 0.830349, proj_loss: -0.385342, loss_mean_cls: 0.084905, grad_norm: 0.453145 +Steps: 0%| | 2887/1000000 [12:00<68:31:08, 4.04it/s, grad_norm=0.453, loss_final=0.53, loss_mean=0.83, loss_mean_cls=0.0849, proj_loss=-0.385][2026-03-26 12:43:03] Step: 2887, Training Logs: loss_final: 0.510248, loss_mean: 0.831025, proj_loss: -0.402083, loss_mean_cls: 0.081306, grad_norm: 0.467455 +Steps: 0%| | 2888/1000000 [12:00<68:29:37, 4.04it/s, grad_norm=0.467, loss_final=0.51, loss_mean=0.831, loss_mean_cls=0.0813, proj_loss=-0.402][2026-03-26 12:43:04] Step: 2888, Training Logs: loss_final: 0.515736, loss_mean: 0.835465, proj_loss: -0.400159, loss_mean_cls: 0.080430, grad_norm: 0.447308 +Steps: 0%| | 2889/1000000 [12:01<68:30:47, 4.04it/s, grad_norm=0.447, loss_final=0.516, loss_mean=0.835, loss_mean_cls=0.0804, proj_loss=-0.4][2026-03-26 12:43:04] Step: 2889, Training Logs: loss_final: 0.528376, loss_mean: 0.850566, proj_loss: -0.402501, loss_mean_cls: 0.080311, grad_norm: 0.497336 +Steps: 0%| | 2890/1000000 [12:01<68:32:18, 4.04it/s, grad_norm=0.497, loss_final=0.528, loss_mean=0.851, loss_mean_cls=0.0803, proj_loss=-0.403][2026-03-26 12:43:04] Step: 2890, Training Logs: loss_final: 0.536672, loss_mean: 0.862033, proj_loss: -0.405272, loss_mean_cls: 0.079912, grad_norm: 0.345902 +Steps: 0%| | 2891/1000000 [12:01<68:32:13, 4.04it/s, grad_norm=0.346, loss_final=0.537, loss_mean=0.862, loss_mean_cls=0.0799, proj_loss=-0.405][2026-03-26 12:43:04] Step: 2891, Training Logs: loss_final: 0.526568, loss_mean: 0.844264, proj_loss: -0.400970, loss_mean_cls: 0.083273, grad_norm: 0.586294 +Steps: 0%| | 2892/1000000 [12:01<68:31:29, 4.04it/s, grad_norm=0.586, loss_final=0.527, loss_mean=0.844, loss_mean_cls=0.0833, proj_loss=-0.401][2026-03-26 12:43:05] Step: 2892, Training Logs: loss_final: 0.519299, loss_mean: 0.836861, proj_loss: -0.399292, loss_mean_cls: 0.081730, grad_norm: 0.516857 +Steps: 0%| | 2893/1000000 [12:02<68:44:11, 4.03it/s, grad_norm=0.517, loss_final=0.519, loss_mean=0.837, loss_mean_cls=0.0817, proj_loss=-0.399][2026-03-26 12:43:05] Step: 2893, Training Logs: loss_final: 0.525928, loss_mean: 0.838499, proj_loss: -0.395239, loss_mean_cls: 0.082668, grad_norm: 0.411485 +Steps: 0%| | 2894/1000000 [12:02<68:42:03, 4.03it/s, grad_norm=0.411, loss_final=0.526, loss_mean=0.838, loss_mean_cls=0.0827, proj_loss=-0.395][2026-03-26 12:43:05] Step: 2894, Training Logs: loss_final: 0.538819, loss_mean: 0.867960, proj_loss: -0.408650, loss_mean_cls: 0.079509, grad_norm: 0.411541 +Steps: 0%| | 2895/1000000 [12:02<68:37:08, 4.04it/s, grad_norm=0.412, loss_final=0.539, loss_mean=0.868, loss_mean_cls=0.0795, proj_loss=-0.409][2026-03-26 12:43:05] Step: 2895, Training Logs: loss_final: 0.530866, loss_mean: 0.859641, proj_loss: -0.407639, loss_mean_cls: 0.078864, grad_norm: 0.400969 +Steps: 0%| | 2896/1000000 [12:02<68:36:55, 4.04it/s, grad_norm=0.401, loss_final=0.531, loss_mean=0.86, loss_mean_cls=0.0789, proj_loss=-0.408][2026-03-26 12:43:06] Step: 2896, Training Logs: loss_final: 0.524952, loss_mean: 0.850129, proj_loss: -0.405852, loss_mean_cls: 0.080675, grad_norm: 0.299807 +Steps: 0%| | 2897/1000000 [12:02<68:35:24, 4.04it/s, grad_norm=0.3, loss_final=0.525, loss_mean=0.85, loss_mean_cls=0.0807, proj_loss=-0.406][2026-03-26 12:43:06] Step: 2897, Training Logs: loss_final: 0.532666, loss_mean: 0.847755, proj_loss: -0.397385, loss_mean_cls: 0.082296, grad_norm: 0.449354 +Steps: 0%| | 2898/1000000 [12:03<68:41:44, 4.03it/s, grad_norm=0.449, loss_final=0.533, loss_mean=0.848, loss_mean_cls=0.0823, proj_loss=-0.397][2026-03-26 12:43:06] Step: 2898, Training Logs: loss_final: 0.545899, loss_mean: 0.873939, proj_loss: -0.408731, loss_mean_cls: 0.080691, grad_norm: 0.473933 +Steps: 0%| | 2899/1000000 [12:03<68:38:18, 4.04it/s, grad_norm=0.474, loss_final=0.546, loss_mean=0.874, loss_mean_cls=0.0807, proj_loss=-0.409][2026-03-26 12:43:06] Step: 2899, Training Logs: loss_final: 0.517735, loss_mean: 0.843509, proj_loss: -0.405602, loss_mean_cls: 0.079827, grad_norm: 0.560491 +Steps: 0%| | 2900/1000000 [12:03<68:41:58, 4.03it/s, grad_norm=0.56, loss_final=0.518, loss_mean=0.844, loss_mean_cls=0.0798, proj_loss=-0.406][2026-03-26 12:43:07] Step: 2900, Training Logs: loss_final: 0.521870, loss_mean: 0.845804, proj_loss: -0.405929, loss_mean_cls: 0.081996, grad_norm: 0.391652 +Steps: 0%| | 2901/1000000 [12:03<68:37:33, 4.04it/s, grad_norm=0.392, loss_final=0.522, loss_mean=0.846, loss_mean_cls=0.082, proj_loss=-0.406][2026-03-26 12:43:07] Step: 2901, Training Logs: loss_final: 0.528392, loss_mean: 0.836559, proj_loss: -0.392372, loss_mean_cls: 0.084204, grad_norm: 0.446088 +Steps: 0%| | 2902/1000000 [12:04<68:37:09, 4.04it/s, grad_norm=0.446, loss_final=0.528, loss_mean=0.837, loss_mean_cls=0.0842, proj_loss=-0.392][2026-03-26 12:43:07] Step: 2902, Training Logs: loss_final: 0.515894, loss_mean: 0.828692, proj_loss: -0.396104, loss_mean_cls: 0.083307, grad_norm: 0.525752 +Steps: 0%| | 2903/1000000 [12:04<68:38:11, 4.04it/s, grad_norm=0.526, loss_final=0.516, loss_mean=0.829, loss_mean_cls=0.0833, proj_loss=-0.396][2026-03-26 12:43:07] Step: 2903, Training Logs: loss_final: 0.502010, loss_mean: 0.815791, proj_loss: -0.397333, loss_mean_cls: 0.083552, grad_norm: 0.375884 +Steps: 0%| | 2904/1000000 [12:04<68:37:51, 4.04it/s, grad_norm=0.376, loss_final=0.502, loss_mean=0.816, loss_mean_cls=0.0836, proj_loss=-0.397][2026-03-26 12:43:08] Step: 2904, Training Logs: loss_final: 0.515835, loss_mean: 0.836255, proj_loss: -0.403727, loss_mean_cls: 0.083308, grad_norm: 0.455801 +Steps: 0%| | 2905/1000000 [12:04<68:42:49, 4.03it/s, grad_norm=0.456, loss_final=0.516, loss_mean=0.836, loss_mean_cls=0.0833, proj_loss=-0.404][2026-03-26 12:43:08] Step: 2905, Training Logs: loss_final: 0.541596, loss_mean: 0.864210, proj_loss: -0.402908, loss_mean_cls: 0.080294, grad_norm: 0.455368 +Steps: 0%| | 2906/1000000 [12:05<68:40:42, 4.03it/s, grad_norm=0.455, loss_final=0.542, loss_mean=0.864, loss_mean_cls=0.0803, proj_loss=-0.403][2026-03-26 12:43:08] Step: 2906, Training Logs: loss_final: 0.529835, loss_mean: 0.839439, proj_loss: -0.391861, loss_mean_cls: 0.082258, grad_norm: 0.407362 +Steps: 0%| | 2907/1000000 [12:05<68:36:37, 4.04it/s, grad_norm=0.407, loss_final=0.53, loss_mean=0.839, loss_mean_cls=0.0823, proj_loss=-0.392][2026-03-26 12:43:08] Step: 2907, Training Logs: loss_final: 0.503077, loss_mean: 0.825478, proj_loss: -0.403055, loss_mean_cls: 0.080655, grad_norm: 0.348451 +Steps: 0%| | 2908/1000000 [12:05<68:34:50, 4.04it/s, grad_norm=0.348, loss_final=0.503, loss_mean=0.825, loss_mean_cls=0.0807, proj_loss=-0.403][2026-03-26 12:43:09] Step: 2908, Training Logs: loss_final: 0.524400, loss_mean: 0.841180, proj_loss: -0.397749, loss_mean_cls: 0.080970, grad_norm: 0.381199 +Steps: 0%| | 2909/1000000 [12:05<68:42:18, 4.03it/s, grad_norm=0.381, loss_final=0.524, loss_mean=0.841, loss_mean_cls=0.081, proj_loss=-0.398][2026-03-26 12:43:09] Step: 2909, Training Logs: loss_final: 0.509549, loss_mean: 0.825790, proj_loss: -0.399710, loss_mean_cls: 0.083470, grad_norm: 0.345416 +Steps: 0%| | 2910/1000000 [12:06<68:43:04, 4.03it/s, grad_norm=0.345, loss_final=0.51, loss_mean=0.826, loss_mean_cls=0.0835, proj_loss=-0.4][2026-03-26 12:43:09] Step: 2910, Training Logs: loss_final: 0.525574, loss_mean: 0.841868, proj_loss: -0.399002, loss_mean_cls: 0.082707, grad_norm: 0.389445 +Steps: 0%| | 2911/1000000 [12:06<68:39:26, 4.03it/s, grad_norm=0.389, loss_final=0.526, loss_mean=0.842, loss_mean_cls=0.0827, proj_loss=-0.399][2026-03-26 12:43:09] Step: 2911, Training Logs: loss_final: 0.522663, loss_mean: 0.849039, proj_loss: -0.406137, loss_mean_cls: 0.079760, grad_norm: 0.297608 +Steps: 0%| | 2912/1000000 [12:06<68:36:57, 4.04it/s, grad_norm=0.298, loss_final=0.523, loss_mean=0.849, loss_mean_cls=0.0798, proj_loss=-0.406][2026-03-26 12:43:10] Step: 2912, Training Logs: loss_final: 0.530637, loss_mean: 0.845187, proj_loss: -0.396874, loss_mean_cls: 0.082324, grad_norm: 0.445180 +Steps: 0%| | 2913/1000000 [12:06<68:44:13, 4.03it/s, grad_norm=0.445, loss_final=0.531, loss_mean=0.845, loss_mean_cls=0.0823, proj_loss=-0.397][2026-03-26 12:43:10] Step: 2913, Training Logs: loss_final: 0.507429, loss_mean: 0.814372, proj_loss: -0.392843, loss_mean_cls: 0.085901, grad_norm: 0.488066 +Steps: 0%| | 2914/1000000 [12:07<68:42:01, 4.03it/s, grad_norm=0.488, loss_final=0.507, loss_mean=0.814, loss_mean_cls=0.0859, proj_loss=-0.393][2026-03-26 12:43:10] Step: 2914, Training Logs: loss_final: 0.527682, loss_mean: 0.853515, proj_loss: -0.406709, loss_mean_cls: 0.080875, grad_norm: 0.279997 +Steps: 0%| | 2915/1000000 [12:07<68:38:43, 4.03it/s, grad_norm=0.28, loss_final=0.528, loss_mean=0.854, loss_mean_cls=0.0809, proj_loss=-0.407][2026-03-26 12:43:10] Step: 2915, Training Logs: loss_final: 0.518443, loss_mean: 0.842228, proj_loss: -0.404396, loss_mean_cls: 0.080611, grad_norm: 0.726456 +Steps: 0%| | 2916/1000000 [12:07<68:39:01, 4.03it/s, grad_norm=0.726, loss_final=0.518, loss_mean=0.842, loss_mean_cls=0.0806, proj_loss=-0.404][2026-03-26 12:43:11] Step: 2916, Training Logs: loss_final: 0.514614, loss_mean: 0.841246, proj_loss: -0.408228, loss_mean_cls: 0.081597, grad_norm: 0.540765 +Steps: 0%| | 2917/1000000 [12:07<68:46:57, 4.03it/s, grad_norm=0.541, loss_final=0.515, loss_mean=0.841, loss_mean_cls=0.0816, proj_loss=-0.408][2026-03-26 12:43:11] Step: 2917, Training Logs: loss_final: 0.517224, loss_mean: 0.825885, proj_loss: -0.394662, loss_mean_cls: 0.086001, grad_norm: 0.380954 +Steps: 0%| | 2918/1000000 [12:08<68:42:20, 4.03it/s, grad_norm=0.381, loss_final=0.517, loss_mean=0.826, loss_mean_cls=0.086, proj_loss=-0.395][2026-03-26 12:43:11] Step: 2918, Training Logs: loss_final: 0.506941, loss_mean: 0.821004, proj_loss: -0.396992, loss_mean_cls: 0.082929, grad_norm: 0.714488 +Steps: 0%| | 2919/1000000 [12:08<68:42:08, 4.03it/s, grad_norm=0.714, loss_final=0.507, loss_mean=0.821, loss_mean_cls=0.0829, proj_loss=-0.397][2026-03-26 12:43:11] Step: 2919, Training Logs: loss_final: 0.512004, loss_mean: 0.827372, proj_loss: -0.399755, loss_mean_cls: 0.084387, grad_norm: 0.358338 +Steps: 0%| | 2920/1000000 [12:08<68:41:44, 4.03it/s, grad_norm=0.358, loss_final=0.512, loss_mean=0.827, loss_mean_cls=0.0844, proj_loss=-0.4][2026-03-26 12:43:12] Step: 2920, Training Logs: loss_final: 0.522666, loss_mean: 0.837439, proj_loss: -0.398571, loss_mean_cls: 0.083798, grad_norm: 0.706827 +Steps: 0%| | 2921/1000000 [12:08<68:52:52, 4.02it/s, grad_norm=0.707, loss_final=0.523, loss_mean=0.837, loss_mean_cls=0.0838, proj_loss=-0.399][2026-03-26 12:43:12] Step: 2921, Training Logs: loss_final: 0.535879, loss_mean: 0.846834, proj_loss: -0.394247, loss_mean_cls: 0.083292, grad_norm: 0.729973 +Steps: 0%| | 2922/1000000 [12:09<68:45:53, 4.03it/s, grad_norm=0.73, loss_final=0.536, loss_mean=0.847, loss_mean_cls=0.0833, proj_loss=-0.394][2026-03-26 12:43:12] Step: 2922, Training Logs: loss_final: 0.524906, loss_mean: 0.844187, proj_loss: -0.402140, loss_mean_cls: 0.082859, grad_norm: 0.291295 +Steps: 0%| | 2923/1000000 [12:09<68:45:39, 4.03it/s, grad_norm=0.291, loss_final=0.525, loss_mean=0.844, loss_mean_cls=0.0829, proj_loss=-0.402][2026-03-26 12:43:12] Step: 2923, Training Logs: loss_final: 0.522258, loss_mean: 0.842457, proj_loss: -0.402433, loss_mean_cls: 0.082234, grad_norm: 0.716867 +Steps: 0%| | 2924/1000000 [12:09<68:41:14, 4.03it/s, grad_norm=0.717, loss_final=0.522, loss_mean=0.842, loss_mean_cls=0.0822, proj_loss=-0.402][2026-03-26 12:43:13] Step: 2924, Training Logs: loss_final: 0.503257, loss_mean: 0.821095, proj_loss: -0.400073, loss_mean_cls: 0.082236, grad_norm: 0.463113 +Steps: 0%| | 2925/1000000 [12:09<68:48:28, 4.03it/s, grad_norm=0.463, loss_final=0.503, loss_mean=0.821, loss_mean_cls=0.0822, proj_loss=-0.4][2026-03-26 12:43:13] Step: 2925, Training Logs: loss_final: 0.530307, loss_mean: 0.842810, proj_loss: -0.394848, loss_mean_cls: 0.082344, grad_norm: 0.592618 +Steps: 0%| | 2926/1000000 [12:10<68:43:13, 4.03it/s, grad_norm=0.593, loss_final=0.53, loss_mean=0.843, loss_mean_cls=0.0823, proj_loss=-0.395][2026-03-26 12:43:13] Step: 2926, Training Logs: loss_final: 0.521055, loss_mean: 0.848574, proj_loss: -0.407553, loss_mean_cls: 0.080034, grad_norm: 0.702790 +Steps: 0%| | 2927/1000000 [12:10<68:41:34, 4.03it/s, grad_norm=0.703, loss_final=0.521, loss_mean=0.849, loss_mean_cls=0.08, proj_loss=-0.408][2026-03-26 12:43:13] Step: 2927, Training Logs: loss_final: 0.518154, loss_mean: 0.836280, proj_loss: -0.399789, loss_mean_cls: 0.081664, grad_norm: 0.312806 +Steps: 0%| | 2928/1000000 [12:10<68:43:27, 4.03it/s, grad_norm=0.313, loss_final=0.518, loss_mean=0.836, loss_mean_cls=0.0817, proj_loss=-0.4][2026-03-26 12:43:14] Step: 2928, Training Logs: loss_final: 0.517406, loss_mean: 0.832721, proj_loss: -0.398989, loss_mean_cls: 0.083675, grad_norm: 0.758728 +Steps: 0%| | 2929/1000000 [12:10<68:51:00, 4.02it/s, grad_norm=0.759, loss_final=0.517, loss_mean=0.833, loss_mean_cls=0.0837, proj_loss=-0.399][2026-03-26 12:43:14] Step: 2929, Training Logs: loss_final: 0.522532, loss_mean: 0.834493, proj_loss: -0.394710, loss_mean_cls: 0.082749, grad_norm: 0.653209 +Steps: 0%| | 2930/1000000 [12:11<68:46:44, 4.03it/s, grad_norm=0.653, loss_final=0.523, loss_mean=0.834, loss_mean_cls=0.0827, proj_loss=-0.395][2026-03-26 12:43:14] Step: 2930, Training Logs: loss_final: 0.520192, loss_mean: 0.853748, proj_loss: -0.411941, loss_mean_cls: 0.078386, grad_norm: 0.341577 +Steps: 0%| | 2931/1000000 [12:11<68:43:42, 4.03it/s, grad_norm=0.342, loss_final=0.52, loss_mean=0.854, loss_mean_cls=0.0784, proj_loss=-0.412][2026-03-26 12:43:14] Step: 2931, Training Logs: loss_final: 0.525630, loss_mean: 0.844831, proj_loss: -0.401760, loss_mean_cls: 0.082560, grad_norm: 0.557369 +Steps: 0%| | 2932/1000000 [12:11<68:37:12, 4.04it/s, grad_norm=0.557, loss_final=0.526, loss_mean=0.845, loss_mean_cls=0.0826, proj_loss=-0.402][2026-03-26 12:43:15] Step: 2932, Training Logs: loss_final: 0.507204, loss_mean: 0.825340, proj_loss: -0.401288, loss_mean_cls: 0.083153, grad_norm: 0.388183 +Steps: 0%| | 2933/1000000 [12:11<68:38:06, 4.04it/s, grad_norm=0.388, loss_final=0.507, loss_mean=0.825, loss_mean_cls=0.0832, proj_loss=-0.401][2026-03-26 12:43:15] Step: 2933, Training Logs: loss_final: 0.535711, loss_mean: 0.847544, proj_loss: -0.394061, loss_mean_cls: 0.082228, grad_norm: 0.505111 +Steps: 0%| | 2934/1000000 [12:12<68:35:22, 4.04it/s, grad_norm=0.505, loss_final=0.536, loss_mean=0.848, loss_mean_cls=0.0822, proj_loss=-0.394][2026-03-26 12:43:15] Step: 2934, Training Logs: loss_final: 0.527867, loss_mean: 0.854148, proj_loss: -0.405675, loss_mean_cls: 0.079394, grad_norm: 0.789869 +Steps: 0%| | 2935/1000000 [12:12<68:35:08, 4.04it/s, grad_norm=0.79, loss_final=0.528, loss_mean=0.854, loss_mean_cls=0.0794, proj_loss=-0.406][2026-03-26 12:43:15] Step: 2935, Training Logs: loss_final: 0.519129, loss_mean: 0.838859, proj_loss: -0.401607, loss_mean_cls: 0.081877, grad_norm: 0.423148 +Steps: 0%| | 2936/1000000 [12:12<68:33:06, 4.04it/s, grad_norm=0.423, loss_final=0.519, loss_mean=0.839, loss_mean_cls=0.0819, proj_loss=-0.402][2026-03-26 12:43:16] Step: 2936, Training Logs: loss_final: 0.526383, loss_mean: 0.843295, proj_loss: -0.399838, loss_mean_cls: 0.082926, grad_norm: 0.631736 +Steps: 0%| | 2937/1000000 [12:12<68:33:11, 4.04it/s, grad_norm=0.632, loss_final=0.526, loss_mean=0.843, loss_mean_cls=0.0829, proj_loss=-0.4][2026-03-26 12:43:16] Step: 2937, Training Logs: loss_final: 0.519258, loss_mean: 0.831227, proj_loss: -0.394979, loss_mean_cls: 0.083009, grad_norm: 0.576444 +Steps: 0%| | 2938/1000000 [12:13<68:33:33, 4.04it/s, grad_norm=0.576, loss_final=0.519, loss_mean=0.831, loss_mean_cls=0.083, proj_loss=-0.395][2026-03-26 12:43:16] Step: 2938, Training Logs: loss_final: 0.530200, loss_mean: 0.847769, proj_loss: -0.398516, loss_mean_cls: 0.080948, grad_norm: 0.370609 +Steps: 0%| | 2939/1000000 [12:13<68:33:07, 4.04it/s, grad_norm=0.371, loss_final=0.53, loss_mean=0.848, loss_mean_cls=0.0809, proj_loss=-0.399][2026-03-26 12:43:16] Step: 2939, Training Logs: loss_final: 0.523025, loss_mean: 0.826419, proj_loss: -0.389840, loss_mean_cls: 0.086446, grad_norm: 0.636643 +Steps: 0%| | 2940/1000000 [12:13<68:33:50, 4.04it/s, grad_norm=0.637, loss_final=0.523, loss_mean=0.826, loss_mean_cls=0.0864, proj_loss=-0.39][2026-03-26 12:43:17] Step: 2940, Training Logs: loss_final: 0.511379, loss_mean: 0.830614, proj_loss: -0.400874, loss_mean_cls: 0.081639, grad_norm: 0.479875 +Steps: 0%| | 2941/1000000 [12:13<68:38:02, 4.04it/s, grad_norm=0.48, loss_final=0.511, loss_mean=0.831, loss_mean_cls=0.0816, proj_loss=-0.401][2026-03-26 12:43:17] Step: 2941, Training Logs: loss_final: 0.526409, loss_mean: 0.841981, proj_loss: -0.398105, loss_mean_cls: 0.082533, grad_norm: 0.516576 +Steps: 0%| | 2942/1000000 [12:14<68:54:20, 4.02it/s, grad_norm=0.517, loss_final=0.526, loss_mean=0.842, loss_mean_cls=0.0825, proj_loss=-0.398][2026-03-26 12:43:17] Step: 2942, Training Logs: loss_final: 0.523997, loss_mean: 0.826164, proj_loss: -0.388288, loss_mean_cls: 0.086121, grad_norm: 0.740225 +Steps: 0%| | 2943/1000000 [12:14<69:15:48, 4.00it/s, grad_norm=0.74, loss_final=0.524, loss_mean=0.826, loss_mean_cls=0.0861, proj_loss=-0.388][2026-03-26 12:43:17] Step: 2943, Training Logs: loss_final: 0.523234, loss_mean: 0.843091, proj_loss: -0.402970, loss_mean_cls: 0.083113, grad_norm: 0.543997 +Steps: 0%| | 2944/1000000 [12:14<68:46:54, 4.03it/s, grad_norm=0.544, loss_final=0.523, loss_mean=0.843, loss_mean_cls=0.0831, proj_loss=-0.403][2026-03-26 12:43:18] Step: 2944, Training Logs: loss_final: 0.537551, loss_mean: 0.863845, proj_loss: -0.405759, loss_mean_cls: 0.079465, grad_norm: 0.530652 +Steps: 0%| | 2945/1000000 [12:14<68:43:48, 4.03it/s, grad_norm=0.531, loss_final=0.538, loss_mean=0.864, loss_mean_cls=0.0795, proj_loss=-0.406][2026-03-26 12:43:18] Step: 2945, Training Logs: loss_final: 0.538497, loss_mean: 0.862758, proj_loss: -0.403853, loss_mean_cls: 0.079592, grad_norm: 0.385319 +Steps: 0%| | 2946/1000000 [12:15<68:43:35, 4.03it/s, grad_norm=0.385, loss_final=0.538, loss_mean=0.863, loss_mean_cls=0.0796, proj_loss=-0.404][2026-03-26 12:43:18] Step: 2946, Training Logs: loss_final: 0.501842, loss_mean: 0.807823, proj_loss: -0.391595, loss_mean_cls: 0.085614, grad_norm: 0.434649 +Steps: 0%| | 2947/1000000 [12:15<68:41:33, 4.03it/s, grad_norm=0.435, loss_final=0.502, loss_mean=0.808, loss_mean_cls=0.0856, proj_loss=-0.392][2026-03-26 12:43:18] Step: 2947, Training Logs: loss_final: 0.519921, loss_mean: 0.845439, proj_loss: -0.406779, loss_mean_cls: 0.081261, grad_norm: 0.438293 +Steps: 0%| | 2948/1000000 [12:15<68:40:31, 4.03it/s, grad_norm=0.438, loss_final=0.52, loss_mean=0.845, loss_mean_cls=0.0813, proj_loss=-0.407][2026-03-26 12:43:19] Step: 2948, Training Logs: loss_final: 0.525960, loss_mean: 0.846462, proj_loss: -0.401999, loss_mean_cls: 0.081497, grad_norm: 0.355289 +Steps: 0%| | 2949/1000000 [12:15<68:37:07, 4.04it/s, grad_norm=0.355, loss_final=0.526, loss_mean=0.846, loss_mean_cls=0.0815, proj_loss=-0.402][2026-03-26 12:43:19] Step: 2949, Training Logs: loss_final: 0.521939, loss_mean: 0.851263, proj_loss: -0.409318, loss_mean_cls: 0.079995, grad_norm: 0.310491 +Steps: 0%| | 2950/1000000 [12:16<68:37:43, 4.04it/s, grad_norm=0.31, loss_final=0.522, loss_mean=0.851, loss_mean_cls=0.08, proj_loss=-0.409][2026-03-26 12:43:19] Step: 2950, Training Logs: loss_final: 0.528202, loss_mean: 0.844285, proj_loss: -0.399432, loss_mean_cls: 0.083349, grad_norm: 0.313567 +Steps: 0%| | 2951/1000000 [12:16<68:33:11, 4.04it/s, grad_norm=0.314, loss_final=0.528, loss_mean=0.844, loss_mean_cls=0.0833, proj_loss=-0.399][2026-03-26 12:43:19] Step: 2951, Training Logs: loss_final: 0.522700, loss_mean: 0.840295, proj_loss: -0.400004, loss_mean_cls: 0.082408, grad_norm: 0.395021 +Steps: 0%| | 2952/1000000 [12:16<68:34:11, 4.04it/s, grad_norm=0.395, loss_final=0.523, loss_mean=0.84, loss_mean_cls=0.0824, proj_loss=-0.4][2026-03-26 12:43:20] Step: 2952, Training Logs: loss_final: 0.506550, loss_mean: 0.828950, proj_loss: -0.402915, loss_mean_cls: 0.080515, grad_norm: 0.419920 +Steps: 0%| | 2953/1000000 [12:16<68:32:06, 4.04it/s, grad_norm=0.42, loss_final=0.507, loss_mean=0.829, loss_mean_cls=0.0805, proj_loss=-0.403][2026-03-26 12:43:20] Step: 2953, Training Logs: loss_final: 0.517579, loss_mean: 0.836883, proj_loss: -0.401037, loss_mean_cls: 0.081733, grad_norm: 0.476640 +Steps: 0%| | 2954/1000000 [12:17<68:33:16, 4.04it/s, grad_norm=0.477, loss_final=0.518, loss_mean=0.837, loss_mean_cls=0.0817, proj_loss=-0.401][2026-03-26 12:43:20] Step: 2954, Training Logs: loss_final: 0.512044, loss_mean: 0.829351, proj_loss: -0.399419, loss_mean_cls: 0.082111, grad_norm: 0.323878 +Steps: 0%| | 2955/1000000 [12:17<68:35:13, 4.04it/s, grad_norm=0.324, loss_final=0.512, loss_mean=0.829, loss_mean_cls=0.0821, proj_loss=-0.399][2026-03-26 12:43:20] Step: 2955, Training Logs: loss_final: 0.529328, loss_mean: 0.846295, proj_loss: -0.400913, loss_mean_cls: 0.083946, grad_norm: 0.530764 +Steps: 0%| | 2956/1000000 [12:17<68:36:17, 4.04it/s, grad_norm=0.531, loss_final=0.529, loss_mean=0.846, loss_mean_cls=0.0839, proj_loss=-0.401][2026-03-26 12:43:21] Step: 2956, Training Logs: loss_final: 0.516180, loss_mean: 0.830933, proj_loss: -0.398566, loss_mean_cls: 0.083812, grad_norm: 0.535248 +Steps: 0%| | 2957/1000000 [12:17<68:36:51, 4.04it/s, grad_norm=0.535, loss_final=0.516, loss_mean=0.831, loss_mean_cls=0.0838, proj_loss=-0.399][2026-03-26 12:43:21] Step: 2957, Training Logs: loss_final: 0.531586, loss_mean: 0.851903, proj_loss: -0.401213, loss_mean_cls: 0.080896, grad_norm: 0.411269 +Steps: 0%| | 2958/1000000 [12:18<68:36:23, 4.04it/s, grad_norm=0.411, loss_final=0.532, loss_mean=0.852, loss_mean_cls=0.0809, proj_loss=-0.401][2026-03-26 12:43:21] Step: 2958, Training Logs: loss_final: 0.520262, loss_mean: 0.845200, proj_loss: -0.405081, loss_mean_cls: 0.080143, grad_norm: 0.664074 +Steps: 0%| | 2959/1000000 [12:18<68:39:52, 4.03it/s, grad_norm=0.664, loss_final=0.52, loss_mean=0.845, loss_mean_cls=0.0801, proj_loss=-0.405][2026-03-26 12:43:21] Step: 2959, Training Logs: loss_final: 0.522542, loss_mean: 0.847851, proj_loss: -0.404334, loss_mean_cls: 0.079026, grad_norm: 0.659119 +Steps: 0%| | 2960/1000000 [12:18<68:36:47, 4.04it/s, grad_norm=0.659, loss_final=0.523, loss_mean=0.848, loss_mean_cls=0.079, proj_loss=-0.404][2026-03-26 12:43:22] Step: 2960, Training Logs: loss_final: 0.524891, loss_mean: 0.845649, proj_loss: -0.402777, loss_mean_cls: 0.082018, grad_norm: 0.416127 +Steps: 0%| | 2961/1000000 [12:18<68:38:03, 4.04it/s, grad_norm=0.416, loss_final=0.525, loss_mean=0.846, loss_mean_cls=0.082, proj_loss=-0.403][2026-03-26 12:43:22] Step: 2961, Training Logs: loss_final: 0.521356, loss_mean: 0.819355, proj_loss: -0.384708, loss_mean_cls: 0.086710, grad_norm: 0.512148 +Steps: 0%| | 2962/1000000 [12:19<68:40:05, 4.03it/s, grad_norm=0.512, loss_final=0.521, loss_mean=0.819, loss_mean_cls=0.0867, proj_loss=-0.385][2026-03-26 12:43:22] Step: 2962, Training Logs: loss_final: 0.523306, loss_mean: 0.845706, proj_loss: -0.402763, loss_mean_cls: 0.080362, grad_norm: 0.332124 +Steps: 0%| | 2963/1000000 [12:19<68:39:27, 4.03it/s, grad_norm=0.332, loss_final=0.523, loss_mean=0.846, loss_mean_cls=0.0804, proj_loss=-0.403][2026-03-26 12:43:22] Step: 2963, Training Logs: loss_final: 0.530361, loss_mean: 0.842164, proj_loss: -0.394987, loss_mean_cls: 0.083183, grad_norm: 0.742135 +Steps: 0%| | 2964/1000000 [12:19<68:37:36, 4.04it/s, grad_norm=0.742, loss_final=0.53, loss_mean=0.842, loss_mean_cls=0.0832, proj_loss=-0.395][2026-03-26 12:43:23] Step: 2964, Training Logs: loss_final: 0.545908, loss_mean: 0.864323, proj_loss: -0.399509, loss_mean_cls: 0.081094, grad_norm: 0.416027 +Steps: 0%| | 2965/1000000 [12:19<68:38:47, 4.03it/s, grad_norm=0.416, loss_final=0.546, loss_mean=0.864, loss_mean_cls=0.0811, proj_loss=-0.4][2026-03-26 12:43:23] Step: 2965, Training Logs: loss_final: 0.528255, loss_mean: 0.850252, proj_loss: -0.402460, loss_mean_cls: 0.080463, grad_norm: 0.513729 +Steps: 0%| | 2966/1000000 [12:20<68:38:39, 4.03it/s, grad_norm=0.514, loss_final=0.528, loss_mean=0.85, loss_mean_cls=0.0805, proj_loss=-0.402][2026-03-26 12:43:23] Step: 2966, Training Logs: loss_final: 0.517098, loss_mean: 0.832346, proj_loss: -0.398630, loss_mean_cls: 0.083382, grad_norm: 0.629588 +Steps: 0%| | 2967/1000000 [12:20<68:38:52, 4.03it/s, grad_norm=0.63, loss_final=0.517, loss_mean=0.832, loss_mean_cls=0.0834, proj_loss=-0.399][2026-03-26 12:43:23] Step: 2967, Training Logs: loss_final: 0.530943, loss_mean: 0.859547, proj_loss: -0.407673, loss_mean_cls: 0.079070, grad_norm: 0.370624 +Steps: 0%| | 2968/1000000 [12:20<68:35:19, 4.04it/s, grad_norm=0.371, loss_final=0.531, loss_mean=0.86, loss_mean_cls=0.0791, proj_loss=-0.408][2026-03-26 12:43:23] Step: 2968, Training Logs: loss_final: 0.500455, loss_mean: 0.823063, proj_loss: -0.403862, loss_mean_cls: 0.081255, grad_norm: 0.407614 +Steps: 0%| | 2969/1000000 [12:20<68:34:01, 4.04it/s, grad_norm=0.408, loss_final=0.5, loss_mean=0.823, loss_mean_cls=0.0813, proj_loss=-0.404][2026-03-26 12:43:24] Step: 2969, Training Logs: loss_final: 0.524639, loss_mean: 0.846536, proj_loss: -0.402405, loss_mean_cls: 0.080508, grad_norm: 0.483259 +Steps: 0%| | 2970/1000000 [12:21<68:32:34, 4.04it/s, grad_norm=0.483, loss_final=0.525, loss_mean=0.847, loss_mean_cls=0.0805, proj_loss=-0.402][2026-03-26 12:43:24] Step: 2970, Training Logs: loss_final: 0.537248, loss_mean: 0.853813, proj_loss: -0.398244, loss_mean_cls: 0.081678, grad_norm: 0.366494 +Steps: 0%| | 2971/1000000 [12:21<68:32:54, 4.04it/s, grad_norm=0.366, loss_final=0.537, loss_mean=0.854, loss_mean_cls=0.0817, proj_loss=-0.398][2026-03-26 12:43:24] Step: 2971, Training Logs: loss_final: 0.519514, loss_mean: 0.839110, proj_loss: -0.400731, loss_mean_cls: 0.081134, grad_norm: 0.344770 +Steps: 0%| | 2972/1000000 [12:21<68:31:11, 4.04it/s, grad_norm=0.345, loss_final=0.52, loss_mean=0.839, loss_mean_cls=0.0811, proj_loss=-0.401][2026-03-26 12:43:24] Step: 2972, Training Logs: loss_final: 0.517630, loss_mean: 0.844849, proj_loss: -0.407645, loss_mean_cls: 0.080427, grad_norm: 0.317395 +Steps: 0%| | 2973/1000000 [12:21<68:32:34, 4.04it/s, grad_norm=0.317, loss_final=0.518, loss_mean=0.845, loss_mean_cls=0.0804, proj_loss=-0.408][2026-03-26 12:43:25] Step: 2973, Training Logs: loss_final: 0.525117, loss_mean: 0.835911, proj_loss: -0.395106, loss_mean_cls: 0.084311, grad_norm: 0.645023 +Steps: 0%| | 2974/1000000 [12:22<68:30:23, 4.04it/s, grad_norm=0.645, loss_final=0.525, loss_mean=0.836, loss_mean_cls=0.0843, proj_loss=-0.395][2026-03-26 12:43:25] Step: 2974, Training Logs: loss_final: 0.519973, loss_mean: 0.842334, proj_loss: -0.403287, loss_mean_cls: 0.080927, grad_norm: 0.659159 +Steps: 0%| | 2975/1000000 [12:22<68:31:07, 4.04it/s, grad_norm=0.659, loss_final=0.52, loss_mean=0.842, loss_mean_cls=0.0809, proj_loss=-0.403][2026-03-26 12:43:25] Step: 2975, Training Logs: loss_final: 0.523403, loss_mean: 0.842867, proj_loss: -0.400484, loss_mean_cls: 0.081019, grad_norm: 0.383177 +Steps: 0%| | 2976/1000000 [12:22<69:17:11, 4.00it/s, grad_norm=0.383, loss_final=0.523, loss_mean=0.843, loss_mean_cls=0.081, proj_loss=-0.4][2026-03-26 12:43:25] Step: 2976, Training Logs: loss_final: 0.524206, loss_mean: 0.834550, proj_loss: -0.394314, loss_mean_cls: 0.083970, grad_norm: 0.429670 +Steps: 0%| | 2977/1000000 [12:22<69:04:51, 4.01it/s, grad_norm=0.43, loss_final=0.524, loss_mean=0.835, loss_mean_cls=0.084, proj_loss=-0.394][2026-03-26 12:43:26] Step: 2977, Training Logs: loss_final: 0.529904, loss_mean: 0.847397, proj_loss: -0.399683, loss_mean_cls: 0.082190, grad_norm: 0.515353 +Steps: 0%| | 2978/1000000 [12:23<68:55:19, 4.02it/s, grad_norm=0.515, loss_final=0.53, loss_mean=0.847, loss_mean_cls=0.0822, proj_loss=-0.4][2026-03-26 12:43:26] Step: 2978, Training Logs: loss_final: 0.528521, loss_mean: 0.856556, proj_loss: -0.407695, loss_mean_cls: 0.079660, grad_norm: 0.446632 +Steps: 0%| | 2979/1000000 [12:23<68:50:08, 4.02it/s, grad_norm=0.447, loss_final=0.529, loss_mean=0.857, loss_mean_cls=0.0797, proj_loss=-0.408][2026-03-26 12:43:26] Step: 2979, Training Logs: loss_final: 0.512054, loss_mean: 0.825197, proj_loss: -0.398367, loss_mean_cls: 0.085225, grad_norm: 0.575347 +Steps: 0%| | 2980/1000000 [12:23<68:42:07, 4.03it/s, grad_norm=0.575, loss_final=0.512, loss_mean=0.825, loss_mean_cls=0.0852, proj_loss=-0.398][2026-03-26 12:43:26] Step: 2980, Training Logs: loss_final: 0.505872, loss_mean: 0.826934, proj_loss: -0.402477, loss_mean_cls: 0.081415, grad_norm: 0.353587 +Steps: 0%| | 2981/1000000 [12:23<68:39:09, 4.03it/s, grad_norm=0.354, loss_final=0.506, loss_mean=0.827, loss_mean_cls=0.0814, proj_loss=-0.402][2026-03-26 12:43:27] Step: 2981, Training Logs: loss_final: 0.522041, loss_mean: 0.835629, proj_loss: -0.396936, loss_mean_cls: 0.083348, grad_norm: 0.458040 +Steps: 0%| | 2982/1000000 [12:24<68:39:37, 4.03it/s, grad_norm=0.458, loss_final=0.522, loss_mean=0.836, loss_mean_cls=0.0833, proj_loss=-0.397][2026-03-26 12:43:27] Step: 2982, Training Logs: loss_final: 0.521696, loss_mean: 0.833390, proj_loss: -0.395031, loss_mean_cls: 0.083338, grad_norm: 0.424562 +Steps: 0%| | 2983/1000000 [12:24<68:37:40, 4.04it/s, grad_norm=0.425, loss_final=0.522, loss_mean=0.833, loss_mean_cls=0.0833, proj_loss=-0.395][2026-03-26 12:43:27] Step: 2983, Training Logs: loss_final: 0.524027, loss_mean: 0.840901, proj_loss: -0.400005, loss_mean_cls: 0.083131, grad_norm: 0.323439 +Steps: 0%| | 2984/1000000 [12:24<68:34:46, 4.04it/s, grad_norm=0.323, loss_final=0.524, loss_mean=0.841, loss_mean_cls=0.0831, proj_loss=-0.4][2026-03-26 12:43:27] Step: 2984, Training Logs: loss_final: 0.509468, loss_mean: 0.839490, proj_loss: -0.409224, loss_mean_cls: 0.079202, grad_norm: 0.409690 +Steps: 0%| | 2985/1000000 [12:24<68:30:49, 4.04it/s, grad_norm=0.41, loss_final=0.509, loss_mean=0.839, loss_mean_cls=0.0792, proj_loss=-0.409][2026-03-26 12:43:28] Step: 2985, Training Logs: loss_final: 0.519263, loss_mean: 0.835028, proj_loss: -0.399127, loss_mean_cls: 0.083362, grad_norm: 0.547274 +Steps: 0%| | 2986/1000000 [12:25<68:31:30, 4.04it/s, grad_norm=0.547, loss_final=0.519, loss_mean=0.835, loss_mean_cls=0.0834, proj_loss=-0.399][2026-03-26 12:43:28] Step: 2986, Training Logs: loss_final: 0.515127, loss_mean: 0.837977, proj_loss: -0.404163, loss_mean_cls: 0.081312, grad_norm: 0.326896 +Steps: 0%| | 2987/1000000 [12:25<68:30:13, 4.04it/s, grad_norm=0.327, loss_final=0.515, loss_mean=0.838, loss_mean_cls=0.0813, proj_loss=-0.404][2026-03-26 12:43:28] Step: 2987, Training Logs: loss_final: 0.522582, loss_mean: 0.846383, proj_loss: -0.404418, loss_mean_cls: 0.080617, grad_norm: 0.274896 +Steps: 0%| | 2988/1000000 [12:25<68:31:53, 4.04it/s, grad_norm=0.275, loss_final=0.523, loss_mean=0.846, loss_mean_cls=0.0806, proj_loss=-0.404][2026-03-26 12:43:28] Step: 2988, Training Logs: loss_final: 0.514339, loss_mean: 0.832606, proj_loss: -0.399963, loss_mean_cls: 0.081696, grad_norm: 0.349482 +Steps: 0%| | 2989/1000000 [12:25<68:31:35, 4.04it/s, grad_norm=0.349, loss_final=0.514, loss_mean=0.833, loss_mean_cls=0.0817, proj_loss=-0.4][2026-03-26 12:43:29] Step: 2989, Training Logs: loss_final: 0.529301, loss_mean: 0.831687, proj_loss: -0.388077, loss_mean_cls: 0.085691, grad_norm: 0.354627 +Steps: 0%| | 2990/1000000 [12:26<68:30:23, 4.04it/s, grad_norm=0.355, loss_final=0.529, loss_mean=0.832, loss_mean_cls=0.0857, proj_loss=-0.388][2026-03-26 12:43:29] Step: 2990, Training Logs: loss_final: 0.519890, loss_mean: 0.834865, proj_loss: -0.398034, loss_mean_cls: 0.083058, grad_norm: 0.465948 +Steps: 0%| | 2991/1000000 [12:26<68:27:52, 4.05it/s, grad_norm=0.466, loss_final=0.52, loss_mean=0.835, loss_mean_cls=0.0831, proj_loss=-0.398][2026-03-26 12:43:29] Step: 2991, Training Logs: loss_final: 0.518001, loss_mean: 0.825511, proj_loss: -0.392990, loss_mean_cls: 0.085480, grad_norm: 0.511100 +Steps: 0%| | 2992/1000000 [12:26<68:34:57, 4.04it/s, grad_norm=0.511, loss_final=0.518, loss_mean=0.826, loss_mean_cls=0.0855, proj_loss=-0.393][2026-03-26 12:43:29] Step: 2992, Training Logs: loss_final: 0.539760, loss_mean: 0.850100, proj_loss: -0.393293, loss_mean_cls: 0.082954, grad_norm: 0.313026 +Steps: 0%| | 2993/1000000 [12:26<68:40:02, 4.03it/s, grad_norm=0.313, loss_final=0.54, loss_mean=0.85, loss_mean_cls=0.083, proj_loss=-0.393][2026-03-26 12:43:30] Step: 2993, Training Logs: loss_final: 0.516731, loss_mean: 0.839499, proj_loss: -0.403380, loss_mean_cls: 0.080612, grad_norm: 0.534535 +Steps: 0%| | 2994/1000000 [12:27<68:40:58, 4.03it/s, grad_norm=0.535, loss_final=0.517, loss_mean=0.839, loss_mean_cls=0.0806, proj_loss=-0.403][2026-03-26 12:43:30] Step: 2994, Training Logs: loss_final: 0.529726, loss_mean: 0.862692, proj_loss: -0.412298, loss_mean_cls: 0.079332, grad_norm: 0.428252 +Steps: 0%| | 2995/1000000 [12:27<68:35:26, 4.04it/s, grad_norm=0.428, loss_final=0.53, loss_mean=0.863, loss_mean_cls=0.0793, proj_loss=-0.412][2026-03-26 12:43:30] Step: 2995, Training Logs: loss_final: 0.530403, loss_mean: 0.844470, proj_loss: -0.398542, loss_mean_cls: 0.084475, grad_norm: 0.336102 +Steps: 0%| | 2996/1000000 [12:27<68:33:37, 4.04it/s, grad_norm=0.336, loss_final=0.53, loss_mean=0.844, loss_mean_cls=0.0845, proj_loss=-0.399][2026-03-26 12:43:30] Step: 2996, Training Logs: loss_final: 0.526787, loss_mean: 0.844703, proj_loss: -0.398275, loss_mean_cls: 0.080359, grad_norm: 0.533293 +Steps: 0%| | 2997/1000000 [12:27<68:29:24, 4.04it/s, grad_norm=0.533, loss_final=0.527, loss_mean=0.845, loss_mean_cls=0.0804, proj_loss=-0.398][2026-03-26 12:43:31] Step: 2997, Training Logs: loss_final: 0.548721, loss_mean: 0.877163, proj_loss: -0.406345, loss_mean_cls: 0.077902, grad_norm: 0.699240 +Steps: 0%| | 2998/1000000 [12:28<68:33:27, 4.04it/s, grad_norm=0.699, loss_final=0.549, loss_mean=0.877, loss_mean_cls=0.0779, proj_loss=-0.406][2026-03-26 12:43:31] Step: 2998, Training Logs: loss_final: 0.533815, loss_mean: 0.854109, proj_loss: -0.401318, loss_mean_cls: 0.081024, grad_norm: 0.351163 +Steps: 0%| | 2999/1000000 [12:28<68:31:43, 4.04it/s, grad_norm=0.351, loss_final=0.534, loss_mean=0.854, loss_mean_cls=0.081, proj_loss=-0.401][2026-03-26 12:43:31] Step: 2999, Training Logs: loss_final: 0.512403, loss_mean: 0.827412, proj_loss: -0.398187, loss_mean_cls: 0.083177, grad_norm: 0.392840 +Steps: 0%| | 3000/1000000 [12:28<68:33:06, 4.04it/s, grad_norm=0.393, loss_final=0.512, loss_mean=0.827, loss_mean_cls=0.0832, proj_loss=-0.398][2026-03-26 12:43:31] Step: 3000, Training Logs: loss_final: 0.545024, loss_mean: 0.872794, proj_loss: -0.406706, loss_mean_cls: 0.078936, grad_norm: 0.479418 +Steps: 0%| | 3001/1000000 [12:28<68:35:28, 4.04it/s, grad_norm=0.479, loss_final=0.545, loss_mean=0.873, loss_mean_cls=0.0789, proj_loss=-0.407][2026-03-26 12:43:32] Step: 3001, Training Logs: loss_final: 0.525194, loss_mean: 0.839840, proj_loss: -0.398190, loss_mean_cls: 0.083545, grad_norm: 0.326844 +Steps: 0%| | 3002/1000000 [12:29<68:37:40, 4.04it/s, grad_norm=0.327, loss_final=0.525, loss_mean=0.84, loss_mean_cls=0.0835, proj_loss=-0.398][2026-03-26 12:43:32] Step: 3002, Training Logs: loss_final: 0.517169, loss_mean: 0.839708, proj_loss: -0.404077, loss_mean_cls: 0.081538, grad_norm: 0.463101 +Steps: 0%| | 3003/1000000 [12:29<68:37:51, 4.04it/s, grad_norm=0.463, loss_final=0.517, loss_mean=0.84, loss_mean_cls=0.0815, proj_loss=-0.404][2026-03-26 12:43:32] Step: 3003, Training Logs: loss_final: 0.547005, loss_mean: 0.861845, proj_loss: -0.396904, loss_mean_cls: 0.082064, grad_norm: 0.392529 +Steps: 0%| | 3004/1000000 [12:29<68:37:04, 4.04it/s, grad_norm=0.393, loss_final=0.547, loss_mean=0.862, loss_mean_cls=0.0821, proj_loss=-0.397][2026-03-26 12:43:32] Step: 3004, Training Logs: loss_final: 0.525136, loss_mean: 0.830109, proj_loss: -0.389719, loss_mean_cls: 0.084746, grad_norm: 0.328533 +Steps: 0%| | 3005/1000000 [12:29<68:37:37, 4.04it/s, grad_norm=0.329, loss_final=0.525, loss_mean=0.83, loss_mean_cls=0.0847, proj_loss=-0.39][2026-03-26 12:43:33] Step: 3005, Training Logs: loss_final: 0.527543, loss_mean: 0.833039, proj_loss: -0.390291, loss_mean_cls: 0.084796, grad_norm: 0.587783 +Steps: 0%| | 3006/1000000 [12:30<68:37:10, 4.04it/s, grad_norm=0.588, loss_final=0.528, loss_mean=0.833, loss_mean_cls=0.0848, proj_loss=-0.39][2026-03-26 12:43:33] Step: 3006, Training Logs: loss_final: 0.503111, loss_mean: 0.819242, proj_loss: -0.398971, loss_mean_cls: 0.082841, grad_norm: 0.567357 +Steps: 0%| | 3007/1000000 [12:30<68:34:44, 4.04it/s, grad_norm=0.567, loss_final=0.503, loss_mean=0.819, loss_mean_cls=0.0828, proj_loss=-0.399][2026-03-26 12:43:33] Step: 3007, Training Logs: loss_final: 0.520538, loss_mean: 0.843586, proj_loss: -0.405534, loss_mean_cls: 0.082486, grad_norm: 0.388995 +Steps: 0%| | 3008/1000000 [12:30<68:36:54, 4.04it/s, grad_norm=0.389, loss_final=0.521, loss_mean=0.844, loss_mean_cls=0.0825, proj_loss=-0.406][2026-03-26 12:43:33] Step: 3008, Training Logs: loss_final: 0.527095, loss_mean: 0.857586, proj_loss: -0.409682, loss_mean_cls: 0.079192, grad_norm: 0.452634 +Steps: 0%| | 3009/1000000 [12:30<68:36:11, 4.04it/s, grad_norm=0.453, loss_final=0.527, loss_mean=0.858, loss_mean_cls=0.0792, proj_loss=-0.41][2026-03-26 12:43:34] Step: 3009, Training Logs: loss_final: 0.499196, loss_mean: 0.814206, proj_loss: -0.399311, loss_mean_cls: 0.084301, grad_norm: 0.342638 +Steps: 0%| | 3010/1000000 [12:31<68:35:58, 4.04it/s, grad_norm=0.343, loss_final=0.499, loss_mean=0.814, loss_mean_cls=0.0843, proj_loss=-0.399][2026-03-26 12:43:34] Step: 3010, Training Logs: loss_final: 0.519881, loss_mean: 0.831426, proj_loss: -0.394650, loss_mean_cls: 0.083105, grad_norm: 0.351345 +Steps: 0%| | 3011/1000000 [12:31<68:37:16, 4.04it/s, grad_norm=0.351, loss_final=0.52, loss_mean=0.831, loss_mean_cls=0.0831, proj_loss=-0.395][2026-03-26 12:43:34] Step: 3011, Training Logs: loss_final: 0.542141, loss_mean: 0.859386, proj_loss: -0.399137, loss_mean_cls: 0.081892, grad_norm: 0.313600 +Steps: 0%| | 3012/1000000 [12:31<68:32:19, 4.04it/s, grad_norm=0.314, loss_final=0.542, loss_mean=0.859, loss_mean_cls=0.0819, proj_loss=-0.399][2026-03-26 12:43:34] Step: 3012, Training Logs: loss_final: 0.523620, loss_mean: 0.834282, proj_loss: -0.395536, loss_mean_cls: 0.084875, grad_norm: 0.380749 +Steps: 0%| | 3013/1000000 [12:31<68:36:19, 4.04it/s, grad_norm=0.381, loss_final=0.524, loss_mean=0.834, loss_mean_cls=0.0849, proj_loss=-0.396][2026-03-26 12:43:35] Step: 3013, Training Logs: loss_final: 0.528930, loss_mean: 0.853714, proj_loss: -0.405362, loss_mean_cls: 0.080578, grad_norm: 0.377250 +Steps: 0%| | 3014/1000000 [12:31<68:34:52, 4.04it/s, grad_norm=0.377, loss_final=0.529, loss_mean=0.854, loss_mean_cls=0.0806, proj_loss=-0.405][2026-03-26 12:43:35] Step: 3014, Training Logs: loss_final: 0.507843, loss_mean: 0.824290, proj_loss: -0.400275, loss_mean_cls: 0.083828, grad_norm: 0.345794 +Steps: 0%| | 3015/1000000 [12:32<68:36:00, 4.04it/s, grad_norm=0.346, loss_final=0.508, loss_mean=0.824, loss_mean_cls=0.0838, proj_loss=-0.4][2026-03-26 12:43:35] Step: 3015, Training Logs: loss_final: 0.501244, loss_mean: 0.814014, proj_loss: -0.398405, loss_mean_cls: 0.085635, grad_norm: 0.349184 +Steps: 0%| | 3016/1000000 [12:32<68:34:04, 4.04it/s, grad_norm=0.349, loss_final=0.501, loss_mean=0.814, loss_mean_cls=0.0856, proj_loss=-0.398][2026-03-26 12:43:35] Step: 3016, Training Logs: loss_final: 0.518834, loss_mean: 0.844078, proj_loss: -0.406685, loss_mean_cls: 0.081440, grad_norm: 0.411961 +Steps: 0%| | 3017/1000000 [12:32<68:34:12, 4.04it/s, grad_norm=0.412, loss_final=0.519, loss_mean=0.844, loss_mean_cls=0.0814, proj_loss=-0.407][2026-03-26 12:43:36] Step: 3017, Training Logs: loss_final: 0.528125, loss_mean: 0.859035, proj_loss: -0.410302, loss_mean_cls: 0.079393, grad_norm: 0.407379 +Steps: 0%| | 3018/1000000 [12:32<68:32:12, 4.04it/s, grad_norm=0.407, loss_final=0.528, loss_mean=0.859, loss_mean_cls=0.0794, proj_loss=-0.41][2026-03-26 12:43:36] Step: 3018, Training Logs: loss_final: 0.519824, loss_mean: 0.832274, proj_loss: -0.395787, loss_mean_cls: 0.083336, grad_norm: 0.368429 +Steps: 0%| | 3019/1000000 [12:33<68:36:07, 4.04it/s, grad_norm=0.368, loss_final=0.52, loss_mean=0.832, loss_mean_cls=0.0833, proj_loss=-0.396][2026-03-26 12:43:36] Step: 3019, Training Logs: loss_final: 0.505997, loss_mean: 0.826528, proj_loss: -0.401404, loss_mean_cls: 0.080873, grad_norm: 0.382752 +Steps: 0%| | 3020/1000000 [12:33<68:35:34, 4.04it/s, grad_norm=0.383, loss_final=0.506, loss_mean=0.827, loss_mean_cls=0.0809, proj_loss=-0.401][2026-03-26 12:43:36] Step: 3020, Training Logs: loss_final: 0.536948, loss_mean: 0.860114, proj_loss: -0.403863, loss_mean_cls: 0.080697, grad_norm: 0.331245 +Steps: 0%| | 3021/1000000 [12:33<68:33:36, 4.04it/s, grad_norm=0.331, loss_final=0.537, loss_mean=0.86, loss_mean_cls=0.0807, proj_loss=-0.404][2026-03-26 12:43:37] Step: 3021, Training Logs: loss_final: 0.536717, loss_mean: 0.845389, proj_loss: -0.392620, loss_mean_cls: 0.083948, grad_norm: 0.447271 +Steps: 0%| | 3022/1000000 [12:33<68:31:31, 4.04it/s, grad_norm=0.447, loss_final=0.537, loss_mean=0.845, loss_mean_cls=0.0839, proj_loss=-0.393][2026-03-26 12:43:37] Step: 3022, Training Logs: loss_final: 0.518471, loss_mean: 0.840002, proj_loss: -0.402561, loss_mean_cls: 0.081030, grad_norm: 0.377736 +Steps: 0%| | 3023/1000000 [12:34<68:33:16, 4.04it/s, grad_norm=0.378, loss_final=0.518, loss_mean=0.84, loss_mean_cls=0.081, proj_loss=-0.403][2026-03-26 12:43:37] Step: 3023, Training Logs: loss_final: 0.519856, loss_mean: 0.835941, proj_loss: -0.398271, loss_mean_cls: 0.082185, grad_norm: 0.416227 +Steps: 0%| | 3024/1000000 [12:34<68:31:35, 4.04it/s, grad_norm=0.416, loss_final=0.52, loss_mean=0.836, loss_mean_cls=0.0822, proj_loss=-0.398][2026-03-26 12:43:37] Step: 3024, Training Logs: loss_final: 0.538141, loss_mean: 0.860142, proj_loss: -0.404285, loss_mean_cls: 0.082284, grad_norm: 0.644480 +Steps: 0%| | 3025/1000000 [12:34<68:28:48, 4.04it/s, grad_norm=0.644, loss_final=0.538, loss_mean=0.86, loss_mean_cls=0.0823, proj_loss=-0.404][2026-03-26 12:43:38] Step: 3025, Training Logs: loss_final: 0.527064, loss_mean: 0.841479, proj_loss: -0.397416, loss_mean_cls: 0.083000, grad_norm: 0.535432 +Steps: 0%| | 3026/1000000 [12:34<68:32:38, 4.04it/s, grad_norm=0.535, loss_final=0.527, loss_mean=0.841, loss_mean_cls=0.083, proj_loss=-0.397][2026-03-26 12:43:38] Step: 3026, Training Logs: loss_final: 0.533323, loss_mean: 0.855165, proj_loss: -0.403277, loss_mean_cls: 0.081435, grad_norm: 0.702549 +Steps: 0%| | 3027/1000000 [12:35<68:30:36, 4.04it/s, grad_norm=0.703, loss_final=0.533, loss_mean=0.855, loss_mean_cls=0.0814, proj_loss=-0.403][2026-03-26 12:43:38] Step: 3027, Training Logs: loss_final: 0.537528, loss_mean: 0.859240, proj_loss: -0.402177, loss_mean_cls: 0.080465, grad_norm: 0.429882 +Steps: 0%| | 3028/1000000 [12:35<68:29:07, 4.04it/s, grad_norm=0.43, loss_final=0.538, loss_mean=0.859, loss_mean_cls=0.0805, proj_loss=-0.402][2026-03-26 12:43:38] Step: 3028, Training Logs: loss_final: 0.529473, loss_mean: 0.849484, proj_loss: -0.400681, loss_mean_cls: 0.080670, grad_norm: 0.583720 +Steps: 0%| | 3029/1000000 [12:35<68:28:31, 4.04it/s, grad_norm=0.584, loss_final=0.529, loss_mean=0.849, loss_mean_cls=0.0807, proj_loss=-0.401][2026-03-26 12:43:39] Step: 3029, Training Logs: loss_final: 0.523623, loss_mean: 0.832160, proj_loss: -0.390700, loss_mean_cls: 0.082163, grad_norm: 0.491219 +Steps: 0%| | 3030/1000000 [12:35<68:28:50, 4.04it/s, grad_norm=0.491, loss_final=0.524, loss_mean=0.832, loss_mean_cls=0.0822, proj_loss=-0.391][2026-03-26 12:43:39] Step: 3030, Training Logs: loss_final: 0.528446, loss_mean: 0.853992, proj_loss: -0.404856, loss_mean_cls: 0.079311, grad_norm: 0.529245 +Steps: 0%| | 3031/1000000 [12:36<68:32:54, 4.04it/s, grad_norm=0.529, loss_final=0.528, loss_mean=0.854, loss_mean_cls=0.0793, proj_loss=-0.405][2026-03-26 12:43:39] Step: 3031, Training Logs: loss_final: 0.534600, loss_mean: 0.842826, proj_loss: -0.391501, loss_mean_cls: 0.083275, grad_norm: 0.494695 +Steps: 0%| | 3032/1000000 [12:36<68:37:14, 4.04it/s, grad_norm=0.495, loss_final=0.535, loss_mean=0.843, loss_mean_cls=0.0833, proj_loss=-0.392][2026-03-26 12:43:39] Step: 3032, Training Logs: loss_final: 0.523257, loss_mean: 0.837932, proj_loss: -0.396920, loss_mean_cls: 0.082245, grad_norm: 0.416851 +Steps: 0%| | 3033/1000000 [12:36<68:36:49, 4.04it/s, grad_norm=0.417, loss_final=0.523, loss_mean=0.838, loss_mean_cls=0.0822, proj_loss=-0.397][2026-03-26 12:43:40] Step: 3033, Training Logs: loss_final: 0.518816, loss_mean: 0.841521, proj_loss: -0.405018, loss_mean_cls: 0.082313, grad_norm: 0.371893 +Steps: 0%| | 3034/1000000 [12:36<68:37:19, 4.04it/s, grad_norm=0.372, loss_final=0.519, loss_mean=0.842, loss_mean_cls=0.0823, proj_loss=-0.405][2026-03-26 12:43:40] Step: 3034, Training Logs: loss_final: 0.515861, loss_mean: 0.832105, proj_loss: -0.398364, loss_mean_cls: 0.082120, grad_norm: 0.375328 +Steps: 0%| | 3035/1000000 [12:37<68:34:45, 4.04it/s, grad_norm=0.375, loss_final=0.516, loss_mean=0.832, loss_mean_cls=0.0821, proj_loss=-0.398][2026-03-26 12:43:40] Step: 3035, Training Logs: loss_final: 0.513520, loss_mean: 0.839566, proj_loss: -0.405205, loss_mean_cls: 0.079158, grad_norm: 0.445186 +Steps: 0%| | 3036/1000000 [12:37<68:36:14, 4.04it/s, grad_norm=0.445, loss_final=0.514, loss_mean=0.84, loss_mean_cls=0.0792, proj_loss=-0.405][2026-03-26 12:43:40] Step: 3036, Training Logs: loss_final: 0.515003, loss_mean: 0.829869, proj_loss: -0.398912, loss_mean_cls: 0.084045, grad_norm: 0.382581 +Steps: 0%| | 3037/1000000 [12:37<68:40:52, 4.03it/s, grad_norm=0.383, loss_final=0.515, loss_mean=0.83, loss_mean_cls=0.084, proj_loss=-0.399][2026-03-26 12:43:41] Step: 3037, Training Logs: loss_final: 0.508701, loss_mean: 0.827154, proj_loss: -0.401345, loss_mean_cls: 0.082891, grad_norm: 0.363348 +Steps: 0%| | 3038/1000000 [12:37<68:37:07, 4.04it/s, grad_norm=0.363, loss_final=0.509, loss_mean=0.827, loss_mean_cls=0.0829, proj_loss=-0.401][2026-03-26 12:43:41] Step: 3038, Training Logs: loss_final: 0.505534, loss_mean: 0.834715, proj_loss: -0.410338, loss_mean_cls: 0.081157, grad_norm: 0.504455 +Steps: 0%| | 3039/1000000 [12:38<68:34:21, 4.04it/s, grad_norm=0.504, loss_final=0.506, loss_mean=0.835, loss_mean_cls=0.0812, proj_loss=-0.41][2026-03-26 12:43:41] Step: 3039, Training Logs: loss_final: 0.501981, loss_mean: 0.812609, proj_loss: -0.396055, loss_mean_cls: 0.085427, grad_norm: 0.491417 +Steps: 0%| | 3040/1000000 [12:38<68:34:37, 4.04it/s, grad_norm=0.491, loss_final=0.502, loss_mean=0.813, loss_mean_cls=0.0854, proj_loss=-0.396][2026-03-26 12:43:41] Step: 3040, Training Logs: loss_final: 0.516582, loss_mean: 0.839291, proj_loss: -0.404132, loss_mean_cls: 0.081423, grad_norm: 0.458869 +Steps: 0%| | 3041/1000000 [12:38<68:32:28, 4.04it/s, grad_norm=0.459, loss_final=0.517, loss_mean=0.839, loss_mean_cls=0.0814, proj_loss=-0.404][2026-03-26 12:43:42] Step: 3041, Training Logs: loss_final: 0.528785, loss_mean: 0.860291, proj_loss: -0.408641, loss_mean_cls: 0.077135, grad_norm: 0.528300 +Steps: 0%| | 3042/1000000 [12:38<68:44:06, 4.03it/s, grad_norm=0.528, loss_final=0.529, loss_mean=0.86, loss_mean_cls=0.0771, proj_loss=-0.409][2026-03-26 12:43:42] Step: 3042, Training Logs: loss_final: 0.518283, loss_mean: 0.840175, proj_loss: -0.401905, loss_mean_cls: 0.080013, grad_norm: 0.405400 +Steps: 0%| | 3043/1000000 [12:39<69:45:43, 3.97it/s, grad_norm=0.405, loss_final=0.518, loss_mean=0.84, loss_mean_cls=0.08, proj_loss=-0.402][2026-03-26 12:43:42] Step: 3043, Training Logs: loss_final: 0.516592, loss_mean: 0.840485, proj_loss: -0.404539, loss_mean_cls: 0.080646, grad_norm: 0.432453 +Steps: 0%| | 3044/1000000 [12:39<69:22:45, 3.99it/s, grad_norm=0.432, loss_final=0.517, loss_mean=0.84, loss_mean_cls=0.0806, proj_loss=-0.405][2026-03-26 12:43:42] Step: 3044, Training Logs: loss_final: 0.538226, loss_mean: 0.862642, proj_loss: -0.404988, loss_mean_cls: 0.080573, grad_norm: 0.464361 +Steps: 0%| | 3045/1000000 [12:39<69:01:28, 4.01it/s, grad_norm=0.464, loss_final=0.538, loss_mean=0.863, loss_mean_cls=0.0806, proj_loss=-0.405][2026-03-26 12:43:43] Step: 3045, Training Logs: loss_final: 0.511696, loss_mean: 0.833816, proj_loss: -0.403526, loss_mean_cls: 0.081406, grad_norm: 0.716033 +Steps: 0%| | 3046/1000000 [12:39<69:25:40, 3.99it/s, grad_norm=0.716, loss_final=0.512, loss_mean=0.834, loss_mean_cls=0.0814, proj_loss=-0.404][2026-03-26 12:43:43] Step: 3046, Training Logs: loss_final: 0.533796, loss_mean: 0.855010, proj_loss: -0.402703, loss_mean_cls: 0.081489, grad_norm: 0.450922 +Steps: 0%| | 3047/1000000 [12:40<71:57:56, 3.85it/s, grad_norm=0.451, loss_final=0.534, loss_mean=0.855, loss_mean_cls=0.0815, proj_loss=-0.403][2026-03-26 12:43:43] Step: 3047, Training Logs: loss_final: 0.527457, loss_mean: 0.844503, proj_loss: -0.401484, loss_mean_cls: 0.084437, grad_norm: 0.328647 +Steps: 0%| | 3048/1000000 [12:40<70:50:59, 3.91it/s, grad_norm=0.329, loss_final=0.527, loss_mean=0.845, loss_mean_cls=0.0844, proj_loss=-0.401][2026-03-26 12:43:43] Step: 3048, Training Logs: loss_final: 0.540041, loss_mean: 0.867810, proj_loss: -0.406415, loss_mean_cls: 0.078646, grad_norm: 0.545048 +Steps: 0%| | 3049/1000000 [12:40<70:06:51, 3.95it/s, grad_norm=0.545, loss_final=0.54, loss_mean=0.868, loss_mean_cls=0.0786, proj_loss=-0.406][2026-03-26 12:43:44] Step: 3049, Training Logs: loss_final: 0.533577, loss_mean: 0.850487, proj_loss: -0.398922, loss_mean_cls: 0.082012, grad_norm: 0.331401 +Steps: 0%| | 3050/1000000 [12:40<69:38:40, 3.98it/s, grad_norm=0.331, loss_final=0.534, loss_mean=0.85, loss_mean_cls=0.082, proj_loss=-0.399][2026-03-26 12:43:44] Step: 3050, Training Logs: loss_final: 0.520785, loss_mean: 0.854006, proj_loss: -0.412507, loss_mean_cls: 0.079287, grad_norm: 0.433553 +Steps: 0%| | 3051/1000000 [12:41<69:20:26, 3.99it/s, grad_norm=0.434, loss_final=0.521, loss_mean=0.854, loss_mean_cls=0.0793, proj_loss=-0.413][2026-03-26 12:43:44] Step: 3051, Training Logs: loss_final: 0.522707, loss_mean: 0.846969, proj_loss: -0.404634, loss_mean_cls: 0.080371, grad_norm: 0.612084 +Steps: 0%| | 3052/1000000 [12:41<69:06:42, 4.01it/s, grad_norm=0.612, loss_final=0.523, loss_mean=0.847, loss_mean_cls=0.0804, proj_loss=-0.405][2026-03-26 12:43:44] Step: 3052, Training Logs: loss_final: 0.519874, loss_mean: 0.845037, proj_loss: -0.404806, loss_mean_cls: 0.079643, grad_norm: 0.323077 +Steps: 0%| | 3053/1000000 [12:41<68:55:28, 4.02it/s, grad_norm=0.323, loss_final=0.52, loss_mean=0.845, loss_mean_cls=0.0796, proj_loss=-0.405][2026-03-26 12:43:45] Step: 3053, Training Logs: loss_final: 0.530127, loss_mean: 0.850615, proj_loss: -0.401580, loss_mean_cls: 0.081092, grad_norm: 0.488958 +Steps: 0%| | 3054/1000000 [12:41<68:49:15, 4.02it/s, grad_norm=0.489, loss_final=0.53, loss_mean=0.851, loss_mean_cls=0.0811, proj_loss=-0.402][2026-03-26 12:43:45] Step: 3054, Training Logs: loss_final: 0.510767, loss_mean: 0.839093, proj_loss: -0.408516, loss_mean_cls: 0.080190, grad_norm: 0.544576 +Steps: 0%| | 3055/1000000 [12:42<68:44:09, 4.03it/s, grad_norm=0.545, loss_final=0.511, loss_mean=0.839, loss_mean_cls=0.0802, proj_loss=-0.409][2026-03-26 12:43:45] Step: 3055, Training Logs: loss_final: 0.531828, loss_mean: 0.861788, proj_loss: -0.409270, loss_mean_cls: 0.079309, grad_norm: 0.295777 +Steps: 0%| | 3056/1000000 [12:42<68:39:39, 4.03it/s, grad_norm=0.296, loss_final=0.532, loss_mean=0.862, loss_mean_cls=0.0793, proj_loss=-0.409][2026-03-26 12:43:45] Step: 3056, Training Logs: loss_final: 0.496144, loss_mean: 0.807756, proj_loss: -0.396680, loss_mean_cls: 0.085068, grad_norm: 0.409713 +Steps: 0%| | 3057/1000000 [12:42<68:36:09, 4.04it/s, grad_norm=0.41, loss_final=0.496, loss_mean=0.808, loss_mean_cls=0.0851, proj_loss=-0.397][2026-03-26 12:43:46] Step: 3057, Training Logs: loss_final: 0.523969, loss_mean: 0.838496, proj_loss: -0.397676, loss_mean_cls: 0.083150, grad_norm: 0.438095 +Steps: 0%| | 3058/1000000 [12:42<68:35:40, 4.04it/s, grad_norm=0.438, loss_final=0.524, loss_mean=0.838, loss_mean_cls=0.0831, proj_loss=-0.398][2026-03-26 12:43:46] Step: 3058, Training Logs: loss_final: 0.518849, loss_mean: 0.841967, proj_loss: -0.404174, loss_mean_cls: 0.081057, grad_norm: 0.337220 +Steps: 0%| | 3059/1000000 [12:43<68:29:22, 4.04it/s, grad_norm=0.337, loss_final=0.519, loss_mean=0.842, loss_mean_cls=0.0811, proj_loss=-0.404][2026-03-26 12:43:46] Step: 3059, Training Logs: loss_final: 0.522754, loss_mean: 0.848051, proj_loss: -0.405916, loss_mean_cls: 0.080620, grad_norm: 0.603921 +Steps: 0%| | 3060/1000000 [12:43<68:30:55, 4.04it/s, grad_norm=0.604, loss_final=0.523, loss_mean=0.848, loss_mean_cls=0.0806, proj_loss=-0.406][2026-03-26 12:43:46] Step: 3060, Training Logs: loss_final: 0.506791, loss_mean: 0.821044, proj_loss: -0.398552, loss_mean_cls: 0.084300, grad_norm: 0.472423 +Steps: 0%| | 3061/1000000 [12:43<68:32:06, 4.04it/s, grad_norm=0.472, loss_final=0.507, loss_mean=0.821, loss_mean_cls=0.0843, proj_loss=-0.399][2026-03-26 12:43:47] Step: 3061, Training Logs: loss_final: 0.514286, loss_mean: 0.835783, proj_loss: -0.404251, loss_mean_cls: 0.082754, grad_norm: 0.382719 +Steps: 0%| | 3062/1000000 [12:43<68:30:08, 4.04it/s, grad_norm=0.383, loss_final=0.514, loss_mean=0.836, loss_mean_cls=0.0828, proj_loss=-0.404][2026-03-26 12:43:47] Step: 3062, Training Logs: loss_final: 0.532096, loss_mean: 0.861551, proj_loss: -0.409396, loss_mean_cls: 0.079941, grad_norm: 0.517479 +Steps: 0%| | 3063/1000000 [12:44<68:27:01, 4.05it/s, grad_norm=0.517, loss_final=0.532, loss_mean=0.862, loss_mean_cls=0.0799, proj_loss=-0.409][2026-03-26 12:43:47] Step: 3063, Training Logs: loss_final: 0.531102, loss_mean: 0.861894, proj_loss: -0.409417, loss_mean_cls: 0.078625, grad_norm: 0.575399 +Steps: 0%| | 3064/1000000 [12:44<68:29:28, 4.04it/s, grad_norm=0.575, loss_final=0.531, loss_mean=0.862, loss_mean_cls=0.0786, proj_loss=-0.409][2026-03-26 12:43:47] Step: 3064, Training Logs: loss_final: 0.520546, loss_mean: 0.841821, proj_loss: -0.403281, loss_mean_cls: 0.082007, grad_norm: 0.352083 +Steps: 0%| | 3065/1000000 [12:44<68:31:47, 4.04it/s, grad_norm=0.352, loss_final=0.521, loss_mean=0.842, loss_mean_cls=0.082, proj_loss=-0.403][2026-03-26 12:43:48] Step: 3065, Training Logs: loss_final: 0.525243, loss_mean: 0.852598, proj_loss: -0.407669, loss_mean_cls: 0.080315, grad_norm: 0.542428 +Steps: 0%| | 3066/1000000 [12:44<68:30:45, 4.04it/s, grad_norm=0.542, loss_final=0.525, loss_mean=0.853, loss_mean_cls=0.0803, proj_loss=-0.408][2026-03-26 12:43:48] Step: 3066, Training Logs: loss_final: 0.529680, loss_mean: 0.849377, proj_loss: -0.402111, loss_mean_cls: 0.082414, grad_norm: 0.408508 +Steps: 0%| | 3067/1000000 [12:45<68:30:34, 4.04it/s, grad_norm=0.409, loss_final=0.53, loss_mean=0.849, loss_mean_cls=0.0824, proj_loss=-0.402][2026-03-26 12:43:48] Step: 3067, Training Logs: loss_final: 0.523263, loss_mean: 0.860883, proj_loss: -0.415450, loss_mean_cls: 0.077831, grad_norm: 0.400844 +Steps: 0%| | 3068/1000000 [12:45<68:31:14, 4.04it/s, grad_norm=0.401, loss_final=0.523, loss_mean=0.861, loss_mean_cls=0.0778, proj_loss=-0.415][2026-03-26 12:43:48] Step: 3068, Training Logs: loss_final: 0.510983, loss_mean: 0.842072, proj_loss: -0.411982, loss_mean_cls: 0.080893, grad_norm: 0.382346 +Steps: 0%| | 3069/1000000 [12:45<68:30:23, 4.04it/s, grad_norm=0.382, loss_final=0.511, loss_mean=0.842, loss_mean_cls=0.0809, proj_loss=-0.412][2026-03-26 12:43:49] Step: 3069, Training Logs: loss_final: 0.532939, loss_mean: 0.851269, proj_loss: -0.400924, loss_mean_cls: 0.082595, grad_norm: 0.392097 +Steps: 0%| | 3070/1000000 [12:45<68:29:57, 4.04it/s, grad_norm=0.392, loss_final=0.533, loss_mean=0.851, loss_mean_cls=0.0826, proj_loss=-0.401][2026-03-26 12:43:49] Step: 3070, Training Logs: loss_final: 0.504110, loss_mean: 0.819996, proj_loss: -0.399500, loss_mean_cls: 0.083614, grad_norm: 0.434978 +Steps: 0%| | 3071/1000000 [12:46<68:32:16, 4.04it/s, grad_norm=0.435, loss_final=0.504, loss_mean=0.82, loss_mean_cls=0.0836, proj_loss=-0.4][2026-03-26 12:43:49] Step: 3071, Training Logs: loss_final: 0.518586, loss_mean: 0.853885, proj_loss: -0.412698, loss_mean_cls: 0.077398, grad_norm: 0.411600 +Steps: 0%| | 3072/1000000 [12:46<68:31:35, 4.04it/s, grad_norm=0.412, loss_final=0.519, loss_mean=0.854, loss_mean_cls=0.0774, proj_loss=-0.413][2026-03-26 12:43:49] Step: 3072, Training Logs: loss_final: 0.523031, loss_mean: 0.847433, proj_loss: -0.403271, loss_mean_cls: 0.078868, grad_norm: 0.319353 +Steps: 0%| | 3073/1000000 [12:46<68:58:12, 4.02it/s, grad_norm=0.319, loss_final=0.523, loss_mean=0.847, loss_mean_cls=0.0789, proj_loss=-0.403][2026-03-26 12:43:50] Step: 3073, Training Logs: loss_final: 0.532011, loss_mean: 0.872079, proj_loss: -0.417313, loss_mean_cls: 0.077244, grad_norm: 0.272434 +Steps: 0%| | 3074/1000000 [12:46<68:33:15, 4.04it/s, grad_norm=0.272, loss_final=0.532, loss_mean=0.872, loss_mean_cls=0.0772, proj_loss=-0.417][2026-03-26 12:43:50] Step: 3074, Training Logs: loss_final: 0.521796, loss_mean: 0.833232, proj_loss: -0.396607, loss_mean_cls: 0.085171, grad_norm: 0.435768 +Steps: 0%| | 3075/1000000 [12:47<68:33:38, 4.04it/s, grad_norm=0.436, loss_final=0.522, loss_mean=0.833, loss_mean_cls=0.0852, proj_loss=-0.397][2026-03-26 12:43:50] Step: 3075, Training Logs: loss_final: 0.510540, loss_mean: 0.834705, proj_loss: -0.405090, loss_mean_cls: 0.080926, grad_norm: 0.319035 +Steps: 0%| | 3076/1000000 [12:47<68:29:55, 4.04it/s, grad_norm=0.319, loss_final=0.511, loss_mean=0.835, loss_mean_cls=0.0809, proj_loss=-0.405][2026-03-26 12:43:50] Step: 3076, Training Logs: loss_final: 0.515522, loss_mean: 0.827771, proj_loss: -0.396094, loss_mean_cls: 0.083845, grad_norm: 0.317744 +Steps: 0%| | 3077/1000000 [12:47<68:31:31, 4.04it/s, grad_norm=0.318, loss_final=0.516, loss_mean=0.828, loss_mean_cls=0.0838, proj_loss=-0.396][2026-03-26 12:43:51] Step: 3077, Training Logs: loss_final: 0.527623, loss_mean: 0.854301, proj_loss: -0.407532, loss_mean_cls: 0.080855, grad_norm: 0.298373 +Steps: 0%| | 3078/1000000 [12:47<68:30:03, 4.04it/s, grad_norm=0.298, loss_final=0.528, loss_mean=0.854, loss_mean_cls=0.0809, proj_loss=-0.408][2026-03-26 12:43:51] Step: 3078, Training Logs: loss_final: 0.516396, loss_mean: 0.835686, proj_loss: -0.402363, loss_mean_cls: 0.083072, grad_norm: 0.379382 +Steps: 0%| | 3079/1000000 [12:48<68:29:46, 4.04it/s, grad_norm=0.379, loss_final=0.516, loss_mean=0.836, loss_mean_cls=0.0831, proj_loss=-0.402][2026-03-26 12:43:51] Step: 3079, Training Logs: loss_final: 0.526096, loss_mean: 0.840180, proj_loss: -0.398891, loss_mean_cls: 0.084806, grad_norm: 0.687400 +Steps: 0%| | 3080/1000000 [12:48<68:29:26, 4.04it/s, grad_norm=0.687, loss_final=0.526, loss_mean=0.84, loss_mean_cls=0.0848, proj_loss=-0.399][2026-03-26 12:43:51] Step: 3080, Training Logs: loss_final: 0.521708, loss_mean: 0.847121, proj_loss: -0.404363, loss_mean_cls: 0.078950, grad_norm: 0.535532 +Steps: 0%| | 3081/1000000 [12:48<68:31:59, 4.04it/s, grad_norm=0.536, loss_final=0.522, loss_mean=0.847, loss_mean_cls=0.079, proj_loss=-0.404][2026-03-26 12:43:52] Step: 3081, Training Logs: loss_final: 0.533142, loss_mean: 0.858183, proj_loss: -0.405609, loss_mean_cls: 0.080568, grad_norm: 0.510390 +Steps: 0%| | 3082/1000000 [12:48<68:37:26, 4.04it/s, grad_norm=0.51, loss_final=0.533, loss_mean=0.858, loss_mean_cls=0.0806, proj_loss=-0.406][2026-03-26 12:43:52] Step: 3082, Training Logs: loss_final: 0.527131, loss_mean: 0.857076, proj_loss: -0.409849, loss_mean_cls: 0.079903, grad_norm: 0.620672 +Steps: 0%| | 3083/1000000 [12:49<68:35:08, 4.04it/s, grad_norm=0.621, loss_final=0.527, loss_mean=0.857, loss_mean_cls=0.0799, proj_loss=-0.41][2026-03-26 12:43:52] Step: 3083, Training Logs: loss_final: 0.521936, loss_mean: 0.854258, proj_loss: -0.410038, loss_mean_cls: 0.077715, grad_norm: 0.471472 +Steps: 0%| | 3084/1000000 [12:49<68:36:09, 4.04it/s, grad_norm=0.471, loss_final=0.522, loss_mean=0.854, loss_mean_cls=0.0777, proj_loss=-0.41][2026-03-26 12:43:52] Step: 3084, Training Logs: loss_final: 0.499484, loss_mean: 0.828922, proj_loss: -0.410774, loss_mean_cls: 0.081336, grad_norm: 0.422820 +Steps: 0%| | 3085/1000000 [12:49<68:36:01, 4.04it/s, grad_norm=0.423, loss_final=0.499, loss_mean=0.829, loss_mean_cls=0.0813, proj_loss=-0.411][2026-03-26 12:43:53] Step: 3085, Training Logs: loss_final: 0.502549, loss_mean: 0.818110, proj_loss: -0.398178, loss_mean_cls: 0.082617, grad_norm: 0.577099 +Steps: 0%| | 3086/1000000 [12:49<68:34:04, 4.04it/s, grad_norm=0.577, loss_final=0.503, loss_mean=0.818, loss_mean_cls=0.0826, proj_loss=-0.398][2026-03-26 12:43:53] Step: 3086, Training Logs: loss_final: 0.504877, loss_mean: 0.818731, proj_loss: -0.397540, loss_mean_cls: 0.083687, grad_norm: 0.453352 +Steps: 0%| | 3087/1000000 [12:50<68:33:04, 4.04it/s, grad_norm=0.453, loss_final=0.505, loss_mean=0.819, loss_mean_cls=0.0837, proj_loss=-0.398][2026-03-26 12:43:53] Step: 3087, Training Logs: loss_final: 0.522043, loss_mean: 0.843201, proj_loss: -0.401165, loss_mean_cls: 0.080006, grad_norm: 0.490270 +Steps: 0%| | 3088/1000000 [12:50<68:32:23, 4.04it/s, grad_norm=0.49, loss_final=0.522, loss_mean=0.843, loss_mean_cls=0.08, proj_loss=-0.401][2026-03-26 12:43:53] Step: 3088, Training Logs: loss_final: 0.522048, loss_mean: 0.843933, proj_loss: -0.403709, loss_mean_cls: 0.081824, grad_norm: 0.381356 +Steps: 0%| | 3089/1000000 [12:50<68:31:59, 4.04it/s, grad_norm=0.381, loss_final=0.522, loss_mean=0.844, loss_mean_cls=0.0818, proj_loss=-0.404][2026-03-26 12:43:54] Step: 3089, Training Logs: loss_final: 0.534387, loss_mean: 0.858253, proj_loss: -0.403947, loss_mean_cls: 0.080081, grad_norm: 0.383733 +Steps: 0%| | 3090/1000000 [12:50<68:33:11, 4.04it/s, grad_norm=0.384, loss_final=0.534, loss_mean=0.858, loss_mean_cls=0.0801, proj_loss=-0.404][2026-03-26 12:43:54] Step: 3090, Training Logs: loss_final: 0.519246, loss_mean: 0.842285, proj_loss: -0.403827, loss_mean_cls: 0.080788, grad_norm: 0.574298 +Steps: 0%| | 3091/1000000 [12:51<68:33:08, 4.04it/s, grad_norm=0.574, loss_final=0.519, loss_mean=0.842, loss_mean_cls=0.0808, proj_loss=-0.404][2026-03-26 12:43:54] Step: 3091, Training Logs: loss_final: 0.525311, loss_mean: 0.837508, proj_loss: -0.394758, loss_mean_cls: 0.082560, grad_norm: 0.436583 +Steps: 0%| | 3092/1000000 [12:51<68:31:46, 4.04it/s, grad_norm=0.437, loss_final=0.525, loss_mean=0.838, loss_mean_cls=0.0826, proj_loss=-0.395][2026-03-26 12:43:54] Step: 3092, Training Logs: loss_final: 0.512497, loss_mean: 0.841936, proj_loss: -0.409681, loss_mean_cls: 0.080241, grad_norm: 0.740536 +Steps: 0%| | 3093/1000000 [12:51<68:51:25, 4.02it/s, grad_norm=0.741, loss_final=0.512, loss_mean=0.842, loss_mean_cls=0.0802, proj_loss=-0.41][2026-03-26 12:43:55] Step: 3093, Training Logs: loss_final: 0.540016, loss_mean: 0.841768, proj_loss: -0.386489, loss_mean_cls: 0.084738, grad_norm: 0.708390 +Steps: 0%| | 3094/1000000 [12:51<68:48:36, 4.02it/s, grad_norm=0.708, loss_final=0.54, loss_mean=0.842, loss_mean_cls=0.0847, proj_loss=-0.386][2026-03-26 12:43:55] Step: 3094, Training Logs: loss_final: 0.512418, loss_mean: 0.820248, proj_loss: -0.392530, loss_mean_cls: 0.084700, grad_norm: 0.546199 +Steps: 0%| | 3095/1000000 [12:52<68:43:15, 4.03it/s, grad_norm=0.546, loss_final=0.512, loss_mean=0.82, loss_mean_cls=0.0847, proj_loss=-0.393][2026-03-26 12:43:55] Step: 3095, Training Logs: loss_final: 0.521344, loss_mean: 0.836219, proj_loss: -0.398456, loss_mean_cls: 0.083581, grad_norm: 0.847265 +Steps: 0%| | 3096/1000000 [12:52<68:43:08, 4.03it/s, grad_norm=0.847, loss_final=0.521, loss_mean=0.836, loss_mean_cls=0.0836, proj_loss=-0.398][2026-03-26 12:43:55] Step: 3096, Training Logs: loss_final: 0.503250, loss_mean: 0.827608, proj_loss: -0.406154, loss_mean_cls: 0.081796, grad_norm: 0.683791 +Steps: 0%| | 3097/1000000 [12:52<68:38:58, 4.03it/s, grad_norm=0.684, loss_final=0.503, loss_mean=0.828, loss_mean_cls=0.0818, proj_loss=-0.406][2026-03-26 12:43:55] Step: 3097, Training Logs: loss_final: 0.538394, loss_mean: 0.871003, proj_loss: -0.411815, loss_mean_cls: 0.079206, grad_norm: 0.460451 +Steps: 0%| | 3098/1000000 [12:52<68:41:14, 4.03it/s, grad_norm=0.46, loss_final=0.538, loss_mean=0.871, loss_mean_cls=0.0792, proj_loss=-0.412][2026-03-26 12:43:56] Step: 3098, Training Logs: loss_final: 0.527427, loss_mean: 0.857655, proj_loss: -0.410236, loss_mean_cls: 0.080008, grad_norm: 0.653407 +Steps: 0%| | 3099/1000000 [12:53<68:39:16, 4.03it/s, grad_norm=0.653, loss_final=0.527, loss_mean=0.858, loss_mean_cls=0.08, proj_loss=-0.41][2026-03-26 12:43:56] Step: 3099, Training Logs: loss_final: 0.501184, loss_mean: 0.828040, proj_loss: -0.408987, loss_mean_cls: 0.082131, grad_norm: 0.353377 +Steps: 0%| | 3100/1000000 [12:53<68:43:40, 4.03it/s, grad_norm=0.353, loss_final=0.501, loss_mean=0.828, loss_mean_cls=0.0821, proj_loss=-0.409][2026-03-26 12:43:56] Step: 3100, Training Logs: loss_final: 0.537780, loss_mean: 0.862477, proj_loss: -0.404738, loss_mean_cls: 0.080041, grad_norm: 0.508935 +Steps: 0%| | 3101/1000000 [12:53<68:38:55, 4.03it/s, grad_norm=0.509, loss_final=0.538, loss_mean=0.862, loss_mean_cls=0.08, proj_loss=-0.405][2026-03-26 12:43:56] Step: 3101, Training Logs: loss_final: 0.529050, loss_mean: 0.852352, proj_loss: -0.405222, loss_mean_cls: 0.081920, grad_norm: 0.566759 +Steps: 0%| | 3102/1000000 [12:53<68:39:14, 4.03it/s, grad_norm=0.567, loss_final=0.529, loss_mean=0.852, loss_mean_cls=0.0819, proj_loss=-0.405][2026-03-26 12:43:57] Step: 3102, Training Logs: loss_final: 0.517234, loss_mean: 0.844501, proj_loss: -0.409169, loss_mean_cls: 0.081902, grad_norm: 0.325681 +Steps: 0%| | 3103/1000000 [12:54<68:37:05, 4.04it/s, grad_norm=0.326, loss_final=0.517, loss_mean=0.845, loss_mean_cls=0.0819, proj_loss=-0.409][2026-03-26 12:43:57] Step: 3103, Training Logs: loss_final: 0.514462, loss_mean: 0.841958, proj_loss: -0.407585, loss_mean_cls: 0.080089, grad_norm: 0.697238 +Steps: 0%| | 3104/1000000 [12:54<68:37:00, 4.04it/s, grad_norm=0.697, loss_final=0.514, loss_mean=0.842, loss_mean_cls=0.0801, proj_loss=-0.408][2026-03-26 12:43:57] Step: 3104, Training Logs: loss_final: 0.520603, loss_mean: 0.845048, proj_loss: -0.405085, loss_mean_cls: 0.080640, grad_norm: 0.540275 +Steps: 0%| | 3105/1000000 [12:54<68:34:35, 4.04it/s, grad_norm=0.54, loss_final=0.521, loss_mean=0.845, loss_mean_cls=0.0806, proj_loss=-0.405][2026-03-26 12:43:57] Step: 3105, Training Logs: loss_final: 0.516777, loss_mean: 0.845013, proj_loss: -0.408147, loss_mean_cls: 0.079910, grad_norm: 0.414353 +Steps: 0%| | 3106/1000000 [12:54<68:35:45, 4.04it/s, grad_norm=0.414, loss_final=0.517, loss_mean=0.845, loss_mean_cls=0.0799, proj_loss=-0.408][2026-03-26 12:43:58] Step: 3106, Training Logs: loss_final: 0.513812, loss_mean: 0.840658, proj_loss: -0.408285, loss_mean_cls: 0.081439, grad_norm: 0.604603 +Steps: 0%| | 3107/1000000 [12:55<68:32:13, 4.04it/s, grad_norm=0.605, loss_final=0.514, loss_mean=0.841, loss_mean_cls=0.0814, proj_loss=-0.408][2026-03-26 12:43:58] Step: 3107, Training Logs: loss_final: 0.501604, loss_mean: 0.823201, proj_loss: -0.403721, loss_mean_cls: 0.082123, grad_norm: 0.349711 +Steps: 0%| | 3108/1000000 [12:55<68:38:07, 4.03it/s, grad_norm=0.35, loss_final=0.502, loss_mean=0.823, loss_mean_cls=0.0821, proj_loss=-0.404][2026-03-26 12:43:58] Step: 3108, Training Logs: loss_final: 0.530006, loss_mean: 0.864408, proj_loss: -0.413245, loss_mean_cls: 0.078844, grad_norm: 0.589361 +Steps: 0%| | 3109/1000000 [12:55<68:38:09, 4.03it/s, grad_norm=0.589, loss_final=0.53, loss_mean=0.864, loss_mean_cls=0.0788, proj_loss=-0.413][2026-03-26 12:43:58] Step: 3109, Training Logs: loss_final: 0.513650, loss_mean: 0.833299, proj_loss: -0.401446, loss_mean_cls: 0.081796, grad_norm: 0.474693 +Steps: 0%| | 3110/1000000 [12:55<68:36:04, 4.04it/s, grad_norm=0.475, loss_final=0.514, loss_mean=0.833, loss_mean_cls=0.0818, proj_loss=-0.401][2026-03-26 12:43:59] Step: 3110, Training Logs: loss_final: 0.538089, loss_mean: 0.853775, proj_loss: -0.396012, loss_mean_cls: 0.080325, grad_norm: 0.351531 +Steps: 0%| | 3111/1000000 [12:56<68:36:55, 4.04it/s, grad_norm=0.352, loss_final=0.538, loss_mean=0.854, loss_mean_cls=0.0803, proj_loss=-0.396][2026-03-26 12:43:59] Step: 3111, Training Logs: loss_final: 0.523623, loss_mean: 0.836805, proj_loss: -0.396458, loss_mean_cls: 0.083277, grad_norm: 0.469507 +Steps: 0%| | 3112/1000000 [12:56<68:36:51, 4.04it/s, grad_norm=0.47, loss_final=0.524, loss_mean=0.837, loss_mean_cls=0.0833, proj_loss=-0.396][2026-03-26 12:43:59] Step: 3112, Training Logs: loss_final: 0.527958, loss_mean: 0.833058, proj_loss: -0.390161, loss_mean_cls: 0.085061, grad_norm: 0.391220 +Steps: 0%| | 3113/1000000 [12:56<68:41:06, 4.03it/s, grad_norm=0.391, loss_final=0.528, loss_mean=0.833, loss_mean_cls=0.0851, proj_loss=-0.39][2026-03-26 12:43:59] Step: 3113, Training Logs: loss_final: 0.507352, loss_mean: 0.833526, proj_loss: -0.407726, loss_mean_cls: 0.081552, grad_norm: 0.425822 +Steps: 0%| | 3114/1000000 [12:56<68:41:26, 4.03it/s, grad_norm=0.426, loss_final=0.507, loss_mean=0.834, loss_mean_cls=0.0816, proj_loss=-0.408][2026-03-26 12:44:00] Step: 3114, Training Logs: loss_final: 0.526965, loss_mean: 0.859190, proj_loss: -0.410816, loss_mean_cls: 0.078591, grad_norm: 0.448007 +Steps: 0%| | 3115/1000000 [12:57<68:38:07, 4.03it/s, grad_norm=0.448, loss_final=0.527, loss_mean=0.859, loss_mean_cls=0.0786, proj_loss=-0.411][2026-03-26 12:44:00] Step: 3115, Training Logs: loss_final: 0.524953, loss_mean: 0.847458, proj_loss: -0.404715, loss_mean_cls: 0.082210, grad_norm: 0.298483 +Steps: 0%| | 3116/1000000 [12:57<68:40:00, 4.03it/s, grad_norm=0.298, loss_final=0.525, loss_mean=0.847, loss_mean_cls=0.0822, proj_loss=-0.405][2026-03-26 12:44:00] Step: 3116, Training Logs: loss_final: 0.500311, loss_mean: 0.825474, proj_loss: -0.406354, loss_mean_cls: 0.081191, grad_norm: 0.380207 +Steps: 0%| | 3117/1000000 [12:57<68:37:56, 4.03it/s, grad_norm=0.38, loss_final=0.5, loss_mean=0.825, loss_mean_cls=0.0812, proj_loss=-0.406][2026-03-26 12:44:00] Step: 3117, Training Logs: loss_final: 0.526750, loss_mean: 0.850898, proj_loss: -0.405491, loss_mean_cls: 0.081343, grad_norm: 0.283157 +Steps: 0%| | 3118/1000000 [12:57<68:38:06, 4.03it/s, grad_norm=0.283, loss_final=0.527, loss_mean=0.851, loss_mean_cls=0.0813, proj_loss=-0.405][2026-03-26 12:44:01] Step: 3118, Training Logs: loss_final: 0.506825, loss_mean: 0.826206, proj_loss: -0.401994, loss_mean_cls: 0.082613, grad_norm: 0.424290 +Steps: 0%| | 3119/1000000 [12:58<68:39:11, 4.03it/s, grad_norm=0.424, loss_final=0.507, loss_mean=0.826, loss_mean_cls=0.0826, proj_loss=-0.402][2026-03-26 12:44:01] Step: 3119, Training Logs: loss_final: 0.521034, loss_mean: 0.846268, proj_loss: -0.404907, loss_mean_cls: 0.079672, grad_norm: 0.348254 +Steps: 0%| | 3120/1000000 [12:58<68:38:31, 4.03it/s, grad_norm=0.348, loss_final=0.521, loss_mean=0.846, loss_mean_cls=0.0797, proj_loss=-0.405][2026-03-26 12:44:01] Step: 3120, Training Logs: loss_final: 0.530377, loss_mean: 0.856943, proj_loss: -0.405381, loss_mean_cls: 0.078815, grad_norm: 0.320984 +Steps: 0%| | 3121/1000000 [12:58<68:38:51, 4.03it/s, grad_norm=0.321, loss_final=0.53, loss_mean=0.857, loss_mean_cls=0.0788, proj_loss=-0.405][2026-03-26 12:44:01] Step: 3121, Training Logs: loss_final: 0.504323, loss_mean: 0.812736, proj_loss: -0.394173, loss_mean_cls: 0.085761, grad_norm: 0.332451 +Steps: 0%| | 3122/1000000 [12:58<68:41:38, 4.03it/s, grad_norm=0.332, loss_final=0.504, loss_mean=0.813, loss_mean_cls=0.0858, proj_loss=-0.394][2026-03-26 12:44:02] Step: 3122, Training Logs: loss_final: 0.529465, loss_mean: 0.856818, proj_loss: -0.407318, loss_mean_cls: 0.079965, grad_norm: 0.332265 +Steps: 0%| | 3123/1000000 [12:59<68:45:36, 4.03it/s, grad_norm=0.332, loss_final=0.529, loss_mean=0.857, loss_mean_cls=0.08, proj_loss=-0.407][2026-03-26 12:44:02] Step: 3123, Training Logs: loss_final: 0.514022, loss_mean: 0.845844, proj_loss: -0.410818, loss_mean_cls: 0.078996, grad_norm: 0.300985 +Steps: 0%| | 3124/1000000 [12:59<68:41:31, 4.03it/s, grad_norm=0.301, loss_final=0.514, loss_mean=0.846, loss_mean_cls=0.079, proj_loss=-0.411][2026-03-26 12:44:02] Step: 3124, Training Logs: loss_final: 0.520819, loss_mean: 0.847003, proj_loss: -0.406012, loss_mean_cls: 0.079828, grad_norm: 0.309758 +Steps: 0%| | 3125/1000000 [12:59<68:36:43, 4.04it/s, grad_norm=0.31, loss_final=0.521, loss_mean=0.847, loss_mean_cls=0.0798, proj_loss=-0.406][2026-03-26 12:44:02] Step: 3125, Training Logs: loss_final: 0.511608, loss_mean: 0.833375, proj_loss: -0.403839, loss_mean_cls: 0.082072, grad_norm: 0.507062 +Steps: 0%| | 3126/1000000 [12:59<68:37:43, 4.03it/s, grad_norm=0.507, loss_final=0.512, loss_mean=0.833, loss_mean_cls=0.0821, proj_loss=-0.404][2026-03-26 12:44:03] Step: 3126, Training Logs: loss_final: 0.525382, loss_mean: 0.848406, proj_loss: -0.404746, loss_mean_cls: 0.081722, grad_norm: 0.394393 +Steps: 0%| | 3127/1000000 [13:00<68:39:02, 4.03it/s, grad_norm=0.394, loss_final=0.525, loss_mean=0.848, loss_mean_cls=0.0817, proj_loss=-0.405][2026-03-26 12:44:03] Step: 3127, Training Logs: loss_final: 0.510707, loss_mean: 0.828626, proj_loss: -0.401520, loss_mean_cls: 0.083600, grad_norm: 0.366374 +Steps: 0%| | 3128/1000000 [13:00<68:38:11, 4.03it/s, grad_norm=0.366, loss_final=0.511, loss_mean=0.829, loss_mean_cls=0.0836, proj_loss=-0.402][2026-03-26 12:44:03] Step: 3128, Training Logs: loss_final: 0.531905, loss_mean: 0.840133, proj_loss: -0.391540, loss_mean_cls: 0.083312, grad_norm: 0.531778 +Steps: 0%| | 3129/1000000 [13:00<68:40:54, 4.03it/s, grad_norm=0.532, loss_final=0.532, loss_mean=0.84, loss_mean_cls=0.0833, proj_loss=-0.392][2026-03-26 12:44:03] Step: 3129, Training Logs: loss_final: 0.525943, loss_mean: 0.849457, proj_loss: -0.403972, loss_mean_cls: 0.080457, grad_norm: 0.419868 +Steps: 0%| | 3130/1000000 [13:00<68:44:19, 4.03it/s, grad_norm=0.42, loss_final=0.526, loss_mean=0.849, loss_mean_cls=0.0805, proj_loss=-0.404][2026-03-26 12:44:04] Step: 3130, Training Logs: loss_final: 0.531505, loss_mean: 0.858324, proj_loss: -0.406940, loss_mean_cls: 0.080122, grad_norm: 0.543498 +Steps: 0%| | 3131/1000000 [13:01<68:41:02, 4.03it/s, grad_norm=0.543, loss_final=0.532, loss_mean=0.858, loss_mean_cls=0.0801, proj_loss=-0.407][2026-03-26 12:44:04] Step: 3131, Training Logs: loss_final: 0.525765, loss_mean: 0.852145, proj_loss: -0.406435, loss_mean_cls: 0.080055, grad_norm: 0.784660 +Steps: 0%| | 3132/1000000 [13:01<68:55:59, 4.02it/s, grad_norm=0.785, loss_final=0.526, loss_mean=0.852, loss_mean_cls=0.0801, proj_loss=-0.406][2026-03-26 12:44:04] Step: 3132, Training Logs: loss_final: 0.531344, loss_mean: 0.861219, proj_loss: -0.410068, loss_mean_cls: 0.080193, grad_norm: 0.460725 +Steps: 0%| | 3133/1000000 [13:01<68:51:04, 4.02it/s, grad_norm=0.461, loss_final=0.531, loss_mean=0.861, loss_mean_cls=0.0802, proj_loss=-0.41][2026-03-26 12:44:04] Step: 3133, Training Logs: loss_final: 0.514994, loss_mean: 0.832204, proj_loss: -0.398602, loss_mean_cls: 0.081391, grad_norm: 0.330729 +Steps: 0%| | 3134/1000000 [13:01<68:47:16, 4.03it/s, grad_norm=0.331, loss_final=0.515, loss_mean=0.832, loss_mean_cls=0.0814, proj_loss=-0.399][2026-03-26 12:44:05] Step: 3134, Training Logs: loss_final: 0.525577, loss_mean: 0.830830, proj_loss: -0.389922, loss_mean_cls: 0.084669, grad_norm: 0.492171 +Steps: 0%| | 3135/1000000 [13:02<68:41:39, 4.03it/s, grad_norm=0.492, loss_final=0.526, loss_mean=0.831, loss_mean_cls=0.0847, proj_loss=-0.39][2026-03-26 12:44:05] Step: 3135, Training Logs: loss_final: 0.511408, loss_mean: 0.837906, proj_loss: -0.406486, loss_mean_cls: 0.079988, grad_norm: 0.438110 +Steps: 0%| | 3136/1000000 [13:02<68:39:27, 4.03it/s, grad_norm=0.438, loss_final=0.511, loss_mean=0.838, loss_mean_cls=0.08, proj_loss=-0.406][2026-03-26 12:44:05] Step: 3136, Training Logs: loss_final: 0.510347, loss_mean: 0.833445, proj_loss: -0.405244, loss_mean_cls: 0.082147, grad_norm: 0.830025 +Steps: 0%| | 3137/1000000 [13:02<68:37:20, 4.04it/s, grad_norm=0.83, loss_final=0.51, loss_mean=0.833, loss_mean_cls=0.0821, proj_loss=-0.405][2026-03-26 12:44:05] Step: 3137, Training Logs: loss_final: 0.506447, loss_mean: 0.831935, proj_loss: -0.406863, loss_mean_cls: 0.081376, grad_norm: 0.688682 +Steps: 0%| | 3138/1000000 [13:02<68:38:08, 4.03it/s, grad_norm=0.689, loss_final=0.506, loss_mean=0.832, loss_mean_cls=0.0814, proj_loss=-0.407][2026-03-26 12:44:06] Step: 3138, Training Logs: loss_final: 0.529542, loss_mean: 0.842096, proj_loss: -0.394849, loss_mean_cls: 0.082295, grad_norm: 0.561321 +Steps: 0%| | 3139/1000000 [13:03<68:42:43, 4.03it/s, grad_norm=0.561, loss_final=0.53, loss_mean=0.842, loss_mean_cls=0.0823, proj_loss=-0.395][2026-03-26 12:44:06] Step: 3139, Training Logs: loss_final: 0.516974, loss_mean: 0.843407, proj_loss: -0.406573, loss_mean_cls: 0.080141, grad_norm: 0.699141 +Steps: 0%| | 3140/1000000 [13:03<68:42:39, 4.03it/s, grad_norm=0.699, loss_final=0.517, loss_mean=0.843, loss_mean_cls=0.0801, proj_loss=-0.407][2026-03-26 12:44:06] Step: 3140, Training Logs: loss_final: 0.496272, loss_mean: 0.819722, proj_loss: -0.405747, loss_mean_cls: 0.082297, grad_norm: 0.622438 +Steps: 0%| | 3141/1000000 [13:03<68:42:45, 4.03it/s, grad_norm=0.622, loss_final=0.496, loss_mean=0.82, loss_mean_cls=0.0823, proj_loss=-0.406][2026-03-26 12:44:06] Step: 3141, Training Logs: loss_final: 0.528948, loss_mean: 0.840522, proj_loss: -0.395917, loss_mean_cls: 0.084343, grad_norm: 0.481177 +Steps: 0%| | 3142/1000000 [13:03<68:43:29, 4.03it/s, grad_norm=0.481, loss_final=0.529, loss_mean=0.841, loss_mean_cls=0.0843, proj_loss=-0.396][2026-03-26 12:44:07] Step: 3142, Training Logs: loss_final: 0.518912, loss_mean: 0.830910, proj_loss: -0.395250, loss_mean_cls: 0.083251, grad_norm: 0.495813 +Steps: 0%| | 3143/1000000 [13:04<68:44:42, 4.03it/s, grad_norm=0.496, loss_final=0.519, loss_mean=0.831, loss_mean_cls=0.0833, proj_loss=-0.395][2026-03-26 12:44:07] Step: 3143, Training Logs: loss_final: 0.519486, loss_mean: 0.844796, proj_loss: -0.406854, loss_mean_cls: 0.081545, grad_norm: 0.314942 +Steps: 0%| | 3144/1000000 [13:04<68:43:20, 4.03it/s, grad_norm=0.315, loss_final=0.519, loss_mean=0.845, loss_mean_cls=0.0815, proj_loss=-0.407][2026-03-26 12:44:07] Step: 3144, Training Logs: loss_final: 0.522153, loss_mean: 0.844797, proj_loss: -0.402658, loss_mean_cls: 0.080014, grad_norm: 0.467507 +Steps: 0%| | 3145/1000000 [13:04<68:42:23, 4.03it/s, grad_norm=0.468, loss_final=0.522, loss_mean=0.845, loss_mean_cls=0.08, proj_loss=-0.403][2026-03-26 12:44:07] Step: 3145, Training Logs: loss_final: 0.512594, loss_mean: 0.837935, proj_loss: -0.406760, loss_mean_cls: 0.081419, grad_norm: 0.588793 +Steps: 0%| | 3146/1000000 [13:04<68:41:04, 4.03it/s, grad_norm=0.589, loss_final=0.513, loss_mean=0.838, loss_mean_cls=0.0814, proj_loss=-0.407][2026-03-26 12:44:08] Step: 3146, Training Logs: loss_final: 0.533536, loss_mean: 0.836986, proj_loss: -0.389411, loss_mean_cls: 0.085961, grad_norm: 0.546689 +Steps: 0%| | 3147/1000000 [13:04<68:36:58, 4.04it/s, grad_norm=0.547, loss_final=0.534, loss_mean=0.837, loss_mean_cls=0.086, proj_loss=-0.389][2026-03-26 12:44:08] Step: 3147, Training Logs: loss_final: 0.522574, loss_mean: 0.851210, proj_loss: -0.410908, loss_mean_cls: 0.082272, grad_norm: 0.386350 +Steps: 0%| | 3148/1000000 [13:05<68:36:07, 4.04it/s, grad_norm=0.386, loss_final=0.523, loss_mean=0.851, loss_mean_cls=0.0823, proj_loss=-0.411][2026-03-26 12:44:08] Step: 3148, Training Logs: loss_final: 0.519657, loss_mean: 0.832514, proj_loss: -0.396094, loss_mean_cls: 0.083237, grad_norm: 0.404520 +Steps: 0%| | 3149/1000000 [13:05<68:34:41, 4.04it/s, grad_norm=0.405, loss_final=0.52, loss_mean=0.833, loss_mean_cls=0.0832, proj_loss=-0.396][2026-03-26 12:44:08] Step: 3149, Training Logs: loss_final: 0.538245, loss_mean: 0.860760, proj_loss: -0.403023, loss_mean_cls: 0.080508, grad_norm: 0.327783 +Steps: 0%| | 3150/1000000 [13:05<68:36:51, 4.04it/s, grad_norm=0.328, loss_final=0.538, loss_mean=0.861, loss_mean_cls=0.0805, proj_loss=-0.403][2026-03-26 12:44:09] Step: 3150, Training Logs: loss_final: 0.507141, loss_mean: 0.830921, proj_loss: -0.404220, loss_mean_cls: 0.080441, grad_norm: 0.349433 +Steps: 0%| | 3151/1000000 [13:05<68:33:45, 4.04it/s, grad_norm=0.349, loss_final=0.507, loss_mean=0.831, loss_mean_cls=0.0804, proj_loss=-0.404][2026-03-26 12:44:09] Step: 3151, Training Logs: loss_final: 0.511944, loss_mean: 0.829972, proj_loss: -0.399524, loss_mean_cls: 0.081497, grad_norm: 0.372966 +Steps: 0%| | 3152/1000000 [13:06<68:35:59, 4.04it/s, grad_norm=0.373, loss_final=0.512, loss_mean=0.83, loss_mean_cls=0.0815, proj_loss=-0.4][2026-03-26 12:44:09] Step: 3152, Training Logs: loss_final: 0.525749, loss_mean: 0.855477, proj_loss: -0.409232, loss_mean_cls: 0.079504, grad_norm: 0.286832 +Steps: 0%| | 3153/1000000 [13:06<68:36:07, 4.04it/s, grad_norm=0.287, loss_final=0.526, loss_mean=0.855, loss_mean_cls=0.0795, proj_loss=-0.409][2026-03-26 12:44:09] Step: 3153, Training Logs: loss_final: 0.529173, loss_mean: 0.848893, proj_loss: -0.400711, loss_mean_cls: 0.080991, grad_norm: 0.478963 +Steps: 0%| | 3154/1000000 [13:06<68:39:02, 4.03it/s, grad_norm=0.479, loss_final=0.529, loss_mean=0.849, loss_mean_cls=0.081, proj_loss=-0.401][2026-03-26 12:44:10] Step: 3154, Training Logs: loss_final: 0.526712, loss_mean: 0.848003, proj_loss: -0.403600, loss_mean_cls: 0.082308, grad_norm: 0.331495 +Steps: 0%| | 3155/1000000 [13:06<68:34:48, 4.04it/s, grad_norm=0.331, loss_final=0.527, loss_mean=0.848, loss_mean_cls=0.0823, proj_loss=-0.404][2026-03-26 12:44:10] Step: 3155, Training Logs: loss_final: 0.511594, loss_mean: 0.837845, proj_loss: -0.408808, loss_mean_cls: 0.082557, grad_norm: 0.450477 +Steps: 0%| | 3156/1000000 [13:07<68:34:54, 4.04it/s, grad_norm=0.45, loss_final=0.512, loss_mean=0.838, loss_mean_cls=0.0826, proj_loss=-0.409][2026-03-26 12:44:10] Step: 3156, Training Logs: loss_final: 0.520184, loss_mean: 0.849972, proj_loss: -0.410193, loss_mean_cls: 0.080405, grad_norm: 0.566710 +Steps: 0%| | 3157/1000000 [13:07<68:34:20, 4.04it/s, grad_norm=0.567, loss_final=0.52, loss_mean=0.85, loss_mean_cls=0.0804, proj_loss=-0.41][2026-03-26 12:44:10] Step: 3157, Training Logs: loss_final: 0.521809, loss_mean: 0.851499, proj_loss: -0.409953, loss_mean_cls: 0.080264, grad_norm: 0.307415 +Steps: 0%| | 3158/1000000 [13:07<68:36:43, 4.04it/s, grad_norm=0.307, loss_final=0.522, loss_mean=0.851, loss_mean_cls=0.0803, proj_loss=-0.41][2026-03-26 12:44:11] Step: 3158, Training Logs: loss_final: 0.516215, loss_mean: 0.838404, proj_loss: -0.404444, loss_mean_cls: 0.082255, grad_norm: 0.516909 +Steps: 0%| | 3159/1000000 [13:07<68:35:40, 4.04it/s, grad_norm=0.517, loss_final=0.516, loss_mean=0.838, loss_mean_cls=0.0823, proj_loss=-0.404][2026-03-26 12:44:11] Step: 3159, Training Logs: loss_final: 0.525474, loss_mean: 0.844150, proj_loss: -0.401204, loss_mean_cls: 0.082528, grad_norm: 0.574276 +Steps: 0%| | 3160/1000000 [13:08<68:33:03, 4.04it/s, grad_norm=0.574, loss_final=0.525, loss_mean=0.844, loss_mean_cls=0.0825, proj_loss=-0.401][2026-03-26 12:44:11] Step: 3160, Training Logs: loss_final: 0.522011, loss_mean: 0.854972, proj_loss: -0.411583, loss_mean_cls: 0.078623, grad_norm: 0.414892 +Steps: 0%| | 3161/1000000 [13:08<68:29:57, 4.04it/s, grad_norm=0.415, loss_final=0.522, loss_mean=0.855, loss_mean_cls=0.0786, proj_loss=-0.412][2026-03-26 12:44:11] Step: 3161, Training Logs: loss_final: 0.510509, loss_mean: 0.805024, proj_loss: -0.380119, loss_mean_cls: 0.085605, grad_norm: 0.772477 +Steps: 0%| | 3162/1000000 [13:08<68:28:33, 4.04it/s, grad_norm=0.772, loss_final=0.511, loss_mean=0.805, loss_mean_cls=0.0856, proj_loss=-0.38][2026-03-26 12:44:12] Step: 3162, Training Logs: loss_final: 0.523075, loss_mean: 0.845085, proj_loss: -0.403884, loss_mean_cls: 0.081874, grad_norm: 0.378226 +Steps: 0%| | 3163/1000000 [13:08<68:28:14, 4.04it/s, grad_norm=0.378, loss_final=0.523, loss_mean=0.845, loss_mean_cls=0.0819, proj_loss=-0.404][2026-03-26 12:44:12] Step: 3163, Training Logs: loss_final: 0.527877, loss_mean: 0.847778, proj_loss: -0.403265, loss_mean_cls: 0.083363, grad_norm: 0.622303 +Steps: 0%| | 3164/1000000 [13:09<68:29:54, 4.04it/s, grad_norm=0.622, loss_final=0.528, loss_mean=0.848, loss_mean_cls=0.0834, proj_loss=-0.403][2026-03-26 12:44:12] Step: 3164, Training Logs: loss_final: 0.519376, loss_mean: 0.840704, proj_loss: -0.403869, loss_mean_cls: 0.082541, grad_norm: 0.777606 +Steps: 0%| | 3165/1000000 [13:09<68:31:44, 4.04it/s, grad_norm=0.778, loss_final=0.519, loss_mean=0.841, loss_mean_cls=0.0825, proj_loss=-0.404][2026-03-26 12:44:12] Step: 3165, Training Logs: loss_final: 0.511201, loss_mean: 0.841455, proj_loss: -0.410299, loss_mean_cls: 0.080045, grad_norm: 0.516625 +Steps: 0%| | 3166/1000000 [13:09<68:32:08, 4.04it/s, grad_norm=0.517, loss_final=0.511, loss_mean=0.841, loss_mean_cls=0.08, proj_loss=-0.41][2026-03-26 12:44:13] Step: 3166, Training Logs: loss_final: 0.513819, loss_mean: 0.834283, proj_loss: -0.403611, loss_mean_cls: 0.083147, grad_norm: 0.343111 +Steps: 0%| | 3167/1000000 [13:09<68:31:06, 4.04it/s, grad_norm=0.343, loss_final=0.514, loss_mean=0.834, loss_mean_cls=0.0831, proj_loss=-0.404][2026-03-26 12:44:13] Step: 3167, Training Logs: loss_final: 0.508595, loss_mean: 0.826313, proj_loss: -0.399525, loss_mean_cls: 0.081807, grad_norm: 0.552730 +Steps: 0%| | 3168/1000000 [13:10<68:30:34, 4.04it/s, grad_norm=0.553, loss_final=0.509, loss_mean=0.826, loss_mean_cls=0.0818, proj_loss=-0.4][2026-03-26 12:44:13] Step: 3168, Training Logs: loss_final: 0.507688, loss_mean: 0.827932, proj_loss: -0.403845, loss_mean_cls: 0.083602, grad_norm: 0.348501 +Steps: 0%| | 3169/1000000 [13:10<68:33:28, 4.04it/s, grad_norm=0.349, loss_final=0.508, loss_mean=0.828, loss_mean_cls=0.0836, proj_loss=-0.404][2026-03-26 12:44:13] Step: 3169, Training Logs: loss_final: 0.508512, loss_mean: 0.838056, proj_loss: -0.409088, loss_mean_cls: 0.079544, grad_norm: 0.550182 +Steps: 0%| | 3170/1000000 [13:10<68:31:53, 4.04it/s, grad_norm=0.55, loss_final=0.509, loss_mean=0.838, loss_mean_cls=0.0795, proj_loss=-0.409][2026-03-26 12:44:14] Step: 3170, Training Logs: loss_final: 0.504242, loss_mean: 0.812571, proj_loss: -0.392805, loss_mean_cls: 0.084476, grad_norm: 0.625241 +Steps: 0%| | 3171/1000000 [13:10<68:33:36, 4.04it/s, grad_norm=0.625, loss_final=0.504, loss_mean=0.813, loss_mean_cls=0.0845, proj_loss=-0.393][2026-03-26 12:44:14] Step: 3171, Training Logs: loss_final: 0.533930, loss_mean: 0.852432, proj_loss: -0.399672, loss_mean_cls: 0.081171, grad_norm: 0.523308 +Steps: 0%| | 3172/1000000 [13:11<68:30:22, 4.04it/s, grad_norm=0.523, loss_final=0.534, loss_mean=0.852, loss_mean_cls=0.0812, proj_loss=-0.4][2026-03-26 12:44:14] Step: 3172, Training Logs: loss_final: 0.509750, loss_mean: 0.825141, proj_loss: -0.397213, loss_mean_cls: 0.081822, grad_norm: 0.284906 +Steps: 0%| | 3173/1000000 [13:11<68:30:26, 4.04it/s, grad_norm=0.285, loss_final=0.51, loss_mean=0.825, loss_mean_cls=0.0818, proj_loss=-0.397][2026-03-26 12:44:14] Step: 3173, Training Logs: loss_final: 0.532294, loss_mean: 0.855901, proj_loss: -0.404005, loss_mean_cls: 0.080398, grad_norm: 0.298203 +Steps: 0%| | 3174/1000000 [13:11<68:30:43, 4.04it/s, grad_norm=0.298, loss_final=0.532, loss_mean=0.856, loss_mean_cls=0.0804, proj_loss=-0.404][2026-03-26 12:44:15] Step: 3174, Training Logs: loss_final: 0.530425, loss_mean: 0.852822, proj_loss: -0.404642, loss_mean_cls: 0.082245, grad_norm: 0.359778 +Steps: 0%| | 3175/1000000 [13:11<68:31:47, 4.04it/s, grad_norm=0.36, loss_final=0.53, loss_mean=0.853, loss_mean_cls=0.0822, proj_loss=-0.405][2026-03-26 12:44:15] Step: 3175, Training Logs: loss_final: 0.523395, loss_mean: 0.835911, proj_loss: -0.396236, loss_mean_cls: 0.083721, grad_norm: 0.398669 +Steps: 0%| | 3176/1000000 [13:12<68:34:48, 4.04it/s, grad_norm=0.399, loss_final=0.523, loss_mean=0.836, loss_mean_cls=0.0837, proj_loss=-0.396][2026-03-26 12:44:15] Step: 3176, Training Logs: loss_final: 0.513366, loss_mean: 0.830911, proj_loss: -0.400112, loss_mean_cls: 0.082567, grad_norm: 0.323301 +Steps: 0%| | 3177/1000000 [13:12<68:33:26, 4.04it/s, grad_norm=0.323, loss_final=0.513, loss_mean=0.831, loss_mean_cls=0.0826, proj_loss=-0.4][2026-03-26 12:44:15] Step: 3177, Training Logs: loss_final: 0.519454, loss_mean: 0.842033, proj_loss: -0.404650, loss_mean_cls: 0.082070, grad_norm: 0.428127 +Steps: 0%| | 3178/1000000 [13:12<68:35:07, 4.04it/s, grad_norm=0.428, loss_final=0.519, loss_mean=0.842, loss_mean_cls=0.0821, proj_loss=-0.405][2026-03-26 12:44:16] Step: 3178, Training Logs: loss_final: 0.527264, loss_mean: 0.854070, proj_loss: -0.405545, loss_mean_cls: 0.078739, grad_norm: 0.801326 +Steps: 0%| | 3179/1000000 [13:12<68:35:49, 4.04it/s, grad_norm=0.801, loss_final=0.527, loss_mean=0.854, loss_mean_cls=0.0787, proj_loss=-0.406][2026-03-26 12:44:16] Step: 3179, Training Logs: loss_final: 0.518862, loss_mean: 0.837868, proj_loss: -0.401381, loss_mean_cls: 0.082375, grad_norm: 0.501177 +Steps: 0%| | 3180/1000000 [13:13<68:36:08, 4.04it/s, grad_norm=0.501, loss_final=0.519, loss_mean=0.838, loss_mean_cls=0.0824, proj_loss=-0.401][2026-03-26 12:44:16] Step: 3180, Training Logs: loss_final: 0.494581, loss_mean: 0.815444, proj_loss: -0.404560, loss_mean_cls: 0.083696, grad_norm: 0.395168 +Steps: 0%| | 3181/1000000 [13:13<68:34:27, 4.04it/s, grad_norm=0.395, loss_final=0.495, loss_mean=0.815, loss_mean_cls=0.0837, proj_loss=-0.405][2026-03-26 12:44:16] Step: 3181, Training Logs: loss_final: 0.515500, loss_mean: 0.812031, proj_loss: -0.384485, loss_mean_cls: 0.087954, grad_norm: 0.611086 +Steps: 0%| | 3182/1000000 [13:13<70:27:35, 3.93it/s, grad_norm=0.611, loss_final=0.516, loss_mean=0.812, loss_mean_cls=0.088, proj_loss=-0.384][2026-03-26 12:44:17] Step: 3182, Training Logs: loss_final: 0.507913, loss_mean: 0.821808, proj_loss: -0.397810, loss_mean_cls: 0.083915, grad_norm: 0.360287 +Steps: 0%| | 3183/1000000 [13:13<69:59:30, 3.96it/s, grad_norm=0.36, loss_final=0.508, loss_mean=0.822, loss_mean_cls=0.0839, proj_loss=-0.398][2026-03-26 12:44:17] Step: 3183, Training Logs: loss_final: 0.532778, loss_mean: 0.863333, proj_loss: -0.409353, loss_mean_cls: 0.078798, grad_norm: 0.545041 +Steps: 0%| | 3184/1000000 [13:14<69:34:00, 3.98it/s, grad_norm=0.545, loss_final=0.533, loss_mean=0.863, loss_mean_cls=0.0788, proj_loss=-0.409][2026-03-26 12:44:17] Step: 3184, Training Logs: loss_final: 0.520305, loss_mean: 0.856223, proj_loss: -0.412816, loss_mean_cls: 0.076898, grad_norm: 0.441978 +Steps: 0%| | 3185/1000000 [13:14<69:17:52, 4.00it/s, grad_norm=0.442, loss_final=0.52, loss_mean=0.856, loss_mean_cls=0.0769, proj_loss=-0.413][2026-03-26 12:44:17] Step: 3185, Training Logs: loss_final: 0.534821, loss_mean: 0.855326, proj_loss: -0.401882, loss_mean_cls: 0.081377, grad_norm: 0.333171 +Steps: 0%| | 3186/1000000 [13:14<69:07:05, 4.01it/s, grad_norm=0.333, loss_final=0.535, loss_mean=0.855, loss_mean_cls=0.0814, proj_loss=-0.402][2026-03-26 12:44:18] Step: 3186, Training Logs: loss_final: 0.524887, loss_mean: 0.855010, proj_loss: -0.408384, loss_mean_cls: 0.078262, grad_norm: 0.526337 +Steps: 0%| | 3187/1000000 [13:14<68:56:50, 4.02it/s, grad_norm=0.526, loss_final=0.525, loss_mean=0.855, loss_mean_cls=0.0783, proj_loss=-0.408][2026-03-26 12:44:18] Step: 3187, Training Logs: loss_final: 0.521302, loss_mean: 0.838410, proj_loss: -0.399089, loss_mean_cls: 0.081981, grad_norm: 0.394490 +Steps: 0%| | 3188/1000000 [13:15<68:49:45, 4.02it/s, grad_norm=0.394, loss_final=0.521, loss_mean=0.838, loss_mean_cls=0.082, proj_loss=-0.399][2026-03-26 12:44:18] Step: 3188, Training Logs: loss_final: 0.506045, loss_mean: 0.838151, proj_loss: -0.412683, loss_mean_cls: 0.080577, grad_norm: 0.383677 +Steps: 0%| | 3189/1000000 [13:15<68:45:03, 4.03it/s, grad_norm=0.384, loss_final=0.506, loss_mean=0.838, loss_mean_cls=0.0806, proj_loss=-0.413][2026-03-26 12:44:18] Step: 3189, Training Logs: loss_final: 0.513913, loss_mean: 0.843393, proj_loss: -0.410040, loss_mean_cls: 0.080560, grad_norm: 0.469890 +Steps: 0%| | 3190/1000000 [13:15<68:42:25, 4.03it/s, grad_norm=0.47, loss_final=0.514, loss_mean=0.843, loss_mean_cls=0.0806, proj_loss=-0.41][2026-03-26 12:44:19] Step: 3190, Training Logs: loss_final: 0.509698, loss_mean: 0.837496, proj_loss: -0.409394, loss_mean_cls: 0.081597, grad_norm: 0.415262 +Steps: 0%| | 3191/1000000 [13:15<68:43:07, 4.03it/s, grad_norm=0.415, loss_final=0.51, loss_mean=0.837, loss_mean_cls=0.0816, proj_loss=-0.409][2026-03-26 12:44:19] Step: 3191, Training Logs: loss_final: 0.515052, loss_mean: 0.843749, proj_loss: -0.411130, loss_mean_cls: 0.082433, grad_norm: 0.811000 +Steps: 0%| | 3192/1000000 [13:16<68:39:21, 4.03it/s, grad_norm=0.811, loss_final=0.515, loss_mean=0.844, loss_mean_cls=0.0824, proj_loss=-0.411][2026-03-26 12:44:19] Step: 3192, Training Logs: loss_final: 0.522037, loss_mean: 0.845308, proj_loss: -0.404969, loss_mean_cls: 0.081698, grad_norm: 0.581374 +Steps: 0%| | 3193/1000000 [13:16<68:36:39, 4.04it/s, grad_norm=0.581, loss_final=0.522, loss_mean=0.845, loss_mean_cls=0.0817, proj_loss=-0.405][2026-03-26 12:44:19] Step: 3193, Training Logs: loss_final: 0.523247, loss_mean: 0.835062, proj_loss: -0.395375, loss_mean_cls: 0.083560, grad_norm: 0.503056 +Steps: 0%| | 3194/1000000 [13:16<68:34:38, 4.04it/s, grad_norm=0.503, loss_final=0.523, loss_mean=0.835, loss_mean_cls=0.0836, proj_loss=-0.395][2026-03-26 12:44:20] Step: 3194, Training Logs: loss_final: 0.528277, loss_mean: 0.837449, proj_loss: -0.393580, loss_mean_cls: 0.084409, grad_norm: 0.745364 +Steps: 0%| | 3195/1000000 [13:16<68:35:32, 4.04it/s, grad_norm=0.745, loss_final=0.528, loss_mean=0.837, loss_mean_cls=0.0844, proj_loss=-0.394][2026-03-26 12:44:20] Step: 3195, Training Logs: loss_final: 0.529863, loss_mean: 0.844044, proj_loss: -0.396954, loss_mean_cls: 0.082774, grad_norm: 0.329488 +Steps: 0%| | 3196/1000000 [13:17<68:34:50, 4.04it/s, grad_norm=0.329, loss_final=0.53, loss_mean=0.844, loss_mean_cls=0.0828, proj_loss=-0.397][2026-03-26 12:44:20] Step: 3196, Training Logs: loss_final: 0.514946, loss_mean: 0.838002, proj_loss: -0.404980, loss_mean_cls: 0.081924, grad_norm: 0.733035 +Steps: 0%| | 3197/1000000 [13:17<68:35:30, 4.04it/s, grad_norm=0.733, loss_final=0.515, loss_mean=0.838, loss_mean_cls=0.0819, proj_loss=-0.405][2026-03-26 12:44:20] Step: 3197, Training Logs: loss_final: 0.520323, loss_mean: 0.846344, proj_loss: -0.405765, loss_mean_cls: 0.079744, grad_norm: 0.745610 +Steps: 0%| | 3198/1000000 [13:17<68:35:46, 4.04it/s, grad_norm=0.746, loss_final=0.52, loss_mean=0.846, loss_mean_cls=0.0797, proj_loss=-0.406][2026-03-26 12:44:21] Step: 3198, Training Logs: loss_final: 0.491881, loss_mean: 0.818624, proj_loss: -0.407842, loss_mean_cls: 0.081099, grad_norm: 0.356098 +Steps: 0%| | 3199/1000000 [13:17<68:33:45, 4.04it/s, grad_norm=0.356, loss_final=0.492, loss_mean=0.819, loss_mean_cls=0.0811, proj_loss=-0.408][2026-03-26 12:44:21] Step: 3199, Training Logs: loss_final: 0.519472, loss_mean: 0.831648, proj_loss: -0.395038, loss_mean_cls: 0.082861, grad_norm: 0.546591 +Steps: 0%| | 3200/1000000 [13:18<68:38:01, 4.03it/s, grad_norm=0.547, loss_final=0.519, loss_mean=0.832, loss_mean_cls=0.0829, proj_loss=-0.395][2026-03-26 12:44:21] Step: 3200, Training Logs: loss_final: 0.496247, loss_mean: 0.825970, proj_loss: -0.410725, loss_mean_cls: 0.081002, grad_norm: 0.347649 +Steps: 0%| | 3201/1000000 [13:18<68:34:44, 4.04it/s, grad_norm=0.348, loss_final=0.496, loss_mean=0.826, loss_mean_cls=0.081, proj_loss=-0.411][2026-03-26 12:44:21] Step: 3201, Training Logs: loss_final: 0.522273, loss_mean: 0.857789, proj_loss: -0.414226, loss_mean_cls: 0.078709, grad_norm: 0.465311 +Steps: 0%| | 3202/1000000 [13:18<68:37:44, 4.03it/s, grad_norm=0.465, loss_final=0.522, loss_mean=0.858, loss_mean_cls=0.0787, proj_loss=-0.414][2026-03-26 12:44:22] Step: 3202, Training Logs: loss_final: 0.522220, loss_mean: 0.843245, proj_loss: -0.402854, loss_mean_cls: 0.081829, grad_norm: 0.574345 +Steps: 0%| | 3203/1000000 [13:18<68:33:02, 4.04it/s, grad_norm=0.574, loss_final=0.522, loss_mean=0.843, loss_mean_cls=0.0818, proj_loss=-0.403][2026-03-26 12:44:22] Step: 3203, Training Logs: loss_final: 0.513201, loss_mean: 0.836151, proj_loss: -0.404534, loss_mean_cls: 0.081584, grad_norm: 0.366579 +Steps: 0%| | 3204/1000000 [13:19<68:35:46, 4.04it/s, grad_norm=0.367, loss_final=0.513, loss_mean=0.836, loss_mean_cls=0.0816, proj_loss=-0.405][2026-03-26 12:44:22] Step: 3204, Training Logs: loss_final: 0.511047, loss_mean: 0.824285, proj_loss: -0.397452, loss_mean_cls: 0.084213, grad_norm: 0.565866 +Steps: 0%| | 3205/1000000 [13:19<68:31:29, 4.04it/s, grad_norm=0.566, loss_final=0.511, loss_mean=0.824, loss_mean_cls=0.0842, proj_loss=-0.397][2026-03-26 12:44:22] Step: 3205, Training Logs: loss_final: 0.525656, loss_mean: 0.851262, proj_loss: -0.406488, loss_mean_cls: 0.080882, grad_norm: 0.529082 +Steps: 0%| | 3206/1000000 [13:19<68:33:50, 4.04it/s, grad_norm=0.529, loss_final=0.526, loss_mean=0.851, loss_mean_cls=0.0809, proj_loss=-0.406][2026-03-26 12:44:23] Step: 3206, Training Logs: loss_final: 0.527454, loss_mean: 0.840958, proj_loss: -0.396482, loss_mean_cls: 0.082978, grad_norm: 0.426948 +Steps: 0%| | 3207/1000000 [13:19<68:31:56, 4.04it/s, grad_norm=0.427, loss_final=0.527, loss_mean=0.841, loss_mean_cls=0.083, proj_loss=-0.396][2026-03-26 12:44:23] Step: 3207, Training Logs: loss_final: 0.514416, loss_mean: 0.830649, proj_loss: -0.398060, loss_mean_cls: 0.081827, grad_norm: 0.599138 +Steps: 0%| | 3208/1000000 [13:20<68:34:36, 4.04it/s, grad_norm=0.599, loss_final=0.514, loss_mean=0.831, loss_mean_cls=0.0818, proj_loss=-0.398][2026-03-26 12:44:23] Step: 3208, Training Logs: loss_final: 0.533603, loss_mean: 0.866697, proj_loss: -0.411228, loss_mean_cls: 0.078133, grad_norm: 0.330551 +Steps: 0%| | 3209/1000000 [13:20<68:33:43, 4.04it/s, grad_norm=0.331, loss_final=0.534, loss_mean=0.867, loss_mean_cls=0.0781, proj_loss=-0.411][2026-03-26 12:44:23] Step: 3209, Training Logs: loss_final: 0.512354, loss_mean: 0.821568, proj_loss: -0.395737, loss_mean_cls: 0.086523, grad_norm: 0.825850 +Steps: 0%| | 3210/1000000 [13:20<68:34:12, 4.04it/s, grad_norm=0.826, loss_final=0.512, loss_mean=0.822, loss_mean_cls=0.0865, proj_loss=-0.396][2026-03-26 12:44:24] Step: 3210, Training Logs: loss_final: 0.519760, loss_mean: 0.838046, proj_loss: -0.401017, loss_mean_cls: 0.082731, grad_norm: 0.565386 +Steps: 0%| | 3211/1000000 [13:20<68:34:27, 4.04it/s, grad_norm=0.565, loss_final=0.52, loss_mean=0.838, loss_mean_cls=0.0827, proj_loss=-0.401][2026-03-26 12:44:24] Step: 3211, Training Logs: loss_final: 0.510484, loss_mean: 0.833143, proj_loss: -0.403332, loss_mean_cls: 0.080673, grad_norm: 0.567360 +Steps: 0%| | 3212/1000000 [13:21<68:34:40, 4.04it/s, grad_norm=0.567, loss_final=0.51, loss_mean=0.833, loss_mean_cls=0.0807, proj_loss=-0.403][2026-03-26 12:44:24] Step: 3212, Training Logs: loss_final: 0.523058, loss_mean: 0.840237, proj_loss: -0.398247, loss_mean_cls: 0.081068, grad_norm: 0.890223 +Steps: 0%| | 3213/1000000 [13:21<68:35:50, 4.04it/s, grad_norm=0.89, loss_final=0.523, loss_mean=0.84, loss_mean_cls=0.0811, proj_loss=-0.398][2026-03-26 12:44:24] Step: 3213, Training Logs: loss_final: 0.509905, loss_mean: 0.817012, proj_loss: -0.391056, loss_mean_cls: 0.083949, grad_norm: 0.434131 +Steps: 0%| | 3214/1000000 [13:21<68:36:06, 4.04it/s, grad_norm=0.434, loss_final=0.51, loss_mean=0.817, loss_mean_cls=0.0839, proj_loss=-0.391][2026-03-26 12:44:25] Step: 3214, Training Logs: loss_final: 0.518201, loss_mean: 0.840403, proj_loss: -0.403587, loss_mean_cls: 0.081385, grad_norm: 0.769017 +Steps: 0%| | 3215/1000000 [13:21<68:35:10, 4.04it/s, grad_norm=0.769, loss_final=0.518, loss_mean=0.84, loss_mean_cls=0.0814, proj_loss=-0.404][2026-03-26 12:44:25] Step: 3215, Training Logs: loss_final: 0.520213, loss_mean: 0.831844, proj_loss: -0.395985, loss_mean_cls: 0.084355, grad_norm: 0.703552 +Steps: 0%| | 3216/1000000 [13:22<68:33:06, 4.04it/s, grad_norm=0.704, loss_final=0.52, loss_mean=0.832, loss_mean_cls=0.0844, proj_loss=-0.396][2026-03-26 12:44:25] Step: 3216, Training Logs: loss_final: 0.511915, loss_mean: 0.820193, proj_loss: -0.392377, loss_mean_cls: 0.084099, grad_norm: 0.553643 +Steps: 0%| | 3217/1000000 [13:22<68:36:25, 4.04it/s, grad_norm=0.554, loss_final=0.512, loss_mean=0.82, loss_mean_cls=0.0841, proj_loss=-0.392][2026-03-26 12:44:25] Step: 3217, Training Logs: loss_final: 0.508802, loss_mean: 0.825234, proj_loss: -0.399690, loss_mean_cls: 0.083259, grad_norm: 0.922576 +Steps: 0%| | 3218/1000000 [13:22<68:33:51, 4.04it/s, grad_norm=0.923, loss_final=0.509, loss_mean=0.825, loss_mean_cls=0.0833, proj_loss=-0.4][2026-03-26 12:44:25] Step: 3218, Training Logs: loss_final: 0.521235, loss_mean: 0.852645, proj_loss: -0.410753, loss_mean_cls: 0.079343, grad_norm: 0.317436 +Steps: 0%| | 3219/1000000 [13:22<68:33:18, 4.04it/s, grad_norm=0.317, loss_final=0.521, loss_mean=0.853, loss_mean_cls=0.0793, proj_loss=-0.411][2026-03-26 12:44:26] Step: 3219, Training Logs: loss_final: 0.521497, loss_mean: 0.838100, proj_loss: -0.400348, loss_mean_cls: 0.083745, grad_norm: 0.922433 +Steps: 0%| | 3220/1000000 [13:23<69:10:52, 4.00it/s, grad_norm=0.922, loss_final=0.521, loss_mean=0.838, loss_mean_cls=0.0837, proj_loss=-0.4][2026-03-26 12:44:26] Step: 3220, Training Logs: loss_final: 0.510444, loss_mean: 0.833232, proj_loss: -0.404679, loss_mean_cls: 0.081891, grad_norm: 0.504521 +Steps: 0%| | 3221/1000000 [13:23<69:03:04, 4.01it/s, grad_norm=0.505, loss_final=0.51, loss_mean=0.833, loss_mean_cls=0.0819, proj_loss=-0.405][2026-03-26 12:44:26] Step: 3221, Training Logs: loss_final: 0.497605, loss_mean: 0.810997, proj_loss: -0.397601, loss_mean_cls: 0.084210, grad_norm: 0.660034 +Steps: 0%| | 3222/1000000 [13:23<68:56:33, 4.02it/s, grad_norm=0.66, loss_final=0.498, loss_mean=0.811, loss_mean_cls=0.0842, proj_loss=-0.398][2026-03-26 12:44:26] Step: 3222, Training Logs: loss_final: 0.516662, loss_mean: 0.836352, proj_loss: -0.401629, loss_mean_cls: 0.081939, grad_norm: 0.542819 +Steps: 0%| | 3223/1000000 [13:23<68:51:45, 4.02it/s, grad_norm=0.543, loss_final=0.517, loss_mean=0.836, loss_mean_cls=0.0819, proj_loss=-0.402][2026-03-26 12:44:27] Step: 3223, Training Logs: loss_final: 0.526524, loss_mean: 0.859137, proj_loss: -0.411570, loss_mean_cls: 0.078956, grad_norm: 0.479139 +Steps: 0%| | 3224/1000000 [13:24<68:48:31, 4.02it/s, grad_norm=0.479, loss_final=0.527, loss_mean=0.859, loss_mean_cls=0.079, proj_loss=-0.412][2026-03-26 12:44:27] Step: 3224, Training Logs: loss_final: 0.500907, loss_mean: 0.825919, proj_loss: -0.407171, loss_mean_cls: 0.082159, grad_norm: 0.671850 +Steps: 0%| | 3225/1000000 [13:24<68:43:55, 4.03it/s, grad_norm=0.672, loss_final=0.501, loss_mean=0.826, loss_mean_cls=0.0822, proj_loss=-0.407][2026-03-26 12:44:27] Step: 3225, Training Logs: loss_final: 0.519908, loss_mean: 0.837266, proj_loss: -0.399777, loss_mean_cls: 0.082419, grad_norm: 0.346331 +Steps: 0%| | 3226/1000000 [13:24<68:39:23, 4.03it/s, grad_norm=0.346, loss_final=0.52, loss_mean=0.837, loss_mean_cls=0.0824, proj_loss=-0.4][2026-03-26 12:44:27] Step: 3226, Training Logs: loss_final: 0.520200, loss_mean: 0.851740, proj_loss: -0.411434, loss_mean_cls: 0.079895, grad_norm: 0.629286 +Steps: 0%| | 3227/1000000 [13:24<68:37:42, 4.03it/s, grad_norm=0.629, loss_final=0.52, loss_mean=0.852, loss_mean_cls=0.0799, proj_loss=-0.411][2026-03-26 12:44:28] Step: 3227, Training Logs: loss_final: 0.524783, loss_mean: 0.856041, proj_loss: -0.409679, loss_mean_cls: 0.078420, grad_norm: 0.438958 +Steps: 0%| | 3228/1000000 [13:25<68:34:49, 4.04it/s, grad_norm=0.439, loss_final=0.525, loss_mean=0.856, loss_mean_cls=0.0784, proj_loss=-0.41][2026-03-26 12:44:28] Step: 3228, Training Logs: loss_final: 0.530370, loss_mean: 0.859799, proj_loss: -0.408948, loss_mean_cls: 0.079519, grad_norm: 0.383437 +Steps: 0%| | 3229/1000000 [13:25<68:38:00, 4.03it/s, grad_norm=0.383, loss_final=0.53, loss_mean=0.86, loss_mean_cls=0.0795, proj_loss=-0.409][2026-03-26 12:44:28] Step: 3229, Training Logs: loss_final: 0.516398, loss_mean: 0.842590, proj_loss: -0.406152, loss_mean_cls: 0.079960, grad_norm: 0.597428 +Steps: 0%| | 3230/1000000 [13:25<68:36:48, 4.04it/s, grad_norm=0.597, loss_final=0.516, loss_mean=0.843, loss_mean_cls=0.08, proj_loss=-0.406][2026-03-26 12:44:28] Step: 3230, Training Logs: loss_final: 0.508134, loss_mean: 0.832763, proj_loss: -0.407222, loss_mean_cls: 0.082593, grad_norm: 0.381602 +Steps: 0%| | 3231/1000000 [13:25<68:35:54, 4.04it/s, grad_norm=0.382, loss_final=0.508, loss_mean=0.833, loss_mean_cls=0.0826, proj_loss=-0.407][2026-03-26 12:44:29] Step: 3231, Training Logs: loss_final: 0.528438, loss_mean: 0.850258, proj_loss: -0.402787, loss_mean_cls: 0.080967, grad_norm: 0.615596 +Steps: 0%| | 3232/1000000 [13:26<68:37:45, 4.03it/s, grad_norm=0.616, loss_final=0.528, loss_mean=0.85, loss_mean_cls=0.081, proj_loss=-0.403][2026-03-26 12:44:29] Step: 3232, Training Logs: loss_final: 0.517790, loss_mean: 0.842671, proj_loss: -0.406590, loss_mean_cls: 0.081709, grad_norm: 0.536846 +Steps: 0%| | 3233/1000000 [13:26<68:35:35, 4.04it/s, grad_norm=0.537, loss_final=0.518, loss_mean=0.843, loss_mean_cls=0.0817, proj_loss=-0.407][2026-03-26 12:44:29] Step: 3233, Training Logs: loss_final: 0.503205, loss_mean: 0.827122, proj_loss: -0.405639, loss_mean_cls: 0.081722, grad_norm: 0.378972 +Steps: 0%| | 3234/1000000 [13:26<68:37:19, 4.03it/s, grad_norm=0.379, loss_final=0.503, loss_mean=0.827, loss_mean_cls=0.0817, proj_loss=-0.406][2026-03-26 12:44:29] Step: 3234, Training Logs: loss_final: 0.522747, loss_mean: 0.831508, proj_loss: -0.393670, loss_mean_cls: 0.084909, grad_norm: 0.792062 +Steps: 0%| | 3235/1000000 [13:26<68:37:29, 4.03it/s, grad_norm=0.792, loss_final=0.523, loss_mean=0.832, loss_mean_cls=0.0849, proj_loss=-0.394][2026-03-26 12:44:30] Step: 3235, Training Logs: loss_final: 0.514629, loss_mean: 0.824108, proj_loss: -0.393949, loss_mean_cls: 0.084470, grad_norm: 0.366727 +Steps: 0%| | 3236/1000000 [13:27<68:39:32, 4.03it/s, grad_norm=0.367, loss_final=0.515, loss_mean=0.824, loss_mean_cls=0.0845, proj_loss=-0.394][2026-03-26 12:44:30] Step: 3236, Training Logs: loss_final: 0.523159, loss_mean: 0.850026, proj_loss: -0.405952, loss_mean_cls: 0.079085, grad_norm: 0.535843 +Steps: 0%| | 3237/1000000 [13:27<68:36:50, 4.04it/s, grad_norm=0.536, loss_final=0.523, loss_mean=0.85, loss_mean_cls=0.0791, proj_loss=-0.406][2026-03-26 12:44:30] Step: 3237, Training Logs: loss_final: 0.519109, loss_mean: 0.830698, proj_loss: -0.394716, loss_mean_cls: 0.083127, grad_norm: 0.515438 +Steps: 0%| | 3238/1000000 [13:27<68:33:12, 4.04it/s, grad_norm=0.515, loss_final=0.519, loss_mean=0.831, loss_mean_cls=0.0831, proj_loss=-0.395][2026-03-26 12:44:30] Step: 3238, Training Logs: loss_final: 0.521761, loss_mean: 0.840693, proj_loss: -0.400766, loss_mean_cls: 0.081834, grad_norm: 0.344504 +Steps: 0%| | 3239/1000000 [13:27<68:32:00, 4.04it/s, grad_norm=0.345, loss_final=0.522, loss_mean=0.841, loss_mean_cls=0.0818, proj_loss=-0.401][2026-03-26 12:44:31] Step: 3239, Training Logs: loss_final: 0.513885, loss_mean: 0.832037, proj_loss: -0.401568, loss_mean_cls: 0.083416, grad_norm: 0.720816 +Steps: 0%| | 3240/1000000 [13:28<68:31:04, 4.04it/s, grad_norm=0.721, loss_final=0.514, loss_mean=0.832, loss_mean_cls=0.0834, proj_loss=-0.402][2026-03-26 12:44:31] Step: 3240, Training Logs: loss_final: 0.528507, loss_mean: 0.854489, proj_loss: -0.407518, loss_mean_cls: 0.081536, grad_norm: 0.675370 +Steps: 0%| | 3241/1000000 [13:28<68:30:45, 4.04it/s, grad_norm=0.675, loss_final=0.529, loss_mean=0.854, loss_mean_cls=0.0815, proj_loss=-0.408][2026-03-26 12:44:31] Step: 3241, Training Logs: loss_final: 0.504039, loss_mean: 0.830426, proj_loss: -0.407607, loss_mean_cls: 0.081220, grad_norm: 0.559805 +Steps: 0%| | 3242/1000000 [13:28<68:33:48, 4.04it/s, grad_norm=0.56, loss_final=0.504, loss_mean=0.83, loss_mean_cls=0.0812, proj_loss=-0.408][2026-03-26 12:44:31] Step: 3242, Training Logs: loss_final: 0.524798, loss_mean: 0.833716, proj_loss: -0.393022, loss_mean_cls: 0.084105, grad_norm: 0.516724 +Steps: 0%| | 3243/1000000 [13:28<68:33:00, 4.04it/s, grad_norm=0.517, loss_final=0.525, loss_mean=0.834, loss_mean_cls=0.0841, proj_loss=-0.393][2026-03-26 12:44:32] Step: 3243, Training Logs: loss_final: 0.505289, loss_mean: 0.827110, proj_loss: -0.404254, loss_mean_cls: 0.082433, grad_norm: 0.468223 +Steps: 0%| | 3244/1000000 [13:29<68:33:29, 4.04it/s, grad_norm=0.468, loss_final=0.505, loss_mean=0.827, loss_mean_cls=0.0824, proj_loss=-0.404][2026-03-26 12:44:32] Step: 3244, Training Logs: loss_final: 0.532771, loss_mean: 0.867923, proj_loss: -0.413355, loss_mean_cls: 0.078203, grad_norm: 0.612854 +Steps: 0%| | 3245/1000000 [13:29<68:29:00, 4.04it/s, grad_norm=0.613, loss_final=0.533, loss_mean=0.868, loss_mean_cls=0.0782, proj_loss=-0.413][2026-03-26 12:44:32] Step: 3245, Training Logs: loss_final: 0.519623, loss_mean: 0.845385, proj_loss: -0.406628, loss_mean_cls: 0.080866, grad_norm: 0.608868 +Steps: 0%| | 3246/1000000 [13:29<68:30:32, 4.04it/s, grad_norm=0.609, loss_final=0.52, loss_mean=0.845, loss_mean_cls=0.0809, proj_loss=-0.407][2026-03-26 12:44:32] Step: 3246, Training Logs: loss_final: 0.512699, loss_mean: 0.831602, proj_loss: -0.401552, loss_mean_cls: 0.082649, grad_norm: 0.326209 +Steps: 0%| | 3247/1000000 [13:29<69:06:13, 4.01it/s, grad_norm=0.326, loss_final=0.513, loss_mean=0.832, loss_mean_cls=0.0826, proj_loss=-0.402][2026-03-26 12:44:33] Step: 3247, Training Logs: loss_final: 0.511469, loss_mean: 0.836078, proj_loss: -0.403988, loss_mean_cls: 0.079380, grad_norm: 0.594826 +Steps: 0%| | 3248/1000000 [13:30<69:16:09, 4.00it/s, grad_norm=0.595, loss_final=0.511, loss_mean=0.836, loss_mean_cls=0.0794, proj_loss=-0.404][2026-03-26 12:44:33] Step: 3248, Training Logs: loss_final: 0.514814, loss_mean: 0.835070, proj_loss: -0.401856, loss_mean_cls: 0.081601, grad_norm: 0.455233 +Steps: 0%| | 3249/1000000 [13:30<68:59:58, 4.01it/s, grad_norm=0.455, loss_final=0.515, loss_mean=0.835, loss_mean_cls=0.0816, proj_loss=-0.402][2026-03-26 12:44:33] Step: 3249, Training Logs: loss_final: 0.515049, loss_mean: 0.835503, proj_loss: -0.403114, loss_mean_cls: 0.082660, grad_norm: 0.607790 +Steps: 0%| | 3250/1000000 [13:30<68:52:30, 4.02it/s, grad_norm=0.608, loss_final=0.515, loss_mean=0.836, loss_mean_cls=0.0827, proj_loss=-0.403][2026-03-26 12:44:33] Step: 3250, Training Logs: loss_final: 0.513996, loss_mean: 0.832274, proj_loss: -0.401536, loss_mean_cls: 0.083258, grad_norm: 0.445076 +Steps: 0%| | 3251/1000000 [13:30<68:43:47, 4.03it/s, grad_norm=0.445, loss_final=0.514, loss_mean=0.832, loss_mean_cls=0.0833, proj_loss=-0.402][2026-03-26 12:44:34] Step: 3251, Training Logs: loss_final: 0.505509, loss_mean: 0.825559, proj_loss: -0.402051, loss_mean_cls: 0.082000, grad_norm: 0.491557 +Steps: 0%| | 3252/1000000 [13:31<68:37:53, 4.03it/s, grad_norm=0.492, loss_final=0.506, loss_mean=0.826, loss_mean_cls=0.082, proj_loss=-0.402][2026-03-26 12:44:34] Step: 3252, Training Logs: loss_final: 0.517966, loss_mean: 0.831846, proj_loss: -0.397235, loss_mean_cls: 0.083355, grad_norm: 0.486529 +Steps: 0%| | 3253/1000000 [13:31<68:33:57, 4.04it/s, grad_norm=0.487, loss_final=0.518, loss_mean=0.832, loss_mean_cls=0.0834, proj_loss=-0.397][2026-03-26 12:44:34] Step: 3253, Training Logs: loss_final: 0.518605, loss_mean: 0.842866, proj_loss: -0.403502, loss_mean_cls: 0.079242, grad_norm: 0.370686 +Steps: 0%| | 3254/1000000 [13:31<68:35:45, 4.04it/s, grad_norm=0.371, loss_final=0.519, loss_mean=0.843, loss_mean_cls=0.0792, proj_loss=-0.404][2026-03-26 12:44:34] Step: 3254, Training Logs: loss_final: 0.519938, loss_mean: 0.845923, proj_loss: -0.404858, loss_mean_cls: 0.078873, grad_norm: 0.567632 +Steps: 0%| | 3255/1000000 [13:31<68:33:15, 4.04it/s, grad_norm=0.568, loss_final=0.52, loss_mean=0.846, loss_mean_cls=0.0789, proj_loss=-0.405][2026-03-26 12:44:35] Step: 3255, Training Logs: loss_final: 0.504664, loss_mean: 0.822821, proj_loss: -0.400583, loss_mean_cls: 0.082426, grad_norm: 0.534576 +Steps: 0%| | 3256/1000000 [13:32<68:35:38, 4.04it/s, grad_norm=0.535, loss_final=0.505, loss_mean=0.823, loss_mean_cls=0.0824, proj_loss=-0.401][2026-03-26 12:44:35] Step: 3256, Training Logs: loss_final: 0.533638, loss_mean: 0.863369, proj_loss: -0.409577, loss_mean_cls: 0.079847, grad_norm: 0.338738 +Steps: 0%| | 3257/1000000 [13:32<68:34:15, 4.04it/s, grad_norm=0.339, loss_final=0.534, loss_mean=0.863, loss_mean_cls=0.0798, proj_loss=-0.41][2026-03-26 12:44:35] Step: 3257, Training Logs: loss_final: 0.521426, loss_mean: 0.841618, proj_loss: -0.401958, loss_mean_cls: 0.081766, grad_norm: 0.508347 +Steps: 0%| | 3258/1000000 [13:32<68:32:13, 4.04it/s, grad_norm=0.508, loss_final=0.521, loss_mean=0.842, loss_mean_cls=0.0818, proj_loss=-0.402][2026-03-26 12:44:35] Step: 3258, Training Logs: loss_final: 0.514098, loss_mean: 0.829863, proj_loss: -0.398693, loss_mean_cls: 0.082928, grad_norm: 0.389751 +Steps: 0%| | 3259/1000000 [13:32<68:34:05, 4.04it/s, grad_norm=0.39, loss_final=0.514, loss_mean=0.83, loss_mean_cls=0.0829, proj_loss=-0.399][2026-03-26 12:44:36] Step: 3259, Training Logs: loss_final: 0.505522, loss_mean: 0.820056, proj_loss: -0.397227, loss_mean_cls: 0.082693, grad_norm: 0.361817 +Steps: 0%| | 3260/1000000 [13:33<68:33:27, 4.04it/s, grad_norm=0.362, loss_final=0.506, loss_mean=0.82, loss_mean_cls=0.0827, proj_loss=-0.397][2026-03-26 12:44:36] Step: 3260, Training Logs: loss_final: 0.527267, loss_mean: 0.848708, proj_loss: -0.402860, loss_mean_cls: 0.081418, grad_norm: 0.486341 +Steps: 0%| | 3261/1000000 [13:33<68:32:03, 4.04it/s, grad_norm=0.486, loss_final=0.527, loss_mean=0.849, loss_mean_cls=0.0814, proj_loss=-0.403][2026-03-26 12:44:36] Step: 3261, Training Logs: loss_final: 0.536598, loss_mean: 0.872573, proj_loss: -0.414324, loss_mean_cls: 0.078349, grad_norm: 0.417823 +Steps: 0%| | 3262/1000000 [13:33<68:31:45, 4.04it/s, grad_norm=0.418, loss_final=0.537, loss_mean=0.873, loss_mean_cls=0.0783, proj_loss=-0.414][2026-03-26 12:44:36] Step: 3262, Training Logs: loss_final: 0.529432, loss_mean: 0.854234, proj_loss: -0.404330, loss_mean_cls: 0.079527, grad_norm: 0.318866 +Steps: 0%| | 3263/1000000 [13:33<68:33:58, 4.04it/s, grad_norm=0.319, loss_final=0.529, loss_mean=0.854, loss_mean_cls=0.0795, proj_loss=-0.404][2026-03-26 12:44:37] Step: 3263, Training Logs: loss_final: 0.508603, loss_mean: 0.831811, proj_loss: -0.405954, loss_mean_cls: 0.082746, grad_norm: 0.498755 +Steps: 0%| | 3264/1000000 [13:34<68:33:52, 4.04it/s, grad_norm=0.499, loss_final=0.509, loss_mean=0.832, loss_mean_cls=0.0827, proj_loss=-0.406][2026-03-26 12:44:37] Step: 3264, Training Logs: loss_final: 0.505080, loss_mean: 0.819451, proj_loss: -0.398262, loss_mean_cls: 0.083891, grad_norm: 0.377922 +Steps: 0%| | 3265/1000000 [13:34<68:32:27, 4.04it/s, grad_norm=0.378, loss_final=0.505, loss_mean=0.819, loss_mean_cls=0.0839, proj_loss=-0.398][2026-03-26 12:44:37] Step: 3265, Training Logs: loss_final: 0.500727, loss_mean: 0.820518, proj_loss: -0.403173, loss_mean_cls: 0.083381, grad_norm: 0.414572 +Steps: 0%| | 3266/1000000 [13:34<68:31:18, 4.04it/s, grad_norm=0.415, loss_final=0.501, loss_mean=0.821, loss_mean_cls=0.0834, proj_loss=-0.403][2026-03-26 12:44:37] Step: 3266, Training Logs: loss_final: 0.531886, loss_mean: 0.853567, proj_loss: -0.403312, loss_mean_cls: 0.081631, grad_norm: 0.413871 +Steps: 0%| | 3267/1000000 [13:34<68:34:43, 4.04it/s, grad_norm=0.414, loss_final=0.532, loss_mean=0.854, loss_mean_cls=0.0816, proj_loss=-0.403][2026-03-26 12:44:38] Step: 3267, Training Logs: loss_final: 0.503215, loss_mean: 0.821734, proj_loss: -0.401386, loss_mean_cls: 0.082867, grad_norm: 0.294838 +Steps: 0%| | 3268/1000000 [13:35<68:34:23, 4.04it/s, grad_norm=0.295, loss_final=0.503, loss_mean=0.822, loss_mean_cls=0.0829, proj_loss=-0.401][2026-03-26 12:44:38] Step: 3268, Training Logs: loss_final: 0.528052, loss_mean: 0.838439, proj_loss: -0.394278, loss_mean_cls: 0.083891, grad_norm: 0.470496 +Steps: 0%| | 3269/1000000 [13:35<68:36:58, 4.04it/s, grad_norm=0.47, loss_final=0.528, loss_mean=0.838, loss_mean_cls=0.0839, proj_loss=-0.394][2026-03-26 12:44:38] Step: 3269, Training Logs: loss_final: 0.511396, loss_mean: 0.828805, proj_loss: -0.400128, loss_mean_cls: 0.082718, grad_norm: 0.287699 +Steps: 0%| | 3270/1000000 [13:35<68:36:10, 4.04it/s, grad_norm=0.288, loss_final=0.511, loss_mean=0.829, loss_mean_cls=0.0827, proj_loss=-0.4][2026-03-26 12:44:38] Step: 3270, Training Logs: loss_final: 0.507098, loss_mean: 0.839606, proj_loss: -0.411197, loss_mean_cls: 0.078689, grad_norm: 0.337351 +Steps: 0%| | 3271/1000000 [13:35<68:37:25, 4.03it/s, grad_norm=0.337, loss_final=0.507, loss_mean=0.84, loss_mean_cls=0.0787, proj_loss=-0.411][2026-03-26 12:44:39] Step: 3271, Training Logs: loss_final: 0.520511, loss_mean: 0.835637, proj_loss: -0.399220, loss_mean_cls: 0.084094, grad_norm: 0.397418 +Steps: 0%| | 3272/1000000 [13:35<68:34:14, 4.04it/s, grad_norm=0.397, loss_final=0.521, loss_mean=0.836, loss_mean_cls=0.0841, proj_loss=-0.399][2026-03-26 12:44:39] Step: 3272, Training Logs: loss_final: 0.518280, loss_mean: 0.829965, proj_loss: -0.393787, loss_mean_cls: 0.082103, grad_norm: 0.363138 +Steps: 0%| | 3273/1000000 [13:36<68:35:46, 4.04it/s, grad_norm=0.363, loss_final=0.518, loss_mean=0.83, loss_mean_cls=0.0821, proj_loss=-0.394][2026-03-26 12:44:39] Step: 3273, Training Logs: loss_final: 0.522598, loss_mean: 0.839415, proj_loss: -0.399083, loss_mean_cls: 0.082266, grad_norm: 0.355100 +Steps: 0%| | 3274/1000000 [13:36<68:34:29, 4.04it/s, grad_norm=0.355, loss_final=0.523, loss_mean=0.839, loss_mean_cls=0.0823, proj_loss=-0.399][2026-03-26 12:44:39] Step: 3274, Training Logs: loss_final: 0.525246, loss_mean: 0.846114, proj_loss: -0.402129, loss_mean_cls: 0.081261, grad_norm: 0.345463 +Steps: 0%| | 3275/1000000 [13:36<68:33:03, 4.04it/s, grad_norm=0.345, loss_final=0.525, loss_mean=0.846, loss_mean_cls=0.0813, proj_loss=-0.402][2026-03-26 12:44:40] Step: 3275, Training Logs: loss_final: 0.519255, loss_mean: 0.831081, proj_loss: -0.396600, loss_mean_cls: 0.084774, grad_norm: 0.359896 +Steps: 0%| | 3276/1000000 [13:36<68:31:56, 4.04it/s, grad_norm=0.36, loss_final=0.519, loss_mean=0.831, loss_mean_cls=0.0848, proj_loss=-0.397][2026-03-26 12:44:40] Step: 3276, Training Logs: loss_final: 0.516071, loss_mean: 0.835681, proj_loss: -0.402076, loss_mean_cls: 0.082467, grad_norm: 0.327147 +Steps: 0%| | 3277/1000000 [13:37<68:31:44, 4.04it/s, grad_norm=0.327, loss_final=0.516, loss_mean=0.836, loss_mean_cls=0.0825, proj_loss=-0.402][2026-03-26 12:44:40] Step: 3277, Training Logs: loss_final: 0.517145, loss_mean: 0.844225, proj_loss: -0.408667, loss_mean_cls: 0.081586, grad_norm: 0.445929 +Steps: 0%| | 3278/1000000 [13:37<68:32:31, 4.04it/s, grad_norm=0.446, loss_final=0.517, loss_mean=0.844, loss_mean_cls=0.0816, proj_loss=-0.409][2026-03-26 12:44:40] Step: 3278, Training Logs: loss_final: 0.497046, loss_mean: 0.818022, proj_loss: -0.404369, loss_mean_cls: 0.083394, grad_norm: 0.415712 +Steps: 0%| | 3279/1000000 [13:37<68:36:51, 4.04it/s, grad_norm=0.416, loss_final=0.497, loss_mean=0.818, loss_mean_cls=0.0834, proj_loss=-0.404][2026-03-26 12:44:41] Step: 3279, Training Logs: loss_final: 0.520591, loss_mean: 0.847458, proj_loss: -0.406713, loss_mean_cls: 0.079845, grad_norm: 0.443678 +Steps: 0%| | 3280/1000000 [13:37<68:36:44, 4.04it/s, grad_norm=0.444, loss_final=0.521, loss_mean=0.847, loss_mean_cls=0.0798, proj_loss=-0.407][2026-03-26 12:44:41] Step: 3280, Training Logs: loss_final: 0.528910, loss_mean: 0.873649, proj_loss: -0.420527, loss_mean_cls: 0.075788, grad_norm: 0.322209 +Steps: 0%| | 3281/1000000 [13:38<68:37:26, 4.03it/s, grad_norm=0.322, loss_final=0.529, loss_mean=0.874, loss_mean_cls=0.0758, proj_loss=-0.421][2026-03-26 12:44:41] Step: 3281, Training Logs: loss_final: 0.533749, loss_mean: 0.850551, proj_loss: -0.400623, loss_mean_cls: 0.083821, grad_norm: 0.369917 +Steps: 0%| | 3282/1000000 [13:38<68:36:50, 4.04it/s, grad_norm=0.37, loss_final=0.534, loss_mean=0.851, loss_mean_cls=0.0838, proj_loss=-0.401][2026-03-26 12:44:41] Step: 3282, Training Logs: loss_final: 0.516012, loss_mean: 0.828245, proj_loss: -0.396857, loss_mean_cls: 0.084624, grad_norm: 0.421657 +Steps: 0%| | 3283/1000000 [13:38<68:34:47, 4.04it/s, grad_norm=0.422, loss_final=0.516, loss_mean=0.828, loss_mean_cls=0.0846, proj_loss=-0.397][2026-03-26 12:44:42] Step: 3283, Training Logs: loss_final: 0.528248, loss_mean: 0.852759, proj_loss: -0.406311, loss_mean_cls: 0.081800, grad_norm: 0.446408 +Steps: 0%| | 3284/1000000 [13:38<68:37:22, 4.03it/s, grad_norm=0.446, loss_final=0.528, loss_mean=0.853, loss_mean_cls=0.0818, proj_loss=-0.406][2026-03-26 12:44:42] Step: 3284, Training Logs: loss_final: 0.499550, loss_mean: 0.824285, proj_loss: -0.406345, loss_mean_cls: 0.081610, grad_norm: 0.410177 +Steps: 0%| | 3285/1000000 [13:39<68:32:16, 4.04it/s, grad_norm=0.41, loss_final=0.5, loss_mean=0.824, loss_mean_cls=0.0816, proj_loss=-0.406][2026-03-26 12:44:42] Step: 3285, Training Logs: loss_final: 0.499586, loss_mean: 0.812406, proj_loss: -0.397393, loss_mean_cls: 0.084574, grad_norm: 0.593107 +Steps: 0%| | 3286/1000000 [13:39<68:34:16, 4.04it/s, grad_norm=0.593, loss_final=0.5, loss_mean=0.812, loss_mean_cls=0.0846, proj_loss=-0.397][2026-03-26 12:44:42] Step: 3286, Training Logs: loss_final: 0.508708, loss_mean: 0.829913, proj_loss: -0.402598, loss_mean_cls: 0.081393, grad_norm: 0.569043 +Steps: 0%| | 3287/1000000 [13:39<68:35:36, 4.04it/s, grad_norm=0.569, loss_final=0.509, loss_mean=0.83, loss_mean_cls=0.0814, proj_loss=-0.403][2026-03-26 12:44:43] Step: 3287, Training Logs: loss_final: 0.515065, loss_mean: 0.842522, proj_loss: -0.407254, loss_mean_cls: 0.079797, grad_norm: 0.395907 +Steps: 0%| | 3288/1000000 [13:39<68:33:16, 4.04it/s, grad_norm=0.396, loss_final=0.515, loss_mean=0.843, loss_mean_cls=0.0798, proj_loss=-0.407][2026-03-26 12:44:43] Step: 3288, Training Logs: loss_final: 0.518124, loss_mean: 0.838543, proj_loss: -0.402008, loss_mean_cls: 0.081589, grad_norm: 0.567289 +Steps: 0%| | 3289/1000000 [13:40<68:33:02, 4.04it/s, grad_norm=0.567, loss_final=0.518, loss_mean=0.839, loss_mean_cls=0.0816, proj_loss=-0.402][2026-03-26 12:44:43] Step: 3289, Training Logs: loss_final: 0.508207, loss_mean: 0.837208, proj_loss: -0.410178, loss_mean_cls: 0.081177, grad_norm: 0.392684 +Steps: 0%| | 3290/1000000 [13:40<68:33:48, 4.04it/s, grad_norm=0.393, loss_final=0.508, loss_mean=0.837, loss_mean_cls=0.0812, proj_loss=-0.41][2026-03-26 12:44:43] Step: 3290, Training Logs: loss_final: 0.505135, loss_mean: 0.820162, proj_loss: -0.398053, loss_mean_cls: 0.083025, grad_norm: 0.398159 +Steps: 0%| | 3291/1000000 [13:40<68:32:29, 4.04it/s, grad_norm=0.398, loss_final=0.505, loss_mean=0.82, loss_mean_cls=0.083, proj_loss=-0.398][2026-03-26 12:44:44] Step: 3291, Training Logs: loss_final: 0.520238, loss_mean: 0.845133, proj_loss: -0.407200, loss_mean_cls: 0.082306, grad_norm: 0.520163 +Steps: 0%| | 3292/1000000 [13:40<68:32:14, 4.04it/s, grad_norm=0.52, loss_final=0.52, loss_mean=0.845, loss_mean_cls=0.0823, proj_loss=-0.407][2026-03-26 12:44:44] Step: 3292, Training Logs: loss_final: 0.517279, loss_mean: 0.850456, proj_loss: -0.412502, loss_mean_cls: 0.079326, grad_norm: 0.319682 +Steps: 0%| | 3293/1000000 [13:41<69:21:13, 3.99it/s, grad_norm=0.32, loss_final=0.517, loss_mean=0.85, loss_mean_cls=0.0793, proj_loss=-0.413][2026-03-26 12:44:44] Step: 3293, Training Logs: loss_final: 0.531162, loss_mean: 0.852584, proj_loss: -0.402264, loss_mean_cls: 0.080842, grad_norm: 0.620092 +Steps: 0%| | 3294/1000000 [13:41<69:08:22, 4.00it/s, grad_norm=0.62, loss_final=0.531, loss_mean=0.853, loss_mean_cls=0.0808, proj_loss=-0.402][2026-03-26 12:44:44] Step: 3294, Training Logs: loss_final: 0.515287, loss_mean: 0.837730, proj_loss: -0.402802, loss_mean_cls: 0.080360, grad_norm: 0.647451 +Steps: 0%| | 3295/1000000 [13:41<68:54:41, 4.02it/s, grad_norm=0.647, loss_final=0.515, loss_mean=0.838, loss_mean_cls=0.0804, proj_loss=-0.403][2026-03-26 12:44:45] Step: 3295, Training Logs: loss_final: 0.511849, loss_mean: 0.842912, proj_loss: -0.411839, loss_mean_cls: 0.080776, grad_norm: 0.463460 +Steps: 0%| | 3296/1000000 [13:41<68:48:05, 4.02it/s, grad_norm=0.463, loss_final=0.512, loss_mean=0.843, loss_mean_cls=0.0808, proj_loss=-0.412][2026-03-26 12:44:45] Step: 3296, Training Logs: loss_final: 0.538401, loss_mean: 0.860731, proj_loss: -0.403063, loss_mean_cls: 0.080733, grad_norm: 0.445641 +Steps: 0%| | 3297/1000000 [13:42<68:40:45, 4.03it/s, grad_norm=0.446, loss_final=0.538, loss_mean=0.861, loss_mean_cls=0.0807, proj_loss=-0.403][2026-03-26 12:44:45] Step: 3297, Training Logs: loss_final: 0.505261, loss_mean: 0.845179, proj_loss: -0.416754, loss_mean_cls: 0.076836, grad_norm: 0.717643 +Steps: 0%| | 3298/1000000 [13:42<68:39:31, 4.03it/s, grad_norm=0.718, loss_final=0.505, loss_mean=0.845, loss_mean_cls=0.0768, proj_loss=-0.417][2026-03-26 12:44:45] Step: 3298, Training Logs: loss_final: 0.515629, loss_mean: 0.840717, proj_loss: -0.406348, loss_mean_cls: 0.081259, grad_norm: 0.423817 +Steps: 0%| | 3299/1000000 [13:42<68:30:38, 4.04it/s, grad_norm=0.424, loss_final=0.516, loss_mean=0.841, loss_mean_cls=0.0813, proj_loss=-0.406][2026-03-26 12:44:46] Step: 3299, Training Logs: loss_final: 0.512622, loss_mean: 0.843961, proj_loss: -0.411556, loss_mean_cls: 0.080217, grad_norm: 0.488847 +Steps: 0%| | 3300/1000000 [13:42<68:31:29, 4.04it/s, grad_norm=0.489, loss_final=0.513, loss_mean=0.844, loss_mean_cls=0.0802, proj_loss=-0.412][2026-03-26 12:44:46] Step: 3300, Training Logs: loss_final: 0.530471, loss_mean: 0.864246, proj_loss: -0.411062, loss_mean_cls: 0.077287, grad_norm: 0.591202 +Steps: 0%| | 3301/1000000 [13:43<68:30:08, 4.04it/s, grad_norm=0.591, loss_final=0.53, loss_mean=0.864, loss_mean_cls=0.0773, proj_loss=-0.411][2026-03-26 12:44:46] Step: 3301, Training Logs: loss_final: 0.511991, loss_mean: 0.838679, proj_loss: -0.408703, loss_mean_cls: 0.082015, grad_norm: 0.426168 +Steps: 0%| | 3302/1000000 [13:43<68:31:39, 4.04it/s, grad_norm=0.426, loss_final=0.512, loss_mean=0.839, loss_mean_cls=0.082, proj_loss=-0.409][2026-03-26 12:44:46] Step: 3302, Training Logs: loss_final: 0.503757, loss_mean: 0.821681, proj_loss: -0.402125, loss_mean_cls: 0.084201, grad_norm: 0.536719 +Steps: 0%| | 3303/1000000 [13:43<68:34:24, 4.04it/s, grad_norm=0.537, loss_final=0.504, loss_mean=0.822, loss_mean_cls=0.0842, proj_loss=-0.402][2026-03-26 12:44:47] Step: 3303, Training Logs: loss_final: 0.511430, loss_mean: 0.840697, proj_loss: -0.410082, loss_mean_cls: 0.080814, grad_norm: 0.480427 +Steps: 0%| | 3304/1000000 [13:43<68:31:35, 4.04it/s, grad_norm=0.48, loss_final=0.511, loss_mean=0.841, loss_mean_cls=0.0808, proj_loss=-0.41][2026-03-26 12:44:47] Step: 3304, Training Logs: loss_final: 0.519674, loss_mean: 0.848339, proj_loss: -0.408475, loss_mean_cls: 0.079810, grad_norm: 0.502371 +Steps: 0%| | 3305/1000000 [13:44<68:30:06, 4.04it/s, grad_norm=0.502, loss_final=0.52, loss_mean=0.848, loss_mean_cls=0.0798, proj_loss=-0.408][2026-03-26 12:44:47] Step: 3305, Training Logs: loss_final: 0.512861, loss_mean: 0.837971, proj_loss: -0.406043, loss_mean_cls: 0.080932, grad_norm: 0.569095 +Steps: 0%| | 3306/1000000 [13:44<68:31:00, 4.04it/s, grad_norm=0.569, loss_final=0.513, loss_mean=0.838, loss_mean_cls=0.0809, proj_loss=-0.406][2026-03-26 12:44:47] Step: 3306, Training Logs: loss_final: 0.544041, loss_mean: 0.867869, proj_loss: -0.404100, loss_mean_cls: 0.080272, grad_norm: 0.404951 +Steps: 0%| | 3307/1000000 [13:44<68:29:02, 4.04it/s, grad_norm=0.405, loss_final=0.544, loss_mean=0.868, loss_mean_cls=0.0803, proj_loss=-0.404][2026-03-26 12:44:48] Step: 3307, Training Logs: loss_final: 0.515585, loss_mean: 0.835307, proj_loss: -0.402411, loss_mean_cls: 0.082689, grad_norm: 0.446250 +Steps: 0%| | 3308/1000000 [13:44<68:29:18, 4.04it/s, grad_norm=0.446, loss_final=0.516, loss_mean=0.835, loss_mean_cls=0.0827, proj_loss=-0.402][2026-03-26 12:44:48] Step: 3308, Training Logs: loss_final: 0.528571, loss_mean: 0.849938, proj_loss: -0.403648, loss_mean_cls: 0.082281, grad_norm: 0.399365 +Steps: 0%| | 3309/1000000 [13:45<68:30:48, 4.04it/s, grad_norm=0.399, loss_final=0.529, loss_mean=0.85, loss_mean_cls=0.0823, proj_loss=-0.404][2026-03-26 12:44:48] Step: 3309, Training Logs: loss_final: 0.508240, loss_mean: 0.804767, proj_loss: -0.382673, loss_mean_cls: 0.086145, grad_norm: 0.471637 +Steps: 0%| | 3310/1000000 [13:45<68:31:03, 4.04it/s, grad_norm=0.472, loss_final=0.508, loss_mean=0.805, loss_mean_cls=0.0861, proj_loss=-0.383][2026-03-26 12:44:48] Step: 3310, Training Logs: loss_final: 0.517559, loss_mean: 0.842538, proj_loss: -0.406528, loss_mean_cls: 0.081548, grad_norm: 0.477762 +Steps: 0%| | 3311/1000000 [13:45<68:32:38, 4.04it/s, grad_norm=0.478, loss_final=0.518, loss_mean=0.843, loss_mean_cls=0.0815, proj_loss=-0.407][2026-03-26 12:44:49] Step: 3311, Training Logs: loss_final: 0.517972, loss_mean: 0.836646, proj_loss: -0.400487, loss_mean_cls: 0.081813, grad_norm: 0.467263 +Steps: 0%| | 3312/1000000 [13:45<68:31:44, 4.04it/s, grad_norm=0.467, loss_final=0.518, loss_mean=0.837, loss_mean_cls=0.0818, proj_loss=-0.4][2026-03-26 12:44:49] Step: 3312, Training Logs: loss_final: 0.527965, loss_mean: 0.851328, proj_loss: -0.404653, loss_mean_cls: 0.081290, grad_norm: 0.441639 +Steps: 0%| | 3313/1000000 [13:46<68:32:09, 4.04it/s, grad_norm=0.442, loss_final=0.528, loss_mean=0.851, loss_mean_cls=0.0813, proj_loss=-0.405][2026-03-26 12:44:49] Step: 3313, Training Logs: loss_final: 0.514424, loss_mean: 0.846576, proj_loss: -0.411984, loss_mean_cls: 0.079832, grad_norm: 0.491538 +Steps: 0%| | 3314/1000000 [13:46<68:29:24, 4.04it/s, grad_norm=0.492, loss_final=0.514, loss_mean=0.847, loss_mean_cls=0.0798, proj_loss=-0.412][2026-03-26 12:44:49] Step: 3314, Training Logs: loss_final: 0.523345, loss_mean: 0.840311, proj_loss: -0.399245, loss_mean_cls: 0.082279, grad_norm: 0.369750 +Steps: 0%| | 3315/1000000 [13:46<68:29:13, 4.04it/s, grad_norm=0.37, loss_final=0.523, loss_mean=0.84, loss_mean_cls=0.0823, proj_loss=-0.399][2026-03-26 12:44:50] Step: 3315, Training Logs: loss_final: 0.524021, loss_mean: 0.846086, proj_loss: -0.403576, loss_mean_cls: 0.081511, grad_norm: 0.395857 +Steps: 0%| | 3316/1000000 [13:46<68:26:36, 4.05it/s, grad_norm=0.396, loss_final=0.524, loss_mean=0.846, loss_mean_cls=0.0815, proj_loss=-0.404][2026-03-26 12:44:50] Step: 3316, Training Logs: loss_final: 0.528114, loss_mean: 0.839265, proj_loss: -0.394483, loss_mean_cls: 0.083331, grad_norm: 0.434877 +Steps: 0%| | 3317/1000000 [13:47<68:27:47, 4.04it/s, grad_norm=0.435, loss_final=0.528, loss_mean=0.839, loss_mean_cls=0.0833, proj_loss=-0.394][2026-03-26 12:44:50] Step: 3317, Training Logs: loss_final: 0.527451, loss_mean: 0.839201, proj_loss: -0.394692, loss_mean_cls: 0.082942, grad_norm: 0.421540 +Steps: 0%| | 3318/1000000 [13:47<68:27:46, 4.04it/s, grad_norm=0.422, loss_final=0.527, loss_mean=0.839, loss_mean_cls=0.0829, proj_loss=-0.395][2026-03-26 12:44:50] Step: 3318, Training Logs: loss_final: 0.532931, loss_mean: 0.860670, proj_loss: -0.407176, loss_mean_cls: 0.079437, grad_norm: 0.301553 +Steps: 0%| | 3319/1000000 [13:47<68:25:28, 4.05it/s, grad_norm=0.302, loss_final=0.533, loss_mean=0.861, loss_mean_cls=0.0794, proj_loss=-0.407][2026-03-26 12:44:51] Step: 3319, Training Logs: loss_final: 0.503342, loss_mean: 0.826350, proj_loss: -0.405291, loss_mean_cls: 0.082283, grad_norm: 0.391685 +Steps: 0%| | 3320/1000000 [13:47<68:30:21, 4.04it/s, grad_norm=0.392, loss_final=0.503, loss_mean=0.826, loss_mean_cls=0.0823, proj_loss=-0.405][2026-03-26 12:44:51] Step: 3320, Training Logs: loss_final: 0.522191, loss_mean: 0.840272, proj_loss: -0.399791, loss_mean_cls: 0.081710, grad_norm: 0.415365 +Steps: 0%| | 3321/1000000 [13:48<68:32:43, 4.04it/s, grad_norm=0.415, loss_final=0.522, loss_mean=0.84, loss_mean_cls=0.0817, proj_loss=-0.4][2026-03-26 12:44:51] Step: 3321, Training Logs: loss_final: 0.513194, loss_mean: 0.841509, proj_loss: -0.409527, loss_mean_cls: 0.081213, grad_norm: 0.427820 +Steps: 0%| | 3322/1000000 [13:48<68:31:49, 4.04it/s, grad_norm=0.428, loss_final=0.513, loss_mean=0.842, loss_mean_cls=0.0812, proj_loss=-0.41][2026-03-26 12:44:51] Step: 3322, Training Logs: loss_final: 0.522855, loss_mean: 0.843300, proj_loss: -0.402353, loss_mean_cls: 0.081908, grad_norm: 0.343175 +Steps: 0%| | 3323/1000000 [13:48<68:29:51, 4.04it/s, grad_norm=0.343, loss_final=0.523, loss_mean=0.843, loss_mean_cls=0.0819, proj_loss=-0.402][2026-03-26 12:44:52] Step: 3323, Training Logs: loss_final: 0.515355, loss_mean: 0.852924, proj_loss: -0.415652, loss_mean_cls: 0.078082, grad_norm: 0.382489 +Steps: 0%| | 3324/1000000 [13:48<68:36:15, 4.04it/s, grad_norm=0.382, loss_final=0.515, loss_mean=0.853, loss_mean_cls=0.0781, proj_loss=-0.416][2026-03-26 12:44:52] Step: 3324, Training Logs: loss_final: 0.523731, loss_mean: 0.841694, proj_loss: -0.400128, loss_mean_cls: 0.082165, grad_norm: 0.389315 +Steps: 0%| | 3325/1000000 [13:49<68:36:10, 4.04it/s, grad_norm=0.389, loss_final=0.524, loss_mean=0.842, loss_mean_cls=0.0822, proj_loss=-0.4][2026-03-26 12:44:52] Step: 3325, Training Logs: loss_final: 0.509863, loss_mean: 0.819414, proj_loss: -0.394015, loss_mean_cls: 0.084464, grad_norm: 0.354117 +Steps: 0%| | 3326/1000000 [13:49<68:35:31, 4.04it/s, grad_norm=0.354, loss_final=0.51, loss_mean=0.819, loss_mean_cls=0.0845, proj_loss=-0.394][2026-03-26 12:44:52] Step: 3326, Training Logs: loss_final: 0.520919, loss_mean: 0.852176, proj_loss: -0.410906, loss_mean_cls: 0.079648, grad_norm: 0.304587 +Steps: 0%| | 3327/1000000 [13:49<69:01:26, 4.01it/s, grad_norm=0.305, loss_final=0.521, loss_mean=0.852, loss_mean_cls=0.0796, proj_loss=-0.411][2026-03-26 12:44:53] Step: 3327, Training Logs: loss_final: 0.529004, loss_mean: 0.844558, proj_loss: -0.396830, loss_mean_cls: 0.081276, grad_norm: 0.450427 +Steps: 0%| | 3328/1000000 [13:49<68:33:38, 4.04it/s, grad_norm=0.45, loss_final=0.529, loss_mean=0.845, loss_mean_cls=0.0813, proj_loss=-0.397][2026-03-26 12:44:53] Step: 3328, Training Logs: loss_final: 0.499385, loss_mean: 0.816656, proj_loss: -0.400597, loss_mean_cls: 0.083326, grad_norm: 0.439049 +Steps: 0%| | 3329/1000000 [13:50<68:30:21, 4.04it/s, grad_norm=0.439, loss_final=0.499, loss_mean=0.817, loss_mean_cls=0.0833, proj_loss=-0.401][2026-03-26 12:44:53] Step: 3329, Training Logs: loss_final: 0.521786, loss_mean: 0.839988, proj_loss: -0.400952, loss_mean_cls: 0.082750, grad_norm: 0.445870 +Steps: 0%| | 3330/1000000 [13:50<68:31:20, 4.04it/s, grad_norm=0.446, loss_final=0.522, loss_mean=0.84, loss_mean_cls=0.0828, proj_loss=-0.401][2026-03-26 12:44:53] Step: 3330, Training Logs: loss_final: 0.518764, loss_mean: 0.850547, proj_loss: -0.411910, loss_mean_cls: 0.080128, grad_norm: 0.432680 +Steps: 0%| | 3331/1000000 [13:50<68:29:54, 4.04it/s, grad_norm=0.433, loss_final=0.519, loss_mean=0.851, loss_mean_cls=0.0801, proj_loss=-0.412][2026-03-26 12:44:54] Step: 3331, Training Logs: loss_final: 0.523983, loss_mean: 0.842893, proj_loss: -0.399588, loss_mean_cls: 0.080678, grad_norm: 0.471910 +Steps: 0%| | 3332/1000000 [13:50<68:36:48, 4.03it/s, grad_norm=0.472, loss_final=0.524, loss_mean=0.843, loss_mean_cls=0.0807, proj_loss=-0.4][2026-03-26 12:44:54] Step: 3332, Training Logs: loss_final: 0.498417, loss_mean: 0.821420, proj_loss: -0.406248, loss_mean_cls: 0.083245, grad_norm: 0.831026 +Steps: 0%| | 3333/1000000 [13:51<68:35:08, 4.04it/s, grad_norm=0.831, loss_final=0.498, loss_mean=0.821, loss_mean_cls=0.0832, proj_loss=-0.406][2026-03-26 12:44:54] Step: 3333, Training Logs: loss_final: 0.517284, loss_mean: 0.849717, proj_loss: -0.411038, loss_mean_cls: 0.078605, grad_norm: 0.282912 +Steps: 0%| | 3334/1000000 [13:51<68:37:05, 4.03it/s, grad_norm=0.283, loss_final=0.517, loss_mean=0.85, loss_mean_cls=0.0786, proj_loss=-0.411][2026-03-26 12:44:54] Step: 3334, Training Logs: loss_final: 0.501421, loss_mean: 0.832050, proj_loss: -0.411363, loss_mean_cls: 0.080734, grad_norm: 0.454570 +Steps: 0%| | 3335/1000000 [13:51<68:35:45, 4.04it/s, grad_norm=0.455, loss_final=0.501, loss_mean=0.832, loss_mean_cls=0.0807, proj_loss=-0.411][2026-03-26 12:44:54] Step: 3335, Training Logs: loss_final: 0.511037, loss_mean: 0.831702, proj_loss: -0.403044, loss_mean_cls: 0.082379, grad_norm: 0.451250 +Steps: 0%| | 3336/1000000 [13:51<68:34:00, 4.04it/s, grad_norm=0.451, loss_final=0.511, loss_mean=0.832, loss_mean_cls=0.0824, proj_loss=-0.403][2026-03-26 12:44:55] Step: 3336, Training Logs: loss_final: 0.506876, loss_mean: 0.832008, proj_loss: -0.406063, loss_mean_cls: 0.080931, grad_norm: 0.300116 +Steps: 0%| | 3337/1000000 [13:52<68:32:13, 4.04it/s, grad_norm=0.3, loss_final=0.507, loss_mean=0.832, loss_mean_cls=0.0809, proj_loss=-0.406][2026-03-26 12:44:55] Step: 3337, Training Logs: loss_final: 0.527961, loss_mean: 0.847976, proj_loss: -0.400771, loss_mean_cls: 0.080756, grad_norm: 0.446101 +Steps: 0%| | 3338/1000000 [13:52<68:30:10, 4.04it/s, grad_norm=0.446, loss_final=0.528, loss_mean=0.848, loss_mean_cls=0.0808, proj_loss=-0.401][2026-03-26 12:44:55] Step: 3338, Training Logs: loss_final: 0.532759, loss_mean: 0.856185, proj_loss: -0.404111, loss_mean_cls: 0.080685, grad_norm: 0.487107 +Steps: 0%| | 3339/1000000 [13:52<68:28:42, 4.04it/s, grad_norm=0.487, loss_final=0.533, loss_mean=0.856, loss_mean_cls=0.0807, proj_loss=-0.404][2026-03-26 12:44:55] Step: 3339, Training Logs: loss_final: 0.532380, loss_mean: 0.856801, proj_loss: -0.405463, loss_mean_cls: 0.081041, grad_norm: 0.357199 +Steps: 0%| | 3340/1000000 [13:52<68:32:55, 4.04it/s, grad_norm=0.357, loss_final=0.532, loss_mean=0.857, loss_mean_cls=0.081, proj_loss=-0.405][2026-03-26 12:44:56] Step: 3340, Training Logs: loss_final: 0.506926, loss_mean: 0.835142, proj_loss: -0.408788, loss_mean_cls: 0.080572, grad_norm: 0.396622 +Steps: 0%| | 3341/1000000 [13:53<68:31:00, 4.04it/s, grad_norm=0.397, loss_final=0.507, loss_mean=0.835, loss_mean_cls=0.0806, proj_loss=-0.409][2026-03-26 12:44:56] Step: 3341, Training Logs: loss_final: 0.519435, loss_mean: 0.851613, proj_loss: -0.411925, loss_mean_cls: 0.079747, grad_norm: 0.390553 +Steps: 0%| | 3342/1000000 [13:53<68:30:54, 4.04it/s, grad_norm=0.391, loss_final=0.519, loss_mean=0.852, loss_mean_cls=0.0797, proj_loss=-0.412][2026-03-26 12:44:56] Step: 3342, Training Logs: loss_final: 0.515823, loss_mean: 0.841702, proj_loss: -0.407313, loss_mean_cls: 0.081434, grad_norm: 0.445340 +Steps: 0%| | 3343/1000000 [13:53<68:29:04, 4.04it/s, grad_norm=0.445, loss_final=0.516, loss_mean=0.842, loss_mean_cls=0.0814, proj_loss=-0.407][2026-03-26 12:44:56] Step: 3343, Training Logs: loss_final: 0.513771, loss_mean: 0.838883, proj_loss: -0.407396, loss_mean_cls: 0.082283, grad_norm: 0.407903 +Steps: 0%| | 3344/1000000 [13:53<68:29:07, 4.04it/s, grad_norm=0.408, loss_final=0.514, loss_mean=0.839, loss_mean_cls=0.0823, proj_loss=-0.407][2026-03-26 12:44:57] Step: 3344, Training Logs: loss_final: 0.499891, loss_mean: 0.811841, proj_loss: -0.398069, loss_mean_cls: 0.086119, grad_norm: 0.413501 +Steps: 0%| | 3345/1000000 [13:54<68:30:53, 4.04it/s, grad_norm=0.414, loss_final=0.5, loss_mean=0.812, loss_mean_cls=0.0861, proj_loss=-0.398][2026-03-26 12:44:57] Step: 3345, Training Logs: loss_final: 0.510289, loss_mean: 0.838195, proj_loss: -0.407283, loss_mean_cls: 0.079377, grad_norm: 0.357647 +Steps: 0%| | 3346/1000000 [13:54<68:54:17, 4.02it/s, grad_norm=0.358, loss_final=0.51, loss_mean=0.838, loss_mean_cls=0.0794, proj_loss=-0.407][2026-03-26 12:44:57] Step: 3346, Training Logs: loss_final: 0.496142, loss_mean: 0.823951, proj_loss: -0.409714, loss_mean_cls: 0.081905, grad_norm: 0.413293 +Steps: 0%| | 3347/1000000 [13:54<68:45:44, 4.03it/s, grad_norm=0.413, loss_final=0.496, loss_mean=0.824, loss_mean_cls=0.0819, proj_loss=-0.41][2026-03-26 12:44:57] Step: 3347, Training Logs: loss_final: 0.522803, loss_mean: 0.848688, proj_loss: -0.405958, loss_mean_cls: 0.080073, grad_norm: 0.390361 +Steps: 0%| | 3348/1000000 [13:54<68:42:38, 4.03it/s, grad_norm=0.39, loss_final=0.523, loss_mean=0.849, loss_mean_cls=0.0801, proj_loss=-0.406][2026-03-26 12:44:58] Step: 3348, Training Logs: loss_final: 0.515619, loss_mean: 0.849075, proj_loss: -0.412343, loss_mean_cls: 0.078886, grad_norm: 0.446594 +Steps: 0%| | 3349/1000000 [13:55<68:42:32, 4.03it/s, grad_norm=0.447, loss_final=0.516, loss_mean=0.849, loss_mean_cls=0.0789, proj_loss=-0.412][2026-03-26 12:44:58] Step: 3349, Training Logs: loss_final: 0.507350, loss_mean: 0.832070, proj_loss: -0.405646, loss_mean_cls: 0.080926, grad_norm: 0.432633 +Steps: 0%| | 3350/1000000 [13:55<72:35:23, 3.81it/s, grad_norm=0.433, loss_final=0.507, loss_mean=0.832, loss_mean_cls=0.0809, proj_loss=-0.406][2026-03-26 12:44:58] Step: 3350, Training Logs: loss_final: 0.524399, loss_mean: 0.847026, proj_loss: -0.404662, loss_mean_cls: 0.082035, grad_norm: 0.466478 +Steps: 0%| | 3351/1000000 [13:55<71:21:48, 3.88it/s, grad_norm=0.466, loss_final=0.524, loss_mean=0.847, loss_mean_cls=0.082, proj_loss=-0.405][2026-03-26 12:44:59] Step: 3351, Training Logs: loss_final: 0.517573, loss_mean: 0.837977, proj_loss: -0.400449, loss_mean_cls: 0.080045, grad_norm: 0.495316 +Steps: 0%| | 3352/1000000 [13:55<70:29:04, 3.93it/s, grad_norm=0.495, loss_final=0.518, loss_mean=0.838, loss_mean_cls=0.08, proj_loss=-0.4][2026-03-26 12:44:59] Step: 3352, Training Logs: loss_final: 0.509449, loss_mean: 0.833557, proj_loss: -0.404433, loss_mean_cls: 0.080324, grad_norm: 0.307412 +Steps: 0%| | 3353/1000000 [13:56<69:53:34, 3.96it/s, grad_norm=0.307, loss_final=0.509, loss_mean=0.834, loss_mean_cls=0.0803, proj_loss=-0.404][2026-03-26 12:44:59] Step: 3353, Training Logs: loss_final: 0.527292, loss_mean: 0.847262, proj_loss: -0.402089, loss_mean_cls: 0.082119, grad_norm: 0.436636 +Steps: 0%| | 3354/1000000 [13:56<69:29:54, 3.98it/s, grad_norm=0.437, loss_final=0.527, loss_mean=0.847, loss_mean_cls=0.0821, proj_loss=-0.402][2026-03-26 12:44:59] Step: 3354, Training Logs: loss_final: 0.505959, loss_mean: 0.843837, proj_loss: -0.417537, loss_mean_cls: 0.079659, grad_norm: 0.302836 +Steps: 0%| | 3355/1000000 [13:56<69:11:11, 4.00it/s, grad_norm=0.303, loss_final=0.506, loss_mean=0.844, loss_mean_cls=0.0797, proj_loss=-0.418][2026-03-26 12:44:59] Step: 3355, Training Logs: loss_final: 0.519879, loss_mean: 0.838721, proj_loss: -0.400507, loss_mean_cls: 0.081665, grad_norm: 0.446872 +Steps: 0%| | 3356/1000000 [13:56<68:59:02, 4.01it/s, grad_norm=0.447, loss_final=0.52, loss_mean=0.839, loss_mean_cls=0.0817, proj_loss=-0.401][2026-03-26 12:45:00] Step: 3356, Training Logs: loss_final: 0.501779, loss_mean: 0.821988, proj_loss: -0.403057, loss_mean_cls: 0.082847, grad_norm: 0.500960 +Steps: 0%| | 3357/1000000 [13:57<68:50:45, 4.02it/s, grad_norm=0.501, loss_final=0.502, loss_mean=0.822, loss_mean_cls=0.0828, proj_loss=-0.403][2026-03-26 12:45:00] Step: 3357, Training Logs: loss_final: 0.510069, loss_mean: 0.829342, proj_loss: -0.402461, loss_mean_cls: 0.083189, grad_norm: 0.342529 +Steps: 0%| | 3358/1000000 [13:57<68:43:28, 4.03it/s, grad_norm=0.343, loss_final=0.51, loss_mean=0.829, loss_mean_cls=0.0832, proj_loss=-0.402][2026-03-26 12:45:00] Step: 3358, Training Logs: loss_final: 0.521660, loss_mean: 0.852771, proj_loss: -0.410783, loss_mean_cls: 0.079672, grad_norm: 0.377467 +Steps: 0%| | 3359/1000000 [13:57<68:36:15, 4.04it/s, grad_norm=0.377, loss_final=0.522, loss_mean=0.853, loss_mean_cls=0.0797, proj_loss=-0.411][2026-03-26 12:45:00] Step: 3359, Training Logs: loss_final: 0.516757, loss_mean: 0.848039, proj_loss: -0.411697, loss_mean_cls: 0.080415, grad_norm: 0.345482 +Steps: 0%| | 3360/1000000 [13:57<68:35:09, 4.04it/s, grad_norm=0.345, loss_final=0.517, loss_mean=0.848, loss_mean_cls=0.0804, proj_loss=-0.412][2026-03-26 12:45:01] Step: 3360, Training Logs: loss_final: 0.497904, loss_mean: 0.815706, proj_loss: -0.400969, loss_mean_cls: 0.083167, grad_norm: 0.356050 +Steps: 0%| | 3361/1000000 [13:58<68:31:36, 4.04it/s, grad_norm=0.356, loss_final=0.498, loss_mean=0.816, loss_mean_cls=0.0832, proj_loss=-0.401][2026-03-26 12:45:01] Step: 3361, Training Logs: loss_final: 0.500562, loss_mean: 0.828670, proj_loss: -0.408495, loss_mean_cls: 0.080388, grad_norm: 0.613016 +Steps: 0%| | 3362/1000000 [13:58<68:31:00, 4.04it/s, grad_norm=0.613, loss_final=0.501, loss_mean=0.829, loss_mean_cls=0.0804, proj_loss=-0.408][2026-03-26 12:45:01] Step: 3362, Training Logs: loss_final: 0.523377, loss_mean: 0.840263, proj_loss: -0.400099, loss_mean_cls: 0.083213, grad_norm: 0.565386 +Steps: 0%| | 3363/1000000 [13:58<68:30:02, 4.04it/s, grad_norm=0.565, loss_final=0.523, loss_mean=0.84, loss_mean_cls=0.0832, proj_loss=-0.4][2026-03-26 12:45:01] Step: 3363, Training Logs: loss_final: 0.528064, loss_mean: 0.852748, proj_loss: -0.406375, loss_mean_cls: 0.081691, grad_norm: 0.368072 +Steps: 0%| | 3364/1000000 [13:58<68:31:14, 4.04it/s, grad_norm=0.368, loss_final=0.528, loss_mean=0.853, loss_mean_cls=0.0817, proj_loss=-0.406][2026-03-26 12:45:02] Step: 3364, Training Logs: loss_final: 0.506849, loss_mean: 0.817009, proj_loss: -0.394762, loss_mean_cls: 0.084602, grad_norm: 0.701075 +Steps: 0%| | 3365/1000000 [13:59<68:30:18, 4.04it/s, grad_norm=0.701, loss_final=0.507, loss_mean=0.817, loss_mean_cls=0.0846, proj_loss=-0.395][2026-03-26 12:45:02] Step: 3365, Training Logs: loss_final: 0.533502, loss_mean: 0.850234, proj_loss: -0.398331, loss_mean_cls: 0.081599, grad_norm: 0.436447 +Steps: 0%| | 3366/1000000 [13:59<68:31:27, 4.04it/s, grad_norm=0.436, loss_final=0.534, loss_mean=0.85, loss_mean_cls=0.0816, proj_loss=-0.398][2026-03-26 12:45:02] Step: 3366, Training Logs: loss_final: 0.521011, loss_mean: 0.833091, proj_loss: -0.394812, loss_mean_cls: 0.082732, grad_norm: 0.788589 +Steps: 0%| | 3367/1000000 [13:59<68:29:37, 4.04it/s, grad_norm=0.789, loss_final=0.521, loss_mean=0.833, loss_mean_cls=0.0827, proj_loss=-0.395][2026-03-26 12:45:02] Step: 3367, Training Logs: loss_final: 0.526126, loss_mean: 0.843897, proj_loss: -0.399630, loss_mean_cls: 0.081859, grad_norm: 0.777159 +Steps: 0%| | 3368/1000000 [13:59<68:29:13, 4.04it/s, grad_norm=0.777, loss_final=0.526, loss_mean=0.844, loss_mean_cls=0.0819, proj_loss=-0.4][2026-03-26 12:45:03] Step: 3368, Training Logs: loss_final: 0.526743, loss_mean: 0.853302, proj_loss: -0.406998, loss_mean_cls: 0.080439, grad_norm: 0.463889 +Steps: 0%| | 3369/1000000 [14:00<68:34:09, 4.04it/s, grad_norm=0.464, loss_final=0.527, loss_mean=0.853, loss_mean_cls=0.0804, proj_loss=-0.407][2026-03-26 12:45:03] Step: 3369, Training Logs: loss_final: 0.509361, loss_mean: 0.848691, proj_loss: -0.417514, loss_mean_cls: 0.078185, grad_norm: 1.024512 +Steps: 0%| | 3370/1000000 [14:00<68:31:09, 4.04it/s, grad_norm=1.02, loss_final=0.509, loss_mean=0.849, loss_mean_cls=0.0782, proj_loss=-0.418][2026-03-26 12:45:03] Step: 3370, Training Logs: loss_final: 0.504128, loss_mean: 0.825176, proj_loss: -0.403182, loss_mean_cls: 0.082134, grad_norm: 0.677827 +Steps: 0%| | 3371/1000000 [14:00<68:30:24, 4.04it/s, grad_norm=0.678, loss_final=0.504, loss_mean=0.825, loss_mean_cls=0.0821, proj_loss=-0.403][2026-03-26 12:45:03] Step: 3371, Training Logs: loss_final: 0.530004, loss_mean: 0.863158, proj_loss: -0.411922, loss_mean_cls: 0.078767, grad_norm: 0.536113 +Steps: 0%| | 3372/1000000 [14:00<68:31:14, 4.04it/s, grad_norm=0.536, loss_final=0.53, loss_mean=0.863, loss_mean_cls=0.0788, proj_loss=-0.412][2026-03-26 12:45:04] Step: 3372, Training Logs: loss_final: 0.520910, loss_mean: 0.832057, proj_loss: -0.393889, loss_mean_cls: 0.082743, grad_norm: 0.774985 +Steps: 0%| | 3373/1000000 [14:01<68:37:08, 4.03it/s, grad_norm=0.775, loss_final=0.521, loss_mean=0.832, loss_mean_cls=0.0827, proj_loss=-0.394][2026-03-26 12:45:04] Step: 3373, Training Logs: loss_final: 0.521771, loss_mean: 0.840398, proj_loss: -0.401068, loss_mean_cls: 0.082441, grad_norm: 0.364370 +Steps: 0%| | 3374/1000000 [14:01<68:35:53, 4.04it/s, grad_norm=0.364, loss_final=0.522, loss_mean=0.84, loss_mean_cls=0.0824, proj_loss=-0.401][2026-03-26 12:45:04] Step: 3374, Training Logs: loss_final: 0.529759, loss_mean: 0.855569, proj_loss: -0.405958, loss_mean_cls: 0.080148, grad_norm: 0.733160 +Steps: 0%| | 3375/1000000 [14:01<68:38:06, 4.03it/s, grad_norm=0.733, loss_final=0.53, loss_mean=0.856, loss_mean_cls=0.0801, proj_loss=-0.406][2026-03-26 12:45:04] Step: 3375, Training Logs: loss_final: 0.517824, loss_mean: 0.840218, proj_loss: -0.404070, loss_mean_cls: 0.081676, grad_norm: 0.484950 +Steps: 0%| | 3376/1000000 [14:01<68:33:45, 4.04it/s, grad_norm=0.485, loss_final=0.518, loss_mean=0.84, loss_mean_cls=0.0817, proj_loss=-0.404][2026-03-26 12:45:05] Step: 3376, Training Logs: loss_final: 0.503312, loss_mean: 0.823809, proj_loss: -0.403938, loss_mean_cls: 0.083441, grad_norm: 0.545720 +Steps: 0%| | 3377/1000000 [14:02<68:33:19, 4.04it/s, grad_norm=0.546, loss_final=0.503, loss_mean=0.824, loss_mean_cls=0.0834, proj_loss=-0.404][2026-03-26 12:45:05] Step: 3377, Training Logs: loss_final: 0.514319, loss_mean: 0.840992, proj_loss: -0.405878, loss_mean_cls: 0.079205, grad_norm: 0.548753 +Steps: 0%| | 3378/1000000 [14:02<68:34:25, 4.04it/s, grad_norm=0.549, loss_final=0.514, loss_mean=0.841, loss_mean_cls=0.0792, proj_loss=-0.406][2026-03-26 12:45:05] Step: 3378, Training Logs: loss_final: 0.516082, loss_mean: 0.838318, proj_loss: -0.403967, loss_mean_cls: 0.081731, grad_norm: 0.335509 +Steps: 0%| | 3379/1000000 [14:02<68:32:34, 4.04it/s, grad_norm=0.336, loss_final=0.516, loss_mean=0.838, loss_mean_cls=0.0817, proj_loss=-0.404][2026-03-26 12:45:05] Step: 3379, Training Logs: loss_final: 0.531123, loss_mean: 0.856406, proj_loss: -0.403710, loss_mean_cls: 0.078427, grad_norm: 0.706271 +Steps: 0%| | 3380/1000000 [14:02<68:29:42, 4.04it/s, grad_norm=0.706, loss_final=0.531, loss_mean=0.856, loss_mean_cls=0.0784, proj_loss=-0.404][2026-03-26 12:45:06] Step: 3380, Training Logs: loss_final: 0.534888, loss_mean: 0.862014, proj_loss: -0.406564, loss_mean_cls: 0.079438, grad_norm: 0.625545 +Steps: 0%| | 3381/1000000 [14:03<68:29:32, 4.04it/s, grad_norm=0.626, loss_final=0.535, loss_mean=0.862, loss_mean_cls=0.0794, proj_loss=-0.407][2026-03-26 12:45:06] Step: 3381, Training Logs: loss_final: 0.502944, loss_mean: 0.820259, proj_loss: -0.399984, loss_mean_cls: 0.082669, grad_norm: 0.324095 +Steps: 0%| | 3382/1000000 [14:03<68:31:44, 4.04it/s, grad_norm=0.324, loss_final=0.503, loss_mean=0.82, loss_mean_cls=0.0827, proj_loss=-0.4][2026-03-26 12:45:06] Step: 3382, Training Logs: loss_final: 0.517741, loss_mean: 0.840114, proj_loss: -0.404158, loss_mean_cls: 0.081785, grad_norm: 0.449597 +Steps: 0%| | 3383/1000000 [14:03<68:33:03, 4.04it/s, grad_norm=0.45, loss_final=0.518, loss_mean=0.84, loss_mean_cls=0.0818, proj_loss=-0.404][2026-03-26 12:45:06] Step: 3383, Training Logs: loss_final: 0.512570, loss_mean: 0.832219, proj_loss: -0.401833, loss_mean_cls: 0.082184, grad_norm: 0.475709 +Steps: 0%| | 3384/1000000 [14:03<68:30:03, 4.04it/s, grad_norm=0.476, loss_final=0.513, loss_mean=0.832, loss_mean_cls=0.0822, proj_loss=-0.402][2026-03-26 12:45:07] Step: 3384, Training Logs: loss_final: 0.515801, loss_mean: 0.844873, proj_loss: -0.409812, loss_mean_cls: 0.080741, grad_norm: 0.370308 +Steps: 0%| | 3385/1000000 [14:04<68:30:06, 4.04it/s, grad_norm=0.37, loss_final=0.516, loss_mean=0.845, loss_mean_cls=0.0807, proj_loss=-0.41][2026-03-26 12:45:07] Step: 3385, Training Logs: loss_final: 0.511829, loss_mean: 0.841965, proj_loss: -0.411153, loss_mean_cls: 0.081017, grad_norm: 0.344647 +Steps: 0%| | 3386/1000000 [14:04<68:32:21, 4.04it/s, grad_norm=0.345, loss_final=0.512, loss_mean=0.842, loss_mean_cls=0.081, proj_loss=-0.411][2026-03-26 12:45:07] Step: 3386, Training Logs: loss_final: 0.510890, loss_mean: 0.835544, proj_loss: -0.405331, loss_mean_cls: 0.080677, grad_norm: 0.470766 +Steps: 0%| | 3387/1000000 [14:04<68:29:47, 4.04it/s, grad_norm=0.471, loss_final=0.511, loss_mean=0.836, loss_mean_cls=0.0807, proj_loss=-0.405][2026-03-26 12:45:07] Step: 3387, Training Logs: loss_final: 0.505496, loss_mean: 0.826612, proj_loss: -0.403051, loss_mean_cls: 0.081935, grad_norm: 0.535074 +Steps: 0%| | 3388/1000000 [14:04<68:32:59, 4.04it/s, grad_norm=0.535, loss_final=0.505, loss_mean=0.827, loss_mean_cls=0.0819, proj_loss=-0.403][2026-03-26 12:45:08] Step: 3388, Training Logs: loss_final: 0.502662, loss_mean: 0.821001, proj_loss: -0.402197, loss_mean_cls: 0.083858, grad_norm: 0.375319 +Steps: 0%| | 3389/1000000 [14:05<68:29:58, 4.04it/s, grad_norm=0.375, loss_final=0.503, loss_mean=0.821, loss_mean_cls=0.0839, proj_loss=-0.402][2026-03-26 12:45:08] Step: 3389, Training Logs: loss_final: 0.525110, loss_mean: 0.845536, proj_loss: -0.403088, loss_mean_cls: 0.082662, grad_norm: 0.534801 +Steps: 0%| | 3390/1000000 [14:05<68:35:38, 4.04it/s, grad_norm=0.535, loss_final=0.525, loss_mean=0.846, loss_mean_cls=0.0827, proj_loss=-0.403][2026-03-26 12:45:08] Step: 3390, Training Logs: loss_final: 0.516626, loss_mean: 0.826199, proj_loss: -0.395095, loss_mean_cls: 0.085522, grad_norm: 0.329321 +Steps: 0%| | 3391/1000000 [14:05<68:31:18, 4.04it/s, grad_norm=0.329, loss_final=0.517, loss_mean=0.826, loss_mean_cls=0.0855, proj_loss=-0.395][2026-03-26 12:45:08] Step: 3391, Training Logs: loss_final: 0.514296, loss_mean: 0.820269, proj_loss: -0.392369, loss_mean_cls: 0.086396, grad_norm: 0.600659 +Steps: 0%| | 3392/1000000 [14:05<68:29:44, 4.04it/s, grad_norm=0.601, loss_final=0.514, loss_mean=0.82, loss_mean_cls=0.0864, proj_loss=-0.392][2026-03-26 12:45:09] Step: 3392, Training Logs: loss_final: 0.506986, loss_mean: 0.821805, proj_loss: -0.398142, loss_mean_cls: 0.083324, grad_norm: 0.504806 +Steps: 0%| | 3393/1000000 [14:06<68:28:16, 4.04it/s, grad_norm=0.505, loss_final=0.507, loss_mean=0.822, loss_mean_cls=0.0833, proj_loss=-0.398][2026-03-26 12:45:09] Step: 3393, Training Logs: loss_final: 0.519328, loss_mean: 0.831814, proj_loss: -0.395189, loss_mean_cls: 0.082703, grad_norm: 0.415441 +Steps: 0%| | 3394/1000000 [14:06<68:31:02, 4.04it/s, grad_norm=0.415, loss_final=0.519, loss_mean=0.832, loss_mean_cls=0.0827, proj_loss=-0.395][2026-03-26 12:45:09] Step: 3394, Training Logs: loss_final: 0.510806, loss_mean: 0.850451, proj_loss: -0.417287, loss_mean_cls: 0.077643, grad_norm: 0.521049 +Steps: 0%| | 3395/1000000 [14:06<68:32:14, 4.04it/s, grad_norm=0.521, loss_final=0.511, loss_mean=0.85, loss_mean_cls=0.0776, proj_loss=-0.417][2026-03-26 12:45:09] Step: 3395, Training Logs: loss_final: 0.520321, loss_mean: 0.855634, proj_loss: -0.414586, loss_mean_cls: 0.079273, grad_norm: 0.546292 +Steps: 0%| | 3396/1000000 [14:06<68:34:50, 4.04it/s, grad_norm=0.546, loss_final=0.52, loss_mean=0.856, loss_mean_cls=0.0793, proj_loss=-0.415][2026-03-26 12:45:10] Step: 3396, Training Logs: loss_final: 0.514204, loss_mean: 0.844426, proj_loss: -0.409432, loss_mean_cls: 0.079209, grad_norm: 0.367019 +Steps: 0%| | 3397/1000000 [14:07<68:33:48, 4.04it/s, grad_norm=0.367, loss_final=0.514, loss_mean=0.844, loss_mean_cls=0.0792, proj_loss=-0.409][2026-03-26 12:45:10] Step: 3397, Training Logs: loss_final: 0.501426, loss_mean: 0.834078, proj_loss: -0.412719, loss_mean_cls: 0.080067, grad_norm: 0.449966 +Steps: 0%| | 3398/1000000 [14:07<68:35:11, 4.04it/s, grad_norm=0.45, loss_final=0.501, loss_mean=0.834, loss_mean_cls=0.0801, proj_loss=-0.413][2026-03-26 12:45:10] Step: 3398, Training Logs: loss_final: 0.518167, loss_mean: 0.832616, proj_loss: -0.398300, loss_mean_cls: 0.083850, grad_norm: 0.603020 +Steps: 0%| | 3399/1000000 [14:07<68:32:36, 4.04it/s, grad_norm=0.603, loss_final=0.518, loss_mean=0.833, loss_mean_cls=0.0839, proj_loss=-0.398][2026-03-26 12:45:10] Step: 3399, Training Logs: loss_final: 0.520575, loss_mean: 0.846211, proj_loss: -0.406950, loss_mean_cls: 0.081314, grad_norm: 0.311537 +Steps: 0%| | 3400/1000000 [14:07<68:33:49, 4.04it/s, grad_norm=0.312, loss_final=0.521, loss_mean=0.846, loss_mean_cls=0.0813, proj_loss=-0.407][2026-03-26 12:45:11] Step: 3400, Training Logs: loss_final: 0.514475, loss_mean: 0.841714, proj_loss: -0.408272, loss_mean_cls: 0.081033, grad_norm: 0.549413 +Steps: 0%| | 3401/1000000 [14:07<68:33:55, 4.04it/s, grad_norm=0.549, loss_final=0.514, loss_mean=0.842, loss_mean_cls=0.081, proj_loss=-0.408][2026-03-26 12:45:11] Step: 3401, Training Logs: loss_final: 0.520506, loss_mean: 0.846127, proj_loss: -0.405736, loss_mean_cls: 0.080115, grad_norm: 0.377847 +Steps: 0%| | 3402/1000000 [14:08<68:29:20, 4.04it/s, grad_norm=0.378, loss_final=0.521, loss_mean=0.846, loss_mean_cls=0.0801, proj_loss=-0.406][2026-03-26 12:45:11] Step: 3402, Training Logs: loss_final: 0.511610, loss_mean: 0.839902, proj_loss: -0.408457, loss_mean_cls: 0.080164, grad_norm: 0.415721 +Steps: 0%| | 3403/1000000 [14:08<68:29:29, 4.04it/s, grad_norm=0.416, loss_final=0.512, loss_mean=0.84, loss_mean_cls=0.0802, proj_loss=-0.408][2026-03-26 12:45:11] Step: 3403, Training Logs: loss_final: 0.503333, loss_mean: 0.828174, proj_loss: -0.406979, loss_mean_cls: 0.082138, grad_norm: 0.479262 +Steps: 0%| | 3404/1000000 [14:08<68:29:17, 4.04it/s, grad_norm=0.479, loss_final=0.503, loss_mean=0.828, loss_mean_cls=0.0821, proj_loss=-0.407][2026-03-26 12:45:12] Step: 3404, Training Logs: loss_final: 0.510695, loss_mean: 0.842621, proj_loss: -0.412696, loss_mean_cls: 0.080769, grad_norm: 0.295044 +Steps: 0%| | 3405/1000000 [14:08<68:29:24, 4.04it/s, grad_norm=0.295, loss_final=0.511, loss_mean=0.843, loss_mean_cls=0.0808, proj_loss=-0.413][2026-03-26 12:45:12] Step: 3405, Training Logs: loss_final: 0.509757, loss_mean: 0.836112, proj_loss: -0.408611, loss_mean_cls: 0.082257, grad_norm: 0.446440 +Steps: 0%| | 3406/1000000 [14:09<68:30:03, 4.04it/s, grad_norm=0.446, loss_final=0.51, loss_mean=0.836, loss_mean_cls=0.0823, proj_loss=-0.409][2026-03-26 12:45:12] Step: 3406, Training Logs: loss_final: 0.538629, loss_mean: 0.856919, proj_loss: -0.399735, loss_mean_cls: 0.081445, grad_norm: 0.433233 +Steps: 0%| | 3407/1000000 [14:09<68:33:02, 4.04it/s, grad_norm=0.433, loss_final=0.539, loss_mean=0.857, loss_mean_cls=0.0814, proj_loss=-0.4][2026-03-26 12:45:12] Step: 3407, Training Logs: loss_final: 0.515044, loss_mean: 0.836994, proj_loss: -0.405213, loss_mean_cls: 0.083264, grad_norm: 0.663584 +Steps: 0%| | 3408/1000000 [14:09<68:32:02, 4.04it/s, grad_norm=0.664, loss_final=0.515, loss_mean=0.837, loss_mean_cls=0.0833, proj_loss=-0.405][2026-03-26 12:45:13] Step: 3408, Training Logs: loss_final: 0.494841, loss_mean: 0.820739, proj_loss: -0.406868, loss_mean_cls: 0.080970, grad_norm: 0.502295 +Steps: 0%| | 3409/1000000 [14:09<68:30:00, 4.04it/s, grad_norm=0.502, loss_final=0.495, loss_mean=0.821, loss_mean_cls=0.081, proj_loss=-0.407][2026-03-26 12:45:13] Step: 3409, Training Logs: loss_final: 0.515110, loss_mean: 0.836257, proj_loss: -0.403725, loss_mean_cls: 0.082578, grad_norm: 0.463661 +Steps: 0%| | 3410/1000000 [14:10<68:43:27, 4.03it/s, grad_norm=0.464, loss_final=0.515, loss_mean=0.836, loss_mean_cls=0.0826, proj_loss=-0.404][2026-03-26 12:45:13] Step: 3410, Training Logs: loss_final: 0.511616, loss_mean: 0.838002, proj_loss: -0.407362, loss_mean_cls: 0.080975, grad_norm: 0.639594 +Steps: 0%| | 3411/1000000 [14:10<68:53:14, 4.02it/s, grad_norm=0.64, loss_final=0.512, loss_mean=0.838, loss_mean_cls=0.081, proj_loss=-0.407][2026-03-26 12:45:13] Step: 3411, Training Logs: loss_final: 0.521519, loss_mean: 0.838571, proj_loss: -0.400308, loss_mean_cls: 0.083256, grad_norm: 0.511267 +Steps: 0%| | 3412/1000000 [14:10<69:01:19, 4.01it/s, grad_norm=0.511, loss_final=0.522, loss_mean=0.839, loss_mean_cls=0.0833, proj_loss=-0.4][2026-03-26 12:45:14] Step: 3412, Training Logs: loss_final: 0.506719, loss_mean: 0.830760, proj_loss: -0.405859, loss_mean_cls: 0.081818, grad_norm: 0.426659 +Steps: 0%| | 3413/1000000 [14:10<69:03:51, 4.01it/s, grad_norm=0.427, loss_final=0.507, loss_mean=0.831, loss_mean_cls=0.0818, proj_loss=-0.406][2026-03-26 12:45:14] Step: 3413, Training Logs: loss_final: 0.505165, loss_mean: 0.825490, proj_loss: -0.402887, loss_mean_cls: 0.082561, grad_norm: 0.549956 +Steps: 0%| | 3414/1000000 [14:11<69:07:11, 4.01it/s, grad_norm=0.55, loss_final=0.505, loss_mean=0.825, loss_mean_cls=0.0826, proj_loss=-0.403][2026-03-26 12:45:14] Step: 3414, Training Logs: loss_final: 0.499382, loss_mean: 0.823697, proj_loss: -0.405518, loss_mean_cls: 0.081202, grad_norm: 0.351051 +Steps: 0%| | 3415/1000000 [14:11<69:09:18, 4.00it/s, grad_norm=0.351, loss_final=0.499, loss_mean=0.824, loss_mean_cls=0.0812, proj_loss=-0.406][2026-03-26 12:45:14] Step: 3415, Training Logs: loss_final: 0.509790, loss_mean: 0.826885, proj_loss: -0.400750, loss_mean_cls: 0.083655, grad_norm: 0.408813 +Steps: 0%| | 3416/1000000 [14:11<69:11:35, 4.00it/s, grad_norm=0.409, loss_final=0.51, loss_mean=0.827, loss_mean_cls=0.0837, proj_loss=-0.401][2026-03-26 12:45:15] Step: 3416, Training Logs: loss_final: 0.512195, loss_mean: 0.846378, proj_loss: -0.412545, loss_mean_cls: 0.078362, grad_norm: 0.366079 +Steps: 0%| | 3417/1000000 [14:11<69:12:47, 4.00it/s, grad_norm=0.366, loss_final=0.512, loss_mean=0.846, loss_mean_cls=0.0784, proj_loss=-0.413][2026-03-26 12:45:15] Step: 3417, Training Logs: loss_final: 0.514877, loss_mean: 0.837604, proj_loss: -0.404715, loss_mean_cls: 0.081988, grad_norm: 0.339383 +Steps: 0%| | 3418/1000000 [14:12<69:12:55, 4.00it/s, grad_norm=0.339, loss_final=0.515, loss_mean=0.838, loss_mean_cls=0.082, proj_loss=-0.405][2026-03-26 12:45:15] Step: 3418, Training Logs: loss_final: 0.518968, loss_mean: 0.842902, proj_loss: -0.406804, loss_mean_cls: 0.082870, grad_norm: 0.329661 +Steps: 0%| | 3419/1000000 [14:12<69:09:25, 4.00it/s, grad_norm=0.33, loss_final=0.519, loss_mean=0.843, loss_mean_cls=0.0829, proj_loss=-0.407][2026-03-26 12:45:15] Step: 3419, Training Logs: loss_final: 0.511870, loss_mean: 0.837554, proj_loss: -0.406577, loss_mean_cls: 0.080894, grad_norm: 0.423502 +Steps: 0%| | 3420/1000000 [14:12<69:48:45, 3.97it/s, grad_norm=0.424, loss_final=0.512, loss_mean=0.838, loss_mean_cls=0.0809, proj_loss=-0.407][2026-03-26 12:45:16] Step: 3420, Training Logs: loss_final: 0.516258, loss_mean: 0.844912, proj_loss: -0.409304, loss_mean_cls: 0.080650, grad_norm: 0.509457 +Steps: 0%| | 3421/1000000 [14:12<69:46:20, 3.97it/s, grad_norm=0.509, loss_final=0.516, loss_mean=0.845, loss_mean_cls=0.0806, proj_loss=-0.409][2026-03-26 12:45:16] Step: 3421, Training Logs: loss_final: 0.516097, loss_mean: 0.829807, proj_loss: -0.398584, loss_mean_cls: 0.084875, grad_norm: 0.458506 +Steps: 0%| | 3422/1000000 [14:13<69:35:08, 3.98it/s, grad_norm=0.459, loss_final=0.516, loss_mean=0.83, loss_mean_cls=0.0849, proj_loss=-0.399][2026-03-26 12:45:16] Step: 3422, Training Logs: loss_final: 0.499034, loss_mean: 0.823140, proj_loss: -0.405134, loss_mean_cls: 0.081029, grad_norm: 0.348043 +Steps: 0%| | 3423/1000000 [14:13<69:29:29, 3.98it/s, grad_norm=0.348, loss_final=0.499, loss_mean=0.823, loss_mean_cls=0.081, proj_loss=-0.405][2026-03-26 12:45:16] Step: 3423, Training Logs: loss_final: 0.507701, loss_mean: 0.822887, proj_loss: -0.398352, loss_mean_cls: 0.083165, grad_norm: 0.421178 +Steps: 0%| | 3424/1000000 [14:13<69:26:08, 3.99it/s, grad_norm=0.421, loss_final=0.508, loss_mean=0.823, loss_mean_cls=0.0832, proj_loss=-0.398][2026-03-26 12:45:17] Step: 3424, Training Logs: loss_final: 0.506344, loss_mean: 0.830963, proj_loss: -0.405199, loss_mean_cls: 0.080581, grad_norm: 0.419008 +Steps: 0%| | 3425/1000000 [14:13<69:23:44, 3.99it/s, grad_norm=0.419, loss_final=0.506, loss_mean=0.831, loss_mean_cls=0.0806, proj_loss=-0.405][2026-03-26 12:45:17] Step: 3425, Training Logs: loss_final: 0.506947, loss_mean: 0.836323, proj_loss: -0.410170, loss_mean_cls: 0.080794, grad_norm: 0.449939 +Steps: 0%| | 3426/1000000 [14:14<69:20:56, 3.99it/s, grad_norm=0.45, loss_final=0.507, loss_mean=0.836, loss_mean_cls=0.0808, proj_loss=-0.41][2026-03-26 12:45:17] Step: 3426, Training Logs: loss_final: 0.506790, loss_mean: 0.831158, proj_loss: -0.407396, loss_mean_cls: 0.083028, grad_norm: 0.426434 +Steps: 0%| | 3427/1000000 [14:14<69:19:06, 3.99it/s, grad_norm=0.426, loss_final=0.507, loss_mean=0.831, loss_mean_cls=0.083, proj_loss=-0.407][2026-03-26 12:45:17] Step: 3427, Training Logs: loss_final: 0.515581, loss_mean: 0.821488, proj_loss: -0.390257, loss_mean_cls: 0.084350, grad_norm: 0.353701 +Steps: 0%| | 3428/1000000 [14:14<69:16:20, 4.00it/s, grad_norm=0.354, loss_final=0.516, loss_mean=0.821, loss_mean_cls=0.0843, proj_loss=-0.39][2026-03-26 12:45:18] Step: 3428, Training Logs: loss_final: 0.504590, loss_mean: 0.816208, proj_loss: -0.397414, loss_mean_cls: 0.085796, grad_norm: 0.372170 +Steps: 0%| | 3429/1000000 [14:14<69:16:05, 4.00it/s, grad_norm=0.372, loss_final=0.505, loss_mean=0.816, loss_mean_cls=0.0858, proj_loss=-0.397][2026-03-26 12:45:18] Step: 3429, Training Logs: loss_final: 0.516090, loss_mean: 0.837623, proj_loss: -0.403575, loss_mean_cls: 0.082042, grad_norm: 0.537372 +Steps: 0%| | 3430/1000000 [14:15<69:15:18, 4.00it/s, grad_norm=0.537, loss_final=0.516, loss_mean=0.838, loss_mean_cls=0.082, proj_loss=-0.404][2026-03-26 12:45:18] Step: 3430, Training Logs: loss_final: 0.527754, loss_mean: 0.848496, proj_loss: -0.401776, loss_mean_cls: 0.081034, grad_norm: 0.362811 +Steps: 0%| | 3431/1000000 [14:15<69:16:02, 4.00it/s, grad_norm=0.363, loss_final=0.528, loss_mean=0.848, loss_mean_cls=0.081, proj_loss=-0.402][2026-03-26 12:45:18] Step: 3431, Training Logs: loss_final: 0.508664, loss_mean: 0.840134, proj_loss: -0.411790, loss_mean_cls: 0.080320, grad_norm: 0.573600 +Steps: 0%| | 3432/1000000 [14:15<69:15:36, 4.00it/s, grad_norm=0.574, loss_final=0.509, loss_mean=0.84, loss_mean_cls=0.0803, proj_loss=-0.412][2026-03-26 12:45:19] Step: 3432, Training Logs: loss_final: 0.507173, loss_mean: 0.841450, proj_loss: -0.413367, loss_mean_cls: 0.079091, grad_norm: 0.474511 +Steps: 0%| | 3433/1000000 [14:15<69:12:03, 4.00it/s, grad_norm=0.475, loss_final=0.507, loss_mean=0.841, loss_mean_cls=0.0791, proj_loss=-0.413][2026-03-26 12:45:19] Step: 3433, Training Logs: loss_final: 0.518469, loss_mean: 0.852798, proj_loss: -0.414055, loss_mean_cls: 0.079726, grad_norm: 0.296512 +Steps: 0%| | 3434/1000000 [14:16<69:14:32, 4.00it/s, grad_norm=0.297, loss_final=0.518, loss_mean=0.853, loss_mean_cls=0.0797, proj_loss=-0.414][2026-03-26 12:45:19] Step: 3434, Training Logs: loss_final: 0.509311, loss_mean: 0.826351, proj_loss: -0.399515, loss_mean_cls: 0.082475, grad_norm: 0.383875 +Steps: 0%| | 3435/1000000 [14:16<69:12:43, 4.00it/s, grad_norm=0.384, loss_final=0.509, loss_mean=0.826, loss_mean_cls=0.0825, proj_loss=-0.4][2026-03-26 12:45:19] Step: 3435, Training Logs: loss_final: 0.522406, loss_mean: 0.854200, proj_loss: -0.411610, loss_mean_cls: 0.079817, grad_norm: 0.419527 +Steps: 0%| | 3436/1000000 [14:16<69:15:42, 4.00it/s, grad_norm=0.42, loss_final=0.522, loss_mean=0.854, loss_mean_cls=0.0798, proj_loss=-0.412][2026-03-26 12:45:20] Step: 3436, Training Logs: loss_final: 0.508839, loss_mean: 0.828854, proj_loss: -0.403604, loss_mean_cls: 0.083588, grad_norm: 0.501466 +Steps: 0%| | 3437/1000000 [14:16<69:16:34, 4.00it/s, grad_norm=0.501, loss_final=0.509, loss_mean=0.829, loss_mean_cls=0.0836, proj_loss=-0.404][2026-03-26 12:45:20] Step: 3437, Training Logs: loss_final: 0.516569, loss_mean: 0.843593, proj_loss: -0.407663, loss_mean_cls: 0.080639, grad_norm: 0.473415 +Steps: 0%| | 3438/1000000 [14:17<69:04:20, 4.01it/s, grad_norm=0.473, loss_final=0.517, loss_mean=0.844, loss_mean_cls=0.0806, proj_loss=-0.408][2026-03-26 12:45:20] Step: 3438, Training Logs: loss_final: 0.499788, loss_mean: 0.830367, proj_loss: -0.411452, loss_mean_cls: 0.080874, grad_norm: 0.552057 +Steps: 0%| | 3439/1000000 [14:17<68:58:01, 4.01it/s, grad_norm=0.552, loss_final=0.5, loss_mean=0.83, loss_mean_cls=0.0809, proj_loss=-0.411][2026-03-26 12:45:20] Step: 3439, Training Logs: loss_final: 0.503749, loss_mean: 0.823410, proj_loss: -0.404073, loss_mean_cls: 0.084411, grad_norm: 0.317399 +Steps: 0%| | 3440/1000000 [14:17<68:49:12, 4.02it/s, grad_norm=0.317, loss_final=0.504, loss_mean=0.823, loss_mean_cls=0.0844, proj_loss=-0.404][2026-03-26 12:45:21] Step: 3440, Training Logs: loss_final: 0.508020, loss_mean: 0.831532, proj_loss: -0.404976, loss_mean_cls: 0.081464, grad_norm: 0.503627 +Steps: 0%| | 3441/1000000 [14:17<68:41:23, 4.03it/s, grad_norm=0.504, loss_final=0.508, loss_mean=0.832, loss_mean_cls=0.0815, proj_loss=-0.405][2026-03-26 12:45:21] Step: 3441, Training Logs: loss_final: 0.501775, loss_mean: 0.827606, proj_loss: -0.406161, loss_mean_cls: 0.080330, grad_norm: 0.597508 +Steps: 0%| | 3442/1000000 [14:18<68:50:46, 4.02it/s, grad_norm=0.598, loss_final=0.502, loss_mean=0.828, loss_mean_cls=0.0803, proj_loss=-0.406][2026-03-26 12:45:21] Step: 3442, Training Logs: loss_final: 0.523935, loss_mean: 0.833703, proj_loss: -0.393523, loss_mean_cls: 0.083756, grad_norm: 0.304197 +Steps: 0%| | 3443/1000000 [14:18<68:52:39, 4.02it/s, grad_norm=0.304, loss_final=0.524, loss_mean=0.834, loss_mean_cls=0.0838, proj_loss=-0.394][2026-03-26 12:45:21] Step: 3443, Training Logs: loss_final: 0.513810, loss_mean: 0.841056, proj_loss: -0.407291, loss_mean_cls: 0.080046, grad_norm: 0.689599 +Steps: 0%| | 3444/1000000 [14:18<68:56:04, 4.02it/s, grad_norm=0.69, loss_final=0.514, loss_mean=0.841, loss_mean_cls=0.08, proj_loss=-0.407][2026-03-26 12:45:22] Step: 3444, Training Logs: loss_final: 0.526838, loss_mean: 0.850187, proj_loss: -0.403577, loss_mean_cls: 0.080228, grad_norm: 0.463317 +Steps: 0%| | 3445/1000000 [14:18<69:00:05, 4.01it/s, grad_norm=0.463, loss_final=0.527, loss_mean=0.85, loss_mean_cls=0.0802, proj_loss=-0.404][2026-03-26 12:45:22] Step: 3445, Training Logs: loss_final: 0.528463, loss_mean: 0.851706, proj_loss: -0.404774, loss_mean_cls: 0.081531, grad_norm: 0.585900 +Steps: 0%| | 3446/1000000 [14:19<69:04:20, 4.01it/s, grad_norm=0.586, loss_final=0.528, loss_mean=0.852, loss_mean_cls=0.0815, proj_loss=-0.405][2026-03-26 12:45:22] Step: 3446, Training Logs: loss_final: 0.525026, loss_mean: 0.856207, proj_loss: -0.411237, loss_mean_cls: 0.080055, grad_norm: 0.759755 +Steps: 0%| | 3447/1000000 [14:19<69:05:55, 4.01it/s, grad_norm=0.76, loss_final=0.525, loss_mean=0.856, loss_mean_cls=0.0801, proj_loss=-0.411][2026-03-26 12:45:22] Step: 3447, Training Logs: loss_final: 0.496332, loss_mean: 0.813899, proj_loss: -0.402846, loss_mean_cls: 0.085278, grad_norm: 0.311361 +Steps: 0%| | 3448/1000000 [14:19<69:06:36, 4.01it/s, grad_norm=0.311, loss_final=0.496, loss_mean=0.814, loss_mean_cls=0.0853, proj_loss=-0.403][2026-03-26 12:45:23] Step: 3448, Training Logs: loss_final: 0.505723, loss_mean: 0.826213, proj_loss: -0.403173, loss_mean_cls: 0.082683, grad_norm: 0.873031 +Steps: 0%| | 3449/1000000 [14:19<69:07:38, 4.00it/s, grad_norm=0.873, loss_final=0.506, loss_mean=0.826, loss_mean_cls=0.0827, proj_loss=-0.403][2026-03-26 12:45:23] Step: 3449, Training Logs: loss_final: 0.513359, loss_mean: 0.836636, proj_loss: -0.404771, loss_mean_cls: 0.081494, grad_norm: 0.601742 +Steps: 0%| | 3450/1000000 [14:20<69:11:50, 4.00it/s, grad_norm=0.602, loss_final=0.513, loss_mean=0.837, loss_mean_cls=0.0815, proj_loss=-0.405][2026-03-26 12:45:23] Step: 3450, Training Logs: loss_final: 0.525513, loss_mean: 0.838000, proj_loss: -0.396826, loss_mean_cls: 0.084339, grad_norm: 0.529992 +Steps: 0%| | 3451/1000000 [14:20<69:10:38, 4.00it/s, grad_norm=0.53, loss_final=0.526, loss_mean=0.838, loss_mean_cls=0.0843, proj_loss=-0.397][2026-03-26 12:45:23] Step: 3451, Training Logs: loss_final: 0.535860, loss_mean: 0.858583, proj_loss: -0.402702, loss_mean_cls: 0.079979, grad_norm: 0.595864 +Steps: 0%| | 3452/1000000 [14:20<69:11:42, 4.00it/s, grad_norm=0.596, loss_final=0.536, loss_mean=0.859, loss_mean_cls=0.08, proj_loss=-0.403][2026-03-26 12:45:24] Step: 3452, Training Logs: loss_final: 0.513924, loss_mean: 0.837580, proj_loss: -0.405893, loss_mean_cls: 0.082237, grad_norm: 0.311079 +Steps: 0%| | 3453/1000000 [14:20<69:13:17, 4.00it/s, grad_norm=0.311, loss_final=0.514, loss_mean=0.838, loss_mean_cls=0.0822, proj_loss=-0.406][2026-03-26 12:45:24] Step: 3453, Training Logs: loss_final: 0.511982, loss_mean: 0.826809, proj_loss: -0.398657, loss_mean_cls: 0.083830, grad_norm: 0.482195 +Steps: 0%| | 3454/1000000 [14:21<69:18:16, 3.99it/s, grad_norm=0.482, loss_final=0.512, loss_mean=0.827, loss_mean_cls=0.0838, proj_loss=-0.399][2026-03-26 12:45:24] Step: 3454, Training Logs: loss_final: 0.506855, loss_mean: 0.829197, proj_loss: -0.405338, loss_mean_cls: 0.082996, grad_norm: 0.390355 +Steps: 0%| | 3455/1000000 [14:21<69:13:21, 4.00it/s, grad_norm=0.39, loss_final=0.507, loss_mean=0.829, loss_mean_cls=0.083, proj_loss=-0.405][2026-03-26 12:45:24] Step: 3455, Training Logs: loss_final: 0.508693, loss_mean: 0.848628, proj_loss: -0.416056, loss_mean_cls: 0.076121, grad_norm: 0.516790 +Steps: 0%| | 3456/1000000 [14:21<69:14:22, 4.00it/s, grad_norm=0.517, loss_final=0.509, loss_mean=0.849, loss_mean_cls=0.0761, proj_loss=-0.416][2026-03-26 12:45:25] Step: 3456, Training Logs: loss_final: 0.518214, loss_mean: 0.844632, proj_loss: -0.407379, loss_mean_cls: 0.080961, grad_norm: 0.822384 +Steps: 0%| | 3457/1000000 [14:21<69:10:11, 4.00it/s, grad_norm=0.822, loss_final=0.518, loss_mean=0.845, loss_mean_cls=0.081, proj_loss=-0.407][2026-03-26 12:45:25] Step: 3457, Training Logs: loss_final: 0.521563, loss_mean: 0.857256, proj_loss: -0.414379, loss_mean_cls: 0.078686, grad_norm: 0.419698 +Steps: 0%| | 3458/1000000 [14:22<69:12:32, 4.00it/s, grad_norm=0.42, loss_final=0.522, loss_mean=0.857, loss_mean_cls=0.0787, proj_loss=-0.414][2026-03-26 12:45:25] Step: 3458, Training Logs: loss_final: 0.514059, loss_mean: 0.840136, proj_loss: -0.405861, loss_mean_cls: 0.079784, grad_norm: 0.424709 +Steps: 0%| | 3459/1000000 [14:22<69:07:26, 4.00it/s, grad_norm=0.425, loss_final=0.514, loss_mean=0.84, loss_mean_cls=0.0798, proj_loss=-0.406][2026-03-26 12:45:25] Step: 3459, Training Logs: loss_final: 0.510099, loss_mean: 0.831215, proj_loss: -0.402455, loss_mean_cls: 0.081338, grad_norm: 0.424230 +Steps: 0%| | 3460/1000000 [14:22<69:06:35, 4.01it/s, grad_norm=0.424, loss_final=0.51, loss_mean=0.831, loss_mean_cls=0.0813, proj_loss=-0.402][2026-03-26 12:45:26] Step: 3460, Training Logs: loss_final: 0.514915, loss_mean: 0.842731, proj_loss: -0.407173, loss_mean_cls: 0.079357, grad_norm: 0.322787 +Steps: 0%| | 3461/1000000 [14:22<69:06:32, 4.01it/s, grad_norm=0.323, loss_final=0.515, loss_mean=0.843, loss_mean_cls=0.0794, proj_loss=-0.407][2026-03-26 12:45:26] Step: 3461, Training Logs: loss_final: 0.501160, loss_mean: 0.829968, proj_loss: -0.410299, loss_mean_cls: 0.081491, grad_norm: 0.438601 +Steps: 0%| | 3462/1000000 [14:23<69:05:41, 4.01it/s, grad_norm=0.439, loss_final=0.501, loss_mean=0.83, loss_mean_cls=0.0815, proj_loss=-0.41][2026-03-26 12:45:26] Step: 3462, Training Logs: loss_final: 0.512354, loss_mean: 0.842201, proj_loss: -0.410558, loss_mean_cls: 0.080710, grad_norm: 0.405409 +Steps: 0%| | 3463/1000000 [14:23<69:09:02, 4.00it/s, grad_norm=0.405, loss_final=0.512, loss_mean=0.842, loss_mean_cls=0.0807, proj_loss=-0.411][2026-03-26 12:45:26] Step: 3463, Training Logs: loss_final: 0.522236, loss_mean: 0.845031, proj_loss: -0.404105, loss_mean_cls: 0.081310, grad_norm: 0.584293 +Steps: 0%| | 3464/1000000 [14:23<69:09:10, 4.00it/s, grad_norm=0.584, loss_final=0.522, loss_mean=0.845, loss_mean_cls=0.0813, proj_loss=-0.404][2026-03-26 12:45:27] Step: 3464, Training Logs: loss_final: 0.509241, loss_mean: 0.840798, proj_loss: -0.409941, loss_mean_cls: 0.078384, grad_norm: 0.352640 +Steps: 0%| | 3465/1000000 [14:23<69:11:05, 4.00it/s, grad_norm=0.353, loss_final=0.509, loss_mean=0.841, loss_mean_cls=0.0784, proj_loss=-0.41][2026-03-26 12:45:27] Step: 3465, Training Logs: loss_final: 0.515882, loss_mean: 0.838746, proj_loss: -0.404102, loss_mean_cls: 0.081239, grad_norm: 0.619034 +Steps: 0%| | 3466/1000000 [14:24<69:08:41, 4.00it/s, grad_norm=0.619, loss_final=0.516, loss_mean=0.839, loss_mean_cls=0.0812, proj_loss=-0.404][2026-03-26 12:45:27] Step: 3466, Training Logs: loss_final: 0.525766, loss_mean: 0.858063, proj_loss: -0.412402, loss_mean_cls: 0.080105, grad_norm: 0.565194 +Steps: 0%| | 3467/1000000 [14:24<69:47:09, 3.97it/s, grad_norm=0.565, loss_final=0.526, loss_mean=0.858, loss_mean_cls=0.0801, proj_loss=-0.412][2026-03-26 12:45:27] Step: 3467, Training Logs: loss_final: 0.529741, loss_mean: 0.855270, proj_loss: -0.405065, loss_mean_cls: 0.079537, grad_norm: 0.401300 +Steps: 0%| | 3468/1000000 [14:24<69:39:30, 3.97it/s, grad_norm=0.401, loss_final=0.53, loss_mean=0.855, loss_mean_cls=0.0795, proj_loss=-0.405][2026-03-26 12:45:28] Step: 3468, Training Logs: loss_final: 0.507145, loss_mean: 0.826951, proj_loss: -0.401387, loss_mean_cls: 0.081580, grad_norm: 0.672472 +Steps: 0%| | 3469/1000000 [14:24<69:29:25, 3.98it/s, grad_norm=0.672, loss_final=0.507, loss_mean=0.827, loss_mean_cls=0.0816, proj_loss=-0.401][2026-03-26 12:45:28] Step: 3469, Training Logs: loss_final: 0.519055, loss_mean: 0.861027, proj_loss: -0.418009, loss_mean_cls: 0.076038, grad_norm: 0.364101 +Steps: 0%| | 3470/1000000 [14:25<69:26:01, 3.99it/s, grad_norm=0.364, loss_final=0.519, loss_mean=0.861, loss_mean_cls=0.076, proj_loss=-0.418][2026-03-26 12:45:28] Step: 3470, Training Logs: loss_final: 0.526341, loss_mean: 0.858091, proj_loss: -0.410862, loss_mean_cls: 0.079112, grad_norm: 0.418203 +Steps: 0%| | 3471/1000000 [14:25<69:21:55, 3.99it/s, grad_norm=0.418, loss_final=0.526, loss_mean=0.858, loss_mean_cls=0.0791, proj_loss=-0.411][2026-03-26 12:45:28] Step: 3471, Training Logs: loss_final: 0.520873, loss_mean: 0.846436, proj_loss: -0.406721, loss_mean_cls: 0.081157, grad_norm: 0.484486 +Steps: 0%| | 3472/1000000 [14:25<69:20:45, 3.99it/s, grad_norm=0.484, loss_final=0.521, loss_mean=0.846, loss_mean_cls=0.0812, proj_loss=-0.407][2026-03-26 12:45:29] Step: 3472, Training Logs: loss_final: 0.521818, loss_mean: 0.847886, proj_loss: -0.407325, loss_mean_cls: 0.081256, grad_norm: 0.401681 +Steps: 0%| | 3473/1000000 [14:25<69:18:18, 3.99it/s, grad_norm=0.402, loss_final=0.522, loss_mean=0.848, loss_mean_cls=0.0813, proj_loss=-0.407][2026-03-26 12:45:29] Step: 3473, Training Logs: loss_final: 0.511532, loss_mean: 0.842131, proj_loss: -0.410902, loss_mean_cls: 0.080303, grad_norm: 0.285693 +Steps: 0%| | 3474/1000000 [14:26<69:14:36, 4.00it/s, grad_norm=0.286, loss_final=0.512, loss_mean=0.842, loss_mean_cls=0.0803, proj_loss=-0.411][2026-03-26 12:45:29] Step: 3474, Training Logs: loss_final: 0.509869, loss_mean: 0.828395, proj_loss: -0.401368, loss_mean_cls: 0.082842, grad_norm: 0.546379 +Steps: 0%| | 3475/1000000 [14:26<69:12:07, 4.00it/s, grad_norm=0.546, loss_final=0.51, loss_mean=0.828, loss_mean_cls=0.0828, proj_loss=-0.401][2026-03-26 12:45:29] Step: 3475, Training Logs: loss_final: 0.510284, loss_mean: 0.843169, proj_loss: -0.412530, loss_mean_cls: 0.079645, grad_norm: 0.335286 +Steps: 0%| | 3476/1000000 [14:26<69:10:49, 4.00it/s, grad_norm=0.335, loss_final=0.51, loss_mean=0.843, loss_mean_cls=0.0796, proj_loss=-0.413][2026-03-26 12:45:30] Step: 3476, Training Logs: loss_final: 0.485662, loss_mean: 0.804098, proj_loss: -0.400666, loss_mean_cls: 0.082230, grad_norm: 0.448660 +Steps: 0%| | 3477/1000000 [14:26<69:09:18, 4.00it/s, grad_norm=0.449, loss_final=0.486, loss_mean=0.804, loss_mean_cls=0.0822, proj_loss=-0.401][2026-03-26 12:45:30] Step: 3477, Training Logs: loss_final: 0.522631, loss_mean: 0.850995, proj_loss: -0.407992, loss_mean_cls: 0.079629, grad_norm: 0.427818 +Steps: 0%| | 3478/1000000 [14:27<69:11:22, 4.00it/s, grad_norm=0.428, loss_final=0.523, loss_mean=0.851, loss_mean_cls=0.0796, proj_loss=-0.408][2026-03-26 12:45:30] Step: 3478, Training Logs: loss_final: 0.522018, loss_mean: 0.846399, proj_loss: -0.405919, loss_mean_cls: 0.081538, grad_norm: 0.556706 +Steps: 0%| | 3479/1000000 [14:27<69:13:51, 4.00it/s, grad_norm=0.557, loss_final=0.522, loss_mean=0.846, loss_mean_cls=0.0815, proj_loss=-0.406][2026-03-26 12:45:30] Step: 3479, Training Logs: loss_final: 0.503139, loss_mean: 0.833493, proj_loss: -0.410775, loss_mean_cls: 0.080421, grad_norm: 0.430579 +Steps: 0%| | 3480/1000000 [14:27<69:14:35, 4.00it/s, grad_norm=0.431, loss_final=0.503, loss_mean=0.833, loss_mean_cls=0.0804, proj_loss=-0.411][2026-03-26 12:45:31] Step: 3480, Training Logs: loss_final: 0.497473, loss_mean: 0.816887, proj_loss: -0.402700, loss_mean_cls: 0.083286, grad_norm: 0.450021 +Steps: 0%| | 3481/1000000 [14:27<69:15:46, 4.00it/s, grad_norm=0.45, loss_final=0.497, loss_mean=0.817, loss_mean_cls=0.0833, proj_loss=-0.403][2026-03-26 12:45:31] Step: 3481, Training Logs: loss_final: 0.512227, loss_mean: 0.845419, proj_loss: -0.413478, loss_mean_cls: 0.080287, grad_norm: 0.577809 +Steps: 0%| | 3482/1000000 [14:28<69:15:24, 4.00it/s, grad_norm=0.578, loss_final=0.512, loss_mean=0.845, loss_mean_cls=0.0803, proj_loss=-0.413][2026-03-26 12:45:31] Step: 3482, Training Logs: loss_final: 0.511720, loss_mean: 0.833130, proj_loss: -0.403443, loss_mean_cls: 0.082033, grad_norm: 0.297161 +Steps: 0%| | 3483/1000000 [14:28<69:15:34, 4.00it/s, grad_norm=0.297, loss_final=0.512, loss_mean=0.833, loss_mean_cls=0.082, proj_loss=-0.403][2026-03-26 12:45:31] Step: 3483, Training Logs: loss_final: 0.519062, loss_mean: 0.850210, proj_loss: -0.410415, loss_mean_cls: 0.079266, grad_norm: 0.418055 +Steps: 0%| | 3484/1000000 [14:28<69:16:21, 4.00it/s, grad_norm=0.418, loss_final=0.519, loss_mean=0.85, loss_mean_cls=0.0793, proj_loss=-0.41][2026-03-26 12:45:32] Step: 3484, Training Logs: loss_final: 0.523323, loss_mean: 0.847383, proj_loss: -0.405248, loss_mean_cls: 0.081187, grad_norm: 0.474292 +Steps: 0%| | 3485/1000000 [14:28<69:15:34, 4.00it/s, grad_norm=0.474, loss_final=0.523, loss_mean=0.847, loss_mean_cls=0.0812, proj_loss=-0.405][2026-03-26 12:45:32] Step: 3485, Training Logs: loss_final: 0.499770, loss_mean: 0.821375, proj_loss: -0.403881, loss_mean_cls: 0.082276, grad_norm: 0.362442 +Steps: 0%| | 3486/1000000 [14:29<69:11:40, 4.00it/s, grad_norm=0.362, loss_final=0.5, loss_mean=0.821, loss_mean_cls=0.0823, proj_loss=-0.404][2026-03-26 12:45:32] Step: 3486, Training Logs: loss_final: 0.502608, loss_mean: 0.825674, proj_loss: -0.405444, loss_mean_cls: 0.082378, grad_norm: 0.675113 +Steps: 0%| | 3487/1000000 [14:29<69:10:33, 4.00it/s, grad_norm=0.675, loss_final=0.503, loss_mean=0.826, loss_mean_cls=0.0824, proj_loss=-0.405][2026-03-26 12:45:32] Step: 3487, Training Logs: loss_final: 0.521889, loss_mean: 0.841361, proj_loss: -0.401622, loss_mean_cls: 0.082150, grad_norm: 0.424097 +Steps: 0%| | 3488/1000000 [14:29<69:11:56, 4.00it/s, grad_norm=0.424, loss_final=0.522, loss_mean=0.841, loss_mean_cls=0.0821, proj_loss=-0.402][2026-03-26 12:45:33] Step: 3488, Training Logs: loss_final: 0.517367, loss_mean: 0.846149, proj_loss: -0.409679, loss_mean_cls: 0.080897, grad_norm: 0.387770 +Steps: 0%| | 3489/1000000 [14:29<69:13:30, 4.00it/s, grad_norm=0.388, loss_final=0.517, loss_mean=0.846, loss_mean_cls=0.0809, proj_loss=-0.41][2026-03-26 12:45:33] Step: 3489, Training Logs: loss_final: 0.516795, loss_mean: 0.831170, proj_loss: -0.397624, loss_mean_cls: 0.083249, grad_norm: 0.562204 +Steps: 0%| | 3490/1000000 [14:30<69:16:29, 4.00it/s, grad_norm=0.562, loss_final=0.517, loss_mean=0.831, loss_mean_cls=0.0832, proj_loss=-0.398][2026-03-26 12:45:33] Step: 3490, Training Logs: loss_final: 0.517590, loss_mean: 0.851515, proj_loss: -0.412727, loss_mean_cls: 0.078802, grad_norm: 0.390226 +Steps: 0%| | 3491/1000000 [14:30<69:16:38, 4.00it/s, grad_norm=0.39, loss_final=0.518, loss_mean=0.852, loss_mean_cls=0.0788, proj_loss=-0.413][2026-03-26 12:45:33] Step: 3491, Training Logs: loss_final: 0.522384, loss_mean: 0.845090, proj_loss: -0.403777, loss_mean_cls: 0.081071, grad_norm: 0.405964 +Steps: 0%| | 3492/1000000 [14:30<69:15:23, 4.00it/s, grad_norm=0.406, loss_final=0.522, loss_mean=0.845, loss_mean_cls=0.0811, proj_loss=-0.404][2026-03-26 12:45:34] Step: 3492, Training Logs: loss_final: 0.510324, loss_mean: 0.836861, proj_loss: -0.407126, loss_mean_cls: 0.080588, grad_norm: 0.425288 +Steps: 0%| | 3493/1000000 [14:30<69:07:36, 4.00it/s, grad_norm=0.425, loss_final=0.51, loss_mean=0.837, loss_mean_cls=0.0806, proj_loss=-0.407][2026-03-26 12:45:34] Step: 3493, Training Logs: loss_final: 0.513022, loss_mean: 0.841534, proj_loss: -0.409598, loss_mean_cls: 0.081085, grad_norm: 0.399118 +Steps: 0%| | 3494/1000000 [14:31<68:58:31, 4.01it/s, grad_norm=0.399, loss_final=0.513, loss_mean=0.842, loss_mean_cls=0.0811, proj_loss=-0.41][2026-03-26 12:45:34] Step: 3494, Training Logs: loss_final: 0.500958, loss_mean: 0.822096, proj_loss: -0.402306, loss_mean_cls: 0.081168, grad_norm: 0.356866 +Steps: 0%| | 3495/1000000 [14:31<68:50:29, 4.02it/s, grad_norm=0.357, loss_final=0.501, loss_mean=0.822, loss_mean_cls=0.0812, proj_loss=-0.402][2026-03-26 12:45:34] Step: 3495, Training Logs: loss_final: 0.497321, loss_mean: 0.806406, proj_loss: -0.395119, loss_mean_cls: 0.086035, grad_norm: 0.423524 +Steps: 0%| | 3496/1000000 [14:31<68:42:46, 4.03it/s, grad_norm=0.424, loss_final=0.497, loss_mean=0.806, loss_mean_cls=0.086, proj_loss=-0.395][2026-03-26 12:45:35] Step: 3496, Training Logs: loss_final: 0.511903, loss_mean: 0.832717, proj_loss: -0.403391, loss_mean_cls: 0.082577, grad_norm: 0.369925 +Steps: 0%| | 3497/1000000 [14:31<68:49:02, 4.02it/s, grad_norm=0.37, loss_final=0.512, loss_mean=0.833, loss_mean_cls=0.0826, proj_loss=-0.403][2026-03-26 12:45:35] Step: 3497, Training Logs: loss_final: 0.503154, loss_mean: 0.845918, proj_loss: -0.420896, loss_mean_cls: 0.078132, grad_norm: 0.281967 +Steps: 0%| | 3498/1000000 [14:32<68:42:36, 4.03it/s, grad_norm=0.282, loss_final=0.503, loss_mean=0.846, loss_mean_cls=0.0781, proj_loss=-0.421][2026-03-26 12:45:35] Step: 3498, Training Logs: loss_final: 0.498299, loss_mean: 0.838834, proj_loss: -0.420308, loss_mean_cls: 0.079773, grad_norm: 0.319608 +Steps: 0%| | 3499/1000000 [14:32<68:39:17, 4.03it/s, grad_norm=0.32, loss_final=0.498, loss_mean=0.839, loss_mean_cls=0.0798, proj_loss=-0.42][2026-03-26 12:45:35] Step: 3499, Training Logs: loss_final: 0.505042, loss_mean: 0.842287, proj_loss: -0.417056, loss_mean_cls: 0.079810, grad_norm: 0.302559 +Steps: 0%| | 3500/1000000 [14:32<68:35:05, 4.04it/s, grad_norm=0.303, loss_final=0.505, loss_mean=0.842, loss_mean_cls=0.0798, proj_loss=-0.417][2026-03-26 12:45:36] Step: 3500, Training Logs: loss_final: 0.515314, loss_mean: 0.851212, proj_loss: -0.416035, loss_mean_cls: 0.080138, grad_norm: 0.438214 +Steps: 0%| | 3501/1000000 [14:32<68:43:44, 4.03it/s, grad_norm=0.438, loss_final=0.515, loss_mean=0.851, loss_mean_cls=0.0801, proj_loss=-0.416][2026-03-26 12:45:36] Step: 3501, Training Logs: loss_final: 0.522519, loss_mean: 0.854318, proj_loss: -0.411196, loss_mean_cls: 0.079398, grad_norm: 0.352443 +Steps: 0%| | 3502/1000000 [14:33<68:42:01, 4.03it/s, grad_norm=0.352, loss_final=0.523, loss_mean=0.854, loss_mean_cls=0.0794, proj_loss=-0.411][2026-03-26 12:45:36] Step: 3502, Training Logs: loss_final: 0.512695, loss_mean: 0.837250, proj_loss: -0.406458, loss_mean_cls: 0.081903, grad_norm: 0.425721 +Steps: 0%| | 3503/1000000 [14:33<68:47:34, 4.02it/s, grad_norm=0.426, loss_final=0.513, loss_mean=0.837, loss_mean_cls=0.0819, proj_loss=-0.406][2026-03-26 12:45:36] Step: 3503, Training Logs: loss_final: 0.504674, loss_mean: 0.833062, proj_loss: -0.410481, loss_mean_cls: 0.082093, grad_norm: 0.432930 +Steps: 0%| | 3504/1000000 [14:33<68:39:10, 4.03it/s, grad_norm=0.433, loss_final=0.505, loss_mean=0.833, loss_mean_cls=0.0821, proj_loss=-0.41][2026-03-26 12:45:37] Step: 3504, Training Logs: loss_final: 0.512248, loss_mean: 0.834169, proj_loss: -0.405074, loss_mean_cls: 0.083154, grad_norm: 0.562674 +Steps: 0%| | 3505/1000000 [14:33<68:45:19, 4.03it/s, grad_norm=0.563, loss_final=0.512, loss_mean=0.834, loss_mean_cls=0.0832, proj_loss=-0.405][2026-03-26 12:45:37] Step: 3505, Training Logs: loss_final: 0.502406, loss_mean: 0.831343, proj_loss: -0.410611, loss_mean_cls: 0.081674, grad_norm: 0.429889 +Steps: 0%| | 3506/1000000 [14:34<68:40:24, 4.03it/s, grad_norm=0.43, loss_final=0.502, loss_mean=0.831, loss_mean_cls=0.0817, proj_loss=-0.411][2026-03-26 12:45:37] Step: 3506, Training Logs: loss_final: 0.511193, loss_mean: 0.836976, proj_loss: -0.405371, loss_mean_cls: 0.079588, grad_norm: 0.493723 +Steps: 0%| | 3507/1000000 [14:34<68:36:43, 4.03it/s, grad_norm=0.494, loss_final=0.511, loss_mean=0.837, loss_mean_cls=0.0796, proj_loss=-0.405][2026-03-26 12:45:37] Step: 3507, Training Logs: loss_final: 0.497642, loss_mean: 0.818197, proj_loss: -0.402752, loss_mean_cls: 0.082196, grad_norm: 0.345665 +Steps: 0%| | 3508/1000000 [14:34<68:32:07, 4.04it/s, grad_norm=0.346, loss_final=0.498, loss_mean=0.818, loss_mean_cls=0.0822, proj_loss=-0.403][2026-03-26 12:45:38] Step: 3508, Training Logs: loss_final: 0.519328, loss_mean: 0.839463, proj_loss: -0.402985, loss_mean_cls: 0.082850, grad_norm: 0.402779 +Steps: 0%| | 3509/1000000 [14:34<68:41:44, 4.03it/s, grad_norm=0.403, loss_final=0.519, loss_mean=0.839, loss_mean_cls=0.0829, proj_loss=-0.403][2026-03-26 12:45:38] Step: 3509, Training Logs: loss_final: 0.527065, loss_mean: 0.856990, proj_loss: -0.408409, loss_mean_cls: 0.078484, grad_norm: 0.523964 +Steps: 0%| | 3510/1000000 [14:35<68:32:31, 4.04it/s, grad_norm=0.524, loss_final=0.527, loss_mean=0.857, loss_mean_cls=0.0785, proj_loss=-0.408][2026-03-26 12:45:38] Step: 3510, Training Logs: loss_final: 0.514491, loss_mean: 0.846490, proj_loss: -0.411794, loss_mean_cls: 0.079795, grad_norm: 0.404615 +Steps: 0%| | 3511/1000000 [14:35<68:29:19, 4.04it/s, grad_norm=0.405, loss_final=0.514, loss_mean=0.846, loss_mean_cls=0.0798, proj_loss=-0.412][2026-03-26 12:45:38] Step: 3511, Training Logs: loss_final: 0.510399, loss_mean: 0.839952, proj_loss: -0.410330, loss_mean_cls: 0.080777, grad_norm: 0.369303 +Steps: 0%| | 3512/1000000 [14:35<68:27:00, 4.04it/s, grad_norm=0.369, loss_final=0.51, loss_mean=0.84, loss_mean_cls=0.0808, proj_loss=-0.41][2026-03-26 12:45:39] Step: 3512, Training Logs: loss_final: 0.507401, loss_mean: 0.819883, proj_loss: -0.398194, loss_mean_cls: 0.085712, grad_norm: 0.495022 +Steps: 0%| | 3513/1000000 [14:35<68:30:56, 4.04it/s, grad_norm=0.495, loss_final=0.507, loss_mean=0.82, loss_mean_cls=0.0857, proj_loss=-0.398][2026-03-26 12:45:39] Step: 3513, Training Logs: loss_final: 0.513326, loss_mean: 0.838419, proj_loss: -0.406642, loss_mean_cls: 0.081549, grad_norm: 0.554173 +Steps: 0%| | 3514/1000000 [14:36<68:27:27, 4.04it/s, grad_norm=0.554, loss_final=0.513, loss_mean=0.838, loss_mean_cls=0.0815, proj_loss=-0.407][2026-03-26 12:45:39] Step: 3514, Training Logs: loss_final: 0.505135, loss_mean: 0.826417, proj_loss: -0.404787, loss_mean_cls: 0.083505, grad_norm: 0.362615 +Steps: 0%| | 3515/1000000 [14:36<68:27:41, 4.04it/s, grad_norm=0.363, loss_final=0.505, loss_mean=0.826, loss_mean_cls=0.0835, proj_loss=-0.405][2026-03-26 12:45:39] Step: 3515, Training Logs: loss_final: 0.529037, loss_mean: 0.854540, proj_loss: -0.405348, loss_mean_cls: 0.079844, grad_norm: 0.741578 +Steps: 0%| | 3516/1000000 [14:36<68:28:01, 4.04it/s, grad_norm=0.742, loss_final=0.529, loss_mean=0.855, loss_mean_cls=0.0798, proj_loss=-0.405][2026-03-26 12:45:40] Step: 3516, Training Logs: loss_final: 0.513843, loss_mean: 0.844252, proj_loss: -0.411017, loss_mean_cls: 0.080608, grad_norm: 0.633680 +Steps: 0%| | 3517/1000000 [14:36<68:43:01, 4.03it/s, grad_norm=0.634, loss_final=0.514, loss_mean=0.844, loss_mean_cls=0.0806, proj_loss=-0.411][2026-03-26 12:45:40] Step: 3517, Training Logs: loss_final: 0.517433, loss_mean: 0.834127, proj_loss: -0.399747, loss_mean_cls: 0.083053, grad_norm: 0.326827 +Steps: 0%| | 3518/1000000 [14:37<68:41:52, 4.03it/s, grad_norm=0.327, loss_final=0.517, loss_mean=0.834, loss_mean_cls=0.0831, proj_loss=-0.4][2026-03-26 12:45:40] Step: 3518, Training Logs: loss_final: 0.508007, loss_mean: 0.835122, proj_loss: -0.408382, loss_mean_cls: 0.081267, grad_norm: 0.699452 +Steps: 0%| | 3519/1000000 [14:37<68:37:36, 4.03it/s, grad_norm=0.699, loss_final=0.508, loss_mean=0.835, loss_mean_cls=0.0813, proj_loss=-0.408][2026-03-26 12:45:40] Step: 3519, Training Logs: loss_final: 0.523624, loss_mean: 0.841935, proj_loss: -0.401643, loss_mean_cls: 0.083332, grad_norm: 0.497060 +Steps: 0%| | 3520/1000000 [14:37<68:37:17, 4.03it/s, grad_norm=0.497, loss_final=0.524, loss_mean=0.842, loss_mean_cls=0.0833, proj_loss=-0.402][2026-03-26 12:45:41] Step: 3520, Training Logs: loss_final: 0.511349, loss_mean: 0.847206, proj_loss: -0.415173, loss_mean_cls: 0.079316, grad_norm: 0.685680 +Steps: 0%| | 3521/1000000 [14:37<68:41:30, 4.03it/s, grad_norm=0.686, loss_final=0.511, loss_mean=0.847, loss_mean_cls=0.0793, proj_loss=-0.415][2026-03-26 12:45:41] Step: 3521, Training Logs: loss_final: 0.518668, loss_mean: 0.841056, proj_loss: -0.403300, loss_mean_cls: 0.080912, grad_norm: 0.717388 +Steps: 0%| | 3522/1000000 [14:38<68:38:53, 4.03it/s, grad_norm=0.717, loss_final=0.519, loss_mean=0.841, loss_mean_cls=0.0809, proj_loss=-0.403][2026-03-26 12:45:41] Step: 3522, Training Logs: loss_final: 0.520295, loss_mean: 0.845518, proj_loss: -0.405959, loss_mean_cls: 0.080735, grad_norm: 0.338984 +Steps: 0%| | 3523/1000000 [14:38<68:34:48, 4.04it/s, grad_norm=0.339, loss_final=0.52, loss_mean=0.846, loss_mean_cls=0.0807, proj_loss=-0.406][2026-03-26 12:45:41] Step: 3523, Training Logs: loss_final: 0.511772, loss_mean: 0.828583, proj_loss: -0.399121, loss_mean_cls: 0.082310, grad_norm: 0.605968 +Steps: 0%| | 3524/1000000 [14:38<68:32:43, 4.04it/s, grad_norm=0.606, loss_final=0.512, loss_mean=0.829, loss_mean_cls=0.0823, proj_loss=-0.399][2026-03-26 12:45:42] Step: 3524, Training Logs: loss_final: 0.496639, loss_mean: 0.818520, proj_loss: -0.403999, loss_mean_cls: 0.082117, grad_norm: 0.402967 +Steps: 0%| | 3525/1000000 [14:38<68:37:08, 4.03it/s, grad_norm=0.403, loss_final=0.497, loss_mean=0.819, loss_mean_cls=0.0821, proj_loss=-0.404][2026-03-26 12:45:42] Step: 3525, Training Logs: loss_final: 0.513624, loss_mean: 0.835251, proj_loss: -0.403668, loss_mean_cls: 0.082041, grad_norm: 0.847632 +Steps: 0%| | 3526/1000000 [14:39<68:37:04, 4.03it/s, grad_norm=0.848, loss_final=0.514, loss_mean=0.835, loss_mean_cls=0.082, proj_loss=-0.404][2026-03-26 12:45:42] Step: 3526, Training Logs: loss_final: 0.506615, loss_mean: 0.832492, proj_loss: -0.406745, loss_mean_cls: 0.080867, grad_norm: 0.741657 +Steps: 0%| | 3527/1000000 [14:39<68:34:18, 4.04it/s, grad_norm=0.742, loss_final=0.507, loss_mean=0.832, loss_mean_cls=0.0809, proj_loss=-0.407][2026-03-26 12:45:42] Step: 3527, Training Logs: loss_final: 0.511228, loss_mean: 0.835890, proj_loss: -0.405308, loss_mean_cls: 0.080645, grad_norm: 0.417021 +Steps: 0%| | 3528/1000000 [14:39<68:34:24, 4.04it/s, grad_norm=0.417, loss_final=0.511, loss_mean=0.836, loss_mean_cls=0.0806, proj_loss=-0.405][2026-03-26 12:45:43] Step: 3528, Training Logs: loss_final: 0.509930, loss_mean: 0.840742, proj_loss: -0.411532, loss_mean_cls: 0.080721, grad_norm: 0.526187 +Steps: 0%| | 3529/1000000 [14:39<68:35:08, 4.04it/s, grad_norm=0.526, loss_final=0.51, loss_mean=0.841, loss_mean_cls=0.0807, proj_loss=-0.412][2026-03-26 12:45:43] Step: 3529, Training Logs: loss_final: 0.529612, loss_mean: 0.855972, proj_loss: -0.406687, loss_mean_cls: 0.080326, grad_norm: 0.389028 +Steps: 0%| | 3530/1000000 [14:40<68:33:16, 4.04it/s, grad_norm=0.389, loss_final=0.53, loss_mean=0.856, loss_mean_cls=0.0803, proj_loss=-0.407][2026-03-26 12:45:43] Step: 3530, Training Logs: loss_final: 0.491312, loss_mean: 0.819721, proj_loss: -0.410346, loss_mean_cls: 0.081937, grad_norm: 0.350330 +Steps: 0%| | 3531/1000000 [14:40<68:29:02, 4.04it/s, grad_norm=0.35, loss_final=0.491, loss_mean=0.82, loss_mean_cls=0.0819, proj_loss=-0.41][2026-03-26 12:45:43] Step: 3531, Training Logs: loss_final: 0.493640, loss_mean: 0.815273, proj_loss: -0.404321, loss_mean_cls: 0.082688, grad_norm: 0.346490 +Steps: 0%| | 3532/1000000 [14:40<68:28:59, 4.04it/s, grad_norm=0.346, loss_final=0.494, loss_mean=0.815, loss_mean_cls=0.0827, proj_loss=-0.404][2026-03-26 12:45:44] Step: 3532, Training Logs: loss_final: 0.499634, loss_mean: 0.823912, proj_loss: -0.406349, loss_mean_cls: 0.082071, grad_norm: 0.311432 +Steps: 0%| | 3533/1000000 [14:40<68:37:18, 4.03it/s, grad_norm=0.311, loss_final=0.5, loss_mean=0.824, loss_mean_cls=0.0821, proj_loss=-0.406][2026-03-26 12:45:44] Step: 3533, Training Logs: loss_final: 0.502420, loss_mean: 0.829289, proj_loss: -0.407312, loss_mean_cls: 0.080443, grad_norm: 0.416877 +Steps: 0%| | 3534/1000000 [14:41<68:33:29, 4.04it/s, grad_norm=0.417, loss_final=0.502, loss_mean=0.829, loss_mean_cls=0.0804, proj_loss=-0.407][2026-03-26 12:45:44] Step: 3534, Training Logs: loss_final: 0.508003, loss_mean: 0.835360, proj_loss: -0.407375, loss_mean_cls: 0.080018, grad_norm: 0.278895 +Steps: 0%| | 3535/1000000 [14:41<68:33:26, 4.04it/s, grad_norm=0.279, loss_final=0.508, loss_mean=0.835, loss_mean_cls=0.08, proj_loss=-0.407][2026-03-26 12:45:44] Step: 3535, Training Logs: loss_final: 0.506776, loss_mean: 0.826882, proj_loss: -0.402168, loss_mean_cls: 0.082062, grad_norm: 0.386588 +Steps: 0%| | 3536/1000000 [14:41<68:29:17, 4.04it/s, grad_norm=0.387, loss_final=0.507, loss_mean=0.827, loss_mean_cls=0.0821, proj_loss=-0.402][2026-03-26 12:45:45] Step: 3536, Training Logs: loss_final: 0.509578, loss_mean: 0.829679, proj_loss: -0.401629, loss_mean_cls: 0.081528, grad_norm: 0.437476 +Steps: 0%| | 3537/1000000 [14:41<68:42:27, 4.03it/s, grad_norm=0.437, loss_final=0.51, loss_mean=0.83, loss_mean_cls=0.0815, proj_loss=-0.402][2026-03-26 12:45:45] Step: 3537, Training Logs: loss_final: 0.517129, loss_mean: 0.838674, proj_loss: -0.403676, loss_mean_cls: 0.082131, grad_norm: 0.375330 +Steps: 0%| | 3538/1000000 [14:42<68:36:08, 4.03it/s, grad_norm=0.375, loss_final=0.517, loss_mean=0.839, loss_mean_cls=0.0821, proj_loss=-0.404][2026-03-26 12:45:45] Step: 3538, Training Logs: loss_final: 0.507930, loss_mean: 0.836418, proj_loss: -0.409564, loss_mean_cls: 0.081076, grad_norm: 0.467114 +Steps: 0%| | 3539/1000000 [14:42<68:35:28, 4.04it/s, grad_norm=0.467, loss_final=0.508, loss_mean=0.836, loss_mean_cls=0.0811, proj_loss=-0.41][2026-03-26 12:45:45] Step: 3539, Training Logs: loss_final: 0.510036, loss_mean: 0.845781, proj_loss: -0.415279, loss_mean_cls: 0.079535, grad_norm: 0.354410 +Steps: 0%| | 3540/1000000 [14:42<68:36:27, 4.03it/s, grad_norm=0.354, loss_final=0.51, loss_mean=0.846, loss_mean_cls=0.0795, proj_loss=-0.415][2026-03-26 12:45:46] Step: 3540, Training Logs: loss_final: 0.524679, loss_mean: 0.849173, proj_loss: -0.404703, loss_mean_cls: 0.080209, grad_norm: 0.338292 +Steps: 0%| | 3541/1000000 [14:42<69:41:55, 3.97it/s, grad_norm=0.338, loss_final=0.525, loss_mean=0.849, loss_mean_cls=0.0802, proj_loss=-0.405][2026-03-26 12:45:46] Step: 3541, Training Logs: loss_final: 0.497081, loss_mean: 0.818901, proj_loss: -0.405307, loss_mean_cls: 0.083487, grad_norm: 0.381346 +Steps: 0%| | 3542/1000000 [14:43<69:17:03, 4.00it/s, grad_norm=0.381, loss_final=0.497, loss_mean=0.819, loss_mean_cls=0.0835, proj_loss=-0.405][2026-03-26 12:45:46] Step: 3542, Training Logs: loss_final: 0.528062, loss_mean: 0.855808, proj_loss: -0.408156, loss_mean_cls: 0.080411, grad_norm: 0.442708 +Steps: 0%| | 3543/1000000 [14:43<69:01:52, 4.01it/s, grad_norm=0.443, loss_final=0.528, loss_mean=0.856, loss_mean_cls=0.0804, proj_loss=-0.408][2026-03-26 12:45:46] Step: 3543, Training Logs: loss_final: 0.518697, loss_mean: 0.850915, proj_loss: -0.411283, loss_mean_cls: 0.079065, grad_norm: 0.391079 +Steps: 0%| | 3544/1000000 [14:43<68:52:53, 4.02it/s, grad_norm=0.391, loss_final=0.519, loss_mean=0.851, loss_mean_cls=0.0791, proj_loss=-0.411][2026-03-26 12:45:47] Step: 3544, Training Logs: loss_final: 0.507750, loss_mean: 0.824040, proj_loss: -0.400793, loss_mean_cls: 0.084504, grad_norm: 0.392529 +Steps: 0%| | 3545/1000000 [14:43<68:59:12, 4.01it/s, grad_norm=0.393, loss_final=0.508, loss_mean=0.824, loss_mean_cls=0.0845, proj_loss=-0.401][2026-03-26 12:45:47] Step: 3545, Training Logs: loss_final: 0.513194, loss_mean: 0.833320, proj_loss: -0.403918, loss_mean_cls: 0.083792, grad_norm: 0.346084 +Steps: 0%| | 3546/1000000 [14:44<68:51:37, 4.02it/s, grad_norm=0.346, loss_final=0.513, loss_mean=0.833, loss_mean_cls=0.0838, proj_loss=-0.404][2026-03-26 12:45:47] Step: 3546, Training Logs: loss_final: 0.523147, loss_mean: 0.854080, proj_loss: -0.409340, loss_mean_cls: 0.078406, grad_norm: 0.317943 +Steps: 0%| | 3547/1000000 [14:44<68:42:43, 4.03it/s, grad_norm=0.318, loss_final=0.523, loss_mean=0.854, loss_mean_cls=0.0784, proj_loss=-0.409][2026-03-26 12:45:47] Step: 3547, Training Logs: loss_final: 0.511958, loss_mean: 0.838211, proj_loss: -0.407907, loss_mean_cls: 0.081654, grad_norm: 0.417127 +Steps: 0%| | 3548/1000000 [14:44<68:40:20, 4.03it/s, grad_norm=0.417, loss_final=0.512, loss_mean=0.838, loss_mean_cls=0.0817, proj_loss=-0.408][2026-03-26 12:45:48] Step: 3548, Training Logs: loss_final: 0.514094, loss_mean: 0.843098, proj_loss: -0.409731, loss_mean_cls: 0.080727, grad_norm: 0.361085 +Steps: 0%| | 3549/1000000 [14:44<68:47:34, 4.02it/s, grad_norm=0.361, loss_final=0.514, loss_mean=0.843, loss_mean_cls=0.0807, proj_loss=-0.41][2026-03-26 12:45:48] Step: 3549, Training Logs: loss_final: 0.522315, loss_mean: 0.844583, proj_loss: -0.404757, loss_mean_cls: 0.082490, grad_norm: 0.339886 +Steps: 0%| | 3550/1000000 [14:45<68:43:28, 4.03it/s, grad_norm=0.34, loss_final=0.522, loss_mean=0.845, loss_mean_cls=0.0825, proj_loss=-0.405][2026-03-26 12:45:48] Step: 3550, Training Logs: loss_final: 0.500092, loss_mean: 0.835335, proj_loss: -0.415265, loss_mean_cls: 0.080022, grad_norm: 0.505184 +Steps: 0%| | 3551/1000000 [14:45<68:39:05, 4.03it/s, grad_norm=0.505, loss_final=0.5, loss_mean=0.835, loss_mean_cls=0.08, proj_loss=-0.415][2026-03-26 12:45:48] Step: 3551, Training Logs: loss_final: 0.502440, loss_mean: 0.831027, proj_loss: -0.409150, loss_mean_cls: 0.080564, grad_norm: 0.327858 +Steps: 0%| | 3552/1000000 [14:45<68:36:32, 4.03it/s, grad_norm=0.328, loss_final=0.502, loss_mean=0.831, loss_mean_cls=0.0806, proj_loss=-0.409][2026-03-26 12:45:49] Step: 3552, Training Logs: loss_final: 0.519830, loss_mean: 0.841401, proj_loss: -0.403959, loss_mean_cls: 0.082389, grad_norm: 0.388587 +Steps: 0%| | 3553/1000000 [14:45<68:33:11, 4.04it/s, grad_norm=0.389, loss_final=0.52, loss_mean=0.841, loss_mean_cls=0.0824, proj_loss=-0.404][2026-03-26 12:45:49] Step: 3553, Training Logs: loss_final: 0.518194, loss_mean: 0.839168, proj_loss: -0.402770, loss_mean_cls: 0.081796, grad_norm: 0.449439 +Steps: 0%| | 3554/1000000 [14:46<68:31:13, 4.04it/s, grad_norm=0.449, loss_final=0.518, loss_mean=0.839, loss_mean_cls=0.0818, proj_loss=-0.403][2026-03-26 12:45:49] Step: 3554, Training Logs: loss_final: 0.517883, loss_mean: 0.844927, proj_loss: -0.408445, loss_mean_cls: 0.081400, grad_norm: 0.280500 +Steps: 0%| | 3555/1000000 [14:46<68:31:33, 4.04it/s, grad_norm=0.28, loss_final=0.518, loss_mean=0.845, loss_mean_cls=0.0814, proj_loss=-0.408][2026-03-26 12:45:49] Step: 3555, Training Logs: loss_final: 0.517477, loss_mean: 0.839751, proj_loss: -0.404174, loss_mean_cls: 0.081899, grad_norm: 0.553329 +Steps: 0%| | 3556/1000000 [14:46<68:32:26, 4.04it/s, grad_norm=0.553, loss_final=0.517, loss_mean=0.84, loss_mean_cls=0.0819, proj_loss=-0.404][2026-03-26 12:45:49] Step: 3556, Training Logs: loss_final: 0.505699, loss_mean: 0.836198, proj_loss: -0.411076, loss_mean_cls: 0.080577, grad_norm: 0.289156 +Steps: 0%| | 3557/1000000 [14:46<68:34:26, 4.04it/s, grad_norm=0.289, loss_final=0.506, loss_mean=0.836, loss_mean_cls=0.0806, proj_loss=-0.411][2026-03-26 12:45:50] Step: 3557, Training Logs: loss_final: 0.507838, loss_mean: 0.843560, proj_loss: -0.414218, loss_mean_cls: 0.078496, grad_norm: 0.422127 +Steps: 0%| | 3558/1000000 [14:47<68:34:36, 4.04it/s, grad_norm=0.422, loss_final=0.508, loss_mean=0.844, loss_mean_cls=0.0785, proj_loss=-0.414][2026-03-26 12:45:50] Step: 3558, Training Logs: loss_final: 0.503096, loss_mean: 0.809302, proj_loss: -0.392193, loss_mean_cls: 0.085987, grad_norm: 0.371066 +Steps: 0%| | 3559/1000000 [14:47<68:33:45, 4.04it/s, grad_norm=0.371, loss_final=0.503, loss_mean=0.809, loss_mean_cls=0.086, proj_loss=-0.392][2026-03-26 12:45:50] Step: 3559, Training Logs: loss_final: 0.515555, loss_mean: 0.841609, proj_loss: -0.407828, loss_mean_cls: 0.081774, grad_norm: 0.329637 +Steps: 0%| | 3560/1000000 [14:47<68:31:18, 4.04it/s, grad_norm=0.33, loss_final=0.516, loss_mean=0.842, loss_mean_cls=0.0818, proj_loss=-0.408][2026-03-26 12:45:50] Step: 3560, Training Logs: loss_final: 0.493548, loss_mean: 0.817455, proj_loss: -0.405050, loss_mean_cls: 0.081143, grad_norm: 0.688756 +Steps: 0%| | 3561/1000000 [14:47<68:29:10, 4.04it/s, grad_norm=0.689, loss_final=0.494, loss_mean=0.817, loss_mean_cls=0.0811, proj_loss=-0.405][2026-03-26 12:45:51] Step: 3561, Training Logs: loss_final: 0.508834, loss_mean: 0.833329, proj_loss: -0.405616, loss_mean_cls: 0.081121, grad_norm: 0.545395 +Steps: 0%| | 3562/1000000 [14:48<68:27:40, 4.04it/s, grad_norm=0.545, loss_final=0.509, loss_mean=0.833, loss_mean_cls=0.0811, proj_loss=-0.406][2026-03-26 12:45:51] Step: 3562, Training Logs: loss_final: 0.521439, loss_mean: 0.844748, proj_loss: -0.405724, loss_mean_cls: 0.082414, grad_norm: 0.296192 +Steps: 0%| | 3563/1000000 [14:48<68:29:58, 4.04it/s, grad_norm=0.296, loss_final=0.521, loss_mean=0.845, loss_mean_cls=0.0824, proj_loss=-0.406][2026-03-26 12:45:51] Step: 3563, Training Logs: loss_final: 0.505655, loss_mean: 0.825389, proj_loss: -0.402907, loss_mean_cls: 0.083173, grad_norm: 0.404471 +Steps: 0%| | 3564/1000000 [14:48<68:29:56, 4.04it/s, grad_norm=0.404, loss_final=0.506, loss_mean=0.825, loss_mean_cls=0.0832, proj_loss=-0.403][2026-03-26 12:45:51] Step: 3564, Training Logs: loss_final: 0.508914, loss_mean: 0.834230, proj_loss: -0.406172, loss_mean_cls: 0.080856, grad_norm: 0.305353 +Steps: 0%| | 3565/1000000 [14:48<68:30:54, 4.04it/s, grad_norm=0.305, loss_final=0.509, loss_mean=0.834, loss_mean_cls=0.0809, proj_loss=-0.406][2026-03-26 12:45:52] Step: 3565, Training Logs: loss_final: 0.527416, loss_mean: 0.847227, proj_loss: -0.402998, loss_mean_cls: 0.083187, grad_norm: 0.421062 +Steps: 0%| | 3566/1000000 [14:49<68:32:42, 4.04it/s, grad_norm=0.421, loss_final=0.527, loss_mean=0.847, loss_mean_cls=0.0832, proj_loss=-0.403][2026-03-26 12:45:52] Step: 3566, Training Logs: loss_final: 0.523527, loss_mean: 0.838632, proj_loss: -0.398190, loss_mean_cls: 0.083084, grad_norm: 0.383648 +Steps: 0%| | 3567/1000000 [14:49<68:29:49, 4.04it/s, grad_norm=0.384, loss_final=0.524, loss_mean=0.839, loss_mean_cls=0.0831, proj_loss=-0.398][2026-03-26 12:45:52] Step: 3567, Training Logs: loss_final: 0.522856, loss_mean: 0.849405, proj_loss: -0.408710, loss_mean_cls: 0.082162, grad_norm: 0.534036 +Steps: 0%| | 3568/1000000 [14:49<68:30:33, 4.04it/s, grad_norm=0.534, loss_final=0.523, loss_mean=0.849, loss_mean_cls=0.0822, proj_loss=-0.409][2026-03-26 12:45:52] Step: 3568, Training Logs: loss_final: 0.524075, loss_mean: 0.848240, proj_loss: -0.406158, loss_mean_cls: 0.081993, grad_norm: 0.365085 +Steps: 0%| | 3569/1000000 [14:49<68:26:44, 4.04it/s, grad_norm=0.365, loss_final=0.524, loss_mean=0.848, loss_mean_cls=0.082, proj_loss=-0.406][2026-03-26 12:45:53] Step: 3569, Training Logs: loss_final: 0.508194, loss_mean: 0.824962, proj_loss: -0.399902, loss_mean_cls: 0.083135, grad_norm: 0.459047 +Steps: 0%| | 3570/1000000 [14:50<68:26:08, 4.04it/s, grad_norm=0.459, loss_final=0.508, loss_mean=0.825, loss_mean_cls=0.0831, proj_loss=-0.4][2026-03-26 12:45:53] Step: 3570, Training Logs: loss_final: 0.505013, loss_mean: 0.829293, proj_loss: -0.406925, loss_mean_cls: 0.082645, grad_norm: 0.549123 +Steps: 0%| | 3571/1000000 [14:50<68:29:00, 4.04it/s, grad_norm=0.549, loss_final=0.505, loss_mean=0.829, loss_mean_cls=0.0826, proj_loss=-0.407][2026-03-26 12:45:53] Step: 3571, Training Logs: loss_final: 0.509695, loss_mean: 0.837042, proj_loss: -0.410285, loss_mean_cls: 0.082939, grad_norm: 0.290254 +Steps: 0%| | 3572/1000000 [14:50<68:27:20, 4.04it/s, grad_norm=0.29, loss_final=0.51, loss_mean=0.837, loss_mean_cls=0.0829, proj_loss=-0.41][2026-03-26 12:45:53] Step: 3572, Training Logs: loss_final: 0.525635, loss_mean: 0.852764, proj_loss: -0.405718, loss_mean_cls: 0.078589, grad_norm: 0.421347 +Steps: 0%| | 3573/1000000 [14:50<68:26:58, 4.04it/s, grad_norm=0.421, loss_final=0.526, loss_mean=0.853, loss_mean_cls=0.0786, proj_loss=-0.406][2026-03-26 12:45:54] Step: 3573, Training Logs: loss_final: 0.520784, loss_mean: 0.838871, proj_loss: -0.400578, loss_mean_cls: 0.082491, grad_norm: 0.586180 +Steps: 0%| | 3574/1000000 [14:51<68:28:58, 4.04it/s, grad_norm=0.586, loss_final=0.521, loss_mean=0.839, loss_mean_cls=0.0825, proj_loss=-0.401][2026-03-26 12:45:54] Step: 3574, Training Logs: loss_final: 0.522617, loss_mean: 0.848198, proj_loss: -0.406268, loss_mean_cls: 0.080687, grad_norm: 0.376924 +Steps: 0%| | 3575/1000000 [14:51<68:27:24, 4.04it/s, grad_norm=0.377, loss_final=0.523, loss_mean=0.848, loss_mean_cls=0.0807, proj_loss=-0.406][2026-03-26 12:45:54] Step: 3575, Training Logs: loss_final: 0.520524, loss_mean: 0.843934, proj_loss: -0.404331, loss_mean_cls: 0.080922, grad_norm: 0.437464 +Steps: 0%| | 3576/1000000 [14:51<68:26:49, 4.04it/s, grad_norm=0.437, loss_final=0.521, loss_mean=0.844, loss_mean_cls=0.0809, proj_loss=-0.404][2026-03-26 12:45:54] Step: 3576, Training Logs: loss_final: 0.517550, loss_mean: 0.845616, proj_loss: -0.407481, loss_mean_cls: 0.079416, grad_norm: 0.563750 +Steps: 0%| | 3577/1000000 [14:51<68:25:37, 4.04it/s, grad_norm=0.564, loss_final=0.518, loss_mean=0.846, loss_mean_cls=0.0794, proj_loss=-0.407][2026-03-26 12:45:55] Step: 3577, Training Logs: loss_final: 0.520900, loss_mean: 0.848744, proj_loss: -0.407705, loss_mean_cls: 0.079861, grad_norm: 0.384412 +Steps: 0%| | 3578/1000000 [14:52<68:25:45, 4.04it/s, grad_norm=0.384, loss_final=0.521, loss_mean=0.849, loss_mean_cls=0.0799, proj_loss=-0.408][2026-03-26 12:45:55] Step: 3578, Training Logs: loss_final: 0.534002, loss_mean: 0.851136, proj_loss: -0.398009, loss_mean_cls: 0.080875, grad_norm: 0.403855 +Steps: 0%| | 3579/1000000 [14:52<68:25:35, 4.04it/s, grad_norm=0.404, loss_final=0.534, loss_mean=0.851, loss_mean_cls=0.0809, proj_loss=-0.398][2026-03-26 12:45:55] Step: 3579, Training Logs: loss_final: 0.496596, loss_mean: 0.813266, proj_loss: -0.402692, loss_mean_cls: 0.086023, grad_norm: 0.350962 +Steps: 0%| | 3580/1000000 [14:52<68:28:53, 4.04it/s, grad_norm=0.351, loss_final=0.497, loss_mean=0.813, loss_mean_cls=0.086, proj_loss=-0.403][2026-03-26 12:45:55] Step: 3580, Training Logs: loss_final: 0.506164, loss_mean: 0.821918, proj_loss: -0.398927, loss_mean_cls: 0.083173, grad_norm: 0.318849 +Steps: 0%| | 3581/1000000 [14:52<68:27:01, 4.04it/s, grad_norm=0.319, loss_final=0.506, loss_mean=0.822, loss_mean_cls=0.0832, proj_loss=-0.399][2026-03-26 12:45:56] Step: 3581, Training Logs: loss_final: 0.519401, loss_mean: 0.838081, proj_loss: -0.402263, loss_mean_cls: 0.083582, grad_norm: 0.347641 +Steps: 0%| | 3582/1000000 [14:53<68:27:54, 4.04it/s, grad_norm=0.348, loss_final=0.519, loss_mean=0.838, loss_mean_cls=0.0836, proj_loss=-0.402][2026-03-26 12:45:56] Step: 3582, Training Logs: loss_final: 0.516409, loss_mean: 0.841288, proj_loss: -0.405863, loss_mean_cls: 0.080984, grad_norm: 0.367881 +Steps: 0%| | 3583/1000000 [14:53<68:25:10, 4.05it/s, grad_norm=0.368, loss_final=0.516, loss_mean=0.841, loss_mean_cls=0.081, proj_loss=-0.406][2026-03-26 12:45:56] Step: 3583, Training Logs: loss_final: 0.515410, loss_mean: 0.844092, proj_loss: -0.408846, loss_mean_cls: 0.080165, grad_norm: 0.253695 +Steps: 0%| | 3584/1000000 [14:53<68:24:37, 4.05it/s, grad_norm=0.254, loss_final=0.515, loss_mean=0.844, loss_mean_cls=0.0802, proj_loss=-0.409][2026-03-26 12:45:56] Step: 3584, Training Logs: loss_final: 0.505651, loss_mean: 0.832924, proj_loss: -0.410059, loss_mean_cls: 0.082786, grad_norm: 0.315499 +Steps: 0%| | 3585/1000000 [14:53<68:22:09, 4.05it/s, grad_norm=0.315, loss_final=0.506, loss_mean=0.833, loss_mean_cls=0.0828, proj_loss=-0.41][2026-03-26 12:45:57] Step: 3585, Training Logs: loss_final: 0.506069, loss_mean: 0.837087, proj_loss: -0.411809, loss_mean_cls: 0.080791, grad_norm: 0.333186 +Steps: 0%| | 3586/1000000 [14:54<68:23:36, 4.05it/s, grad_norm=0.333, loss_final=0.506, loss_mean=0.837, loss_mean_cls=0.0808, proj_loss=-0.412][2026-03-26 12:45:57] Step: 3586, Training Logs: loss_final: 0.509582, loss_mean: 0.846769, proj_loss: -0.416285, loss_mean_cls: 0.079099, grad_norm: 0.300653 +Steps: 0%| | 3587/1000000 [14:54<68:20:54, 4.05it/s, grad_norm=0.301, loss_final=0.51, loss_mean=0.847, loss_mean_cls=0.0791, proj_loss=-0.416][2026-03-26 12:45:57] Step: 3587, Training Logs: loss_final: 0.505109, loss_mean: 0.837580, proj_loss: -0.412093, loss_mean_cls: 0.079622, grad_norm: 0.265846 +Steps: 0%| | 3588/1000000 [14:54<68:24:02, 4.05it/s, grad_norm=0.266, loss_final=0.505, loss_mean=0.838, loss_mean_cls=0.0796, proj_loss=-0.412][2026-03-26 12:45:57] Step: 3588, Training Logs: loss_final: 0.498787, loss_mean: 0.830976, proj_loss: -0.413601, loss_mean_cls: 0.081412, grad_norm: 0.286567 +Steps: 0%| | 3589/1000000 [14:54<68:25:42, 4.04it/s, grad_norm=0.287, loss_final=0.499, loss_mean=0.831, loss_mean_cls=0.0814, proj_loss=-0.414][2026-03-26 12:45:58] Step: 3589, Training Logs: loss_final: 0.509306, loss_mean: 0.834751, proj_loss: -0.407535, loss_mean_cls: 0.082090, grad_norm: 0.333661 +Steps: 0%| | 3590/1000000 [14:55<68:26:52, 4.04it/s, grad_norm=0.334, loss_final=0.509, loss_mean=0.835, loss_mean_cls=0.0821, proj_loss=-0.408][2026-03-26 12:45:58] Step: 3590, Training Logs: loss_final: 0.517653, loss_mean: 0.844880, proj_loss: -0.407447, loss_mean_cls: 0.080220, grad_norm: 0.509025 +Steps: 0%| | 3591/1000000 [14:55<68:31:27, 4.04it/s, grad_norm=0.509, loss_final=0.518, loss_mean=0.845, loss_mean_cls=0.0802, proj_loss=-0.407][2026-03-26 12:45:58] Step: 3591, Training Logs: loss_final: 0.518435, loss_mean: 0.850389, proj_loss: -0.411996, loss_mean_cls: 0.080041, grad_norm: 0.492812 +Steps: 0%| | 3592/1000000 [14:55<68:50:21, 4.02it/s, grad_norm=0.493, loss_final=0.518, loss_mean=0.85, loss_mean_cls=0.08, proj_loss=-0.412][2026-03-26 12:45:58] Step: 3592, Training Logs: loss_final: 0.495742, loss_mean: 0.827510, proj_loss: -0.412127, loss_mean_cls: 0.080360, grad_norm: 0.389477 +Steps: 0%| | 3593/1000000 [14:55<68:46:26, 4.02it/s, grad_norm=0.389, loss_final=0.496, loss_mean=0.828, loss_mean_cls=0.0804, proj_loss=-0.412][2026-03-26 12:45:59] Step: 3593, Training Logs: loss_final: 0.507025, loss_mean: 0.829123, proj_loss: -0.404597, loss_mean_cls: 0.082499, grad_norm: 0.307458 +Steps: 0%| | 3594/1000000 [14:56<68:41:06, 4.03it/s, grad_norm=0.307, loss_final=0.507, loss_mean=0.829, loss_mean_cls=0.0825, proj_loss=-0.405][2026-03-26 12:45:59] Step: 3594, Training Logs: loss_final: 0.516588, loss_mean: 0.834362, proj_loss: -0.401215, loss_mean_cls: 0.083442, grad_norm: 0.377413 +Steps: 0%| | 3595/1000000 [14:56<68:39:04, 4.03it/s, grad_norm=0.377, loss_final=0.517, loss_mean=0.834, loss_mean_cls=0.0834, proj_loss=-0.401][2026-03-26 12:45:59] Step: 3595, Training Logs: loss_final: 0.499437, loss_mean: 0.819987, proj_loss: -0.404499, loss_mean_cls: 0.083949, grad_norm: 0.403236 +Steps: 0%| | 3596/1000000 [14:56<68:33:53, 4.04it/s, grad_norm=0.403, loss_final=0.499, loss_mean=0.82, loss_mean_cls=0.0839, proj_loss=-0.404][2026-03-26 12:45:59] Step: 3596, Training Logs: loss_final: 0.516658, loss_mean: 0.848080, proj_loss: -0.411957, loss_mean_cls: 0.080536, grad_norm: 0.335094 +Steps: 0%| | 3597/1000000 [14:56<68:31:56, 4.04it/s, grad_norm=0.335, loss_final=0.517, loss_mean=0.848, loss_mean_cls=0.0805, proj_loss=-0.412][2026-03-26 12:46:00] Step: 3597, Training Logs: loss_final: 0.506130, loss_mean: 0.833648, proj_loss: -0.407586, loss_mean_cls: 0.080069, grad_norm: 0.389369 +Steps: 0%| | 3598/1000000 [14:56<68:27:28, 4.04it/s, grad_norm=0.389, loss_final=0.506, loss_mean=0.834, loss_mean_cls=0.0801, proj_loss=-0.408][2026-03-26 12:46:00] Step: 3598, Training Logs: loss_final: 0.512993, loss_mean: 0.844749, proj_loss: -0.410385, loss_mean_cls: 0.078629, grad_norm: 0.356500 +Steps: 0%| | 3599/1000000 [14:57<68:27:48, 4.04it/s, grad_norm=0.356, loss_final=0.513, loss_mean=0.845, loss_mean_cls=0.0786, proj_loss=-0.41][2026-03-26 12:46:00] Step: 3599, Training Logs: loss_final: 0.514482, loss_mean: 0.837017, proj_loss: -0.403787, loss_mean_cls: 0.081252, grad_norm: 0.381094 +Steps: 0%| | 3600/1000000 [14:57<68:25:56, 4.04it/s, grad_norm=0.381, loss_final=0.514, loss_mean=0.837, loss_mean_cls=0.0813, proj_loss=-0.404][2026-03-26 12:46:00] Step: 3600, Training Logs: loss_final: 0.501264, loss_mean: 0.838855, proj_loss: -0.416451, loss_mean_cls: 0.078860, grad_norm: 0.325663 +Steps: 0%| | 3601/1000000 [14:57<68:28:12, 4.04it/s, grad_norm=0.326, loss_final=0.501, loss_mean=0.839, loss_mean_cls=0.0789, proj_loss=-0.416][2026-03-26 12:46:01] Step: 3601, Training Logs: loss_final: 0.506743, loss_mean: 0.834163, proj_loss: -0.408740, loss_mean_cls: 0.081320, grad_norm: 0.317091 +Steps: 0%| | 3602/1000000 [14:57<68:28:07, 4.04it/s, grad_norm=0.317, loss_final=0.507, loss_mean=0.834, loss_mean_cls=0.0813, proj_loss=-0.409][2026-03-26 12:46:01] Step: 3602, Training Logs: loss_final: 0.503082, loss_mean: 0.838269, proj_loss: -0.415503, loss_mean_cls: 0.080316, grad_norm: 0.297723 +Steps: 0%| | 3603/1000000 [14:58<68:30:45, 4.04it/s, grad_norm=0.298, loss_final=0.503, loss_mean=0.838, loss_mean_cls=0.0803, proj_loss=-0.416][2026-03-26 12:46:01] Step: 3603, Training Logs: loss_final: 0.507620, loss_mean: 0.847431, proj_loss: -0.416942, loss_mean_cls: 0.077130, grad_norm: 0.378232 +Steps: 0%| | 3604/1000000 [14:58<68:24:24, 4.05it/s, grad_norm=0.378, loss_final=0.508, loss_mean=0.847, loss_mean_cls=0.0771, proj_loss=-0.417][2026-03-26 12:46:01] Step: 3604, Training Logs: loss_final: 0.509237, loss_mean: 0.826494, proj_loss: -0.400230, loss_mean_cls: 0.082972, grad_norm: 0.280495 +Steps: 0%| | 3605/1000000 [14:58<68:27:57, 4.04it/s, grad_norm=0.28, loss_final=0.509, loss_mean=0.826, loss_mean_cls=0.083, proj_loss=-0.4][2026-03-26 12:46:02] Step: 3605, Training Logs: loss_final: 0.509303, loss_mean: 0.830662, proj_loss: -0.404107, loss_mean_cls: 0.082748, grad_norm: 0.241192 +Steps: 0%| | 3606/1000000 [14:58<68:26:37, 4.04it/s, grad_norm=0.241, loss_final=0.509, loss_mean=0.831, loss_mean_cls=0.0827, proj_loss=-0.404][2026-03-26 12:46:02] Step: 3606, Training Logs: loss_final: 0.513110, loss_mean: 0.834304, proj_loss: -0.403151, loss_mean_cls: 0.081957, grad_norm: 0.406226 +Steps: 0%| | 3607/1000000 [14:59<68:28:24, 4.04it/s, grad_norm=0.406, loss_final=0.513, loss_mean=0.834, loss_mean_cls=0.082, proj_loss=-0.403][2026-03-26 12:46:02] Step: 3607, Training Logs: loss_final: 0.506515, loss_mean: 0.840009, proj_loss: -0.413498, loss_mean_cls: 0.080004, grad_norm: 0.305815 +Steps: 0%| | 3608/1000000 [14:59<68:30:35, 4.04it/s, grad_norm=0.306, loss_final=0.507, loss_mean=0.84, loss_mean_cls=0.08, proj_loss=-0.413][2026-03-26 12:46:02] Step: 3608, Training Logs: loss_final: 0.516086, loss_mean: 0.844160, proj_loss: -0.408995, loss_mean_cls: 0.080921, grad_norm: 0.431707 +Steps: 0%| | 3609/1000000 [14:59<68:30:58, 4.04it/s, grad_norm=0.432, loss_final=0.516, loss_mean=0.844, loss_mean_cls=0.0809, proj_loss=-0.409][2026-03-26 12:46:03] Step: 3609, Training Logs: loss_final: 0.510473, loss_mean: 0.834061, proj_loss: -0.404770, loss_mean_cls: 0.081182, grad_norm: 0.416426 +Steps: 0%| | 3610/1000000 [14:59<68:35:45, 4.03it/s, grad_norm=0.416, loss_final=0.51, loss_mean=0.834, loss_mean_cls=0.0812, proj_loss=-0.405][2026-03-26 12:46:03] Step: 3610, Training Logs: loss_final: 0.502053, loss_mean: 0.825950, proj_loss: -0.406481, loss_mean_cls: 0.082584, grad_norm: 0.355970 +Steps: 0%| | 3611/1000000 [15:00<68:35:13, 4.04it/s, grad_norm=0.356, loss_final=0.502, loss_mean=0.826, loss_mean_cls=0.0826, proj_loss=-0.406][2026-03-26 12:46:03] Step: 3611, Training Logs: loss_final: 0.503562, loss_mean: 0.834486, proj_loss: -0.410799, loss_mean_cls: 0.079876, grad_norm: 0.304747 +Steps: 0%| | 3612/1000000 [15:00<69:18:33, 3.99it/s, grad_norm=0.305, loss_final=0.504, loss_mean=0.834, loss_mean_cls=0.0799, proj_loss=-0.411][2026-03-26 12:46:03] Step: 3612, Training Logs: loss_final: 0.508637, loss_mean: 0.832248, proj_loss: -0.404869, loss_mean_cls: 0.081258, grad_norm: 0.425133 +Steps: 0%| | 3613/1000000 [15:00<69:00:28, 4.01it/s, grad_norm=0.425, loss_final=0.509, loss_mean=0.832, loss_mean_cls=0.0813, proj_loss=-0.405][2026-03-26 12:46:04] Step: 3613, Training Logs: loss_final: 0.507895, loss_mean: 0.833864, proj_loss: -0.407588, loss_mean_cls: 0.081619, grad_norm: 0.281559 +Steps: 0%| | 3614/1000000 [15:00<68:49:40, 4.02it/s, grad_norm=0.282, loss_final=0.508, loss_mean=0.834, loss_mean_cls=0.0816, proj_loss=-0.408][2026-03-26 12:46:04] Step: 3614, Training Logs: loss_final: 0.509972, loss_mean: 0.828229, proj_loss: -0.400746, loss_mean_cls: 0.082489, grad_norm: 0.651949 +Steps: 0%| | 3615/1000000 [15:01<68:44:21, 4.03it/s, grad_norm=0.652, loss_final=0.51, loss_mean=0.828, loss_mean_cls=0.0825, proj_loss=-0.401][2026-03-26 12:46:04] Step: 3615, Training Logs: loss_final: 0.502415, loss_mean: 0.834817, proj_loss: -0.411954, loss_mean_cls: 0.079553, grad_norm: 0.494465 +Steps: 0%| | 3616/1000000 [15:01<68:40:28, 4.03it/s, grad_norm=0.494, loss_final=0.502, loss_mean=0.835, loss_mean_cls=0.0796, proj_loss=-0.412][2026-03-26 12:46:04] Step: 3616, Training Logs: loss_final: 0.498880, loss_mean: 0.818519, proj_loss: -0.403479, loss_mean_cls: 0.083841, grad_norm: 0.328616 +Steps: 0%| | 3617/1000000 [15:01<68:38:54, 4.03it/s, grad_norm=0.329, loss_final=0.499, loss_mean=0.819, loss_mean_cls=0.0838, proj_loss=-0.403][2026-03-26 12:46:05] Step: 3617, Training Logs: loss_final: 0.523574, loss_mean: 0.854860, proj_loss: -0.409965, loss_mean_cls: 0.078679, grad_norm: 0.456010 +Steps: 0%| | 3618/1000000 [15:01<68:36:53, 4.03it/s, grad_norm=0.456, loss_final=0.524, loss_mean=0.855, loss_mean_cls=0.0787, proj_loss=-0.41][2026-03-26 12:46:05] Step: 3618, Training Logs: loss_final: 0.513549, loss_mean: 0.845415, proj_loss: -0.412124, loss_mean_cls: 0.080259, grad_norm: 0.445165 +Steps: 0%| | 3619/1000000 [15:02<68:35:04, 4.04it/s, grad_norm=0.445, loss_final=0.514, loss_mean=0.845, loss_mean_cls=0.0803, proj_loss=-0.412][2026-03-26 12:46:05] Step: 3619, Training Logs: loss_final: 0.489786, loss_mean: 0.809552, proj_loss: -0.404980, loss_mean_cls: 0.085214, grad_norm: 0.327229 +Steps: 0%| | 3620/1000000 [15:02<68:33:30, 4.04it/s, grad_norm=0.327, loss_final=0.49, loss_mean=0.81, loss_mean_cls=0.0852, proj_loss=-0.405][2026-03-26 12:46:05] Step: 3620, Training Logs: loss_final: 0.510837, loss_mean: 0.828909, proj_loss: -0.400795, loss_mean_cls: 0.082723, grad_norm: 0.288548 +Steps: 0%| | 3621/1000000 [15:02<68:29:12, 4.04it/s, grad_norm=0.289, loss_final=0.511, loss_mean=0.829, loss_mean_cls=0.0827, proj_loss=-0.401][2026-03-26 12:46:06] Step: 3621, Training Logs: loss_final: 0.517293, loss_mean: 0.841015, proj_loss: -0.405705, loss_mean_cls: 0.081984, grad_norm: 0.468167 +Steps: 0%| | 3622/1000000 [15:02<68:28:45, 4.04it/s, grad_norm=0.468, loss_final=0.517, loss_mean=0.841, loss_mean_cls=0.082, proj_loss=-0.406][2026-03-26 12:46:06] Step: 3622, Training Logs: loss_final: 0.508610, loss_mean: 0.835925, proj_loss: -0.408756, loss_mean_cls: 0.081442, grad_norm: 0.290723 +Steps: 0%| | 3623/1000000 [15:03<68:23:20, 4.05it/s, grad_norm=0.291, loss_final=0.509, loss_mean=0.836, loss_mean_cls=0.0814, proj_loss=-0.409][2026-03-26 12:46:06] Step: 3623, Training Logs: loss_final: 0.516975, loss_mean: 0.836577, proj_loss: -0.402684, loss_mean_cls: 0.083082, grad_norm: 0.369020 +Steps: 0%| | 3624/1000000 [15:03<68:23:49, 4.05it/s, grad_norm=0.369, loss_final=0.517, loss_mean=0.837, loss_mean_cls=0.0831, proj_loss=-0.403][2026-03-26 12:46:06] Step: 3624, Training Logs: loss_final: 0.523297, loss_mean: 0.849811, proj_loss: -0.406169, loss_mean_cls: 0.079656, grad_norm: 0.425110 +Steps: 0%| | 3625/1000000 [15:03<68:21:19, 4.05it/s, grad_norm=0.425, loss_final=0.523, loss_mean=0.85, loss_mean_cls=0.0797, proj_loss=-0.406][2026-03-26 12:46:07] Step: 3625, Training Logs: loss_final: 0.526461, loss_mean: 0.854550, proj_loss: -0.407943, loss_mean_cls: 0.079855, grad_norm: 0.406799 +Steps: 0%| | 3626/1000000 [15:03<68:23:58, 4.05it/s, grad_norm=0.407, loss_final=0.526, loss_mean=0.855, loss_mean_cls=0.0799, proj_loss=-0.408][2026-03-26 12:46:07] Step: 3626, Training Logs: loss_final: 0.516079, loss_mean: 0.838558, proj_loss: -0.403138, loss_mean_cls: 0.080658, grad_norm: 0.360289 +Steps: 0%| | 3627/1000000 [15:04<68:25:12, 4.05it/s, grad_norm=0.36, loss_final=0.516, loss_mean=0.839, loss_mean_cls=0.0807, proj_loss=-0.403][2026-03-26 12:46:07] Step: 3627, Training Logs: loss_final: 0.519080, loss_mean: 0.838098, proj_loss: -0.402261, loss_mean_cls: 0.083243, grad_norm: 0.331728 +Steps: 0%| | 3628/1000000 [15:04<68:26:30, 4.04it/s, grad_norm=0.332, loss_final=0.519, loss_mean=0.838, loss_mean_cls=0.0832, proj_loss=-0.402][2026-03-26 12:46:07] Step: 3628, Training Logs: loss_final: 0.516297, loss_mean: 0.848801, proj_loss: -0.411199, loss_mean_cls: 0.078696, grad_norm: 0.302684 +Steps: 0%| | 3629/1000000 [15:04<68:25:44, 4.04it/s, grad_norm=0.303, loss_final=0.516, loss_mean=0.849, loss_mean_cls=0.0787, proj_loss=-0.411][2026-03-26 12:46:08] Step: 3629, Training Logs: loss_final: 0.505223, loss_mean: 0.842562, proj_loss: -0.415881, loss_mean_cls: 0.078542, grad_norm: 0.390954 +Steps: 0%| | 3630/1000000 [15:04<68:28:53, 4.04it/s, grad_norm=0.391, loss_final=0.505, loss_mean=0.843, loss_mean_cls=0.0785, proj_loss=-0.416][2026-03-26 12:46:08] Step: 3630, Training Logs: loss_final: 0.521803, loss_mean: 0.840277, proj_loss: -0.400645, loss_mean_cls: 0.082171, grad_norm: 0.491999 +Steps: 0%| | 3631/1000000 [15:05<68:27:50, 4.04it/s, grad_norm=0.492, loss_final=0.522, loss_mean=0.84, loss_mean_cls=0.0822, proj_loss=-0.401][2026-03-26 12:46:08] Step: 3631, Training Logs: loss_final: 0.503996, loss_mean: 0.821358, proj_loss: -0.399777, loss_mean_cls: 0.082415, grad_norm: 0.303307 +Steps: 0%| | 3632/1000000 [15:05<68:30:58, 4.04it/s, grad_norm=0.303, loss_final=0.504, loss_mean=0.821, loss_mean_cls=0.0824, proj_loss=-0.4][2026-03-26 12:46:08] Step: 3632, Training Logs: loss_final: 0.501193, loss_mean: 0.822352, proj_loss: -0.404764, loss_mean_cls: 0.083606, grad_norm: 0.522518 +Steps: 0%| | 3633/1000000 [15:05<68:29:33, 4.04it/s, grad_norm=0.523, loss_final=0.501, loss_mean=0.822, loss_mean_cls=0.0836, proj_loss=-0.405][2026-03-26 12:46:09] Step: 3633, Training Logs: loss_final: 0.501503, loss_mean: 0.837831, proj_loss: -0.415156, loss_mean_cls: 0.078828, grad_norm: 0.529379 +Steps: 0%| | 3634/1000000 [15:05<68:29:11, 4.04it/s, grad_norm=0.529, loss_final=0.502, loss_mean=0.838, loss_mean_cls=0.0788, proj_loss=-0.415][2026-03-26 12:46:09] Step: 3634, Training Logs: loss_final: 0.509354, loss_mean: 0.833627, proj_loss: -0.406056, loss_mean_cls: 0.081784, grad_norm: 0.364891 +Steps: 0%| | 3635/1000000 [15:06<68:26:47, 4.04it/s, grad_norm=0.365, loss_final=0.509, loss_mean=0.834, loss_mean_cls=0.0818, proj_loss=-0.406][2026-03-26 12:46:09] Step: 3635, Training Logs: loss_final: 0.490892, loss_mean: 0.822764, proj_loss: -0.412578, loss_mean_cls: 0.080705, grad_norm: 0.600151 +Steps: 0%| | 3636/1000000 [15:06<68:24:26, 4.05it/s, grad_norm=0.6, loss_final=0.491, loss_mean=0.823, loss_mean_cls=0.0807, proj_loss=-0.413][2026-03-26 12:46:09] Step: 3636, Training Logs: loss_final: 0.511779, loss_mean: 0.842936, proj_loss: -0.411720, loss_mean_cls: 0.080563, grad_norm: 0.623882 +Steps: 0%| | 3637/1000000 [15:06<68:25:07, 4.05it/s, grad_norm=0.624, loss_final=0.512, loss_mean=0.843, loss_mean_cls=0.0806, proj_loss=-0.412][2026-03-26 12:46:10] Step: 3637, Training Logs: loss_final: 0.515481, loss_mean: 0.834282, proj_loss: -0.400602, loss_mean_cls: 0.081801, grad_norm: 0.288171 +Steps: 0%| | 3638/1000000 [15:06<68:23:49, 4.05it/s, grad_norm=0.288, loss_final=0.515, loss_mean=0.834, loss_mean_cls=0.0818, proj_loss=-0.401][2026-03-26 12:46:10] Step: 3638, Training Logs: loss_final: 0.518390, loss_mean: 0.856958, proj_loss: -0.416724, loss_mean_cls: 0.078155, grad_norm: 0.639968 +Steps: 0%| | 3639/1000000 [15:07<68:26:12, 4.04it/s, grad_norm=0.64, loss_final=0.518, loss_mean=0.857, loss_mean_cls=0.0782, proj_loss=-0.417][2026-03-26 12:46:10] Step: 3639, Training Logs: loss_final: 0.515337, loss_mean: 0.840741, proj_loss: -0.409317, loss_mean_cls: 0.083913, grad_norm: 0.628118 +Steps: 0%| | 3640/1000000 [15:07<68:22:59, 4.05it/s, grad_norm=0.628, loss_final=0.515, loss_mean=0.841, loss_mean_cls=0.0839, proj_loss=-0.409][2026-03-26 12:46:10] Step: 3640, Training Logs: loss_final: 0.515522, loss_mean: 0.847598, proj_loss: -0.412999, loss_mean_cls: 0.080922, grad_norm: 0.387914 +Steps: 0%| | 3641/1000000 [15:07<68:20:38, 4.05it/s, grad_norm=0.388, loss_final=0.516, loss_mean=0.848, loss_mean_cls=0.0809, proj_loss=-0.413][2026-03-26 12:46:11] Step: 3641, Training Logs: loss_final: 0.509836, loss_mean: 0.842061, proj_loss: -0.412449, loss_mean_cls: 0.080224, grad_norm: 0.493965 +Steps: 0%| | 3642/1000000 [15:07<68:16:46, 4.05it/s, grad_norm=0.494, loss_final=0.51, loss_mean=0.842, loss_mean_cls=0.0802, proj_loss=-0.412][2026-03-26 12:46:11] Step: 3642, Training Logs: loss_final: 0.500348, loss_mean: 0.827497, proj_loss: -0.408632, loss_mean_cls: 0.081483, grad_norm: 0.399540 +Steps: 0%| | 3643/1000000 [15:08<68:21:07, 4.05it/s, grad_norm=0.4, loss_final=0.5, loss_mean=0.827, loss_mean_cls=0.0815, proj_loss=-0.409][2026-03-26 12:46:11] Step: 3643, Training Logs: loss_final: 0.515548, loss_mean: 0.845882, proj_loss: -0.409835, loss_mean_cls: 0.079501, grad_norm: 0.427334 +Steps: 0%| | 3644/1000000 [15:08<68:21:14, 4.05it/s, grad_norm=0.427, loss_final=0.516, loss_mean=0.846, loss_mean_cls=0.0795, proj_loss=-0.41][2026-03-26 12:46:11] Step: 3644, Training Logs: loss_final: 0.504749, loss_mean: 0.824605, proj_loss: -0.402776, loss_mean_cls: 0.082920, grad_norm: 0.427363 +Steps: 0%| | 3645/1000000 [15:08<68:20:39, 4.05it/s, grad_norm=0.427, loss_final=0.505, loss_mean=0.825, loss_mean_cls=0.0829, proj_loss=-0.403][2026-03-26 12:46:12] Step: 3645, Training Logs: loss_final: 0.519233, loss_mean: 0.842829, proj_loss: -0.405227, loss_mean_cls: 0.081631, grad_norm: 0.305770 +Steps: 0%| | 3646/1000000 [15:08<68:24:23, 4.05it/s, grad_norm=0.306, loss_final=0.519, loss_mean=0.843, loss_mean_cls=0.0816, proj_loss=-0.405][2026-03-26 12:46:12] Step: 3646, Training Logs: loss_final: 0.509962, loss_mean: 0.843556, proj_loss: -0.412565, loss_mean_cls: 0.078971, grad_norm: 0.649659 +Steps: 0%| | 3647/1000000 [15:09<68:26:09, 4.04it/s, grad_norm=0.65, loss_final=0.51, loss_mean=0.844, loss_mean_cls=0.079, proj_loss=-0.413][2026-03-26 12:46:12] Step: 3647, Training Logs: loss_final: 0.508841, loss_mean: 0.840046, proj_loss: -0.410456, loss_mean_cls: 0.079251, grad_norm: 0.454125 +Steps: 0%| | 3648/1000000 [15:09<68:29:22, 4.04it/s, grad_norm=0.454, loss_final=0.509, loss_mean=0.84, loss_mean_cls=0.0793, proj_loss=-0.41][2026-03-26 12:46:12] Step: 3648, Training Logs: loss_final: 0.497773, loss_mean: 0.812910, proj_loss: -0.397109, loss_mean_cls: 0.081972, grad_norm: 0.404485 +Steps: 0%| | 3649/1000000 [15:09<68:27:16, 4.04it/s, grad_norm=0.404, loss_final=0.498, loss_mean=0.813, loss_mean_cls=0.082, proj_loss=-0.397][2026-03-26 12:46:13] Step: 3649, Training Logs: loss_final: 0.493291, loss_mean: 0.810283, proj_loss: -0.401587, loss_mean_cls: 0.084595, grad_norm: 0.517919 +Steps: 0%| | 3650/1000000 [15:09<68:28:05, 4.04it/s, grad_norm=0.518, loss_final=0.493, loss_mean=0.81, loss_mean_cls=0.0846, proj_loss=-0.402][2026-03-26 12:46:13] Step: 3650, Training Logs: loss_final: 0.505249, loss_mean: 0.840188, proj_loss: -0.413077, loss_mean_cls: 0.078138, grad_norm: 0.349303 +Steps: 0%| | 3651/1000000 [15:10<68:26:41, 4.04it/s, grad_norm=0.349, loss_final=0.505, loss_mean=0.84, loss_mean_cls=0.0781, proj_loss=-0.413][2026-03-26 12:46:13] Step: 3651, Training Logs: loss_final: 0.515416, loss_mean: 0.842802, proj_loss: -0.407683, loss_mean_cls: 0.080298, grad_norm: 0.375435 +Steps: 0%| | 3652/1000000 [15:10<68:26:55, 4.04it/s, grad_norm=0.375, loss_final=0.515, loss_mean=0.843, loss_mean_cls=0.0803, proj_loss=-0.408][2026-03-26 12:46:13] Step: 3652, Training Logs: loss_final: 0.518581, loss_mean: 0.852360, proj_loss: -0.414251, loss_mean_cls: 0.080472, grad_norm: 0.361793 +Steps: 0%| | 3653/1000000 [15:10<68:25:07, 4.05it/s, grad_norm=0.362, loss_final=0.519, loss_mean=0.852, loss_mean_cls=0.0805, proj_loss=-0.414][2026-03-26 12:46:13] Step: 3653, Training Logs: loss_final: 0.498738, loss_mean: 0.823169, proj_loss: -0.407809, loss_mean_cls: 0.083379, grad_norm: 0.321815 +Steps: 0%| | 3654/1000000 [15:10<68:26:08, 4.04it/s, grad_norm=0.322, loss_final=0.499, loss_mean=0.823, loss_mean_cls=0.0834, proj_loss=-0.408][2026-03-26 12:46:14] Step: 3654, Training Logs: loss_final: 0.515710, loss_mean: 0.836263, proj_loss: -0.403419, loss_mean_cls: 0.082866, grad_norm: 0.369792 +Steps: 0%| | 3655/1000000 [15:11<68:23:30, 4.05it/s, grad_norm=0.37, loss_final=0.516, loss_mean=0.836, loss_mean_cls=0.0829, proj_loss=-0.403][2026-03-26 12:46:14] Step: 3655, Training Logs: loss_final: 0.496367, loss_mean: 0.821567, proj_loss: -0.407314, loss_mean_cls: 0.082114, grad_norm: 0.342390 +Steps: 0%| | 3656/1000000 [15:11<68:25:07, 4.05it/s, grad_norm=0.342, loss_final=0.496, loss_mean=0.822, loss_mean_cls=0.0821, proj_loss=-0.407][2026-03-26 12:46:14] Step: 3656, Training Logs: loss_final: 0.513864, loss_mean: 0.842950, proj_loss: -0.408476, loss_mean_cls: 0.079391, grad_norm: 0.279485 +Steps: 0%| | 3657/1000000 [15:11<68:32:04, 4.04it/s, grad_norm=0.279, loss_final=0.514, loss_mean=0.843, loss_mean_cls=0.0794, proj_loss=-0.408][2026-03-26 12:46:14] Step: 3657, Training Logs: loss_final: 0.515321, loss_mean: 0.856624, proj_loss: -0.418899, loss_mean_cls: 0.077596, grad_norm: 0.367865 +Steps: 0%| | 3658/1000000 [15:11<68:29:24, 4.04it/s, grad_norm=0.368, loss_final=0.515, loss_mean=0.857, loss_mean_cls=0.0776, proj_loss=-0.419][2026-03-26 12:46:15] Step: 3658, Training Logs: loss_final: 0.496990, loss_mean: 0.819974, proj_loss: -0.404594, loss_mean_cls: 0.081609, grad_norm: 0.407800 +Steps: 0%| | 3659/1000000 [15:12<68:27:28, 4.04it/s, grad_norm=0.408, loss_final=0.497, loss_mean=0.82, loss_mean_cls=0.0816, proj_loss=-0.405][2026-03-26 12:46:15] Step: 3659, Training Logs: loss_final: 0.520025, loss_mean: 0.863755, proj_loss: -0.421201, loss_mean_cls: 0.077471, grad_norm: 0.340981 +Steps: 0%| | 3660/1000000 [15:12<68:26:14, 4.04it/s, grad_norm=0.341, loss_final=0.52, loss_mean=0.864, loss_mean_cls=0.0775, proj_loss=-0.421][2026-03-26 12:46:15] Step: 3660, Training Logs: loss_final: 0.481756, loss_mean: 0.810186, proj_loss: -0.408017, loss_mean_cls: 0.079587, grad_norm: 0.433009 +Steps: 0%| | 3661/1000000 [15:12<68:26:34, 4.04it/s, grad_norm=0.433, loss_final=0.482, loss_mean=0.81, loss_mean_cls=0.0796, proj_loss=-0.408][2026-03-26 12:46:15] Step: 3661, Training Logs: loss_final: 0.524835, loss_mean: 0.847893, proj_loss: -0.404576, loss_mean_cls: 0.081519, grad_norm: 0.572744 +Steps: 0%| | 3662/1000000 [15:12<68:30:26, 4.04it/s, grad_norm=0.573, loss_final=0.525, loss_mean=0.848, loss_mean_cls=0.0815, proj_loss=-0.405][2026-03-26 12:46:16] Step: 3662, Training Logs: loss_final: 0.516929, loss_mean: 0.839828, proj_loss: -0.406424, loss_mean_cls: 0.083525, grad_norm: 0.466473 +Steps: 0%| | 3663/1000000 [15:13<68:28:50, 4.04it/s, grad_norm=0.466, loss_final=0.517, loss_mean=0.84, loss_mean_cls=0.0835, proj_loss=-0.406][2026-03-26 12:46:16] Step: 3663, Training Logs: loss_final: 0.516145, loss_mean: 0.851068, proj_loss: -0.414218, loss_mean_cls: 0.079295, grad_norm: 0.505654 +Steps: 0%| | 3664/1000000 [15:13<68:28:57, 4.04it/s, grad_norm=0.506, loss_final=0.516, loss_mean=0.851, loss_mean_cls=0.0793, proj_loss=-0.414][2026-03-26 12:46:16] Step: 3664, Training Logs: loss_final: 0.514428, loss_mean: 0.831470, proj_loss: -0.401323, loss_mean_cls: 0.084282, grad_norm: 0.450356 +Steps: 0%| | 3665/1000000 [15:13<68:30:13, 4.04it/s, grad_norm=0.45, loss_final=0.514, loss_mean=0.831, loss_mean_cls=0.0843, proj_loss=-0.401][2026-03-26 12:46:16] Step: 3665, Training Logs: loss_final: 0.506290, loss_mean: 0.847358, proj_loss: -0.418719, loss_mean_cls: 0.077650, grad_norm: 0.427560 +Steps: 0%| | 3666/1000000 [15:13<68:31:21, 4.04it/s, grad_norm=0.428, loss_final=0.506, loss_mean=0.847, loss_mean_cls=0.0777, proj_loss=-0.419][2026-03-26 12:46:17] Step: 3666, Training Logs: loss_final: 0.520507, loss_mean: 0.856064, proj_loss: -0.414429, loss_mean_cls: 0.078872, grad_norm: 0.428214 +Steps: 0%| | 3667/1000000 [15:14<68:29:10, 4.04it/s, grad_norm=0.428, loss_final=0.521, loss_mean=0.856, loss_mean_cls=0.0789, proj_loss=-0.414][2026-03-26 12:46:17] Step: 3667, Training Logs: loss_final: 0.490444, loss_mean: 0.801100, proj_loss: -0.397223, loss_mean_cls: 0.086567, grad_norm: 0.312623 +Steps: 0%| | 3668/1000000 [15:14<68:27:38, 4.04it/s, grad_norm=0.313, loss_final=0.49, loss_mean=0.801, loss_mean_cls=0.0866, proj_loss=-0.397][2026-03-26 12:46:17] Step: 3668, Training Logs: loss_final: 0.520766, loss_mean: 0.826318, proj_loss: -0.390094, loss_mean_cls: 0.084542, grad_norm: 0.521776 +Steps: 0%| | 3669/1000000 [15:14<68:27:06, 4.04it/s, grad_norm=0.522, loss_final=0.521, loss_mean=0.826, loss_mean_cls=0.0845, proj_loss=-0.39][2026-03-26 12:46:17] Step: 3669, Training Logs: loss_final: 0.521476, loss_mean: 0.839975, proj_loss: -0.400453, loss_mean_cls: 0.081954, grad_norm: 0.415819 +Steps: 0%| | 3670/1000000 [15:14<68:28:58, 4.04it/s, grad_norm=0.416, loss_final=0.521, loss_mean=0.84, loss_mean_cls=0.082, proj_loss=-0.4][2026-03-26 12:46:18] Step: 3670, Training Logs: loss_final: 0.509445, loss_mean: 0.826983, proj_loss: -0.400017, loss_mean_cls: 0.082479, grad_norm: 0.395535 +Steps: 0%| | 3671/1000000 [15:15<68:27:53, 4.04it/s, grad_norm=0.396, loss_final=0.509, loss_mean=0.827, loss_mean_cls=0.0825, proj_loss=-0.4][2026-03-26 12:46:18] Step: 3671, Training Logs: loss_final: 0.498261, loss_mean: 0.835891, proj_loss: -0.416829, loss_mean_cls: 0.079199, grad_norm: 0.290920 +Steps: 0%| | 3672/1000000 [15:15<68:26:40, 4.04it/s, grad_norm=0.291, loss_final=0.498, loss_mean=0.836, loss_mean_cls=0.0792, proj_loss=-0.417][2026-03-26 12:46:18] Step: 3672, Training Logs: loss_final: 0.526369, loss_mean: 0.848074, proj_loss: -0.401951, loss_mean_cls: 0.080246, grad_norm: 0.377447 +Steps: 0%| | 3673/1000000 [15:15<68:27:44, 4.04it/s, grad_norm=0.377, loss_final=0.526, loss_mean=0.848, loss_mean_cls=0.0802, proj_loss=-0.402][2026-03-26 12:46:18] Step: 3673, Training Logs: loss_final: 0.504162, loss_mean: 0.831229, proj_loss: -0.406568, loss_mean_cls: 0.079501, grad_norm: 0.370499 +Steps: 0%| | 3674/1000000 [15:15<68:27:59, 4.04it/s, grad_norm=0.37, loss_final=0.504, loss_mean=0.831, loss_mean_cls=0.0795, proj_loss=-0.407][2026-03-26 12:46:19] Step: 3674, Training Logs: loss_final: 0.514451, loss_mean: 0.826804, proj_loss: -0.396210, loss_mean_cls: 0.083857, grad_norm: 0.392346 +Steps: 0%| | 3675/1000000 [15:16<68:28:52, 4.04it/s, grad_norm=0.392, loss_final=0.514, loss_mean=0.827, loss_mean_cls=0.0839, proj_loss=-0.396][2026-03-26 12:46:19] Step: 3675, Training Logs: loss_final: 0.512457, loss_mean: 0.848530, proj_loss: -0.414588, loss_mean_cls: 0.078514, grad_norm: 0.479160 +Steps: 0%| | 3676/1000000 [15:16<68:27:04, 4.04it/s, grad_norm=0.479, loss_final=0.512, loss_mean=0.849, loss_mean_cls=0.0785, proj_loss=-0.415][2026-03-26 12:46:19] Step: 3676, Training Logs: loss_final: 0.516316, loss_mean: 0.846082, proj_loss: -0.410617, loss_mean_cls: 0.080851, grad_norm: 0.490136 +Steps: 0%| | 3677/1000000 [15:16<68:26:31, 4.04it/s, grad_norm=0.49, loss_final=0.516, loss_mean=0.846, loss_mean_cls=0.0809, proj_loss=-0.411][2026-03-26 12:46:19] Step: 3677, Training Logs: loss_final: 0.504920, loss_mean: 0.836001, proj_loss: -0.409743, loss_mean_cls: 0.078663, grad_norm: 0.446161 +Steps: 0%| | 3678/1000000 [15:16<68:25:24, 4.04it/s, grad_norm=0.446, loss_final=0.505, loss_mean=0.836, loss_mean_cls=0.0787, proj_loss=-0.41][2026-03-26 12:46:20] Step: 3678, Training Logs: loss_final: 0.512254, loss_mean: 0.849448, proj_loss: -0.415257, loss_mean_cls: 0.078063, grad_norm: 0.526381 +Steps: 0%| | 3679/1000000 [15:17<68:29:12, 4.04it/s, grad_norm=0.526, loss_final=0.512, loss_mean=0.849, loss_mean_cls=0.0781, proj_loss=-0.415][2026-03-26 12:46:20] Step: 3679, Training Logs: loss_final: 0.511112, loss_mean: 0.830837, proj_loss: -0.401404, loss_mean_cls: 0.081679, grad_norm: 0.331136 +Steps: 0%| | 3680/1000000 [15:17<68:29:10, 4.04it/s, grad_norm=0.331, loss_final=0.511, loss_mean=0.831, loss_mean_cls=0.0817, proj_loss=-0.401][2026-03-26 12:46:20] Step: 3680, Training Logs: loss_final: 0.499331, loss_mean: 0.820276, proj_loss: -0.403861, loss_mean_cls: 0.082916, grad_norm: 0.575475 +Steps: 0%| | 3681/1000000 [15:17<68:29:37, 4.04it/s, grad_norm=0.575, loss_final=0.499, loss_mean=0.82, loss_mean_cls=0.0829, proj_loss=-0.404][2026-03-26 12:46:20] Step: 3681, Training Logs: loss_final: 0.512160, loss_mean: 0.833132, proj_loss: -0.402448, loss_mean_cls: 0.081476, grad_norm: 0.481361 +Steps: 0%| | 3682/1000000 [15:17<68:25:38, 4.04it/s, grad_norm=0.481, loss_final=0.512, loss_mean=0.833, loss_mean_cls=0.0815, proj_loss=-0.402][2026-03-26 12:46:21] Step: 3682, Training Logs: loss_final: 0.509443, loss_mean: 0.820669, proj_loss: -0.396666, loss_mean_cls: 0.085440, grad_norm: 0.540426 +Steps: 0%| | 3683/1000000 [15:18<68:23:16, 4.05it/s, grad_norm=0.54, loss_final=0.509, loss_mean=0.821, loss_mean_cls=0.0854, proj_loss=-0.397][2026-03-26 12:46:21] Step: 3683, Training Logs: loss_final: 0.499233, loss_mean: 0.824947, proj_loss: -0.408174, loss_mean_cls: 0.082460, grad_norm: 0.365110 +Steps: 0%| | 3684/1000000 [15:18<68:24:35, 4.05it/s, grad_norm=0.365, loss_final=0.499, loss_mean=0.825, loss_mean_cls=0.0825, proj_loss=-0.408][2026-03-26 12:46:21] Step: 3684, Training Logs: loss_final: 0.516437, loss_mean: 0.837068, proj_loss: -0.402717, loss_mean_cls: 0.082085, grad_norm: 0.829270 +Steps: 0%| | 3685/1000000 [15:18<68:24:37, 4.05it/s, grad_norm=0.829, loss_final=0.516, loss_mean=0.837, loss_mean_cls=0.0821, proj_loss=-0.403][2026-03-26 12:46:21] Step: 3685, Training Logs: loss_final: 0.506563, loss_mean: 0.839581, proj_loss: -0.413483, loss_mean_cls: 0.080465, grad_norm: 0.576264 +Steps: 0%| | 3686/1000000 [15:18<68:27:16, 4.04it/s, grad_norm=0.576, loss_final=0.507, loss_mean=0.84, loss_mean_cls=0.0805, proj_loss=-0.413][2026-03-26 12:46:22] Step: 3686, Training Logs: loss_final: 0.517863, loss_mean: 0.849077, proj_loss: -0.410370, loss_mean_cls: 0.079157, grad_norm: 0.460474 +Steps: 0%| | 3687/1000000 [15:19<68:27:46, 4.04it/s, grad_norm=0.46, loss_final=0.518, loss_mean=0.849, loss_mean_cls=0.0792, proj_loss=-0.41][2026-03-26 12:46:22] Step: 3687, Training Logs: loss_final: 0.507903, loss_mean: 0.836389, proj_loss: -0.407988, loss_mean_cls: 0.079502, grad_norm: 0.725172 +Steps: 0%| | 3688/1000000 [15:19<68:30:43, 4.04it/s, grad_norm=0.725, loss_final=0.508, loss_mean=0.836, loss_mean_cls=0.0795, proj_loss=-0.408][2026-03-26 12:46:22] Step: 3688, Training Logs: loss_final: 0.514005, loss_mean: 0.834114, proj_loss: -0.401999, loss_mean_cls: 0.081890, grad_norm: 0.405191 +Steps: 0%| | 3689/1000000 [15:19<68:26:43, 4.04it/s, grad_norm=0.405, loss_final=0.514, loss_mean=0.834, loss_mean_cls=0.0819, proj_loss=-0.402][2026-03-26 12:46:22] Step: 3689, Training Logs: loss_final: 0.503801, loss_mean: 0.819329, proj_loss: -0.399327, loss_mean_cls: 0.083800, grad_norm: 0.537071 +Steps: 0%| | 3690/1000000 [15:20<97:28:25, 2.84it/s, grad_norm=0.537, loss_final=0.504, loss_mean=0.819, loss_mean_cls=0.0838, proj_loss=-0.399][2026-03-26 12:46:23] Step: 3690, Training Logs: loss_final: 0.503142, loss_mean: 0.824180, proj_loss: -0.403107, loss_mean_cls: 0.082068, grad_norm: 0.631816 +Steps: 0%| | 3691/1000000 [15:20<88:45:22, 3.12it/s, grad_norm=0.632, loss_final=0.503, loss_mean=0.824, loss_mean_cls=0.0821, proj_loss=-0.403][2026-03-26 12:46:23] Step: 3691, Training Logs: loss_final: 0.508353, loss_mean: 0.839503, proj_loss: -0.409999, loss_mean_cls: 0.078849, grad_norm: 0.397450 +Steps: 0%| | 3692/1000000 [15:20<82:38:20, 3.35it/s, grad_norm=0.397, loss_final=0.508, loss_mean=0.84, loss_mean_cls=0.0788, proj_loss=-0.41][2026-03-26 12:46:23] Step: 3692, Training Logs: loss_final: 0.497252, loss_mean: 0.825662, proj_loss: -0.408728, loss_mean_cls: 0.080318, grad_norm: 0.300160 +Steps: 0%| | 3693/1000000 [15:20<78:23:19, 3.53it/s, grad_norm=0.3, loss_final=0.497, loss_mean=0.826, loss_mean_cls=0.0803, proj_loss=-0.409][2026-03-26 12:46:24] Step: 3693, Training Logs: loss_final: 0.516822, loss_mean: 0.850870, proj_loss: -0.413671, loss_mean_cls: 0.079624, grad_norm: 0.473265 +Steps: 0%| | 3694/1000000 [15:21<75:22:16, 3.67it/s, grad_norm=0.473, loss_final=0.517, loss_mean=0.851, loss_mean_cls=0.0796, proj_loss=-0.414][2026-03-26 12:46:24] Step: 3694, Training Logs: loss_final: 0.518587, loss_mean: 0.842851, proj_loss: -0.405321, loss_mean_cls: 0.081057, grad_norm: 0.333054 +Steps: 0%| | 3695/1000000 [15:21<73:19:17, 3.77it/s, grad_norm=0.333, loss_final=0.519, loss_mean=0.843, loss_mean_cls=0.0811, proj_loss=-0.405][2026-03-26 12:46:24] Step: 3695, Training Logs: loss_final: 0.498049, loss_mean: 0.815719, proj_loss: -0.402447, loss_mean_cls: 0.084777, grad_norm: 0.299502 +Steps: 0%| | 3696/1000000 [15:21<71:49:50, 3.85it/s, grad_norm=0.3, loss_final=0.498, loss_mean=0.816, loss_mean_cls=0.0848, proj_loss=-0.402][2026-03-26 12:46:24] Step: 3696, Training Logs: loss_final: 0.509392, loss_mean: 0.832623, proj_loss: -0.404548, loss_mean_cls: 0.081317, grad_norm: 0.402490 +Steps: 0%| | 3697/1000000 [15:21<70:50:54, 3.91it/s, grad_norm=0.402, loss_final=0.509, loss_mean=0.833, loss_mean_cls=0.0813, proj_loss=-0.405][2026-03-26 12:46:25] Step: 3697, Training Logs: loss_final: 0.498176, loss_mean: 0.832613, proj_loss: -0.413843, loss_mean_cls: 0.079406, grad_norm: 0.370299 +Steps: 0%| | 3698/1000000 [15:22<70:03:35, 3.95it/s, grad_norm=0.37, loss_final=0.498, loss_mean=0.833, loss_mean_cls=0.0794, proj_loss=-0.414][2026-03-26 12:46:25] Step: 3698, Training Logs: loss_final: 0.512687, loss_mean: 0.842861, proj_loss: -0.409491, loss_mean_cls: 0.079317, grad_norm: 0.377211 +Steps: 0%| | 3699/1000000 [15:22<69:33:30, 3.98it/s, grad_norm=0.377, loss_final=0.513, loss_mean=0.843, loss_mean_cls=0.0793, proj_loss=-0.409][2026-03-26 12:46:25] Step: 3699, Training Logs: loss_final: 0.514566, loss_mean: 0.832241, proj_loss: -0.402095, loss_mean_cls: 0.084420, grad_norm: 0.347758 +Steps: 0%| | 3700/1000000 [15:22<69:12:03, 4.00it/s, grad_norm=0.348, loss_final=0.515, loss_mean=0.832, loss_mean_cls=0.0844, proj_loss=-0.402][2026-03-26 12:46:25] Step: 3700, Training Logs: loss_final: 0.501979, loss_mean: 0.837671, proj_loss: -0.416639, loss_mean_cls: 0.080948, grad_norm: 0.376840 +Steps: 0%| | 3701/1000000 [15:22<68:57:31, 4.01it/s, grad_norm=0.377, loss_final=0.502, loss_mean=0.838, loss_mean_cls=0.0809, proj_loss=-0.417][2026-03-26 12:46:26] Step: 3701, Training Logs: loss_final: 0.513980, loss_mean: 0.848575, proj_loss: -0.414458, loss_mean_cls: 0.079864, grad_norm: 0.456526 +Steps: 0%| | 3702/1000000 [15:23<68:48:17, 4.02it/s, grad_norm=0.457, loss_final=0.514, loss_mean=0.849, loss_mean_cls=0.0799, proj_loss=-0.414][2026-03-26 12:46:26] Step: 3702, Training Logs: loss_final: 0.493191, loss_mean: 0.826387, proj_loss: -0.412990, loss_mean_cls: 0.079794, grad_norm: 0.402338 +Steps: 0%| | 3703/1000000 [15:23<68:43:41, 4.03it/s, grad_norm=0.402, loss_final=0.493, loss_mean=0.826, loss_mean_cls=0.0798, proj_loss=-0.413][2026-03-26 12:46:26] Step: 3703, Training Logs: loss_final: 0.495409, loss_mean: 0.821944, proj_loss: -0.408153, loss_mean_cls: 0.081617, grad_norm: 0.297124 +Steps: 0%| | 3704/1000000 [15:23<68:38:30, 4.03it/s, grad_norm=0.297, loss_final=0.495, loss_mean=0.822, loss_mean_cls=0.0816, proj_loss=-0.408][2026-03-26 12:46:26] Step: 3704, Training Logs: loss_final: 0.512872, loss_mean: 0.839567, proj_loss: -0.408706, loss_mean_cls: 0.082011, grad_norm: 0.372543 +Steps: 0%| | 3705/1000000 [15:23<68:36:18, 4.03it/s, grad_norm=0.373, loss_final=0.513, loss_mean=0.84, loss_mean_cls=0.082, proj_loss=-0.409][2026-03-26 12:46:27] Step: 3705, Training Logs: loss_final: 0.508263, loss_mean: 0.843527, proj_loss: -0.413917, loss_mean_cls: 0.078653, grad_norm: 0.401499 +Steps: 0%| | 3706/1000000 [15:24<68:34:44, 4.04it/s, grad_norm=0.401, loss_final=0.508, loss_mean=0.844, loss_mean_cls=0.0787, proj_loss=-0.414][2026-03-26 12:46:27] Step: 3706, Training Logs: loss_final: 0.507050, loss_mean: 0.845083, proj_loss: -0.416335, loss_mean_cls: 0.078301, grad_norm: 0.556517 +Steps: 0%| | 3707/1000000 [15:24<68:33:46, 4.04it/s, grad_norm=0.557, loss_final=0.507, loss_mean=0.845, loss_mean_cls=0.0783, proj_loss=-0.416][2026-03-26 12:46:27] Step: 3707, Training Logs: loss_final: 0.516537, loss_mean: 0.849051, proj_loss: -0.411255, loss_mean_cls: 0.078740, grad_norm: 0.389617 +Steps: 0%| | 3708/1000000 [15:24<68:30:06, 4.04it/s, grad_norm=0.39, loss_final=0.517, loss_mean=0.849, loss_mean_cls=0.0787, proj_loss=-0.411][2026-03-26 12:46:27] Step: 3708, Training Logs: loss_final: 0.515252, loss_mean: 0.844053, proj_loss: -0.409016, loss_mean_cls: 0.080216, grad_norm: 0.745104 +Steps: 0%| | 3709/1000000 [15:24<68:53:40, 4.02it/s, grad_norm=0.745, loss_final=0.515, loss_mean=0.844, loss_mean_cls=0.0802, proj_loss=-0.409][2026-03-26 12:46:28] Step: 3709, Training Logs: loss_final: 0.523318, loss_mean: 0.846046, proj_loss: -0.405364, loss_mean_cls: 0.082636, grad_norm: 0.537929 +Steps: 0%| | 3710/1000000 [15:25<68:30:50, 4.04it/s, grad_norm=0.538, loss_final=0.523, loss_mean=0.846, loss_mean_cls=0.0826, proj_loss=-0.405][2026-03-26 12:46:28] Step: 3710, Training Logs: loss_final: 0.511749, loss_mean: 0.831085, proj_loss: -0.402761, loss_mean_cls: 0.083425, grad_norm: 0.436595 +Steps: 0%| | 3711/1000000 [15:25<68:31:15, 4.04it/s, grad_norm=0.437, loss_final=0.512, loss_mean=0.831, loss_mean_cls=0.0834, proj_loss=-0.403][2026-03-26 12:46:28] Step: 3711, Training Logs: loss_final: 0.517764, loss_mean: 0.852040, proj_loss: -0.412901, loss_mean_cls: 0.078625, grad_norm: 0.665687 +Steps: 0%| | 3712/1000000 [15:25<68:31:49, 4.04it/s, grad_norm=0.666, loss_final=0.518, loss_mean=0.852, loss_mean_cls=0.0786, proj_loss=-0.413][2026-03-26 12:46:28] Step: 3712, Training Logs: loss_final: 0.515552, loss_mean: 0.850618, proj_loss: -0.413891, loss_mean_cls: 0.078825, grad_norm: 0.490885 +Steps: 0%| | 3713/1000000 [15:25<68:28:25, 4.04it/s, grad_norm=0.491, loss_final=0.516, loss_mean=0.851, loss_mean_cls=0.0788, proj_loss=-0.414][2026-03-26 12:46:29] Step: 3713, Training Logs: loss_final: 0.517387, loss_mean: 0.850957, proj_loss: -0.412300, loss_mean_cls: 0.078731, grad_norm: 0.386050 +Steps: 0%| | 3714/1000000 [15:26<68:28:54, 4.04it/s, grad_norm=0.386, loss_final=0.517, loss_mean=0.851, loss_mean_cls=0.0787, proj_loss=-0.412][2026-03-26 12:46:29] Step: 3714, Training Logs: loss_final: 0.497873, loss_mean: 0.816468, proj_loss: -0.401859, loss_mean_cls: 0.083264, grad_norm: 0.381562 +Steps: 0%| | 3715/1000000 [15:26<69:11:58, 4.00it/s, grad_norm=0.382, loss_final=0.498, loss_mean=0.816, loss_mean_cls=0.0833, proj_loss=-0.402][2026-03-26 12:46:29] Step: 3715, Training Logs: loss_final: 0.520193, loss_mean: 0.845618, proj_loss: -0.406657, loss_mean_cls: 0.081232, grad_norm: 0.603565 +Steps: 0%| | 3716/1000000 [15:26<68:59:07, 4.01it/s, grad_norm=0.604, loss_final=0.52, loss_mean=0.846, loss_mean_cls=0.0812, proj_loss=-0.407][2026-03-26 12:46:29] Step: 3716, Training Logs: loss_final: 0.513803, loss_mean: 0.833971, proj_loss: -0.401956, loss_mean_cls: 0.081788, grad_norm: 0.383745 +Steps: 0%| | 3717/1000000 [15:26<68:46:10, 4.02it/s, grad_norm=0.384, loss_final=0.514, loss_mean=0.834, loss_mean_cls=0.0818, proj_loss=-0.402][2026-03-26 12:46:30] Step: 3717, Training Logs: loss_final: 0.510625, loss_mean: 0.834068, proj_loss: -0.405733, loss_mean_cls: 0.082291, grad_norm: 0.436289 +Steps: 0%| | 3718/1000000 [15:27<68:43:44, 4.03it/s, grad_norm=0.436, loss_final=0.511, loss_mean=0.834, loss_mean_cls=0.0823, proj_loss=-0.406][2026-03-26 12:46:30] Step: 3718, Training Logs: loss_final: 0.498388, loss_mean: 0.825137, proj_loss: -0.409287, loss_mean_cls: 0.082538, grad_norm: 0.373674 +Steps: 0%| | 3719/1000000 [15:27<68:36:37, 4.03it/s, grad_norm=0.374, loss_final=0.498, loss_mean=0.825, loss_mean_cls=0.0825, proj_loss=-0.409][2026-03-26 12:46:30] Step: 3719, Training Logs: loss_final: 0.515294, loss_mean: 0.856400, proj_loss: -0.418990, loss_mean_cls: 0.077885, grad_norm: 0.436565 +Steps: 0%| | 3720/1000000 [15:27<68:35:39, 4.03it/s, grad_norm=0.437, loss_final=0.515, loss_mean=0.856, loss_mean_cls=0.0779, proj_loss=-0.419][2026-03-26 12:46:30] Step: 3720, Training Logs: loss_final: 0.501786, loss_mean: 0.819665, proj_loss: -0.402815, loss_mean_cls: 0.084936, grad_norm: 0.439109 +Steps: 0%| | 3721/1000000 [15:27<68:32:10, 4.04it/s, grad_norm=0.439, loss_final=0.502, loss_mean=0.82, loss_mean_cls=0.0849, proj_loss=-0.403][2026-03-26 12:46:31] Step: 3721, Training Logs: loss_final: 0.490494, loss_mean: 0.808234, proj_loss: -0.401931, loss_mean_cls: 0.084191, grad_norm: 0.331775 +Steps: 0%| | 3722/1000000 [15:28<68:32:21, 4.04it/s, grad_norm=0.332, loss_final=0.49, loss_mean=0.808, loss_mean_cls=0.0842, proj_loss=-0.402][2026-03-26 12:46:31] Step: 3722, Training Logs: loss_final: 0.512018, loss_mean: 0.841222, proj_loss: -0.409761, loss_mean_cls: 0.080557, grad_norm: 0.499545 +Steps: 0%| | 3723/1000000 [15:28<68:27:28, 4.04it/s, grad_norm=0.5, loss_final=0.512, loss_mean=0.841, loss_mean_cls=0.0806, proj_loss=-0.41][2026-03-26 12:46:31] Step: 3723, Training Logs: loss_final: 0.510285, loss_mean: 0.840260, proj_loss: -0.409921, loss_mean_cls: 0.079946, grad_norm: 0.372344 +Steps: 0%| | 3724/1000000 [15:28<68:27:37, 4.04it/s, grad_norm=0.372, loss_final=0.51, loss_mean=0.84, loss_mean_cls=0.0799, proj_loss=-0.41][2026-03-26 12:46:31] Step: 3724, Training Logs: loss_final: 0.507946, loss_mean: 0.844912, proj_loss: -0.417053, loss_mean_cls: 0.080087, grad_norm: 0.353838 +Steps: 0%| | 3725/1000000 [15:28<68:27:08, 4.04it/s, grad_norm=0.354, loss_final=0.508, loss_mean=0.845, loss_mean_cls=0.0801, proj_loss=-0.417][2026-03-26 12:46:32] Step: 3725, Training Logs: loss_final: 0.494761, loss_mean: 0.819170, proj_loss: -0.406093, loss_mean_cls: 0.081684, grad_norm: 0.629479 +Steps: 0%| | 3726/1000000 [15:29<68:27:03, 4.04it/s, grad_norm=0.629, loss_final=0.495, loss_mean=0.819, loss_mean_cls=0.0817, proj_loss=-0.406][2026-03-26 12:46:32] Step: 3726, Training Logs: loss_final: 0.508995, loss_mean: 0.835787, proj_loss: -0.407560, loss_mean_cls: 0.080768, grad_norm: 0.391582 +Steps: 0%| | 3727/1000000 [15:29<68:26:29, 4.04it/s, grad_norm=0.392, loss_final=0.509, loss_mean=0.836, loss_mean_cls=0.0808, proj_loss=-0.408][2026-03-26 12:46:32] Step: 3727, Training Logs: loss_final: 0.517942, loss_mean: 0.837270, proj_loss: -0.402681, loss_mean_cls: 0.083353, grad_norm: 0.462560 +Steps: 0%| | 3728/1000000 [15:29<68:28:32, 4.04it/s, grad_norm=0.463, loss_final=0.518, loss_mean=0.837, loss_mean_cls=0.0834, proj_loss=-0.403][2026-03-26 12:46:32] Step: 3728, Training Logs: loss_final: 0.504470, loss_mean: 0.830004, proj_loss: -0.406116, loss_mean_cls: 0.080583, grad_norm: 0.572378 +Steps: 0%| | 3729/1000000 [15:29<68:25:21, 4.04it/s, grad_norm=0.572, loss_final=0.504, loss_mean=0.83, loss_mean_cls=0.0806, proj_loss=-0.406][2026-03-26 12:46:33] Step: 3729, Training Logs: loss_final: 0.499472, loss_mean: 0.833981, proj_loss: -0.415199, loss_mean_cls: 0.080690, grad_norm: 0.297996 +Steps: 0%| | 3730/1000000 [15:30<68:26:12, 4.04it/s, grad_norm=0.298, loss_final=0.499, loss_mean=0.834, loss_mean_cls=0.0807, proj_loss=-0.415][2026-03-26 12:46:33] Step: 3730, Training Logs: loss_final: 0.520234, loss_mean: 0.847072, proj_loss: -0.405494, loss_mean_cls: 0.078656, grad_norm: 0.560183 +Steps: 0%| | 3731/1000000 [15:30<68:26:19, 4.04it/s, grad_norm=0.56, loss_final=0.52, loss_mean=0.847, loss_mean_cls=0.0787, proj_loss=-0.405][2026-03-26 12:46:33] Step: 3731, Training Logs: loss_final: 0.515998, loss_mean: 0.848838, proj_loss: -0.412410, loss_mean_cls: 0.079569, grad_norm: 0.482248 +Steps: 0%| | 3732/1000000 [15:30<68:26:54, 4.04it/s, grad_norm=0.482, loss_final=0.516, loss_mean=0.849, loss_mean_cls=0.0796, proj_loss=-0.412][2026-03-26 12:46:33] Step: 3732, Training Logs: loss_final: 0.516245, loss_mean: 0.850410, proj_loss: -0.413158, loss_mean_cls: 0.078994, grad_norm: 0.326452 +Steps: 0%| | 3733/1000000 [15:30<68:29:28, 4.04it/s, grad_norm=0.326, loss_final=0.516, loss_mean=0.85, loss_mean_cls=0.079, proj_loss=-0.413][2026-03-26 12:46:34] Step: 3733, Training Logs: loss_final: 0.521743, loss_mean: 0.851897, proj_loss: -0.411860, loss_mean_cls: 0.081706, grad_norm: 0.705732 +Steps: 0%| | 3734/1000000 [15:31<68:27:57, 4.04it/s, grad_norm=0.706, loss_final=0.522, loss_mean=0.852, loss_mean_cls=0.0817, proj_loss=-0.412][2026-03-26 12:46:34] Step: 3734, Training Logs: loss_final: 0.512632, loss_mean: 0.840670, proj_loss: -0.408582, loss_mean_cls: 0.080544, grad_norm: 0.377625 +Steps: 0%| | 3735/1000000 [15:31<68:27:55, 4.04it/s, grad_norm=0.378, loss_final=0.513, loss_mean=0.841, loss_mean_cls=0.0805, proj_loss=-0.409][2026-03-26 12:46:34] Step: 3735, Training Logs: loss_final: 0.506586, loss_mean: 0.838684, proj_loss: -0.412335, loss_mean_cls: 0.080238, grad_norm: 0.521202 +Steps: 0%| | 3736/1000000 [15:31<68:24:46, 4.05it/s, grad_norm=0.521, loss_final=0.507, loss_mean=0.839, loss_mean_cls=0.0802, proj_loss=-0.412][2026-03-26 12:46:34] Step: 3736, Training Logs: loss_final: 0.501559, loss_mean: 0.833959, proj_loss: -0.411121, loss_mean_cls: 0.078721, grad_norm: 0.502338 +Steps: 0%| | 3737/1000000 [15:31<68:26:04, 4.04it/s, grad_norm=0.502, loss_final=0.502, loss_mean=0.834, loss_mean_cls=0.0787, proj_loss=-0.411][2026-03-26 12:46:35] Step: 3737, Training Logs: loss_final: 0.492425, loss_mean: 0.827415, proj_loss: -0.414341, loss_mean_cls: 0.079350, grad_norm: 0.279690 +Steps: 0%| | 3738/1000000 [15:31<68:34:51, 4.04it/s, grad_norm=0.28, loss_final=0.492, loss_mean=0.827, loss_mean_cls=0.0794, proj_loss=-0.414][2026-03-26 12:46:35] Step: 3738, Training Logs: loss_final: 0.497676, loss_mean: 0.825570, proj_loss: -0.409276, loss_mean_cls: 0.081382, grad_norm: 0.664807 +Steps: 0%| | 3739/1000000 [15:32<68:29:57, 4.04it/s, grad_norm=0.665, loss_final=0.498, loss_mean=0.826, loss_mean_cls=0.0814, proj_loss=-0.409][2026-03-26 12:46:35] Step: 3739, Training Logs: loss_final: 0.517390, loss_mean: 0.851567, proj_loss: -0.413604, loss_mean_cls: 0.079427, grad_norm: 0.430988 +Steps: 0%| | 3740/1000000 [15:32<68:25:40, 4.04it/s, grad_norm=0.431, loss_final=0.517, loss_mean=0.852, loss_mean_cls=0.0794, proj_loss=-0.414][2026-03-26 12:46:35] Step: 3740, Training Logs: loss_final: 0.512219, loss_mean: 0.836869, proj_loss: -0.407401, loss_mean_cls: 0.082751, grad_norm: 0.483499 +Steps: 0%| | 3741/1000000 [15:32<68:31:14, 4.04it/s, grad_norm=0.483, loss_final=0.512, loss_mean=0.837, loss_mean_cls=0.0828, proj_loss=-0.407][2026-03-26 12:46:36] Step: 3741, Training Logs: loss_final: 0.500022, loss_mean: 0.827008, proj_loss: -0.406365, loss_mean_cls: 0.079379, grad_norm: 0.563740 +Steps: 0%| | 3742/1000000 [15:32<68:27:30, 4.04it/s, grad_norm=0.564, loss_final=0.5, loss_mean=0.827, loss_mean_cls=0.0794, proj_loss=-0.406][2026-03-26 12:46:36] Step: 3742, Training Logs: loss_final: 0.515998, loss_mean: 0.835865, proj_loss: -0.403014, loss_mean_cls: 0.083146, grad_norm: 0.314977 +Steps: 0%| | 3743/1000000 [15:33<68:28:34, 4.04it/s, grad_norm=0.315, loss_final=0.516, loss_mean=0.836, loss_mean_cls=0.0831, proj_loss=-0.403][2026-03-26 12:46:36] Step: 3743, Training Logs: loss_final: 0.512998, loss_mean: 0.843389, proj_loss: -0.410041, loss_mean_cls: 0.079650, grad_norm: 0.624110 +Steps: 0%| | 3744/1000000 [15:33<68:29:23, 4.04it/s, grad_norm=0.624, loss_final=0.513, loss_mean=0.843, loss_mean_cls=0.0796, proj_loss=-0.41][2026-03-26 12:46:36] Step: 3744, Training Logs: loss_final: 0.520071, loss_mean: 0.848800, proj_loss: -0.408839, loss_mean_cls: 0.080109, grad_norm: 0.578880 +Steps: 0%| | 3745/1000000 [15:33<68:32:27, 4.04it/s, grad_norm=0.579, loss_final=0.52, loss_mean=0.849, loss_mean_cls=0.0801, proj_loss=-0.409][2026-03-26 12:46:37] Step: 3745, Training Logs: loss_final: 0.516501, loss_mean: 0.848863, proj_loss: -0.411849, loss_mean_cls: 0.079487, grad_norm: 0.369311 +Steps: 0%| | 3746/1000000 [15:33<68:29:40, 4.04it/s, grad_norm=0.369, loss_final=0.517, loss_mean=0.849, loss_mean_cls=0.0795, proj_loss=-0.412][2026-03-26 12:46:37] Step: 3746, Training Logs: loss_final: 0.515365, loss_mean: 0.845039, proj_loss: -0.410434, loss_mean_cls: 0.080760, grad_norm: 0.717310 +Steps: 0%| | 3747/1000000 [15:34<68:30:04, 4.04it/s, grad_norm=0.717, loss_final=0.515, loss_mean=0.845, loss_mean_cls=0.0808, proj_loss=-0.41][2026-03-26 12:46:37] Step: 3747, Training Logs: loss_final: 0.520649, loss_mean: 0.849771, proj_loss: -0.410301, loss_mean_cls: 0.081178, grad_norm: 0.543471 +Steps: 0%| | 3748/1000000 [15:34<68:28:54, 4.04it/s, grad_norm=0.543, loss_final=0.521, loss_mean=0.85, loss_mean_cls=0.0812, proj_loss=-0.41][2026-03-26 12:46:37] Step: 3748, Training Logs: loss_final: 0.508111, loss_mean: 0.842366, proj_loss: -0.414206, loss_mean_cls: 0.079951, grad_norm: 0.323639 +Steps: 0%| | 3749/1000000 [15:34<68:28:52, 4.04it/s, grad_norm=0.324, loss_final=0.508, loss_mean=0.842, loss_mean_cls=0.08, proj_loss=-0.414][2026-03-26 12:46:38] Step: 3749, Training Logs: loss_final: 0.516664, loss_mean: 0.848035, proj_loss: -0.412926, loss_mean_cls: 0.081555, grad_norm: 0.497104 +Steps: 0%| | 3750/1000000 [15:34<68:31:10, 4.04it/s, grad_norm=0.497, loss_final=0.517, loss_mean=0.848, loss_mean_cls=0.0816, proj_loss=-0.413][2026-03-26 12:46:38] Step: 3750, Training Logs: loss_final: 0.509605, loss_mean: 0.837291, proj_loss: -0.408026, loss_mean_cls: 0.080340, grad_norm: 0.390610 +Steps: 0%| | 3751/1000000 [15:35<68:32:45, 4.04it/s, grad_norm=0.391, loss_final=0.51, loss_mean=0.837, loss_mean_cls=0.0803, proj_loss=-0.408][2026-03-26 12:46:38] Step: 3751, Training Logs: loss_final: 0.506764, loss_mean: 0.828218, proj_loss: -0.404393, loss_mean_cls: 0.082939, grad_norm: 0.364500 +Steps: 0%| | 3752/1000000 [15:35<68:32:51, 4.04it/s, grad_norm=0.365, loss_final=0.507, loss_mean=0.828, loss_mean_cls=0.0829, proj_loss=-0.404][2026-03-26 12:46:38] Step: 3752, Training Logs: loss_final: 0.507412, loss_mean: 0.831644, proj_loss: -0.406168, loss_mean_cls: 0.081936, grad_norm: 0.437573 +Steps: 0%| | 3753/1000000 [15:35<68:31:51, 4.04it/s, grad_norm=0.438, loss_final=0.507, loss_mean=0.832, loss_mean_cls=0.0819, proj_loss=-0.406][2026-03-26 12:46:39] Step: 3753, Training Logs: loss_final: 0.521895, loss_mean: 0.848260, proj_loss: -0.405590, loss_mean_cls: 0.079225, grad_norm: 0.488879 +Steps: 0%| | 3754/1000000 [15:35<68:31:38, 4.04it/s, grad_norm=0.489, loss_final=0.522, loss_mean=0.848, loss_mean_cls=0.0792, proj_loss=-0.406][2026-03-26 12:46:39] Step: 3754, Training Logs: loss_final: 0.504420, loss_mean: 0.820943, proj_loss: -0.401940, loss_mean_cls: 0.085417, grad_norm: 0.306043 +Steps: 0%| | 3755/1000000 [15:36<68:28:15, 4.04it/s, grad_norm=0.306, loss_final=0.504, loss_mean=0.821, loss_mean_cls=0.0854, proj_loss=-0.402][2026-03-26 12:46:39] Step: 3755, Training Logs: loss_final: 0.506739, loss_mean: 0.832224, proj_loss: -0.404969, loss_mean_cls: 0.079483, grad_norm: 0.647530 +Steps: 0%| | 3756/1000000 [15:36<70:00:37, 3.95it/s, grad_norm=0.648, loss_final=0.507, loss_mean=0.832, loss_mean_cls=0.0795, proj_loss=-0.405][2026-03-26 12:46:39] Step: 3756, Training Logs: loss_final: 0.497421, loss_mean: 0.821341, proj_loss: -0.406956, loss_mean_cls: 0.083036, grad_norm: 0.487239 +Steps: 0%| | 3757/1000000 [15:36<69:41:44, 3.97it/s, grad_norm=0.487, loss_final=0.497, loss_mean=0.821, loss_mean_cls=0.083, proj_loss=-0.407][2026-03-26 12:46:40] Step: 3757, Training Logs: loss_final: 0.519221, loss_mean: 0.837855, proj_loss: -0.402256, loss_mean_cls: 0.083621, grad_norm: 0.528007 +Steps: 0%| | 3758/1000000 [15:36<69:20:17, 3.99it/s, grad_norm=0.528, loss_final=0.519, loss_mean=0.838, loss_mean_cls=0.0836, proj_loss=-0.402][2026-03-26 12:46:40] Step: 3758, Training Logs: loss_final: 0.522807, loss_mean: 0.851825, proj_loss: -0.407385, loss_mean_cls: 0.078367, grad_norm: 0.637033 +Steps: 0%| | 3759/1000000 [15:37<69:08:01, 4.00it/s, grad_norm=0.637, loss_final=0.523, loss_mean=0.852, loss_mean_cls=0.0784, proj_loss=-0.407][2026-03-26 12:46:40] Step: 3759, Training Logs: loss_final: 0.498161, loss_mean: 0.835528, proj_loss: -0.416230, loss_mean_cls: 0.078864, grad_norm: 0.506827 +Steps: 0%| | 3760/1000000 [15:37<68:55:09, 4.02it/s, grad_norm=0.507, loss_final=0.498, loss_mean=0.836, loss_mean_cls=0.0789, proj_loss=-0.416][2026-03-26 12:46:40] Step: 3760, Training Logs: loss_final: 0.511333, loss_mean: 0.842412, proj_loss: -0.411474, loss_mean_cls: 0.080396, grad_norm: 0.306267 +Steps: 0%| | 3761/1000000 [15:37<68:44:30, 4.03it/s, grad_norm=0.306, loss_final=0.511, loss_mean=0.842, loss_mean_cls=0.0804, proj_loss=-0.411][2026-03-26 12:46:41] Step: 3761, Training Logs: loss_final: 0.513041, loss_mean: 0.851316, proj_loss: -0.415427, loss_mean_cls: 0.077152, grad_norm: 0.319578 +Steps: 0%| | 3762/1000000 [15:37<68:37:20, 4.03it/s, grad_norm=0.32, loss_final=0.513, loss_mean=0.851, loss_mean_cls=0.0772, proj_loss=-0.415][2026-03-26 12:46:41] Step: 3762, Training Logs: loss_final: 0.491579, loss_mean: 0.820871, proj_loss: -0.409620, loss_mean_cls: 0.080328, grad_norm: 0.261870 +Steps: 0%| | 3763/1000000 [15:38<68:31:28, 4.04it/s, grad_norm=0.262, loss_final=0.492, loss_mean=0.821, loss_mean_cls=0.0803, proj_loss=-0.41][2026-03-26 12:46:41] Step: 3763, Training Logs: loss_final: 0.509482, loss_mean: 0.842505, proj_loss: -0.414191, loss_mean_cls: 0.081168, grad_norm: 0.368771 +Steps: 0%| | 3764/1000000 [15:38<68:33:42, 4.04it/s, grad_norm=0.369, loss_final=0.509, loss_mean=0.843, loss_mean_cls=0.0812, proj_loss=-0.414][2026-03-26 12:46:41] Step: 3764, Training Logs: loss_final: 0.523775, loss_mean: 0.857212, proj_loss: -0.411873, loss_mean_cls: 0.078436, grad_norm: 0.366832 +Steps: 0%| | 3765/1000000 [15:38<68:30:11, 4.04it/s, grad_norm=0.367, loss_final=0.524, loss_mean=0.857, loss_mean_cls=0.0784, proj_loss=-0.412][2026-03-26 12:46:42] Step: 3765, Training Logs: loss_final: 0.492283, loss_mean: 0.830229, proj_loss: -0.416824, loss_mean_cls: 0.078878, grad_norm: 0.421404 +Steps: 0%| | 3766/1000000 [15:38<68:28:44, 4.04it/s, grad_norm=0.421, loss_final=0.492, loss_mean=0.83, loss_mean_cls=0.0789, proj_loss=-0.417][2026-03-26 12:46:42] Step: 3766, Training Logs: loss_final: 0.534670, loss_mean: 0.863214, proj_loss: -0.408830, loss_mean_cls: 0.080286, grad_norm: 0.476361 +Steps: 0%| | 3767/1000000 [15:39<68:27:24, 4.04it/s, grad_norm=0.476, loss_final=0.535, loss_mean=0.863, loss_mean_cls=0.0803, proj_loss=-0.409][2026-03-26 12:46:42] Step: 3767, Training Logs: loss_final: 0.518748, loss_mean: 0.846732, proj_loss: -0.409920, loss_mean_cls: 0.081935, grad_norm: 0.435199 +Steps: 0%| | 3768/1000000 [15:39<68:28:28, 4.04it/s, grad_norm=0.435, loss_final=0.519, loss_mean=0.847, loss_mean_cls=0.0819, proj_loss=-0.41][2026-03-26 12:46:42] Step: 3768, Training Logs: loss_final: 0.504831, loss_mean: 0.832799, proj_loss: -0.408956, loss_mean_cls: 0.080988, grad_norm: 0.430419 +Steps: 0%| | 3769/1000000 [15:39<68:25:58, 4.04it/s, grad_norm=0.43, loss_final=0.505, loss_mean=0.833, loss_mean_cls=0.081, proj_loss=-0.409][2026-03-26 12:46:43] Step: 3769, Training Logs: loss_final: 0.503978, loss_mean: 0.823403, proj_loss: -0.401262, loss_mean_cls: 0.081837, grad_norm: 0.474534 +Steps: 0%| | 3770/1000000 [15:39<68:27:21, 4.04it/s, grad_norm=0.475, loss_final=0.504, loss_mean=0.823, loss_mean_cls=0.0818, proj_loss=-0.401][2026-03-26 12:46:43] Step: 3770, Training Logs: loss_final: 0.519394, loss_mean: 0.847247, proj_loss: -0.409053, loss_mean_cls: 0.081199, grad_norm: 0.494089 +Steps: 0%| | 3771/1000000 [15:40<68:26:17, 4.04it/s, grad_norm=0.494, loss_final=0.519, loss_mean=0.847, loss_mean_cls=0.0812, proj_loss=-0.409][2026-03-26 12:46:43] Step: 3771, Training Logs: loss_final: 0.499472, loss_mean: 0.828308, proj_loss: -0.410199, loss_mean_cls: 0.081362, grad_norm: 0.435064 +Steps: 0%| | 3772/1000000 [15:40<68:25:05, 4.04it/s, grad_norm=0.435, loss_final=0.499, loss_mean=0.828, loss_mean_cls=0.0814, proj_loss=-0.41][2026-03-26 12:46:43] Step: 3772, Training Logs: loss_final: 0.502911, loss_mean: 0.834594, proj_loss: -0.412430, loss_mean_cls: 0.080747, grad_norm: 0.540700 +Steps: 0%| | 3773/1000000 [15:40<68:25:03, 4.04it/s, grad_norm=0.541, loss_final=0.503, loss_mean=0.835, loss_mean_cls=0.0807, proj_loss=-0.412][2026-03-26 12:46:44] Step: 3773, Training Logs: loss_final: 0.517091, loss_mean: 0.849716, proj_loss: -0.411469, loss_mean_cls: 0.078844, grad_norm: 0.825675 +Steps: 0%| | 3774/1000000 [15:40<68:23:54, 4.05it/s, grad_norm=0.826, loss_final=0.517, loss_mean=0.85, loss_mean_cls=0.0788, proj_loss=-0.411][2026-03-26 12:46:44] Step: 3774, Training Logs: loss_final: 0.503206, loss_mean: 0.831776, proj_loss: -0.409673, loss_mean_cls: 0.081103, grad_norm: 0.668460 +Steps: 0%| | 3775/1000000 [15:41<68:26:37, 4.04it/s, grad_norm=0.668, loss_final=0.503, loss_mean=0.832, loss_mean_cls=0.0811, proj_loss=-0.41][2026-03-26 12:46:44] Step: 3775, Training Logs: loss_final: 0.518816, loss_mean: 0.844752, proj_loss: -0.408347, loss_mean_cls: 0.082411, grad_norm: 0.658966 +Steps: 0%| | 3776/1000000 [15:41<68:23:49, 4.05it/s, grad_norm=0.659, loss_final=0.519, loss_mean=0.845, loss_mean_cls=0.0824, proj_loss=-0.408][2026-03-26 12:46:44] Step: 3776, Training Logs: loss_final: 0.513617, loss_mean: 0.848764, proj_loss: -0.415499, loss_mean_cls: 0.080352, grad_norm: 0.942432 +Steps: 0%| | 3777/1000000 [15:41<68:23:51, 4.05it/s, grad_norm=0.942, loss_final=0.514, loss_mean=0.849, loss_mean_cls=0.0804, proj_loss=-0.415][2026-03-26 12:46:45] Step: 3777, Training Logs: loss_final: 0.496459, loss_mean: 0.831661, proj_loss: -0.415516, loss_mean_cls: 0.080314, grad_norm: 0.573146 +Steps: 0%| | 3778/1000000 [15:41<68:21:57, 4.05it/s, grad_norm=0.573, loss_final=0.496, loss_mean=0.832, loss_mean_cls=0.0803, proj_loss=-0.416][2026-03-26 12:46:45] Step: 3778, Training Logs: loss_final: 0.520952, loss_mean: 0.851741, proj_loss: -0.410308, loss_mean_cls: 0.079519, grad_norm: 0.492082 +Steps: 0%| | 3779/1000000 [15:42<68:23:56, 4.05it/s, grad_norm=0.492, loss_final=0.521, loss_mean=0.852, loss_mean_cls=0.0795, proj_loss=-0.41][2026-03-26 12:46:45] Step: 3779, Training Logs: loss_final: 0.507375, loss_mean: 0.840065, proj_loss: -0.412493, loss_mean_cls: 0.079803, grad_norm: 0.815063 +Steps: 0%| | 3780/1000000 [15:42<68:24:14, 4.05it/s, grad_norm=0.815, loss_final=0.507, loss_mean=0.84, loss_mean_cls=0.0798, proj_loss=-0.412][2026-03-26 12:46:45] Step: 3780, Training Logs: loss_final: 0.518214, loss_mean: 0.848285, proj_loss: -0.408678, loss_mean_cls: 0.078607, grad_norm: 0.462385 +Steps: 0%| | 3781/1000000 [15:42<68:28:16, 4.04it/s, grad_norm=0.462, loss_final=0.518, loss_mean=0.848, loss_mean_cls=0.0786, proj_loss=-0.409][2026-03-26 12:46:46] Step: 3781, Training Logs: loss_final: 0.508988, loss_mean: 0.834270, proj_loss: -0.408113, loss_mean_cls: 0.082831, grad_norm: 1.067303 +Steps: 0%| | 3782/1000000 [15:42<68:26:23, 4.04it/s, grad_norm=1.07, loss_final=0.509, loss_mean=0.834, loss_mean_cls=0.0828, proj_loss=-0.408][2026-03-26 12:46:46] Step: 3782, Training Logs: loss_final: 0.500891, loss_mean: 0.824604, proj_loss: -0.406839, loss_mean_cls: 0.083126, grad_norm: 0.663782 +Steps: 0%| | 3783/1000000 [15:43<68:29:17, 4.04it/s, grad_norm=0.664, loss_final=0.501, loss_mean=0.825, loss_mean_cls=0.0831, proj_loss=-0.407][2026-03-26 12:46:46] Step: 3783, Training Logs: loss_final: 0.513696, loss_mean: 0.834066, proj_loss: -0.404015, loss_mean_cls: 0.083644, grad_norm: 0.705210 +Steps: 0%| | 3784/1000000 [15:43<68:26:50, 4.04it/s, grad_norm=0.705, loss_final=0.514, loss_mean=0.834, loss_mean_cls=0.0836, proj_loss=-0.404][2026-03-26 12:46:46] Step: 3784, Training Logs: loss_final: 0.516397, loss_mean: 0.850790, proj_loss: -0.413781, loss_mean_cls: 0.079388, grad_norm: 0.529619 +Steps: 0%| | 3785/1000000 [15:43<68:26:37, 4.04it/s, grad_norm=0.53, loss_final=0.516, loss_mean=0.851, loss_mean_cls=0.0794, proj_loss=-0.414][2026-03-26 12:46:47] Step: 3785, Training Logs: loss_final: 0.534647, loss_mean: 0.856933, proj_loss: -0.403966, loss_mean_cls: 0.081680, grad_norm: 0.358897 +Steps: 0%| | 3786/1000000 [15:43<68:25:18, 4.04it/s, grad_norm=0.359, loss_final=0.535, loss_mean=0.857, loss_mean_cls=0.0817, proj_loss=-0.404][2026-03-26 12:46:47] Step: 3786, Training Logs: loss_final: 0.518086, loss_mean: 0.850149, proj_loss: -0.411973, loss_mean_cls: 0.079911, grad_norm: 0.647590 +Steps: 0%| | 3787/1000000 [15:44<68:24:10, 4.05it/s, grad_norm=0.648, loss_final=0.518, loss_mean=0.85, loss_mean_cls=0.0799, proj_loss=-0.412][2026-03-26 12:46:47] Step: 3787, Training Logs: loss_final: 0.526940, loss_mean: 0.840319, proj_loss: -0.397074, loss_mean_cls: 0.083695, grad_norm: 0.544725 +Steps: 0%| | 3788/1000000 [15:44<68:22:40, 4.05it/s, grad_norm=0.545, loss_final=0.527, loss_mean=0.84, loss_mean_cls=0.0837, proj_loss=-0.397][2026-03-26 12:46:47] Step: 3788, Training Logs: loss_final: 0.512359, loss_mean: 0.838893, proj_loss: -0.407326, loss_mean_cls: 0.080792, grad_norm: 0.459046 +Steps: 0%| | 3789/1000000 [15:44<68:22:56, 4.05it/s, grad_norm=0.459, loss_final=0.512, loss_mean=0.839, loss_mean_cls=0.0808, proj_loss=-0.407][2026-03-26 12:46:48] Step: 3789, Training Logs: loss_final: 0.521385, loss_mean: 0.861029, proj_loss: -0.415872, loss_mean_cls: 0.076228, grad_norm: 0.597502 +Steps: 0%| | 3790/1000000 [15:44<68:23:04, 4.05it/s, grad_norm=0.598, loss_final=0.521, loss_mean=0.861, loss_mean_cls=0.0762, proj_loss=-0.416][2026-03-26 12:46:48] Step: 3790, Training Logs: loss_final: 0.508768, loss_mean: 0.824180, proj_loss: -0.398888, loss_mean_cls: 0.083476, grad_norm: 0.293015 +Steps: 0%| | 3791/1000000 [15:45<68:24:08, 4.05it/s, grad_norm=0.293, loss_final=0.509, loss_mean=0.824, loss_mean_cls=0.0835, proj_loss=-0.399][2026-03-26 12:46:48] Step: 3791, Training Logs: loss_final: 0.505833, loss_mean: 0.830721, proj_loss: -0.406558, loss_mean_cls: 0.081670, grad_norm: 0.316209 +Steps: 0%| | 3792/1000000 [15:45<69:25:09, 3.99it/s, grad_norm=0.316, loss_final=0.506, loss_mean=0.831, loss_mean_cls=0.0817, proj_loss=-0.407][2026-03-26 12:46:48] Step: 3792, Training Logs: loss_final: 0.514358, loss_mean: 0.837577, proj_loss: -0.405344, loss_mean_cls: 0.082124, grad_norm: 0.335020 +Steps: 0%| | 3793/1000000 [15:45<69:06:31, 4.00it/s, grad_norm=0.335, loss_final=0.514, loss_mean=0.838, loss_mean_cls=0.0821, proj_loss=-0.405][2026-03-26 12:46:49] Step: 3793, Training Logs: loss_final: 0.508717, loss_mean: 0.816645, proj_loss: -0.391350, loss_mean_cls: 0.083422, grad_norm: 0.354505 +Steps: 0%| | 3794/1000000 [15:45<68:54:53, 4.02it/s, grad_norm=0.355, loss_final=0.509, loss_mean=0.817, loss_mean_cls=0.0834, proj_loss=-0.391][2026-03-26 12:46:49] Step: 3794, Training Logs: loss_final: 0.512841, loss_mean: 0.854936, proj_loss: -0.418766, loss_mean_cls: 0.076671, grad_norm: 0.261312 +Steps: 0%| | 3795/1000000 [15:46<68:46:04, 4.02it/s, grad_norm=0.261, loss_final=0.513, loss_mean=0.855, loss_mean_cls=0.0767, proj_loss=-0.419][2026-03-26 12:46:49] Step: 3795, Training Logs: loss_final: 0.500733, loss_mean: 0.836336, proj_loss: -0.414554, loss_mean_cls: 0.078951, grad_norm: 0.436886 +Steps: 0%| | 3796/1000000 [15:46<68:42:24, 4.03it/s, grad_norm=0.437, loss_final=0.501, loss_mean=0.836, loss_mean_cls=0.079, proj_loss=-0.415][2026-03-26 12:46:49] Step: 3796, Training Logs: loss_final: 0.510697, loss_mean: 0.834386, proj_loss: -0.403957, loss_mean_cls: 0.080269, grad_norm: 0.289105 +Steps: 0%| | 3797/1000000 [15:46<68:38:15, 4.03it/s, grad_norm=0.289, loss_final=0.511, loss_mean=0.834, loss_mean_cls=0.0803, proj_loss=-0.404][2026-03-26 12:46:50] Step: 3797, Training Logs: loss_final: 0.517315, loss_mean: 0.834297, proj_loss: -0.399793, loss_mean_cls: 0.082810, grad_norm: 0.506540 +Steps: 0%| | 3798/1000000 [15:46<68:36:16, 4.03it/s, grad_norm=0.507, loss_final=0.517, loss_mean=0.834, loss_mean_cls=0.0828, proj_loss=-0.4][2026-03-26 12:46:50] Step: 3798, Training Logs: loss_final: 0.502896, loss_mean: 0.833855, proj_loss: -0.410861, loss_mean_cls: 0.079902, grad_norm: 0.476275 +Steps: 0%| | 3799/1000000 [15:47<68:32:12, 4.04it/s, grad_norm=0.476, loss_final=0.503, loss_mean=0.834, loss_mean_cls=0.0799, proj_loss=-0.411][2026-03-26 12:46:50] Step: 3799, Training Logs: loss_final: 0.517576, loss_mean: 0.839228, proj_loss: -0.403740, loss_mean_cls: 0.082088, grad_norm: 0.317493 +Steps: 0%| | 3800/1000000 [15:47<68:30:43, 4.04it/s, grad_norm=0.317, loss_final=0.518, loss_mean=0.839, loss_mean_cls=0.0821, proj_loss=-0.404][2026-03-26 12:46:50] Step: 3800, Training Logs: loss_final: 0.497256, loss_mean: 0.833551, proj_loss: -0.415154, loss_mean_cls: 0.078859, grad_norm: 0.321574 +Steps: 0%| | 3801/1000000 [15:47<68:30:05, 4.04it/s, grad_norm=0.322, loss_final=0.497, loss_mean=0.834, loss_mean_cls=0.0789, proj_loss=-0.415][2026-03-26 12:46:50] Step: 3801, Training Logs: loss_final: 0.516589, loss_mean: 0.845302, proj_loss: -0.409711, loss_mean_cls: 0.080998, grad_norm: 0.315549 +Steps: 0%| | 3802/1000000 [15:47<68:29:04, 4.04it/s, grad_norm=0.316, loss_final=0.517, loss_mean=0.845, loss_mean_cls=0.081, proj_loss=-0.41][2026-03-26 12:46:51] Step: 3802, Training Logs: loss_final: 0.482322, loss_mean: 0.802093, proj_loss: -0.402413, loss_mean_cls: 0.082642, grad_norm: 0.427113 +Steps: 0%| | 3803/1000000 [15:48<68:26:32, 4.04it/s, grad_norm=0.427, loss_final=0.482, loss_mean=0.802, loss_mean_cls=0.0826, proj_loss=-0.402][2026-03-26 12:46:51] Step: 3803, Training Logs: loss_final: 0.508324, loss_mean: 0.826779, proj_loss: -0.399981, loss_mean_cls: 0.081526, grad_norm: 0.368002 +Steps: 0%| | 3804/1000000 [15:48<68:28:17, 4.04it/s, grad_norm=0.368, loss_final=0.508, loss_mean=0.827, loss_mean_cls=0.0815, proj_loss=-0.4][2026-03-26 12:46:51] Step: 3804, Training Logs: loss_final: 0.500624, loss_mean: 0.838033, proj_loss: -0.416667, loss_mean_cls: 0.079258, grad_norm: 0.309876 +Steps: 0%| | 3805/1000000 [15:48<68:25:42, 4.04it/s, grad_norm=0.31, loss_final=0.501, loss_mean=0.838, loss_mean_cls=0.0793, proj_loss=-0.417][2026-03-26 12:46:51] Step: 3805, Training Logs: loss_final: 0.516127, loss_mean: 0.837642, proj_loss: -0.403674, loss_mean_cls: 0.082159, grad_norm: 0.300041 +Steps: 0%| | 3806/1000000 [15:48<68:27:52, 4.04it/s, grad_norm=0.3, loss_final=0.516, loss_mean=0.838, loss_mean_cls=0.0822, proj_loss=-0.404][2026-03-26 12:46:52] Step: 3806, Training Logs: loss_final: 0.512273, loss_mean: 0.831998, proj_loss: -0.401378, loss_mean_cls: 0.081653, grad_norm: 0.241679 +Steps: 0%| | 3807/1000000 [15:49<68:29:59, 4.04it/s, grad_norm=0.242, loss_final=0.512, loss_mean=0.832, loss_mean_cls=0.0817, proj_loss=-0.401][2026-03-26 12:46:52] Step: 3807, Training Logs: loss_final: 0.505475, loss_mean: 0.843176, proj_loss: -0.417503, loss_mean_cls: 0.079802, grad_norm: 0.385053 +Steps: 0%| | 3808/1000000 [15:49<69:09:10, 4.00it/s, grad_norm=0.385, loss_final=0.505, loss_mean=0.843, loss_mean_cls=0.0798, proj_loss=-0.418][2026-03-26 12:46:52] Step: 3808, Training Logs: loss_final: 0.488573, loss_mean: 0.806563, proj_loss: -0.401410, loss_mean_cls: 0.083419, grad_norm: 0.363640 +Steps: 0%| | 3809/1000000 [15:49<71:48:41, 3.85it/s, grad_norm=0.364, loss_final=0.489, loss_mean=0.807, loss_mean_cls=0.0834, proj_loss=-0.401][2026-03-26 12:46:53] Step: 3809, Training Logs: loss_final: 0.495254, loss_mean: 0.821359, proj_loss: -0.408487, loss_mean_cls: 0.082383, grad_norm: 0.298672 +Steps: 0%| | 3810/1000000 [15:49<70:47:12, 3.91it/s, grad_norm=0.299, loss_final=0.495, loss_mean=0.821, loss_mean_cls=0.0824, proj_loss=-0.408][2026-03-26 12:46:53] Step: 3810, Training Logs: loss_final: 0.504425, loss_mean: 0.829146, proj_loss: -0.408187, loss_mean_cls: 0.083467, grad_norm: 0.292961 +Steps: 0%| | 3811/1000000 [15:50<70:08:21, 3.95it/s, grad_norm=0.293, loss_final=0.504, loss_mean=0.829, loss_mean_cls=0.0835, proj_loss=-0.408][2026-03-26 12:46:53] Step: 3811, Training Logs: loss_final: 0.495943, loss_mean: 0.817755, proj_loss: -0.404676, loss_mean_cls: 0.082864, grad_norm: 0.395606 +Steps: 0%| | 3812/1000000 [15:50<69:37:36, 3.97it/s, grad_norm=0.396, loss_final=0.496, loss_mean=0.818, loss_mean_cls=0.0829, proj_loss=-0.405][2026-03-26 12:46:53] Step: 3812, Training Logs: loss_final: 0.524105, loss_mean: 0.840128, proj_loss: -0.399200, loss_mean_cls: 0.083177, grad_norm: 0.419737 +Steps: 0%| | 3813/1000000 [15:50<69:18:24, 3.99it/s, grad_norm=0.42, loss_final=0.524, loss_mean=0.84, loss_mean_cls=0.0832, proj_loss=-0.399][2026-03-26 12:46:54] Step: 3813, Training Logs: loss_final: 0.515439, loss_mean: 0.835024, proj_loss: -0.402629, loss_mean_cls: 0.083045, grad_norm: 0.396152 +Steps: 0%| | 3814/1000000 [15:50<69:04:41, 4.01it/s, grad_norm=0.396, loss_final=0.515, loss_mean=0.835, loss_mean_cls=0.083, proj_loss=-0.403][2026-03-26 12:46:54] Step: 3814, Training Logs: loss_final: 0.509603, loss_mean: 0.830897, proj_loss: -0.405192, loss_mean_cls: 0.083899, grad_norm: 0.346036 +Steps: 0%| | 3815/1000000 [15:51<68:54:32, 4.02it/s, grad_norm=0.346, loss_final=0.51, loss_mean=0.831, loss_mean_cls=0.0839, proj_loss=-0.405][2026-03-26 12:46:54] Step: 3815, Training Logs: loss_final: 0.523913, loss_mean: 0.841672, proj_loss: -0.400020, loss_mean_cls: 0.082262, grad_norm: 0.396304 +Steps: 0%| | 3816/1000000 [15:51<68:44:32, 4.03it/s, grad_norm=0.396, loss_final=0.524, loss_mean=0.842, loss_mean_cls=0.0823, proj_loss=-0.4][2026-03-26 12:46:54] Step: 3816, Training Logs: loss_final: 0.497094, loss_mean: 0.830387, proj_loss: -0.413226, loss_mean_cls: 0.079933, grad_norm: 0.314418 +Steps: 0%| | 3817/1000000 [15:51<68:38:25, 4.03it/s, grad_norm=0.314, loss_final=0.497, loss_mean=0.83, loss_mean_cls=0.0799, proj_loss=-0.413][2026-03-26 12:46:55] Step: 3817, Training Logs: loss_final: 0.488241, loss_mean: 0.805586, proj_loss: -0.400554, loss_mean_cls: 0.083209, grad_norm: 0.473051 +Steps: 0%| | 3818/1000000 [15:51<68:37:20, 4.03it/s, grad_norm=0.473, loss_final=0.488, loss_mean=0.806, loss_mean_cls=0.0832, proj_loss=-0.401][2026-03-26 12:46:55] Step: 3818, Training Logs: loss_final: 0.510280, loss_mean: 0.837404, proj_loss: -0.408159, loss_mean_cls: 0.081035, grad_norm: 0.408222 +Steps: 0%| | 3819/1000000 [15:52<68:35:16, 4.03it/s, grad_norm=0.408, loss_final=0.51, loss_mean=0.837, loss_mean_cls=0.081, proj_loss=-0.408][2026-03-26 12:46:55] Step: 3819, Training Logs: loss_final: 0.493742, loss_mean: 0.812632, proj_loss: -0.402074, loss_mean_cls: 0.083184, grad_norm: 0.365185 +Steps: 0%| | 3820/1000000 [15:52<68:33:49, 4.04it/s, grad_norm=0.365, loss_final=0.494, loss_mean=0.813, loss_mean_cls=0.0832, proj_loss=-0.402][2026-03-26 12:46:55] Step: 3820, Training Logs: loss_final: 0.507083, loss_mean: 0.831035, proj_loss: -0.405294, loss_mean_cls: 0.081342, grad_norm: 0.336682 +Steps: 0%| | 3821/1000000 [15:52<68:31:48, 4.04it/s, grad_norm=0.337, loss_final=0.507, loss_mean=0.831, loss_mean_cls=0.0813, proj_loss=-0.405][2026-03-26 12:46:55] Step: 3821, Training Logs: loss_final: 0.507573, loss_mean: 0.846391, proj_loss: -0.417078, loss_mean_cls: 0.078260, grad_norm: 0.404742 +Steps: 0%| | 3822/1000000 [15:52<68:27:46, 4.04it/s, grad_norm=0.405, loss_final=0.508, loss_mean=0.846, loss_mean_cls=0.0783, proj_loss=-0.417][2026-03-26 12:46:56] Step: 3822, Training Logs: loss_final: 0.520038, loss_mean: 0.846776, proj_loss: -0.407237, loss_mean_cls: 0.080499, grad_norm: 0.356240 +Steps: 0%| | 3823/1000000 [15:53<68:25:20, 4.04it/s, grad_norm=0.356, loss_final=0.52, loss_mean=0.847, loss_mean_cls=0.0805, proj_loss=-0.407][2026-03-26 12:46:56] Step: 3823, Training Logs: loss_final: 0.511202, loss_mean: 0.830436, proj_loss: -0.402237, loss_mean_cls: 0.083002, grad_norm: 0.515361 +Steps: 0%| | 3824/1000000 [15:53<68:29:41, 4.04it/s, grad_norm=0.515, loss_final=0.511, loss_mean=0.83, loss_mean_cls=0.083, proj_loss=-0.402][2026-03-26 12:46:56] Step: 3824, Training Logs: loss_final: 0.505917, loss_mean: 0.821786, proj_loss: -0.400185, loss_mean_cls: 0.084316, grad_norm: 0.354568 +Steps: 0%| | 3825/1000000 [15:53<68:25:41, 4.04it/s, grad_norm=0.355, loss_final=0.506, loss_mean=0.822, loss_mean_cls=0.0843, proj_loss=-0.4][2026-03-26 12:46:56] Step: 3825, Training Logs: loss_final: 0.515739, loss_mean: 0.851934, proj_loss: -0.414227, loss_mean_cls: 0.078032, grad_norm: 0.282836 +Steps: 0%| | 3826/1000000 [15:53<68:28:08, 4.04it/s, grad_norm=0.283, loss_final=0.516, loss_mean=0.852, loss_mean_cls=0.078, proj_loss=-0.414][2026-03-26 12:46:57] Step: 3826, Training Logs: loss_final: 0.507971, loss_mean: 0.838614, proj_loss: -0.411297, loss_mean_cls: 0.080654, grad_norm: 0.489658 +Steps: 0%| | 3827/1000000 [15:54<68:26:02, 4.04it/s, grad_norm=0.49, loss_final=0.508, loss_mean=0.839, loss_mean_cls=0.0807, proj_loss=-0.411][2026-03-26 12:46:57] Step: 3827, Training Logs: loss_final: 0.505004, loss_mean: 0.826271, proj_loss: -0.403448, loss_mean_cls: 0.082182, grad_norm: 0.420603 +Steps: 0%| | 3828/1000000 [15:54<68:27:24, 4.04it/s, grad_norm=0.421, loss_final=0.505, loss_mean=0.826, loss_mean_cls=0.0822, proj_loss=-0.403][2026-03-26 12:46:57] Step: 3828, Training Logs: loss_final: 0.503899, loss_mean: 0.830307, proj_loss: -0.407045, loss_mean_cls: 0.080636, grad_norm: 0.315068 +Steps: 0%| | 3829/1000000 [15:54<68:24:57, 4.04it/s, grad_norm=0.315, loss_final=0.504, loss_mean=0.83, loss_mean_cls=0.0806, proj_loss=-0.407][2026-03-26 12:46:57] Step: 3829, Training Logs: loss_final: 0.516938, loss_mean: 0.859018, proj_loss: -0.419459, loss_mean_cls: 0.077379, grad_norm: 0.512190 +Steps: 0%| | 3830/1000000 [15:54<68:27:01, 4.04it/s, grad_norm=0.512, loss_final=0.517, loss_mean=0.859, loss_mean_cls=0.0774, proj_loss=-0.419][2026-03-26 12:46:58] Step: 3830, Training Logs: loss_final: 0.517238, loss_mean: 0.854096, proj_loss: -0.416237, loss_mean_cls: 0.079379, grad_norm: 0.316990 +Steps: 0%| | 3831/1000000 [15:55<68:26:08, 4.04it/s, grad_norm=0.317, loss_final=0.517, loss_mean=0.854, loss_mean_cls=0.0794, proj_loss=-0.416][2026-03-26 12:46:58] Step: 3831, Training Logs: loss_final: 0.481000, loss_mean: 0.800626, proj_loss: -0.402644, loss_mean_cls: 0.083018, grad_norm: 0.362280 +Steps: 0%| | 3832/1000000 [15:55<68:25:58, 4.04it/s, grad_norm=0.362, loss_final=0.481, loss_mean=0.801, loss_mean_cls=0.083, proj_loss=-0.403][2026-03-26 12:46:58] Step: 3832, Training Logs: loss_final: 0.500523, loss_mean: 0.820396, proj_loss: -0.402135, loss_mean_cls: 0.082262, grad_norm: 0.426128 +Steps: 0%| | 3833/1000000 [15:55<68:26:23, 4.04it/s, grad_norm=0.426, loss_final=0.501, loss_mean=0.82, loss_mean_cls=0.0823, proj_loss=-0.402][2026-03-26 12:46:58] Step: 3833, Training Logs: loss_final: 0.498687, loss_mean: 0.822667, proj_loss: -0.407541, loss_mean_cls: 0.083561, grad_norm: 0.234370 +Steps: 0%| | 3834/1000000 [15:55<68:26:53, 4.04it/s, grad_norm=0.234, loss_final=0.499, loss_mean=0.823, loss_mean_cls=0.0836, proj_loss=-0.408][2026-03-26 12:46:59] Step: 3834, Training Logs: loss_final: 0.504944, loss_mean: 0.829722, proj_loss: -0.405640, loss_mean_cls: 0.080862, grad_norm: 0.391003 +Steps: 0%| | 3835/1000000 [15:56<68:35:41, 4.03it/s, grad_norm=0.391, loss_final=0.505, loss_mean=0.83, loss_mean_cls=0.0809, proj_loss=-0.406][2026-03-26 12:46:59] Step: 3835, Training Logs: loss_final: 0.511788, loss_mean: 0.831888, proj_loss: -0.401741, loss_mean_cls: 0.081641, grad_norm: 0.364841 +Steps: 0%| | 3836/1000000 [15:56<68:35:28, 4.03it/s, grad_norm=0.365, loss_final=0.512, loss_mean=0.832, loss_mean_cls=0.0816, proj_loss=-0.402][2026-03-26 12:46:59] Step: 3836, Training Logs: loss_final: 0.511020, loss_mean: 0.843735, proj_loss: -0.413106, loss_mean_cls: 0.080391, grad_norm: 0.284306 +Steps: 0%| | 3837/1000000 [15:56<68:32:12, 4.04it/s, grad_norm=0.284, loss_final=0.511, loss_mean=0.844, loss_mean_cls=0.0804, proj_loss=-0.413][2026-03-26 12:46:59] Step: 3837, Training Logs: loss_final: 0.508755, loss_mean: 0.834330, proj_loss: -0.407752, loss_mean_cls: 0.082178, grad_norm: 0.415948 +Steps: 0%| | 3838/1000000 [15:56<68:30:11, 4.04it/s, grad_norm=0.416, loss_final=0.509, loss_mean=0.834, loss_mean_cls=0.0822, proj_loss=-0.408][2026-03-26 12:47:00] Step: 3838, Training Logs: loss_final: 0.483407, loss_mean: 0.800979, proj_loss: -0.401946, loss_mean_cls: 0.084374, grad_norm: 0.583204 +Steps: 0%| | 3839/1000000 [15:57<68:38:51, 4.03it/s, grad_norm=0.583, loss_final=0.483, loss_mean=0.801, loss_mean_cls=0.0844, proj_loss=-0.402][2026-03-26 12:47:00] Step: 3839, Training Logs: loss_final: 0.502766, loss_mean: 0.833294, proj_loss: -0.412318, loss_mean_cls: 0.081791, grad_norm: 0.400847 +Steps: 0%| | 3840/1000000 [15:57<68:33:29, 4.04it/s, grad_norm=0.401, loss_final=0.503, loss_mean=0.833, loss_mean_cls=0.0818, proj_loss=-0.412][2026-03-26 12:47:00] Step: 3840, Training Logs: loss_final: 0.508881, loss_mean: 0.841203, proj_loss: -0.412663, loss_mean_cls: 0.080342, grad_norm: 0.410354 +Steps: 0%| | 3841/1000000 [15:57<68:31:07, 4.04it/s, grad_norm=0.41, loss_final=0.509, loss_mean=0.841, loss_mean_cls=0.0803, proj_loss=-0.413][2026-03-26 12:47:00] Step: 3841, Training Logs: loss_final: 0.498368, loss_mean: 0.827075, proj_loss: -0.410387, loss_mean_cls: 0.081680, grad_norm: 0.520951 +Steps: 0%| | 3842/1000000 [15:57<68:28:38, 4.04it/s, grad_norm=0.521, loss_final=0.498, loss_mean=0.827, loss_mean_cls=0.0817, proj_loss=-0.41][2026-03-26 12:47:01] Step: 3842, Training Logs: loss_final: 0.505432, loss_mean: 0.821799, proj_loss: -0.399690, loss_mean_cls: 0.083322, grad_norm: 0.386302 +Steps: 0%| | 3843/1000000 [15:58<68:37:39, 4.03it/s, grad_norm=0.386, loss_final=0.505, loss_mean=0.822, loss_mean_cls=0.0833, proj_loss=-0.4][2026-03-26 12:47:01] Step: 3843, Training Logs: loss_final: 0.526989, loss_mean: 0.858581, proj_loss: -0.410753, loss_mean_cls: 0.079161, grad_norm: 0.438141 +Steps: 0%| | 3844/1000000 [15:58<68:33:34, 4.04it/s, grad_norm=0.438, loss_final=0.527, loss_mean=0.859, loss_mean_cls=0.0792, proj_loss=-0.411][2026-03-26 12:47:01] Step: 3844, Training Logs: loss_final: 0.517441, loss_mean: 0.844478, proj_loss: -0.407321, loss_mean_cls: 0.080284, grad_norm: 0.547701 +Steps: 0%| | 3845/1000000 [15:58<68:48:21, 4.02it/s, grad_norm=0.548, loss_final=0.517, loss_mean=0.844, loss_mean_cls=0.0803, proj_loss=-0.407][2026-03-26 12:47:01] Step: 3845, Training Logs: loss_final: 0.499833, loss_mean: 0.809052, proj_loss: -0.394836, loss_mean_cls: 0.085617, grad_norm: 0.311697 +Steps: 0%| | 3846/1000000 [15:58<68:42:28, 4.03it/s, grad_norm=0.312, loss_final=0.5, loss_mean=0.809, loss_mean_cls=0.0856, proj_loss=-0.395][2026-03-26 12:47:02] Step: 3846, Training Logs: loss_final: 0.510915, loss_mean: 0.843621, proj_loss: -0.411639, loss_mean_cls: 0.078933, grad_norm: 0.640228 +Steps: 0%| | 3847/1000000 [15:59<68:50:37, 4.02it/s, grad_norm=0.64, loss_final=0.511, loss_mean=0.844, loss_mean_cls=0.0789, proj_loss=-0.412][2026-03-26 12:47:02] Step: 3847, Training Logs: loss_final: 0.484229, loss_mean: 0.802854, proj_loss: -0.403572, loss_mean_cls: 0.084947, grad_norm: 0.474680 +Steps: 0%| | 3848/1000000 [15:59<68:42:55, 4.03it/s, grad_norm=0.475, loss_final=0.484, loss_mean=0.803, loss_mean_cls=0.0849, proj_loss=-0.404][2026-03-26 12:47:02] Step: 3848, Training Logs: loss_final: 0.496660, loss_mean: 0.818809, proj_loss: -0.405261, loss_mean_cls: 0.083112, grad_norm: 0.406242 +Steps: 0%| | 3849/1000000 [15:59<68:37:45, 4.03it/s, grad_norm=0.406, loss_final=0.497, loss_mean=0.819, loss_mean_cls=0.0831, proj_loss=-0.405][2026-03-26 12:47:02] Step: 3849, Training Logs: loss_final: 0.505940, loss_mean: 0.831841, proj_loss: -0.406641, loss_mean_cls: 0.080740, grad_norm: 0.577740 +Steps: 0%| | 3850/1000000 [15:59<68:32:34, 4.04it/s, grad_norm=0.578, loss_final=0.506, loss_mean=0.832, loss_mean_cls=0.0807, proj_loss=-0.407][2026-03-26 12:47:03] Step: 3850, Training Logs: loss_final: 0.513906, loss_mean: 0.843436, proj_loss: -0.411772, loss_mean_cls: 0.082243, grad_norm: 0.373584 +Steps: 0%| | 3851/1000000 [16:00<87:09:13, 3.17it/s, grad_norm=0.374, loss_final=0.514, loss_mean=0.843, loss_mean_cls=0.0822, proj_loss=-0.412][2026-03-26 12:47:03] Step: 3851, Training Logs: loss_final: 0.517364, loss_mean: 0.836031, proj_loss: -0.401544, loss_mean_cls: 0.082877, grad_norm: 0.387897 +Steps: 0%| | 3852/1000000 [16:00<81:32:33, 3.39it/s, grad_norm=0.388, loss_final=0.517, loss_mean=0.836, loss_mean_cls=0.0829, proj_loss=-0.402][2026-03-26 12:47:03] Step: 3852, Training Logs: loss_final: 0.502098, loss_mean: 0.831520, proj_loss: -0.411058, loss_mean_cls: 0.081635, grad_norm: 0.472379 +Steps: 0%| | 3853/1000000 [16:00<77:36:31, 3.57it/s, grad_norm=0.472, loss_final=0.502, loss_mean=0.832, loss_mean_cls=0.0816, proj_loss=-0.411][2026-03-26 12:47:04] Step: 3853, Training Logs: loss_final: 0.514240, loss_mean: 0.846626, proj_loss: -0.410845, loss_mean_cls: 0.078459, grad_norm: 0.605301 +Steps: 0%| | 3854/1000000 [16:00<74:51:23, 3.70it/s, grad_norm=0.605, loss_final=0.514, loss_mean=0.847, loss_mean_cls=0.0785, proj_loss=-0.411][2026-03-26 12:47:04] Step: 3854, Training Logs: loss_final: 0.522768, loss_mean: 0.854742, proj_loss: -0.410410, loss_mean_cls: 0.078437, grad_norm: 0.456022 +Steps: 0%| | 3855/1000000 [16:01<73:05:00, 3.79it/s, grad_norm=0.456, loss_final=0.523, loss_mean=0.855, loss_mean_cls=0.0784, proj_loss=-0.41][2026-03-26 12:47:04] Step: 3855, Training Logs: loss_final: 0.522312, loss_mean: 0.860869, proj_loss: -0.416002, loss_mean_cls: 0.077445, grad_norm: 0.505496 +Steps: 0%| | 3856/1000000 [16:01<71:38:56, 3.86it/s, grad_norm=0.505, loss_final=0.522, loss_mean=0.861, loss_mean_cls=0.0774, proj_loss=-0.416][2026-03-26 12:47:04] Step: 3856, Training Logs: loss_final: 0.511926, loss_mean: 0.831604, proj_loss: -0.402394, loss_mean_cls: 0.082716, grad_norm: 0.742211 +Steps: 0%| | 3857/1000000 [16:01<70:40:36, 3.92it/s, grad_norm=0.742, loss_final=0.512, loss_mean=0.832, loss_mean_cls=0.0827, proj_loss=-0.402][2026-03-26 12:47:05] Step: 3857, Training Logs: loss_final: 0.501532, loss_mean: 0.827761, proj_loss: -0.408687, loss_mean_cls: 0.082457, grad_norm: 0.584922 +Steps: 0%| | 3858/1000000 [16:01<69:56:43, 3.96it/s, grad_norm=0.585, loss_final=0.502, loss_mean=0.828, loss_mean_cls=0.0825, proj_loss=-0.409][2026-03-26 12:47:05] Step: 3858, Training Logs: loss_final: 0.510360, loss_mean: 0.837648, proj_loss: -0.408835, loss_mean_cls: 0.081547, grad_norm: 0.477977 +Steps: 0%| | 3859/1000000 [16:02<69:28:52, 3.98it/s, grad_norm=0.478, loss_final=0.51, loss_mean=0.838, loss_mean_cls=0.0815, proj_loss=-0.409][2026-03-26 12:47:05] Step: 3859, Training Logs: loss_final: 0.503113, loss_mean: 0.813441, proj_loss: -0.394840, loss_mean_cls: 0.084513, grad_norm: 0.494977 +Steps: 0%| | 3860/1000000 [16:02<69:07:02, 4.00it/s, grad_norm=0.495, loss_final=0.503, loss_mean=0.813, loss_mean_cls=0.0845, proj_loss=-0.395][2026-03-26 12:47:05] Step: 3860, Training Logs: loss_final: 0.503716, loss_mean: 0.835686, proj_loss: -0.412450, loss_mean_cls: 0.080480, grad_norm: 0.412812 +Steps: 0%| | 3861/1000000 [16:02<68:56:11, 4.01it/s, grad_norm=0.413, loss_final=0.504, loss_mean=0.836, loss_mean_cls=0.0805, proj_loss=-0.412][2026-03-26 12:47:06] Step: 3861, Training Logs: loss_final: 0.524168, loss_mean: 0.853000, proj_loss: -0.407481, loss_mean_cls: 0.078649, grad_norm: 0.418133 +Steps: 0%| | 3862/1000000 [16:02<68:45:06, 4.02it/s, grad_norm=0.418, loss_final=0.524, loss_mean=0.853, loss_mean_cls=0.0786, proj_loss=-0.407][2026-03-26 12:47:06] Step: 3862, Training Logs: loss_final: 0.515831, loss_mean: 0.832339, proj_loss: -0.399717, loss_mean_cls: 0.083208, grad_norm: 0.644129 +Steps: 0%| | 3863/1000000 [16:03<68:40:26, 4.03it/s, grad_norm=0.644, loss_final=0.516, loss_mean=0.832, loss_mean_cls=0.0832, proj_loss=-0.4][2026-03-26 12:47:06] Step: 3863, Training Logs: loss_final: 0.510760, loss_mean: 0.853498, proj_loss: -0.422035, loss_mean_cls: 0.079298, grad_norm: 0.381289 +Steps: 0%| | 3864/1000000 [16:03<85:00:48, 3.25it/s, grad_norm=0.381, loss_final=0.511, loss_mean=0.853, loss_mean_cls=0.0793, proj_loss=-0.422][2026-03-26 12:47:07] Step: 3864, Training Logs: loss_final: 0.503284, loss_mean: 0.837896, proj_loss: -0.413673, loss_mean_cls: 0.079061, grad_norm: 0.453075 +Steps: 0%| | 3865/1000000 [16:03<80:01:37, 3.46it/s, grad_norm=0.453, loss_final=0.503, loss_mean=0.838, loss_mean_cls=0.0791, proj_loss=-0.414][2026-03-26 12:47:07] Step: 3865, Training Logs: loss_final: 0.504777, loss_mean: 0.826243, proj_loss: -0.402714, loss_mean_cls: 0.081248, grad_norm: 0.374938 +Steps: 0%| | 3866/1000000 [16:04<76:31:13, 3.62it/s, grad_norm=0.375, loss_final=0.505, loss_mean=0.826, loss_mean_cls=0.0812, proj_loss=-0.403][2026-03-26 12:47:07] Step: 3866, Training Logs: loss_final: 0.503721, loss_mean: 0.823139, proj_loss: -0.401978, loss_mean_cls: 0.082559, grad_norm: 0.335173 +Steps: 0%| | 3867/1000000 [16:04<74:05:01, 3.74it/s, grad_norm=0.335, loss_final=0.504, loss_mean=0.823, loss_mean_cls=0.0826, proj_loss=-0.402][2026-03-26 12:47:07] Step: 3867, Training Logs: loss_final: 0.493599, loss_mean: 0.807614, proj_loss: -0.399262, loss_mean_cls: 0.085247, grad_norm: 0.389314 +Steps: 0%| | 3868/1000000 [16:04<72:23:47, 3.82it/s, grad_norm=0.389, loss_final=0.494, loss_mean=0.808, loss_mean_cls=0.0852, proj_loss=-0.399][2026-03-26 12:47:08] Step: 3868, Training Logs: loss_final: 0.508699, loss_mean: 0.843818, proj_loss: -0.415954, loss_mean_cls: 0.080835, grad_norm: 0.317971 +Steps: 0%| | 3869/1000000 [16:04<71:09:33, 3.89it/s, grad_norm=0.318, loss_final=0.509, loss_mean=0.844, loss_mean_cls=0.0808, proj_loss=-0.416][2026-03-26 12:47:08] Step: 3869, Training Logs: loss_final: 0.504181, loss_mean: 0.828781, proj_loss: -0.405210, loss_mean_cls: 0.080610, grad_norm: 0.399682 +Steps: 0%| | 3870/1000000 [16:05<70:18:50, 3.94it/s, grad_norm=0.4, loss_final=0.504, loss_mean=0.829, loss_mean_cls=0.0806, proj_loss=-0.405][2026-03-26 12:47:08] Step: 3870, Training Logs: loss_final: 0.508757, loss_mean: 0.827463, proj_loss: -0.401027, loss_mean_cls: 0.082320, grad_norm: 0.430041 +Steps: 0%| | 3871/1000000 [16:05<69:44:59, 3.97it/s, grad_norm=0.43, loss_final=0.509, loss_mean=0.827, loss_mean_cls=0.0823, proj_loss=-0.401][2026-03-26 12:47:08] Step: 3871, Training Logs: loss_final: 0.500702, loss_mean: 0.833994, proj_loss: -0.413700, loss_mean_cls: 0.080408, grad_norm: 0.301744 +Steps: 0%| | 3872/1000000 [16:05<69:21:45, 3.99it/s, grad_norm=0.302, loss_final=0.501, loss_mean=0.834, loss_mean_cls=0.0804, proj_loss=-0.414][2026-03-26 12:47:09] Step: 3872, Training Logs: loss_final: 0.511792, loss_mean: 0.833184, proj_loss: -0.403516, loss_mean_cls: 0.082124, grad_norm: 0.522946 +Steps: 0%| | 3873/1000000 [16:05<69:02:31, 4.01it/s, grad_norm=0.523, loss_final=0.512, loss_mean=0.833, loss_mean_cls=0.0821, proj_loss=-0.404][2026-03-26 12:47:09] Step: 3873, Training Logs: loss_final: 0.500074, loss_mean: 0.826760, proj_loss: -0.409116, loss_mean_cls: 0.082429, grad_norm: 0.393052 +Steps: 0%| | 3874/1000000 [16:06<68:50:39, 4.02it/s, grad_norm=0.393, loss_final=0.5, loss_mean=0.827, loss_mean_cls=0.0824, proj_loss=-0.409][2026-03-26 12:47:09] Step: 3874, Training Logs: loss_final: 0.504709, loss_mean: 0.816992, proj_loss: -0.396784, loss_mean_cls: 0.084501, grad_norm: 0.329964 +Steps: 0%| | 3875/1000000 [16:06<68:42:49, 4.03it/s, grad_norm=0.33, loss_final=0.505, loss_mean=0.817, loss_mean_cls=0.0845, proj_loss=-0.397][2026-03-26 12:47:09] Step: 3875, Training Logs: loss_final: 0.511190, loss_mean: 0.828183, proj_loss: -0.398910, loss_mean_cls: 0.081918, grad_norm: 0.318842 +Steps: 0%| | 3876/1000000 [16:06<86:52:53, 3.18it/s, grad_norm=0.319, loss_final=0.511, loss_mean=0.828, loss_mean_cls=0.0819, proj_loss=-0.399][2026-03-26 12:47:10] Step: 3876, Training Logs: loss_final: 0.529227, loss_mean: 0.851862, proj_loss: -0.404493, loss_mean_cls: 0.081857, grad_norm: 0.300253 +Steps: 0%| | 3877/1000000 [16:07<81:16:18, 3.40it/s, grad_norm=0.3, loss_final=0.529, loss_mean=0.852, loss_mean_cls=0.0819, proj_loss=-0.404][2026-03-26 12:47:10] Step: 3877, Training Logs: loss_final: 0.527759, loss_mean: 0.863200, proj_loss: -0.413349, loss_mean_cls: 0.077908, grad_norm: 0.269644 +Steps: 0%| | 3878/1000000 [16:07<77:25:37, 3.57it/s, grad_norm=0.27, loss_final=0.528, loss_mean=0.863, loss_mean_cls=0.0779, proj_loss=-0.413][2026-03-26 12:47:10] Step: 3878, Training Logs: loss_final: 0.509029, loss_mean: 0.849759, proj_loss: -0.416854, loss_mean_cls: 0.076124, grad_norm: 0.365778 +Steps: 0%| | 3879/1000000 [16:07<74:40:56, 3.71it/s, grad_norm=0.366, loss_final=0.509, loss_mean=0.85, loss_mean_cls=0.0761, proj_loss=-0.417][2026-03-26 12:47:10] Step: 3879, Training Logs: loss_final: 0.504092, loss_mean: 0.835683, proj_loss: -0.413889, loss_mean_cls: 0.082298, grad_norm: 0.358414 +Steps: 0%| | 3880/1000000 [16:07<72:50:09, 3.80it/s, grad_norm=0.358, loss_final=0.504, loss_mean=0.836, loss_mean_cls=0.0823, proj_loss=-0.414][2026-03-26 12:47:11] Step: 3880, Training Logs: loss_final: 0.510123, loss_mean: 0.833517, proj_loss: -0.405081, loss_mean_cls: 0.081687, grad_norm: 0.397471 +Steps: 0%| | 3881/1000000 [16:08<71:27:26, 3.87it/s, grad_norm=0.397, loss_final=0.51, loss_mean=0.834, loss_mean_cls=0.0817, proj_loss=-0.405][2026-03-26 12:47:11] Step: 3881, Training Logs: loss_final: 0.504665, loss_mean: 0.838499, proj_loss: -0.413994, loss_mean_cls: 0.080160, grad_norm: 0.384915 +Steps: 0%| | 3882/1000000 [16:08<70:34:38, 3.92it/s, grad_norm=0.385, loss_final=0.505, loss_mean=0.838, loss_mean_cls=0.0802, proj_loss=-0.414][2026-03-26 12:47:11] Step: 3882, Training Logs: loss_final: 0.499235, loss_mean: 0.822813, proj_loss: -0.406191, loss_mean_cls: 0.082613, grad_norm: 0.349000 +Steps: 0%| | 3883/1000000 [16:08<69:54:23, 3.96it/s, grad_norm=0.349, loss_final=0.499, loss_mean=0.823, loss_mean_cls=0.0826, proj_loss=-0.406][2026-03-26 12:47:11] Step: 3883, Training Logs: loss_final: 0.520935, loss_mean: 0.855454, proj_loss: -0.413036, loss_mean_cls: 0.078517, grad_norm: 0.520461 +Steps: 0%| | 3884/1000000 [16:08<69:27:11, 3.98it/s, grad_norm=0.52, loss_final=0.521, loss_mean=0.855, loss_mean_cls=0.0785, proj_loss=-0.413][2026-03-26 12:47:12] Step: 3884, Training Logs: loss_final: 0.522207, loss_mean: 0.854572, proj_loss: -0.412384, loss_mean_cls: 0.080019, grad_norm: 0.406136 +Steps: 0%| | 3885/1000000 [16:09<69:05:28, 4.00it/s, grad_norm=0.406, loss_final=0.522, loss_mean=0.855, loss_mean_cls=0.08, proj_loss=-0.412][2026-03-26 12:47:12] Step: 3885, Training Logs: loss_final: 0.504022, loss_mean: 0.838159, proj_loss: -0.414282, loss_mean_cls: 0.080145, grad_norm: 0.418831 +Steps: 0%| | 3886/1000000 [16:09<68:54:38, 4.02it/s, grad_norm=0.419, loss_final=0.504, loss_mean=0.838, loss_mean_cls=0.0801, proj_loss=-0.414][2026-03-26 12:47:12] Step: 3886, Training Logs: loss_final: 0.498786, loss_mean: 0.823267, proj_loss: -0.405715, loss_mean_cls: 0.081234, grad_norm: 0.411676 +Steps: 0%| | 3887/1000000 [16:09<68:45:16, 4.02it/s, grad_norm=0.412, loss_final=0.499, loss_mean=0.823, loss_mean_cls=0.0812, proj_loss=-0.406][2026-03-26 12:47:12] Step: 3887, Training Logs: loss_final: 0.504924, loss_mean: 0.833401, proj_loss: -0.408839, loss_mean_cls: 0.080362, grad_norm: 0.507098 +Steps: 0%| | 3888/1000000 [16:09<68:39:39, 4.03it/s, grad_norm=0.507, loss_final=0.505, loss_mean=0.833, loss_mean_cls=0.0804, proj_loss=-0.409][2026-03-26 12:47:13] Step: 3888, Training Logs: loss_final: 0.507856, loss_mean: 0.834672, proj_loss: -0.407601, loss_mean_cls: 0.080785, grad_norm: 0.578507 +Steps: 0%| | 3889/1000000 [16:10<68:34:54, 4.03it/s, grad_norm=0.579, loss_final=0.508, loss_mean=0.835, loss_mean_cls=0.0808, proj_loss=-0.408][2026-03-26 12:47:13] Step: 3889, Training Logs: loss_final: 0.505040, loss_mean: 0.841610, proj_loss: -0.415661, loss_mean_cls: 0.079091, grad_norm: 0.451996 +Steps: 0%| | 3890/1000000 [16:10<68:31:05, 4.04it/s, grad_norm=0.452, loss_final=0.505, loss_mean=0.842, loss_mean_cls=0.0791, proj_loss=-0.416][2026-03-26 12:47:13] Step: 3890, Training Logs: loss_final: 0.490321, loss_mean: 0.801433, proj_loss: -0.396627, loss_mean_cls: 0.085516, grad_norm: 0.291007 +Steps: 0%| | 3891/1000000 [16:10<68:28:52, 4.04it/s, grad_norm=0.291, loss_final=0.49, loss_mean=0.801, loss_mean_cls=0.0855, proj_loss=-0.397][2026-03-26 12:47:13] Step: 3891, Training Logs: loss_final: 0.512822, loss_mean: 0.831903, proj_loss: -0.401849, loss_mean_cls: 0.082768, grad_norm: 0.473961 +Steps: 0%| | 3892/1000000 [16:10<68:29:46, 4.04it/s, grad_norm=0.474, loss_final=0.513, loss_mean=0.832, loss_mean_cls=0.0828, proj_loss=-0.402][2026-03-26 12:47:14] Step: 3892, Training Logs: loss_final: 0.526840, loss_mean: 0.863989, proj_loss: -0.416607, loss_mean_cls: 0.079458, grad_norm: 0.646913 +Steps: 0%| | 3893/1000000 [16:11<68:31:23, 4.04it/s, grad_norm=0.647, loss_final=0.527, loss_mean=0.864, loss_mean_cls=0.0795, proj_loss=-0.417][2026-03-26 12:47:14] Step: 3893, Training Logs: loss_final: 0.515369, loss_mean: 0.857647, proj_loss: -0.420618, loss_mean_cls: 0.078341, grad_norm: 0.398442 +Steps: 0%| | 3894/1000000 [16:11<68:27:55, 4.04it/s, grad_norm=0.398, loss_final=0.515, loss_mean=0.858, loss_mean_cls=0.0783, proj_loss=-0.421][2026-03-26 12:47:14] Step: 3894, Training Logs: loss_final: 0.516641, loss_mean: 0.836250, proj_loss: -0.402869, loss_mean_cls: 0.083260, grad_norm: 0.321371 +Steps: 0%| | 3895/1000000 [16:11<68:27:40, 4.04it/s, grad_norm=0.321, loss_final=0.517, loss_mean=0.836, loss_mean_cls=0.0833, proj_loss=-0.403][2026-03-26 12:47:14] Step: 3895, Training Logs: loss_final: 0.516623, loss_mean: 0.841229, proj_loss: -0.405089, loss_mean_cls: 0.080483, grad_norm: 0.438262 +Steps: 0%| | 3896/1000000 [16:11<68:26:33, 4.04it/s, grad_norm=0.438, loss_final=0.517, loss_mean=0.841, loss_mean_cls=0.0805, proj_loss=-0.405][2026-03-26 12:47:15] Step: 3896, Training Logs: loss_final: 0.514125, loss_mean: 0.843173, proj_loss: -0.408898, loss_mean_cls: 0.079850, grad_norm: 0.382844 +Steps: 0%| | 3897/1000000 [16:12<68:27:30, 4.04it/s, grad_norm=0.383, loss_final=0.514, loss_mean=0.843, loss_mean_cls=0.0798, proj_loss=-0.409][2026-03-26 12:47:15] Step: 3897, Training Logs: loss_final: 0.500636, loss_mean: 0.839101, proj_loss: -0.418074, loss_mean_cls: 0.079609, grad_norm: 0.513285 +Steps: 0%| | 3898/1000000 [16:12<68:26:09, 4.04it/s, grad_norm=0.513, loss_final=0.501, loss_mean=0.839, loss_mean_cls=0.0796, proj_loss=-0.418][2026-03-26 12:47:15] Step: 3898, Training Logs: loss_final: 0.505587, loss_mean: 0.836948, proj_loss: -0.411920, loss_mean_cls: 0.080560, grad_norm: 0.493610 +Steps: 0%| | 3899/1000000 [16:12<68:23:48, 4.05it/s, grad_norm=0.494, loss_final=0.506, loss_mean=0.837, loss_mean_cls=0.0806, proj_loss=-0.412][2026-03-26 12:47:15] Step: 3899, Training Logs: loss_final: 0.515826, loss_mean: 0.833037, proj_loss: -0.400252, loss_mean_cls: 0.083042, grad_norm: 0.318595 +Steps: 0%| | 3900/1000000 [16:12<68:24:12, 4.05it/s, grad_norm=0.319, loss_final=0.516, loss_mean=0.833, loss_mean_cls=0.083, proj_loss=-0.4][2026-03-26 12:47:16] Step: 3900, Training Logs: loss_final: 0.506846, loss_mean: 0.848648, proj_loss: -0.418360, loss_mean_cls: 0.076558, grad_norm: 0.713394 +Steps: 0%| | 3901/1000000 [16:13<68:24:15, 4.04it/s, grad_norm=0.713, loss_final=0.507, loss_mean=0.849, loss_mean_cls=0.0766, proj_loss=-0.418][2026-03-26 12:47:16] Step: 3901, Training Logs: loss_final: 0.505841, loss_mean: 0.839429, proj_loss: -0.412654, loss_mean_cls: 0.079066, grad_norm: 0.886887 +Steps: 0%| | 3902/1000000 [16:13<68:23:54, 4.05it/s, grad_norm=0.887, loss_final=0.506, loss_mean=0.839, loss_mean_cls=0.0791, proj_loss=-0.413][2026-03-26 12:47:16] Step: 3902, Training Logs: loss_final: 0.498761, loss_mean: 0.822981, proj_loss: -0.406342, loss_mean_cls: 0.082121, grad_norm: 0.314809 +Steps: 0%| | 3903/1000000 [16:13<68:24:44, 4.04it/s, grad_norm=0.315, loss_final=0.499, loss_mean=0.823, loss_mean_cls=0.0821, proj_loss=-0.406][2026-03-26 12:47:16] Step: 3903, Training Logs: loss_final: 0.512581, loss_mean: 0.838196, proj_loss: -0.406465, loss_mean_cls: 0.080850, grad_norm: 0.671409 +Steps: 0%| | 3904/1000000 [16:13<68:26:18, 4.04it/s, grad_norm=0.671, loss_final=0.513, loss_mean=0.838, loss_mean_cls=0.0809, proj_loss=-0.406][2026-03-26 12:47:17] Step: 3904, Training Logs: loss_final: 0.517145, loss_mean: 0.852581, proj_loss: -0.414921, loss_mean_cls: 0.079485, grad_norm: 0.440086 +Steps: 0%| | 3905/1000000 [16:14<68:24:25, 4.04it/s, grad_norm=0.44, loss_final=0.517, loss_mean=0.853, loss_mean_cls=0.0795, proj_loss=-0.415][2026-03-26 12:47:17] Step: 3905, Training Logs: loss_final: 0.517394, loss_mean: 0.854974, proj_loss: -0.415525, loss_mean_cls: 0.077946, grad_norm: 0.358476 +Steps: 0%| | 3906/1000000 [16:14<68:24:20, 4.04it/s, grad_norm=0.358, loss_final=0.517, loss_mean=0.855, loss_mean_cls=0.0779, proj_loss=-0.416][2026-03-26 12:47:17] Step: 3906, Training Logs: loss_final: 0.508465, loss_mean: 0.836021, proj_loss: -0.408220, loss_mean_cls: 0.080664, grad_norm: 0.482463 +Steps: 0%| | 3907/1000000 [16:14<68:23:20, 4.05it/s, grad_norm=0.482, loss_final=0.508, loss_mean=0.836, loss_mean_cls=0.0807, proj_loss=-0.408][2026-03-26 12:47:17] Step: 3907, Training Logs: loss_final: 0.494770, loss_mean: 0.832544, proj_loss: -0.418063, loss_mean_cls: 0.080288, grad_norm: 0.297859 +Steps: 0%| | 3908/1000000 [16:14<68:25:22, 4.04it/s, grad_norm=0.298, loss_final=0.495, loss_mean=0.833, loss_mean_cls=0.0803, proj_loss=-0.418][2026-03-26 12:47:18] Step: 3908, Training Logs: loss_final: 0.515366, loss_mean: 0.830517, proj_loss: -0.399114, loss_mean_cls: 0.083963, grad_norm: 0.410437 +Steps: 0%| | 3909/1000000 [16:15<68:30:20, 4.04it/s, grad_norm=0.41, loss_final=0.515, loss_mean=0.831, loss_mean_cls=0.084, proj_loss=-0.399][2026-03-26 12:47:18] Step: 3909, Training Logs: loss_final: 0.501976, loss_mean: 0.809779, proj_loss: -0.391765, loss_mean_cls: 0.083962, grad_norm: 0.504640 +Steps: 0%| | 3910/1000000 [16:15<68:26:58, 4.04it/s, grad_norm=0.505, loss_final=0.502, loss_mean=0.81, loss_mean_cls=0.084, proj_loss=-0.392][2026-03-26 12:47:18] Step: 3910, Training Logs: loss_final: 0.506808, loss_mean: 0.841563, proj_loss: -0.415216, loss_mean_cls: 0.080460, grad_norm: 0.287285 +Steps: 0%| | 3911/1000000 [16:15<68:25:40, 4.04it/s, grad_norm=0.287, loss_final=0.507, loss_mean=0.842, loss_mean_cls=0.0805, proj_loss=-0.415][2026-03-26 12:47:18] Step: 3911, Training Logs: loss_final: 0.489490, loss_mean: 0.808018, proj_loss: -0.401556, loss_mean_cls: 0.083029, grad_norm: 0.814532 +Steps: 0%| | 3912/1000000 [16:15<68:28:20, 4.04it/s, grad_norm=0.815, loss_final=0.489, loss_mean=0.808, loss_mean_cls=0.083, proj_loss=-0.402][2026-03-26 12:47:19] Step: 3912, Training Logs: loss_final: 0.493114, loss_mean: 0.809220, proj_loss: -0.401083, loss_mean_cls: 0.084977, grad_norm: 0.649409 +Steps: 0%| | 3913/1000000 [16:16<68:26:35, 4.04it/s, grad_norm=0.649, loss_final=0.493, loss_mean=0.809, loss_mean_cls=0.085, proj_loss=-0.401][2026-03-26 12:47:19] Step: 3913, Training Logs: loss_final: 0.490017, loss_mean: 0.810311, proj_loss: -0.405062, loss_mean_cls: 0.084768, grad_norm: 0.343317 +Steps: 0%| | 3914/1000000 [16:16<68:27:34, 4.04it/s, grad_norm=0.343, loss_final=0.49, loss_mean=0.81, loss_mean_cls=0.0848, proj_loss=-0.405][2026-03-26 12:47:19] Step: 3914, Training Logs: loss_final: 0.506352, loss_mean: 0.826115, proj_loss: -0.402794, loss_mean_cls: 0.083031, grad_norm: 0.490510 +Steps: 0%| | 3915/1000000 [16:16<68:25:47, 4.04it/s, grad_norm=0.491, loss_final=0.506, loss_mean=0.826, loss_mean_cls=0.083, proj_loss=-0.403][2026-03-26 12:47:19] Step: 3915, Training Logs: loss_final: 0.499001, loss_mean: 0.843982, proj_loss: -0.421382, loss_mean_cls: 0.076401, grad_norm: 0.322008 +Steps: 0%| | 3916/1000000 [16:16<68:25:34, 4.04it/s, grad_norm=0.322, loss_final=0.499, loss_mean=0.844, loss_mean_cls=0.0764, proj_loss=-0.421][2026-03-26 12:47:20] Step: 3916, Training Logs: loss_final: 0.512036, loss_mean: 0.851528, proj_loss: -0.417337, loss_mean_cls: 0.077845, grad_norm: 0.668811 +Steps: 0%| | 3917/1000000 [16:16<68:23:11, 4.05it/s, grad_norm=0.669, loss_final=0.512, loss_mean=0.852, loss_mean_cls=0.0778, proj_loss=-0.417][2026-03-26 12:47:20] Step: 3917, Training Logs: loss_final: 0.506006, loss_mean: 0.838115, proj_loss: -0.410758, loss_mean_cls: 0.078649, grad_norm: 1.051877 +Steps: 0%| | 3918/1000000 [16:17<68:25:21, 4.04it/s, grad_norm=1.05, loss_final=0.506, loss_mean=0.838, loss_mean_cls=0.0786, proj_loss=-0.411][2026-03-26 12:47:20] Step: 3918, Training Logs: loss_final: 0.495477, loss_mean: 0.823412, proj_loss: -0.410184, loss_mean_cls: 0.082249, grad_norm: 0.380793 +Steps: 0%| | 3919/1000000 [16:17<68:23:32, 4.05it/s, grad_norm=0.381, loss_final=0.495, loss_mean=0.823, loss_mean_cls=0.0822, proj_loss=-0.41][2026-03-26 12:47:20] Step: 3919, Training Logs: loss_final: 0.507502, loss_mean: 0.835292, proj_loss: -0.408402, loss_mean_cls: 0.080612, grad_norm: 0.535653 +Steps: 0%| | 3920/1000000 [16:17<68:24:34, 4.04it/s, grad_norm=0.536, loss_final=0.508, loss_mean=0.835, loss_mean_cls=0.0806, proj_loss=-0.408][2026-03-26 12:47:21] Step: 3920, Training Logs: loss_final: 0.503428, loss_mean: 0.833691, proj_loss: -0.410563, loss_mean_cls: 0.080301, grad_norm: 0.403923 +Steps: 0%| | 3921/1000000 [16:17<68:21:28, 4.05it/s, grad_norm=0.404, loss_final=0.503, loss_mean=0.834, loss_mean_cls=0.0803, proj_loss=-0.411][2026-03-26 12:47:21] Step: 3921, Training Logs: loss_final: 0.507813, loss_mean: 0.828920, proj_loss: -0.402118, loss_mean_cls: 0.081010, grad_norm: 0.455543 +Steps: 0%| | 3922/1000000 [16:18<68:23:56, 4.05it/s, grad_norm=0.456, loss_final=0.508, loss_mean=0.829, loss_mean_cls=0.081, proj_loss=-0.402][2026-03-26 12:47:21] Step: 3922, Training Logs: loss_final: 0.529128, loss_mean: 0.857707, proj_loss: -0.408800, loss_mean_cls: 0.080221, grad_norm: 0.668409 +Steps: 0%| | 3923/1000000 [16:18<68:23:55, 4.05it/s, grad_norm=0.668, loss_final=0.529, loss_mean=0.858, loss_mean_cls=0.0802, proj_loss=-0.409][2026-03-26 12:47:21] Step: 3923, Training Logs: loss_final: 0.511961, loss_mean: 0.844866, proj_loss: -0.413368, loss_mean_cls: 0.080464, grad_norm: 0.475415 +Steps: 0%| | 3924/1000000 [16:18<68:23:50, 4.05it/s, grad_norm=0.475, loss_final=0.512, loss_mean=0.845, loss_mean_cls=0.0805, proj_loss=-0.413][2026-03-26 12:47:22] Step: 3924, Training Logs: loss_final: 0.496819, loss_mean: 0.827342, proj_loss: -0.410966, loss_mean_cls: 0.080443, grad_norm: 0.663759 +Steps: 0%| | 3925/1000000 [16:18<68:26:17, 4.04it/s, grad_norm=0.664, loss_final=0.497, loss_mean=0.827, loss_mean_cls=0.0804, proj_loss=-0.411][2026-03-26 12:47:22] Step: 3925, Training Logs: loss_final: 0.511609, loss_mean: 0.848319, proj_loss: -0.414821, loss_mean_cls: 0.078111, grad_norm: 0.479111 +Steps: 0%| | 3926/1000000 [16:19<68:26:29, 4.04it/s, grad_norm=0.479, loss_final=0.512, loss_mean=0.848, loss_mean_cls=0.0781, proj_loss=-0.415][2026-03-26 12:47:22] Step: 3926, Training Logs: loss_final: 0.500541, loss_mean: 0.833560, proj_loss: -0.414010, loss_mean_cls: 0.080991, grad_norm: 0.549674 +Steps: 0%| | 3927/1000000 [16:19<68:26:51, 4.04it/s, grad_norm=0.55, loss_final=0.501, loss_mean=0.834, loss_mean_cls=0.081, proj_loss=-0.414][2026-03-26 12:47:22] Step: 3927, Training Logs: loss_final: 0.515562, loss_mean: 0.834168, proj_loss: -0.403026, loss_mean_cls: 0.084421, grad_norm: 0.446564 +Steps: 0%| | 3928/1000000 [16:19<68:26:20, 4.04it/s, grad_norm=0.447, loss_final=0.516, loss_mean=0.834, loss_mean_cls=0.0844, proj_loss=-0.403][2026-03-26 12:47:23] Step: 3928, Training Logs: loss_final: 0.521511, loss_mean: 0.861694, proj_loss: -0.417834, loss_mean_cls: 0.077651, grad_norm: 0.393755 +Steps: 0%| | 3929/1000000 [16:19<68:25:39, 4.04it/s, grad_norm=0.394, loss_final=0.522, loss_mean=0.862, loss_mean_cls=0.0777, proj_loss=-0.418][2026-03-26 12:47:23] Step: 3929, Training Logs: loss_final: 0.507112, loss_mean: 0.825888, proj_loss: -0.401818, loss_mean_cls: 0.083042, grad_norm: 0.779615 +Steps: 0%| | 3930/1000000 [16:20<68:27:19, 4.04it/s, grad_norm=0.78, loss_final=0.507, loss_mean=0.826, loss_mean_cls=0.083, proj_loss=-0.402][2026-03-26 12:47:23] Step: 3930, Training Logs: loss_final: 0.510427, loss_mean: 0.840750, proj_loss: -0.409734, loss_mean_cls: 0.079410, grad_norm: 0.459519 +Steps: 0%| | 3931/1000000 [16:20<68:25:49, 4.04it/s, grad_norm=0.46, loss_final=0.51, loss_mean=0.841, loss_mean_cls=0.0794, proj_loss=-0.41][2026-03-26 12:47:23] Step: 3931, Training Logs: loss_final: 0.503262, loss_mean: 0.833194, proj_loss: -0.411049, loss_mean_cls: 0.081117, grad_norm: 0.667912 +Steps: 0%| | 3932/1000000 [16:20<68:24:33, 4.04it/s, grad_norm=0.668, loss_final=0.503, loss_mean=0.833, loss_mean_cls=0.0811, proj_loss=-0.411][2026-03-26 12:47:24] Step: 3932, Training Logs: loss_final: 0.508940, loss_mean: 0.840610, proj_loss: -0.412720, loss_mean_cls: 0.081050, grad_norm: 0.561638 +Steps: 0%| | 3933/1000000 [16:20<68:28:08, 4.04it/s, grad_norm=0.562, loss_final=0.509, loss_mean=0.841, loss_mean_cls=0.081, proj_loss=-0.413][2026-03-26 12:47:24] Step: 3933, Training Logs: loss_final: 0.481482, loss_mean: 0.795874, proj_loss: -0.399632, loss_mean_cls: 0.085240, grad_norm: 0.432007 +Steps: 0%| | 3934/1000000 [16:21<68:25:46, 4.04it/s, grad_norm=0.432, loss_final=0.481, loss_mean=0.796, loss_mean_cls=0.0852, proj_loss=-0.4][2026-03-26 12:47:24] Step: 3934, Training Logs: loss_final: 0.496858, loss_mean: 0.824354, proj_loss: -0.409047, loss_mean_cls: 0.081551, grad_norm: 0.462454 +Steps: 0%| | 3935/1000000 [16:21<68:25:30, 4.04it/s, grad_norm=0.462, loss_final=0.497, loss_mean=0.824, loss_mean_cls=0.0816, proj_loss=-0.409][2026-03-26 12:47:24] Step: 3935, Training Logs: loss_final: 0.510856, loss_mean: 0.839948, proj_loss: -0.409531, loss_mean_cls: 0.080440, grad_norm: 0.444725 +Steps: 0%| | 3936/1000000 [16:21<68:23:25, 4.05it/s, grad_norm=0.445, loss_final=0.511, loss_mean=0.84, loss_mean_cls=0.0804, proj_loss=-0.41][2026-03-26 12:47:25] Step: 3936, Training Logs: loss_final: 0.510280, loss_mean: 0.839326, proj_loss: -0.408883, loss_mean_cls: 0.079837, grad_norm: 0.646522 +Steps: 0%| | 3937/1000000 [16:21<68:26:18, 4.04it/s, grad_norm=0.647, loss_final=0.51, loss_mean=0.839, loss_mean_cls=0.0798, proj_loss=-0.409][2026-03-26 12:47:25] Step: 3937, Training Logs: loss_final: 0.512709, loss_mean: 0.843486, proj_loss: -0.409704, loss_mean_cls: 0.078927, grad_norm: 0.406487 +Steps: 0%| | 3938/1000000 [16:22<68:23:40, 4.05it/s, grad_norm=0.406, loss_final=0.513, loss_mean=0.843, loss_mean_cls=0.0789, proj_loss=-0.41][2026-03-26 12:47:25] Step: 3938, Training Logs: loss_final: 0.510625, loss_mean: 0.840133, proj_loss: -0.409408, loss_mean_cls: 0.079900, grad_norm: 0.638141 +Steps: 0%| | 3939/1000000 [16:22<68:24:45, 4.04it/s, grad_norm=0.638, loss_final=0.511, loss_mean=0.84, loss_mean_cls=0.0799, proj_loss=-0.409][2026-03-26 12:47:25] Step: 3939, Training Logs: loss_final: 0.510918, loss_mean: 0.842591, proj_loss: -0.411556, loss_mean_cls: 0.079882, grad_norm: 0.576133 +Steps: 0%| | 3940/1000000 [16:22<68:25:24, 4.04it/s, grad_norm=0.576, loss_final=0.511, loss_mean=0.843, loss_mean_cls=0.0799, proj_loss=-0.412][2026-03-26 12:47:26] Step: 3940, Training Logs: loss_final: 0.516544, loss_mean: 0.839809, proj_loss: -0.405260, loss_mean_cls: 0.081995, grad_norm: 0.497250 +Steps: 0%| | 3941/1000000 [16:22<68:27:25, 4.04it/s, grad_norm=0.497, loss_final=0.517, loss_mean=0.84, loss_mean_cls=0.082, proj_loss=-0.405][2026-03-26 12:47:26] Step: 3941, Training Logs: loss_final: 0.499180, loss_mean: 0.832828, proj_loss: -0.412863, loss_mean_cls: 0.079215, grad_norm: 0.462373 +Steps: 0%| | 3942/1000000 [16:23<68:26:05, 4.04it/s, grad_norm=0.462, loss_final=0.499, loss_mean=0.833, loss_mean_cls=0.0792, proj_loss=-0.413][2026-03-26 12:47:26] Step: 3942, Training Logs: loss_final: 0.498086, loss_mean: 0.827731, proj_loss: -0.410487, loss_mean_cls: 0.080841, grad_norm: 0.297282 +Steps: 0%| | 3943/1000000 [16:23<68:25:25, 4.04it/s, grad_norm=0.297, loss_final=0.498, loss_mean=0.828, loss_mean_cls=0.0808, proj_loss=-0.41][2026-03-26 12:47:26] Step: 3943, Training Logs: loss_final: 0.516440, loss_mean: 0.840046, proj_loss: -0.405027, loss_mean_cls: 0.081422, grad_norm: 0.393010 +Steps: 0%| | 3944/1000000 [16:23<68:27:03, 4.04it/s, grad_norm=0.393, loss_final=0.516, loss_mean=0.84, loss_mean_cls=0.0814, proj_loss=-0.405][2026-03-26 12:47:27] Step: 3944, Training Logs: loss_final: 0.512172, loss_mean: 0.826741, proj_loss: -0.398982, loss_mean_cls: 0.084413, grad_norm: 0.435282 +Steps: 0%| | 3945/1000000 [16:23<68:26:21, 4.04it/s, grad_norm=0.435, loss_final=0.512, loss_mean=0.827, loss_mean_cls=0.0844, proj_loss=-0.399][2026-03-26 12:47:27] Step: 3945, Training Logs: loss_final: 0.521050, loss_mean: 0.851780, proj_loss: -0.409880, loss_mean_cls: 0.079150, grad_norm: 0.303329 +Steps: 0%| | 3946/1000000 [16:24<68:25:41, 4.04it/s, grad_norm=0.303, loss_final=0.521, loss_mean=0.852, loss_mean_cls=0.0791, proj_loss=-0.41][2026-03-26 12:47:27] Step: 3946, Training Logs: loss_final: 0.505670, loss_mean: 0.837100, proj_loss: -0.410471, loss_mean_cls: 0.079042, grad_norm: 0.337745 +Steps: 0%| | 3947/1000000 [16:24<68:29:01, 4.04it/s, grad_norm=0.338, loss_final=0.506, loss_mean=0.837, loss_mean_cls=0.079, proj_loss=-0.41][2026-03-26 12:47:27] Step: 3947, Training Logs: loss_final: 0.504843, loss_mean: 0.824705, proj_loss: -0.402640, loss_mean_cls: 0.082778, grad_norm: 0.534886 +Steps: 0%| | 3948/1000000 [16:24<68:31:42, 4.04it/s, grad_norm=0.535, loss_final=0.505, loss_mean=0.825, loss_mean_cls=0.0828, proj_loss=-0.403][2026-03-26 12:47:28] Step: 3948, Training Logs: loss_final: 0.481641, loss_mean: 0.807443, proj_loss: -0.406688, loss_mean_cls: 0.080886, grad_norm: 0.366532 +Steps: 0%| | 3949/1000000 [16:24<68:29:39, 4.04it/s, grad_norm=0.367, loss_final=0.482, loss_mean=0.807, loss_mean_cls=0.0809, proj_loss=-0.407][2026-03-26 12:47:28] Step: 3949, Training Logs: loss_final: 0.502972, loss_mean: 0.838835, proj_loss: -0.415582, loss_mean_cls: 0.079718, grad_norm: 0.633194 +Steps: 0%| | 3950/1000000 [16:25<68:28:42, 4.04it/s, grad_norm=0.633, loss_final=0.503, loss_mean=0.839, loss_mean_cls=0.0797, proj_loss=-0.416][2026-03-26 12:47:28] Step: 3950, Training Logs: loss_final: 0.499229, loss_mean: 0.835559, proj_loss: -0.416944, loss_mean_cls: 0.080614, grad_norm: 0.528566 +Steps: 0%| | 3951/1000000 [16:25<68:26:55, 4.04it/s, grad_norm=0.529, loss_final=0.499, loss_mean=0.836, loss_mean_cls=0.0806, proj_loss=-0.417][2026-03-26 12:47:28] Step: 3951, Training Logs: loss_final: 0.504599, loss_mean: 0.829110, proj_loss: -0.405625, loss_mean_cls: 0.081113, grad_norm: 0.366582 +Steps: 0%| | 3952/1000000 [16:25<68:27:42, 4.04it/s, grad_norm=0.367, loss_final=0.505, loss_mean=0.829, loss_mean_cls=0.0811, proj_loss=-0.406][2026-03-26 12:47:29] Step: 3952, Training Logs: loss_final: 0.508905, loss_mean: 0.847495, proj_loss: -0.415558, loss_mean_cls: 0.076969, grad_norm: 0.428007 +Steps: 0%| | 3953/1000000 [16:25<68:24:59, 4.04it/s, grad_norm=0.428, loss_final=0.509, loss_mean=0.847, loss_mean_cls=0.077, proj_loss=-0.416][2026-03-26 12:47:29] Step: 3953, Training Logs: loss_final: 0.502267, loss_mean: 0.829211, proj_loss: -0.408766, loss_mean_cls: 0.081822, grad_norm: 0.430562 +Steps: 0%| | 3954/1000000 [16:26<68:24:01, 4.04it/s, grad_norm=0.431, loss_final=0.502, loss_mean=0.829, loss_mean_cls=0.0818, proj_loss=-0.409][2026-03-26 12:47:29] Step: 3954, Training Logs: loss_final: 0.511938, loss_mean: 0.834841, proj_loss: -0.405565, loss_mean_cls: 0.082663, grad_norm: 0.552462 +Steps: 0%| | 3955/1000000 [16:26<68:21:19, 4.05it/s, grad_norm=0.552, loss_final=0.512, loss_mean=0.835, loss_mean_cls=0.0827, proj_loss=-0.406][2026-03-26 12:47:29] Step: 3955, Training Logs: loss_final: 0.483663, loss_mean: 0.812012, proj_loss: -0.410292, loss_mean_cls: 0.081944, grad_norm: 0.305913 +Steps: 0%| | 3956/1000000 [16:26<68:22:26, 4.05it/s, grad_norm=0.306, loss_final=0.484, loss_mean=0.812, loss_mean_cls=0.0819, proj_loss=-0.41][2026-03-26 12:47:30] Step: 3956, Training Logs: loss_final: 0.502459, loss_mean: 0.829549, proj_loss: -0.408248, loss_mean_cls: 0.081158, grad_norm: 0.263589 +Steps: 0%| | 3957/1000000 [16:26<68:24:35, 4.04it/s, grad_norm=0.264, loss_final=0.502, loss_mean=0.83, loss_mean_cls=0.0812, proj_loss=-0.408][2026-03-26 12:47:30] Step: 3957, Training Logs: loss_final: 0.508639, loss_mean: 0.845827, proj_loss: -0.415412, loss_mean_cls: 0.078224, grad_norm: 0.422057 +Steps: 0%| | 3958/1000000 [16:27<68:24:52, 4.04it/s, grad_norm=0.422, loss_final=0.509, loss_mean=0.846, loss_mean_cls=0.0782, proj_loss=-0.415][2026-03-26 12:47:30] Step: 3958, Training Logs: loss_final: 0.503733, loss_mean: 0.828665, proj_loss: -0.406047, loss_mean_cls: 0.081116, grad_norm: 0.352198 +Steps: 0%| | 3959/1000000 [16:27<68:26:21, 4.04it/s, grad_norm=0.352, loss_final=0.504, loss_mean=0.829, loss_mean_cls=0.0811, proj_loss=-0.406][2026-03-26 12:47:30] Step: 3959, Training Logs: loss_final: 0.503005, loss_mean: 0.821242, proj_loss: -0.400986, loss_mean_cls: 0.082749, grad_norm: 0.354403 +Steps: 0%| | 3960/1000000 [16:27<68:30:30, 4.04it/s, grad_norm=0.354, loss_final=0.503, loss_mean=0.821, loss_mean_cls=0.0827, proj_loss=-0.401][2026-03-26 12:47:31] Step: 3960, Training Logs: loss_final: 0.502359, loss_mean: 0.829809, proj_loss: -0.407755, loss_mean_cls: 0.080305, grad_norm: 0.465649 +Steps: 0%| | 3961/1000000 [16:27<68:31:43, 4.04it/s, grad_norm=0.466, loss_final=0.502, loss_mean=0.83, loss_mean_cls=0.0803, proj_loss=-0.408][2026-03-26 12:47:31] Step: 3961, Training Logs: loss_final: 0.496576, loss_mean: 0.831991, proj_loss: -0.414802, loss_mean_cls: 0.079387, grad_norm: 0.342786 +Steps: 0%| | 3962/1000000 [16:28<68:28:38, 4.04it/s, grad_norm=0.343, loss_final=0.497, loss_mean=0.832, loss_mean_cls=0.0794, proj_loss=-0.415][2026-03-26 12:47:31] Step: 3962, Training Logs: loss_final: 0.499205, loss_mean: 0.829983, proj_loss: -0.411201, loss_mean_cls: 0.080424, grad_norm: 0.251304 +Steps: 0%| | 3963/1000000 [16:28<68:27:41, 4.04it/s, grad_norm=0.251, loss_final=0.499, loss_mean=0.83, loss_mean_cls=0.0804, proj_loss=-0.411][2026-03-26 12:47:31] Step: 3963, Training Logs: loss_final: 0.514944, loss_mean: 0.832126, proj_loss: -0.399830, loss_mean_cls: 0.082647, grad_norm: 0.319446 +Steps: 0%| | 3964/1000000 [16:28<69:14:42, 4.00it/s, grad_norm=0.319, loss_final=0.515, loss_mean=0.832, loss_mean_cls=0.0826, proj_loss=-0.4][2026-03-26 12:47:32] Step: 3964, Training Logs: loss_final: 0.517665, loss_mean: 0.849745, proj_loss: -0.412123, loss_mean_cls: 0.080042, grad_norm: 0.286572 +Steps: 0%| | 3965/1000000 [16:28<69:00:36, 4.01it/s, grad_norm=0.287, loss_final=0.518, loss_mean=0.85, loss_mean_cls=0.08, proj_loss=-0.412][2026-03-26 12:47:32] Step: 3965, Training Logs: loss_final: 0.494194, loss_mean: 0.804403, proj_loss: -0.394616, loss_mean_cls: 0.084407, grad_norm: 0.451749 +Steps: 0%| | 3966/1000000 [16:29<68:48:54, 4.02it/s, grad_norm=0.452, loss_final=0.494, loss_mean=0.804, loss_mean_cls=0.0844, proj_loss=-0.395][2026-03-26 12:47:32] Step: 3966, Training Logs: loss_final: 0.501841, loss_mean: 0.823689, proj_loss: -0.405236, loss_mean_cls: 0.083388, grad_norm: 0.363499 +Steps: 0%| | 3967/1000000 [16:29<68:43:04, 4.03it/s, grad_norm=0.363, loss_final=0.502, loss_mean=0.824, loss_mean_cls=0.0834, proj_loss=-0.405][2026-03-26 12:47:32] Step: 3967, Training Logs: loss_final: 0.499931, loss_mean: 0.835122, proj_loss: -0.416393, loss_mean_cls: 0.081202, grad_norm: 0.521720 +Steps: 0%| | 3968/1000000 [16:29<68:38:46, 4.03it/s, grad_norm=0.522, loss_final=0.5, loss_mean=0.835, loss_mean_cls=0.0812, proj_loss=-0.416][2026-03-26 12:47:33] Step: 3968, Training Logs: loss_final: 0.517379, loss_mean: 0.838138, proj_loss: -0.402950, loss_mean_cls: 0.082192, grad_norm: 0.482670 +Steps: 0%| | 3969/1000000 [16:29<68:33:06, 4.04it/s, grad_norm=0.483, loss_final=0.517, loss_mean=0.838, loss_mean_cls=0.0822, proj_loss=-0.403][2026-03-26 12:47:33] Step: 3969, Training Logs: loss_final: 0.510044, loss_mean: 0.847899, proj_loss: -0.416178, loss_mean_cls: 0.078323, grad_norm: 0.557123 +Steps: 0%| | 3970/1000000 [16:30<68:30:34, 4.04it/s, grad_norm=0.557, loss_final=0.51, loss_mean=0.848, loss_mean_cls=0.0783, proj_loss=-0.416][2026-03-26 12:47:33] Step: 3970, Training Logs: loss_final: 0.504601, loss_mean: 0.837276, proj_loss: -0.412226, loss_mean_cls: 0.079550, grad_norm: 0.437279 +Steps: 0%| | 3971/1000000 [16:30<68:27:16, 4.04it/s, grad_norm=0.437, loss_final=0.505, loss_mean=0.837, loss_mean_cls=0.0796, proj_loss=-0.412][2026-03-26 12:47:33] Step: 3971, Training Logs: loss_final: 0.497754, loss_mean: 0.829925, proj_loss: -0.413549, loss_mean_cls: 0.081379, grad_norm: 0.350852 +Steps: 0%| | 3972/1000000 [16:30<68:25:22, 4.04it/s, grad_norm=0.351, loss_final=0.498, loss_mean=0.83, loss_mean_cls=0.0814, proj_loss=-0.414][2026-03-26 12:47:33] Step: 3972, Training Logs: loss_final: 0.507895, loss_mean: 0.836184, proj_loss: -0.409754, loss_mean_cls: 0.081465, grad_norm: 0.470296 +Steps: 0%| | 3973/1000000 [16:30<68:24:50, 4.04it/s, grad_norm=0.47, loss_final=0.508, loss_mean=0.836, loss_mean_cls=0.0815, proj_loss=-0.41][2026-03-26 12:47:34] Step: 3973, Training Logs: loss_final: 0.506001, loss_mean: 0.834173, proj_loss: -0.409185, loss_mean_cls: 0.081014, grad_norm: 0.441754 +Steps: 0%| | 3974/1000000 [16:31<68:24:40, 4.04it/s, grad_norm=0.442, loss_final=0.506, loss_mean=0.834, loss_mean_cls=0.081, proj_loss=-0.409][2026-03-26 12:47:34] Step: 3974, Training Logs: loss_final: 0.503667, loss_mean: 0.836111, proj_loss: -0.413150, loss_mean_cls: 0.080705, grad_norm: 0.370420 +Steps: 0%| | 3975/1000000 [16:31<68:24:57, 4.04it/s, grad_norm=0.37, loss_final=0.504, loss_mean=0.836, loss_mean_cls=0.0807, proj_loss=-0.413][2026-03-26 12:47:34] Step: 3975, Training Logs: loss_final: 0.516289, loss_mean: 0.832538, proj_loss: -0.400700, loss_mean_cls: 0.084450, grad_norm: 0.426237 +Steps: 0%| | 3976/1000000 [16:31<68:23:48, 4.05it/s, grad_norm=0.426, loss_final=0.516, loss_mean=0.833, loss_mean_cls=0.0845, proj_loss=-0.401][2026-03-26 12:47:34] Step: 3976, Training Logs: loss_final: 0.501393, loss_mean: 0.832643, proj_loss: -0.411986, loss_mean_cls: 0.080737, grad_norm: 0.313001 +Steps: 0%| | 3977/1000000 [16:31<68:26:03, 4.04it/s, grad_norm=0.313, loss_final=0.501, loss_mean=0.833, loss_mean_cls=0.0807, proj_loss=-0.412][2026-03-26 12:47:35] Step: 3977, Training Logs: loss_final: 0.510623, loss_mean: 0.832219, proj_loss: -0.403421, loss_mean_cls: 0.081826, grad_norm: 0.543560 +Steps: 0%| | 3978/1000000 [16:32<68:24:10, 4.04it/s, grad_norm=0.544, loss_final=0.511, loss_mean=0.832, loss_mean_cls=0.0818, proj_loss=-0.403][2026-03-26 12:47:35] Step: 3978, Training Logs: loss_final: 0.508641, loss_mean: 0.841577, proj_loss: -0.411365, loss_mean_cls: 0.078429, grad_norm: 0.478482 +Steps: 0%| | 3979/1000000 [16:32<68:23:26, 4.05it/s, grad_norm=0.478, loss_final=0.509, loss_mean=0.842, loss_mean_cls=0.0784, proj_loss=-0.411][2026-03-26 12:47:35] Step: 3979, Training Logs: loss_final: 0.504520, loss_mean: 0.831515, proj_loss: -0.407963, loss_mean_cls: 0.080969, grad_norm: 0.291999 +Steps: 0%| | 3980/1000000 [16:32<68:24:43, 4.04it/s, grad_norm=0.292, loss_final=0.505, loss_mean=0.832, loss_mean_cls=0.081, proj_loss=-0.408][2026-03-26 12:47:35] Step: 3980, Training Logs: loss_final: 0.498610, loss_mean: 0.827113, proj_loss: -0.410287, loss_mean_cls: 0.081784, grad_norm: 0.526376 +Steps: 0%| | 3981/1000000 [16:32<68:25:14, 4.04it/s, grad_norm=0.526, loss_final=0.499, loss_mean=0.827, loss_mean_cls=0.0818, proj_loss=-0.41][2026-03-26 12:47:36] Step: 3981, Training Logs: loss_final: 0.509573, loss_mean: 0.833681, proj_loss: -0.406045, loss_mean_cls: 0.081937, grad_norm: 0.524719 +Steps: 0%| | 3982/1000000 [16:33<68:28:16, 4.04it/s, grad_norm=0.525, loss_final=0.51, loss_mean=0.834, loss_mean_cls=0.0819, proj_loss=-0.406][2026-03-26 12:47:36] Step: 3982, Training Logs: loss_final: 0.500523, loss_mean: 0.836851, proj_loss: -0.415069, loss_mean_cls: 0.078741, grad_norm: 0.346078 +Steps: 0%| | 3983/1000000 [16:33<68:28:13, 4.04it/s, grad_norm=0.346, loss_final=0.501, loss_mean=0.837, loss_mean_cls=0.0787, proj_loss=-0.415][2026-03-26 12:47:36] Step: 3983, Training Logs: loss_final: 0.498143, loss_mean: 0.825698, proj_loss: -0.408022, loss_mean_cls: 0.080467, grad_norm: 0.602880 +Steps: 0%| | 3984/1000000 [16:33<68:28:12, 4.04it/s, grad_norm=0.603, loss_final=0.498, loss_mean=0.826, loss_mean_cls=0.0805, proj_loss=-0.408][2026-03-26 12:47:36] Step: 3984, Training Logs: loss_final: 0.501140, loss_mean: 0.818670, proj_loss: -0.398730, loss_mean_cls: 0.081201, grad_norm: 0.444243 +Steps: 0%| | 3985/1000000 [16:33<68:27:09, 4.04it/s, grad_norm=0.444, loss_final=0.501, loss_mean=0.819, loss_mean_cls=0.0812, proj_loss=-0.399][2026-03-26 12:47:37] Step: 3985, Training Logs: loss_final: 0.512890, loss_mean: 0.851056, proj_loss: -0.416177, loss_mean_cls: 0.078010, grad_norm: 0.350948 +Steps: 0%| | 3986/1000000 [16:34<68:28:05, 4.04it/s, grad_norm=0.351, loss_final=0.513, loss_mean=0.851, loss_mean_cls=0.078, proj_loss=-0.416][2026-03-26 12:47:37] Step: 3986, Training Logs: loss_final: 0.501064, loss_mean: 0.828809, proj_loss: -0.409364, loss_mean_cls: 0.081619, grad_norm: 0.502226 +Steps: 0%| | 3987/1000000 [16:34<68:27:04, 4.04it/s, grad_norm=0.502, loss_final=0.501, loss_mean=0.829, loss_mean_cls=0.0816, proj_loss=-0.409][2026-03-26 12:47:37] Step: 3987, Training Logs: loss_final: 0.515515, loss_mean: 0.840585, proj_loss: -0.405427, loss_mean_cls: 0.080356, grad_norm: 0.400063 +Steps: 0%| | 3988/1000000 [16:34<68:24:27, 4.04it/s, grad_norm=0.4, loss_final=0.516, loss_mean=0.841, loss_mean_cls=0.0804, proj_loss=-0.405][2026-03-26 12:47:37] Step: 3988, Training Logs: loss_final: 0.497321, loss_mean: 0.827309, proj_loss: -0.411719, loss_mean_cls: 0.081731, grad_norm: 0.344217 +Steps: 0%| | 3989/1000000 [16:34<68:24:54, 4.04it/s, grad_norm=0.344, loss_final=0.497, loss_mean=0.827, loss_mean_cls=0.0817, proj_loss=-0.412][2026-03-26 12:47:38] Step: 3989, Training Logs: loss_final: 0.494532, loss_mean: 0.817945, proj_loss: -0.405974, loss_mean_cls: 0.082562, grad_norm: 0.550067 +Steps: 0%| | 3990/1000000 [16:35<68:26:51, 4.04it/s, grad_norm=0.55, loss_final=0.495, loss_mean=0.818, loss_mean_cls=0.0826, proj_loss=-0.406][2026-03-26 12:47:38] Step: 3990, Training Logs: loss_final: 0.497544, loss_mean: 0.822535, proj_loss: -0.407399, loss_mean_cls: 0.082408, grad_norm: 0.337474 +Steps: 0%| | 3991/1000000 [16:35<68:23:53, 4.04it/s, grad_norm=0.337, loss_final=0.498, loss_mean=0.823, loss_mean_cls=0.0824, proj_loss=-0.407][2026-03-26 12:47:38] Step: 3991, Training Logs: loss_final: 0.495437, loss_mean: 0.817095, proj_loss: -0.404970, loss_mean_cls: 0.083311, grad_norm: 0.389327 +Steps: 0%| | 3992/1000000 [16:35<68:25:14, 4.04it/s, grad_norm=0.389, loss_final=0.495, loss_mean=0.817, loss_mean_cls=0.0833, proj_loss=-0.405][2026-03-26 12:47:38] Step: 3992, Training Logs: loss_final: 0.507317, loss_mean: 0.842045, proj_loss: -0.413658, loss_mean_cls: 0.078930, grad_norm: 0.405941 +Steps: 0%| | 3993/1000000 [16:35<68:24:02, 4.04it/s, grad_norm=0.406, loss_final=0.507, loss_mean=0.842, loss_mean_cls=0.0789, proj_loss=-0.414][2026-03-26 12:47:39] Step: 3993, Training Logs: loss_final: 0.527053, loss_mean: 0.857356, proj_loss: -0.408738, loss_mean_cls: 0.078434, grad_norm: 0.379092 +Steps: 0%| | 3994/1000000 [16:36<68:26:43, 4.04it/s, grad_norm=0.379, loss_final=0.527, loss_mean=0.857, loss_mean_cls=0.0784, proj_loss=-0.409][2026-03-26 12:47:39] Step: 3994, Training Logs: loss_final: 0.502119, loss_mean: 0.827619, proj_loss: -0.409074, loss_mean_cls: 0.083574, grad_norm: 0.331766 +Steps: 0%| | 3995/1000000 [16:36<68:24:12, 4.04it/s, grad_norm=0.332, loss_final=0.502, loss_mean=0.828, loss_mean_cls=0.0836, proj_loss=-0.409][2026-03-26 12:47:39] Step: 3995, Training Logs: loss_final: 0.498350, loss_mean: 0.825568, proj_loss: -0.407748, loss_mean_cls: 0.080530, grad_norm: 0.292367 +Steps: 0%| | 3996/1000000 [16:36<68:23:21, 4.05it/s, grad_norm=0.292, loss_final=0.498, loss_mean=0.826, loss_mean_cls=0.0805, proj_loss=-0.408][2026-03-26 12:47:39] Step: 3996, Training Logs: loss_final: 0.520573, loss_mean: 0.856839, proj_loss: -0.416663, loss_mean_cls: 0.080397, grad_norm: 0.338860 +Steps: 0%| | 3997/1000000 [16:36<68:22:34, 4.05it/s, grad_norm=0.339, loss_final=0.521, loss_mean=0.857, loss_mean_cls=0.0804, proj_loss=-0.417][2026-03-26 12:47:40] Step: 3997, Training Logs: loss_final: 0.496344, loss_mean: 0.818587, proj_loss: -0.403925, loss_mean_cls: 0.081682, grad_norm: 0.389735 +Steps: 0%| | 3998/1000000 [16:37<68:22:46, 4.05it/s, grad_norm=0.39, loss_final=0.496, loss_mean=0.819, loss_mean_cls=0.0817, proj_loss=-0.404][2026-03-26 12:47:40] Step: 3998, Training Logs: loss_final: 0.507982, loss_mean: 0.829688, proj_loss: -0.404074, loss_mean_cls: 0.082368, grad_norm: 0.350100 +Steps: 0%| | 3999/1000000 [16:37<68:22:37, 4.05it/s, grad_norm=0.35, loss_final=0.508, loss_mean=0.83, loss_mean_cls=0.0824, proj_loss=-0.404][2026-03-26 12:47:40] Step: 3999, Training Logs: loss_final: 0.486555, loss_mean: 0.819014, proj_loss: -0.413477, loss_mean_cls: 0.081018, grad_norm: 0.331208 +Steps: 0%| | 4000/1000000 [16:37<68:25:11, 4.04it/s, grad_norm=0.331, loss_final=0.487, loss_mean=0.819, loss_mean_cls=0.081, proj_loss=-0.413][2026-03-26 12:47:40] Step: 4000, Training Logs: loss_final: 0.519185, loss_mean: 0.846284, proj_loss: -0.408433, loss_mean_cls: 0.081334, grad_norm: 0.343679 +Steps: 0%| | 4001/1000000 [16:37<68:26:51, 4.04it/s, grad_norm=0.344, loss_final=0.519, loss_mean=0.846, loss_mean_cls=0.0813, proj_loss=-0.408][2026-03-26 12:47:41] Step: 4001, Training Logs: loss_final: 0.523934, loss_mean: 0.851945, proj_loss: -0.407032, loss_mean_cls: 0.079021, grad_norm: 0.304440 +Steps: 0%| | 4002/1000000 [16:38<68:23:36, 4.05it/s, grad_norm=0.304, loss_final=0.524, loss_mean=0.852, loss_mean_cls=0.079, proj_loss=-0.407][2026-03-26 12:47:41] Step: 4002, Training Logs: loss_final: 0.508669, loss_mean: 0.838055, proj_loss: -0.409829, loss_mean_cls: 0.080443, grad_norm: 0.300771 +Steps: 0%| | 4003/1000000 [16:38<68:23:57, 4.04it/s, grad_norm=0.301, loss_final=0.509, loss_mean=0.838, loss_mean_cls=0.0804, proj_loss=-0.41][2026-03-26 12:47:41] Step: 4003, Training Logs: loss_final: 0.506791, loss_mean: 0.842204, proj_loss: -0.414601, loss_mean_cls: 0.079189, grad_norm: 0.301251 +Steps: 0%| | 4004/1000000 [16:38<68:25:40, 4.04it/s, grad_norm=0.301, loss_final=0.507, loss_mean=0.842, loss_mean_cls=0.0792, proj_loss=-0.415][2026-03-26 12:47:41] Step: 4004, Training Logs: loss_final: 0.493940, loss_mean: 0.826012, proj_loss: -0.412890, loss_mean_cls: 0.080818, grad_norm: 0.271173 +Steps: 0%| | 4005/1000000 [16:38<68:27:15, 4.04it/s, grad_norm=0.271, loss_final=0.494, loss_mean=0.826, loss_mean_cls=0.0808, proj_loss=-0.413][2026-03-26 12:47:42] Step: 4005, Training Logs: loss_final: 0.509698, loss_mean: 0.842834, proj_loss: -0.413308, loss_mean_cls: 0.080173, grad_norm: 0.325222 +Steps: 0%| | 4006/1000000 [16:39<68:27:41, 4.04it/s, grad_norm=0.325, loss_final=0.51, loss_mean=0.843, loss_mean_cls=0.0802, proj_loss=-0.413][2026-03-26 12:47:42] Step: 4006, Training Logs: loss_final: 0.498120, loss_mean: 0.819734, proj_loss: -0.404748, loss_mean_cls: 0.083134, grad_norm: 0.285428 +Steps: 0%| | 4007/1000000 [16:39<68:26:13, 4.04it/s, grad_norm=0.285, loss_final=0.498, loss_mean=0.82, loss_mean_cls=0.0831, proj_loss=-0.405][2026-03-26 12:47:42] Step: 4007, Training Logs: loss_final: 0.508109, loss_mean: 0.831581, proj_loss: -0.404689, loss_mean_cls: 0.081217, grad_norm: 0.257355 +Steps: 0%| | 4008/1000000 [16:39<68:27:31, 4.04it/s, grad_norm=0.257, loss_final=0.508, loss_mean=0.832, loss_mean_cls=0.0812, proj_loss=-0.405][2026-03-26 12:47:42] Step: 4008, Training Logs: loss_final: 0.512697, loss_mean: 0.837021, proj_loss: -0.404468, loss_mean_cls: 0.080145, grad_norm: 0.320409 +Steps: 0%| | 4009/1000000 [16:39<68:29:39, 4.04it/s, grad_norm=0.32, loss_final=0.513, loss_mean=0.837, loss_mean_cls=0.0801, proj_loss=-0.404][2026-03-26 12:47:43] Step: 4009, Training Logs: loss_final: 0.491024, loss_mean: 0.819381, proj_loss: -0.410987, loss_mean_cls: 0.082631, grad_norm: 0.321243 +Steps: 0%| | 4010/1000000 [16:40<68:30:15, 4.04it/s, grad_norm=0.321, loss_final=0.491, loss_mean=0.819, loss_mean_cls=0.0826, proj_loss=-0.411][2026-03-26 12:47:43] Step: 4010, Training Logs: loss_final: 0.503840, loss_mean: 0.820396, proj_loss: -0.400308, loss_mean_cls: 0.083753, grad_norm: 0.369984 +Steps: 0%| | 4011/1000000 [16:40<68:34:31, 4.03it/s, grad_norm=0.37, loss_final=0.504, loss_mean=0.82, loss_mean_cls=0.0838, proj_loss=-0.4][2026-03-26 12:47:43] Step: 4011, Training Logs: loss_final: 0.507234, loss_mean: 0.835052, proj_loss: -0.408557, loss_mean_cls: 0.080739, grad_norm: 0.346542 +Steps: 0%| | 4012/1000000 [16:40<68:22:08, 4.05it/s, grad_norm=0.347, loss_final=0.507, loss_mean=0.835, loss_mean_cls=0.0807, proj_loss=-0.409][2026-03-26 12:47:43] Step: 4012, Training Logs: loss_final: 0.516938, loss_mean: 0.830324, proj_loss: -0.397628, loss_mean_cls: 0.084243, grad_norm: 0.376311 +Steps: 0%| | 4013/1000000 [16:40<68:24:01, 4.04it/s, grad_norm=0.376, loss_final=0.517, loss_mean=0.83, loss_mean_cls=0.0842, proj_loss=-0.398][2026-03-26 12:47:44] Step: 4013, Training Logs: loss_final: 0.513145, loss_mean: 0.848390, proj_loss: -0.414367, loss_mean_cls: 0.079123, grad_norm: 0.418628 +Steps: 0%| | 4014/1000000 [16:40<68:23:55, 4.04it/s, grad_norm=0.419, loss_final=0.513, loss_mean=0.848, loss_mean_cls=0.0791, proj_loss=-0.414][2026-03-26 12:47:44] Step: 4014, Training Logs: loss_final: 0.506066, loss_mean: 0.823162, proj_loss: -0.400912, loss_mean_cls: 0.083816, grad_norm: 0.563970 +Steps: 0%| | 4015/1000000 [16:41<68:26:25, 4.04it/s, grad_norm=0.564, loss_final=0.506, loss_mean=0.823, loss_mean_cls=0.0838, proj_loss=-0.401][2026-03-26 12:47:44] Step: 4015, Training Logs: loss_final: 0.495180, loss_mean: 0.826441, proj_loss: -0.413308, loss_mean_cls: 0.082047, grad_norm: 0.425919 +Steps: 0%| | 4016/1000000 [16:41<68:25:53, 4.04it/s, grad_norm=0.426, loss_final=0.495, loss_mean=0.826, loss_mean_cls=0.082, proj_loss=-0.413][2026-03-26 12:47:44] Step: 4016, Training Logs: loss_final: 0.515114, loss_mean: 0.845058, proj_loss: -0.409308, loss_mean_cls: 0.079364, grad_norm: 0.360359 +Steps: 0%| | 4017/1000000 [16:41<68:27:38, 4.04it/s, grad_norm=0.36, loss_final=0.515, loss_mean=0.845, loss_mean_cls=0.0794, proj_loss=-0.409][2026-03-26 12:47:45] Step: 4017, Training Logs: loss_final: 0.504588, loss_mean: 0.819750, proj_loss: -0.398879, loss_mean_cls: 0.083716, grad_norm: 0.383941 +Steps: 0%| | 4018/1000000 [16:41<68:29:15, 4.04it/s, grad_norm=0.384, loss_final=0.505, loss_mean=0.82, loss_mean_cls=0.0837, proj_loss=-0.399][2026-03-26 12:47:45] Step: 4018, Training Logs: loss_final: 0.522634, loss_mean: 0.865380, proj_loss: -0.419021, loss_mean_cls: 0.076275, grad_norm: 0.377405 +Steps: 0%| | 4019/1000000 [16:42<68:29:51, 4.04it/s, grad_norm=0.377, loss_final=0.523, loss_mean=0.865, loss_mean_cls=0.0763, proj_loss=-0.419][2026-03-26 12:47:45] Step: 4019, Training Logs: loss_final: 0.499930, loss_mean: 0.826208, proj_loss: -0.408809, loss_mean_cls: 0.082530, grad_norm: 0.354958 +Steps: 0%| | 4020/1000000 [16:42<68:30:15, 4.04it/s, grad_norm=0.355, loss_final=0.5, loss_mean=0.826, loss_mean_cls=0.0825, proj_loss=-0.409][2026-03-26 12:47:45] Step: 4020, Training Logs: loss_final: 0.507797, loss_mean: 0.843437, proj_loss: -0.414853, loss_mean_cls: 0.079212, grad_norm: 0.235806 +Steps: 0%| | 4021/1000000 [16:42<68:29:40, 4.04it/s, grad_norm=0.236, loss_final=0.508, loss_mean=0.843, loss_mean_cls=0.0792, proj_loss=-0.415][2026-03-26 12:47:46] Step: 4021, Training Logs: loss_final: 0.505600, loss_mean: 0.845837, proj_loss: -0.418651, loss_mean_cls: 0.078414, grad_norm: 0.336938 +Steps: 0%| | 4022/1000000 [16:42<68:29:21, 4.04it/s, grad_norm=0.337, loss_final=0.506, loss_mean=0.846, loss_mean_cls=0.0784, proj_loss=-0.419][2026-03-26 12:47:46] Step: 4022, Training Logs: loss_final: 0.499881, loss_mean: 0.827782, proj_loss: -0.408393, loss_mean_cls: 0.080493, grad_norm: 0.417015 +Steps: 0%| | 4023/1000000 [16:43<68:28:50, 4.04it/s, grad_norm=0.417, loss_final=0.5, loss_mean=0.828, loss_mean_cls=0.0805, proj_loss=-0.408][2026-03-26 12:47:46] Step: 4023, Training Logs: loss_final: 0.509531, loss_mean: 0.839255, proj_loss: -0.410416, loss_mean_cls: 0.080692, grad_norm: 0.339247 +Steps: 0%| | 4024/1000000 [16:43<68:28:12, 4.04it/s, grad_norm=0.339, loss_final=0.51, loss_mean=0.839, loss_mean_cls=0.0807, proj_loss=-0.41][2026-03-26 12:47:46] Step: 4024, Training Logs: loss_final: 0.488946, loss_mean: 0.823322, proj_loss: -0.414615, loss_mean_cls: 0.080238, grad_norm: 0.368017 +Steps: 0%| | 4025/1000000 [16:43<68:26:44, 4.04it/s, grad_norm=0.368, loss_final=0.489, loss_mean=0.823, loss_mean_cls=0.0802, proj_loss=-0.415][2026-03-26 12:47:47] Step: 4025, Training Logs: loss_final: 0.506182, loss_mean: 0.832546, proj_loss: -0.407749, loss_mean_cls: 0.081386, grad_norm: 0.393794 +Steps: 0%| | 4026/1000000 [16:43<68:29:22, 4.04it/s, grad_norm=0.394, loss_final=0.506, loss_mean=0.833, loss_mean_cls=0.0814, proj_loss=-0.408][2026-03-26 12:47:47] Step: 4026, Training Logs: loss_final: 0.501607, loss_mean: 0.837116, proj_loss: -0.414807, loss_mean_cls: 0.079297, grad_norm: 0.295384 +Steps: 0%| | 4027/1000000 [16:44<68:27:00, 4.04it/s, grad_norm=0.295, loss_final=0.502, loss_mean=0.837, loss_mean_cls=0.0793, proj_loss=-0.415][2026-03-26 12:47:47] Step: 4027, Training Logs: loss_final: 0.510870, loss_mean: 0.853889, proj_loss: -0.420474, loss_mean_cls: 0.077455, grad_norm: 0.397995 +Steps: 0%| | 4028/1000000 [16:44<68:26:16, 4.04it/s, grad_norm=0.398, loss_final=0.511, loss_mean=0.854, loss_mean_cls=0.0775, proj_loss=-0.42][2026-03-26 12:47:47] Step: 4028, Training Logs: loss_final: 0.513616, loss_mean: 0.839800, proj_loss: -0.405713, loss_mean_cls: 0.079530, grad_norm: 0.626983 +Steps: 0%| | 4029/1000000 [16:44<68:23:11, 4.05it/s, grad_norm=0.627, loss_final=0.514, loss_mean=0.84, loss_mean_cls=0.0795, proj_loss=-0.406][2026-03-26 12:47:48] Step: 4029, Training Logs: loss_final: 0.496012, loss_mean: 0.825234, proj_loss: -0.411403, loss_mean_cls: 0.082181, grad_norm: 0.372393 +Steps: 0%| | 4030/1000000 [16:44<68:27:32, 4.04it/s, grad_norm=0.372, loss_final=0.496, loss_mean=0.825, loss_mean_cls=0.0822, proj_loss=-0.411][2026-03-26 12:47:48] Step: 4030, Training Logs: loss_final: 0.492240, loss_mean: 0.803175, proj_loss: -0.396705, loss_mean_cls: 0.085769, grad_norm: 0.530455 +Steps: 0%| | 4031/1000000 [16:45<68:25:29, 4.04it/s, grad_norm=0.53, loss_final=0.492, loss_mean=0.803, loss_mean_cls=0.0858, proj_loss=-0.397][2026-03-26 12:47:48] Step: 4031, Training Logs: loss_final: 0.504578, loss_mean: 0.842732, proj_loss: -0.416696, loss_mean_cls: 0.078542, grad_norm: 0.648980 +Steps: 0%| | 4032/1000000 [16:45<68:27:52, 4.04it/s, grad_norm=0.649, loss_final=0.505, loss_mean=0.843, loss_mean_cls=0.0785, proj_loss=-0.417][2026-03-26 12:47:48] Step: 4032, Training Logs: loss_final: 0.492752, loss_mean: 0.818968, proj_loss: -0.406675, loss_mean_cls: 0.080459, grad_norm: 0.443799 +Steps: 0%| | 4033/1000000 [16:45<68:26:02, 4.04it/s, grad_norm=0.444, loss_final=0.493, loss_mean=0.819, loss_mean_cls=0.0805, proj_loss=-0.407][2026-03-26 12:47:49] Step: 4033, Training Logs: loss_final: 0.496596, loss_mean: 0.816464, proj_loss: -0.402521, loss_mean_cls: 0.082653, grad_norm: 0.311773 +Steps: 0%| | 4034/1000000 [16:45<68:26:26, 4.04it/s, grad_norm=0.312, loss_final=0.497, loss_mean=0.816, loss_mean_cls=0.0827, proj_loss=-0.403][2026-03-26 12:47:49] Step: 4034, Training Logs: loss_final: 0.497337, loss_mean: 0.833739, proj_loss: -0.416601, loss_mean_cls: 0.080199, grad_norm: 0.423933 +Steps: 0%| | 4035/1000000 [16:46<68:25:29, 4.04it/s, grad_norm=0.424, loss_final=0.497, loss_mean=0.834, loss_mean_cls=0.0802, proj_loss=-0.417][2026-03-26 12:47:49] Step: 4035, Training Logs: loss_final: 0.503358, loss_mean: 0.833775, proj_loss: -0.410307, loss_mean_cls: 0.079890, grad_norm: 0.408817 +Steps: 0%| | 4036/1000000 [16:46<68:24:36, 4.04it/s, grad_norm=0.409, loss_final=0.503, loss_mean=0.834, loss_mean_cls=0.0799, proj_loss=-0.41][2026-03-26 12:47:49] Step: 4036, Training Logs: loss_final: 0.505022, loss_mean: 0.818957, proj_loss: -0.398991, loss_mean_cls: 0.085056, grad_norm: 0.445026 +Steps: 0%| | 4037/1000000 [16:46<68:22:09, 4.05it/s, grad_norm=0.445, loss_final=0.505, loss_mean=0.819, loss_mean_cls=0.0851, proj_loss=-0.399][2026-03-26 12:47:50] Step: 4037, Training Logs: loss_final: 0.496499, loss_mean: 0.832808, proj_loss: -0.416105, loss_mean_cls: 0.079796, grad_norm: 0.435641 +Steps: 0%| | 4038/1000000 [16:46<68:24:17, 4.04it/s, grad_norm=0.436, loss_final=0.496, loss_mean=0.833, loss_mean_cls=0.0798, proj_loss=-0.416][2026-03-26 12:47:50] Step: 4038, Training Logs: loss_final: 0.525322, loss_mean: 0.849452, proj_loss: -0.406078, loss_mean_cls: 0.081948, grad_norm: 0.376759 +Steps: 0%| | 4039/1000000 [16:47<68:22:59, 4.05it/s, grad_norm=0.377, loss_final=0.525, loss_mean=0.849, loss_mean_cls=0.0819, proj_loss=-0.406][2026-03-26 12:47:50] Step: 4039, Training Logs: loss_final: 0.496196, loss_mean: 0.834998, proj_loss: -0.418211, loss_mean_cls: 0.079408, grad_norm: 0.306233 +Steps: 0%| | 4040/1000000 [16:47<68:21:23, 4.05it/s, grad_norm=0.306, loss_final=0.496, loss_mean=0.835, loss_mean_cls=0.0794, proj_loss=-0.418][2026-03-26 12:47:50] Step: 4040, Training Logs: loss_final: 0.508473, loss_mean: 0.831955, proj_loss: -0.405519, loss_mean_cls: 0.082037, grad_norm: 0.363540 +Steps: 0%| | 4041/1000000 [16:47<68:21:41, 4.05it/s, grad_norm=0.364, loss_final=0.508, loss_mean=0.832, loss_mean_cls=0.082, proj_loss=-0.406][2026-03-26 12:47:51] Step: 4041, Training Logs: loss_final: 0.515306, loss_mean: 0.857144, proj_loss: -0.418536, loss_mean_cls: 0.076698, grad_norm: 0.389905 +Steps: 0%| | 4042/1000000 [16:47<69:16:51, 3.99it/s, grad_norm=0.39, loss_final=0.515, loss_mean=0.857, loss_mean_cls=0.0767, proj_loss=-0.419][2026-03-26 12:47:51] Step: 4042, Training Logs: loss_final: 0.504729, loss_mean: 0.835098, proj_loss: -0.409704, loss_mean_cls: 0.079336, grad_norm: 0.322356 +Steps: 0%| | 4043/1000000 [16:48<69:00:51, 4.01it/s, grad_norm=0.322, loss_final=0.505, loss_mean=0.835, loss_mean_cls=0.0793, proj_loss=-0.41][2026-03-26 12:47:51] Step: 4043, Training Logs: loss_final: 0.512960, loss_mean: 0.848924, proj_loss: -0.413926, loss_mean_cls: 0.077962, grad_norm: 0.344328 +Steps: 0%| | 4044/1000000 [16:48<68:50:26, 4.02it/s, grad_norm=0.344, loss_final=0.513, loss_mean=0.849, loss_mean_cls=0.078, proj_loss=-0.414][2026-03-26 12:47:51] Step: 4044, Training Logs: loss_final: 0.513800, loss_mean: 0.855342, proj_loss: -0.418143, loss_mean_cls: 0.076601, grad_norm: 0.500628 +Steps: 0%| | 4045/1000000 [16:48<68:42:35, 4.03it/s, grad_norm=0.501, loss_final=0.514, loss_mean=0.855, loss_mean_cls=0.0766, proj_loss=-0.418][2026-03-26 12:47:52] Step: 4045, Training Logs: loss_final: 0.520786, loss_mean: 0.861267, proj_loss: -0.418337, loss_mean_cls: 0.077855, grad_norm: 0.392331 +Steps: 0%| | 4046/1000000 [16:48<68:42:38, 4.03it/s, grad_norm=0.392, loss_final=0.521, loss_mean=0.861, loss_mean_cls=0.0779, proj_loss=-0.418][2026-03-26 12:47:52] Step: 4046, Training Logs: loss_final: 0.510579, loss_mean: 0.827457, proj_loss: -0.398174, loss_mean_cls: 0.081296, grad_norm: 0.235018 +Steps: 0%| | 4047/1000000 [16:49<68:39:01, 4.03it/s, grad_norm=0.235, loss_final=0.511, loss_mean=0.827, loss_mean_cls=0.0813, proj_loss=-0.398][2026-03-26 12:47:52] Step: 4047, Training Logs: loss_final: 0.515027, loss_mean: 0.850459, proj_loss: -0.415045, loss_mean_cls: 0.079613, grad_norm: 0.423408 +Steps: 0%| | 4048/1000000 [16:49<68:35:21, 4.03it/s, grad_norm=0.423, loss_final=0.515, loss_mean=0.85, loss_mean_cls=0.0796, proj_loss=-0.415][2026-03-26 12:47:52] Step: 4048, Training Logs: loss_final: 0.528181, loss_mean: 0.859854, proj_loss: -0.410126, loss_mean_cls: 0.078453, grad_norm: 0.373094 +Steps: 0%| | 4049/1000000 [16:49<68:33:31, 4.04it/s, grad_norm=0.373, loss_final=0.528, loss_mean=0.86, loss_mean_cls=0.0785, proj_loss=-0.41][2026-03-26 12:47:53] Step: 4049, Training Logs: loss_final: 0.501495, loss_mean: 0.828377, proj_loss: -0.409735, loss_mean_cls: 0.082853, grad_norm: 0.367882 +Steps: 0%| | 4050/1000000 [16:49<68:31:46, 4.04it/s, grad_norm=0.368, loss_final=0.501, loss_mean=0.828, loss_mean_cls=0.0829, proj_loss=-0.41][2026-03-26 12:47:53] Step: 4050, Training Logs: loss_final: 0.525342, loss_mean: 0.852190, proj_loss: -0.407135, loss_mean_cls: 0.080287, grad_norm: 0.396884 +Steps: 0%| | 4051/1000000 [16:50<68:30:33, 4.04it/s, grad_norm=0.397, loss_final=0.525, loss_mean=0.852, loss_mean_cls=0.0803, proj_loss=-0.407][2026-03-26 12:47:53] Step: 4051, Training Logs: loss_final: 0.513860, loss_mean: 0.846814, proj_loss: -0.413462, loss_mean_cls: 0.080508, grad_norm: 0.291723 +Steps: 0%| | 4052/1000000 [16:50<68:28:57, 4.04it/s, grad_norm=0.292, loss_final=0.514, loss_mean=0.847, loss_mean_cls=0.0805, proj_loss=-0.413][2026-03-26 12:47:53] Step: 4052, Training Logs: loss_final: 0.509198, loss_mean: 0.838299, proj_loss: -0.410482, loss_mean_cls: 0.081381, grad_norm: 0.380583 +Steps: 0%| | 4053/1000000 [16:50<68:25:49, 4.04it/s, grad_norm=0.381, loss_final=0.509, loss_mean=0.838, loss_mean_cls=0.0814, proj_loss=-0.41][2026-03-26 12:47:54] Step: 4053, Training Logs: loss_final: 0.516656, loss_mean: 0.842256, proj_loss: -0.406902, loss_mean_cls: 0.081302, grad_norm: 0.372490 +Steps: 0%| | 4054/1000000 [16:50<68:27:00, 4.04it/s, grad_norm=0.372, loss_final=0.517, loss_mean=0.842, loss_mean_cls=0.0813, proj_loss=-0.407][2026-03-26 12:47:54] Step: 4054, Training Logs: loss_final: 0.508079, loss_mean: 0.839309, proj_loss: -0.412285, loss_mean_cls: 0.081055, grad_norm: 0.304642 +Steps: 0%| | 4055/1000000 [16:51<68:26:55, 4.04it/s, grad_norm=0.305, loss_final=0.508, loss_mean=0.839, loss_mean_cls=0.0811, proj_loss=-0.412][2026-03-26 12:47:54] Step: 4055, Training Logs: loss_final: 0.490598, loss_mean: 0.807809, proj_loss: -0.401955, loss_mean_cls: 0.084744, grad_norm: 0.478536 +Steps: 0%| | 4056/1000000 [16:51<68:24:40, 4.04it/s, grad_norm=0.479, loss_final=0.491, loss_mean=0.808, loss_mean_cls=0.0847, proj_loss=-0.402][2026-03-26 12:47:54] Step: 4056, Training Logs: loss_final: 0.509980, loss_mean: 0.836202, proj_loss: -0.407948, loss_mean_cls: 0.081726, grad_norm: 0.351983 +Steps: 0%| | 4057/1000000 [16:51<68:24:26, 4.04it/s, grad_norm=0.352, loss_final=0.51, loss_mean=0.836, loss_mean_cls=0.0817, proj_loss=-0.408][2026-03-26 12:47:55] Step: 4057, Training Logs: loss_final: 0.504602, loss_mean: 0.832454, proj_loss: -0.408385, loss_mean_cls: 0.080534, grad_norm: 0.318886 +Steps: 0%| | 4058/1000000 [16:51<68:22:24, 4.05it/s, grad_norm=0.319, loss_final=0.505, loss_mean=0.832, loss_mean_cls=0.0805, proj_loss=-0.408][2026-03-26 12:47:55] Step: 4058, Training Logs: loss_final: 0.509340, loss_mean: 0.852208, proj_loss: -0.420691, loss_mean_cls: 0.077823, grad_norm: 0.437864 +Steps: 0%| | 4059/1000000 [16:52<68:24:26, 4.04it/s, grad_norm=0.438, loss_final=0.509, loss_mean=0.852, loss_mean_cls=0.0778, proj_loss=-0.421][2026-03-26 12:47:55] Step: 4059, Training Logs: loss_final: 0.505402, loss_mean: 0.836940, proj_loss: -0.411024, loss_mean_cls: 0.079486, grad_norm: 0.500908 +Steps: 0%| | 4060/1000000 [16:52<68:25:46, 4.04it/s, grad_norm=0.501, loss_final=0.505, loss_mean=0.837, loss_mean_cls=0.0795, proj_loss=-0.411][2026-03-26 12:47:55] Step: 4060, Training Logs: loss_final: 0.505661, loss_mean: 0.835328, proj_loss: -0.411073, loss_mean_cls: 0.081407, grad_norm: 0.349709 +Steps: 0%| | 4061/1000000 [16:52<74:20:31, 3.72it/s, grad_norm=0.35, loss_final=0.506, loss_mean=0.835, loss_mean_cls=0.0814, proj_loss=-0.411][2026-03-26 12:47:56] Step: 4061, Training Logs: loss_final: 0.521316, loss_mean: 0.852077, proj_loss: -0.411355, loss_mean_cls: 0.080595, grad_norm: 0.364098 +Steps: 0%| | 4062/1000000 [16:52<72:53:31, 3.80it/s, grad_norm=0.364, loss_final=0.521, loss_mean=0.852, loss_mean_cls=0.0806, proj_loss=-0.411][2026-03-26 12:47:56] Step: 4062, Training Logs: loss_final: 0.507517, loss_mean: 0.825628, proj_loss: -0.400816, loss_mean_cls: 0.082706, grad_norm: 0.406073 +Steps: 0%| | 4063/1000000 [16:53<71:21:11, 3.88it/s, grad_norm=0.406, loss_final=0.508, loss_mean=0.826, loss_mean_cls=0.0827, proj_loss=-0.401][2026-03-26 12:47:56] Step: 4063, Training Logs: loss_final: 0.502284, loss_mean: 0.832383, proj_loss: -0.411325, loss_mean_cls: 0.081226, grad_norm: 0.389273 +Steps: 0%| | 4064/1000000 [16:53<70:30:07, 3.92it/s, grad_norm=0.389, loss_final=0.502, loss_mean=0.832, loss_mean_cls=0.0812, proj_loss=-0.411][2026-03-26 12:47:56] Step: 4064, Training Logs: loss_final: 0.496581, loss_mean: 0.826398, proj_loss: -0.410667, loss_mean_cls: 0.080849, grad_norm: 0.304810 +Steps: 0%| | 4065/1000000 [16:53<69:50:08, 3.96it/s, grad_norm=0.305, loss_final=0.497, loss_mean=0.826, loss_mean_cls=0.0808, proj_loss=-0.411][2026-03-26 12:47:57] Step: 4065, Training Logs: loss_final: 0.506786, loss_mean: 0.838036, proj_loss: -0.412311, loss_mean_cls: 0.081061, grad_norm: 0.287414 +Steps: 0%| | 4066/1000000 [16:53<69:24:17, 3.99it/s, grad_norm=0.287, loss_final=0.507, loss_mean=0.838, loss_mean_cls=0.0811, proj_loss=-0.412][2026-03-26 12:47:57] Step: 4066, Training Logs: loss_final: 0.521632, loss_mean: 0.848151, proj_loss: -0.406177, loss_mean_cls: 0.079658, grad_norm: 0.391819 +Steps: 0%| | 4067/1000000 [16:54<69:06:11, 4.00it/s, grad_norm=0.392, loss_final=0.522, loss_mean=0.848, loss_mean_cls=0.0797, proj_loss=-0.406][2026-03-26 12:47:57] Step: 4067, Training Logs: loss_final: 0.504732, loss_mean: 0.832719, proj_loss: -0.409229, loss_mean_cls: 0.081242, grad_norm: 0.335471 +Steps: 0%| | 4068/1000000 [16:54<68:53:48, 4.02it/s, grad_norm=0.335, loss_final=0.505, loss_mean=0.833, loss_mean_cls=0.0812, proj_loss=-0.409][2026-03-26 12:47:57] Step: 4068, Training Logs: loss_final: 0.506839, loss_mean: 0.832276, proj_loss: -0.408605, loss_mean_cls: 0.083167, grad_norm: 0.537793 +Steps: 0%| | 4069/1000000 [16:54<68:44:44, 4.02it/s, grad_norm=0.538, loss_final=0.507, loss_mean=0.832, loss_mean_cls=0.0832, proj_loss=-0.409][2026-03-26 12:47:58] Step: 4069, Training Logs: loss_final: 0.521028, loss_mean: 0.857437, proj_loss: -0.414924, loss_mean_cls: 0.078515, grad_norm: 0.533610 +Steps: 0%| | 4070/1000000 [16:54<68:39:27, 4.03it/s, grad_norm=0.534, loss_final=0.521, loss_mean=0.857, loss_mean_cls=0.0785, proj_loss=-0.415][2026-03-26 12:47:58] Step: 4070, Training Logs: loss_final: 0.514918, loss_mean: 0.847205, proj_loss: -0.412662, loss_mean_cls: 0.080375, grad_norm: 0.353627 +Steps: 0%| | 4071/1000000 [16:55<68:35:57, 4.03it/s, grad_norm=0.354, loss_final=0.515, loss_mean=0.847, loss_mean_cls=0.0804, proj_loss=-0.413][2026-03-26 12:47:58] Step: 4071, Training Logs: loss_final: 0.507033, loss_mean: 0.831810, proj_loss: -0.406494, loss_mean_cls: 0.081717, grad_norm: 0.558604 +Steps: 0%| | 4072/1000000 [16:55<68:30:35, 4.04it/s, grad_norm=0.559, loss_final=0.507, loss_mean=0.832, loss_mean_cls=0.0817, proj_loss=-0.406][2026-03-26 12:47:58] Step: 4072, Training Logs: loss_final: 0.493618, loss_mean: 0.814966, proj_loss: -0.403550, loss_mean_cls: 0.082201, grad_norm: 0.512663 +Steps: 0%| | 4073/1000000 [16:55<68:28:10, 4.04it/s, grad_norm=0.513, loss_final=0.494, loss_mean=0.815, loss_mean_cls=0.0822, proj_loss=-0.404][2026-03-26 12:47:59] Step: 4073, Training Logs: loss_final: 0.515410, loss_mean: 0.838661, proj_loss: -0.405632, loss_mean_cls: 0.082380, grad_norm: 0.342907 +Steps: 0%| | 4074/1000000 [16:55<68:26:34, 4.04it/s, grad_norm=0.343, loss_final=0.515, loss_mean=0.839, loss_mean_cls=0.0824, proj_loss=-0.406][2026-03-26 12:47:59] Step: 4074, Training Logs: loss_final: 0.514840, loss_mean: 0.841957, proj_loss: -0.407565, loss_mean_cls: 0.080448, grad_norm: 0.383369 +Steps: 0%| | 4075/1000000 [16:56<68:26:42, 4.04it/s, grad_norm=0.383, loss_final=0.515, loss_mean=0.842, loss_mean_cls=0.0804, proj_loss=-0.408][2026-03-26 12:47:59] Step: 4075, Training Logs: loss_final: 0.506853, loss_mean: 0.841777, proj_loss: -0.413933, loss_mean_cls: 0.079009, grad_norm: 0.282794 +Steps: 0%| | 4076/1000000 [16:56<68:27:41, 4.04it/s, grad_norm=0.283, loss_final=0.507, loss_mean=0.842, loss_mean_cls=0.079, proj_loss=-0.414][2026-03-26 12:47:59] Step: 4076, Training Logs: loss_final: 0.508826, loss_mean: 0.845559, proj_loss: -0.417033, loss_mean_cls: 0.080300, grad_norm: 0.425514 +Steps: 0%| | 4077/1000000 [16:56<68:30:14, 4.04it/s, grad_norm=0.426, loss_final=0.509, loss_mean=0.846, loss_mean_cls=0.0803, proj_loss=-0.417][2026-03-26 12:48:00] Step: 4077, Training Logs: loss_final: 0.508893, loss_mean: 0.841347, proj_loss: -0.412157, loss_mean_cls: 0.079703, grad_norm: 0.326939 +Steps: 0%| | 4078/1000000 [16:56<68:29:37, 4.04it/s, grad_norm=0.327, loss_final=0.509, loss_mean=0.841, loss_mean_cls=0.0797, proj_loss=-0.412][2026-03-26 12:48:00] Step: 4078, Training Logs: loss_final: 0.513408, loss_mean: 0.828701, proj_loss: -0.398459, loss_mean_cls: 0.083165, grad_norm: 0.416540 +Steps: 0%| | 4079/1000000 [16:57<68:28:22, 4.04it/s, grad_norm=0.417, loss_final=0.513, loss_mean=0.829, loss_mean_cls=0.0832, proj_loss=-0.398][2026-03-26 12:48:00] Step: 4079, Training Logs: loss_final: 0.515606, loss_mean: 0.835737, proj_loss: -0.402927, loss_mean_cls: 0.082795, grad_norm: 0.558526 +Steps: 0%| | 4080/1000000 [16:57<68:25:42, 4.04it/s, grad_norm=0.559, loss_final=0.516, loss_mean=0.836, loss_mean_cls=0.0828, proj_loss=-0.403][2026-03-26 12:48:00] Step: 4080, Training Logs: loss_final: 0.515425, loss_mean: 0.848054, proj_loss: -0.411634, loss_mean_cls: 0.079005, grad_norm: 0.435745 +Steps: 0%| | 4081/1000000 [16:57<68:25:37, 4.04it/s, grad_norm=0.436, loss_final=0.515, loss_mean=0.848, loss_mean_cls=0.079, proj_loss=-0.412][2026-03-26 12:48:01] Step: 4081, Training Logs: loss_final: 0.492857, loss_mean: 0.811136, proj_loss: -0.401513, loss_mean_cls: 0.083233, grad_norm: 0.419491 +Steps: 0%| | 4082/1000000 [16:57<68:25:09, 4.04it/s, grad_norm=0.419, loss_final=0.493, loss_mean=0.811, loss_mean_cls=0.0832, proj_loss=-0.402][2026-03-26 12:48:01] Step: 4082, Training Logs: loss_final: 0.505193, loss_mean: 0.839822, proj_loss: -0.413795, loss_mean_cls: 0.079167, grad_norm: 0.610141 +Steps: 0%| | 4083/1000000 [16:58<68:26:04, 4.04it/s, grad_norm=0.61, loss_final=0.505, loss_mean=0.84, loss_mean_cls=0.0792, proj_loss=-0.414][2026-03-26 12:48:01] Step: 4083, Training Logs: loss_final: 0.478722, loss_mean: 0.798879, proj_loss: -0.403074, loss_mean_cls: 0.082917, grad_norm: 0.350552 +Steps: 0%| | 4084/1000000 [16:58<68:24:54, 4.04it/s, grad_norm=0.351, loss_final=0.479, loss_mean=0.799, loss_mean_cls=0.0829, proj_loss=-0.403][2026-03-26 12:48:01] Step: 4084, Training Logs: loss_final: 0.510240, loss_mean: 0.843576, proj_loss: -0.411630, loss_mean_cls: 0.078294, grad_norm: 0.419910 +Steps: 0%| | 4085/1000000 [16:58<68:26:39, 4.04it/s, grad_norm=0.42, loss_final=0.51, loss_mean=0.844, loss_mean_cls=0.0783, proj_loss=-0.412][2026-03-26 12:48:02] Step: 4085, Training Logs: loss_final: 0.494903, loss_mean: 0.821274, proj_loss: -0.407138, loss_mean_cls: 0.080768, grad_norm: 0.356804 +Steps: 0%| | 4086/1000000 [16:58<68:11:19, 4.06it/s, grad_norm=0.357, loss_final=0.495, loss_mean=0.821, loss_mean_cls=0.0808, proj_loss=-0.407][2026-03-26 12:48:02] Step: 4086, Training Logs: loss_final: 0.510111, loss_mean: 0.833903, proj_loss: -0.407096, loss_mean_cls: 0.083304, grad_norm: inf +Steps: 0%| | 4087/1000000 [16:59<67:07:00, 4.12it/s, grad_norm=inf, loss_final=0.51, loss_mean=0.834, loss_mean_cls=0.0833, proj_loss=-0.407][2026-03-26 12:48:02] Step: 4087, Training Logs: loss_final: 0.496358, loss_mean: 0.812524, proj_loss: -0.400027, loss_mean_cls: 0.083861, grad_norm: 0.466063 +Steps: 0%| | 4088/1000000 [16:59<67:30:17, 4.10it/s, grad_norm=0.466, loss_final=0.496, loss_mean=0.813, loss_mean_cls=0.0839, proj_loss=-0.4][2026-03-26 12:48:02] Step: 4088, Training Logs: loss_final: 0.501426, loss_mean: 0.832291, proj_loss: -0.411307, loss_mean_cls: 0.080442, grad_norm: 0.303365 +Steps: 0%| | 4089/1000000 [16:59<67:47:08, 4.08it/s, grad_norm=0.303, loss_final=0.501, loss_mean=0.832, loss_mean_cls=0.0804, proj_loss=-0.411][2026-03-26 12:48:03] Step: 4089, Training Logs: loss_final: 0.504023, loss_mean: 0.836677, proj_loss: -0.411854, loss_mean_cls: 0.079201, grad_norm: 0.440736 +Steps: 0%| | 4090/1000000 [16:59<67:57:17, 4.07it/s, grad_norm=0.441, loss_final=0.504, loss_mean=0.837, loss_mean_cls=0.0792, proj_loss=-0.412][2026-03-26 12:48:03] Step: 4090, Training Logs: loss_final: 0.501761, loss_mean: 0.842292, proj_loss: -0.418806, loss_mean_cls: 0.078276, grad_norm: 0.472110 +Steps: 0%| | 4091/1000000 [17:00<68:04:53, 4.06it/s, grad_norm=0.472, loss_final=0.502, loss_mean=0.842, loss_mean_cls=0.0783, proj_loss=-0.419][2026-03-26 12:48:03] Step: 4091, Training Logs: loss_final: 0.499161, loss_mean: 0.830098, proj_loss: -0.412356, loss_mean_cls: 0.081419, grad_norm: 0.422976 +Steps: 0%| | 4092/1000000 [17:00<69:21:18, 3.99it/s, grad_norm=0.423, loss_final=0.499, loss_mean=0.83, loss_mean_cls=0.0814, proj_loss=-0.412][2026-03-26 12:48:03] Step: 4092, Training Logs: loss_final: 0.504383, loss_mean: 0.821705, proj_loss: -0.401742, loss_mean_cls: 0.084420, grad_norm: 0.283744 +Steps: 0%| | 4093/1000000 [17:00<69:07:56, 4.00it/s, grad_norm=0.284, loss_final=0.504, loss_mean=0.822, loss_mean_cls=0.0844, proj_loss=-0.402][2026-03-26 12:48:04] Step: 4093, Training Logs: loss_final: 0.508451, loss_mean: 0.837395, proj_loss: -0.409168, loss_mean_cls: 0.080224, grad_norm: 0.373229 +Steps: 0%| | 4094/1000000 [17:00<68:56:28, 4.01it/s, grad_norm=0.373, loss_final=0.508, loss_mean=0.837, loss_mean_cls=0.0802, proj_loss=-0.409][2026-03-26 12:48:04] Step: 4094, Training Logs: loss_final: 0.516653, loss_mean: 0.841659, proj_loss: -0.406448, loss_mean_cls: 0.081442, grad_norm: 0.377930 +Steps: 0%| | 4095/1000000 [17:01<68:47:48, 4.02it/s, grad_norm=0.378, loss_final=0.517, loss_mean=0.842, loss_mean_cls=0.0814, proj_loss=-0.406][2026-03-26 12:48:04] Step: 4095, Training Logs: loss_final: 0.516094, loss_mean: 0.830646, proj_loss: -0.398771, loss_mean_cls: 0.084219, grad_norm: 0.362253 +Steps: 0%| | 4096/1000000 [17:01<68:40:41, 4.03it/s, grad_norm=0.362, loss_final=0.516, loss_mean=0.831, loss_mean_cls=0.0842, proj_loss=-0.399][2026-03-26 12:48:04] Step: 4096, Training Logs: loss_final: 0.494104, loss_mean: 0.831526, proj_loss: -0.417228, loss_mean_cls: 0.079805, grad_norm: 0.374349 +Steps: 0%| | 4097/1000000 [17:01<68:36:00, 4.03it/s, grad_norm=0.374, loss_final=0.494, loss_mean=0.832, loss_mean_cls=0.0798, proj_loss=-0.417][2026-03-26 12:48:04] Step: 4097, Training Logs: loss_final: 0.492979, loss_mean: 0.823512, proj_loss: -0.411202, loss_mean_cls: 0.080669, grad_norm: 0.356289 +Steps: 0%| | 4098/1000000 [17:01<68:34:48, 4.03it/s, grad_norm=0.356, loss_final=0.493, loss_mean=0.824, loss_mean_cls=0.0807, proj_loss=-0.411][2026-03-26 12:48:05] Step: 4098, Training Logs: loss_final: 0.501576, loss_mean: 0.817486, proj_loss: -0.400631, loss_mean_cls: 0.084721, grad_norm: 0.376684 +Steps: 0%| | 4099/1000000 [17:02<68:30:52, 4.04it/s, grad_norm=0.377, loss_final=0.502, loss_mean=0.817, loss_mean_cls=0.0847, proj_loss=-0.401][2026-03-26 12:48:05] Step: 4099, Training Logs: loss_final: 0.503322, loss_mean: 0.832925, proj_loss: -0.410713, loss_mean_cls: 0.081110, grad_norm: 0.273140 +Steps: 0%| | 4100/1000000 [17:02<68:33:28, 4.04it/s, grad_norm=0.273, loss_final=0.503, loss_mean=0.833, loss_mean_cls=0.0811, proj_loss=-0.411][2026-03-26 12:48:05] Step: 4100, Training Logs: loss_final: 0.522605, loss_mean: 0.859303, proj_loss: -0.415629, loss_mean_cls: 0.078931, grad_norm: 0.384412 +Steps: 0%| | 4101/1000000 [17:02<68:30:15, 4.04it/s, grad_norm=0.384, loss_final=0.523, loss_mean=0.859, loss_mean_cls=0.0789, proj_loss=-0.416][2026-03-26 12:48:05] Step: 4101, Training Logs: loss_final: 0.510852, loss_mean: 0.834759, proj_loss: -0.404393, loss_mean_cls: 0.080485, grad_norm: 0.343496 +Steps: 0%| | 4102/1000000 [17:02<68:28:06, 4.04it/s, grad_norm=0.343, loss_final=0.511, loss_mean=0.835, loss_mean_cls=0.0805, proj_loss=-0.404][2026-03-26 12:48:06] Step: 4102, Training Logs: loss_final: 0.499435, loss_mean: 0.835172, proj_loss: -0.414593, loss_mean_cls: 0.078856, grad_norm: 0.311611 +Steps: 0%| | 4103/1000000 [17:03<68:25:14, 4.04it/s, grad_norm=0.312, loss_final=0.499, loss_mean=0.835, loss_mean_cls=0.0789, proj_loss=-0.415][2026-03-26 12:48:06] Step: 4103, Training Logs: loss_final: 0.517989, loss_mean: 0.856137, proj_loss: -0.417055, loss_mean_cls: 0.078906, grad_norm: 0.385506 +Steps: 0%| | 4104/1000000 [17:03<68:26:43, 4.04it/s, grad_norm=0.386, loss_final=0.518, loss_mean=0.856, loss_mean_cls=0.0789, proj_loss=-0.417][2026-03-26 12:48:06] Step: 4104, Training Logs: loss_final: 0.498567, loss_mean: 0.825703, proj_loss: -0.408758, loss_mean_cls: 0.081622, grad_norm: 0.489995 +Steps: 0%| | 4105/1000000 [17:03<68:23:02, 4.05it/s, grad_norm=0.49, loss_final=0.499, loss_mean=0.826, loss_mean_cls=0.0816, proj_loss=-0.409][2026-03-26 12:48:06] Step: 4105, Training Logs: loss_final: 0.512870, loss_mean: 0.821510, proj_loss: -0.393792, loss_mean_cls: 0.085152, grad_norm: 0.320331 +Steps: 0%| | 4106/1000000 [17:03<68:22:43, 4.05it/s, grad_norm=0.32, loss_final=0.513, loss_mean=0.822, loss_mean_cls=0.0852, proj_loss=-0.394][2026-03-26 12:48:07] Step: 4106, Training Logs: loss_final: 0.509087, loss_mean: 0.842026, proj_loss: -0.412379, loss_mean_cls: 0.079440, grad_norm: 0.632154 +Steps: 0%| | 4107/1000000 [17:04<68:25:25, 4.04it/s, grad_norm=0.632, loss_final=0.509, loss_mean=0.842, loss_mean_cls=0.0794, proj_loss=-0.412][2026-03-26 12:48:07] Step: 4107, Training Logs: loss_final: 0.490469, loss_mean: 0.815234, proj_loss: -0.405930, loss_mean_cls: 0.081165, grad_norm: 0.630026 +Steps: 0%| | 4108/1000000 [17:04<68:28:17, 4.04it/s, grad_norm=0.63, loss_final=0.49, loss_mean=0.815, loss_mean_cls=0.0812, proj_loss=-0.406][2026-03-26 12:48:07] Step: 4108, Training Logs: loss_final: 0.518717, loss_mean: 0.854130, proj_loss: -0.413905, loss_mean_cls: 0.078492, grad_norm: 0.374713 +Steps: 0%| | 4109/1000000 [17:04<68:29:53, 4.04it/s, grad_norm=0.375, loss_final=0.519, loss_mean=0.854, loss_mean_cls=0.0785, proj_loss=-0.414][2026-03-26 12:48:07] Step: 4109, Training Logs: loss_final: 0.509285, loss_mean: 0.839631, proj_loss: -0.410179, loss_mean_cls: 0.079833, grad_norm: 0.919531 +Steps: 0%| | 4110/1000000 [17:04<68:46:31, 4.02it/s, grad_norm=0.92, loss_final=0.509, loss_mean=0.84, loss_mean_cls=0.0798, proj_loss=-0.41][2026-03-26 12:48:08] Step: 4110, Training Logs: loss_final: 0.516278, loss_mean: 0.836345, proj_loss: -0.403343, loss_mean_cls: 0.083277, grad_norm: 0.846760 +Steps: 0%| | 4111/1000000 [17:05<68:39:13, 4.03it/s, grad_norm=0.847, loss_final=0.516, loss_mean=0.836, loss_mean_cls=0.0833, proj_loss=-0.403][2026-03-26 12:48:08] Step: 4111, Training Logs: loss_final: 0.503844, loss_mean: 0.837573, proj_loss: -0.413856, loss_mean_cls: 0.080128, grad_norm: 0.329262 +Steps: 0%| | 4112/1000000 [17:05<68:35:10, 4.03it/s, grad_norm=0.329, loss_final=0.504, loss_mean=0.838, loss_mean_cls=0.0801, proj_loss=-0.414][2026-03-26 12:48:08] Step: 4112, Training Logs: loss_final: 0.496109, loss_mean: 0.819037, proj_loss: -0.405666, loss_mean_cls: 0.082737, grad_norm: 0.743581 +Steps: 0%| | 4113/1000000 [17:05<68:32:56, 4.04it/s, grad_norm=0.744, loss_final=0.496, loss_mean=0.819, loss_mean_cls=0.0827, proj_loss=-0.406][2026-03-26 12:48:08] Step: 4113, Training Logs: loss_final: 0.519763, loss_mean: 0.861391, proj_loss: -0.418802, loss_mean_cls: 0.077173, grad_norm: 0.489120 +Steps: 0%| | 4114/1000000 [17:05<68:31:15, 4.04it/s, grad_norm=0.489, loss_final=0.52, loss_mean=0.861, loss_mean_cls=0.0772, proj_loss=-0.419][2026-03-26 12:48:09] Step: 4114, Training Logs: loss_final: 0.508995, loss_mean: 0.819198, proj_loss: -0.395564, loss_mean_cls: 0.085360, grad_norm: 0.507834 +Steps: 0%| | 4115/1000000 [17:06<68:30:31, 4.04it/s, grad_norm=0.508, loss_final=0.509, loss_mean=0.819, loss_mean_cls=0.0854, proj_loss=-0.396][2026-03-26 12:48:09] Step: 4115, Training Logs: loss_final: 0.493549, loss_mean: 0.825945, proj_loss: -0.412927, loss_mean_cls: 0.080531, grad_norm: 0.557581 +Steps: 0%| | 4116/1000000 [17:06<68:29:08, 4.04it/s, grad_norm=0.558, loss_final=0.494, loss_mean=0.826, loss_mean_cls=0.0805, proj_loss=-0.413][2026-03-26 12:48:09] Step: 4116, Training Logs: loss_final: 0.498045, loss_mean: 0.832610, proj_loss: -0.415189, loss_mean_cls: 0.080623, grad_norm: 0.296495 +Steps: 0%| | 4117/1000000 [17:06<68:30:22, 4.04it/s, grad_norm=0.296, loss_final=0.498, loss_mean=0.833, loss_mean_cls=0.0806, proj_loss=-0.415][2026-03-26 12:48:09] Step: 4117, Training Logs: loss_final: 0.501352, loss_mean: 0.833346, proj_loss: -0.411476, loss_mean_cls: 0.079482, grad_norm: 0.572957 +Steps: 0%| | 4118/1000000 [17:06<68:27:17, 4.04it/s, grad_norm=0.573, loss_final=0.501, loss_mean=0.833, loss_mean_cls=0.0795, proj_loss=-0.411][2026-03-26 12:48:10] Step: 4118, Training Logs: loss_final: 0.494366, loss_mean: 0.819622, proj_loss: -0.406750, loss_mean_cls: 0.081495, grad_norm: 0.544173 +Steps: 0%| | 4119/1000000 [17:07<68:26:05, 4.04it/s, grad_norm=0.544, loss_final=0.494, loss_mean=0.82, loss_mean_cls=0.0815, proj_loss=-0.407][2026-03-26 12:48:10] Step: 4119, Training Logs: loss_final: 0.499926, loss_mean: 0.822449, proj_loss: -0.405537, loss_mean_cls: 0.083014, grad_norm: 0.299064 +Steps: 0%| | 4120/1000000 [17:07<68:24:55, 4.04it/s, grad_norm=0.299, loss_final=0.5, loss_mean=0.822, loss_mean_cls=0.083, proj_loss=-0.406][2026-03-26 12:48:10] Step: 4120, Training Logs: loss_final: 0.507297, loss_mean: 0.830412, proj_loss: -0.405335, loss_mean_cls: 0.082220, grad_norm: 0.446025 +Steps: 0%| | 4121/1000000 [17:07<68:25:52, 4.04it/s, grad_norm=0.446, loss_final=0.507, loss_mean=0.83, loss_mean_cls=0.0822, proj_loss=-0.405][2026-03-26 12:48:10] Step: 4121, Training Logs: loss_final: 0.492820, loss_mean: 0.821495, proj_loss: -0.410264, loss_mean_cls: 0.081588, grad_norm: 0.337312 +Steps: 0%| | 4122/1000000 [17:07<68:24:18, 4.04it/s, grad_norm=0.337, loss_final=0.493, loss_mean=0.821, loss_mean_cls=0.0816, proj_loss=-0.41][2026-03-26 12:48:11] Step: 4122, Training Logs: loss_final: 0.505882, loss_mean: 0.828658, proj_loss: -0.405465, loss_mean_cls: 0.082688, grad_norm: 0.590944 +Steps: 0%| | 4123/1000000 [17:08<68:24:34, 4.04it/s, grad_norm=0.591, loss_final=0.506, loss_mean=0.829, loss_mean_cls=0.0827, proj_loss=-0.405][2026-03-26 12:48:11] Step: 4123, Training Logs: loss_final: 0.496443, loss_mean: 0.826896, proj_loss: -0.409483, loss_mean_cls: 0.079030, grad_norm: 0.341806 +Steps: 0%| | 4124/1000000 [17:08<68:22:37, 4.05it/s, grad_norm=0.342, loss_final=0.496, loss_mean=0.827, loss_mean_cls=0.079, proj_loss=-0.409][2026-03-26 12:48:11] Step: 4124, Training Logs: loss_final: 0.511986, loss_mean: 0.843537, proj_loss: -0.411443, loss_mean_cls: 0.079893, grad_norm: 0.634510 +Steps: 0%| | 4125/1000000 [17:08<68:23:59, 4.04it/s, grad_norm=0.635, loss_final=0.512, loss_mean=0.844, loss_mean_cls=0.0799, proj_loss=-0.411][2026-03-26 12:48:11] Step: 4125, Training Logs: loss_final: 0.499978, loss_mean: 0.836044, proj_loss: -0.414648, loss_mean_cls: 0.078582, grad_norm: 0.820491 +Steps: 0%| | 4126/1000000 [17:08<68:26:12, 4.04it/s, grad_norm=0.82, loss_final=0.5, loss_mean=0.836, loss_mean_cls=0.0786, proj_loss=-0.415][2026-03-26 12:48:12] Step: 4126, Training Logs: loss_final: 0.520155, loss_mean: 0.845187, proj_loss: -0.405576, loss_mean_cls: 0.080544, grad_norm: 0.507661 +Steps: 0%| | 4127/1000000 [17:09<68:24:14, 4.04it/s, grad_norm=0.508, loss_final=0.52, loss_mean=0.845, loss_mean_cls=0.0805, proj_loss=-0.406][2026-03-26 12:48:12] Step: 4127, Training Logs: loss_final: 0.521573, loss_mean: 0.845433, proj_loss: -0.404806, loss_mean_cls: 0.080946, grad_norm: 0.405850 +Steps: 0%| | 4128/1000000 [17:09<68:23:14, 4.05it/s, grad_norm=0.406, loss_final=0.522, loss_mean=0.845, loss_mean_cls=0.0809, proj_loss=-0.405][2026-03-26 12:48:12] Step: 4128, Training Logs: loss_final: 0.515822, loss_mean: 0.842046, proj_loss: -0.406154, loss_mean_cls: 0.079930, grad_norm: 0.507175 +Steps: 0%| | 4129/1000000 [17:09<68:24:42, 4.04it/s, grad_norm=0.507, loss_final=0.516, loss_mean=0.842, loss_mean_cls=0.0799, proj_loss=-0.406][2026-03-26 12:48:12] Step: 4129, Training Logs: loss_final: 0.500031, loss_mean: 0.828351, proj_loss: -0.408234, loss_mean_cls: 0.079913, grad_norm: 0.388661 +Steps: 0%| | 4130/1000000 [17:09<68:23:26, 4.04it/s, grad_norm=0.389, loss_final=0.5, loss_mean=0.828, loss_mean_cls=0.0799, proj_loss=-0.408][2026-03-26 12:48:13] Step: 4130, Training Logs: loss_final: 0.518259, loss_mean: 0.840797, proj_loss: -0.401823, loss_mean_cls: 0.079285, grad_norm: 0.565689 +Steps: 0%| | 4131/1000000 [17:10<68:23:45, 4.04it/s, grad_norm=0.566, loss_final=0.518, loss_mean=0.841, loss_mean_cls=0.0793, proj_loss=-0.402][2026-03-26 12:48:13] Step: 4131, Training Logs: loss_final: 0.496416, loss_mean: 0.825836, proj_loss: -0.410611, loss_mean_cls: 0.081191, grad_norm: 0.262012 +Steps: 0%| | 4132/1000000 [17:10<68:26:30, 4.04it/s, grad_norm=0.262, loss_final=0.496, loss_mean=0.826, loss_mean_cls=0.0812, proj_loss=-0.411][2026-03-26 12:48:13] Step: 4132, Training Logs: loss_final: 0.514670, loss_mean: 0.841260, proj_loss: -0.406247, loss_mean_cls: 0.079657, grad_norm: 0.405644 +Steps: 0%| | 4133/1000000 [17:10<68:25:27, 4.04it/s, grad_norm=0.406, loss_final=0.515, loss_mean=0.841, loss_mean_cls=0.0797, proj_loss=-0.406][2026-03-26 12:48:13] Step: 4133, Training Logs: loss_final: 0.502226, loss_mean: 0.834536, proj_loss: -0.411896, loss_mean_cls: 0.079586, grad_norm: 0.266055 +Steps: 0%| | 4134/1000000 [17:10<68:26:49, 4.04it/s, grad_norm=0.266, loss_final=0.502, loss_mean=0.835, loss_mean_cls=0.0796, proj_loss=-0.412][2026-03-26 12:48:14] Step: 4134, Training Logs: loss_final: 0.494216, loss_mean: 0.810553, proj_loss: -0.399686, loss_mean_cls: 0.083350, grad_norm: 0.352561 +Steps: 0%| | 4135/1000000 [17:11<68:25:32, 4.04it/s, grad_norm=0.353, loss_final=0.494, loss_mean=0.811, loss_mean_cls=0.0833, proj_loss=-0.4][2026-03-26 12:48:14] Step: 4135, Training Logs: loss_final: 0.507843, loss_mean: 0.837327, proj_loss: -0.409881, loss_mean_cls: 0.080396, grad_norm: 0.422637 +Steps: 0%| | 4136/1000000 [17:11<68:25:58, 4.04it/s, grad_norm=0.423, loss_final=0.508, loss_mean=0.837, loss_mean_cls=0.0804, proj_loss=-0.41][2026-03-26 12:48:14] Step: 4136, Training Logs: loss_final: 0.502477, loss_mean: 0.836715, proj_loss: -0.412856, loss_mean_cls: 0.078618, grad_norm: 0.259400 +Steps: 0%| | 4137/1000000 [17:11<68:24:32, 4.04it/s, grad_norm=0.259, loss_final=0.502, loss_mean=0.837, loss_mean_cls=0.0786, proj_loss=-0.413][2026-03-26 12:48:14] Step: 4137, Training Logs: loss_final: 0.510753, loss_mean: 0.837344, proj_loss: -0.406586, loss_mean_cls: 0.079996, grad_norm: 0.281544 +Steps: 0%| | 4138/1000000 [17:11<68:24:13, 4.04it/s, grad_norm=0.282, loss_final=0.511, loss_mean=0.837, loss_mean_cls=0.08, proj_loss=-0.407][2026-03-26 12:48:15] Step: 4138, Training Logs: loss_final: 0.485430, loss_mean: 0.820725, proj_loss: -0.415253, loss_mean_cls: 0.079959, grad_norm: 0.254853 +Steps: 0%| | 4139/1000000 [17:11<68:23:49, 4.04it/s, grad_norm=0.255, loss_final=0.485, loss_mean=0.821, loss_mean_cls=0.08, proj_loss=-0.415][2026-03-26 12:48:15] Step: 4139, Training Logs: loss_final: 0.513844, loss_mean: 0.839008, proj_loss: -0.408056, loss_mean_cls: 0.082892, grad_norm: 0.260717 +Steps: 0%| | 4140/1000000 [17:12<68:27:17, 4.04it/s, grad_norm=0.261, loss_final=0.514, loss_mean=0.839, loss_mean_cls=0.0829, proj_loss=-0.408][2026-03-26 12:48:15] Step: 4140, Training Logs: loss_final: 0.511698, loss_mean: 0.840222, proj_loss: -0.409740, loss_mean_cls: 0.081216, grad_norm: 0.313909 +Steps: 0%| | 4141/1000000 [17:12<68:23:39, 4.04it/s, grad_norm=0.314, loss_final=0.512, loss_mean=0.84, loss_mean_cls=0.0812, proj_loss=-0.41][2026-03-26 12:48:15] Step: 4141, Training Logs: loss_final: 0.491685, loss_mean: 0.815802, proj_loss: -0.405987, loss_mean_cls: 0.081870, grad_norm: 0.416346 +Steps: 0%| | 4142/1000000 [17:12<68:25:18, 4.04it/s, grad_norm=0.416, loss_final=0.492, loss_mean=0.816, loss_mean_cls=0.0819, proj_loss=-0.406][2026-03-26 12:48:16] Step: 4142, Training Logs: loss_final: 0.506175, loss_mean: 0.841362, proj_loss: -0.414145, loss_mean_cls: 0.078958, grad_norm: 0.277745 +Steps: 0%| | 4143/1000000 [17:12<68:27:37, 4.04it/s, grad_norm=0.278, loss_final=0.506, loss_mean=0.841, loss_mean_cls=0.079, proj_loss=-0.414][2026-03-26 12:48:16] Step: 4143, Training Logs: loss_final: 0.521603, loss_mean: 0.848764, proj_loss: -0.408039, loss_mean_cls: 0.080878, grad_norm: 0.435106 +Steps: 0%| | 4144/1000000 [17:13<68:33:52, 4.03it/s, grad_norm=0.435, loss_final=0.522, loss_mean=0.849, loss_mean_cls=0.0809, proj_loss=-0.408][2026-03-26 12:48:16] Step: 4144, Training Logs: loss_final: 0.526616, loss_mean: 0.864887, proj_loss: -0.415540, loss_mean_cls: 0.077269, grad_norm: 0.348928 +Steps: 0%| | 4145/1000000 [17:13<68:37:15, 4.03it/s, grad_norm=0.349, loss_final=0.527, loss_mean=0.865, loss_mean_cls=0.0773, proj_loss=-0.416][2026-03-26 12:48:16] Step: 4145, Training Logs: loss_final: 0.500636, loss_mean: 0.826571, proj_loss: -0.407715, loss_mean_cls: 0.081780, grad_norm: 0.305360 +Steps: 0%| | 4146/1000000 [17:13<68:34:48, 4.03it/s, grad_norm=0.305, loss_final=0.501, loss_mean=0.827, loss_mean_cls=0.0818, proj_loss=-0.408][2026-03-26 12:48:17] Step: 4146, Training Logs: loss_final: 0.491288, loss_mean: 0.815221, proj_loss: -0.407608, loss_mean_cls: 0.083674, grad_norm: 0.579148 +Steps: 0%| | 4147/1000000 [17:13<68:33:26, 4.03it/s, grad_norm=0.579, loss_final=0.491, loss_mean=0.815, loss_mean_cls=0.0837, proj_loss=-0.408][2026-03-26 12:48:17] Step: 4147, Training Logs: loss_final: 0.518443, loss_mean: 0.838532, proj_loss: -0.403146, loss_mean_cls: 0.083057, grad_norm: 0.394394 +Steps: 0%| | 4148/1000000 [17:14<68:33:32, 4.03it/s, grad_norm=0.394, loss_final=0.518, loss_mean=0.839, loss_mean_cls=0.0831, proj_loss=-0.403][2026-03-26 12:48:17] Step: 4148, Training Logs: loss_final: 0.498888, loss_mean: 0.825447, proj_loss: -0.408351, loss_mean_cls: 0.081791, grad_norm: 0.464532 +Steps: 0%| | 4149/1000000 [17:14<68:31:36, 4.04it/s, grad_norm=0.465, loss_final=0.499, loss_mean=0.825, loss_mean_cls=0.0818, proj_loss=-0.408][2026-03-26 12:48:17] Step: 4149, Training Logs: loss_final: 0.511330, loss_mean: 0.843087, proj_loss: -0.410834, loss_mean_cls: 0.079077, grad_norm: 0.408770 +Steps: 0%| | 4150/1000000 [17:14<68:26:39, 4.04it/s, grad_norm=0.409, loss_final=0.511, loss_mean=0.843, loss_mean_cls=0.0791, proj_loss=-0.411][2026-03-26 12:48:18] Step: 4150, Training Logs: loss_final: 0.496328, loss_mean: 0.820532, proj_loss: -0.405550, loss_mean_cls: 0.081346, grad_norm: 0.366613 +Steps: 0%| | 4151/1000000 [17:14<68:25:13, 4.04it/s, grad_norm=0.367, loss_final=0.496, loss_mean=0.821, loss_mean_cls=0.0813, proj_loss=-0.406][2026-03-26 12:48:18] Step: 4151, Training Logs: loss_final: 0.516639, loss_mean: 0.837338, proj_loss: -0.403983, loss_mean_cls: 0.083285, grad_norm: 0.641736 +Steps: 0%| | 4152/1000000 [17:15<68:29:27, 4.04it/s, grad_norm=0.642, loss_final=0.517, loss_mean=0.837, loss_mean_cls=0.0833, proj_loss=-0.404][2026-03-26 12:48:18] Step: 4152, Training Logs: loss_final: 0.515435, loss_mean: 0.844128, proj_loss: -0.406534, loss_mean_cls: 0.077840, grad_norm: 0.416755 +Steps: 0%| | 4153/1000000 [17:15<68:27:43, 4.04it/s, grad_norm=0.417, loss_final=0.515, loss_mean=0.844, loss_mean_cls=0.0778, proj_loss=-0.407][2026-03-26 12:48:18] Step: 4153, Training Logs: loss_final: 0.497784, loss_mean: 0.833873, proj_loss: -0.415215, loss_mean_cls: 0.079127, grad_norm: 0.337203 +Steps: 0%| | 4154/1000000 [17:15<68:27:49, 4.04it/s, grad_norm=0.337, loss_final=0.498, loss_mean=0.834, loss_mean_cls=0.0791, proj_loss=-0.415][2026-03-26 12:48:19] Step: 4154, Training Logs: loss_final: 0.517476, loss_mean: 0.847978, proj_loss: -0.410992, loss_mean_cls: 0.080491, grad_norm: 0.376131 +Steps: 0%| | 4155/1000000 [17:15<68:28:19, 4.04it/s, grad_norm=0.376, loss_final=0.517, loss_mean=0.848, loss_mean_cls=0.0805, proj_loss=-0.411][2026-03-26 12:48:19] Step: 4155, Training Logs: loss_final: 0.491038, loss_mean: 0.806565, proj_loss: -0.401158, loss_mean_cls: 0.085632, grad_norm: 0.275558 +Steps: 0%| | 4156/1000000 [17:16<68:27:42, 4.04it/s, grad_norm=0.276, loss_final=0.491, loss_mean=0.807, loss_mean_cls=0.0856, proj_loss=-0.401][2026-03-26 12:48:19] Step: 4156, Training Logs: loss_final: 0.497377, loss_mean: 0.818948, proj_loss: -0.403938, loss_mean_cls: 0.082368, grad_norm: 0.337397 +Steps: 0%| | 4157/1000000 [17:16<68:29:56, 4.04it/s, grad_norm=0.337, loss_final=0.497, loss_mean=0.819, loss_mean_cls=0.0824, proj_loss=-0.404][2026-03-26 12:48:19] Step: 4157, Training Logs: loss_final: 0.493572, loss_mean: 0.826397, proj_loss: -0.412273, loss_mean_cls: 0.079448, grad_norm: 0.401611 +Steps: 0%| | 4158/1000000 [17:16<68:28:52, 4.04it/s, grad_norm=0.402, loss_final=0.494, loss_mean=0.826, loss_mean_cls=0.0794, proj_loss=-0.412][2026-03-26 12:48:20] Step: 4158, Training Logs: loss_final: 0.502604, loss_mean: 0.819357, proj_loss: -0.400601, loss_mean_cls: 0.083847, grad_norm: 0.359234 +Steps: 0%| | 4159/1000000 [17:16<68:28:55, 4.04it/s, grad_norm=0.359, loss_final=0.503, loss_mean=0.819, loss_mean_cls=0.0838, proj_loss=-0.401][2026-03-26 12:48:20] Step: 4159, Training Logs: loss_final: 0.525434, loss_mean: 0.854570, proj_loss: -0.408320, loss_mean_cls: 0.079185, grad_norm: 0.258707 +Steps: 0%| | 4160/1000000 [17:17<68:32:38, 4.04it/s, grad_norm=0.259, loss_final=0.525, loss_mean=0.855, loss_mean_cls=0.0792, proj_loss=-0.408][2026-03-26 12:48:20] Step: 4160, Training Logs: loss_final: 0.501584, loss_mean: 0.823519, proj_loss: -0.403853, loss_mean_cls: 0.081919, grad_norm: 0.517288 +Steps: 0%| | 4161/1000000 [17:17<68:32:14, 4.04it/s, grad_norm=0.517, loss_final=0.502, loss_mean=0.824, loss_mean_cls=0.0819, proj_loss=-0.404][2026-03-26 12:48:20] Step: 4161, Training Logs: loss_final: 0.521092, loss_mean: 0.844373, proj_loss: -0.405983, loss_mean_cls: 0.082702, grad_norm: 0.415641 +Steps: 0%| | 4162/1000000 [17:17<68:28:08, 4.04it/s, grad_norm=0.416, loss_final=0.521, loss_mean=0.844, loss_mean_cls=0.0827, proj_loss=-0.406][2026-03-26 12:48:21] Step: 4162, Training Logs: loss_final: 0.498297, loss_mean: 0.817973, proj_loss: -0.404002, loss_mean_cls: 0.084326, grad_norm: 0.275955 +Steps: 0%| | 4163/1000000 [17:17<68:26:30, 4.04it/s, grad_norm=0.276, loss_final=0.498, loss_mean=0.818, loss_mean_cls=0.0843, proj_loss=-0.404][2026-03-26 12:48:21] Step: 4163, Training Logs: loss_final: 0.513326, loss_mean: 0.837608, proj_loss: -0.405880, loss_mean_cls: 0.081599, grad_norm: 0.426411 +Steps: 0%| | 4164/1000000 [17:18<68:25:01, 4.04it/s, grad_norm=0.426, loss_final=0.513, loss_mean=0.838, loss_mean_cls=0.0816, proj_loss=-0.406][2026-03-26 12:48:21] Step: 4164, Training Logs: loss_final: 0.513301, loss_mean: 0.839825, proj_loss: -0.407715, loss_mean_cls: 0.081191, grad_norm: 0.321694 +Steps: 0%| | 4165/1000000 [17:18<68:25:50, 4.04it/s, grad_norm=0.322, loss_final=0.513, loss_mean=0.84, loss_mean_cls=0.0812, proj_loss=-0.408][2026-03-26 12:48:21] Step: 4165, Training Logs: loss_final: 0.480970, loss_mean: 0.804689, proj_loss: -0.406682, loss_mean_cls: 0.082963, grad_norm: 0.339692 +Steps: 0%| | 4166/1000000 [17:18<68:25:02, 4.04it/s, grad_norm=0.34, loss_final=0.481, loss_mean=0.805, loss_mean_cls=0.083, proj_loss=-0.407][2026-03-26 12:48:22] Step: 4166, Training Logs: loss_final: 0.507760, loss_mean: 0.847641, proj_loss: -0.418672, loss_mean_cls: 0.078790, grad_norm: 0.347723 +Steps: 0%| | 4167/1000000 [17:18<68:24:24, 4.04it/s, grad_norm=0.348, loss_final=0.508, loss_mean=0.848, loss_mean_cls=0.0788, proj_loss=-0.419][2026-03-26 12:48:22] Step: 4167, Training Logs: loss_final: 0.511526, loss_mean: 0.833726, proj_loss: -0.403354, loss_mean_cls: 0.081155, grad_norm: 0.412687 +Steps: 0%| | 4168/1000000 [17:19<68:26:01, 4.04it/s, grad_norm=0.413, loss_final=0.512, loss_mean=0.834, loss_mean_cls=0.0812, proj_loss=-0.403][2026-03-26 12:48:22] Step: 4168, Training Logs: loss_final: 0.505288, loss_mean: 0.822065, proj_loss: -0.400430, loss_mean_cls: 0.083653, grad_norm: 0.346013 +Steps: 0%| | 4169/1000000 [17:19<68:25:41, 4.04it/s, grad_norm=0.346, loss_final=0.505, loss_mean=0.822, loss_mean_cls=0.0837, proj_loss=-0.4][2026-03-26 12:48:22] Step: 4169, Training Logs: loss_final: 0.514714, loss_mean: 0.833856, proj_loss: -0.400272, loss_mean_cls: 0.081130, grad_norm: 0.492798 +Steps: 0%| | 4170/1000000 [17:19<68:25:52, 4.04it/s, grad_norm=0.493, loss_final=0.515, loss_mean=0.834, loss_mean_cls=0.0811, proj_loss=-0.4][2026-03-26 12:48:23] Step: 4170, Training Logs: loss_final: 0.502605, loss_mean: 0.841348, proj_loss: -0.416801, loss_mean_cls: 0.078058, grad_norm: 0.473377 +Steps: 0%| | 4171/1000000 [17:19<68:26:05, 4.04it/s, grad_norm=0.473, loss_final=0.503, loss_mean=0.841, loss_mean_cls=0.0781, proj_loss=-0.417][2026-03-26 12:48:23] Step: 4171, Training Logs: loss_final: 0.498929, loss_mean: 0.826213, proj_loss: -0.408687, loss_mean_cls: 0.081404, grad_norm: 0.378199 +Steps: 0%| | 4172/1000000 [17:20<68:31:53, 4.04it/s, grad_norm=0.378, loss_final=0.499, loss_mean=0.826, loss_mean_cls=0.0814, proj_loss=-0.409][2026-03-26 12:48:23] Step: 4172, Training Logs: loss_final: 0.505006, loss_mean: 0.839825, proj_loss: -0.413301, loss_mean_cls: 0.078482, grad_norm: 0.489875 +Steps: 0%| | 4173/1000000 [17:20<68:28:43, 4.04it/s, grad_norm=0.49, loss_final=0.505, loss_mean=0.84, loss_mean_cls=0.0785, proj_loss=-0.413][2026-03-26 12:48:23] Step: 4173, Training Logs: loss_final: 0.499036, loss_mean: 0.834833, proj_loss: -0.415037, loss_mean_cls: 0.079240, grad_norm: 0.311491 +Steps: 0%| | 4174/1000000 [17:20<68:29:41, 4.04it/s, grad_norm=0.311, loss_final=0.499, loss_mean=0.835, loss_mean_cls=0.0792, proj_loss=-0.415][2026-03-26 12:48:24] Step: 4174, Training Logs: loss_final: 0.523878, loss_mean: 0.850476, proj_loss: -0.406876, loss_mean_cls: 0.080278, grad_norm: 0.357537 +Steps: 0%| | 4175/1000000 [17:20<68:27:45, 4.04it/s, grad_norm=0.358, loss_final=0.524, loss_mean=0.85, loss_mean_cls=0.0803, proj_loss=-0.407][2026-03-26 12:48:24] Step: 4175, Training Logs: loss_final: 0.509972, loss_mean: 0.835160, proj_loss: -0.405656, loss_mean_cls: 0.080467, grad_norm: 0.409595 +Steps: 0%| | 4176/1000000 [17:21<68:27:03, 4.04it/s, grad_norm=0.41, loss_final=0.51, loss_mean=0.835, loss_mean_cls=0.0805, proj_loss=-0.406][2026-03-26 12:48:24] Step: 4176, Training Logs: loss_final: 0.520842, loss_mean: 0.845659, proj_loss: -0.406111, loss_mean_cls: 0.081294, grad_norm: 0.318385 +Steps: 0%| | 4177/1000000 [17:21<68:23:56, 4.04it/s, grad_norm=0.318, loss_final=0.521, loss_mean=0.846, loss_mean_cls=0.0813, proj_loss=-0.406][2026-03-26 12:48:24] Step: 4177, Training Logs: loss_final: 0.495934, loss_mean: 0.834458, proj_loss: -0.417433, loss_mean_cls: 0.078909, grad_norm: 0.352528 +Steps: 0%| | 4178/1000000 [17:21<68:23:46, 4.04it/s, grad_norm=0.353, loss_final=0.496, loss_mean=0.834, loss_mean_cls=0.0789, proj_loss=-0.417][2026-03-26 12:48:25] Step: 4178, Training Logs: loss_final: 0.482544, loss_mean: 0.793305, proj_loss: -0.399975, loss_mean_cls: 0.089214, grad_norm: 0.327905 +Steps: 0%| | 4179/1000000 [17:21<68:24:31, 4.04it/s, grad_norm=0.328, loss_final=0.483, loss_mean=0.793, loss_mean_cls=0.0892, proj_loss=-0.4][2026-03-26 12:48:25] Step: 4179, Training Logs: loss_final: 0.501891, loss_mean: 0.840121, proj_loss: -0.417789, loss_mean_cls: 0.079558, grad_norm: 0.449474 +Steps: 0%| | 4180/1000000 [17:22<68:25:47, 4.04it/s, grad_norm=0.449, loss_final=0.502, loss_mean=0.84, loss_mean_cls=0.0796, proj_loss=-0.418][2026-03-26 12:48:25] Step: 4180, Training Logs: loss_final: 0.510458, loss_mean: 0.834069, proj_loss: -0.405396, loss_mean_cls: 0.081784, grad_norm: 0.662145 +Steps: 0%| | 4181/1000000 [17:22<68:22:23, 4.05it/s, grad_norm=0.662, loss_final=0.51, loss_mean=0.834, loss_mean_cls=0.0818, proj_loss=-0.405][2026-03-26 12:48:25] Step: 4181, Training Logs: loss_final: 0.505952, loss_mean: 0.818753, proj_loss: -0.396778, loss_mean_cls: 0.083977, grad_norm: 0.320784 +Steps: 0%| | 4182/1000000 [17:22<68:23:23, 4.04it/s, grad_norm=0.321, loss_final=0.506, loss_mean=0.819, loss_mean_cls=0.084, proj_loss=-0.397][2026-03-26 12:48:26] Step: 4182, Training Logs: loss_final: 0.492655, loss_mean: 0.820736, proj_loss: -0.409592, loss_mean_cls: 0.081511, grad_norm: 0.629381 +Steps: 0%| | 4183/1000000 [17:22<68:20:52, 4.05it/s, grad_norm=0.629, loss_final=0.493, loss_mean=0.821, loss_mean_cls=0.0815, proj_loss=-0.41][2026-03-26 12:48:26] Step: 4183, Training Logs: loss_final: 0.493897, loss_mean: 0.818259, proj_loss: -0.406943, loss_mean_cls: 0.082581, grad_norm: 0.615597 +Steps: 0%| | 4184/1000000 [17:23<68:21:17, 4.05it/s, grad_norm=0.616, loss_final=0.494, loss_mean=0.818, loss_mean_cls=0.0826, proj_loss=-0.407][2026-03-26 12:48:26] Step: 4184, Training Logs: loss_final: 0.505482, loss_mean: 0.835384, proj_loss: -0.409276, loss_mean_cls: 0.079374, grad_norm: 0.248122 +Steps: 0%| | 4185/1000000 [17:23<68:23:38, 4.04it/s, grad_norm=0.248, loss_final=0.505, loss_mean=0.835, loss_mean_cls=0.0794, proj_loss=-0.409][2026-03-26 12:48:26] Step: 4185, Training Logs: loss_final: 0.508948, loss_mean: 0.845434, proj_loss: -0.416223, loss_mean_cls: 0.079736, grad_norm: 0.724474 +Steps: 0%| | 4186/1000000 [17:23<68:23:22, 4.04it/s, grad_norm=0.724, loss_final=0.509, loss_mean=0.845, loss_mean_cls=0.0797, proj_loss=-0.416][2026-03-26 12:48:27] Step: 4186, Training Logs: loss_final: 0.512646, loss_mean: 0.857718, proj_loss: -0.422886, loss_mean_cls: 0.077813, grad_norm: 0.412257 +Steps: 0%| | 4187/1000000 [17:23<68:23:11, 4.04it/s, grad_norm=0.412, loss_final=0.513, loss_mean=0.858, loss_mean_cls=0.0778, proj_loss=-0.423][2026-03-26 12:48:27] Step: 4187, Training Logs: loss_final: 0.503110, loss_mean: 0.829002, proj_loss: -0.405946, loss_mean_cls: 0.080054, grad_norm: 0.505844 +Steps: 0%| | 4188/1000000 [17:24<68:26:02, 4.04it/s, grad_norm=0.506, loss_final=0.503, loss_mean=0.829, loss_mean_cls=0.0801, proj_loss=-0.406][2026-03-26 12:48:27] Step: 4188, Training Logs: loss_final: 0.489074, loss_mean: 0.815077, proj_loss: -0.408419, loss_mean_cls: 0.082415, grad_norm: 0.545284 +Steps: 0%| | 4189/1000000 [17:24<68:24:51, 4.04it/s, grad_norm=0.545, loss_final=0.489, loss_mean=0.815, loss_mean_cls=0.0824, proj_loss=-0.408][2026-03-26 12:48:27] Step: 4189, Training Logs: loss_final: 0.511419, loss_mean: 0.850032, proj_loss: -0.417082, loss_mean_cls: 0.078470, grad_norm: 0.371603 +Steps: 0%| | 4190/1000000 [17:24<68:25:18, 4.04it/s, grad_norm=0.372, loss_final=0.511, loss_mean=0.85, loss_mean_cls=0.0785, proj_loss=-0.417][2026-03-26 12:48:28] Step: 4190, Training Logs: loss_final: 0.502367, loss_mean: 0.845748, proj_loss: -0.421151, loss_mean_cls: 0.077770, grad_norm: 0.955274 +Steps: 0%| | 4191/1000000 [17:24<68:25:20, 4.04it/s, grad_norm=0.955, loss_final=0.502, loss_mean=0.846, loss_mean_cls=0.0778, proj_loss=-0.421][2026-03-26 12:48:28] Step: 4191, Training Logs: loss_final: 0.518431, loss_mean: 0.854063, proj_loss: -0.414001, loss_mean_cls: 0.078369, grad_norm: 0.779381 +Steps: 0%| | 4192/1000000 [17:25<68:25:00, 4.04it/s, grad_norm=0.779, loss_final=0.518, loss_mean=0.854, loss_mean_cls=0.0784, proj_loss=-0.414][2026-03-26 12:48:28] Step: 4192, Training Logs: loss_final: 0.500055, loss_mean: 0.833463, proj_loss: -0.413634, loss_mean_cls: 0.080226, grad_norm: 0.335733 +Steps: 0%| | 4193/1000000 [17:25<68:27:12, 4.04it/s, grad_norm=0.336, loss_final=0.5, loss_mean=0.833, loss_mean_cls=0.0802, proj_loss=-0.414][2026-03-26 12:48:28] Step: 4193, Training Logs: loss_final: 0.507483, loss_mean: 0.849591, proj_loss: -0.418687, loss_mean_cls: 0.076579, grad_norm: 0.793970 +Steps: 0%| | 4194/1000000 [17:25<68:25:50, 4.04it/s, grad_norm=0.794, loss_final=0.507, loss_mean=0.85, loss_mean_cls=0.0766, proj_loss=-0.419][2026-03-26 12:48:28] Step: 4194, Training Logs: loss_final: 0.511313, loss_mean: 0.839478, proj_loss: -0.407838, loss_mean_cls: 0.079673, grad_norm: 0.530228 +Steps: 0%| | 4195/1000000 [17:25<68:27:24, 4.04it/s, grad_norm=0.53, loss_final=0.511, loss_mean=0.839, loss_mean_cls=0.0797, proj_loss=-0.408][2026-03-26 12:48:29] Step: 4195, Training Logs: loss_final: 0.487069, loss_mean: 0.814316, proj_loss: -0.407520, loss_mean_cls: 0.080273, grad_norm: 0.611524 +Steps: 0%| | 4196/1000000 [17:26<68:26:41, 4.04it/s, grad_norm=0.612, loss_final=0.487, loss_mean=0.814, loss_mean_cls=0.0803, proj_loss=-0.408][2026-03-26 12:48:29] Step: 4196, Training Logs: loss_final: 0.501137, loss_mean: 0.835592, proj_loss: -0.413162, loss_mean_cls: 0.078707, grad_norm: 0.767475 +Steps: 0%| | 4197/1000000 [17:26<68:25:58, 4.04it/s, grad_norm=0.767, loss_final=0.501, loss_mean=0.836, loss_mean_cls=0.0787, proj_loss=-0.413][2026-03-26 12:48:29] Step: 4197, Training Logs: loss_final: 0.497229, loss_mean: 0.831886, proj_loss: -0.414068, loss_mean_cls: 0.079411, grad_norm: 0.335793 +Steps: 0%| | 4198/1000000 [17:26<68:25:12, 4.04it/s, grad_norm=0.336, loss_final=0.497, loss_mean=0.832, loss_mean_cls=0.0794, proj_loss=-0.414][2026-03-26 12:48:29] Step: 4198, Training Logs: loss_final: 0.509933, loss_mean: 0.827666, proj_loss: -0.402101, loss_mean_cls: 0.084367, grad_norm: 0.512757 +Steps: 0%| | 4199/1000000 [17:26<68:26:59, 4.04it/s, grad_norm=0.513, loss_final=0.51, loss_mean=0.828, loss_mean_cls=0.0844, proj_loss=-0.402][2026-03-26 12:48:30] Step: 4199, Training Logs: loss_final: 0.497218, loss_mean: 0.822621, proj_loss: -0.407811, loss_mean_cls: 0.082408, grad_norm: 0.410976 +Steps: 0%| | 4200/1000000 [17:27<68:25:48, 4.04it/s, grad_norm=0.411, loss_final=0.497, loss_mean=0.823, loss_mean_cls=0.0824, proj_loss=-0.408][2026-03-26 12:48:30] Step: 4200, Training Logs: loss_final: 0.505846, loss_mean: 0.830752, proj_loss: -0.407541, loss_mean_cls: 0.082634, grad_norm: 0.575009 +Steps: 0%| | 4201/1000000 [17:27<68:27:23, 4.04it/s, grad_norm=0.575, loss_final=0.506, loss_mean=0.831, loss_mean_cls=0.0826, proj_loss=-0.408][2026-03-26 12:48:30] Step: 4201, Training Logs: loss_final: 0.504899, loss_mean: 0.830672, proj_loss: -0.407453, loss_mean_cls: 0.081679, grad_norm: 0.806079 +Steps: 0%| | 4202/1000000 [17:27<68:28:12, 4.04it/s, grad_norm=0.806, loss_final=0.505, loss_mean=0.831, loss_mean_cls=0.0817, proj_loss=-0.407][2026-03-26 12:48:30] Step: 4202, Training Logs: loss_final: 0.509016, loss_mean: 0.836856, proj_loss: -0.409725, loss_mean_cls: 0.081886, grad_norm: 0.432538 +Steps: 0%| | 4203/1000000 [17:27<68:26:48, 4.04it/s, grad_norm=0.433, loss_final=0.509, loss_mean=0.837, loss_mean_cls=0.0819, proj_loss=-0.41][2026-03-26 12:48:31] Step: 4203, Training Logs: loss_final: 0.514899, loss_mean: 0.852041, proj_loss: -0.416389, loss_mean_cls: 0.079247, grad_norm: 0.860143 +Steps: 0%| | 4204/1000000 [17:28<68:24:13, 4.04it/s, grad_norm=0.86, loss_final=0.515, loss_mean=0.852, loss_mean_cls=0.0792, proj_loss=-0.416][2026-03-26 12:48:31] Step: 4204, Training Logs: loss_final: 0.500597, loss_mean: 0.830692, proj_loss: -0.411080, loss_mean_cls: 0.080985, grad_norm: 0.741850 +Steps: 0%| | 4205/1000000 [17:28<68:26:02, 4.04it/s, grad_norm=0.742, loss_final=0.501, loss_mean=0.831, loss_mean_cls=0.081, proj_loss=-0.411][2026-03-26 12:48:31] Step: 4205, Training Logs: loss_final: 0.520412, loss_mean: 0.851444, proj_loss: -0.410907, loss_mean_cls: 0.079875, grad_norm: 0.435930 +Steps: 0%| | 4206/1000000 [17:28<69:21:36, 3.99it/s, grad_norm=0.436, loss_final=0.52, loss_mean=0.851, loss_mean_cls=0.0799, proj_loss=-0.411][2026-03-26 12:48:31] Step: 4206, Training Logs: loss_final: 0.501946, loss_mean: 0.838583, proj_loss: -0.416440, loss_mean_cls: 0.079804, grad_norm: 0.772985 +Steps: 0%| | 4207/1000000 [17:28<69:12:50, 4.00it/s, grad_norm=0.773, loss_final=0.502, loss_mean=0.839, loss_mean_cls=0.0798, proj_loss=-0.416][2026-03-26 12:48:32] Step: 4207, Training Logs: loss_final: 0.509309, loss_mean: 0.848694, proj_loss: -0.419456, loss_mean_cls: 0.080071, grad_norm: 0.305090 +Steps: 0%| | 4208/1000000 [17:29<68:58:04, 4.01it/s, grad_norm=0.305, loss_final=0.509, loss_mean=0.849, loss_mean_cls=0.0801, proj_loss=-0.419][2026-03-26 12:48:32] Step: 4208, Training Logs: loss_final: 0.517142, loss_mean: 0.824820, proj_loss: -0.390278, loss_mean_cls: 0.082599, grad_norm: 0.700096 +Steps: 0%| | 4209/1000000 [17:29<68:49:30, 4.02it/s, grad_norm=0.7, loss_final=0.517, loss_mean=0.825, loss_mean_cls=0.0826, proj_loss=-0.39][2026-03-26 12:48:32] Step: 4209, Training Logs: loss_final: 0.496341, loss_mean: 0.819899, proj_loss: -0.405274, loss_mean_cls: 0.081716, grad_norm: 0.398600 +Steps: 0%| | 4210/1000000 [17:29<68:42:15, 4.03it/s, grad_norm=0.399, loss_final=0.496, loss_mean=0.82, loss_mean_cls=0.0817, proj_loss=-0.405][2026-03-26 12:48:32] Step: 4210, Training Logs: loss_final: 0.506260, loss_mean: 0.828230, proj_loss: -0.403514, loss_mean_cls: 0.081543, grad_norm: 0.599009 +Steps: 0%| | 4211/1000000 [17:29<68:36:05, 4.03it/s, grad_norm=0.599, loss_final=0.506, loss_mean=0.828, loss_mean_cls=0.0815, proj_loss=-0.404][2026-03-26 12:48:33] Step: 4211, Training Logs: loss_final: 0.526524, loss_mean: 0.859232, proj_loss: -0.411561, loss_mean_cls: 0.078853, grad_norm: 0.613554 +Steps: 0%| | 4212/1000000 [17:30<68:32:25, 4.04it/s, grad_norm=0.614, loss_final=0.527, loss_mean=0.859, loss_mean_cls=0.0789, proj_loss=-0.412][2026-03-26 12:48:33] Step: 4212, Training Logs: loss_final: 0.502088, loss_mean: 0.834831, proj_loss: -0.411840, loss_mean_cls: 0.079096, grad_norm: 0.302842 +Steps: 0%| | 4213/1000000 [17:30<68:27:28, 4.04it/s, grad_norm=0.303, loss_final=0.502, loss_mean=0.835, loss_mean_cls=0.0791, proj_loss=-0.412][2026-03-26 12:48:33] Step: 4213, Training Logs: loss_final: 0.513583, loss_mean: 0.847339, proj_loss: -0.412909, loss_mean_cls: 0.079153, grad_norm: 0.474608 +Steps: 0%| | 4214/1000000 [17:30<68:26:40, 4.04it/s, grad_norm=0.475, loss_final=0.514, loss_mean=0.847, loss_mean_cls=0.0792, proj_loss=-0.413][2026-03-26 12:48:33] Step: 4214, Training Logs: loss_final: 0.518713, loss_mean: 0.837754, proj_loss: -0.402571, loss_mean_cls: 0.083530, grad_norm: 0.279194 +Steps: 0%| | 4215/1000000 [17:30<68:25:29, 4.04it/s, grad_norm=0.279, loss_final=0.519, loss_mean=0.838, loss_mean_cls=0.0835, proj_loss=-0.403][2026-03-26 12:48:34] Step: 4215, Training Logs: loss_final: 0.500689, loss_mean: 0.820166, proj_loss: -0.402433, loss_mean_cls: 0.082956, grad_norm: 0.483922 +Steps: 0%| | 4216/1000000 [17:31<68:24:36, 4.04it/s, grad_norm=0.484, loss_final=0.501, loss_mean=0.82, loss_mean_cls=0.083, proj_loss=-0.402][2026-03-26 12:48:34] Step: 4216, Training Logs: loss_final: 0.497819, loss_mean: 0.825446, proj_loss: -0.409601, loss_mean_cls: 0.081973, grad_norm: 0.590734 +Steps: 0%| | 4217/1000000 [17:31<68:23:00, 4.04it/s, grad_norm=0.591, loss_final=0.498, loss_mean=0.825, loss_mean_cls=0.082, proj_loss=-0.41][2026-03-26 12:48:34] Step: 4217, Training Logs: loss_final: 0.490627, loss_mean: 0.804258, proj_loss: -0.398776, loss_mean_cls: 0.085146, grad_norm: 0.261092 +Steps: 0%| | 4218/1000000 [17:31<68:23:38, 4.04it/s, grad_norm=0.261, loss_final=0.491, loss_mean=0.804, loss_mean_cls=0.0851, proj_loss=-0.399][2026-03-26 12:48:34] Step: 4218, Training Logs: loss_final: 0.505839, loss_mean: 0.839217, proj_loss: -0.413441, loss_mean_cls: 0.080063, grad_norm: 0.652689 +Steps: 0%| | 4219/1000000 [17:31<68:44:52, 4.02it/s, grad_norm=0.653, loss_final=0.506, loss_mean=0.839, loss_mean_cls=0.0801, proj_loss=-0.413][2026-03-26 12:48:35] Step: 4219, Training Logs: loss_final: 0.490999, loss_mean: 0.815465, proj_loss: -0.408264, loss_mean_cls: 0.083798, grad_norm: 0.520545 +Steps: 0%| | 4220/1000000 [17:32<68:19:24, 4.05it/s, grad_norm=0.521, loss_final=0.491, loss_mean=0.815, loss_mean_cls=0.0838, proj_loss=-0.408][2026-03-26 12:48:35] Step: 4220, Training Logs: loss_final: 0.512579, loss_mean: 0.852612, proj_loss: -0.418246, loss_mean_cls: 0.078213, grad_norm: 0.332359 +Steps: 0%| | 4221/1000000 [17:32<68:21:37, 4.05it/s, grad_norm=0.332, loss_final=0.513, loss_mean=0.853, loss_mean_cls=0.0782, proj_loss=-0.418][2026-03-26 12:48:35] Step: 4221, Training Logs: loss_final: 0.498961, loss_mean: 0.833042, proj_loss: -0.414337, loss_mean_cls: 0.080256, grad_norm: 0.476165 +Steps: 0%| | 4222/1000000 [17:32<68:25:03, 4.04it/s, grad_norm=0.476, loss_final=0.499, loss_mean=0.833, loss_mean_cls=0.0803, proj_loss=-0.414][2026-03-26 12:48:35] Step: 4222, Training Logs: loss_final: 0.506047, loss_mean: 0.808630, proj_loss: -0.389313, loss_mean_cls: 0.086731, grad_norm: 0.357666 +Steps: 0%| | 4223/1000000 [17:32<68:22:28, 4.05it/s, grad_norm=0.358, loss_final=0.506, loss_mean=0.809, loss_mean_cls=0.0867, proj_loss=-0.389][2026-03-26 12:48:36] Step: 4223, Training Logs: loss_final: 0.485867, loss_mean: 0.818894, proj_loss: -0.410238, loss_mean_cls: 0.077211, grad_norm: 0.583160 +Steps: 0%| | 4224/1000000 [17:33<68:23:41, 4.04it/s, grad_norm=0.583, loss_final=0.486, loss_mean=0.819, loss_mean_cls=0.0772, proj_loss=-0.41][2026-03-26 12:48:36] Step: 4224, Training Logs: loss_final: 0.501437, loss_mean: 0.837796, proj_loss: -0.415896, loss_mean_cls: 0.079538, grad_norm: 0.634198 +Steps: 0%| | 4225/1000000 [17:33<68:23:00, 4.04it/s, grad_norm=0.634, loss_final=0.501, loss_mean=0.838, loss_mean_cls=0.0795, proj_loss=-0.416][2026-03-26 12:48:36] Step: 4225, Training Logs: loss_final: 0.518295, loss_mean: 0.855252, proj_loss: -0.416026, loss_mean_cls: 0.079069, grad_norm: 0.379683 +Steps: 0%| | 4226/1000000 [17:33<68:24:52, 4.04it/s, grad_norm=0.38, loss_final=0.518, loss_mean=0.855, loss_mean_cls=0.0791, proj_loss=-0.416][2026-03-26 12:48:36] Step: 4226, Training Logs: loss_final: 0.511971, loss_mean: 0.835185, proj_loss: -0.405356, loss_mean_cls: 0.082142, grad_norm: 0.754916 +Steps: 0%| | 4227/1000000 [17:33<68:24:15, 4.04it/s, grad_norm=0.755, loss_final=0.512, loss_mean=0.835, loss_mean_cls=0.0821, proj_loss=-0.405][2026-03-26 12:48:37] Step: 4227, Training Logs: loss_final: 0.517500, loss_mean: 0.851362, proj_loss: -0.413100, loss_mean_cls: 0.079239, grad_norm: 0.396896 +Steps: 0%| | 4228/1000000 [17:34<68:24:27, 4.04it/s, grad_norm=0.397, loss_final=0.518, loss_mean=0.851, loss_mean_cls=0.0792, proj_loss=-0.413][2026-03-26 12:48:37] Step: 4228, Training Logs: loss_final: 0.499172, loss_mean: 0.831957, proj_loss: -0.412138, loss_mean_cls: 0.079354, grad_norm: 0.457841 +Steps: 0%| | 4229/1000000 [17:34<68:27:33, 4.04it/s, grad_norm=0.458, loss_final=0.499, loss_mean=0.832, loss_mean_cls=0.0794, proj_loss=-0.412][2026-03-26 12:48:37] Step: 4229, Training Logs: loss_final: 0.511291, loss_mean: 0.839298, proj_loss: -0.408790, loss_mean_cls: 0.080783, grad_norm: 0.636953 +Steps: 0%| | 4230/1000000 [17:34<68:26:18, 4.04it/s, grad_norm=0.637, loss_final=0.511, loss_mean=0.839, loss_mean_cls=0.0808, proj_loss=-0.409][2026-03-26 12:48:37] Step: 4230, Training Logs: loss_final: 0.513275, loss_mean: 0.842983, proj_loss: -0.410282, loss_mean_cls: 0.080574, grad_norm: 0.377089 +Steps: 0%| | 4231/1000000 [17:34<68:25:06, 4.04it/s, grad_norm=0.377, loss_final=0.513, loss_mean=0.843, loss_mean_cls=0.0806, proj_loss=-0.41][2026-03-26 12:48:38] Step: 4231, Training Logs: loss_final: 0.512423, loss_mean: 0.831786, proj_loss: -0.401265, loss_mean_cls: 0.081902, grad_norm: 0.531001 +Steps: 0%| | 4232/1000000 [17:35<68:26:02, 4.04it/s, grad_norm=0.531, loss_final=0.512, loss_mean=0.832, loss_mean_cls=0.0819, proj_loss=-0.401][2026-03-26 12:48:38] Step: 4232, Training Logs: loss_final: 0.509474, loss_mean: 0.845927, proj_loss: -0.416282, loss_mean_cls: 0.079829, grad_norm: 0.565635 +Steps: 0%| | 4233/1000000 [17:35<68:26:44, 4.04it/s, grad_norm=0.566, loss_final=0.509, loss_mean=0.846, loss_mean_cls=0.0798, proj_loss=-0.416][2026-03-26 12:48:38] Step: 4233, Training Logs: loss_final: 0.502065, loss_mean: 0.837340, proj_loss: -0.415400, loss_mean_cls: 0.080125, grad_norm: 0.234002 +Steps: 0%| | 4234/1000000 [17:35<68:26:01, 4.04it/s, grad_norm=0.234, loss_final=0.502, loss_mean=0.837, loss_mean_cls=0.0801, proj_loss=-0.415][2026-03-26 12:48:38] Step: 4234, Training Logs: loss_final: 0.507707, loss_mean: 0.831874, proj_loss: -0.405391, loss_mean_cls: 0.081225, grad_norm: 0.391070 +Steps: 0%| | 4235/1000000 [17:35<68:23:32, 4.04it/s, grad_norm=0.391, loss_final=0.508, loss_mean=0.832, loss_mean_cls=0.0812, proj_loss=-0.405][2026-03-26 12:48:39] Step: 4235, Training Logs: loss_final: 0.506980, loss_mean: 0.833169, proj_loss: -0.406284, loss_mean_cls: 0.080095, grad_norm: 0.399223 +Steps: 0%| | 4236/1000000 [17:36<68:23:30, 4.04it/s, grad_norm=0.399, loss_final=0.507, loss_mean=0.833, loss_mean_cls=0.0801, proj_loss=-0.406][2026-03-26 12:48:39] Step: 4236, Training Logs: loss_final: 0.525034, loss_mean: 0.843468, proj_loss: -0.400427, loss_mean_cls: 0.081992, grad_norm: 0.357927 +Steps: 0%| | 4237/1000000 [17:36<68:26:59, 4.04it/s, grad_norm=0.358, loss_final=0.525, loss_mean=0.843, loss_mean_cls=0.082, proj_loss=-0.4][2026-03-26 12:48:39] Step: 4237, Training Logs: loss_final: 0.508187, loss_mean: 0.835077, proj_loss: -0.407775, loss_mean_cls: 0.080886, grad_norm: 0.445363 +Steps: 0%| | 4238/1000000 [17:36<68:29:52, 4.04it/s, grad_norm=0.445, loss_final=0.508, loss_mean=0.835, loss_mean_cls=0.0809, proj_loss=-0.408][2026-03-26 12:48:39] Step: 4238, Training Logs: loss_final: 0.506988, loss_mean: 0.843882, proj_loss: -0.414256, loss_mean_cls: 0.077362, grad_norm: 0.388605 +Steps: 0%| | 4239/1000000 [17:36<68:28:03, 4.04it/s, grad_norm=0.389, loss_final=0.507, loss_mean=0.844, loss_mean_cls=0.0774, proj_loss=-0.414][2026-03-26 12:48:40] Step: 4239, Training Logs: loss_final: 0.515772, loss_mean: 0.841148, proj_loss: -0.406444, loss_mean_cls: 0.081068, grad_norm: 0.259647 +Steps: 0%| | 4240/1000000 [17:36<68:25:03, 4.04it/s, grad_norm=0.26, loss_final=0.516, loss_mean=0.841, loss_mean_cls=0.0811, proj_loss=-0.406][2026-03-26 12:48:40] Step: 4240, Training Logs: loss_final: 0.517098, loss_mean: 0.843260, proj_loss: -0.406525, loss_mean_cls: 0.080364, grad_norm: 0.244629 +Steps: 0%| | 4241/1000000 [17:37<68:25:19, 4.04it/s, grad_norm=0.245, loss_final=0.517, loss_mean=0.843, loss_mean_cls=0.0804, proj_loss=-0.407][2026-03-26 12:48:40] Step: 4241, Training Logs: loss_final: 0.509701, loss_mean: 0.850020, proj_loss: -0.419104, loss_mean_cls: 0.078785, grad_norm: 0.338203 +Steps: 0%| | 4242/1000000 [17:37<68:22:07, 4.05it/s, grad_norm=0.338, loss_final=0.51, loss_mean=0.85, loss_mean_cls=0.0788, proj_loss=-0.419][2026-03-26 12:48:40] Step: 4242, Training Logs: loss_final: 0.495387, loss_mean: 0.809736, proj_loss: -0.397738, loss_mean_cls: 0.083390, grad_norm: 0.309323 +Steps: 0%| | 4243/1000000 [17:37<68:20:05, 4.05it/s, grad_norm=0.309, loss_final=0.495, loss_mean=0.81, loss_mean_cls=0.0834, proj_loss=-0.398][2026-03-26 12:48:41] Step: 4243, Training Logs: loss_final: 0.501229, loss_mean: 0.838337, proj_loss: -0.416520, loss_mean_cls: 0.079412, grad_norm: 0.273932 +Steps: 0%| | 4244/1000000 [17:37<68:20:51, 4.05it/s, grad_norm=0.274, loss_final=0.501, loss_mean=0.838, loss_mean_cls=0.0794, proj_loss=-0.417][2026-03-26 12:48:41] Step: 4244, Training Logs: loss_final: 0.509183, loss_mean: 0.844200, proj_loss: -0.414259, loss_mean_cls: 0.079243, grad_norm: 0.335543 +Steps: 0%| | 4245/1000000 [17:38<68:21:22, 4.05it/s, grad_norm=0.336, loss_final=0.509, loss_mean=0.844, loss_mean_cls=0.0792, proj_loss=-0.414][2026-03-26 12:48:41] Step: 4245, Training Logs: loss_final: 0.513145, loss_mean: 0.840920, proj_loss: -0.408641, loss_mean_cls: 0.080867, grad_norm: 0.325580 +Steps: 0%| | 4246/1000000 [17:38<68:23:27, 4.04it/s, grad_norm=0.326, loss_final=0.513, loss_mean=0.841, loss_mean_cls=0.0809, proj_loss=-0.409][2026-03-26 12:48:41] Step: 4246, Training Logs: loss_final: 0.508558, loss_mean: 0.838767, proj_loss: -0.410754, loss_mean_cls: 0.080545, grad_norm: 0.315009 +Steps: 0%| | 4247/1000000 [17:38<68:23:45, 4.04it/s, grad_norm=0.315, loss_final=0.509, loss_mean=0.839, loss_mean_cls=0.0805, proj_loss=-0.411][2026-03-26 12:48:42] Step: 4247, Training Logs: loss_final: 0.488655, loss_mean: 0.822787, proj_loss: -0.413790, loss_mean_cls: 0.079658, grad_norm: 0.265102 +Steps: 0%| | 4248/1000000 [17:38<68:24:02, 4.04it/s, grad_norm=0.265, loss_final=0.489, loss_mean=0.823, loss_mean_cls=0.0797, proj_loss=-0.414][2026-03-26 12:48:42] Step: 4248, Training Logs: loss_final: 0.489418, loss_mean: 0.816388, proj_loss: -0.407408, loss_mean_cls: 0.080438, grad_norm: 0.457042 +Steps: 0%| | 4249/1000000 [17:39<68:22:49, 4.04it/s, grad_norm=0.457, loss_final=0.489, loss_mean=0.816, loss_mean_cls=0.0804, proj_loss=-0.407][2026-03-26 12:48:42] Step: 4249, Training Logs: loss_final: 0.519135, loss_mean: 0.853172, proj_loss: -0.412529, loss_mean_cls: 0.078492, grad_norm: 0.368225 +Steps: 0%| | 4250/1000000 [17:39<68:23:15, 4.04it/s, grad_norm=0.368, loss_final=0.519, loss_mean=0.853, loss_mean_cls=0.0785, proj_loss=-0.413][2026-03-26 12:48:42] Step: 4250, Training Logs: loss_final: 0.505813, loss_mean: 0.837094, proj_loss: -0.411061, loss_mean_cls: 0.079780, grad_norm: 0.276027 +Steps: 0%| | 4251/1000000 [17:39<68:22:30, 4.05it/s, grad_norm=0.276, loss_final=0.506, loss_mean=0.837, loss_mean_cls=0.0798, proj_loss=-0.411][2026-03-26 12:48:43] Step: 4251, Training Logs: loss_final: 0.508951, loss_mean: 0.837565, proj_loss: -0.409841, loss_mean_cls: 0.081227, grad_norm: 0.309136 +Steps: 0%| | 4252/1000000 [17:39<68:24:37, 4.04it/s, grad_norm=0.309, loss_final=0.509, loss_mean=0.838, loss_mean_cls=0.0812, proj_loss=-0.41][2026-03-26 12:48:43] Step: 4252, Training Logs: loss_final: 0.512267, loss_mean: 0.843137, proj_loss: -0.410228, loss_mean_cls: 0.079359, grad_norm: 0.377421 +Steps: 0%| | 4253/1000000 [17:40<68:25:41, 4.04it/s, grad_norm=0.377, loss_final=0.512, loss_mean=0.843, loss_mean_cls=0.0794, proj_loss=-0.41][2026-03-26 12:48:43] Step: 4253, Training Logs: loss_final: 0.494183, loss_mean: 0.820530, proj_loss: -0.408542, loss_mean_cls: 0.082194, grad_norm: 0.303363 +Steps: 0%| | 4254/1000000 [17:40<68:25:22, 4.04it/s, grad_norm=0.303, loss_final=0.494, loss_mean=0.821, loss_mean_cls=0.0822, proj_loss=-0.409][2026-03-26 12:48:43] Step: 4254, Training Logs: loss_final: 0.494068, loss_mean: 0.820499, proj_loss: -0.410271, loss_mean_cls: 0.083840, grad_norm: 0.363706 +Steps: 0%| | 4255/1000000 [17:40<68:23:31, 4.04it/s, grad_norm=0.364, loss_final=0.494, loss_mean=0.82, loss_mean_cls=0.0838, proj_loss=-0.41][2026-03-26 12:48:44] Step: 4255, Training Logs: loss_final: 0.507691, loss_mean: 0.845067, proj_loss: -0.416142, loss_mean_cls: 0.078766, grad_norm: 0.338279 +Steps: 0%| | 4256/1000000 [17:40<68:24:51, 4.04it/s, grad_norm=0.338, loss_final=0.508, loss_mean=0.845, loss_mean_cls=0.0788, proj_loss=-0.416][2026-03-26 12:48:44] Step: 4256, Training Logs: loss_final: 0.499701, loss_mean: 0.831073, proj_loss: -0.410417, loss_mean_cls: 0.079045, grad_norm: 0.269513 +Steps: 0%| | 4257/1000000 [17:41<68:21:12, 4.05it/s, grad_norm=0.27, loss_final=0.5, loss_mean=0.831, loss_mean_cls=0.079, proj_loss=-0.41][2026-03-26 12:48:44] Step: 4257, Training Logs: loss_final: 0.516300, loss_mean: 0.849972, proj_loss: -0.413344, loss_mean_cls: 0.079673, grad_norm: 0.387360 +Steps: 0%| | 4258/1000000 [17:41<68:23:16, 4.04it/s, grad_norm=0.387, loss_final=0.516, loss_mean=0.85, loss_mean_cls=0.0797, proj_loss=-0.413][2026-03-26 12:48:44] Step: 4258, Training Logs: loss_final: 0.491601, loss_mean: 0.819066, proj_loss: -0.407593, loss_mean_cls: 0.080129, grad_norm: 0.338678 +Steps: 0%| | 4259/1000000 [17:41<68:21:40, 4.05it/s, grad_norm=0.339, loss_final=0.492, loss_mean=0.819, loss_mean_cls=0.0801, proj_loss=-0.408][2026-03-26 12:48:45] Step: 4259, Training Logs: loss_final: 0.498365, loss_mean: 0.833941, proj_loss: -0.415210, loss_mean_cls: 0.079634, grad_norm: 0.351433 +Steps: 0%| | 4260/1000000 [17:41<68:22:22, 4.05it/s, grad_norm=0.351, loss_final=0.498, loss_mean=0.834, loss_mean_cls=0.0796, proj_loss=-0.415][2026-03-26 12:48:45] Step: 4260, Training Logs: loss_final: 0.506729, loss_mean: 0.835241, proj_loss: -0.409007, loss_mean_cls: 0.080495, grad_norm: 0.374714 +Steps: 0%| | 4261/1000000 [17:42<68:20:58, 4.05it/s, grad_norm=0.375, loss_final=0.507, loss_mean=0.835, loss_mean_cls=0.0805, proj_loss=-0.409][2026-03-26 12:48:45] Step: 4261, Training Logs: loss_final: 0.505657, loss_mean: 0.828580, proj_loss: -0.405574, loss_mean_cls: 0.082651, grad_norm: 0.297173 +Steps: 0%| | 4262/1000000 [17:42<68:23:49, 4.04it/s, grad_norm=0.297, loss_final=0.506, loss_mean=0.829, loss_mean_cls=0.0827, proj_loss=-0.406][2026-03-26 12:48:45] Step: 4262, Training Logs: loss_final: 0.520597, loss_mean: 0.858020, proj_loss: -0.416204, loss_mean_cls: 0.078781, grad_norm: 0.273383 +Steps: 0%| | 4263/1000000 [17:42<68:21:15, 4.05it/s, grad_norm=0.273, loss_final=0.521, loss_mean=0.858, loss_mean_cls=0.0788, proj_loss=-0.416][2026-03-26 12:48:46] Step: 4263, Training Logs: loss_final: 0.501972, loss_mean: 0.834508, proj_loss: -0.411237, loss_mean_cls: 0.078701, grad_norm: 0.313387 +Steps: 0%| | 4264/1000000 [17:42<68:20:54, 4.05it/s, grad_norm=0.313, loss_final=0.502, loss_mean=0.835, loss_mean_cls=0.0787, proj_loss=-0.411][2026-03-26 12:48:46] Step: 4264, Training Logs: loss_final: 0.504165, loss_mean: 0.831915, proj_loss: -0.408551, loss_mean_cls: 0.080801, grad_norm: 0.349540 +Steps: 0%| | 4265/1000000 [17:43<68:22:00, 4.05it/s, grad_norm=0.35, loss_final=0.504, loss_mean=0.832, loss_mean_cls=0.0808, proj_loss=-0.409][2026-03-26 12:48:46] Step: 4265, Training Logs: loss_final: 0.504912, loss_mean: 0.840650, proj_loss: -0.414404, loss_mean_cls: 0.078666, grad_norm: 0.270261 +Steps: 0%| | 4266/1000000 [17:43<68:21:57, 4.05it/s, grad_norm=0.27, loss_final=0.505, loss_mean=0.841, loss_mean_cls=0.0787, proj_loss=-0.414][2026-03-26 12:48:46] Step: 4266, Training Logs: loss_final: 0.499585, loss_mean: 0.832610, proj_loss: -0.414317, loss_mean_cls: 0.081292, grad_norm: 0.381507 +Steps: 0%| | 4267/1000000 [17:43<68:23:04, 4.04it/s, grad_norm=0.382, loss_final=0.5, loss_mean=0.833, loss_mean_cls=0.0813, proj_loss=-0.414][2026-03-26 12:48:47] Step: 4267, Training Logs: loss_final: 0.507470, loss_mean: 0.830622, proj_loss: -0.405764, loss_mean_cls: 0.082612, grad_norm: 0.447582 +Steps: 0%| | 4268/1000000 [17:43<68:26:43, 4.04it/s, grad_norm=0.448, loss_final=0.507, loss_mean=0.831, loss_mean_cls=0.0826, proj_loss=-0.406][2026-03-26 12:48:47] Step: 4268, Training Logs: loss_final: 0.495825, loss_mean: 0.827827, proj_loss: -0.412028, loss_mean_cls: 0.080026, grad_norm: 0.320486 +Steps: 0%| | 4269/1000000 [17:44<68:25:21, 4.04it/s, grad_norm=0.32, loss_final=0.496, loss_mean=0.828, loss_mean_cls=0.08, proj_loss=-0.412][2026-03-26 12:48:47] Step: 4269, Training Logs: loss_final: 0.503152, loss_mean: 0.843532, proj_loss: -0.416798, loss_mean_cls: 0.076418, grad_norm: 0.387213 +Steps: 0%| | 4270/1000000 [17:44<68:25:53, 4.04it/s, grad_norm=0.387, loss_final=0.503, loss_mean=0.844, loss_mean_cls=0.0764, proj_loss=-0.417][2026-03-26 12:48:47] Step: 4270, Training Logs: loss_final: 0.505233, loss_mean: 0.841158, proj_loss: -0.415051, loss_mean_cls: 0.079126, grad_norm: 0.314230 +Steps: 0%| | 4271/1000000 [17:44<68:22:50, 4.04it/s, grad_norm=0.314, loss_final=0.505, loss_mean=0.841, loss_mean_cls=0.0791, proj_loss=-0.415][2026-03-26 12:48:48] Step: 4271, Training Logs: loss_final: 0.502722, loss_mean: 0.829110, proj_loss: -0.408469, loss_mean_cls: 0.082080, grad_norm: 0.252015 +Steps: 0%| | 4272/1000000 [17:44<68:24:02, 4.04it/s, grad_norm=0.252, loss_final=0.503, loss_mean=0.829, loss_mean_cls=0.0821, proj_loss=-0.408][2026-03-26 12:48:48] Step: 4272, Training Logs: loss_final: 0.517447, loss_mean: 0.837227, proj_loss: -0.401786, loss_mean_cls: 0.082007, grad_norm: 0.390827 +Steps: 0%| | 4273/1000000 [17:45<68:25:22, 4.04it/s, grad_norm=0.391, loss_final=0.517, loss_mean=0.837, loss_mean_cls=0.082, proj_loss=-0.402][2026-03-26 12:48:48] Step: 4273, Training Logs: loss_final: 0.482952, loss_mean: 0.812627, proj_loss: -0.410432, loss_mean_cls: 0.080757, grad_norm: 0.342503 +Steps: 0%| | 4274/1000000 [17:45<68:25:31, 4.04it/s, grad_norm=0.343, loss_final=0.483, loss_mean=0.813, loss_mean_cls=0.0808, proj_loss=-0.41][2026-03-26 12:48:48] Step: 4274, Training Logs: loss_final: 0.500903, loss_mean: 0.827145, proj_loss: -0.408385, loss_mean_cls: 0.082144, grad_norm: 0.362524 +Steps: 0%| | 4275/1000000 [17:45<68:24:25, 4.04it/s, grad_norm=0.363, loss_final=0.501, loss_mean=0.827, loss_mean_cls=0.0821, proj_loss=-0.408][2026-03-26 12:48:49] Step: 4275, Training Logs: loss_final: 0.493945, loss_mean: 0.818865, proj_loss: -0.406382, loss_mean_cls: 0.081462, grad_norm: 0.242496 +Steps: 0%| | 4276/1000000 [17:45<68:22:25, 4.05it/s, grad_norm=0.242, loss_final=0.494, loss_mean=0.819, loss_mean_cls=0.0815, proj_loss=-0.406][2026-03-26 12:48:49] Step: 4276, Training Logs: loss_final: 0.510318, loss_mean: 0.856802, proj_loss: -0.423310, loss_mean_cls: 0.076825, grad_norm: 0.272617 +Steps: 0%| | 4277/1000000 [17:46<68:25:57, 4.04it/s, grad_norm=0.273, loss_final=0.51, loss_mean=0.857, loss_mean_cls=0.0768, proj_loss=-0.423][2026-03-26 12:48:49] Step: 4277, Training Logs: loss_final: 0.491532, loss_mean: 0.827712, proj_loss: -0.415285, loss_mean_cls: 0.079104, grad_norm: 0.363477 +Steps: 0%| | 4278/1000000 [17:46<68:25:56, 4.04it/s, grad_norm=0.363, loss_final=0.492, loss_mean=0.828, loss_mean_cls=0.0791, proj_loss=-0.415][2026-03-26 12:48:49] Step: 4278, Training Logs: loss_final: 0.491195, loss_mean: 0.822674, proj_loss: -0.414063, loss_mean_cls: 0.082585, grad_norm: 0.322087 +Steps: 0%| | 4279/1000000 [17:46<68:25:50, 4.04it/s, grad_norm=0.322, loss_final=0.491, loss_mean=0.823, loss_mean_cls=0.0826, proj_loss=-0.414][2026-03-26 12:48:50] Step: 4279, Training Logs: loss_final: 0.511529, loss_mean: 0.838348, proj_loss: -0.407499, loss_mean_cls: 0.080680, grad_norm: 0.475845 +Steps: 0%| | 4280/1000000 [17:46<68:24:48, 4.04it/s, grad_norm=0.476, loss_final=0.512, loss_mean=0.838, loss_mean_cls=0.0807, proj_loss=-0.407][2026-03-26 12:48:50] Step: 4280, Training Logs: loss_final: 0.507948, loss_mean: 0.840509, proj_loss: -0.412669, loss_mean_cls: 0.080108, grad_norm: 0.271508 +Steps: 0%| | 4281/1000000 [17:47<68:27:43, 4.04it/s, grad_norm=0.272, loss_final=0.508, loss_mean=0.841, loss_mean_cls=0.0801, proj_loss=-0.413][2026-03-26 12:48:50] Step: 4281, Training Logs: loss_final: 0.488191, loss_mean: 0.821157, proj_loss: -0.413798, loss_mean_cls: 0.080832, grad_norm: 0.304482 +Steps: 0%| | 4282/1000000 [17:47<68:30:34, 4.04it/s, grad_norm=0.304, loss_final=0.488, loss_mean=0.821, loss_mean_cls=0.0808, proj_loss=-0.414][2026-03-26 12:48:50] Step: 4282, Training Logs: loss_final: 0.510473, loss_mean: 0.852246, proj_loss: -0.417098, loss_mean_cls: 0.075326, grad_norm: 0.415534 +Steps: 0%| | 4283/1000000 [17:47<68:26:58, 4.04it/s, grad_norm=0.416, loss_final=0.51, loss_mean=0.852, loss_mean_cls=0.0753, proj_loss=-0.417][2026-03-26 12:48:51] Step: 4283, Training Logs: loss_final: 0.493780, loss_mean: 0.821793, proj_loss: -0.408265, loss_mean_cls: 0.080252, grad_norm: 0.506382 +Steps: 0%| | 4284/1000000 [17:47<68:24:55, 4.04it/s, grad_norm=0.506, loss_final=0.494, loss_mean=0.822, loss_mean_cls=0.0803, proj_loss=-0.408][2026-03-26 12:48:51] Step: 4284, Training Logs: loss_final: 0.492042, loss_mean: 0.818498, proj_loss: -0.409113, loss_mean_cls: 0.082657, grad_norm: 0.496823 +Steps: 0%| | 4285/1000000 [17:48<68:24:45, 4.04it/s, grad_norm=0.497, loss_final=0.492, loss_mean=0.818, loss_mean_cls=0.0827, proj_loss=-0.409][2026-03-26 12:48:51] Step: 4285, Training Logs: loss_final: 0.513111, loss_mean: 0.850510, proj_loss: -0.415855, loss_mean_cls: 0.078455, grad_norm: 0.375750 +Steps: 0%| | 4286/1000000 [17:48<68:24:30, 4.04it/s, grad_norm=0.376, loss_final=0.513, loss_mean=0.851, loss_mean_cls=0.0785, proj_loss=-0.416][2026-03-26 12:48:51] Step: 4286, Training Logs: loss_final: 0.499616, loss_mean: 0.827388, proj_loss: -0.409112, loss_mean_cls: 0.081340, grad_norm: 0.619321 +Steps: 0%| | 4287/1000000 [17:48<68:22:03, 4.05it/s, grad_norm=0.619, loss_final=0.5, loss_mean=0.827, loss_mean_cls=0.0813, proj_loss=-0.409][2026-03-26 12:48:52] Step: 4287, Training Logs: loss_final: 0.498783, loss_mean: 0.838009, proj_loss: -0.419443, loss_mean_cls: 0.080218, grad_norm: 0.490998 +Steps: 0%| | 4288/1000000 [17:48<68:24:21, 4.04it/s, grad_norm=0.491, loss_final=0.499, loss_mean=0.838, loss_mean_cls=0.0802, proj_loss=-0.419][2026-03-26 12:48:52] Step: 4288, Training Logs: loss_final: 0.486034, loss_mean: 0.823668, proj_loss: -0.417109, loss_mean_cls: 0.079475, grad_norm: 0.309838 +Steps: 0%| | 4289/1000000 [17:49<68:25:53, 4.04it/s, grad_norm=0.31, loss_final=0.486, loss_mean=0.824, loss_mean_cls=0.0795, proj_loss=-0.417][2026-03-26 12:48:52] Step: 4289, Training Logs: loss_final: 0.509820, loss_mean: 0.836589, proj_loss: -0.406546, loss_mean_cls: 0.079777, grad_norm: 0.480271 +Steps: 0%| | 4290/1000000 [17:49<68:26:48, 4.04it/s, grad_norm=0.48, loss_final=0.51, loss_mean=0.837, loss_mean_cls=0.0798, proj_loss=-0.407][2026-03-26 12:48:52] Step: 4290, Training Logs: loss_final: 0.495112, loss_mean: 0.808941, proj_loss: -0.398009, loss_mean_cls: 0.084180, grad_norm: 0.410447 +Steps: 0%| | 4291/1000000 [17:49<68:26:52, 4.04it/s, grad_norm=0.41, loss_final=0.495, loss_mean=0.809, loss_mean_cls=0.0842, proj_loss=-0.398][2026-03-26 12:48:53] Step: 4291, Training Logs: loss_final: 0.508093, loss_mean: 0.837632, proj_loss: -0.409170, loss_mean_cls: 0.079631, grad_norm: 0.420192 +Steps: 0%| | 4292/1000000 [17:49<69:12:22, 4.00it/s, grad_norm=0.42, loss_final=0.508, loss_mean=0.838, loss_mean_cls=0.0796, proj_loss=-0.409][2026-03-26 12:48:53] Step: 4292, Training Logs: loss_final: 0.489248, loss_mean: 0.817780, proj_loss: -0.410639, loss_mean_cls: 0.082107, grad_norm: 0.691900 +Steps: 0%| | 4293/1000000 [17:50<68:58:25, 4.01it/s, grad_norm=0.692, loss_final=0.489, loss_mean=0.818, loss_mean_cls=0.0821, proj_loss=-0.411][2026-03-26 12:48:53] Step: 4293, Training Logs: loss_final: 0.515960, loss_mean: 0.841118, proj_loss: -0.405155, loss_mean_cls: 0.079997, grad_norm: 0.375966 +Steps: 0%| | 4294/1000000 [17:50<68:48:37, 4.02it/s, grad_norm=0.376, loss_final=0.516, loss_mean=0.841, loss_mean_cls=0.08, proj_loss=-0.405][2026-03-26 12:48:53] Step: 4294, Training Logs: loss_final: 0.507928, loss_mean: 0.838116, proj_loss: -0.411415, loss_mean_cls: 0.081228, grad_norm: 0.516700 +Steps: 0%| | 4295/1000000 [17:50<68:40:09, 4.03it/s, grad_norm=0.517, loss_final=0.508, loss_mean=0.838, loss_mean_cls=0.0812, proj_loss=-0.411][2026-03-26 12:48:54] Step: 4295, Training Logs: loss_final: 0.502767, loss_mean: 0.818348, proj_loss: -0.398898, loss_mean_cls: 0.083316, grad_norm: 0.720195 +Steps: 0%| | 4296/1000000 [17:50<68:36:10, 4.03it/s, grad_norm=0.72, loss_final=0.503, loss_mean=0.818, loss_mean_cls=0.0833, proj_loss=-0.399][2026-03-26 12:48:54] Step: 4296, Training Logs: loss_final: 0.501923, loss_mean: 0.825185, proj_loss: -0.405262, loss_mean_cls: 0.081999, grad_norm: 0.367807 +Steps: 0%| | 4297/1000000 [17:51<68:30:36, 4.04it/s, grad_norm=0.368, loss_final=0.502, loss_mean=0.825, loss_mean_cls=0.082, proj_loss=-0.405][2026-03-26 12:48:54] Step: 4297, Training Logs: loss_final: 0.503715, loss_mean: 0.819602, proj_loss: -0.399600, loss_mean_cls: 0.083713, grad_norm: 0.733105 +Steps: 0%| | 4298/1000000 [17:51<68:31:21, 4.04it/s, grad_norm=0.733, loss_final=0.504, loss_mean=0.82, loss_mean_cls=0.0837, proj_loss=-0.4][2026-03-26 12:48:54] Step: 4298, Training Logs: loss_final: 0.507482, loss_mean: 0.828855, proj_loss: -0.402466, loss_mean_cls: 0.081094, grad_norm: 0.388657 +Steps: 0%| | 4299/1000000 [17:51<68:26:18, 4.04it/s, grad_norm=0.389, loss_final=0.507, loss_mean=0.829, loss_mean_cls=0.0811, proj_loss=-0.402][2026-03-26 12:48:54] Step: 4299, Training Logs: loss_final: 0.502626, loss_mean: 0.835808, proj_loss: -0.413967, loss_mean_cls: 0.080785, grad_norm: 0.608242 +Steps: 0%| | 4300/1000000 [17:51<68:29:34, 4.04it/s, grad_norm=0.608, loss_final=0.503, loss_mean=0.836, loss_mean_cls=0.0808, proj_loss=-0.414][2026-03-26 12:48:55] Step: 4300, Training Logs: loss_final: 0.520288, loss_mean: 0.847687, proj_loss: -0.407520, loss_mean_cls: 0.080121, grad_norm: 0.685290 +Steps: 0%| | 4301/1000000 [17:52<68:28:24, 4.04it/s, grad_norm=0.685, loss_final=0.52, loss_mean=0.848, loss_mean_cls=0.0801, proj_loss=-0.408][2026-03-26 12:48:55] Step: 4301, Training Logs: loss_final: 0.506871, loss_mean: 0.834614, proj_loss: -0.408917, loss_mean_cls: 0.081174, grad_norm: 0.268683 +Steps: 0%| | 4302/1000000 [17:52<68:26:56, 4.04it/s, grad_norm=0.269, loss_final=0.507, loss_mean=0.835, loss_mean_cls=0.0812, proj_loss=-0.409][2026-03-26 12:48:55] Step: 4302, Training Logs: loss_final: 0.494095, loss_mean: 0.827513, proj_loss: -0.414623, loss_mean_cls: 0.081205, grad_norm: 0.709534 +Steps: 0%| | 4303/1000000 [17:52<68:25:42, 4.04it/s, grad_norm=0.71, loss_final=0.494, loss_mean=0.828, loss_mean_cls=0.0812, proj_loss=-0.415][2026-03-26 12:48:55] Step: 4303, Training Logs: loss_final: 0.501663, loss_mean: 0.827853, proj_loss: -0.406684, loss_mean_cls: 0.080495, grad_norm: 0.734423 +Steps: 0%| | 4304/1000000 [17:52<68:27:48, 4.04it/s, grad_norm=0.734, loss_final=0.502, loss_mean=0.828, loss_mean_cls=0.0805, proj_loss=-0.407][2026-03-26 12:48:56] Step: 4304, Training Logs: loss_final: 0.498892, loss_mean: 0.830527, proj_loss: -0.411667, loss_mean_cls: 0.080032, grad_norm: 0.318916 +Steps: 0%| | 4305/1000000 [17:53<68:25:11, 4.04it/s, grad_norm=0.319, loss_final=0.499, loss_mean=0.831, loss_mean_cls=0.08, proj_loss=-0.412][2026-03-26 12:48:56] Step: 4305, Training Logs: loss_final: 0.489347, loss_mean: 0.816101, proj_loss: -0.408710, loss_mean_cls: 0.081956, grad_norm: 0.514042 +Steps: 0%| | 4306/1000000 [17:53<68:25:22, 4.04it/s, grad_norm=0.514, loss_final=0.489, loss_mean=0.816, loss_mean_cls=0.082, proj_loss=-0.409][2026-03-26 12:48:56] Step: 4306, Training Logs: loss_final: 0.487612, loss_mean: 0.828007, proj_loss: -0.419604, loss_mean_cls: 0.079209, grad_norm: 0.417917 +Steps: 0%| | 4307/1000000 [17:53<68:23:04, 4.04it/s, grad_norm=0.418, loss_final=0.488, loss_mean=0.828, loss_mean_cls=0.0792, proj_loss=-0.42][2026-03-26 12:48:56] Step: 4307, Training Logs: loss_final: 0.507151, loss_mean: 0.834237, proj_loss: -0.408357, loss_mean_cls: 0.081271, grad_norm: 0.319332 +Steps: 0%| | 4308/1000000 [17:53<68:25:06, 4.04it/s, grad_norm=0.319, loss_final=0.507, loss_mean=0.834, loss_mean_cls=0.0813, proj_loss=-0.408][2026-03-26 12:48:57] Step: 4308, Training Logs: loss_final: 0.504829, loss_mean: 0.839922, proj_loss: -0.414053, loss_mean_cls: 0.078960, grad_norm: 0.536931 +Steps: 0%| | 4309/1000000 [17:54<68:24:00, 4.04it/s, grad_norm=0.537, loss_final=0.505, loss_mean=0.84, loss_mean_cls=0.079, proj_loss=-0.414][2026-03-26 12:48:57] Step: 4309, Training Logs: loss_final: 0.502049, loss_mean: 0.827259, proj_loss: -0.406907, loss_mean_cls: 0.081698, grad_norm: 0.315099 +Steps: 0%| | 4310/1000000 [17:54<68:24:05, 4.04it/s, grad_norm=0.315, loss_final=0.502, loss_mean=0.827, loss_mean_cls=0.0817, proj_loss=-0.407][2026-03-26 12:48:57] Step: 4310, Training Logs: loss_final: 0.505409, loss_mean: 0.853998, proj_loss: -0.426303, loss_mean_cls: 0.077715, grad_norm: 0.517837 +Steps: 0%| | 4311/1000000 [17:54<68:24:21, 4.04it/s, grad_norm=0.518, loss_final=0.505, loss_mean=0.854, loss_mean_cls=0.0777, proj_loss=-0.426][2026-03-26 12:48:57] Step: 4311, Training Logs: loss_final: 0.496998, loss_mean: 0.816645, proj_loss: -0.401915, loss_mean_cls: 0.082267, grad_norm: 0.371860 +Steps: 0%| | 4312/1000000 [17:54<68:24:49, 4.04it/s, grad_norm=0.372, loss_final=0.497, loss_mean=0.817, loss_mean_cls=0.0823, proj_loss=-0.402][2026-03-26 12:48:58] Step: 4312, Training Logs: loss_final: 0.519613, loss_mean: 0.847281, proj_loss: -0.408679, loss_mean_cls: 0.081011, grad_norm: 0.512146 +Steps: 0%| | 4313/1000000 [17:55<68:24:02, 4.04it/s, grad_norm=0.512, loss_final=0.52, loss_mean=0.847, loss_mean_cls=0.081, proj_loss=-0.409][2026-03-26 12:48:58] Step: 4313, Training Logs: loss_final: 0.510283, loss_mean: 0.830230, proj_loss: -0.403143, loss_mean_cls: 0.083196, grad_norm: 0.438881 +Steps: 0%| | 4314/1000000 [17:55<68:24:29, 4.04it/s, grad_norm=0.439, loss_final=0.51, loss_mean=0.83, loss_mean_cls=0.0832, proj_loss=-0.403][2026-03-26 12:48:58] Step: 4314, Training Logs: loss_final: 0.489597, loss_mean: 0.816020, proj_loss: -0.409213, loss_mean_cls: 0.082790, grad_norm: 0.529309 +Steps: 0%| | 4315/1000000 [17:55<68:24:56, 4.04it/s, grad_norm=0.529, loss_final=0.49, loss_mean=0.816, loss_mean_cls=0.0828, proj_loss=-0.409][2026-03-26 12:48:58] Step: 4315, Training Logs: loss_final: 0.513359, loss_mean: 0.846590, proj_loss: -0.412642, loss_mean_cls: 0.079411, grad_norm: 0.806543 +Steps: 0%| | 4316/1000000 [17:55<73:50:16, 3.75it/s, grad_norm=0.807, loss_final=0.513, loss_mean=0.847, loss_mean_cls=0.0794, proj_loss=-0.413][2026-03-26 12:48:59] Step: 4316, Training Logs: loss_final: 0.507540, loss_mean: 0.837754, proj_loss: -0.411016, loss_mean_cls: 0.080802, grad_norm: 0.369340 +Steps: 0%| | 4317/1000000 [17:56<72:15:42, 3.83it/s, grad_norm=0.369, loss_final=0.508, loss_mean=0.838, loss_mean_cls=0.0808, proj_loss=-0.411][2026-03-26 12:48:59] Step: 4317, Training Logs: loss_final: 0.508692, loss_mean: 0.834070, proj_loss: -0.407413, loss_mean_cls: 0.082035, grad_norm: 0.504144 +Steps: 0%| | 4318/1000000 [17:56<71:07:38, 3.89it/s, grad_norm=0.504, loss_final=0.509, loss_mean=0.834, loss_mean_cls=0.082, proj_loss=-0.407][2026-03-26 12:48:59] Step: 4318, Training Logs: loss_final: 0.499750, loss_mean: 0.841247, proj_loss: -0.419498, loss_mean_cls: 0.078001, grad_norm: 0.603152 +Steps: 0%| | 4319/1000000 [17:56<70:17:51, 3.93it/s, grad_norm=0.603, loss_final=0.5, loss_mean=0.841, loss_mean_cls=0.078, proj_loss=-0.419][2026-03-26 12:49:00] Step: 4319, Training Logs: loss_final: 0.510889, loss_mean: 0.842115, proj_loss: -0.412429, loss_mean_cls: 0.081203, grad_norm: 0.305978 +Steps: 0%| | 4320/1000000 [17:56<69:44:17, 3.97it/s, grad_norm=0.306, loss_final=0.511, loss_mean=0.842, loss_mean_cls=0.0812, proj_loss=-0.412][2026-03-26 12:49:00] Step: 4320, Training Logs: loss_final: 0.498826, loss_mean: 0.840663, proj_loss: -0.419644, loss_mean_cls: 0.077807, grad_norm: 0.472731 +Steps: 0%| | 4321/1000000 [17:57<69:19:05, 3.99it/s, grad_norm=0.473, loss_final=0.499, loss_mean=0.841, loss_mean_cls=0.0778, proj_loss=-0.42][2026-03-26 12:49:00] Step: 4321, Training Logs: loss_final: 0.495550, loss_mean: 0.831726, proj_loss: -0.415386, loss_mean_cls: 0.079210, grad_norm: 0.353123 +Steps: 0%| | 4322/1000000 [17:57<69:02:28, 4.01it/s, grad_norm=0.353, loss_final=0.496, loss_mean=0.832, loss_mean_cls=0.0792, proj_loss=-0.415][2026-03-26 12:49:00] Step: 4322, Training Logs: loss_final: 0.492419, loss_mean: 0.823693, proj_loss: -0.412273, loss_mean_cls: 0.080999, grad_norm: 0.536367 +Steps: 0%| | 4323/1000000 [17:57<68:47:35, 4.02it/s, grad_norm=0.536, loss_final=0.492, loss_mean=0.824, loss_mean_cls=0.081, proj_loss=-0.412][2026-03-26 12:49:00] Step: 4323, Training Logs: loss_final: 0.488219, loss_mean: 0.821685, proj_loss: -0.414111, loss_mean_cls: 0.080645, grad_norm: 0.555660 +Steps: 0%| | 4324/1000000 [17:57<68:39:25, 4.03it/s, grad_norm=0.556, loss_final=0.488, loss_mean=0.822, loss_mean_cls=0.0806, proj_loss=-0.414][2026-03-26 12:49:01] Step: 4324, Training Logs: loss_final: 0.510093, loss_mean: 0.843920, proj_loss: -0.412987, loss_mean_cls: 0.079161, grad_norm: 0.305400 +Steps: 0%| | 4325/1000000 [17:58<68:35:39, 4.03it/s, grad_norm=0.305, loss_final=0.51, loss_mean=0.844, loss_mean_cls=0.0792, proj_loss=-0.413][2026-03-26 12:49:01] Step: 4325, Training Logs: loss_final: 0.504352, loss_mean: 0.836398, proj_loss: -0.413193, loss_mean_cls: 0.081146, grad_norm: 0.464575 +Steps: 0%| | 4326/1000000 [17:58<68:36:18, 4.03it/s, grad_norm=0.465, loss_final=0.504, loss_mean=0.836, loss_mean_cls=0.0811, proj_loss=-0.413][2026-03-26 12:49:01] Step: 4326, Training Logs: loss_final: 0.502042, loss_mean: 0.820202, proj_loss: -0.401600, loss_mean_cls: 0.083441, grad_norm: 0.549644 +Steps: 0%| | 4327/1000000 [17:58<68:30:46, 4.04it/s, grad_norm=0.55, loss_final=0.502, loss_mean=0.82, loss_mean_cls=0.0834, proj_loss=-0.402][2026-03-26 12:49:01] Step: 4327, Training Logs: loss_final: 0.509490, loss_mean: 0.834380, proj_loss: -0.406204, loss_mean_cls: 0.081315, grad_norm: 0.311938 +Steps: 0%| | 4328/1000000 [17:58<68:29:52, 4.04it/s, grad_norm=0.312, loss_final=0.509, loss_mean=0.834, loss_mean_cls=0.0813, proj_loss=-0.406][2026-03-26 12:49:02] Step: 4328, Training Logs: loss_final: 0.501326, loss_mean: 0.842407, proj_loss: -0.418923, loss_mean_cls: 0.077842, grad_norm: 0.519257 +Steps: 0%| | 4329/1000000 [17:59<68:26:18, 4.04it/s, grad_norm=0.519, loss_final=0.501, loss_mean=0.842, loss_mean_cls=0.0778, proj_loss=-0.419][2026-03-26 12:49:02] Step: 4329, Training Logs: loss_final: 0.493700, loss_mean: 0.824000, proj_loss: -0.411600, loss_mean_cls: 0.081299, grad_norm: 0.546975 +Steps: 0%| | 4330/1000000 [17:59<68:29:49, 4.04it/s, grad_norm=0.547, loss_final=0.494, loss_mean=0.824, loss_mean_cls=0.0813, proj_loss=-0.412][2026-03-26 12:49:02] Step: 4330, Training Logs: loss_final: 0.504904, loss_mean: 0.839053, proj_loss: -0.414974, loss_mean_cls: 0.080825, grad_norm: 0.389594 +Steps: 0%| | 4331/1000000 [17:59<68:27:53, 4.04it/s, grad_norm=0.39, loss_final=0.505, loss_mean=0.839, loss_mean_cls=0.0808, proj_loss=-0.415][2026-03-26 12:49:02] Step: 4331, Training Logs: loss_final: 0.521051, loss_mean: 0.862032, proj_loss: -0.418831, loss_mean_cls: 0.077851, grad_norm: 0.446581 +Steps: 0%| | 4332/1000000 [17:59<68:26:20, 4.04it/s, grad_norm=0.447, loss_final=0.521, loss_mean=0.862, loss_mean_cls=0.0779, proj_loss=-0.419][2026-03-26 12:49:03] Step: 4332, Training Logs: loss_final: 0.504477, loss_mean: 0.821431, proj_loss: -0.401249, loss_mean_cls: 0.084295, grad_norm: 0.451998 +Steps: 0%| | 4333/1000000 [18:00<68:26:00, 4.04it/s, grad_norm=0.452, loss_final=0.504, loss_mean=0.821, loss_mean_cls=0.0843, proj_loss=-0.401][2026-03-26 12:49:03] Step: 4333, Training Logs: loss_final: 0.507124, loss_mean: 0.841445, proj_loss: -0.413673, loss_mean_cls: 0.079352, grad_norm: 0.406528 +Steps: 0%| | 4334/1000000 [18:00<68:25:45, 4.04it/s, grad_norm=0.407, loss_final=0.507, loss_mean=0.841, loss_mean_cls=0.0794, proj_loss=-0.414][2026-03-26 12:49:03] Step: 4334, Training Logs: loss_final: 0.502583, loss_mean: 0.838259, proj_loss: -0.415543, loss_mean_cls: 0.079868, grad_norm: 0.302933 +Steps: 0%| | 4335/1000000 [18:00<68:24:14, 4.04it/s, grad_norm=0.303, loss_final=0.503, loss_mean=0.838, loss_mean_cls=0.0799, proj_loss=-0.416][2026-03-26 12:49:03] Step: 4335, Training Logs: loss_final: 0.501205, loss_mean: 0.825898, proj_loss: -0.406902, loss_mean_cls: 0.082209, grad_norm: 0.388634 +Steps: 0%| | 4336/1000000 [18:00<68:22:25, 4.05it/s, grad_norm=0.389, loss_final=0.501, loss_mean=0.826, loss_mean_cls=0.0822, proj_loss=-0.407][2026-03-26 12:49:04] Step: 4336, Training Logs: loss_final: 0.516959, loss_mean: 0.836487, proj_loss: -0.401729, loss_mean_cls: 0.082200, grad_norm: 0.476508 +Steps: 0%| | 4337/1000000 [18:01<68:23:03, 4.04it/s, grad_norm=0.477, loss_final=0.517, loss_mean=0.836, loss_mean_cls=0.0822, proj_loss=-0.402][2026-03-26 12:49:04] Step: 4337, Training Logs: loss_final: 0.521312, loss_mean: 0.851453, proj_loss: -0.410273, loss_mean_cls: 0.080131, grad_norm: 0.338778 +Steps: 0%| | 4338/1000000 [18:01<68:24:33, 4.04it/s, grad_norm=0.339, loss_final=0.521, loss_mean=0.851, loss_mean_cls=0.0801, proj_loss=-0.41][2026-03-26 12:49:04] Step: 4338, Training Logs: loss_final: 0.505761, loss_mean: 0.836214, proj_loss: -0.410491, loss_mean_cls: 0.080038, grad_norm: 0.434217 +Steps: 0%| | 4339/1000000 [18:01<68:25:59, 4.04it/s, grad_norm=0.434, loss_final=0.506, loss_mean=0.836, loss_mean_cls=0.08, proj_loss=-0.41][2026-03-26 12:49:04] Step: 4339, Training Logs: loss_final: 0.493134, loss_mean: 0.822632, proj_loss: -0.410218, loss_mean_cls: 0.080720, grad_norm: 0.507243 +Steps: 0%| | 4340/1000000 [18:01<68:26:24, 4.04it/s, grad_norm=0.507, loss_final=0.493, loss_mean=0.823, loss_mean_cls=0.0807, proj_loss=-0.41][2026-03-26 12:49:05] Step: 4340, Training Logs: loss_final: 0.513328, loss_mean: 0.850792, proj_loss: -0.417231, loss_mean_cls: 0.079767, grad_norm: 0.438381 +Steps: 0%| | 4341/1000000 [18:02<68:27:34, 4.04it/s, grad_norm=0.438, loss_final=0.513, loss_mean=0.851, loss_mean_cls=0.0798, proj_loss=-0.417][2026-03-26 12:49:05] Step: 4341, Training Logs: loss_final: 0.527104, loss_mean: 0.860095, proj_loss: -0.412229, loss_mean_cls: 0.079238, grad_norm: 0.618579 +Steps: 0%| | 4342/1000000 [18:02<68:28:19, 4.04it/s, grad_norm=0.619, loss_final=0.527, loss_mean=0.86, loss_mean_cls=0.0792, proj_loss=-0.412][2026-03-26 12:49:05] Step: 4342, Training Logs: loss_final: 0.500489, loss_mean: 0.832313, proj_loss: -0.411811, loss_mean_cls: 0.079987, grad_norm: 0.558549 +Steps: 0%| | 4343/1000000 [18:02<68:27:06, 4.04it/s, grad_norm=0.559, loss_final=0.5, loss_mean=0.832, loss_mean_cls=0.08, proj_loss=-0.412][2026-03-26 12:49:05] Step: 4343, Training Logs: loss_final: 0.513295, loss_mean: 0.843028, proj_loss: -0.409846, loss_mean_cls: 0.080113, grad_norm: 0.555273 +Steps: 0%| | 4344/1000000 [18:02<68:42:02, 4.03it/s, grad_norm=0.555, loss_final=0.513, loss_mean=0.843, loss_mean_cls=0.0801, proj_loss=-0.41][2026-03-26 12:49:06] Step: 4344, Training Logs: loss_final: 0.505704, loss_mean: 0.830559, proj_loss: -0.405627, loss_mean_cls: 0.080772, grad_norm: 0.606880 +Steps: 0%| | 4345/1000000 [18:03<68:39:27, 4.03it/s, grad_norm=0.607, loss_final=0.506, loss_mean=0.831, loss_mean_cls=0.0808, proj_loss=-0.406][2026-03-26 12:49:06] Step: 4345, Training Logs: loss_final: 0.493485, loss_mean: 0.814746, proj_loss: -0.403622, loss_mean_cls: 0.082361, grad_norm: 0.377689 +Steps: 0%| | 4346/1000000 [18:03<68:35:37, 4.03it/s, grad_norm=0.378, loss_final=0.493, loss_mean=0.815, loss_mean_cls=0.0824, proj_loss=-0.404][2026-03-26 12:49:06] Step: 4346, Training Logs: loss_final: 0.480614, loss_mean: 0.807255, proj_loss: -0.409555, loss_mean_cls: 0.082914, grad_norm: 0.635067 +Steps: 0%| | 4347/1000000 [18:03<68:31:06, 4.04it/s, grad_norm=0.635, loss_final=0.481, loss_mean=0.807, loss_mean_cls=0.0829, proj_loss=-0.41][2026-03-26 12:49:06] Step: 4347, Training Logs: loss_final: 0.497183, loss_mean: 0.827481, proj_loss: -0.410141, loss_mean_cls: 0.079843, grad_norm: 0.307442 +Steps: 0%| | 4348/1000000 [18:03<68:28:31, 4.04it/s, grad_norm=0.307, loss_final=0.497, loss_mean=0.827, loss_mean_cls=0.0798, proj_loss=-0.41][2026-03-26 12:49:07] Step: 4348, Training Logs: loss_final: 0.491984, loss_mean: 0.830671, proj_loss: -0.418349, loss_mean_cls: 0.079661, grad_norm: 0.691773 +Steps: 0%| | 4349/1000000 [18:04<68:29:54, 4.04it/s, grad_norm=0.692, loss_final=0.492, loss_mean=0.831, loss_mean_cls=0.0797, proj_loss=-0.418][2026-03-26 12:49:07] Step: 4349, Training Logs: loss_final: 0.505820, loss_mean: 0.837612, proj_loss: -0.411233, loss_mean_cls: 0.079441, grad_norm: 0.709144 +Steps: 0%| | 4350/1000000 [18:04<68:24:23, 4.04it/s, grad_norm=0.709, loss_final=0.506, loss_mean=0.838, loss_mean_cls=0.0794, proj_loss=-0.411][2026-03-26 12:49:07] Step: 4350, Training Logs: loss_final: 0.492733, loss_mean: 0.827279, proj_loss: -0.414621, loss_mean_cls: 0.080075, grad_norm: 0.299510 +Steps: 0%| | 4351/1000000 [18:04<68:25:31, 4.04it/s, grad_norm=0.3, loss_final=0.493, loss_mean=0.827, loss_mean_cls=0.0801, proj_loss=-0.415][2026-03-26 12:49:07] Step: 4351, Training Logs: loss_final: 0.506840, loss_mean: 0.844724, proj_loss: -0.416867, loss_mean_cls: 0.078984, grad_norm: 0.644419 +Steps: 0%| | 4352/1000000 [18:04<68:23:06, 4.04it/s, grad_norm=0.644, loss_final=0.507, loss_mean=0.845, loss_mean_cls=0.079, proj_loss=-0.417][2026-03-26 12:49:08] Step: 4352, Training Logs: loss_final: 0.489762, loss_mean: 0.816932, proj_loss: -0.410528, loss_mean_cls: 0.083358, grad_norm: 0.366074 +Steps: 0%| | 4353/1000000 [18:05<68:25:55, 4.04it/s, grad_norm=0.366, loss_final=0.49, loss_mean=0.817, loss_mean_cls=0.0834, proj_loss=-0.411][2026-03-26 12:49:08] Step: 4353, Training Logs: loss_final: 0.501923, loss_mean: 0.820261, proj_loss: -0.401936, loss_mean_cls: 0.083598, grad_norm: 0.433003 +Steps: 0%| | 4354/1000000 [18:05<68:25:03, 4.04it/s, grad_norm=0.433, loss_final=0.502, loss_mean=0.82, loss_mean_cls=0.0836, proj_loss=-0.402][2026-03-26 12:49:08] Step: 4354, Training Logs: loss_final: 0.502354, loss_mean: 0.847902, proj_loss: -0.422945, loss_mean_cls: 0.077397, grad_norm: 0.492024 +Steps: 0%| | 4355/1000000 [18:05<68:24:38, 4.04it/s, grad_norm=0.492, loss_final=0.502, loss_mean=0.848, loss_mean_cls=0.0774, proj_loss=-0.423][2026-03-26 12:49:08] Step: 4355, Training Logs: loss_final: 0.507039, loss_mean: 0.839425, proj_loss: -0.412672, loss_mean_cls: 0.080287, grad_norm: 0.300882 +Steps: 0%| | 4356/1000000 [18:05<68:23:24, 4.04it/s, grad_norm=0.301, loss_final=0.507, loss_mean=0.839, loss_mean_cls=0.0803, proj_loss=-0.413][2026-03-26 12:49:09] Step: 4356, Training Logs: loss_final: 0.497300, loss_mean: 0.820910, proj_loss: -0.405224, loss_mean_cls: 0.081614, grad_norm: 0.498140 +Steps: 0%| | 4357/1000000 [18:06<68:23:10, 4.04it/s, grad_norm=0.498, loss_final=0.497, loss_mean=0.821, loss_mean_cls=0.0816, proj_loss=-0.405][2026-03-26 12:49:09] Step: 4357, Training Logs: loss_final: 0.499777, loss_mean: 0.840989, proj_loss: -0.418706, loss_mean_cls: 0.077494, grad_norm: 0.352024 +Steps: 0%| | 4358/1000000 [18:06<68:23:37, 4.04it/s, grad_norm=0.352, loss_final=0.5, loss_mean=0.841, loss_mean_cls=0.0775, proj_loss=-0.419][2026-03-26 12:49:09] Step: 4358, Training Logs: loss_final: 0.495983, loss_mean: 0.823735, proj_loss: -0.408439, loss_mean_cls: 0.080687, grad_norm: 0.433944 +Steps: 0%| | 4359/1000000 [18:06<68:21:26, 4.05it/s, grad_norm=0.434, loss_final=0.496, loss_mean=0.824, loss_mean_cls=0.0807, proj_loss=-0.408][2026-03-26 12:49:09] Step: 4359, Training Logs: loss_final: 0.511213, loss_mean: 0.846345, proj_loss: -0.413996, loss_mean_cls: 0.078864, grad_norm: 0.433219 +Steps: 0%| | 4360/1000000 [18:06<68:28:20, 4.04it/s, grad_norm=0.433, loss_final=0.511, loss_mean=0.846, loss_mean_cls=0.0789, proj_loss=-0.414][2026-03-26 12:49:10] Step: 4360, Training Logs: loss_final: 0.519100, loss_mean: 0.843238, proj_loss: -0.404016, loss_mean_cls: 0.079878, grad_norm: 0.463532 +Steps: 0%| | 4361/1000000 [18:07<68:28:23, 4.04it/s, grad_norm=0.464, loss_final=0.519, loss_mean=0.843, loss_mean_cls=0.0799, proj_loss=-0.404][2026-03-26 12:49:10] Step: 4361, Training Logs: loss_final: 0.497100, loss_mean: 0.827332, proj_loss: -0.410952, loss_mean_cls: 0.080720, grad_norm: 0.524842 +Steps: 0%| | 4362/1000000 [18:07<68:28:36, 4.04it/s, grad_norm=0.525, loss_final=0.497, loss_mean=0.827, loss_mean_cls=0.0807, proj_loss=-0.411][2026-03-26 12:49:10] Step: 4362, Training Logs: loss_final: 0.496797, loss_mean: 0.822710, proj_loss: -0.408350, loss_mean_cls: 0.082437, grad_norm: 0.358778 +Steps: 0%| | 4363/1000000 [18:07<68:26:54, 4.04it/s, grad_norm=0.359, loss_final=0.497, loss_mean=0.823, loss_mean_cls=0.0824, proj_loss=-0.408][2026-03-26 12:49:10] Step: 4363, Training Logs: loss_final: 0.505894, loss_mean: 0.847953, proj_loss: -0.419044, loss_mean_cls: 0.076985, grad_norm: 0.516993 +Steps: 0%| | 4364/1000000 [18:07<68:25:58, 4.04it/s, grad_norm=0.517, loss_final=0.506, loss_mean=0.848, loss_mean_cls=0.077, proj_loss=-0.419][2026-03-26 12:49:11] Step: 4364, Training Logs: loss_final: 0.515291, loss_mean: 0.846972, proj_loss: -0.411203, loss_mean_cls: 0.079523, grad_norm: 0.462670 +Steps: 0%| | 4365/1000000 [18:07<68:33:12, 4.03it/s, grad_norm=0.463, loss_final=0.515, loss_mean=0.847, loss_mean_cls=0.0795, proj_loss=-0.411][2026-03-26 12:49:11] Step: 4365, Training Logs: loss_final: 0.503034, loss_mean: 0.830311, proj_loss: -0.407587, loss_mean_cls: 0.080309, grad_norm: 0.296252 +Steps: 0%| | 4366/1000000 [18:08<68:58:38, 4.01it/s, grad_norm=0.296, loss_final=0.503, loss_mean=0.83, loss_mean_cls=0.0803, proj_loss=-0.408][2026-03-26 12:49:11] Step: 4366, Training Logs: loss_final: 0.488003, loss_mean: 0.828985, proj_loss: -0.419596, loss_mean_cls: 0.078613, grad_norm: 0.740422 +Steps: 0%| | 4367/1000000 [18:08<68:48:24, 4.02it/s, grad_norm=0.74, loss_final=0.488, loss_mean=0.829, loss_mean_cls=0.0786, proj_loss=-0.42][2026-03-26 12:49:11] Step: 4367, Training Logs: loss_final: 0.490331, loss_mean: 0.829351, proj_loss: -0.416645, loss_mean_cls: 0.077625, grad_norm: 0.733865 +Steps: 0%| | 4368/1000000 [18:08<68:44:43, 4.02it/s, grad_norm=0.734, loss_final=0.49, loss_mean=0.829, loss_mean_cls=0.0776, proj_loss=-0.417][2026-03-26 12:49:12] Step: 4368, Training Logs: loss_final: 0.498390, loss_mean: 0.838589, proj_loss: -0.418121, loss_mean_cls: 0.077922, grad_norm: 0.226607 +Steps: 0%| | 4369/1000000 [18:08<68:40:33, 4.03it/s, grad_norm=0.227, loss_final=0.498, loss_mean=0.839, loss_mean_cls=0.0779, proj_loss=-0.418][2026-03-26 12:49:12] Step: 4369, Training Logs: loss_final: 0.501138, loss_mean: 0.841985, proj_loss: -0.419869, loss_mean_cls: 0.079021, grad_norm: 0.378556 +Steps: 0%| | 4370/1000000 [18:09<68:37:59, 4.03it/s, grad_norm=0.379, loss_final=0.501, loss_mean=0.842, loss_mean_cls=0.079, proj_loss=-0.42][2026-03-26 12:49:12] Step: 4370, Training Logs: loss_final: 0.491491, loss_mean: 0.816682, proj_loss: -0.408015, loss_mean_cls: 0.082824, grad_norm: 0.361586 +Steps: 0%| | 4371/1000000 [18:09<68:33:27, 4.03it/s, grad_norm=0.362, loss_final=0.491, loss_mean=0.817, loss_mean_cls=0.0828, proj_loss=-0.408][2026-03-26 12:49:12] Step: 4371, Training Logs: loss_final: 0.509353, loss_mean: 0.843197, proj_loss: -0.414313, loss_mean_cls: 0.080470, grad_norm: 0.290216 +Steps: 0%| | 4372/1000000 [18:09<68:32:33, 4.03it/s, grad_norm=0.29, loss_final=0.509, loss_mean=0.843, loss_mean_cls=0.0805, proj_loss=-0.414][2026-03-26 12:49:13] Step: 4372, Training Logs: loss_final: 0.500584, loss_mean: 0.814609, proj_loss: -0.396935, loss_mean_cls: 0.082910, grad_norm: 0.313231 +Steps: 0%| | 4373/1000000 [18:09<68:28:43, 4.04it/s, grad_norm=0.313, loss_final=0.501, loss_mean=0.815, loss_mean_cls=0.0829, proj_loss=-0.397][2026-03-26 12:49:13] Step: 4373, Training Logs: loss_final: 0.492527, loss_mean: 0.840428, proj_loss: -0.424025, loss_mean_cls: 0.076124, grad_norm: 0.311088 +Steps: 0%| | 4374/1000000 [18:10<68:31:04, 4.04it/s, grad_norm=0.311, loss_final=0.493, loss_mean=0.84, loss_mean_cls=0.0761, proj_loss=-0.424][2026-03-26 12:49:13] Step: 4374, Training Logs: loss_final: 0.491666, loss_mean: 0.825665, proj_loss: -0.414626, loss_mean_cls: 0.080627, grad_norm: 0.247002 +Steps: 0%| | 4375/1000000 [18:10<68:31:07, 4.04it/s, grad_norm=0.247, loss_final=0.492, loss_mean=0.826, loss_mean_cls=0.0806, proj_loss=-0.415][2026-03-26 12:49:13] Step: 4375, Training Logs: loss_final: 0.503986, loss_mean: 0.839909, proj_loss: -0.416744, loss_mean_cls: 0.080821, grad_norm: 0.310383 +Steps: 0%| | 4376/1000000 [18:10<68:29:05, 4.04it/s, grad_norm=0.31, loss_final=0.504, loss_mean=0.84, loss_mean_cls=0.0808, proj_loss=-0.417][2026-03-26 12:49:14] Step: 4376, Training Logs: loss_final: 0.520162, loss_mean: 0.849309, proj_loss: -0.408290, loss_mean_cls: 0.079143, grad_norm: 0.423335 +Steps: 0%| | 4377/1000000 [18:10<68:28:20, 4.04it/s, grad_norm=0.423, loss_final=0.52, loss_mean=0.849, loss_mean_cls=0.0791, proj_loss=-0.408][2026-03-26 12:49:14] Step: 4377, Training Logs: loss_final: 0.497959, loss_mean: 0.825144, proj_loss: -0.407077, loss_mean_cls: 0.079892, grad_norm: 0.340325 +Steps: 0%| | 4378/1000000 [18:11<68:31:20, 4.04it/s, grad_norm=0.34, loss_final=0.498, loss_mean=0.825, loss_mean_cls=0.0799, proj_loss=-0.407][2026-03-26 12:49:14] Step: 4378, Training Logs: loss_final: 0.497926, loss_mean: 0.838609, proj_loss: -0.419514, loss_mean_cls: 0.078830, grad_norm: 0.370160 +Steps: 0%| | 4379/1000000 [18:11<68:50:24, 4.02it/s, grad_norm=0.37, loss_final=0.498, loss_mean=0.839, loss_mean_cls=0.0788, proj_loss=-0.42][2026-03-26 12:49:14] Step: 4379, Training Logs: loss_final: 0.505249, loss_mean: 0.838261, proj_loss: -0.412203, loss_mean_cls: 0.079191, grad_norm: 0.347477 +Steps: 0%| | 4380/1000000 [18:11<68:45:05, 4.02it/s, grad_norm=0.347, loss_final=0.505, loss_mean=0.838, loss_mean_cls=0.0792, proj_loss=-0.412][2026-03-26 12:49:15] Step: 4380, Training Logs: loss_final: 0.512182, loss_mean: 0.842586, proj_loss: -0.411220, loss_mean_cls: 0.080817, grad_norm: 0.499251 +Steps: 0%| | 4381/1000000 [18:11<68:41:55, 4.03it/s, grad_norm=0.499, loss_final=0.512, loss_mean=0.843, loss_mean_cls=0.0808, proj_loss=-0.411][2026-03-26 12:49:15] Step: 4381, Training Logs: loss_final: 0.499549, loss_mean: 0.831495, proj_loss: -0.411777, loss_mean_cls: 0.079831, grad_norm: 0.638113 +Steps: 0%| | 4382/1000000 [18:12<68:36:39, 4.03it/s, grad_norm=0.638, loss_final=0.5, loss_mean=0.831, loss_mean_cls=0.0798, proj_loss=-0.412][2026-03-26 12:49:15] Step: 4382, Training Logs: loss_final: 0.498544, loss_mean: 0.836210, proj_loss: -0.416060, loss_mean_cls: 0.078394, grad_norm: 0.528542 +Steps: 0%| | 4383/1000000 [18:12<68:34:28, 4.03it/s, grad_norm=0.529, loss_final=0.499, loss_mean=0.836, loss_mean_cls=0.0784, proj_loss=-0.416][2026-03-26 12:49:15] Step: 4383, Training Logs: loss_final: 0.497102, loss_mean: 0.833637, proj_loss: -0.416909, loss_mean_cls: 0.080373, grad_norm: 0.309419 +Steps: 0%| | 4384/1000000 [18:12<68:33:48, 4.03it/s, grad_norm=0.309, loss_final=0.497, loss_mean=0.834, loss_mean_cls=0.0804, proj_loss=-0.417][2026-03-26 12:49:16] Step: 4384, Training Logs: loss_final: 0.506850, loss_mean: 0.827820, proj_loss: -0.403150, loss_mean_cls: 0.082181, grad_norm: 0.474064 +Steps: 0%| | 4385/1000000 [18:12<68:32:12, 4.04it/s, grad_norm=0.474, loss_final=0.507, loss_mean=0.828, loss_mean_cls=0.0822, proj_loss=-0.403][2026-03-26 12:49:16] Step: 4385, Training Logs: loss_final: 0.501374, loss_mean: 0.826337, proj_loss: -0.407267, loss_mean_cls: 0.082303, grad_norm: 0.325966 +Steps: 0%| | 4386/1000000 [18:13<68:32:30, 4.03it/s, grad_norm=0.326, loss_final=0.501, loss_mean=0.826, loss_mean_cls=0.0823, proj_loss=-0.407][2026-03-26 12:49:16] Step: 4386, Training Logs: loss_final: 0.495520, loss_mean: 0.826860, proj_loss: -0.413599, loss_mean_cls: 0.082258, grad_norm: 0.477533 +Steps: 0%| | 4387/1000000 [18:13<68:34:07, 4.03it/s, grad_norm=0.478, loss_final=0.496, loss_mean=0.827, loss_mean_cls=0.0823, proj_loss=-0.414][2026-03-26 12:49:16] Step: 4387, Training Logs: loss_final: 0.500472, loss_mean: 0.829624, proj_loss: -0.410443, loss_mean_cls: 0.081292, grad_norm: 0.568709 +Steps: 0%| | 4388/1000000 [18:13<68:30:07, 4.04it/s, grad_norm=0.569, loss_final=0.5, loss_mean=0.83, loss_mean_cls=0.0813, proj_loss=-0.41][2026-03-26 12:49:17] Step: 4388, Training Logs: loss_final: 0.494705, loss_mean: 0.830569, proj_loss: -0.415230, loss_mean_cls: 0.079365, grad_norm: 0.288487 +Steps: 0%| | 4389/1000000 [18:13<68:29:29, 4.04it/s, grad_norm=0.288, loss_final=0.495, loss_mean=0.831, loss_mean_cls=0.0794, proj_loss=-0.415][2026-03-26 12:49:17] Step: 4389, Training Logs: loss_final: 0.516988, loss_mean: 0.850357, proj_loss: -0.412400, loss_mean_cls: 0.079032, grad_norm: 0.556896 +Steps: 0%| | 4390/1000000 [18:14<68:26:40, 4.04it/s, grad_norm=0.557, loss_final=0.517, loss_mean=0.85, loss_mean_cls=0.079, proj_loss=-0.412][2026-03-26 12:49:17] Step: 4390, Training Logs: loss_final: 0.496292, loss_mean: 0.837278, proj_loss: -0.419587, loss_mean_cls: 0.078601, grad_norm: 0.574123 +Steps: 0%| | 4391/1000000 [18:14<68:30:20, 4.04it/s, grad_norm=0.574, loss_final=0.496, loss_mean=0.837, loss_mean_cls=0.0786, proj_loss=-0.42][2026-03-26 12:49:17] Step: 4391, Training Logs: loss_final: 0.493554, loss_mean: 0.820597, proj_loss: -0.408164, loss_mean_cls: 0.081121, grad_norm: 0.329620 +Steps: 0%| | 4392/1000000 [18:14<68:31:36, 4.04it/s, grad_norm=0.33, loss_final=0.494, loss_mean=0.821, loss_mean_cls=0.0811, proj_loss=-0.408][2026-03-26 12:49:18] Step: 4392, Training Logs: loss_final: 0.486053, loss_mean: 0.815993, proj_loss: -0.411149, loss_mean_cls: 0.081209, grad_norm: 0.562042 +Steps: 0%| | 4393/1000000 [18:14<68:27:47, 4.04it/s, grad_norm=0.562, loss_final=0.486, loss_mean=0.816, loss_mean_cls=0.0812, proj_loss=-0.411][2026-03-26 12:49:18] Step: 4393, Training Logs: loss_final: 0.509392, loss_mean: 0.844348, proj_loss: -0.412943, loss_mean_cls: 0.077987, grad_norm: 0.610930 +Steps: 0%| | 4394/1000000 [18:15<68:23:49, 4.04it/s, grad_norm=0.611, loss_final=0.509, loss_mean=0.844, loss_mean_cls=0.078, proj_loss=-0.413][2026-03-26 12:49:18] Step: 4394, Training Logs: loss_final: 0.502541, loss_mean: 0.831371, proj_loss: -0.409692, loss_mean_cls: 0.080862, grad_norm: 0.309086 +Steps: 0%| | 4395/1000000 [18:15<68:25:51, 4.04it/s, grad_norm=0.309, loss_final=0.503, loss_mean=0.831, loss_mean_cls=0.0809, proj_loss=-0.41][2026-03-26 12:49:18] Step: 4395, Training Logs: loss_final: 0.508630, loss_mean: 0.842136, proj_loss: -0.414705, loss_mean_cls: 0.081199, grad_norm: 0.431704 +Steps: 0%| | 4396/1000000 [18:15<68:23:47, 4.04it/s, grad_norm=0.432, loss_final=0.509, loss_mean=0.842, loss_mean_cls=0.0812, proj_loss=-0.415][2026-03-26 12:49:19] Step: 4396, Training Logs: loss_final: 0.490308, loss_mean: 0.822634, proj_loss: -0.411801, loss_mean_cls: 0.079475, grad_norm: 0.321532 +Steps: 0%| | 4397/1000000 [18:15<68:25:22, 4.04it/s, grad_norm=0.322, loss_final=0.49, loss_mean=0.823, loss_mean_cls=0.0795, proj_loss=-0.412][2026-03-26 12:49:19] Step: 4397, Training Logs: loss_final: 0.504495, loss_mean: 0.837549, proj_loss: -0.411670, loss_mean_cls: 0.078616, grad_norm: 0.412506 +Steps: 0%| | 4398/1000000 [18:16<68:22:37, 4.04it/s, grad_norm=0.413, loss_final=0.504, loss_mean=0.838, loss_mean_cls=0.0786, proj_loss=-0.412][2026-03-26 12:49:19] Step: 4398, Training Logs: loss_final: 0.501111, loss_mean: 0.831467, proj_loss: -0.410114, loss_mean_cls: 0.079758, grad_norm: 0.338137 +Steps: 0%| | 4399/1000000 [18:16<68:24:29, 4.04it/s, grad_norm=0.338, loss_final=0.501, loss_mean=0.831, loss_mean_cls=0.0798, proj_loss=-0.41][2026-03-26 12:49:19] Step: 4399, Training Logs: loss_final: 0.504515, loss_mean: 0.832572, proj_loss: -0.408412, loss_mean_cls: 0.080354, grad_norm: 0.381504 +Steps: 0%| | 4400/1000000 [18:16<68:25:37, 4.04it/s, grad_norm=0.382, loss_final=0.505, loss_mean=0.833, loss_mean_cls=0.0804, proj_loss=-0.408][2026-03-26 12:49:20] Step: 4400, Training Logs: loss_final: 0.503745, loss_mean: 0.835888, proj_loss: -0.412957, loss_mean_cls: 0.080813, grad_norm: 0.310797 +Steps: 0%| | 4401/1000000 [18:16<68:24:42, 4.04it/s, grad_norm=0.311, loss_final=0.504, loss_mean=0.836, loss_mean_cls=0.0808, proj_loss=-0.413][2026-03-26 12:49:20] Step: 4401, Training Logs: loss_final: 0.506332, loss_mean: 0.833478, proj_loss: -0.407751, loss_mean_cls: 0.080605, grad_norm: 0.285907 +Steps: 0%| | 4402/1000000 [18:17<68:27:30, 4.04it/s, grad_norm=0.286, loss_final=0.506, loss_mean=0.833, loss_mean_cls=0.0806, proj_loss=-0.408][2026-03-26 12:49:20] Step: 4402, Training Logs: loss_final: 0.479595, loss_mean: 0.808473, proj_loss: -0.411154, loss_mean_cls: 0.082276, grad_norm: 0.295777 +Steps: 0%| | 4403/1000000 [18:17<68:24:54, 4.04it/s, grad_norm=0.296, loss_final=0.48, loss_mean=0.808, loss_mean_cls=0.0823, proj_loss=-0.411][2026-03-26 12:49:20] Step: 4403, Training Logs: loss_final: 0.488678, loss_mean: 0.823034, proj_loss: -0.414494, loss_mean_cls: 0.080138, grad_norm: 0.287465 +Steps: 0%| | 4404/1000000 [18:17<68:27:32, 4.04it/s, grad_norm=0.287, loss_final=0.489, loss_mean=0.823, loss_mean_cls=0.0801, proj_loss=-0.414][2026-03-26 12:49:21] Step: 4404, Training Logs: loss_final: 0.486232, loss_mean: 0.815682, proj_loss: -0.410133, loss_mean_cls: 0.080683, grad_norm: 0.312804 +Steps: 0%| | 4405/1000000 [18:17<68:26:30, 4.04it/s, grad_norm=0.313, loss_final=0.486, loss_mean=0.816, loss_mean_cls=0.0807, proj_loss=-0.41][2026-03-26 12:49:21] Step: 4405, Training Logs: loss_final: 0.494069, loss_mean: 0.817608, proj_loss: -0.405993, loss_mean_cls: 0.082454, grad_norm: 0.296175 +Steps: 0%| | 4406/1000000 [18:18<68:27:39, 4.04it/s, grad_norm=0.296, loss_final=0.494, loss_mean=0.818, loss_mean_cls=0.0825, proj_loss=-0.406][2026-03-26 12:49:21] Step: 4406, Training Logs: loss_final: 0.497983, loss_mean: 0.838546, proj_loss: -0.417916, loss_mean_cls: 0.077353, grad_norm: 0.304136 +Steps: 0%| | 4407/1000000 [18:18<68:24:43, 4.04it/s, grad_norm=0.304, loss_final=0.498, loss_mean=0.839, loss_mean_cls=0.0774, proj_loss=-0.418][2026-03-26 12:49:21] Step: 4407, Training Logs: loss_final: 0.496712, loss_mean: 0.826577, proj_loss: -0.410715, loss_mean_cls: 0.080850, grad_norm: 0.286284 +Steps: 0%| | 4408/1000000 [18:18<68:25:46, 4.04it/s, grad_norm=0.286, loss_final=0.497, loss_mean=0.827, loss_mean_cls=0.0809, proj_loss=-0.411][2026-03-26 12:49:22] Step: 4408, Training Logs: loss_final: 0.504388, loss_mean: 0.838304, proj_loss: -0.414276, loss_mean_cls: 0.080360, grad_norm: 0.362002 +Steps: 0%| | 4409/1000000 [18:18<68:23:28, 4.04it/s, grad_norm=0.362, loss_final=0.504, loss_mean=0.838, loss_mean_cls=0.0804, proj_loss=-0.414][2026-03-26 12:49:22] Step: 4409, Training Logs: loss_final: 0.492885, loss_mean: 0.815872, proj_loss: -0.406350, loss_mean_cls: 0.083363, grad_norm: 0.357117 +Steps: 0%| | 4410/1000000 [18:19<68:28:25, 4.04it/s, grad_norm=0.357, loss_final=0.493, loss_mean=0.816, loss_mean_cls=0.0834, proj_loss=-0.406][2026-03-26 12:49:22] Step: 4410, Training Logs: loss_final: 0.497774, loss_mean: 0.833406, proj_loss: -0.415264, loss_mean_cls: 0.079632, grad_norm: 0.198820 +Steps: 0%| | 4411/1000000 [18:19<68:23:49, 4.04it/s, grad_norm=0.199, loss_final=0.498, loss_mean=0.833, loss_mean_cls=0.0796, proj_loss=-0.415][2026-03-26 12:49:22] Step: 4411, Training Logs: loss_final: 0.509004, loss_mean: 0.828828, proj_loss: -0.402170, loss_mean_cls: 0.082346, grad_norm: 0.381319 +Steps: 0%| | 4412/1000000 [18:19<68:23:09, 4.04it/s, grad_norm=0.381, loss_final=0.509, loss_mean=0.829, loss_mean_cls=0.0823, proj_loss=-0.402][2026-03-26 12:49:23] Step: 4412, Training Logs: loss_final: 0.511989, loss_mean: 0.842594, proj_loss: -0.411174, loss_mean_cls: 0.080569, grad_norm: 0.316818 +Steps: 0%| | 4413/1000000 [18:19<68:22:41, 4.04it/s, grad_norm=0.317, loss_final=0.512, loss_mean=0.843, loss_mean_cls=0.0806, proj_loss=-0.411][2026-03-26 12:49:23] Step: 4413, Training Logs: loss_final: 0.495973, loss_mean: 0.828016, proj_loss: -0.412350, loss_mean_cls: 0.080307, grad_norm: 0.278376 +Steps: 0%| | 4414/1000000 [18:20<68:22:14, 4.04it/s, grad_norm=0.278, loss_final=0.496, loss_mean=0.828, loss_mean_cls=0.0803, proj_loss=-0.412][2026-03-26 12:49:23] Step: 4414, Training Logs: loss_final: 0.511517, loss_mean: 0.830312, proj_loss: -0.399745, loss_mean_cls: 0.080949, grad_norm: 0.519154 +Steps: 0%| | 4415/1000000 [18:20<68:23:25, 4.04it/s, grad_norm=0.519, loss_final=0.512, loss_mean=0.83, loss_mean_cls=0.0809, proj_loss=-0.4][2026-03-26 12:49:23] Step: 4415, Training Logs: loss_final: 0.499722, loss_mean: 0.826409, proj_loss: -0.409072, loss_mean_cls: 0.082384, grad_norm: 0.282532 +Steps: 0%| | 4416/1000000 [18:20<68:24:19, 4.04it/s, grad_norm=0.283, loss_final=0.5, loss_mean=0.826, loss_mean_cls=0.0824, proj_loss=-0.409][2026-03-26 12:49:24] Step: 4416, Training Logs: loss_final: 0.508597, loss_mean: 0.843650, proj_loss: -0.414625, loss_mean_cls: 0.079573, grad_norm: 0.469315 +Steps: 0%| | 4417/1000000 [18:20<68:26:37, 4.04it/s, grad_norm=0.469, loss_final=0.509, loss_mean=0.844, loss_mean_cls=0.0796, proj_loss=-0.415][2026-03-26 12:49:24] Step: 4417, Training Logs: loss_final: 0.519346, loss_mean: 0.860776, proj_loss: -0.418877, loss_mean_cls: 0.077446, grad_norm: 0.368374 +Steps: 0%| | 4418/1000000 [18:21<68:25:52, 4.04it/s, grad_norm=0.368, loss_final=0.519, loss_mean=0.861, loss_mean_cls=0.0774, proj_loss=-0.419][2026-03-26 12:49:24] Step: 4418, Training Logs: loss_final: 0.506557, loss_mean: 0.834061, proj_loss: -0.409741, loss_mean_cls: 0.082237, grad_norm: 0.310146 +Steps: 0%| | 4419/1000000 [18:21<68:25:30, 4.04it/s, grad_norm=0.31, loss_final=0.507, loss_mean=0.834, loss_mean_cls=0.0822, proj_loss=-0.41][2026-03-26 12:49:24] Step: 4419, Training Logs: loss_final: 0.506331, loss_mean: 0.831632, proj_loss: -0.407158, loss_mean_cls: 0.081856, grad_norm: 0.565731 +Steps: 0%| | 4420/1000000 [18:21<68:26:49, 4.04it/s, grad_norm=0.566, loss_final=0.506, loss_mean=0.832, loss_mean_cls=0.0819, proj_loss=-0.407][2026-03-26 12:49:25] Step: 4420, Training Logs: loss_final: 0.501102, loss_mean: 0.835315, proj_loss: -0.414804, loss_mean_cls: 0.080590, grad_norm: 0.393265 +Steps: 0%| | 4421/1000000 [18:21<68:24:55, 4.04it/s, grad_norm=0.393, loss_final=0.501, loss_mean=0.835, loss_mean_cls=0.0806, proj_loss=-0.415][2026-03-26 12:49:25] Step: 4421, Training Logs: loss_final: 0.494792, loss_mean: 0.821969, proj_loss: -0.409463, loss_mean_cls: 0.082286, grad_norm: 0.329853 +Steps: 0%| | 4422/1000000 [18:22<68:24:43, 4.04it/s, grad_norm=0.33, loss_final=0.495, loss_mean=0.822, loss_mean_cls=0.0823, proj_loss=-0.409][2026-03-26 12:49:25] Step: 4422, Training Logs: loss_final: 0.482034, loss_mean: 0.810618, proj_loss: -0.408711, loss_mean_cls: 0.080128, grad_norm: 0.315719 +Steps: 0%| | 4423/1000000 [18:22<68:23:10, 4.04it/s, grad_norm=0.316, loss_final=0.482, loss_mean=0.811, loss_mean_cls=0.0801, proj_loss=-0.409][2026-03-26 12:49:25] Step: 4423, Training Logs: loss_final: 0.485535, loss_mean: 0.806940, proj_loss: -0.405330, loss_mean_cls: 0.083925, grad_norm: 0.350415 +Steps: 0%| | 4424/1000000 [18:22<68:23:26, 4.04it/s, grad_norm=0.35, loss_final=0.486, loss_mean=0.807, loss_mean_cls=0.0839, proj_loss=-0.405][2026-03-26 12:49:25] Step: 4424, Training Logs: loss_final: 0.494971, loss_mean: 0.831546, proj_loss: -0.414297, loss_mean_cls: 0.077722, grad_norm: 0.327659 +Steps: 0%| | 4425/1000000 [18:22<68:24:55, 4.04it/s, grad_norm=0.328, loss_final=0.495, loss_mean=0.832, loss_mean_cls=0.0777, proj_loss=-0.414][2026-03-26 12:49:26] Step: 4425, Training Logs: loss_final: 0.503604, loss_mean: 0.834459, proj_loss: -0.409573, loss_mean_cls: 0.078718, grad_norm: 0.292810 +Steps: 0%| | 4426/1000000 [18:23<68:22:57, 4.04it/s, grad_norm=0.293, loss_final=0.504, loss_mean=0.834, loss_mean_cls=0.0787, proj_loss=-0.41][2026-03-26 12:49:26] Step: 4426, Training Logs: loss_final: 0.507384, loss_mean: 0.832843, proj_loss: -0.406049, loss_mean_cls: 0.080590, grad_norm: 0.381629 +Steps: 0%| | 4427/1000000 [18:23<68:22:36, 4.04it/s, grad_norm=0.382, loss_final=0.507, loss_mean=0.833, loss_mean_cls=0.0806, proj_loss=-0.406][2026-03-26 12:49:26] Step: 4427, Training Logs: loss_final: 0.485148, loss_mean: 0.826494, proj_loss: -0.418822, loss_mean_cls: 0.077477, grad_norm: 0.331400 +Steps: 0%| | 4428/1000000 [18:23<68:24:34, 4.04it/s, grad_norm=0.331, loss_final=0.485, loss_mean=0.826, loss_mean_cls=0.0775, proj_loss=-0.419][2026-03-26 12:49:26] Step: 4428, Training Logs: loss_final: 0.466090, loss_mean: 0.794309, proj_loss: -0.410903, loss_mean_cls: 0.082684, grad_norm: 0.302771 +Steps: 0%| | 4429/1000000 [18:23<68:24:14, 4.04it/s, grad_norm=0.303, loss_final=0.466, loss_mean=0.794, loss_mean_cls=0.0827, proj_loss=-0.411][2026-03-26 12:49:27] Step: 4429, Training Logs: loss_final: 0.499909, loss_mean: 0.831301, proj_loss: -0.412562, loss_mean_cls: 0.081170, grad_norm: 0.282353 +Steps: 0%| | 4430/1000000 [18:24<68:23:51, 4.04it/s, grad_norm=0.282, loss_final=0.5, loss_mean=0.831, loss_mean_cls=0.0812, proj_loss=-0.413][2026-03-26 12:49:27] Step: 4430, Training Logs: loss_final: 0.522982, loss_mean: 0.851921, proj_loss: -0.409859, loss_mean_cls: 0.080921, grad_norm: 0.324279 +Steps: 0%| | 4431/1000000 [18:24<68:25:11, 4.04it/s, grad_norm=0.324, loss_final=0.523, loss_mean=0.852, loss_mean_cls=0.0809, proj_loss=-0.41][2026-03-26 12:49:27] Step: 4431, Training Logs: loss_final: 0.495480, loss_mean: 0.825276, proj_loss: -0.409401, loss_mean_cls: 0.079606, grad_norm: 0.379504 +Steps: 0%| | 4432/1000000 [18:24<68:22:04, 4.04it/s, grad_norm=0.38, loss_final=0.495, loss_mean=0.825, loss_mean_cls=0.0796, proj_loss=-0.409][2026-03-26 12:49:27] Step: 4432, Training Logs: loss_final: 0.506963, loss_mean: 0.831460, proj_loss: -0.405159, loss_mean_cls: 0.080662, grad_norm: 0.311568 +Steps: 0%| | 4433/1000000 [18:24<68:22:01, 4.05it/s, grad_norm=0.312, loss_final=0.507, loss_mean=0.831, loss_mean_cls=0.0807, proj_loss=-0.405][2026-03-26 12:49:28] Step: 4433, Training Logs: loss_final: 0.500484, loss_mean: 0.826676, proj_loss: -0.407556, loss_mean_cls: 0.081364, grad_norm: 0.341094 +Steps: 0%| | 4434/1000000 [18:25<68:17:57, 4.05it/s, grad_norm=0.341, loss_final=0.5, loss_mean=0.827, loss_mean_cls=0.0814, proj_loss=-0.408][2026-03-26 12:49:28] Step: 4434, Training Logs: loss_final: 0.497184, loss_mean: 0.831275, proj_loss: -0.413452, loss_mean_cls: 0.079362, grad_norm: 0.313703 +Steps: 0%| | 4435/1000000 [18:25<68:19:56, 4.05it/s, grad_norm=0.314, loss_final=0.497, loss_mean=0.831, loss_mean_cls=0.0794, proj_loss=-0.413][2026-03-26 12:49:28] Step: 4435, Training Logs: loss_final: 0.510681, loss_mean: 0.848528, proj_loss: -0.416321, loss_mean_cls: 0.078474, grad_norm: 0.365888 +Steps: 0%| | 4436/1000000 [18:25<68:21:37, 4.05it/s, grad_norm=0.366, loss_final=0.511, loss_mean=0.849, loss_mean_cls=0.0785, proj_loss=-0.416][2026-03-26 12:49:28] Step: 4436, Training Logs: loss_final: 0.484938, loss_mean: 0.817115, proj_loss: -0.413933, loss_mean_cls: 0.081756, grad_norm: 0.379173 +Steps: 0%| | 4437/1000000 [18:25<68:21:17, 4.05it/s, grad_norm=0.379, loss_final=0.485, loss_mean=0.817, loss_mean_cls=0.0818, proj_loss=-0.414][2026-03-26 12:49:29] Step: 4437, Training Logs: loss_final: 0.487610, loss_mean: 0.812746, proj_loss: -0.407769, loss_mean_cls: 0.082633, grad_norm: 0.355483 +Steps: 0%| | 4438/1000000 [18:26<68:21:44, 4.05it/s, grad_norm=0.355, loss_final=0.488, loss_mean=0.813, loss_mean_cls=0.0826, proj_loss=-0.408][2026-03-26 12:49:29] Step: 4438, Training Logs: loss_final: 0.508337, loss_mean: 0.838385, proj_loss: -0.409985, loss_mean_cls: 0.079937, grad_norm: 0.238391 +Steps: 0%| | 4439/1000000 [18:26<68:25:08, 4.04it/s, grad_norm=0.238, loss_final=0.508, loss_mean=0.838, loss_mean_cls=0.0799, proj_loss=-0.41][2026-03-26 12:49:29] Step: 4439, Training Logs: loss_final: 0.501029, loss_mean: 0.831595, proj_loss: -0.409855, loss_mean_cls: 0.079289, grad_norm: 0.350931 +Steps: 0%| | 4440/1000000 [18:26<68:24:20, 4.04it/s, grad_norm=0.351, loss_final=0.501, loss_mean=0.832, loss_mean_cls=0.0793, proj_loss=-0.41][2026-03-26 12:49:29] Step: 4440, Training Logs: loss_final: 0.509280, loss_mean: 0.836121, proj_loss: -0.407410, loss_mean_cls: 0.080568, grad_norm: 0.334547 +Steps: 0%| | 4441/1000000 [18:26<68:24:16, 4.04it/s, grad_norm=0.335, loss_final=0.509, loss_mean=0.836, loss_mean_cls=0.0806, proj_loss=-0.407][2026-03-26 12:49:30] Step: 4441, Training Logs: loss_final: 0.505222, loss_mean: 0.825206, proj_loss: -0.402489, loss_mean_cls: 0.082505, grad_norm: 0.435438 +Steps: 0%| | 4442/1000000 [18:27<68:25:36, 4.04it/s, grad_norm=0.435, loss_final=0.505, loss_mean=0.825, loss_mean_cls=0.0825, proj_loss=-0.402][2026-03-26 12:49:30] Step: 4442, Training Logs: loss_final: 0.493810, loss_mean: 0.819867, proj_loss: -0.408682, loss_mean_cls: 0.082625, grad_norm: 0.318185 +Steps: 0%| | 4443/1000000 [18:27<68:30:00, 4.04it/s, grad_norm=0.318, loss_final=0.494, loss_mean=0.82, loss_mean_cls=0.0826, proj_loss=-0.409][2026-03-26 12:49:30] Step: 4443, Training Logs: loss_final: 0.484979, loss_mean: 0.821746, proj_loss: -0.415790, loss_mean_cls: 0.079023, grad_norm: 0.338696 +Steps: 0%| | 4444/1000000 [18:27<68:28:59, 4.04it/s, grad_norm=0.339, loss_final=0.485, loss_mean=0.822, loss_mean_cls=0.079, proj_loss=-0.416][2026-03-26 12:49:30] Step: 4444, Training Logs: loss_final: 0.488630, loss_mean: 0.824574, proj_loss: -0.413559, loss_mean_cls: 0.077615, grad_norm: 0.280750 +Steps: 0%| | 4445/1000000 [18:27<68:26:28, 4.04it/s, grad_norm=0.281, loss_final=0.489, loss_mean=0.825, loss_mean_cls=0.0776, proj_loss=-0.414][2026-03-26 12:49:31] Step: 4445, Training Logs: loss_final: 0.489353, loss_mean: 0.821536, proj_loss: -0.415196, loss_mean_cls: 0.083014, grad_norm: 0.302765 +Steps: 0%| | 4446/1000000 [18:28<68:26:35, 4.04it/s, grad_norm=0.303, loss_final=0.489, loss_mean=0.822, loss_mean_cls=0.083, proj_loss=-0.415][2026-03-26 12:49:31] Step: 4446, Training Logs: loss_final: 0.492846, loss_mean: 0.827846, proj_loss: -0.413800, loss_mean_cls: 0.078800, grad_norm: 0.327076 +Steps: 0%| | 4447/1000000 [18:28<68:24:40, 4.04it/s, grad_norm=0.327, loss_final=0.493, loss_mean=0.828, loss_mean_cls=0.0788, proj_loss=-0.414][2026-03-26 12:49:31] Step: 4447, Training Logs: loss_final: 0.498152, loss_mean: 0.833643, proj_loss: -0.414587, loss_mean_cls: 0.079097, grad_norm: 0.416379 +Steps: 0%| | 4448/1000000 [18:28<68:26:39, 4.04it/s, grad_norm=0.416, loss_final=0.498, loss_mean=0.834, loss_mean_cls=0.0791, proj_loss=-0.415][2026-03-26 12:49:31] Step: 4448, Training Logs: loss_final: 0.490706, loss_mean: 0.820192, proj_loss: -0.411445, loss_mean_cls: 0.081959, grad_norm: 0.441539 +Steps: 0%| | 4449/1000000 [18:28<68:24:37, 4.04it/s, grad_norm=0.442, loss_final=0.491, loss_mean=0.82, loss_mean_cls=0.082, proj_loss=-0.411][2026-03-26 12:49:32] Step: 4449, Training Logs: loss_final: 0.507171, loss_mean: 0.837834, proj_loss: -0.411498, loss_mean_cls: 0.080835, grad_norm: 0.421390 +Steps: 0%| | 4450/1000000 [18:29<68:29:42, 4.04it/s, grad_norm=0.421, loss_final=0.507, loss_mean=0.838, loss_mean_cls=0.0808, proj_loss=-0.411][2026-03-26 12:49:32] Step: 4450, Training Logs: loss_final: 0.505330, loss_mean: 0.833331, proj_loss: -0.408885, loss_mean_cls: 0.080884, grad_norm: 0.369606 +Steps: 0%| | 4451/1000000 [18:29<68:26:51, 4.04it/s, grad_norm=0.37, loss_final=0.505, loss_mean=0.833, loss_mean_cls=0.0809, proj_loss=-0.409][2026-03-26 12:49:32] Step: 4451, Training Logs: loss_final: 0.495496, loss_mean: 0.820728, proj_loss: -0.406513, loss_mean_cls: 0.081281, grad_norm: 0.392413 +Steps: 0%| | 4452/1000000 [18:29<68:27:13, 4.04it/s, grad_norm=0.392, loss_final=0.495, loss_mean=0.821, loss_mean_cls=0.0813, proj_loss=-0.407][2026-03-26 12:49:32] Step: 4452, Training Logs: loss_final: 0.503358, loss_mean: 0.824381, proj_loss: -0.402314, loss_mean_cls: 0.081292, grad_norm: 0.473105 +Steps: 0%| | 4453/1000000 [18:29<69:05:00, 4.00it/s, grad_norm=0.473, loss_final=0.503, loss_mean=0.824, loss_mean_cls=0.0813, proj_loss=-0.402][2026-03-26 12:49:33] Step: 4453, Training Logs: loss_final: 0.501925, loss_mean: 0.837515, proj_loss: -0.413848, loss_mean_cls: 0.078258, grad_norm: 0.344703 +Steps: 0%| | 4454/1000000 [18:30<68:56:31, 4.01it/s, grad_norm=0.345, loss_final=0.502, loss_mean=0.838, loss_mean_cls=0.0783, proj_loss=-0.414][2026-03-26 12:49:33] Step: 4454, Training Logs: loss_final: 0.500398, loss_mean: 0.823348, proj_loss: -0.405576, loss_mean_cls: 0.082627, grad_norm: 0.356220 +Steps: 0%| | 4455/1000000 [18:30<68:45:48, 4.02it/s, grad_norm=0.356, loss_final=0.5, loss_mean=0.823, loss_mean_cls=0.0826, proj_loss=-0.406][2026-03-26 12:49:33] Step: 4455, Training Logs: loss_final: 0.474000, loss_mean: 0.795521, proj_loss: -0.405519, loss_mean_cls: 0.083998, grad_norm: 0.315622 +Steps: 0%| | 4456/1000000 [18:30<68:40:10, 4.03it/s, grad_norm=0.316, loss_final=0.474, loss_mean=0.796, loss_mean_cls=0.084, proj_loss=-0.406][2026-03-26 12:49:33] Step: 4456, Training Logs: loss_final: 0.507729, loss_mean: 0.828744, proj_loss: -0.403700, loss_mean_cls: 0.082685, grad_norm: 0.327314 +Steps: 0%| | 4457/1000000 [18:30<68:37:33, 4.03it/s, grad_norm=0.327, loss_final=0.508, loss_mean=0.829, loss_mean_cls=0.0827, proj_loss=-0.404][2026-03-26 12:49:34] Step: 4457, Training Logs: loss_final: 0.497762, loss_mean: 0.821466, proj_loss: -0.404892, loss_mean_cls: 0.081188, grad_norm: 0.397248 +Steps: 0%| | 4458/1000000 [18:31<68:34:42, 4.03it/s, grad_norm=0.397, loss_final=0.498, loss_mean=0.821, loss_mean_cls=0.0812, proj_loss=-0.405][2026-03-26 12:49:34] Step: 4458, Training Logs: loss_final: 0.513414, loss_mean: 0.852706, proj_loss: -0.417541, loss_mean_cls: 0.078249, grad_norm: 0.293309 +Steps: 0%| | 4459/1000000 [18:31<68:29:41, 4.04it/s, grad_norm=0.293, loss_final=0.513, loss_mean=0.853, loss_mean_cls=0.0782, proj_loss=-0.418][2026-03-26 12:49:34] Step: 4459, Training Logs: loss_final: 0.497533, loss_mean: 0.830577, proj_loss: -0.412490, loss_mean_cls: 0.079446, grad_norm: 0.286225 +Steps: 0%| | 4460/1000000 [18:31<68:28:13, 4.04it/s, grad_norm=0.286, loss_final=0.498, loss_mean=0.831, loss_mean_cls=0.0794, proj_loss=-0.412][2026-03-26 12:49:34] Step: 4460, Training Logs: loss_final: 0.503622, loss_mean: 0.831043, proj_loss: -0.408043, loss_mean_cls: 0.080621, grad_norm: 0.308311 +Steps: 0%| | 4461/1000000 [18:31<68:25:56, 4.04it/s, grad_norm=0.308, loss_final=0.504, loss_mean=0.831, loss_mean_cls=0.0806, proj_loss=-0.408][2026-03-26 12:49:35] Step: 4461, Training Logs: loss_final: 0.482232, loss_mean: 0.802951, proj_loss: -0.403439, loss_mean_cls: 0.082720, grad_norm: 0.245548 +Steps: 0%| | 4462/1000000 [18:32<68:26:46, 4.04it/s, grad_norm=0.246, loss_final=0.482, loss_mean=0.803, loss_mean_cls=0.0827, proj_loss=-0.403][2026-03-26 12:49:35] Step: 4462, Training Logs: loss_final: 0.493306, loss_mean: 0.839558, proj_loss: -0.422487, loss_mean_cls: 0.076236, grad_norm: 0.395718 +Steps: 0%| | 4463/1000000 [18:32<68:27:19, 4.04it/s, grad_norm=0.396, loss_final=0.493, loss_mean=0.84, loss_mean_cls=0.0762, proj_loss=-0.422][2026-03-26 12:49:35] Step: 4463, Training Logs: loss_final: 0.488590, loss_mean: 0.822340, proj_loss: -0.412654, loss_mean_cls: 0.078904, grad_norm: 0.378705 +Steps: 0%| | 4464/1000000 [18:32<68:28:00, 4.04it/s, grad_norm=0.379, loss_final=0.489, loss_mean=0.822, loss_mean_cls=0.0789, proj_loss=-0.413][2026-03-26 12:49:35] Step: 4464, Training Logs: loss_final: 0.501937, loss_mean: 0.841058, proj_loss: -0.418009, loss_mean_cls: 0.078888, grad_norm: 0.244580 +Steps: 0%| | 4465/1000000 [18:32<68:25:08, 4.04it/s, grad_norm=0.245, loss_final=0.502, loss_mean=0.841, loss_mean_cls=0.0789, proj_loss=-0.418][2026-03-26 12:49:36] Step: 4465, Training Logs: loss_final: 0.504769, loss_mean: 0.848774, proj_loss: -0.421061, loss_mean_cls: 0.077056, grad_norm: 0.434996 +Steps: 0%| | 4466/1000000 [18:33<68:25:05, 4.04it/s, grad_norm=0.435, loss_final=0.505, loss_mean=0.849, loss_mean_cls=0.0771, proj_loss=-0.421][2026-03-26 12:49:36] Step: 4466, Training Logs: loss_final: 0.498825, loss_mean: 0.823389, proj_loss: -0.406637, loss_mean_cls: 0.082073, grad_norm: 0.454995 +Steps: 0%| | 4467/1000000 [18:33<68:27:09, 4.04it/s, grad_norm=0.455, loss_final=0.499, loss_mean=0.823, loss_mean_cls=0.0821, proj_loss=-0.407][2026-03-26 12:49:36] Step: 4467, Training Logs: loss_final: 0.495009, loss_mean: 0.836039, proj_loss: -0.418573, loss_mean_cls: 0.077542, grad_norm: 0.214972 +Steps: 0%| | 4468/1000000 [18:33<68:24:56, 4.04it/s, grad_norm=0.215, loss_final=0.495, loss_mean=0.836, loss_mean_cls=0.0775, proj_loss=-0.419][2026-03-26 12:49:36] Step: 4468, Training Logs: loss_final: 0.501677, loss_mean: 0.837675, proj_loss: -0.414838, loss_mean_cls: 0.078840, grad_norm: 0.510016 +Steps: 0%| | 4469/1000000 [18:33<68:25:58, 4.04it/s, grad_norm=0.51, loss_final=0.502, loss_mean=0.838, loss_mean_cls=0.0788, proj_loss=-0.415][2026-03-26 12:49:37] Step: 4469, Training Logs: loss_final: 0.501595, loss_mean: 0.829692, proj_loss: -0.409107, loss_mean_cls: 0.081010, grad_norm: 0.643850 +Steps: 0%| | 4470/1000000 [18:33<68:24:03, 4.04it/s, grad_norm=0.644, loss_final=0.502, loss_mean=0.83, loss_mean_cls=0.081, proj_loss=-0.409][2026-03-26 12:49:37] Step: 4470, Training Logs: loss_final: 0.507481, loss_mean: 0.842645, proj_loss: -0.414066, loss_mean_cls: 0.078902, grad_norm: 0.262652 +Steps: 0%| | 4471/1000000 [18:34<68:29:34, 4.04it/s, grad_norm=0.263, loss_final=0.507, loss_mean=0.843, loss_mean_cls=0.0789, proj_loss=-0.414][2026-03-26 12:49:37] Step: 4471, Training Logs: loss_final: 0.503549, loss_mean: 0.829651, proj_loss: -0.408299, loss_mean_cls: 0.082197, grad_norm: 0.322439 +Steps: 0%| | 4472/1000000 [18:34<68:25:02, 4.04it/s, grad_norm=0.322, loss_final=0.504, loss_mean=0.83, loss_mean_cls=0.0822, proj_loss=-0.408][2026-03-26 12:49:37] Step: 4472, Training Logs: loss_final: 0.503086, loss_mean: 0.834784, proj_loss: -0.410925, loss_mean_cls: 0.079228, grad_norm: 0.342064 +Steps: 0%| | 4473/1000000 [18:34<69:04:04, 4.00it/s, grad_norm=0.342, loss_final=0.503, loss_mean=0.835, loss_mean_cls=0.0792, proj_loss=-0.411][2026-03-26 12:49:38] Step: 4473, Training Logs: loss_final: 0.501662, loss_mean: 0.835811, proj_loss: -0.412642, loss_mean_cls: 0.078494, grad_norm: 0.572822 +Steps: 0%| | 4474/1000000 [18:34<69:04:15, 4.00it/s, grad_norm=0.573, loss_final=0.502, loss_mean=0.836, loss_mean_cls=0.0785, proj_loss=-0.413][2026-03-26 12:49:38] Step: 4474, Training Logs: loss_final: 0.506974, loss_mean: 0.836360, proj_loss: -0.408806, loss_mean_cls: 0.079420, grad_norm: 0.401962 +Steps: 0%| | 4475/1000000 [18:35<68:52:00, 4.02it/s, grad_norm=0.402, loss_final=0.507, loss_mean=0.836, loss_mean_cls=0.0794, proj_loss=-0.409][2026-03-26 12:49:38] Step: 4475, Training Logs: loss_final: 0.490669, loss_mean: 0.823188, proj_loss: -0.410742, loss_mean_cls: 0.078223, grad_norm: 0.334847 +Steps: 0%| | 4476/1000000 [18:35<68:43:52, 4.02it/s, grad_norm=0.335, loss_final=0.491, loss_mean=0.823, loss_mean_cls=0.0782, proj_loss=-0.411][2026-03-26 12:49:38] Step: 4476, Training Logs: loss_final: 0.509396, loss_mean: 0.825137, proj_loss: -0.399142, loss_mean_cls: 0.083401, grad_norm: 0.322485 +Steps: 0%| | 4477/1000000 [18:35<68:35:53, 4.03it/s, grad_norm=0.322, loss_final=0.509, loss_mean=0.825, loss_mean_cls=0.0834, proj_loss=-0.399][2026-03-26 12:49:39] Step: 4477, Training Logs: loss_final: 0.500210, loss_mean: 0.829196, proj_loss: -0.411063, loss_mean_cls: 0.082077, grad_norm: 0.542871 +Steps: 0%| | 4478/1000000 [18:35<68:31:23, 4.04it/s, grad_norm=0.543, loss_final=0.5, loss_mean=0.829, loss_mean_cls=0.0821, proj_loss=-0.411][2026-03-26 12:49:39] Step: 4478, Training Logs: loss_final: 0.506208, loss_mean: 0.851500, proj_loss: -0.422821, loss_mean_cls: 0.077529, grad_norm: 0.885045 +Steps: 0%| | 4479/1000000 [18:36<68:28:21, 4.04it/s, grad_norm=0.885, loss_final=0.506, loss_mean=0.851, loss_mean_cls=0.0775, proj_loss=-0.423][2026-03-26 12:49:39] Step: 4479, Training Logs: loss_final: 0.509159, loss_mean: 0.835907, proj_loss: -0.408987, loss_mean_cls: 0.082239, grad_norm: 0.392948 +Steps: 0%| | 4480/1000000 [18:36<68:29:34, 4.04it/s, grad_norm=0.393, loss_final=0.509, loss_mean=0.836, loss_mean_cls=0.0822, proj_loss=-0.409][2026-03-26 12:49:39] Step: 4480, Training Logs: loss_final: 0.494793, loss_mean: 0.822517, proj_loss: -0.408302, loss_mean_cls: 0.080577, grad_norm: 0.851098 +Steps: 0%| | 4481/1000000 [18:36<68:28:28, 4.04it/s, grad_norm=0.851, loss_final=0.495, loss_mean=0.823, loss_mean_cls=0.0806, proj_loss=-0.408][2026-03-26 12:49:40] Step: 4481, Training Logs: loss_final: 0.493681, loss_mean: 0.825744, proj_loss: -0.412495, loss_mean_cls: 0.080431, grad_norm: 0.782717 +Steps: 0%| | 4482/1000000 [18:36<68:29:45, 4.04it/s, grad_norm=0.783, loss_final=0.494, loss_mean=0.826, loss_mean_cls=0.0804, proj_loss=-0.412][2026-03-26 12:49:40] Step: 4482, Training Logs: loss_final: 0.517001, loss_mean: 0.852752, proj_loss: -0.414487, loss_mean_cls: 0.078736, grad_norm: 0.395933 +Steps: 0%| | 4483/1000000 [18:37<68:30:31, 4.04it/s, grad_norm=0.396, loss_final=0.517, loss_mean=0.853, loss_mean_cls=0.0787, proj_loss=-0.414][2026-03-26 12:49:40] Step: 4483, Training Logs: loss_final: 0.501598, loss_mean: 0.840570, proj_loss: -0.418912, loss_mean_cls: 0.079940, grad_norm: 0.934376 +Steps: 0%| | 4484/1000000 [18:37<68:30:19, 4.04it/s, grad_norm=0.934, loss_final=0.502, loss_mean=0.841, loss_mean_cls=0.0799, proj_loss=-0.419][2026-03-26 12:49:40] Step: 4484, Training Logs: loss_final: 0.495358, loss_mean: 0.825454, proj_loss: -0.409987, loss_mean_cls: 0.079891, grad_norm: 0.542537 +Steps: 0%| | 4485/1000000 [18:37<68:29:42, 4.04it/s, grad_norm=0.543, loss_final=0.495, loss_mean=0.825, loss_mean_cls=0.0799, proj_loss=-0.41][2026-03-26 12:49:41] Step: 4485, Training Logs: loss_final: 0.494406, loss_mean: 0.827239, proj_loss: -0.413532, loss_mean_cls: 0.080699, grad_norm: 0.609644 +Steps: 0%| | 4486/1000000 [18:37<68:28:11, 4.04it/s, grad_norm=0.61, loss_final=0.494, loss_mean=0.827, loss_mean_cls=0.0807, proj_loss=-0.414][2026-03-26 12:49:41] Step: 4486, Training Logs: loss_final: 0.499524, loss_mean: 0.832945, proj_loss: -0.414011, loss_mean_cls: 0.080590, grad_norm: 0.577294 +Steps: 0%| | 4487/1000000 [18:38<68:25:56, 4.04it/s, grad_norm=0.577, loss_final=0.5, loss_mean=0.833, loss_mean_cls=0.0806, proj_loss=-0.414][2026-03-26 12:49:41] Step: 4487, Training Logs: loss_final: 0.502016, loss_mean: 0.838687, proj_loss: -0.416047, loss_mean_cls: 0.079376, grad_norm: 0.450218 +Steps: 0%| | 4488/1000000 [18:38<68:26:14, 4.04it/s, grad_norm=0.45, loss_final=0.502, loss_mean=0.839, loss_mean_cls=0.0794, proj_loss=-0.416][2026-03-26 12:49:41] Step: 4488, Training Logs: loss_final: 0.500878, loss_mean: 0.842053, proj_loss: -0.419769, loss_mean_cls: 0.078593, grad_norm: 0.825081 +Steps: 0%| | 4489/1000000 [18:38<68:24:38, 4.04it/s, grad_norm=0.825, loss_final=0.501, loss_mean=0.842, loss_mean_cls=0.0786, proj_loss=-0.42][2026-03-26 12:49:42] Step: 4489, Training Logs: loss_final: 0.495029, loss_mean: 0.809888, proj_loss: -0.396984, loss_mean_cls: 0.082126, grad_norm: 0.611416 +Steps: 0%| | 4490/1000000 [18:38<68:24:47, 4.04it/s, grad_norm=0.611, loss_final=0.495, loss_mean=0.81, loss_mean_cls=0.0821, proj_loss=-0.397][2026-03-26 12:49:42] Step: 4490, Training Logs: loss_final: 0.519502, loss_mean: 0.847077, proj_loss: -0.408239, loss_mean_cls: 0.080665, grad_norm: 0.355784 +Steps: 0%| | 4491/1000000 [18:39<68:26:12, 4.04it/s, grad_norm=0.356, loss_final=0.52, loss_mean=0.847, loss_mean_cls=0.0807, proj_loss=-0.408][2026-03-26 12:49:42] Step: 4491, Training Logs: loss_final: 0.495108, loss_mean: 0.833008, proj_loss: -0.418236, loss_mean_cls: 0.080335, grad_norm: 0.447459 +Steps: 0%| | 4492/1000000 [18:39<68:27:03, 4.04it/s, grad_norm=0.447, loss_final=0.495, loss_mean=0.833, loss_mean_cls=0.0803, proj_loss=-0.418][2026-03-26 12:49:42] Step: 4492, Training Logs: loss_final: 0.492955, loss_mean: 0.829276, proj_loss: -0.416348, loss_mean_cls: 0.080027, grad_norm: 0.300012 +Steps: 0%| | 4493/1000000 [18:39<68:25:48, 4.04it/s, grad_norm=0.3, loss_final=0.493, loss_mean=0.829, loss_mean_cls=0.08, proj_loss=-0.416][2026-03-26 12:49:43] Step: 4493, Training Logs: loss_final: 0.482910, loss_mean: 0.815762, proj_loss: -0.413738, loss_mean_cls: 0.080885, grad_norm: 0.362972 +Steps: 0%| | 4494/1000000 [18:39<68:27:38, 4.04it/s, grad_norm=0.363, loss_final=0.483, loss_mean=0.816, loss_mean_cls=0.0809, proj_loss=-0.414][2026-03-26 12:49:43] Step: 4494, Training Logs: loss_final: 0.500762, loss_mean: 0.836259, proj_loss: -0.415284, loss_mean_cls: 0.079787, grad_norm: 0.351329 +Steps: 0%| | 4495/1000000 [18:40<68:26:01, 4.04it/s, grad_norm=0.351, loss_final=0.501, loss_mean=0.836, loss_mean_cls=0.0798, proj_loss=-0.415][2026-03-26 12:49:43] Step: 4495, Training Logs: loss_final: 0.516836, loss_mean: 0.839755, proj_loss: -0.404098, loss_mean_cls: 0.081179, grad_norm: 0.407928 +Steps: 0%| | 4496/1000000 [18:40<68:25:51, 4.04it/s, grad_norm=0.408, loss_final=0.517, loss_mean=0.84, loss_mean_cls=0.0812, proj_loss=-0.404][2026-03-26 12:49:43] Step: 4496, Training Logs: loss_final: 0.496000, loss_mean: 0.836537, proj_loss: -0.417535, loss_mean_cls: 0.076998, grad_norm: 0.341473 +Steps: 0%| | 4497/1000000 [18:40<68:22:03, 4.04it/s, grad_norm=0.341, loss_final=0.496, loss_mean=0.837, loss_mean_cls=0.077, proj_loss=-0.418][2026-03-26 12:49:44] Step: 4497, Training Logs: loss_final: 0.499051, loss_mean: 0.814498, proj_loss: -0.400389, loss_mean_cls: 0.084942, grad_norm: 0.469061 +Steps: 0%| | 4498/1000000 [18:40<68:23:46, 4.04it/s, grad_norm=0.469, loss_final=0.499, loss_mean=0.814, loss_mean_cls=0.0849, proj_loss=-0.4][2026-03-26 12:49:44] Step: 4498, Training Logs: loss_final: 0.514497, loss_mean: 0.852836, proj_loss: -0.415395, loss_mean_cls: 0.077057, grad_norm: 0.301446 +Steps: 0%| | 4499/1000000 [18:41<68:22:24, 4.04it/s, grad_norm=0.301, loss_final=0.514, loss_mean=0.853, loss_mean_cls=0.0771, proj_loss=-0.415][2026-03-26 12:49:44] Step: 4499, Training Logs: loss_final: 0.504550, loss_mean: 0.823457, proj_loss: -0.401319, loss_mean_cls: 0.082412, grad_norm: 0.390552 +Steps: 0%| | 4500/1000000 [18:41<68:24:00, 4.04it/s, grad_norm=0.391, loss_final=0.505, loss_mean=0.823, loss_mean_cls=0.0824, proj_loss=-0.401][2026-03-26 12:49:44] Step: 4500, Training Logs: loss_final: 0.500176, loss_mean: 0.842479, proj_loss: -0.420151, loss_mean_cls: 0.077848, grad_norm: 0.490729 +Steps: 0%| | 4501/1000000 [18:41<68:24:35, 4.04it/s, grad_norm=0.491, loss_final=0.5, loss_mean=0.842, loss_mean_cls=0.0778, proj_loss=-0.42][2026-03-26 12:49:45] Step: 4501, Training Logs: loss_final: 0.517638, loss_mean: 0.846346, proj_loss: -0.407323, loss_mean_cls: 0.078615, grad_norm: 0.405676 +Steps: 0%| | 4502/1000000 [18:41<68:26:09, 4.04it/s, grad_norm=0.406, loss_final=0.518, loss_mean=0.846, loss_mean_cls=0.0786, proj_loss=-0.407][2026-03-26 12:49:45] Step: 4502, Training Logs: loss_final: 0.500479, loss_mean: 0.832426, proj_loss: -0.413008, loss_mean_cls: 0.081062, grad_norm: 0.382910 +Steps: 0%| | 4503/1000000 [18:42<68:23:02, 4.04it/s, grad_norm=0.383, loss_final=0.5, loss_mean=0.832, loss_mean_cls=0.0811, proj_loss=-0.413][2026-03-26 12:49:45] Step: 4503, Training Logs: loss_final: 0.491783, loss_mean: 0.817978, proj_loss: -0.408446, loss_mean_cls: 0.082251, grad_norm: 0.445137 +Steps: 0%| | 4504/1000000 [18:42<68:23:32, 4.04it/s, grad_norm=0.445, loss_final=0.492, loss_mean=0.818, loss_mean_cls=0.0823, proj_loss=-0.408][2026-03-26 12:49:45] Step: 4504, Training Logs: loss_final: 0.499728, loss_mean: 0.839748, proj_loss: -0.417350, loss_mean_cls: 0.077330, grad_norm: 0.580290 +Steps: 0%| | 4505/1000000 [18:42<68:23:24, 4.04it/s, grad_norm=0.58, loss_final=0.5, loss_mean=0.84, loss_mean_cls=0.0773, proj_loss=-0.417][2026-03-26 12:49:46] Step: 4505, Training Logs: loss_final: 0.501443, loss_mean: 0.828176, proj_loss: -0.407266, loss_mean_cls: 0.080533, grad_norm: 0.436766 +Steps: 0%| | 4506/1000000 [18:42<68:28:07, 4.04it/s, grad_norm=0.437, loss_final=0.501, loss_mean=0.828, loss_mean_cls=0.0805, proj_loss=-0.407][2026-03-26 12:49:46] Step: 4506, Training Logs: loss_final: 0.496074, loss_mean: 0.816558, proj_loss: -0.404297, loss_mean_cls: 0.083813, grad_norm: 0.427870 +Steps: 0%| | 4507/1000000 [18:43<68:24:40, 4.04it/s, grad_norm=0.428, loss_final=0.496, loss_mean=0.817, loss_mean_cls=0.0838, proj_loss=-0.404][2026-03-26 12:49:46] Step: 4507, Training Logs: loss_final: 0.505258, loss_mean: 0.829948, proj_loss: -0.406967, loss_mean_cls: 0.082277, grad_norm: 0.559076 +Steps: 0%| | 4508/1000000 [18:43<68:23:50, 4.04it/s, grad_norm=0.559, loss_final=0.505, loss_mean=0.83, loss_mean_cls=0.0823, proj_loss=-0.407][2026-03-26 12:49:46] Step: 4508, Training Logs: loss_final: 0.505679, loss_mean: 0.841844, proj_loss: -0.415875, loss_mean_cls: 0.079710, grad_norm: 0.556601 +Steps: 0%| | 4509/1000000 [18:43<68:25:15, 4.04it/s, grad_norm=0.557, loss_final=0.506, loss_mean=0.842, loss_mean_cls=0.0797, proj_loss=-0.416][2026-03-26 12:49:47] Step: 4509, Training Logs: loss_final: 0.480618, loss_mean: 0.809078, proj_loss: -0.410153, loss_mean_cls: 0.081693, grad_norm: 0.323306 +Steps: 0%| | 4510/1000000 [18:43<68:25:10, 4.04it/s, grad_norm=0.323, loss_final=0.481, loss_mean=0.809, loss_mean_cls=0.0817, proj_loss=-0.41][2026-03-26 12:49:47] Step: 4510, Training Logs: loss_final: 0.512500, loss_mean: 0.845644, proj_loss: -0.413044, loss_mean_cls: 0.079900, grad_norm: 0.489336 +Steps: 0%| | 4511/1000000 [18:44<68:27:57, 4.04it/s, grad_norm=0.489, loss_final=0.513, loss_mean=0.846, loss_mean_cls=0.0799, proj_loss=-0.413][2026-03-26 12:49:47] Step: 4511, Training Logs: loss_final: 0.499826, loss_mean: 0.829400, proj_loss: -0.410129, loss_mean_cls: 0.080555, grad_norm: 0.375984 +Steps: 0%| | 4512/1000000 [18:44<68:27:57, 4.04it/s, grad_norm=0.376, loss_final=0.5, loss_mean=0.829, loss_mean_cls=0.0806, proj_loss=-0.41][2026-03-26 12:49:47] Step: 4512, Training Logs: loss_final: 0.492155, loss_mean: 0.816139, proj_loss: -0.405141, loss_mean_cls: 0.081158, grad_norm: 0.375782 +Steps: 0%| | 4513/1000000 [18:44<68:26:52, 4.04it/s, grad_norm=0.376, loss_final=0.492, loss_mean=0.816, loss_mean_cls=0.0812, proj_loss=-0.405][2026-03-26 12:49:48] Step: 4513, Training Logs: loss_final: 0.504126, loss_mean: 0.830864, proj_loss: -0.407179, loss_mean_cls: 0.080441, grad_norm: 0.300520 +Steps: 0%| | 4514/1000000 [18:44<68:25:46, 4.04it/s, grad_norm=0.301, loss_final=0.504, loss_mean=0.831, loss_mean_cls=0.0804, proj_loss=-0.407][2026-03-26 12:49:48] Step: 4514, Training Logs: loss_final: 0.498382, loss_mean: 0.834898, proj_loss: -0.415823, loss_mean_cls: 0.079307, grad_norm: 0.324880 +Steps: 0%| | 4515/1000000 [18:45<68:27:22, 4.04it/s, grad_norm=0.325, loss_final=0.498, loss_mean=0.835, loss_mean_cls=0.0793, proj_loss=-0.416][2026-03-26 12:49:48] Step: 4515, Training Logs: loss_final: 0.474861, loss_mean: 0.807692, proj_loss: -0.414207, loss_mean_cls: 0.081376, grad_norm: 0.677723 +Steps: 0%| | 4516/1000000 [18:45<68:23:47, 4.04it/s, grad_norm=0.678, loss_final=0.475, loss_mean=0.808, loss_mean_cls=0.0814, proj_loss=-0.414][2026-03-26 12:49:48] Step: 4516, Training Logs: loss_final: 0.504196, loss_mean: 0.824254, proj_loss: -0.402326, loss_mean_cls: 0.082268, grad_norm: 0.493691 +Steps: 0%| | 4517/1000000 [18:45<68:24:27, 4.04it/s, grad_norm=0.494, loss_final=0.504, loss_mean=0.824, loss_mean_cls=0.0823, proj_loss=-0.402][2026-03-26 12:49:49] Step: 4517, Training Logs: loss_final: 0.495313, loss_mean: 0.827976, proj_loss: -0.413944, loss_mean_cls: 0.081281, grad_norm: 0.397570 +Steps: 0%| | 4518/1000000 [18:45<68:23:18, 4.04it/s, grad_norm=0.398, loss_final=0.495, loss_mean=0.828, loss_mean_cls=0.0813, proj_loss=-0.414][2026-03-26 12:49:49] Step: 4518, Training Logs: loss_final: 0.517426, loss_mean: 0.847251, proj_loss: -0.408978, loss_mean_cls: 0.079152, grad_norm: 0.437871 +Steps: 0%| | 4519/1000000 [18:46<68:24:49, 4.04it/s, grad_norm=0.438, loss_final=0.517, loss_mean=0.847, loss_mean_cls=0.0792, proj_loss=-0.409][2026-03-26 12:49:49] Step: 4519, Training Logs: loss_final: 0.494788, loss_mean: 0.825729, proj_loss: -0.410950, loss_mean_cls: 0.080009, grad_norm: 0.259189 +Steps: 0%| | 4520/1000000 [18:46<68:24:44, 4.04it/s, grad_norm=0.259, loss_final=0.495, loss_mean=0.826, loss_mean_cls=0.08, proj_loss=-0.411][2026-03-26 12:49:49] Step: 4520, Training Logs: loss_final: 0.494617, loss_mean: 0.816144, proj_loss: -0.403961, loss_mean_cls: 0.082434, grad_norm: 0.351329 +Steps: 0%| | 4521/1000000 [18:46<68:25:55, 4.04it/s, grad_norm=0.351, loss_final=0.495, loss_mean=0.816, loss_mean_cls=0.0824, proj_loss=-0.404][2026-03-26 12:49:50] Step: 4521, Training Logs: loss_final: 0.504282, loss_mean: 0.833471, proj_loss: -0.411204, loss_mean_cls: 0.082015, grad_norm: 0.257299 +Steps: 0%| | 4522/1000000 [18:46<68:28:51, 4.04it/s, grad_norm=0.257, loss_final=0.504, loss_mean=0.833, loss_mean_cls=0.082, proj_loss=-0.411][2026-03-26 12:49:50] Step: 4522, Training Logs: loss_final: 0.504045, loss_mean: 0.844573, proj_loss: -0.417284, loss_mean_cls: 0.076757, grad_norm: 0.553253 +Steps: 0%| | 4523/1000000 [18:47<68:29:03, 4.04it/s, grad_norm=0.553, loss_final=0.504, loss_mean=0.845, loss_mean_cls=0.0768, proj_loss=-0.417][2026-03-26 12:49:50] Step: 4523, Training Logs: loss_final: 0.498823, loss_mean: 0.825821, proj_loss: -0.407164, loss_mean_cls: 0.080166, grad_norm: 0.713980 +Steps: 0%| | 4524/1000000 [18:47<68:28:15, 4.04it/s, grad_norm=0.714, loss_final=0.499, loss_mean=0.826, loss_mean_cls=0.0802, proj_loss=-0.407][2026-03-26 12:49:50] Step: 4524, Training Logs: loss_final: 0.510152, loss_mean: 0.821950, proj_loss: -0.396786, loss_mean_cls: 0.084988, grad_norm: 0.531417 +Steps: 0%| | 4525/1000000 [18:47<68:27:39, 4.04it/s, grad_norm=0.531, loss_final=0.51, loss_mean=0.822, loss_mean_cls=0.085, proj_loss=-0.397][2026-03-26 12:49:51] Step: 4525, Training Logs: loss_final: 0.503090, loss_mean: 0.826537, proj_loss: -0.404842, loss_mean_cls: 0.081396, grad_norm: 0.622986 +Steps: 0%| | 4526/1000000 [18:47<68:27:29, 4.04it/s, grad_norm=0.623, loss_final=0.503, loss_mean=0.827, loss_mean_cls=0.0814, proj_loss=-0.405][2026-03-26 12:49:51] Step: 4526, Training Logs: loss_final: 0.496192, loss_mean: 0.834150, proj_loss: -0.416468, loss_mean_cls: 0.078510, grad_norm: 0.309463 +Steps: 0%| | 4527/1000000 [18:48<68:27:30, 4.04it/s, grad_norm=0.309, loss_final=0.496, loss_mean=0.834, loss_mean_cls=0.0785, proj_loss=-0.416][2026-03-26 12:49:51] Step: 4527, Training Logs: loss_final: 0.504897, loss_mean: 0.833993, proj_loss: -0.408250, loss_mean_cls: 0.079154, grad_norm: 0.851582 +Steps: 0%| | 4528/1000000 [18:48<68:28:47, 4.04it/s, grad_norm=0.852, loss_final=0.505, loss_mean=0.834, loss_mean_cls=0.0792, proj_loss=-0.408][2026-03-26 12:49:51] Step: 4528, Training Logs: loss_final: 0.505406, loss_mean: 0.831307, proj_loss: -0.406604, loss_mean_cls: 0.080703, grad_norm: 0.751155 +Steps: 0%| | 4529/1000000 [18:48<68:27:07, 4.04it/s, grad_norm=0.751, loss_final=0.505, loss_mean=0.831, loss_mean_cls=0.0807, proj_loss=-0.407][2026-03-26 12:49:51] Step: 4529, Training Logs: loss_final: 0.506332, loss_mean: 0.836444, proj_loss: -0.409252, loss_mean_cls: 0.079140, grad_norm: 0.293482 +Steps: 0%| | 4530/1000000 [18:48<68:26:35, 4.04it/s, grad_norm=0.293, loss_final=0.506, loss_mean=0.836, loss_mean_cls=0.0791, proj_loss=-0.409][2026-03-26 12:49:52] Step: 4530, Training Logs: loss_final: 0.503408, loss_mean: 0.825079, proj_loss: -0.402357, loss_mean_cls: 0.080685, grad_norm: 0.716369 +Steps: 0%| | 4531/1000000 [18:49<68:30:32, 4.04it/s, grad_norm=0.716, loss_final=0.503, loss_mean=0.825, loss_mean_cls=0.0807, proj_loss=-0.402][2026-03-26 12:49:52] Step: 4531, Training Logs: loss_final: 0.503109, loss_mean: 0.839454, proj_loss: -0.416446, loss_mean_cls: 0.080100, grad_norm: 0.467509 +Steps: 0%| | 4532/1000000 [18:49<68:29:20, 4.04it/s, grad_norm=0.468, loss_final=0.503, loss_mean=0.839, loss_mean_cls=0.0801, proj_loss=-0.416][2026-03-26 12:49:52] Step: 4532, Training Logs: loss_final: 0.504985, loss_mean: 0.837389, proj_loss: -0.412016, loss_mean_cls: 0.079612, grad_norm: 0.848318 +Steps: 0%| | 4533/1000000 [18:49<68:27:57, 4.04it/s, grad_norm=0.848, loss_final=0.505, loss_mean=0.837, loss_mean_cls=0.0796, proj_loss=-0.412][2026-03-26 12:49:52] Step: 4533, Training Logs: loss_final: 0.493819, loss_mean: 0.828713, proj_loss: -0.413741, loss_mean_cls: 0.078846, grad_norm: 0.621639 +Steps: 0%| | 4534/1000000 [18:49<68:30:08, 4.04it/s, grad_norm=0.622, loss_final=0.494, loss_mean=0.829, loss_mean_cls=0.0788, proj_loss=-0.414][2026-03-26 12:49:53] Step: 4534, Training Logs: loss_final: 0.506179, loss_mean: 0.833953, proj_loss: -0.408654, loss_mean_cls: 0.080880, grad_norm: 0.585414 +Steps: 0%| | 4535/1000000 [18:50<68:26:09, 4.04it/s, grad_norm=0.585, loss_final=0.506, loss_mean=0.834, loss_mean_cls=0.0809, proj_loss=-0.409][2026-03-26 12:49:53] Step: 4535, Training Logs: loss_final: 0.518256, loss_mean: 0.859570, proj_loss: -0.419187, loss_mean_cls: 0.077873, grad_norm: 0.886594 +Steps: 0%| | 4536/1000000 [18:50<68:24:30, 4.04it/s, grad_norm=0.887, loss_final=0.518, loss_mean=0.86, loss_mean_cls=0.0779, proj_loss=-0.419][2026-03-26 12:49:53] Step: 4536, Training Logs: loss_final: 0.509142, loss_mean: 0.850640, proj_loss: -0.419394, loss_mean_cls: 0.077896, grad_norm: 0.380608 +Steps: 0%| | 4537/1000000 [18:50<68:22:04, 4.04it/s, grad_norm=0.381, loss_final=0.509, loss_mean=0.851, loss_mean_cls=0.0779, proj_loss=-0.419][2026-03-26 12:49:53] Step: 4537, Training Logs: loss_final: 0.520776, loss_mean: 0.853573, proj_loss: -0.413512, loss_mean_cls: 0.080714, grad_norm: 0.644688 +Steps: 0%| | 4538/1000000 [18:50<68:24:49, 4.04it/s, grad_norm=0.645, loss_final=0.521, loss_mean=0.854, loss_mean_cls=0.0807, proj_loss=-0.414][2026-03-26 12:49:54] Step: 4538, Training Logs: loss_final: 0.498207, loss_mean: 0.830723, proj_loss: -0.413747, loss_mean_cls: 0.081231, grad_norm: 0.546385 +Steps: 0%| | 4539/1000000 [18:51<68:22:31, 4.04it/s, grad_norm=0.546, loss_final=0.498, loss_mean=0.831, loss_mean_cls=0.0812, proj_loss=-0.414][2026-03-26 12:49:54] Step: 4539, Training Logs: loss_final: 0.492504, loss_mean: 0.802196, proj_loss: -0.394918, loss_mean_cls: 0.085225, grad_norm: 0.491896 +Steps: 0%| | 4540/1000000 [18:51<69:19:35, 3.99it/s, grad_norm=0.492, loss_final=0.493, loss_mean=0.802, loss_mean_cls=0.0852, proj_loss=-0.395][2026-03-26 12:49:54] Step: 4540, Training Logs: loss_final: 0.497251, loss_mean: 0.832188, proj_loss: -0.414963, loss_mean_cls: 0.080026, grad_norm: 0.720622 +Steps: 0%| | 4541/1000000 [18:51<69:02:31, 4.01it/s, grad_norm=0.721, loss_final=0.497, loss_mean=0.832, loss_mean_cls=0.08, proj_loss=-0.415][2026-03-26 12:49:54] Step: 4541, Training Logs: loss_final: 0.497189, loss_mean: 0.830455, proj_loss: -0.414629, loss_mean_cls: 0.081363, grad_norm: 0.341104 +Steps: 0%| | 4542/1000000 [18:51<68:49:39, 4.02it/s, grad_norm=0.341, loss_final=0.497, loss_mean=0.83, loss_mean_cls=0.0814, proj_loss=-0.415][2026-03-26 12:49:55] Step: 4542, Training Logs: loss_final: 0.528922, loss_mean: 0.857946, proj_loss: -0.408964, loss_mean_cls: 0.079940, grad_norm: 0.475437 +Steps: 0%| | 4543/1000000 [18:52<68:39:21, 4.03it/s, grad_norm=0.475, loss_final=0.529, loss_mean=0.858, loss_mean_cls=0.0799, proj_loss=-0.409][2026-03-26 12:49:55] Step: 4543, Training Logs: loss_final: 0.506692, loss_mean: 0.835344, proj_loss: -0.410043, loss_mean_cls: 0.081390, grad_norm: 0.259829 +Steps: 0%| | 4544/1000000 [18:52<68:37:10, 4.03it/s, grad_norm=0.26, loss_final=0.507, loss_mean=0.835, loss_mean_cls=0.0814, proj_loss=-0.41][2026-03-26 12:49:55] Step: 4544, Training Logs: loss_final: 0.500039, loss_mean: 0.832357, proj_loss: -0.411907, loss_mean_cls: 0.079590, grad_norm: 0.517181 +Steps: 0%| | 4545/1000000 [18:52<68:32:47, 4.03it/s, grad_norm=0.517, loss_final=0.5, loss_mean=0.832, loss_mean_cls=0.0796, proj_loss=-0.412][2026-03-26 12:49:55] Step: 4545, Training Logs: loss_final: 0.504684, loss_mean: 0.843573, proj_loss: -0.415872, loss_mean_cls: 0.076983, grad_norm: 0.382962 +Steps: 0%| | 4546/1000000 [18:52<68:31:30, 4.04it/s, grad_norm=0.383, loss_final=0.505, loss_mean=0.844, loss_mean_cls=0.077, proj_loss=-0.416][2026-03-26 12:49:56] Step: 4546, Training Logs: loss_final: 0.494149, loss_mean: 0.812665, proj_loss: -0.401000, loss_mean_cls: 0.082484, grad_norm: 0.390464 +Steps: 0%| | 4547/1000000 [18:53<68:28:35, 4.04it/s, grad_norm=0.39, loss_final=0.494, loss_mean=0.813, loss_mean_cls=0.0825, proj_loss=-0.401][2026-03-26 12:49:56] Step: 4547, Training Logs: loss_final: 0.498143, loss_mean: 0.832023, proj_loss: -0.413438, loss_mean_cls: 0.079558, grad_norm: 0.508127 +Steps: 0%| | 4548/1000000 [18:53<68:28:03, 4.04it/s, grad_norm=0.508, loss_final=0.498, loss_mean=0.832, loss_mean_cls=0.0796, proj_loss=-0.413][2026-03-26 12:49:56] Step: 4548, Training Logs: loss_final: 0.493297, loss_mean: 0.820645, proj_loss: -0.408434, loss_mean_cls: 0.081087, grad_norm: 0.354713 +Steps: 0%| | 4549/1000000 [18:53<68:29:49, 4.04it/s, grad_norm=0.355, loss_final=0.493, loss_mean=0.821, loss_mean_cls=0.0811, proj_loss=-0.408][2026-03-26 12:49:56] Step: 4549, Training Logs: loss_final: 0.494242, loss_mean: 0.835515, proj_loss: -0.419425, loss_mean_cls: 0.078152, grad_norm: 0.361321 +Steps: 0%| | 4550/1000000 [18:53<68:27:54, 4.04it/s, grad_norm=0.361, loss_final=0.494, loss_mean=0.836, loss_mean_cls=0.0782, proj_loss=-0.419][2026-03-26 12:49:57] Step: 4550, Training Logs: loss_final: 0.500120, loss_mean: 0.830991, proj_loss: -0.410615, loss_mean_cls: 0.079743, grad_norm: 0.309646 +Steps: 0%| | 4551/1000000 [18:54<68:26:37, 4.04it/s, grad_norm=0.31, loss_final=0.5, loss_mean=0.831, loss_mean_cls=0.0797, proj_loss=-0.411][2026-03-26 12:49:57] Step: 4551, Training Logs: loss_final: 0.506114, loss_mean: 0.835227, proj_loss: -0.408660, loss_mean_cls: 0.079547, grad_norm: 0.423291 +Steps: 0%| | 4552/1000000 [18:54<68:26:29, 4.04it/s, grad_norm=0.423, loss_final=0.506, loss_mean=0.835, loss_mean_cls=0.0795, proj_loss=-0.409][2026-03-26 12:49:57] Step: 4552, Training Logs: loss_final: 0.487836, loss_mean: 0.818039, proj_loss: -0.411113, loss_mean_cls: 0.080910, grad_norm: 0.528403 +Steps: 0%| | 4553/1000000 [18:54<68:27:03, 4.04it/s, grad_norm=0.528, loss_final=0.488, loss_mean=0.818, loss_mean_cls=0.0809, proj_loss=-0.411][2026-03-26 12:49:57] Step: 4553, Training Logs: loss_final: 0.487123, loss_mean: 0.823825, proj_loss: -0.414544, loss_mean_cls: 0.077843, grad_norm: 0.277602 +Steps: 0%| | 4554/1000000 [18:54<68:26:13, 4.04it/s, grad_norm=0.278, loss_final=0.487, loss_mean=0.824, loss_mean_cls=0.0778, proj_loss=-0.415][2026-03-26 12:49:58] Step: 4554, Training Logs: loss_final: 0.521744, loss_mean: 0.849596, proj_loss: -0.408534, loss_mean_cls: 0.080681, grad_norm: 0.459535 +Steps: 0%| | 4555/1000000 [18:55<68:26:57, 4.04it/s, grad_norm=0.46, loss_final=0.522, loss_mean=0.85, loss_mean_cls=0.0807, proj_loss=-0.409][2026-03-26 12:49:58] Step: 4555, Training Logs: loss_final: 0.500244, loss_mean: 0.831202, proj_loss: -0.411286, loss_mean_cls: 0.080328, grad_norm: 0.350152 +Steps: 0%| | 4556/1000000 [18:55<68:27:26, 4.04it/s, grad_norm=0.35, loss_final=0.5, loss_mean=0.831, loss_mean_cls=0.0803, proj_loss=-0.411][2026-03-26 12:49:58] Step: 4556, Training Logs: loss_final: 0.521545, loss_mean: 0.837297, proj_loss: -0.397028, loss_mean_cls: 0.081275, grad_norm: 0.327365 +Steps: 0%| | 4557/1000000 [18:55<68:25:37, 4.04it/s, grad_norm=0.327, loss_final=0.522, loss_mean=0.837, loss_mean_cls=0.0813, proj_loss=-0.397][2026-03-26 12:49:58] Step: 4557, Training Logs: loss_final: 0.497802, loss_mean: 0.838827, proj_loss: -0.419178, loss_mean_cls: 0.078153, grad_norm: 0.404800 +Steps: 0%| | 4558/1000000 [18:55<68:22:32, 4.04it/s, grad_norm=0.405, loss_final=0.498, loss_mean=0.839, loss_mean_cls=0.0782, proj_loss=-0.419][2026-03-26 12:49:59] Step: 4558, Training Logs: loss_final: 0.499272, loss_mean: 0.844301, proj_loss: -0.420426, loss_mean_cls: 0.075398, grad_norm: 0.340214 +Steps: 0%| | 4559/1000000 [18:56<68:20:24, 4.05it/s, grad_norm=0.34, loss_final=0.499, loss_mean=0.844, loss_mean_cls=0.0754, proj_loss=-0.42][2026-03-26 12:49:59] Step: 4559, Training Logs: loss_final: 0.484896, loss_mean: 0.813887, proj_loss: -0.409565, loss_mean_cls: 0.080574, grad_norm: 0.314138 +Steps: 0%| | 4560/1000000 [18:56<68:21:24, 4.05it/s, grad_norm=0.314, loss_final=0.485, loss_mean=0.814, loss_mean_cls=0.0806, proj_loss=-0.41][2026-03-26 12:49:59] Step: 4560, Training Logs: loss_final: 0.492828, loss_mean: 0.814349, proj_loss: -0.403166, loss_mean_cls: 0.081645, grad_norm: 0.470530 +Steps: 0%| | 4561/1000000 [18:56<68:22:23, 4.04it/s, grad_norm=0.471, loss_final=0.493, loss_mean=0.814, loss_mean_cls=0.0816, proj_loss=-0.403][2026-03-26 12:49:59] Step: 4561, Training Logs: loss_final: 0.499608, loss_mean: 0.815337, proj_loss: -0.398348, loss_mean_cls: 0.082619, grad_norm: 0.283410 +Steps: 0%| | 4562/1000000 [18:56<68:21:19, 4.05it/s, grad_norm=0.283, loss_final=0.5, loss_mean=0.815, loss_mean_cls=0.0826, proj_loss=-0.398][2026-03-26 12:50:00] Step: 4562, Training Logs: loss_final: 0.508106, loss_mean: 0.835950, proj_loss: -0.408843, loss_mean_cls: 0.081000, grad_norm: 0.424049 +Steps: 0%| | 4563/1000000 [18:57<68:22:06, 4.04it/s, grad_norm=0.424, loss_final=0.508, loss_mean=0.836, loss_mean_cls=0.081, proj_loss=-0.409][2026-03-26 12:50:00] Step: 4563, Training Logs: loss_final: 0.499580, loss_mean: 0.824425, proj_loss: -0.406592, loss_mean_cls: 0.081747, grad_norm: 0.452706 +Steps: 0%| | 4564/1000000 [18:57<68:24:28, 4.04it/s, grad_norm=0.453, loss_final=0.5, loss_mean=0.824, loss_mean_cls=0.0817, proj_loss=-0.407][2026-03-26 12:50:00] Step: 4564, Training Logs: loss_final: 0.499933, loss_mean: 0.827835, proj_loss: -0.409546, loss_mean_cls: 0.081644, grad_norm: 0.423096 +Steps: 0%| | 4565/1000000 [18:57<68:21:57, 4.04it/s, grad_norm=0.423, loss_final=0.5, loss_mean=0.828, loss_mean_cls=0.0816, proj_loss=-0.41][2026-03-26 12:50:00] Step: 4565, Training Logs: loss_final: 0.505422, loss_mean: 0.821313, proj_loss: -0.399411, loss_mean_cls: 0.083521, grad_norm: 0.392778 +Steps: 0%| | 4566/1000000 [18:57<68:22:56, 4.04it/s, grad_norm=0.393, loss_final=0.505, loss_mean=0.821, loss_mean_cls=0.0835, proj_loss=-0.399][2026-03-26 12:50:01] Step: 4566, Training Logs: loss_final: 0.505392, loss_mean: 0.846364, proj_loss: -0.417949, loss_mean_cls: 0.076977, grad_norm: 0.365576 +Steps: 0%| | 4567/1000000 [18:58<68:23:21, 4.04it/s, grad_norm=0.366, loss_final=0.505, loss_mean=0.846, loss_mean_cls=0.077, proj_loss=-0.418][2026-03-26 12:50:01] Step: 4567, Training Logs: loss_final: 0.492464, loss_mean: 0.818221, proj_loss: -0.407581, loss_mean_cls: 0.081824, grad_norm: 0.394308 +Steps: 0%| | 4568/1000000 [18:58<68:26:19, 4.04it/s, grad_norm=0.394, loss_final=0.492, loss_mean=0.818, loss_mean_cls=0.0818, proj_loss=-0.408][2026-03-26 12:50:01] Step: 4568, Training Logs: loss_final: 0.502643, loss_mean: 0.832560, proj_loss: -0.411592, loss_mean_cls: 0.081675, grad_norm: 0.355107 +Steps: 0%| | 4569/1000000 [18:58<68:26:31, 4.04it/s, grad_norm=0.355, loss_final=0.503, loss_mean=0.833, loss_mean_cls=0.0817, proj_loss=-0.412][2026-03-26 12:50:01] Step: 4569, Training Logs: loss_final: 0.507008, loss_mean: 0.835102, proj_loss: -0.408545, loss_mean_cls: 0.080452, grad_norm: 0.340913 +Steps: 0%| | 4570/1000000 [18:58<68:55:53, 4.01it/s, grad_norm=0.341, loss_final=0.507, loss_mean=0.835, loss_mean_cls=0.0805, proj_loss=-0.409][2026-03-26 12:50:02] Step: 4570, Training Logs: loss_final: 0.505517, loss_mean: 0.837752, proj_loss: -0.412087, loss_mean_cls: 0.079852, grad_norm: 0.398923 +Steps: 0%| | 4571/1000000 [18:59<72:10:01, 3.83it/s, grad_norm=0.399, loss_final=0.506, loss_mean=0.838, loss_mean_cls=0.0799, proj_loss=-0.412][2026-03-26 12:50:02] Step: 4571, Training Logs: loss_final: 0.508317, loss_mean: 0.856998, proj_loss: -0.424946, loss_mean_cls: 0.076265, grad_norm: 0.333849 +Steps: 0%| | 4572/1000000 [18:59<71:01:30, 3.89it/s, grad_norm=0.334, loss_final=0.508, loss_mean=0.857, loss_mean_cls=0.0763, proj_loss=-0.425][2026-03-26 12:50:02] Step: 4572, Training Logs: loss_final: 0.514821, loss_mean: 0.836479, proj_loss: -0.401598, loss_mean_cls: 0.079941, grad_norm: 0.437904 +Steps: 0%| | 4573/1000000 [18:59<70:15:55, 3.94it/s, grad_norm=0.438, loss_final=0.515, loss_mean=0.836, loss_mean_cls=0.0799, proj_loss=-0.402][2026-03-26 12:50:02] Step: 4573, Training Logs: loss_final: 0.477677, loss_mean: 0.804366, proj_loss: -0.409120, loss_mean_cls: 0.082432, grad_norm: 0.447929 +Steps: 0%| | 4574/1000000 [18:59<69:41:35, 3.97it/s, grad_norm=0.448, loss_final=0.478, loss_mean=0.804, loss_mean_cls=0.0824, proj_loss=-0.409][2026-03-26 12:50:03] Step: 4574, Training Logs: loss_final: 0.493045, loss_mean: 0.828799, proj_loss: -0.417203, loss_mean_cls: 0.081449, grad_norm: 0.328973 +Steps: 0%| | 4575/1000000 [19:00<69:19:27, 3.99it/s, grad_norm=0.329, loss_final=0.493, loss_mean=0.829, loss_mean_cls=0.0814, proj_loss=-0.417][2026-03-26 12:50:03] Step: 4575, Training Logs: loss_final: 0.494231, loss_mean: 0.817384, proj_loss: -0.405450, loss_mean_cls: 0.082297, grad_norm: 0.543854 +Steps: 0%| | 4576/1000000 [19:00<69:03:13, 4.00it/s, grad_norm=0.544, loss_final=0.494, loss_mean=0.817, loss_mean_cls=0.0823, proj_loss=-0.405][2026-03-26 12:50:03] Step: 4576, Training Logs: loss_final: 0.509651, loss_mean: 0.846023, proj_loss: -0.414942, loss_mean_cls: 0.078570, grad_norm: 0.448539 +Steps: 0%| | 4577/1000000 [19:00<68:51:04, 4.02it/s, grad_norm=0.449, loss_final=0.51, loss_mean=0.846, loss_mean_cls=0.0786, proj_loss=-0.415][2026-03-26 12:50:03] Step: 4577, Training Logs: loss_final: 0.493633, loss_mean: 0.830387, proj_loss: -0.416125, loss_mean_cls: 0.079370, grad_norm: 0.360263 +Steps: 0%| | 4578/1000000 [19:00<68:41:40, 4.03it/s, grad_norm=0.36, loss_final=0.494, loss_mean=0.83, loss_mean_cls=0.0794, proj_loss=-0.416][2026-03-26 12:50:04] Step: 4578, Training Logs: loss_final: 0.495664, loss_mean: 0.837785, proj_loss: -0.420160, loss_mean_cls: 0.078039, grad_norm: 0.352341 +Steps: 0%| | 4579/1000000 [19:01<68:39:57, 4.03it/s, grad_norm=0.352, loss_final=0.496, loss_mean=0.838, loss_mean_cls=0.078, proj_loss=-0.42][2026-03-26 12:50:04] Step: 4579, Training Logs: loss_final: 0.496293, loss_mean: 0.826068, proj_loss: -0.409114, loss_mean_cls: 0.079339, grad_norm: 0.332861 +Steps: 0%| | 4580/1000000 [19:01<68:35:00, 4.03it/s, grad_norm=0.333, loss_final=0.496, loss_mean=0.826, loss_mean_cls=0.0793, proj_loss=-0.409][2026-03-26 12:50:04] Step: 4580, Training Logs: loss_final: 0.516604, loss_mean: 0.853774, proj_loss: -0.414258, loss_mean_cls: 0.077089, grad_norm: 0.308092 +Steps: 0%| | 4581/1000000 [19:01<68:30:29, 4.04it/s, grad_norm=0.308, loss_final=0.517, loss_mean=0.854, loss_mean_cls=0.0771, proj_loss=-0.414][2026-03-26 12:50:04] Step: 4581, Training Logs: loss_final: 0.493539, loss_mean: 0.832253, proj_loss: -0.416954, loss_mean_cls: 0.078240, grad_norm: 0.406883 +Steps: 0%| | 4582/1000000 [19:01<68:26:53, 4.04it/s, grad_norm=0.407, loss_final=0.494, loss_mean=0.832, loss_mean_cls=0.0782, proj_loss=-0.417][2026-03-26 12:50:05] Step: 4582, Training Logs: loss_final: 0.504520, loss_mean: 0.841769, proj_loss: -0.415667, loss_mean_cls: 0.078418, grad_norm: 0.314706 +Steps: 0%| | 4583/1000000 [19:02<68:29:17, 4.04it/s, grad_norm=0.315, loss_final=0.505, loss_mean=0.842, loss_mean_cls=0.0784, proj_loss=-0.416][2026-03-26 12:50:05] Step: 4583, Training Logs: loss_final: 0.483659, loss_mean: 0.814818, proj_loss: -0.412874, loss_mean_cls: 0.081715, grad_norm: 0.343459 +Steps: 0%| | 4584/1000000 [19:02<68:28:13, 4.04it/s, grad_norm=0.343, loss_final=0.484, loss_mean=0.815, loss_mean_cls=0.0817, proj_loss=-0.413][2026-03-26 12:50:05] Step: 4584, Training Logs: loss_final: 0.492200, loss_mean: 0.802744, proj_loss: -0.396725, loss_mean_cls: 0.086181, grad_norm: 0.398363 +Steps: 0%| | 4585/1000000 [19:02<68:26:48, 4.04it/s, grad_norm=0.398, loss_final=0.492, loss_mean=0.803, loss_mean_cls=0.0862, proj_loss=-0.397][2026-03-26 12:50:05] Step: 4585, Training Logs: loss_final: 0.505998, loss_mean: 0.838120, proj_loss: -0.412227, loss_mean_cls: 0.080106, grad_norm: 0.292217 +Steps: 0%| | 4586/1000000 [19:02<68:26:27, 4.04it/s, grad_norm=0.292, loss_final=0.506, loss_mean=0.838, loss_mean_cls=0.0801, proj_loss=-0.412][2026-03-26 12:50:06] Step: 4586, Training Logs: loss_final: 0.512282, loss_mean: 0.848345, proj_loss: -0.414519, loss_mean_cls: 0.078456, grad_norm: 0.376926 +Steps: 0%| | 4587/1000000 [19:03<68:27:38, 4.04it/s, grad_norm=0.377, loss_final=0.512, loss_mean=0.848, loss_mean_cls=0.0785, proj_loss=-0.415][2026-03-26 12:50:06] Step: 4587, Training Logs: loss_final: 0.490690, loss_mean: 0.825817, proj_loss: -0.415538, loss_mean_cls: 0.080411, grad_norm: 0.323514 +Steps: 0%| | 4588/1000000 [19:03<68:23:57, 4.04it/s, grad_norm=0.324, loss_final=0.491, loss_mean=0.826, loss_mean_cls=0.0804, proj_loss=-0.416][2026-03-26 12:50:06] Step: 4588, Training Logs: loss_final: 0.493598, loss_mean: 0.818157, proj_loss: -0.405499, loss_mean_cls: 0.080940, grad_norm: 0.407211 +Steps: 0%| | 4589/1000000 [19:03<68:23:56, 4.04it/s, grad_norm=0.407, loss_final=0.494, loss_mean=0.818, loss_mean_cls=0.0809, proj_loss=-0.405][2026-03-26 12:50:06] Step: 4589, Training Logs: loss_final: 0.488589, loss_mean: 0.823029, proj_loss: -0.414878, loss_mean_cls: 0.080438, grad_norm: 0.293902 +Steps: 0%| | 4590/1000000 [19:03<68:38:12, 4.03it/s, grad_norm=0.294, loss_final=0.489, loss_mean=0.823, loss_mean_cls=0.0804, proj_loss=-0.415][2026-03-26 12:50:07] Step: 4590, Training Logs: loss_final: 0.492952, loss_mean: 0.831639, proj_loss: -0.417074, loss_mean_cls: 0.078387, grad_norm: 0.419565 +Steps: 0%| | 4591/1000000 [19:04<68:35:28, 4.03it/s, grad_norm=0.42, loss_final=0.493, loss_mean=0.832, loss_mean_cls=0.0784, proj_loss=-0.417][2026-03-26 12:50:07] Step: 4591, Training Logs: loss_final: 0.499506, loss_mean: 0.828395, proj_loss: -0.409669, loss_mean_cls: 0.080780, grad_norm: 0.557282 +Steps: 0%| | 4592/1000000 [19:04<68:35:38, 4.03it/s, grad_norm=0.557, loss_final=0.5, loss_mean=0.828, loss_mean_cls=0.0808, proj_loss=-0.41][2026-03-26 12:50:07] Step: 4592, Training Logs: loss_final: 0.508895, loss_mean: 0.844956, proj_loss: -0.415167, loss_mean_cls: 0.079106, grad_norm: 0.336689 +Steps: 0%| | 4593/1000000 [19:04<68:33:24, 4.03it/s, grad_norm=0.337, loss_final=0.509, loss_mean=0.845, loss_mean_cls=0.0791, proj_loss=-0.415][2026-03-26 12:50:07] Step: 4593, Training Logs: loss_final: 0.493334, loss_mean: 0.813017, proj_loss: -0.403699, loss_mean_cls: 0.084016, grad_norm: 0.988680 +Steps: 0%| | 4594/1000000 [19:04<68:30:22, 4.04it/s, grad_norm=0.989, loss_final=0.493, loss_mean=0.813, loss_mean_cls=0.084, proj_loss=-0.404][2026-03-26 12:50:08] Step: 4594, Training Logs: loss_final: 0.495327, loss_mean: 0.833929, proj_loss: -0.418056, loss_mean_cls: 0.079454, grad_norm: 0.570548 +Steps: 0%| | 4595/1000000 [19:04<68:27:37, 4.04it/s, grad_norm=0.571, loss_final=0.495, loss_mean=0.834, loss_mean_cls=0.0795, proj_loss=-0.418][2026-03-26 12:50:08] Step: 4595, Training Logs: loss_final: 0.507015, loss_mean: 0.844442, proj_loss: -0.417044, loss_mean_cls: 0.079616, grad_norm: 0.422079 +Steps: 0%| | 4596/1000000 [19:05<68:30:08, 4.04it/s, grad_norm=0.422, loss_final=0.507, loss_mean=0.844, loss_mean_cls=0.0796, proj_loss=-0.417][2026-03-26 12:50:08] Step: 4596, Training Logs: loss_final: 0.489862, loss_mean: 0.828952, proj_loss: -0.417805, loss_mean_cls: 0.078715, grad_norm: 0.617648 +Steps: 0%| | 4597/1000000 [19:05<68:28:43, 4.04it/s, grad_norm=0.618, loss_final=0.49, loss_mean=0.829, loss_mean_cls=0.0787, proj_loss=-0.418][2026-03-26 12:50:08] Step: 4597, Training Logs: loss_final: 0.501204, loss_mean: 0.835070, proj_loss: -0.414047, loss_mean_cls: 0.080181, grad_norm: 0.523223 +Steps: 0%| | 4598/1000000 [19:05<68:29:12, 4.04it/s, grad_norm=0.523, loss_final=0.501, loss_mean=0.835, loss_mean_cls=0.0802, proj_loss=-0.414][2026-03-26 12:50:09] Step: 4598, Training Logs: loss_final: 0.513324, loss_mean: 0.846617, proj_loss: -0.413132, loss_mean_cls: 0.079840, grad_norm: 0.580264 +Steps: 0%| | 4599/1000000 [19:05<68:41:05, 4.03it/s, grad_norm=0.58, loss_final=0.513, loss_mean=0.847, loss_mean_cls=0.0798, proj_loss=-0.413][2026-03-26 12:50:09] Step: 4599, Training Logs: loss_final: 0.496416, loss_mean: 0.829784, proj_loss: -0.414602, loss_mean_cls: 0.081234, grad_norm: 0.587611 +Steps: 0%| | 4600/1000000 [19:06<68:37:01, 4.03it/s, grad_norm=0.588, loss_final=0.496, loss_mean=0.83, loss_mean_cls=0.0812, proj_loss=-0.415][2026-03-26 12:50:09] Step: 4600, Training Logs: loss_final: 0.504427, loss_mean: 0.843591, proj_loss: -0.417793, loss_mean_cls: 0.078629, grad_norm: 0.515712 +Steps: 0%| | 4601/1000000 [19:06<68:31:43, 4.03it/s, grad_norm=0.516, loss_final=0.504, loss_mean=0.844, loss_mean_cls=0.0786, proj_loss=-0.418][2026-03-26 12:50:09] Step: 4601, Training Logs: loss_final: 0.505807, loss_mean: 0.840925, proj_loss: -0.415296, loss_mean_cls: 0.080179, grad_norm: 0.382055 +Steps: 0%| | 4602/1000000 [19:06<68:48:12, 4.02it/s, grad_norm=0.382, loss_final=0.506, loss_mean=0.841, loss_mean_cls=0.0802, proj_loss=-0.415][2026-03-26 12:50:10] Step: 4602, Training Logs: loss_final: 0.482253, loss_mean: 0.802761, proj_loss: -0.405374, loss_mean_cls: 0.084866, grad_norm: 0.364666 +Steps: 0%| | 4603/1000000 [19:06<68:41:54, 4.02it/s, grad_norm=0.365, loss_final=0.482, loss_mean=0.803, loss_mean_cls=0.0849, proj_loss=-0.405][2026-03-26 12:50:10] Step: 4603, Training Logs: loss_final: 0.486416, loss_mean: 0.810541, proj_loss: -0.407232, loss_mean_cls: 0.083107, grad_norm: 0.475575 +Steps: 0%| | 4604/1000000 [19:07<68:36:30, 4.03it/s, grad_norm=0.476, loss_final=0.486, loss_mean=0.811, loss_mean_cls=0.0831, proj_loss=-0.407][2026-03-26 12:50:10] Step: 4604, Training Logs: loss_final: 0.502893, loss_mean: 0.824377, proj_loss: -0.404438, loss_mean_cls: 0.082954, grad_norm: 0.430905 +Steps: 0%| | 4605/1000000 [19:07<68:45:34, 4.02it/s, grad_norm=0.431, loss_final=0.503, loss_mean=0.824, loss_mean_cls=0.083, proj_loss=-0.404][2026-03-26 12:50:10] Step: 4605, Training Logs: loss_final: 0.504017, loss_mean: 0.833960, proj_loss: -0.410085, loss_mean_cls: 0.080142, grad_norm: 0.417004 +Steps: 0%| | 4606/1000000 [19:07<68:39:11, 4.03it/s, grad_norm=0.417, loss_final=0.504, loss_mean=0.834, loss_mean_cls=0.0801, proj_loss=-0.41][2026-03-26 12:50:11] Step: 4606, Training Logs: loss_final: 0.526423, loss_mean: 0.856762, proj_loss: -0.409605, loss_mean_cls: 0.079266, grad_norm: 0.388397 +Steps: 0%| | 4607/1000000 [19:07<68:35:19, 4.03it/s, grad_norm=0.388, loss_final=0.526, loss_mean=0.857, loss_mean_cls=0.0793, proj_loss=-0.41][2026-03-26 12:50:11] Step: 4607, Training Logs: loss_final: 0.532639, loss_mean: 0.860485, proj_loss: -0.407311, loss_mean_cls: 0.079464, grad_norm: 0.496224 +Steps: 0%| | 4608/1000000 [19:08<68:49:10, 4.02it/s, grad_norm=0.496, loss_final=0.533, loss_mean=0.86, loss_mean_cls=0.0795, proj_loss=-0.407][2026-03-26 12:50:11] Step: 4608, Training Logs: loss_final: 0.495764, loss_mean: 0.835030, proj_loss: -0.417787, loss_mean_cls: 0.078521, grad_norm: 0.375887 +Steps: 0%| | 4609/1000000 [19:08<68:43:33, 4.02it/s, grad_norm=0.376, loss_final=0.496, loss_mean=0.835, loss_mean_cls=0.0785, proj_loss=-0.418][2026-03-26 12:50:11] Step: 4609, Training Logs: loss_final: 0.500131, loss_mean: 0.826132, proj_loss: -0.407003, loss_mean_cls: 0.081002, grad_norm: 0.452628 +Steps: 0%| | 4610/1000000 [19:08<68:39:57, 4.03it/s, grad_norm=0.453, loss_final=0.5, loss_mean=0.826, loss_mean_cls=0.081, proj_loss=-0.407][2026-03-26 12:50:12] Step: 4610, Training Logs: loss_final: 0.481009, loss_mean: 0.807926, proj_loss: -0.409083, loss_mean_cls: 0.082166, grad_norm: 0.451041 +Steps: 0%| | 4611/1000000 [19:08<68:36:58, 4.03it/s, grad_norm=0.451, loss_final=0.481, loss_mean=0.808, loss_mean_cls=0.0822, proj_loss=-0.409][2026-03-26 12:50:12] Step: 4611, Training Logs: loss_final: 0.510339, loss_mean: 0.844421, proj_loss: -0.412182, loss_mean_cls: 0.078100, grad_norm: 0.371816 +Steps: 0%| | 4612/1000000 [19:09<68:34:40, 4.03it/s, grad_norm=0.372, loss_final=0.51, loss_mean=0.844, loss_mean_cls=0.0781, proj_loss=-0.412][2026-03-26 12:50:12] Step: 4612, Training Logs: loss_final: 0.493720, loss_mean: 0.820584, proj_loss: -0.408601, loss_mean_cls: 0.081738, grad_norm: 0.377145 +Steps: 0%| | 4613/1000000 [19:09<68:47:19, 4.02it/s, grad_norm=0.377, loss_final=0.494, loss_mean=0.821, loss_mean_cls=0.0817, proj_loss=-0.409][2026-03-26 12:50:12] Step: 4613, Training Logs: loss_final: 0.511882, loss_mean: 0.854329, proj_loss: -0.419779, loss_mean_cls: 0.077333, grad_norm: 0.341281 +Steps: 0%| | 4614/1000000 [19:09<68:39:25, 4.03it/s, grad_norm=0.341, loss_final=0.512, loss_mean=0.854, loss_mean_cls=0.0773, proj_loss=-0.42][2026-03-26 12:50:13] Step: 4614, Training Logs: loss_final: 0.510002, loss_mean: 0.846770, proj_loss: -0.417045, loss_mean_cls: 0.080276, grad_norm: 0.215304 +Steps: 0%| | 4615/1000000 [19:09<68:34:20, 4.03it/s, grad_norm=0.215, loss_final=0.51, loss_mean=0.847, loss_mean_cls=0.0803, proj_loss=-0.417][2026-03-26 12:50:13] Step: 4615, Training Logs: loss_final: 0.497671, loss_mean: 0.818793, proj_loss: -0.403964, loss_mean_cls: 0.082842, grad_norm: 0.436101 +Steps: 0%| | 4616/1000000 [19:10<68:32:21, 4.03it/s, grad_norm=0.436, loss_final=0.498, loss_mean=0.819, loss_mean_cls=0.0828, proj_loss=-0.404][2026-03-26 12:50:13] Step: 4616, Training Logs: loss_final: 0.502069, loss_mean: 0.828362, proj_loss: -0.407893, loss_mean_cls: 0.081599, grad_norm: 0.350119 +Steps: 0%| | 4617/1000000 [19:10<68:29:28, 4.04it/s, grad_norm=0.35, loss_final=0.502, loss_mean=0.828, loss_mean_cls=0.0816, proj_loss=-0.408][2026-03-26 12:50:13] Step: 4617, Training Logs: loss_final: 0.503875, loss_mean: 0.830499, proj_loss: -0.407806, loss_mean_cls: 0.081182, grad_norm: 0.349936 +Steps: 0%| | 4618/1000000 [19:10<68:27:05, 4.04it/s, grad_norm=0.35, loss_final=0.504, loss_mean=0.83, loss_mean_cls=0.0812, proj_loss=-0.408][2026-03-26 12:50:14] Step: 4618, Training Logs: loss_final: 0.495260, loss_mean: 0.828994, proj_loss: -0.413275, loss_mean_cls: 0.079541, grad_norm: 0.470268 +Steps: 0%| | 4619/1000000 [19:10<68:27:05, 4.04it/s, grad_norm=0.47, loss_final=0.495, loss_mean=0.829, loss_mean_cls=0.0795, proj_loss=-0.413][2026-03-26 12:50:14] Step: 4619, Training Logs: loss_final: 0.515237, loss_mean: 0.849213, proj_loss: -0.412415, loss_mean_cls: 0.078439, grad_norm: 0.433159 +Steps: 0%| | 4620/1000000 [19:11<68:47:05, 4.02it/s, grad_norm=0.433, loss_final=0.515, loss_mean=0.849, loss_mean_cls=0.0784, proj_loss=-0.412][2026-03-26 12:50:14] Step: 4620, Training Logs: loss_final: 0.508692, loss_mean: 0.844858, proj_loss: -0.414541, loss_mean_cls: 0.078375, grad_norm: 0.277928 +Steps: 0%| | 4621/1000000 [19:11<68:39:04, 4.03it/s, grad_norm=0.278, loss_final=0.509, loss_mean=0.845, loss_mean_cls=0.0784, proj_loss=-0.415][2026-03-26 12:50:14] Step: 4621, Training Logs: loss_final: 0.510454, loss_mean: 0.842474, proj_loss: -0.410073, loss_mean_cls: 0.078053, grad_norm: 0.543977 +Steps: 0%| | 4622/1000000 [19:11<68:42:19, 4.02it/s, grad_norm=0.544, loss_final=0.51, loss_mean=0.842, loss_mean_cls=0.0781, proj_loss=-0.41][2026-03-26 12:50:15] Step: 4622, Training Logs: loss_final: 0.493092, loss_mean: 0.831235, proj_loss: -0.416849, loss_mean_cls: 0.078706, grad_norm: 0.513774 +Steps: 0%| | 4623/1000000 [19:11<68:36:01, 4.03it/s, grad_norm=0.514, loss_final=0.493, loss_mean=0.831, loss_mean_cls=0.0787, proj_loss=-0.417][2026-03-26 12:50:15] Step: 4623, Training Logs: loss_final: 0.488933, loss_mean: 0.820821, proj_loss: -0.411981, loss_mean_cls: 0.080092, grad_norm: 0.256448 +Steps: 0%| | 4624/1000000 [19:12<68:38:06, 4.03it/s, grad_norm=0.256, loss_final=0.489, loss_mean=0.821, loss_mean_cls=0.0801, proj_loss=-0.412][2026-03-26 12:50:15] Step: 4624, Training Logs: loss_final: 0.517113, loss_mean: 0.856066, proj_loss: -0.416086, loss_mean_cls: 0.077133, grad_norm: 0.258474 +Steps: 0%| | 4625/1000000 [19:12<68:36:46, 4.03it/s, grad_norm=0.258, loss_final=0.517, loss_mean=0.856, loss_mean_cls=0.0771, proj_loss=-0.416][2026-03-26 12:50:15] Step: 4625, Training Logs: loss_final: 0.491483, loss_mean: 0.820159, proj_loss: -0.408929, loss_mean_cls: 0.080253, grad_norm: 0.321759 +Steps: 0%| | 4626/1000000 [19:12<68:31:55, 4.03it/s, grad_norm=0.322, loss_final=0.491, loss_mean=0.82, loss_mean_cls=0.0803, proj_loss=-0.409][2026-03-26 12:50:16] Step: 4626, Training Logs: loss_final: 0.499503, loss_mean: 0.836136, proj_loss: -0.417653, loss_mean_cls: 0.081020, grad_norm: 0.214693 +Steps: 0%| | 4627/1000000 [19:12<68:31:21, 4.04it/s, grad_norm=0.215, loss_final=0.5, loss_mean=0.836, loss_mean_cls=0.081, proj_loss=-0.418][2026-03-26 12:50:16] Step: 4627, Training Logs: loss_final: 0.478599, loss_mean: 0.812861, proj_loss: -0.415163, loss_mean_cls: 0.080901, grad_norm: 0.354953 +Steps: 0%| | 4628/1000000 [19:13<68:27:48, 4.04it/s, grad_norm=0.355, loss_final=0.479, loss_mean=0.813, loss_mean_cls=0.0809, proj_loss=-0.415][2026-03-26 12:50:16] Step: 4628, Training Logs: loss_final: 0.507263, loss_mean: 0.845570, proj_loss: -0.418450, loss_mean_cls: 0.080143, grad_norm: 0.400021 +Steps: 0%| | 4629/1000000 [19:13<68:27:20, 4.04it/s, grad_norm=0.4, loss_final=0.507, loss_mean=0.846, loss_mean_cls=0.0801, proj_loss=-0.418][2026-03-26 12:50:16] Step: 4629, Training Logs: loss_final: 0.487343, loss_mean: 0.811916, proj_loss: -0.406675, loss_mean_cls: 0.082102, grad_norm: 0.357316 +Steps: 0%| | 4630/1000000 [19:13<68:24:18, 4.04it/s, grad_norm=0.357, loss_final=0.487, loss_mean=0.812, loss_mean_cls=0.0821, proj_loss=-0.407][2026-03-26 12:50:17] Step: 4630, Training Logs: loss_final: 0.505442, loss_mean: 0.844088, proj_loss: -0.417642, loss_mean_cls: 0.078996, grad_norm: 0.241680 +Steps: 0%| | 4631/1000000 [19:13<68:23:26, 4.04it/s, grad_norm=0.242, loss_final=0.505, loss_mean=0.844, loss_mean_cls=0.079, proj_loss=-0.418][2026-03-26 12:50:17] Step: 4631, Training Logs: loss_final: 0.485917, loss_mean: 0.811164, proj_loss: -0.406639, loss_mean_cls: 0.081392, grad_norm: 0.329262 +Steps: 0%| | 4632/1000000 [19:14<68:25:17, 4.04it/s, grad_norm=0.329, loss_final=0.486, loss_mean=0.811, loss_mean_cls=0.0814, proj_loss=-0.407][2026-03-26 12:50:17] Step: 4632, Training Logs: loss_final: 0.502487, loss_mean: 0.833311, proj_loss: -0.411047, loss_mean_cls: 0.080222, grad_norm: 0.270960 +Steps: 0%| | 4633/1000000 [19:14<68:22:20, 4.04it/s, grad_norm=0.271, loss_final=0.502, loss_mean=0.833, loss_mean_cls=0.0802, proj_loss=-0.411][2026-03-26 12:50:17] Step: 4633, Training Logs: loss_final: 0.501044, loss_mean: 0.828977, proj_loss: -0.409235, loss_mean_cls: 0.081302, grad_norm: 0.279529 +Steps: 0%| | 4634/1000000 [19:14<68:23:00, 4.04it/s, grad_norm=0.28, loss_final=0.501, loss_mean=0.829, loss_mean_cls=0.0813, proj_loss=-0.409][2026-03-26 12:50:18] Step: 4634, Training Logs: loss_final: 0.491573, loss_mean: 0.827823, proj_loss: -0.415193, loss_mean_cls: 0.078942, grad_norm: 0.263698 +Steps: 0%| | 4635/1000000 [19:14<68:25:34, 4.04it/s, grad_norm=0.264, loss_final=0.492, loss_mean=0.828, loss_mean_cls=0.0789, proj_loss=-0.415][2026-03-26 12:50:18] Step: 4635, Training Logs: loss_final: 0.499224, loss_mean: 0.822040, proj_loss: -0.405328, loss_mean_cls: 0.082512, grad_norm: 0.469560 +Steps: 0%| | 4636/1000000 [19:15<68:26:58, 4.04it/s, grad_norm=0.47, loss_final=0.499, loss_mean=0.822, loss_mean_cls=0.0825, proj_loss=-0.405][2026-03-26 12:50:18] Step: 4636, Training Logs: loss_final: 0.518512, loss_mean: 0.848976, proj_loss: -0.410097, loss_mean_cls: 0.079633, grad_norm: 0.359217 +Steps: 0%| | 4637/1000000 [19:15<68:25:32, 4.04it/s, grad_norm=0.359, loss_final=0.519, loss_mean=0.849, loss_mean_cls=0.0796, proj_loss=-0.41][2026-03-26 12:50:18] Step: 4637, Training Logs: loss_final: 0.495461, loss_mean: 0.836064, proj_loss: -0.419539, loss_mean_cls: 0.078936, grad_norm: 0.514781 +Steps: 0%| | 4638/1000000 [19:15<68:26:13, 4.04it/s, grad_norm=0.515, loss_final=0.495, loss_mean=0.836, loss_mean_cls=0.0789, proj_loss=-0.42][2026-03-26 12:50:19] Step: 4638, Training Logs: loss_final: 0.519660, loss_mean: 0.853193, proj_loss: -0.412827, loss_mean_cls: 0.079294, grad_norm: 0.425515 +Steps: 0%| | 4639/1000000 [19:15<68:23:19, 4.04it/s, grad_norm=0.426, loss_final=0.52, loss_mean=0.853, loss_mean_cls=0.0793, proj_loss=-0.413][2026-03-26 12:50:19] Step: 4639, Training Logs: loss_final: 0.505472, loss_mean: 0.837393, proj_loss: -0.413621, loss_mean_cls: 0.081700, grad_norm: 0.572894 +Steps: 0%| | 4640/1000000 [19:16<68:23:24, 4.04it/s, grad_norm=0.573, loss_final=0.505, loss_mean=0.837, loss_mean_cls=0.0817, proj_loss=-0.414][2026-03-26 12:50:19] Step: 4640, Training Logs: loss_final: 0.498330, loss_mean: 0.825996, proj_loss: -0.408203, loss_mean_cls: 0.080537, grad_norm: 0.397368 +Steps: 0%| | 4641/1000000 [19:16<68:20:06, 4.05it/s, grad_norm=0.397, loss_final=0.498, loss_mean=0.826, loss_mean_cls=0.0805, proj_loss=-0.408][2026-03-26 12:50:19] Step: 4641, Training Logs: loss_final: 0.505493, loss_mean: 0.831705, proj_loss: -0.407668, loss_mean_cls: 0.081457, grad_norm: 0.559226 +Steps: 0%| | 4642/1000000 [19:16<68:23:22, 4.04it/s, grad_norm=0.559, loss_final=0.505, loss_mean=0.832, loss_mean_cls=0.0815, proj_loss=-0.408][2026-03-26 12:50:20] Step: 4642, Training Logs: loss_final: 0.491523, loss_mean: 0.825848, proj_loss: -0.414680, loss_mean_cls: 0.080355, grad_norm: 0.505866 +Steps: 0%| | 4643/1000000 [19:16<68:22:50, 4.04it/s, grad_norm=0.506, loss_final=0.492, loss_mean=0.826, loss_mean_cls=0.0804, proj_loss=-0.415][2026-03-26 12:50:20] Step: 4643, Training Logs: loss_final: 0.499529, loss_mean: 0.826815, proj_loss: -0.406123, loss_mean_cls: 0.078838, grad_norm: 0.476263 +Steps: 0%| | 4644/1000000 [19:17<68:29:21, 4.04it/s, grad_norm=0.476, loss_final=0.5, loss_mean=0.827, loss_mean_cls=0.0788, proj_loss=-0.406][2026-03-26 12:50:20] Step: 4644, Training Logs: loss_final: 0.489009, loss_mean: 0.819164, proj_loss: -0.411704, loss_mean_cls: 0.081549, grad_norm: 0.442308 +Steps: 0%| | 4645/1000000 [19:17<68:24:16, 4.04it/s, grad_norm=0.442, loss_final=0.489, loss_mean=0.819, loss_mean_cls=0.0815, proj_loss=-0.412][2026-03-26 12:50:20] Step: 4645, Training Logs: loss_final: 0.483846, loss_mean: 0.823083, proj_loss: -0.419293, loss_mean_cls: 0.080055, grad_norm: 0.405561 +Steps: 0%| | 4646/1000000 [19:17<68:24:50, 4.04it/s, grad_norm=0.406, loss_final=0.484, loss_mean=0.823, loss_mean_cls=0.0801, proj_loss=-0.419][2026-03-26 12:50:21] Step: 4646, Training Logs: loss_final: 0.492746, loss_mean: 0.818615, proj_loss: -0.406940, loss_mean_cls: 0.081071, grad_norm: 0.391741 +Steps: 0%| | 4647/1000000 [19:17<68:23:28, 4.04it/s, grad_norm=0.392, loss_final=0.493, loss_mean=0.819, loss_mean_cls=0.0811, proj_loss=-0.407][2026-03-26 12:50:21] Step: 4647, Training Logs: loss_final: 0.499439, loss_mean: 0.836093, proj_loss: -0.417420, loss_mean_cls: 0.080766, grad_norm: 0.344089 +Steps: 0%| | 4648/1000000 [19:18<68:24:11, 4.04it/s, grad_norm=0.344, loss_final=0.499, loss_mean=0.836, loss_mean_cls=0.0808, proj_loss=-0.417][2026-03-26 12:50:21] Step: 4648, Training Logs: loss_final: 0.487826, loss_mean: 0.827954, proj_loss: -0.419331, loss_mean_cls: 0.079203, grad_norm: 0.306740 +Steps: 0%| | 4649/1000000 [19:18<68:27:35, 4.04it/s, grad_norm=0.307, loss_final=0.488, loss_mean=0.828, loss_mean_cls=0.0792, proj_loss=-0.419][2026-03-26 12:50:21] Step: 4649, Training Logs: loss_final: 0.489590, loss_mean: 0.825454, proj_loss: -0.415890, loss_mean_cls: 0.080027, grad_norm: 0.305992 +Steps: 0%| | 4650/1000000 [19:18<68:26:57, 4.04it/s, grad_norm=0.306, loss_final=0.49, loss_mean=0.825, loss_mean_cls=0.08, proj_loss=-0.416][2026-03-26 12:50:22] Step: 4650, Training Logs: loss_final: 0.502923, loss_mean: 0.847021, proj_loss: -0.420476, loss_mean_cls: 0.076378, grad_norm: 0.242240 +Steps: 0%| | 4651/1000000 [19:18<68:31:14, 4.04it/s, grad_norm=0.242, loss_final=0.503, loss_mean=0.847, loss_mean_cls=0.0764, proj_loss=-0.42][2026-03-26 12:50:22] Step: 4651, Training Logs: loss_final: 0.493882, loss_mean: 0.827398, proj_loss: -0.414672, loss_mean_cls: 0.081156, grad_norm: 0.300463 +Steps: 0%| | 4652/1000000 [19:19<68:32:14, 4.03it/s, grad_norm=0.3, loss_final=0.494, loss_mean=0.827, loss_mean_cls=0.0812, proj_loss=-0.415][2026-03-26 12:50:22] Step: 4652, Training Logs: loss_final: 0.503220, loss_mean: 0.824928, proj_loss: -0.405071, loss_mean_cls: 0.083362, grad_norm: 0.276211 +Steps: 0%| | 4653/1000000 [19:19<68:32:10, 4.03it/s, grad_norm=0.276, loss_final=0.503, loss_mean=0.825, loss_mean_cls=0.0834, proj_loss=-0.405][2026-03-26 12:50:22] Step: 4653, Training Logs: loss_final: 0.494648, loss_mean: 0.822991, proj_loss: -0.410878, loss_mean_cls: 0.082536, grad_norm: 0.408671 +Steps: 0%| | 4654/1000000 [19:19<68:28:06, 4.04it/s, grad_norm=0.409, loss_final=0.495, loss_mean=0.823, loss_mean_cls=0.0825, proj_loss=-0.411][2026-03-26 12:50:23] Step: 4654, Training Logs: loss_final: 0.499673, loss_mean: 0.826788, proj_loss: -0.408749, loss_mean_cls: 0.081633, grad_norm: 0.323109 +Steps: 0%| | 4655/1000000 [19:19<68:26:00, 4.04it/s, grad_norm=0.323, loss_final=0.5, loss_mean=0.827, loss_mean_cls=0.0816, proj_loss=-0.409][2026-03-26 12:50:23] Step: 4655, Training Logs: loss_final: 0.502752, loss_mean: 0.839499, proj_loss: -0.415096, loss_mean_cls: 0.078350, grad_norm: 0.413570 +Steps: 0%| | 4656/1000000 [19:20<68:27:56, 4.04it/s, grad_norm=0.414, loss_final=0.503, loss_mean=0.839, loss_mean_cls=0.0783, proj_loss=-0.415][2026-03-26 12:50:23] Step: 4656, Training Logs: loss_final: 0.507691, loss_mean: 0.830708, proj_loss: -0.404282, loss_mean_cls: 0.081265, grad_norm: 0.401099 +Steps: 0%| | 4657/1000000 [19:20<68:28:44, 4.04it/s, grad_norm=0.401, loss_final=0.508, loss_mean=0.831, loss_mean_cls=0.0813, proj_loss=-0.404][2026-03-26 12:50:23] Step: 4657, Training Logs: loss_final: 0.518027, loss_mean: 0.854877, proj_loss: -0.414814, loss_mean_cls: 0.077963, grad_norm: 0.232601 +Steps: 0%| | 4658/1000000 [19:20<68:26:54, 4.04it/s, grad_norm=0.233, loss_final=0.518, loss_mean=0.855, loss_mean_cls=0.078, proj_loss=-0.415][2026-03-26 12:50:24] Step: 4658, Training Logs: loss_final: 0.510401, loss_mean: 0.846572, proj_loss: -0.414625, loss_mean_cls: 0.078454, grad_norm: 0.542328 +Steps: 0%| | 4659/1000000 [19:20<68:22:45, 4.04it/s, grad_norm=0.542, loss_final=0.51, loss_mean=0.847, loss_mean_cls=0.0785, proj_loss=-0.415][2026-03-26 12:50:24] Step: 4659, Training Logs: loss_final: 0.517480, loss_mean: 0.847932, proj_loss: -0.409946, loss_mean_cls: 0.079494, grad_norm: 0.563628 +Steps: 0%| | 4660/1000000 [19:21<68:24:35, 4.04it/s, grad_norm=0.564, loss_final=0.517, loss_mean=0.848, loss_mean_cls=0.0795, proj_loss=-0.41][2026-03-26 12:50:24] Step: 4660, Training Logs: loss_final: 0.486627, loss_mean: 0.815980, proj_loss: -0.410144, loss_mean_cls: 0.080791, grad_norm: 0.297749 +Steps: 0%| | 4661/1000000 [19:21<68:24:01, 4.04it/s, grad_norm=0.298, loss_final=0.487, loss_mean=0.816, loss_mean_cls=0.0808, proj_loss=-0.41][2026-03-26 12:50:24] Step: 4661, Training Logs: loss_final: 0.504744, loss_mean: 0.831528, proj_loss: -0.408786, loss_mean_cls: 0.082001, grad_norm: 0.361264 +Steps: 0%| | 4662/1000000 [19:21<68:20:13, 4.05it/s, grad_norm=0.361, loss_final=0.505, loss_mean=0.832, loss_mean_cls=0.082, proj_loss=-0.409][2026-03-26 12:50:24] Step: 4662, Training Logs: loss_final: 0.483197, loss_mean: 0.824847, proj_loss: -0.420489, loss_mean_cls: 0.078838, grad_norm: 0.249570 +Steps: 0%| | 4663/1000000 [19:21<68:19:00, 4.05it/s, grad_norm=0.25, loss_final=0.483, loss_mean=0.825, loss_mean_cls=0.0788, proj_loss=-0.42][2026-03-26 12:50:25] Step: 4663, Training Logs: loss_final: 0.489452, loss_mean: 0.826107, proj_loss: -0.415961, loss_mean_cls: 0.079306, grad_norm: 0.373364 +Steps: 0%| | 4664/1000000 [19:22<68:18:27, 4.05it/s, grad_norm=0.373, loss_final=0.489, loss_mean=0.826, loss_mean_cls=0.0793, proj_loss=-0.416][2026-03-26 12:50:25] Step: 4664, Training Logs: loss_final: 0.500884, loss_mean: 0.834771, proj_loss: -0.412980, loss_mean_cls: 0.079094, grad_norm: 0.413617 +Steps: 0%| | 4665/1000000 [19:22<68:21:36, 4.04it/s, grad_norm=0.414, loss_final=0.501, loss_mean=0.835, loss_mean_cls=0.0791, proj_loss=-0.413][2026-03-26 12:50:25] Step: 4665, Training Logs: loss_final: 0.491658, loss_mean: 0.822522, proj_loss: -0.412512, loss_mean_cls: 0.081648, grad_norm: 0.274155 +Steps: 0%| | 4666/1000000 [19:22<68:22:40, 4.04it/s, grad_norm=0.274, loss_final=0.492, loss_mean=0.823, loss_mean_cls=0.0816, proj_loss=-0.413][2026-03-26 12:50:25] Step: 4666, Training Logs: loss_final: 0.504558, loss_mean: 0.832186, proj_loss: -0.408187, loss_mean_cls: 0.080558, grad_norm: 0.738859 +Steps: 0%| | 4667/1000000 [19:22<68:25:10, 4.04it/s, grad_norm=0.739, loss_final=0.505, loss_mean=0.832, loss_mean_cls=0.0806, proj_loss=-0.408][2026-03-26 12:50:26] Step: 4667, Training Logs: loss_final: 0.511286, loss_mean: 0.843485, proj_loss: -0.411524, loss_mean_cls: 0.079325, grad_norm: 0.646341 +Steps: 0%| | 4668/1000000 [19:23<68:24:59, 4.04it/s, grad_norm=0.646, loss_final=0.511, loss_mean=0.843, loss_mean_cls=0.0793, proj_loss=-0.412][2026-03-26 12:50:26] Step: 4668, Training Logs: loss_final: 0.485015, loss_mean: 0.814265, proj_loss: -0.410215, loss_mean_cls: 0.080964, grad_norm: 0.250340 +Steps: 0%| | 4669/1000000 [19:23<68:24:57, 4.04it/s, grad_norm=0.25, loss_final=0.485, loss_mean=0.814, loss_mean_cls=0.081, proj_loss=-0.41][2026-03-26 12:50:26] Step: 4669, Training Logs: loss_final: 0.507367, loss_mean: 0.841607, proj_loss: -0.411241, loss_mean_cls: 0.077001, grad_norm: 0.700855 +Steps: 0%| | 4670/1000000 [19:23<69:57:17, 3.95it/s, grad_norm=0.701, loss_final=0.507, loss_mean=0.842, loss_mean_cls=0.077, proj_loss=-0.411][2026-03-26 12:50:26] Step: 4670, Training Logs: loss_final: 0.489167, loss_mean: 0.822333, proj_loss: -0.413472, loss_mean_cls: 0.080306, grad_norm: 0.522147 +Steps: 0%| | 4671/1000000 [19:23<69:45:38, 3.96it/s, grad_norm=0.522, loss_final=0.489, loss_mean=0.822, loss_mean_cls=0.0803, proj_loss=-0.413][2026-03-26 12:50:27] Step: 4671, Training Logs: loss_final: 0.516545, loss_mean: 0.854329, proj_loss: -0.415529, loss_mean_cls: 0.077745, grad_norm: 0.338850 +Steps: 0%| | 4672/1000000 [19:24<69:20:56, 3.99it/s, grad_norm=0.339, loss_final=0.517, loss_mean=0.854, loss_mean_cls=0.0777, proj_loss=-0.416][2026-03-26 12:50:27] Step: 4672, Training Logs: loss_final: 0.498045, loss_mean: 0.824636, proj_loss: -0.406841, loss_mean_cls: 0.080250, grad_norm: 0.670397 +Steps: 0%| | 4673/1000000 [19:24<69:03:38, 4.00it/s, grad_norm=0.67, loss_final=0.498, loss_mean=0.825, loss_mean_cls=0.0802, proj_loss=-0.407][2026-03-26 12:50:27] Step: 4673, Training Logs: loss_final: 0.497575, loss_mean: 0.819538, proj_loss: -0.404507, loss_mean_cls: 0.082544, grad_norm: 0.314803 +Steps: 0%| | 4674/1000000 [19:24<68:50:55, 4.02it/s, grad_norm=0.315, loss_final=0.498, loss_mean=0.82, loss_mean_cls=0.0825, proj_loss=-0.405][2026-03-26 12:50:27] Step: 4674, Training Logs: loss_final: 0.506647, loss_mean: 0.821784, proj_loss: -0.397779, loss_mean_cls: 0.082642, grad_norm: 0.525377 +Steps: 0%| | 4675/1000000 [19:24<68:40:37, 4.03it/s, grad_norm=0.525, loss_final=0.507, loss_mean=0.822, loss_mean_cls=0.0826, proj_loss=-0.398][2026-03-26 12:50:28] Step: 4675, Training Logs: loss_final: 0.509261, loss_mean: 0.845986, proj_loss: -0.415853, loss_mean_cls: 0.079128, grad_norm: 0.433965 +Steps: 0%| | 4676/1000000 [19:25<68:37:13, 4.03it/s, grad_norm=0.434, loss_final=0.509, loss_mean=0.846, loss_mean_cls=0.0791, proj_loss=-0.416][2026-03-26 12:50:28] Step: 4676, Training Logs: loss_final: 0.500641, loss_mean: 0.819425, proj_loss: -0.403870, loss_mean_cls: 0.085085, grad_norm: 0.429689 +Steps: 0%| | 4677/1000000 [19:25<68:32:18, 4.03it/s, grad_norm=0.43, loss_final=0.501, loss_mean=0.819, loss_mean_cls=0.0851, proj_loss=-0.404][2026-03-26 12:50:28] Step: 4677, Training Logs: loss_final: 0.504071, loss_mean: 0.835146, proj_loss: -0.410187, loss_mean_cls: 0.079112, grad_norm: 0.385557 +Steps: 0%| | 4678/1000000 [19:25<68:28:42, 4.04it/s, grad_norm=0.386, loss_final=0.504, loss_mean=0.835, loss_mean_cls=0.0791, proj_loss=-0.41][2026-03-26 12:50:28] Step: 4678, Training Logs: loss_final: 0.507732, loss_mean: 0.840390, proj_loss: -0.410156, loss_mean_cls: 0.077498, grad_norm: 0.424199 +Steps: 0%| | 4679/1000000 [19:25<68:28:33, 4.04it/s, grad_norm=0.424, loss_final=0.508, loss_mean=0.84, loss_mean_cls=0.0775, proj_loss=-0.41][2026-03-26 12:50:29] Step: 4679, Training Logs: loss_final: 0.504464, loss_mean: 0.837205, proj_loss: -0.413206, loss_mean_cls: 0.080464, grad_norm: 0.509618 +Steps: 0%| | 4680/1000000 [19:26<68:29:29, 4.04it/s, grad_norm=0.51, loss_final=0.504, loss_mean=0.837, loss_mean_cls=0.0805, proj_loss=-0.413][2026-03-26 12:50:29] Step: 4680, Training Logs: loss_final: 0.498899, loss_mean: 0.839907, proj_loss: -0.419342, loss_mean_cls: 0.078334, grad_norm: 0.421770 +Steps: 0%| | 4681/1000000 [19:26<68:26:47, 4.04it/s, grad_norm=0.422, loss_final=0.499, loss_mean=0.84, loss_mean_cls=0.0783, proj_loss=-0.419][2026-03-26 12:50:29] Step: 4681, Training Logs: loss_final: 0.482930, loss_mean: 0.819250, proj_loss: -0.416875, loss_mean_cls: 0.080554, grad_norm: 0.347773 +Steps: 0%| | 4682/1000000 [19:26<68:28:00, 4.04it/s, grad_norm=0.348, loss_final=0.483, loss_mean=0.819, loss_mean_cls=0.0806, proj_loss=-0.417][2026-03-26 12:50:29] Step: 4682, Training Logs: loss_final: 0.496171, loss_mean: 0.836498, proj_loss: -0.420570, loss_mean_cls: 0.080243, grad_norm: 0.424282 +Steps: 0%| | 4683/1000000 [19:26<68:26:06, 4.04it/s, grad_norm=0.424, loss_final=0.496, loss_mean=0.836, loss_mean_cls=0.0802, proj_loss=-0.421][2026-03-26 12:50:30] Step: 4683, Training Logs: loss_final: 0.472818, loss_mean: 0.798500, proj_loss: -0.409863, loss_mean_cls: 0.084182, grad_norm: 0.341063 +Steps: 0%| | 4684/1000000 [19:27<68:24:53, 4.04it/s, grad_norm=0.341, loss_final=0.473, loss_mean=0.798, loss_mean_cls=0.0842, proj_loss=-0.41][2026-03-26 12:50:30] Step: 4684, Training Logs: loss_final: 0.500317, loss_mean: 0.834584, proj_loss: -0.412711, loss_mean_cls: 0.078444, grad_norm: 0.377734 +Steps: 0%| | 4685/1000000 [19:27<68:24:46, 4.04it/s, grad_norm=0.378, loss_final=0.5, loss_mean=0.835, loss_mean_cls=0.0784, proj_loss=-0.413][2026-03-26 12:50:30] Step: 4685, Training Logs: loss_final: 0.490898, loss_mean: 0.834153, proj_loss: -0.421621, loss_mean_cls: 0.078366, grad_norm: 0.454661 +Steps: 0%| | 4686/1000000 [19:27<68:23:59, 4.04it/s, grad_norm=0.455, loss_final=0.491, loss_mean=0.834, loss_mean_cls=0.0784, proj_loss=-0.422][2026-03-26 12:50:30] Step: 4686, Training Logs: loss_final: 0.493720, loss_mean: 0.830406, proj_loss: -0.417794, loss_mean_cls: 0.081107, grad_norm: 0.331210 +Steps: 0%| | 4687/1000000 [19:27<68:23:25, 4.04it/s, grad_norm=0.331, loss_final=0.494, loss_mean=0.83, loss_mean_cls=0.0811, proj_loss=-0.418][2026-03-26 12:50:31] Step: 4687, Training Logs: loss_final: 0.494766, loss_mean: 0.838933, proj_loss: -0.420743, loss_mean_cls: 0.076576, grad_norm: 0.465328 +Steps: 0%| | 4688/1000000 [19:28<68:25:12, 4.04it/s, grad_norm=0.465, loss_final=0.495, loss_mean=0.839, loss_mean_cls=0.0766, proj_loss=-0.421][2026-03-26 12:50:31] Step: 4688, Training Logs: loss_final: 0.481763, loss_mean: 0.818525, proj_loss: -0.416383, loss_mean_cls: 0.079622, grad_norm: 0.322745 +Steps: 0%| | 4689/1000000 [19:28<68:23:21, 4.04it/s, grad_norm=0.323, loss_final=0.482, loss_mean=0.819, loss_mean_cls=0.0796, proj_loss=-0.416][2026-03-26 12:50:31] Step: 4689, Training Logs: loss_final: 0.495018, loss_mean: 0.832578, proj_loss: -0.415701, loss_mean_cls: 0.078141, grad_norm: 0.344880 +Steps: 0%| | 4690/1000000 [19:28<68:24:41, 4.04it/s, grad_norm=0.345, loss_final=0.495, loss_mean=0.833, loss_mean_cls=0.0781, proj_loss=-0.416][2026-03-26 12:50:31] Step: 4690, Training Logs: loss_final: 0.489750, loss_mean: 0.822694, proj_loss: -0.414194, loss_mean_cls: 0.081250, grad_norm: 0.371930 +Steps: 0%| | 4691/1000000 [19:28<68:24:31, 4.04it/s, grad_norm=0.372, loss_final=0.49, loss_mean=0.823, loss_mean_cls=0.0812, proj_loss=-0.414][2026-03-26 12:50:32] Step: 4691, Training Logs: loss_final: 0.495718, loss_mean: 0.831491, proj_loss: -0.414094, loss_mean_cls: 0.078321, grad_norm: 0.304491 +Steps: 0%| | 4692/1000000 [19:29<68:27:07, 4.04it/s, grad_norm=0.304, loss_final=0.496, loss_mean=0.831, loss_mean_cls=0.0783, proj_loss=-0.414][2026-03-26 12:50:32] Step: 4692, Training Logs: loss_final: 0.498010, loss_mean: 0.833103, proj_loss: -0.414319, loss_mean_cls: 0.079226, grad_norm: 0.565633 +Steps: 0%| | 4693/1000000 [19:29<68:25:33, 4.04it/s, grad_norm=0.566, loss_final=0.498, loss_mean=0.833, loss_mean_cls=0.0792, proj_loss=-0.414][2026-03-26 12:50:32] Step: 4693, Training Logs: loss_final: 0.506396, loss_mean: 0.841019, proj_loss: -0.413807, loss_mean_cls: 0.079184, grad_norm: 0.318803 +Steps: 0%| | 4694/1000000 [19:29<68:26:38, 4.04it/s, grad_norm=0.319, loss_final=0.506, loss_mean=0.841, loss_mean_cls=0.0792, proj_loss=-0.414][2026-03-26 12:50:32] Step: 4694, Training Logs: loss_final: 0.500828, loss_mean: 0.832718, proj_loss: -0.410913, loss_mean_cls: 0.079022, grad_norm: 0.494264 +Steps: 0%| | 4695/1000000 [19:29<68:24:29, 4.04it/s, grad_norm=0.494, loss_final=0.501, loss_mean=0.833, loss_mean_cls=0.079, proj_loss=-0.411][2026-03-26 12:50:33] Step: 4695, Training Logs: loss_final: 0.507102, loss_mean: 0.842005, proj_loss: -0.414159, loss_mean_cls: 0.079257, grad_norm: 0.526543 +Steps: 0%| | 4696/1000000 [19:30<68:25:49, 4.04it/s, grad_norm=0.527, loss_final=0.507, loss_mean=0.842, loss_mean_cls=0.0793, proj_loss=-0.414][2026-03-26 12:50:33] Step: 4696, Training Logs: loss_final: 0.495863, loss_mean: 0.814454, proj_loss: -0.401425, loss_mean_cls: 0.082834, grad_norm: 0.251839 +Steps: 0%| | 4697/1000000 [19:30<68:22:15, 4.04it/s, grad_norm=0.252, loss_final=0.496, loss_mean=0.814, loss_mean_cls=0.0828, proj_loss=-0.401][2026-03-26 12:50:33] Step: 4697, Training Logs: loss_final: 0.487712, loss_mean: 0.813080, proj_loss: -0.407383, loss_mean_cls: 0.082016, grad_norm: 0.336120 +Steps: 0%| | 4698/1000000 [19:30<68:24:18, 4.04it/s, grad_norm=0.336, loss_final=0.488, loss_mean=0.813, loss_mean_cls=0.082, proj_loss=-0.407][2026-03-26 12:50:33] Step: 4698, Training Logs: loss_final: 0.512399, loss_mean: 0.844021, proj_loss: -0.411585, loss_mean_cls: 0.079963, grad_norm: 0.413961 +Steps: 0%| | 4699/1000000 [19:30<69:05:39, 4.00it/s, grad_norm=0.414, loss_final=0.512, loss_mean=0.844, loss_mean_cls=0.08, proj_loss=-0.412][2026-03-26 12:50:34] Step: 4699, Training Logs: loss_final: 0.499655, loss_mean: 0.833362, proj_loss: -0.414301, loss_mean_cls: 0.080594, grad_norm: 0.334700 +Steps: 0%| | 4700/1000000 [19:31<68:56:07, 4.01it/s, grad_norm=0.335, loss_final=0.5, loss_mean=0.833, loss_mean_cls=0.0806, proj_loss=-0.414][2026-03-26 12:50:34] Step: 4700, Training Logs: loss_final: 0.475878, loss_mean: 0.824172, proj_loss: -0.425945, loss_mean_cls: 0.077651, grad_norm: 0.303394 +Steps: 0%| | 4701/1000000 [19:31<68:45:57, 4.02it/s, grad_norm=0.303, loss_final=0.476, loss_mean=0.824, loss_mean_cls=0.0777, proj_loss=-0.426][2026-03-26 12:50:34] Step: 4701, Training Logs: loss_final: 0.486326, loss_mean: 0.833798, proj_loss: -0.423459, loss_mean_cls: 0.075987, grad_norm: 0.385364 +Steps: 0%| | 4702/1000000 [19:31<68:39:13, 4.03it/s, grad_norm=0.385, loss_final=0.486, loss_mean=0.834, loss_mean_cls=0.076, proj_loss=-0.423][2026-03-26 12:50:34] Step: 4702, Training Logs: loss_final: 0.485822, loss_mean: 0.827401, proj_loss: -0.420940, loss_mean_cls: 0.079361, grad_norm: 0.359416 +Steps: 0%| | 4703/1000000 [19:31<68:38:52, 4.03it/s, grad_norm=0.359, loss_final=0.486, loss_mean=0.827, loss_mean_cls=0.0794, proj_loss=-0.421][2026-03-26 12:50:35] Step: 4703, Training Logs: loss_final: 0.489099, loss_mean: 0.828269, proj_loss: -0.417103, loss_mean_cls: 0.077933, grad_norm: 0.296558 +Steps: 0%| | 4704/1000000 [19:32<68:34:56, 4.03it/s, grad_norm=0.297, loss_final=0.489, loss_mean=0.828, loss_mean_cls=0.0779, proj_loss=-0.417][2026-03-26 12:50:35] Step: 4704, Training Logs: loss_final: 0.509998, loss_mean: 0.846004, proj_loss: -0.414032, loss_mean_cls: 0.078026, grad_norm: 0.246741 +Steps: 0%| | 4705/1000000 [19:32<68:30:47, 4.04it/s, grad_norm=0.247, loss_final=0.51, loss_mean=0.846, loss_mean_cls=0.078, proj_loss=-0.414][2026-03-26 12:50:35] Step: 4705, Training Logs: loss_final: 0.492310, loss_mean: 0.834985, proj_loss: -0.421666, loss_mean_cls: 0.078991, grad_norm: 0.438402 +Steps: 0%| | 4706/1000000 [19:32<68:29:11, 4.04it/s, grad_norm=0.438, loss_final=0.492, loss_mean=0.835, loss_mean_cls=0.079, proj_loss=-0.422][2026-03-26 12:50:35] Step: 4706, Training Logs: loss_final: 0.493974, loss_mean: 0.831079, proj_loss: -0.414021, loss_mean_cls: 0.076917, grad_norm: 0.303166 +Steps: 0%| | 4707/1000000 [19:32<68:22:22, 4.04it/s, grad_norm=0.303, loss_final=0.494, loss_mean=0.831, loss_mean_cls=0.0769, proj_loss=-0.414][2026-03-26 12:50:36] Step: 4707, Training Logs: loss_final: 0.487125, loss_mean: 0.809831, proj_loss: -0.404854, loss_mean_cls: 0.082148, grad_norm: 0.366348 +Steps: 0%| | 4708/1000000 [19:33<68:23:50, 4.04it/s, grad_norm=0.366, loss_final=0.487, loss_mean=0.81, loss_mean_cls=0.0821, proj_loss=-0.405][2026-03-26 12:50:36] Step: 4708, Training Logs: loss_final: 0.481535, loss_mean: 0.813761, proj_loss: -0.412801, loss_mean_cls: 0.080575, grad_norm: 0.390531 +Steps: 0%| | 4709/1000000 [19:33<68:23:53, 4.04it/s, grad_norm=0.391, loss_final=0.482, loss_mean=0.814, loss_mean_cls=0.0806, proj_loss=-0.413][2026-03-26 12:50:36] Step: 4709, Training Logs: loss_final: 0.497237, loss_mean: 0.825167, proj_loss: -0.407711, loss_mean_cls: 0.079780, grad_norm: 0.248255 +Steps: 0%| | 4710/1000000 [19:33<68:22:56, 4.04it/s, grad_norm=0.248, loss_final=0.497, loss_mean=0.825, loss_mean_cls=0.0798, proj_loss=-0.408][2026-03-26 12:50:36] Step: 4710, Training Logs: loss_final: 0.495255, loss_mean: 0.830653, proj_loss: -0.414376, loss_mean_cls: 0.078978, grad_norm: 0.438009 +Steps: 0%| | 4711/1000000 [19:33<68:24:11, 4.04it/s, grad_norm=0.438, loss_final=0.495, loss_mean=0.831, loss_mean_cls=0.079, proj_loss=-0.414][2026-03-26 12:50:37] Step: 4711, Training Logs: loss_final: 0.500218, loss_mean: 0.833595, proj_loss: -0.411522, loss_mean_cls: 0.078144, grad_norm: 0.456951 +Steps: 0%| | 4712/1000000 [19:34<68:21:13, 4.04it/s, grad_norm=0.457, loss_final=0.5, loss_mean=0.834, loss_mean_cls=0.0781, proj_loss=-0.412][2026-03-26 12:50:37] Step: 4712, Training Logs: loss_final: 0.497227, loss_mean: 0.811430, proj_loss: -0.397785, loss_mean_cls: 0.083582, grad_norm: 0.324934 +Steps: 0%| | 4713/1000000 [19:34<68:23:46, 4.04it/s, grad_norm=0.325, loss_final=0.497, loss_mean=0.811, loss_mean_cls=0.0836, proj_loss=-0.398][2026-03-26 12:50:37] Step: 4713, Training Logs: loss_final: 0.495011, loss_mean: 0.825654, proj_loss: -0.410140, loss_mean_cls: 0.079496, grad_norm: 0.387929 +Steps: 0%| | 4714/1000000 [19:34<68:24:52, 4.04it/s, grad_norm=0.388, loss_final=0.495, loss_mean=0.826, loss_mean_cls=0.0795, proj_loss=-0.41][2026-03-26 12:50:37] Step: 4714, Training Logs: loss_final: 0.495339, loss_mean: 0.833205, proj_loss: -0.416221, loss_mean_cls: 0.078355, grad_norm: 0.340860 +Steps: 0%| | 4715/1000000 [19:34<68:23:20, 4.04it/s, grad_norm=0.341, loss_final=0.495, loss_mean=0.833, loss_mean_cls=0.0784, proj_loss=-0.416][2026-03-26 12:50:38] Step: 4715, Training Logs: loss_final: 0.514463, loss_mean: 0.841142, proj_loss: -0.408750, loss_mean_cls: 0.082070, grad_norm: 0.384458 +Steps: 0%| | 4716/1000000 [19:34<68:23:36, 4.04it/s, grad_norm=0.384, loss_final=0.514, loss_mean=0.841, loss_mean_cls=0.0821, proj_loss=-0.409][2026-03-26 12:50:38] Step: 4716, Training Logs: loss_final: 0.495630, loss_mean: 0.823918, proj_loss: -0.410763, loss_mean_cls: 0.082476, grad_norm: 0.472759 +Steps: 0%| | 4717/1000000 [19:35<68:24:23, 4.04it/s, grad_norm=0.473, loss_final=0.496, loss_mean=0.824, loss_mean_cls=0.0825, proj_loss=-0.411][2026-03-26 12:50:38] Step: 4717, Training Logs: loss_final: 0.492393, loss_mean: 0.811427, proj_loss: -0.400699, loss_mean_cls: 0.081665, grad_norm: 0.343478 +Steps: 0%| | 4718/1000000 [19:35<68:22:27, 4.04it/s, grad_norm=0.343, loss_final=0.492, loss_mean=0.811, loss_mean_cls=0.0817, proj_loss=-0.401][2026-03-26 12:50:38] Step: 4718, Training Logs: loss_final: 0.487927, loss_mean: 0.817127, proj_loss: -0.412082, loss_mean_cls: 0.082882, grad_norm: 0.300486 +Steps: 0%| | 4719/1000000 [19:35<68:25:07, 4.04it/s, grad_norm=0.3, loss_final=0.488, loss_mean=0.817, loss_mean_cls=0.0829, proj_loss=-0.412][2026-03-26 12:50:39] Step: 4719, Training Logs: loss_final: 0.515297, loss_mean: 0.836974, proj_loss: -0.403392, loss_mean_cls: 0.081715, grad_norm: 0.347807 +Steps: 0%| | 4720/1000000 [19:35<68:23:02, 4.04it/s, grad_norm=0.348, loss_final=0.515, loss_mean=0.837, loss_mean_cls=0.0817, proj_loss=-0.403][2026-03-26 12:50:39] Step: 4720, Training Logs: loss_final: 0.489163, loss_mean: 0.828803, proj_loss: -0.418608, loss_mean_cls: 0.078967, grad_norm: 0.277631 +Steps: 0%| | 4721/1000000 [19:36<68:23:14, 4.04it/s, grad_norm=0.278, loss_final=0.489, loss_mean=0.829, loss_mean_cls=0.079, proj_loss=-0.419][2026-03-26 12:50:39] Step: 4721, Training Logs: loss_final: 0.486632, loss_mean: 0.830654, proj_loss: -0.420369, loss_mean_cls: 0.076347, grad_norm: 0.436800 +Steps: 0%| | 4722/1000000 [19:36<68:25:31, 4.04it/s, grad_norm=0.437, loss_final=0.487, loss_mean=0.831, loss_mean_cls=0.0763, proj_loss=-0.42][2026-03-26 12:50:39] Step: 4722, Training Logs: loss_final: 0.490671, loss_mean: 0.818265, proj_loss: -0.410474, loss_mean_cls: 0.082881, grad_norm: 0.423938 +Steps: 0%| | 4723/1000000 [19:36<68:23:58, 4.04it/s, grad_norm=0.424, loss_final=0.491, loss_mean=0.818, loss_mean_cls=0.0829, proj_loss=-0.41][2026-03-26 12:50:40] Step: 4723, Training Logs: loss_final: 0.508352, loss_mean: 0.834296, proj_loss: -0.407317, loss_mean_cls: 0.081373, grad_norm: 0.290909 +Steps: 0%| | 4724/1000000 [19:37<71:28:26, 3.87it/s, grad_norm=0.291, loss_final=0.508, loss_mean=0.834, loss_mean_cls=0.0814, proj_loss=-0.407][2026-03-26 12:50:40] Step: 4724, Training Logs: loss_final: 0.488743, loss_mean: 0.817612, proj_loss: -0.410267, loss_mean_cls: 0.081398, grad_norm: 0.448021 +Steps: 0%| | 4725/1000000 [19:37<70:45:27, 3.91it/s, grad_norm=0.448, loss_final=0.489, loss_mean=0.818, loss_mean_cls=0.0814, proj_loss=-0.41][2026-03-26 12:50:40] Step: 4725, Training Logs: loss_final: 0.495701, loss_mean: 0.827413, proj_loss: -0.413732, loss_mean_cls: 0.082019, grad_norm: 0.240046 +Steps: 0%| | 4726/1000000 [19:37<70:05:20, 3.94it/s, grad_norm=0.24, loss_final=0.496, loss_mean=0.827, loss_mean_cls=0.082, proj_loss=-0.414][2026-03-26 12:50:40] Step: 4726, Training Logs: loss_final: 0.498270, loss_mean: 0.833720, proj_loss: -0.416503, loss_mean_cls: 0.081052, grad_norm: 0.464741 +Steps: 0%| | 4727/1000000 [19:37<69:34:11, 3.97it/s, grad_norm=0.465, loss_final=0.498, loss_mean=0.834, loss_mean_cls=0.0811, proj_loss=-0.417][2026-03-26 12:50:41] Step: 4727, Training Logs: loss_final: 0.510301, loss_mean: 0.841913, proj_loss: -0.410282, loss_mean_cls: 0.078669, grad_norm: 0.428709 +Steps: 0%| | 4728/1000000 [19:38<69:10:15, 4.00it/s, grad_norm=0.429, loss_final=0.51, loss_mean=0.842, loss_mean_cls=0.0787, proj_loss=-0.41][2026-03-26 12:50:41] Step: 4728, Training Logs: loss_final: 0.486396, loss_mean: 0.816870, proj_loss: -0.411758, loss_mean_cls: 0.081283, grad_norm: 0.246550 +Steps: 0%| | 4729/1000000 [19:38<68:56:01, 4.01it/s, grad_norm=0.247, loss_final=0.486, loss_mean=0.817, loss_mean_cls=0.0813, proj_loss=-0.412][2026-03-26 12:50:41] Step: 4729, Training Logs: loss_final: 0.494080, loss_mean: 0.825460, proj_loss: -0.411137, loss_mean_cls: 0.079757, grad_norm: 0.522987 +Steps: 0%| | 4730/1000000 [19:38<68:49:32, 4.02it/s, grad_norm=0.523, loss_final=0.494, loss_mean=0.825, loss_mean_cls=0.0798, proj_loss=-0.411][2026-03-26 12:50:41] Step: 4730, Training Logs: loss_final: 0.504007, loss_mean: 0.843356, proj_loss: -0.418966, loss_mean_cls: 0.079616, grad_norm: 0.616622 +Steps: 0%| | 4731/1000000 [19:38<68:43:11, 4.02it/s, grad_norm=0.617, loss_final=0.504, loss_mean=0.843, loss_mean_cls=0.0796, proj_loss=-0.419][2026-03-26 12:50:42] Step: 4731, Training Logs: loss_final: 0.499554, loss_mean: 0.822511, proj_loss: -0.404884, loss_mean_cls: 0.081927, grad_norm: 0.371724 +Steps: 0%| | 4732/1000000 [19:38<68:38:36, 4.03it/s, grad_norm=0.372, loss_final=0.5, loss_mean=0.823, loss_mean_cls=0.0819, proj_loss=-0.405][2026-03-26 12:50:42] Step: 4732, Training Logs: loss_final: 0.496385, loss_mean: 0.835967, proj_loss: -0.417408, loss_mean_cls: 0.077827, grad_norm: 0.670492 +Steps: 0%| | 4733/1000000 [19:39<68:35:11, 4.03it/s, grad_norm=0.67, loss_final=0.496, loss_mean=0.836, loss_mean_cls=0.0778, proj_loss=-0.417][2026-03-26 12:50:42] Step: 4733, Training Logs: loss_final: 0.514044, loss_mean: 0.841021, proj_loss: -0.407661, loss_mean_cls: 0.080683, grad_norm: 0.561747 +Steps: 0%| | 4734/1000000 [19:39<68:40:51, 4.03it/s, grad_norm=0.562, loss_final=0.514, loss_mean=0.841, loss_mean_cls=0.0807, proj_loss=-0.408][2026-03-26 12:50:42] Step: 4734, Training Logs: loss_final: 0.503572, loss_mean: 0.835027, proj_loss: -0.410101, loss_mean_cls: 0.078646, grad_norm: 0.309976 +Steps: 0%| | 4735/1000000 [19:39<68:35:13, 4.03it/s, grad_norm=0.31, loss_final=0.504, loss_mean=0.835, loss_mean_cls=0.0786, proj_loss=-0.41][2026-03-26 12:50:43] Step: 4735, Training Logs: loss_final: 0.498717, loss_mean: 0.838218, proj_loss: -0.417984, loss_mean_cls: 0.078482, grad_norm: 0.730534 +Steps: 0%| | 4736/1000000 [19:39<68:28:54, 4.04it/s, grad_norm=0.731, loss_final=0.499, loss_mean=0.838, loss_mean_cls=0.0785, proj_loss=-0.418][2026-03-26 12:50:43] Step: 4736, Training Logs: loss_final: 0.510546, loss_mean: 0.839238, proj_loss: -0.409049, loss_mean_cls: 0.080357, grad_norm: 0.430877 +Steps: 0%| | 4737/1000000 [19:40<68:26:59, 4.04it/s, grad_norm=0.431, loss_final=0.511, loss_mean=0.839, loss_mean_cls=0.0804, proj_loss=-0.409][2026-03-26 12:50:43] Step: 4737, Training Logs: loss_final: 0.492636, loss_mean: 0.828085, proj_loss: -0.414101, loss_mean_cls: 0.078653, grad_norm: 0.470312 +Steps: 0%| | 4738/1000000 [19:40<68:36:20, 4.03it/s, grad_norm=0.47, loss_final=0.493, loss_mean=0.828, loss_mean_cls=0.0787, proj_loss=-0.414][2026-03-26 12:50:43] Step: 4738, Training Logs: loss_final: 0.519483, loss_mean: 0.855340, proj_loss: -0.412240, loss_mean_cls: 0.076384, grad_norm: 0.745509 +Steps: 0%| | 4739/1000000 [19:40<68:33:33, 4.03it/s, grad_norm=0.746, loss_final=0.519, loss_mean=0.855, loss_mean_cls=0.0764, proj_loss=-0.412][2026-03-26 12:50:44] Step: 4739, Training Logs: loss_final: 0.493930, loss_mean: 0.829509, proj_loss: -0.413540, loss_mean_cls: 0.077962, grad_norm: 0.450550 +Steps: 0%| | 4740/1000000 [19:40<68:31:10, 4.03it/s, grad_norm=0.451, loss_final=0.494, loss_mean=0.83, loss_mean_cls=0.078, proj_loss=-0.414][2026-03-26 12:50:44] Step: 4740, Training Logs: loss_final: 0.500452, loss_mean: 0.837159, proj_loss: -0.416506, loss_mean_cls: 0.079799, grad_norm: 0.414305 +Steps: 0%| | 4741/1000000 [19:41<68:28:27, 4.04it/s, grad_norm=0.414, loss_final=0.5, loss_mean=0.837, loss_mean_cls=0.0798, proj_loss=-0.417][2026-03-26 12:50:44] Step: 4741, Training Logs: loss_final: 0.507015, loss_mean: 0.829335, proj_loss: -0.403844, loss_mean_cls: 0.081524, grad_norm: 0.429674 +Steps: 0%| | 4742/1000000 [19:41<68:37:19, 4.03it/s, grad_norm=0.43, loss_final=0.507, loss_mean=0.829, loss_mean_cls=0.0815, proj_loss=-0.404][2026-03-26 12:50:44] Step: 4742, Training Logs: loss_final: 0.506367, loss_mean: 0.850401, proj_loss: -0.421971, loss_mean_cls: 0.077936, grad_norm: 0.424134 +Steps: 0%| | 4743/1000000 [19:41<68:31:49, 4.03it/s, grad_norm=0.424, loss_final=0.506, loss_mean=0.85, loss_mean_cls=0.0779, proj_loss=-0.422][2026-03-26 12:50:45] Step: 4743, Training Logs: loss_final: 0.474282, loss_mean: 0.812064, proj_loss: -0.416636, loss_mean_cls: 0.078853, grad_norm: 0.485654 +Steps: 0%| | 4744/1000000 [19:41<68:32:27, 4.03it/s, grad_norm=0.486, loss_final=0.474, loss_mean=0.812, loss_mean_cls=0.0789, proj_loss=-0.417][2026-03-26 12:50:45] Step: 4744, Training Logs: loss_final: 0.495216, loss_mean: 0.820926, proj_loss: -0.406206, loss_mean_cls: 0.080496, grad_norm: 0.392554 +Steps: 0%| | 4745/1000000 [19:42<68:29:38, 4.04it/s, grad_norm=0.393, loss_final=0.495, loss_mean=0.821, loss_mean_cls=0.0805, proj_loss=-0.406][2026-03-26 12:50:45] Step: 4745, Training Logs: loss_final: 0.517060, loss_mean: 0.849110, proj_loss: -0.412193, loss_mean_cls: 0.080143, grad_norm: 0.444891 +Steps: 0%| | 4746/1000000 [19:42<68:37:47, 4.03it/s, grad_norm=0.445, loss_final=0.517, loss_mean=0.849, loss_mean_cls=0.0801, proj_loss=-0.412][2026-03-26 12:50:45] Step: 4746, Training Logs: loss_final: 0.495903, loss_mean: 0.823531, proj_loss: -0.408917, loss_mean_cls: 0.081289, grad_norm: 0.548246 +Steps: 0%| | 4747/1000000 [19:42<68:31:54, 4.03it/s, grad_norm=0.548, loss_final=0.496, loss_mean=0.824, loss_mean_cls=0.0813, proj_loss=-0.409][2026-03-26 12:50:46] Step: 4747, Training Logs: loss_final: 0.492026, loss_mean: 0.834370, proj_loss: -0.420287, loss_mean_cls: 0.077943, grad_norm: 0.353661 +Steps: 0%| | 4748/1000000 [19:42<68:30:31, 4.04it/s, grad_norm=0.354, loss_final=0.492, loss_mean=0.834, loss_mean_cls=0.0779, proj_loss=-0.42][2026-03-26 12:50:46] Step: 4748, Training Logs: loss_final: 0.503164, loss_mean: 0.841520, proj_loss: -0.415930, loss_mean_cls: 0.077573, grad_norm: 0.517058 +Steps: 0%| | 4749/1000000 [19:43<68:28:30, 4.04it/s, grad_norm=0.517, loss_final=0.503, loss_mean=0.842, loss_mean_cls=0.0776, proj_loss=-0.416][2026-03-26 12:50:46] Step: 4749, Training Logs: loss_final: 0.510371, loss_mean: 0.844776, proj_loss: -0.413724, loss_mean_cls: 0.079319, grad_norm: 0.370473 +Steps: 0%| | 4750/1000000 [19:43<68:34:23, 4.03it/s, grad_norm=0.37, loss_final=0.51, loss_mean=0.845, loss_mean_cls=0.0793, proj_loss=-0.414][2026-03-26 12:50:46] Step: 4750, Training Logs: loss_final: 0.494001, loss_mean: 0.829015, proj_loss: -0.414812, loss_mean_cls: 0.079798, grad_norm: 0.338561 +Steps: 0%| | 4751/1000000 [19:43<68:31:04, 4.03it/s, grad_norm=0.339, loss_final=0.494, loss_mean=0.829, loss_mean_cls=0.0798, proj_loss=-0.415][2026-03-26 12:50:47] Step: 4751, Training Logs: loss_final: 0.505518, loss_mean: 0.847052, proj_loss: -0.419133, loss_mean_cls: 0.077599, grad_norm: 0.564074 +Steps: 0%| | 4752/1000000 [19:43<68:35:19, 4.03it/s, grad_norm=0.564, loss_final=0.506, loss_mean=0.847, loss_mean_cls=0.0776, proj_loss=-0.419][2026-03-26 12:50:47] Step: 4752, Training Logs: loss_final: 0.488087, loss_mean: 0.824478, proj_loss: -0.416062, loss_mean_cls: 0.079671, grad_norm: 0.424038 +Steps: 0%| | 4753/1000000 [19:44<68:31:54, 4.03it/s, grad_norm=0.424, loss_final=0.488, loss_mean=0.824, loss_mean_cls=0.0797, proj_loss=-0.416][2026-03-26 12:50:47] Step: 4753, Training Logs: loss_final: 0.494346, loss_mean: 0.829179, proj_loss: -0.414583, loss_mean_cls: 0.079750, grad_norm: 0.456960 +Steps: 0%| | 4754/1000000 [19:44<68:40:30, 4.03it/s, grad_norm=0.457, loss_final=0.494, loss_mean=0.829, loss_mean_cls=0.0797, proj_loss=-0.415][2026-03-26 12:50:47] Step: 4754, Training Logs: loss_final: 0.499639, loss_mean: 0.832745, proj_loss: -0.413946, loss_mean_cls: 0.080841, grad_norm: 0.579240 +Steps: 0%| | 4755/1000000 [19:44<68:35:15, 4.03it/s, grad_norm=0.579, loss_final=0.5, loss_mean=0.833, loss_mean_cls=0.0808, proj_loss=-0.414][2026-03-26 12:50:48] Step: 4755, Training Logs: loss_final: 0.478172, loss_mean: 0.805114, proj_loss: -0.408941, loss_mean_cls: 0.081998, grad_norm: 0.388783 +Steps: 0%| | 4756/1000000 [19:44<68:32:10, 4.03it/s, grad_norm=0.389, loss_final=0.478, loss_mean=0.805, loss_mean_cls=0.082, proj_loss=-0.409][2026-03-26 12:50:48] Step: 4756, Training Logs: loss_final: 0.485416, loss_mean: 0.813852, proj_loss: -0.408668, loss_mean_cls: 0.080232, grad_norm: 0.373863 +Steps: 0%| | 4757/1000000 [19:45<68:29:21, 4.04it/s, grad_norm=0.374, loss_final=0.485, loss_mean=0.814, loss_mean_cls=0.0802, proj_loss=-0.409][2026-03-26 12:50:48] Step: 4757, Training Logs: loss_final: 0.506404, loss_mean: 0.851370, proj_loss: -0.422498, loss_mean_cls: 0.077532, grad_norm: 0.447484 +Steps: 0%| | 4758/1000000 [19:45<68:28:46, 4.04it/s, grad_norm=0.447, loss_final=0.506, loss_mean=0.851, loss_mean_cls=0.0775, proj_loss=-0.422][2026-03-26 12:50:48] Step: 4758, Training Logs: loss_final: 0.490059, loss_mean: 0.823864, proj_loss: -0.412853, loss_mean_cls: 0.079048, grad_norm: 0.340847 +Steps: 0%| | 4759/1000000 [19:45<68:25:19, 4.04it/s, grad_norm=0.341, loss_final=0.49, loss_mean=0.824, loss_mean_cls=0.079, proj_loss=-0.413][2026-03-26 12:50:49] Step: 4759, Training Logs: loss_final: 0.491349, loss_mean: 0.824503, proj_loss: -0.412820, loss_mean_cls: 0.079665, grad_norm: 0.338543 +Steps: 0%| | 4760/1000000 [19:45<68:22:09, 4.04it/s, grad_norm=0.339, loss_final=0.491, loss_mean=0.825, loss_mean_cls=0.0797, proj_loss=-0.413][2026-03-26 12:50:49] Step: 4760, Training Logs: loss_final: 0.504046, loss_mean: 0.833137, proj_loss: -0.410079, loss_mean_cls: 0.080989, grad_norm: 0.314502 +Steps: 0%| | 4761/1000000 [19:46<68:20:32, 4.05it/s, grad_norm=0.315, loss_final=0.504, loss_mean=0.833, loss_mean_cls=0.081, proj_loss=-0.41][2026-03-26 12:50:49] Step: 4761, Training Logs: loss_final: 0.489211, loss_mean: 0.824152, proj_loss: -0.414603, loss_mean_cls: 0.079662, grad_norm: 0.437625 +Steps: 0%| | 4762/1000000 [19:46<68:21:31, 4.04it/s, grad_norm=0.438, loss_final=0.489, loss_mean=0.824, loss_mean_cls=0.0797, proj_loss=-0.415][2026-03-26 12:50:49] Step: 4762, Training Logs: loss_final: 0.504488, loss_mean: 0.846904, proj_loss: -0.420228, loss_mean_cls: 0.077812, grad_norm: 0.430939 +Steps: 0%| | 4763/1000000 [19:46<68:23:51, 4.04it/s, grad_norm=0.431, loss_final=0.504, loss_mean=0.847, loss_mean_cls=0.0778, proj_loss=-0.42][2026-03-26 12:50:50] Step: 4763, Training Logs: loss_final: 0.471422, loss_mean: 0.805597, proj_loss: -0.414720, loss_mean_cls: 0.080546, grad_norm: 0.311439 +Steps: 0%| | 4764/1000000 [19:46<68:25:19, 4.04it/s, grad_norm=0.311, loss_final=0.471, loss_mean=0.806, loss_mean_cls=0.0805, proj_loss=-0.415][2026-03-26 12:50:50] Step: 4764, Training Logs: loss_final: 0.504652, loss_mean: 0.842658, proj_loss: -0.416753, loss_mean_cls: 0.078747, grad_norm: 0.556525 +Steps: 0%| | 4765/1000000 [19:47<68:25:52, 4.04it/s, grad_norm=0.557, loss_final=0.505, loss_mean=0.843, loss_mean_cls=0.0787, proj_loss=-0.417][2026-03-26 12:50:50] Step: 4765, Training Logs: loss_final: 0.507119, loss_mean: 0.842079, proj_loss: -0.412867, loss_mean_cls: 0.077907, grad_norm: 0.601666 +Steps: 0%| | 4766/1000000 [19:47<68:26:26, 4.04it/s, grad_norm=0.602, loss_final=0.507, loss_mean=0.842, loss_mean_cls=0.0779, proj_loss=-0.413][2026-03-26 12:50:50] Step: 4766, Training Logs: loss_final: 0.485325, loss_mean: 0.817143, proj_loss: -0.410992, loss_mean_cls: 0.079175, grad_norm: 0.332983 +Steps: 0%| | 4767/1000000 [19:47<68:26:58, 4.04it/s, grad_norm=0.333, loss_final=0.485, loss_mean=0.817, loss_mean_cls=0.0792, proj_loss=-0.411][2026-03-26 12:50:51] Step: 4767, Training Logs: loss_final: 0.479492, loss_mean: 0.814246, proj_loss: -0.414672, loss_mean_cls: 0.079918, grad_norm: 0.543937 +Steps: 0%| | 4768/1000000 [19:47<68:27:23, 4.04it/s, grad_norm=0.544, loss_final=0.479, loss_mean=0.814, loss_mean_cls=0.0799, proj_loss=-0.415][2026-03-26 12:50:51] Step: 4768, Training Logs: loss_final: 0.488567, loss_mean: 0.824178, proj_loss: -0.414708, loss_mean_cls: 0.079097, grad_norm: 0.521414 +Steps: 0%| | 4769/1000000 [19:48<68:27:20, 4.04it/s, grad_norm=0.521, loss_final=0.489, loss_mean=0.824, loss_mean_cls=0.0791, proj_loss=-0.415][2026-03-26 12:50:51] Step: 4769, Training Logs: loss_final: 0.498278, loss_mean: 0.818794, proj_loss: -0.403003, loss_mean_cls: 0.082487, grad_norm: 0.310372 +Steps: 0%| | 4770/1000000 [19:48<68:25:41, 4.04it/s, grad_norm=0.31, loss_final=0.498, loss_mean=0.819, loss_mean_cls=0.0825, proj_loss=-0.403][2026-03-26 12:50:51] Step: 4770, Training Logs: loss_final: 0.514080, loss_mean: 0.837458, proj_loss: -0.404442, loss_mean_cls: 0.081064, grad_norm: 0.662755 +Steps: 0%| | 4771/1000000 [19:48<68:24:28, 4.04it/s, grad_norm=0.663, loss_final=0.514, loss_mean=0.837, loss_mean_cls=0.0811, proj_loss=-0.404][2026-03-26 12:50:52] Step: 4771, Training Logs: loss_final: 0.498186, loss_mean: 0.837384, proj_loss: -0.417926, loss_mean_cls: 0.078728, grad_norm: 0.492210 +Steps: 0%| | 4772/1000000 [19:48<68:25:20, 4.04it/s, grad_norm=0.492, loss_final=0.498, loss_mean=0.837, loss_mean_cls=0.0787, proj_loss=-0.418][2026-03-26 12:50:52] Step: 4772, Training Logs: loss_final: 0.496500, loss_mean: 0.829230, proj_loss: -0.412858, loss_mean_cls: 0.080128, grad_norm: 0.410114 +Steps: 0%| | 4773/1000000 [19:49<68:25:02, 4.04it/s, grad_norm=0.41, loss_final=0.497, loss_mean=0.829, loss_mean_cls=0.0801, proj_loss=-0.413][2026-03-26 12:50:52] Step: 4773, Training Logs: loss_final: 0.489351, loss_mean: 0.809065, proj_loss: -0.401695, loss_mean_cls: 0.081982, grad_norm: 0.933138 +Steps: 0%| | 4774/1000000 [19:49<68:22:44, 4.04it/s, grad_norm=0.933, loss_final=0.489, loss_mean=0.809, loss_mean_cls=0.082, proj_loss=-0.402][2026-03-26 12:50:52] Step: 4774, Training Logs: loss_final: 0.511872, loss_mean: 0.845857, proj_loss: -0.411386, loss_mean_cls: 0.077401, grad_norm: 0.535793 +Steps: 0%| | 4775/1000000 [19:49<68:24:20, 4.04it/s, grad_norm=0.536, loss_final=0.512, loss_mean=0.846, loss_mean_cls=0.0774, proj_loss=-0.411][2026-03-26 12:50:53] Step: 4775, Training Logs: loss_final: 0.493577, loss_mean: 0.834132, proj_loss: -0.416993, loss_mean_cls: 0.076437, grad_norm: 0.518030 +Steps: 0%| | 4776/1000000 [19:49<68:23:14, 4.04it/s, grad_norm=0.518, loss_final=0.494, loss_mean=0.834, loss_mean_cls=0.0764, proj_loss=-0.417][2026-03-26 12:50:53] Step: 4776, Training Logs: loss_final: 0.499219, loss_mean: 0.834377, proj_loss: -0.415091, loss_mean_cls: 0.079933, grad_norm: 0.688082 +Steps: 0%| | 4777/1000000 [19:50<68:26:11, 4.04it/s, grad_norm=0.688, loss_final=0.499, loss_mean=0.834, loss_mean_cls=0.0799, proj_loss=-0.415][2026-03-26 12:50:53] Step: 4777, Training Logs: loss_final: 0.463315, loss_mean: 0.794310, proj_loss: -0.413144, loss_mean_cls: 0.082149, grad_norm: 0.514726 +Steps: 0%| | 4778/1000000 [19:50<68:26:13, 4.04it/s, grad_norm=0.515, loss_final=0.463, loss_mean=0.794, loss_mean_cls=0.0821, proj_loss=-0.413][2026-03-26 12:50:53] Step: 4778, Training Logs: loss_final: 0.508590, loss_mean: 0.849325, proj_loss: -0.416753, loss_mean_cls: 0.076018, grad_norm: 0.445499 +Steps: 0%| | 4779/1000000 [19:50<68:23:22, 4.04it/s, grad_norm=0.445, loss_final=0.509, loss_mean=0.849, loss_mean_cls=0.076, proj_loss=-0.417][2026-03-26 12:50:54] Step: 4779, Training Logs: loss_final: 0.502566, loss_mean: 0.831749, proj_loss: -0.410116, loss_mean_cls: 0.080933, grad_norm: 0.520045 +Steps: 0%| | 4780/1000000 [19:50<68:21:46, 4.04it/s, grad_norm=0.52, loss_final=0.503, loss_mean=0.832, loss_mean_cls=0.0809, proj_loss=-0.41][2026-03-26 12:50:54] Step: 4780, Training Logs: loss_final: 0.486475, loss_mean: 0.823829, proj_loss: -0.415220, loss_mean_cls: 0.077866, grad_norm: 0.398440 +Steps: 0%| | 4781/1000000 [19:51<68:25:01, 4.04it/s, grad_norm=0.398, loss_final=0.486, loss_mean=0.824, loss_mean_cls=0.0779, proj_loss=-0.415][2026-03-26 12:50:54] Step: 4781, Training Logs: loss_final: 0.499022, loss_mean: 0.841500, proj_loss: -0.419985, loss_mean_cls: 0.077507, grad_norm: 0.548938 +Steps: 0%| | 4782/1000000 [19:51<68:25:03, 4.04it/s, grad_norm=0.549, loss_final=0.499, loss_mean=0.842, loss_mean_cls=0.0775, proj_loss=-0.42][2026-03-26 12:50:54] Step: 4782, Training Logs: loss_final: 0.502002, loss_mean: 0.827291, proj_loss: -0.406258, loss_mean_cls: 0.080970, grad_norm: 0.278509 +Steps: 0%| | 4783/1000000 [19:51<68:26:16, 4.04it/s, grad_norm=0.279, loss_final=0.502, loss_mean=0.827, loss_mean_cls=0.081, proj_loss=-0.406][2026-03-26 12:50:55] Step: 4783, Training Logs: loss_final: 0.483685, loss_mean: 0.808891, proj_loss: -0.406898, loss_mean_cls: 0.081691, grad_norm: 0.738575 +Steps: 0%| | 4784/1000000 [19:51<68:25:36, 4.04it/s, grad_norm=0.739, loss_final=0.484, loss_mean=0.809, loss_mean_cls=0.0817, proj_loss=-0.407][2026-03-26 12:50:55] Step: 4784, Training Logs: loss_final: 0.488612, loss_mean: 0.825472, proj_loss: -0.416177, loss_mean_cls: 0.079318, grad_norm: 0.559728 +Steps: 0%| | 4785/1000000 [19:52<68:25:54, 4.04it/s, grad_norm=0.56, loss_final=0.489, loss_mean=0.825, loss_mean_cls=0.0793, proj_loss=-0.416][2026-03-26 12:50:55] Step: 4785, Training Logs: loss_final: 0.482679, loss_mean: 0.825684, proj_loss: -0.421766, loss_mean_cls: 0.078760, grad_norm: 0.378026 +Steps: 0%| | 4786/1000000 [19:52<68:21:24, 4.04it/s, grad_norm=0.378, loss_final=0.483, loss_mean=0.826, loss_mean_cls=0.0788, proj_loss=-0.422][2026-03-26 12:50:55] Step: 4786, Training Logs: loss_final: 0.517988, loss_mean: 0.854437, proj_loss: -0.415449, loss_mean_cls: 0.078999, grad_norm: 0.611123 +Steps: 0%| | 4787/1000000 [19:52<68:20:00, 4.05it/s, grad_norm=0.611, loss_final=0.518, loss_mean=0.854, loss_mean_cls=0.079, proj_loss=-0.415][2026-03-26 12:50:56] Step: 4787, Training Logs: loss_final: 0.494559, loss_mean: 0.828377, proj_loss: -0.414146, loss_mean_cls: 0.080329, grad_norm: 0.286305 +Steps: 0%| | 4788/1000000 [19:52<68:19:25, 4.05it/s, grad_norm=0.286, loss_final=0.495, loss_mean=0.828, loss_mean_cls=0.0803, proj_loss=-0.414][2026-03-26 12:50:56] Step: 4788, Training Logs: loss_final: 0.478970, loss_mean: 0.805164, proj_loss: -0.408939, loss_mean_cls: 0.082744, grad_norm: 0.420072 +Steps: 0%| | 4789/1000000 [19:53<69:14:23, 3.99it/s, grad_norm=0.42, loss_final=0.479, loss_mean=0.805, loss_mean_cls=0.0827, proj_loss=-0.409][2026-03-26 12:50:56] Step: 4789, Training Logs: loss_final: 0.489989, loss_mean: 0.813006, proj_loss: -0.406172, loss_mean_cls: 0.083154, grad_norm: 0.373993 +Steps: 0%| | 4790/1000000 [19:53<68:59:55, 4.01it/s, grad_norm=0.374, loss_final=0.49, loss_mean=0.813, loss_mean_cls=0.0832, proj_loss=-0.406][2026-03-26 12:50:56] Step: 4790, Training Logs: loss_final: 0.500338, loss_mean: 0.831823, proj_loss: -0.410947, loss_mean_cls: 0.079462, grad_norm: 0.487408 +Steps: 0%| | 4791/1000000 [19:53<68:49:27, 4.02it/s, grad_norm=0.487, loss_final=0.5, loss_mean=0.832, loss_mean_cls=0.0795, proj_loss=-0.411][2026-03-26 12:50:57] Step: 4791, Training Logs: loss_final: 0.514173, loss_mean: 0.856927, proj_loss: -0.420725, loss_mean_cls: 0.077971, grad_norm: 0.456811 +Steps: 0%| | 4792/1000000 [19:53<68:42:43, 4.02it/s, grad_norm=0.457, loss_final=0.514, loss_mean=0.857, loss_mean_cls=0.078, proj_loss=-0.421][2026-03-26 12:50:57] Step: 4792, Training Logs: loss_final: 0.491128, loss_mean: 0.819405, proj_loss: -0.409783, loss_mean_cls: 0.081506, grad_norm: 0.305792 +Steps: 0%| | 4793/1000000 [19:54<68:34:55, 4.03it/s, grad_norm=0.306, loss_final=0.491, loss_mean=0.819, loss_mean_cls=0.0815, proj_loss=-0.41][2026-03-26 12:50:57] Step: 4793, Training Logs: loss_final: 0.501900, loss_mean: 0.827544, proj_loss: -0.408518, loss_mean_cls: 0.082874, grad_norm: 0.344867 +Steps: 0%| | 4794/1000000 [19:54<68:33:59, 4.03it/s, grad_norm=0.345, loss_final=0.502, loss_mean=0.828, loss_mean_cls=0.0829, proj_loss=-0.409][2026-03-26 12:50:57] Step: 4794, Training Logs: loss_final: 0.502660, loss_mean: 0.829118, proj_loss: -0.409680, loss_mean_cls: 0.083222, grad_norm: 0.404135 +Steps: 0%| | 4795/1000000 [19:54<68:29:27, 4.04it/s, grad_norm=0.404, loss_final=0.503, loss_mean=0.829, loss_mean_cls=0.0832, proj_loss=-0.41][2026-03-26 12:50:57] Step: 4795, Training Logs: loss_final: 0.493200, loss_mean: 0.822564, proj_loss: -0.409560, loss_mean_cls: 0.080197, grad_norm: 0.404979 +Steps: 0%| | 4796/1000000 [19:54<68:28:53, 4.04it/s, grad_norm=0.405, loss_final=0.493, loss_mean=0.823, loss_mean_cls=0.0802, proj_loss=-0.41][2026-03-26 12:50:58] Step: 4796, Training Logs: loss_final: 0.503187, loss_mean: 0.826584, proj_loss: -0.404746, loss_mean_cls: 0.081349, grad_norm: 0.370321 +Steps: 0%| | 4797/1000000 [19:55<68:23:07, 4.04it/s, grad_norm=0.37, loss_final=0.503, loss_mean=0.827, loss_mean_cls=0.0813, proj_loss=-0.405][2026-03-26 12:50:58] Step: 4797, Training Logs: loss_final: 0.511643, loss_mean: 0.840611, proj_loss: -0.407722, loss_mean_cls: 0.078754, grad_norm: 0.448179 +Steps: 0%| | 4798/1000000 [19:55<68:22:44, 4.04it/s, grad_norm=0.448, loss_final=0.512, loss_mean=0.841, loss_mean_cls=0.0788, proj_loss=-0.408][2026-03-26 12:50:58] Step: 4798, Training Logs: loss_final: 0.494784, loss_mean: 0.816708, proj_loss: -0.405278, loss_mean_cls: 0.083354, grad_norm: 0.321453 +Steps: 0%| | 4799/1000000 [19:55<68:18:05, 4.05it/s, grad_norm=0.321, loss_final=0.495, loss_mean=0.817, loss_mean_cls=0.0834, proj_loss=-0.405][2026-03-26 12:50:58] Step: 4799, Training Logs: loss_final: 0.502690, loss_mean: 0.824930, proj_loss: -0.404885, loss_mean_cls: 0.082645, grad_norm: 0.348886 +Steps: 0%| | 4800/1000000 [19:55<68:20:15, 4.05it/s, grad_norm=0.349, loss_final=0.503, loss_mean=0.825, loss_mean_cls=0.0826, proj_loss=-0.405][2026-03-26 12:50:59] Step: 4800, Training Logs: loss_final: 0.525805, loss_mean: 0.858858, proj_loss: -0.410313, loss_mean_cls: 0.077260, grad_norm: 0.304875 +Steps: 0%| | 4801/1000000 [19:56<68:19:10, 4.05it/s, grad_norm=0.305, loss_final=0.526, loss_mean=0.859, loss_mean_cls=0.0773, proj_loss=-0.41][2026-03-26 12:50:59] Step: 4801, Training Logs: loss_final: 0.506477, loss_mean: 0.840140, proj_loss: -0.413859, loss_mean_cls: 0.080196, grad_norm: 0.270142 +Steps: 0%| | 4802/1000000 [19:56<68:22:52, 4.04it/s, grad_norm=0.27, loss_final=0.506, loss_mean=0.84, loss_mean_cls=0.0802, proj_loss=-0.414][2026-03-26 12:50:59] Step: 4802, Training Logs: loss_final: 0.490464, loss_mean: 0.819277, proj_loss: -0.410864, loss_mean_cls: 0.082051, grad_norm: 0.376882 +Steps: 0%| | 4803/1000000 [19:56<68:19:02, 4.05it/s, grad_norm=0.377, loss_final=0.49, loss_mean=0.819, loss_mean_cls=0.0821, proj_loss=-0.411][2026-03-26 12:50:59] Step: 4803, Training Logs: loss_final: 0.504812, loss_mean: 0.829739, proj_loss: -0.407415, loss_mean_cls: 0.082488, grad_norm: 0.282143 +Steps: 0%| | 4804/1000000 [19:56<68:20:15, 4.05it/s, grad_norm=0.282, loss_final=0.505, loss_mean=0.83, loss_mean_cls=0.0825, proj_loss=-0.407][2026-03-26 12:51:00] Step: 4804, Training Logs: loss_final: 0.503816, loss_mean: 0.826980, proj_loss: -0.405681, loss_mean_cls: 0.082518, grad_norm: 0.379658 +Steps: 0%| | 4805/1000000 [19:57<68:21:11, 4.04it/s, grad_norm=0.38, loss_final=0.504, loss_mean=0.827, loss_mean_cls=0.0825, proj_loss=-0.406][2026-03-26 12:51:00] Step: 4805, Training Logs: loss_final: 0.498196, loss_mean: 0.830329, proj_loss: -0.411726, loss_mean_cls: 0.079593, grad_norm: 0.446844 +Steps: 0%| | 4806/1000000 [19:57<68:24:53, 4.04it/s, grad_norm=0.447, loss_final=0.498, loss_mean=0.83, loss_mean_cls=0.0796, proj_loss=-0.412][2026-03-26 12:51:00] Step: 4806, Training Logs: loss_final: 0.508287, loss_mean: 0.847951, proj_loss: -0.418199, loss_mean_cls: 0.078536, grad_norm: 0.507582 +Steps: 0%| | 4807/1000000 [19:57<68:23:19, 4.04it/s, grad_norm=0.508, loss_final=0.508, loss_mean=0.848, loss_mean_cls=0.0785, proj_loss=-0.418][2026-03-26 12:51:00] Step: 4807, Training Logs: loss_final: 0.489750, loss_mean: 0.821172, proj_loss: -0.411247, loss_mean_cls: 0.079826, grad_norm: 0.372515 +Steps: 0%| | 4808/1000000 [19:57<68:24:59, 4.04it/s, grad_norm=0.373, loss_final=0.49, loss_mean=0.821, loss_mean_cls=0.0798, proj_loss=-0.411][2026-03-26 12:51:01] Step: 4808, Training Logs: loss_final: 0.488189, loss_mean: 0.817063, proj_loss: -0.409832, loss_mean_cls: 0.080958, grad_norm: 0.384617 +Steps: 0%| | 4809/1000000 [19:58<68:25:25, 4.04it/s, grad_norm=0.385, loss_final=0.488, loss_mean=0.817, loss_mean_cls=0.081, proj_loss=-0.41][2026-03-26 12:51:01] Step: 4809, Training Logs: loss_final: 0.488173, loss_mean: 0.819995, proj_loss: -0.412218, loss_mean_cls: 0.080397, grad_norm: 0.553944 +Steps: 0%| | 4810/1000000 [19:58<68:27:35, 4.04it/s, grad_norm=0.554, loss_final=0.488, loss_mean=0.82, loss_mean_cls=0.0804, proj_loss=-0.412][2026-03-26 12:51:01] Step: 4810, Training Logs: loss_final: 0.479106, loss_mean: 0.802674, proj_loss: -0.407218, loss_mean_cls: 0.083650, grad_norm: 0.458682 +Steps: 0%| | 4811/1000000 [19:58<68:31:51, 4.03it/s, grad_norm=0.459, loss_final=0.479, loss_mean=0.803, loss_mean_cls=0.0837, proj_loss=-0.407][2026-03-26 12:51:01] Step: 4811, Training Logs: loss_final: 0.497155, loss_mean: 0.822735, proj_loss: -0.407198, loss_mean_cls: 0.081618, grad_norm: 0.274798 +Steps: 0%| | 4812/1000000 [19:58<69:01:54, 4.00it/s, grad_norm=0.275, loss_final=0.497, loss_mean=0.823, loss_mean_cls=0.0816, proj_loss=-0.407][2026-03-26 12:51:02] Step: 4812, Training Logs: loss_final: 0.496192, loss_mean: 0.818147, proj_loss: -0.403300, loss_mean_cls: 0.081346, grad_norm: 0.310127 +Steps: 0%| | 4813/1000000 [19:59<68:52:54, 4.01it/s, grad_norm=0.31, loss_final=0.496, loss_mean=0.818, loss_mean_cls=0.0813, proj_loss=-0.403][2026-03-26 12:51:02] Step: 4813, Training Logs: loss_final: 0.485240, loss_mean: 0.806794, proj_loss: -0.404985, loss_mean_cls: 0.083432, grad_norm: 0.574688 +Steps: 0%| | 4814/1000000 [19:59<68:45:04, 4.02it/s, grad_norm=0.575, loss_final=0.485, loss_mean=0.807, loss_mean_cls=0.0834, proj_loss=-0.405][2026-03-26 12:51:02] Step: 4814, Training Logs: loss_final: 0.494227, loss_mean: 0.819530, proj_loss: -0.407293, loss_mean_cls: 0.081989, grad_norm: 0.720346 +Steps: 0%| | 4815/1000000 [19:59<68:41:53, 4.02it/s, grad_norm=0.72, loss_final=0.494, loss_mean=0.82, loss_mean_cls=0.082, proj_loss=-0.407][2026-03-26 12:51:02] Step: 4815, Training Logs: loss_final: 0.496788, loss_mean: 0.830425, proj_loss: -0.414628, loss_mean_cls: 0.080992, grad_norm: 0.353844 +Steps: 0%| | 4816/1000000 [19:59<68:36:37, 4.03it/s, grad_norm=0.354, loss_final=0.497, loss_mean=0.83, loss_mean_cls=0.081, proj_loss=-0.415][2026-03-26 12:51:03] Step: 4816, Training Logs: loss_final: 0.498013, loss_mean: 0.845797, proj_loss: -0.424287, loss_mean_cls: 0.076503, grad_norm: 0.744501 +Steps: 0%| | 4817/1000000 [20:00<68:34:53, 4.03it/s, grad_norm=0.745, loss_final=0.498, loss_mean=0.846, loss_mean_cls=0.0765, proj_loss=-0.424][2026-03-26 12:51:03] Step: 4817, Training Logs: loss_final: 0.484314, loss_mean: 0.829064, proj_loss: -0.424100, loss_mean_cls: 0.079350, grad_norm: 0.722518 +Steps: 0%| | 4818/1000000 [20:00<68:30:11, 4.04it/s, grad_norm=0.723, loss_final=0.484, loss_mean=0.829, loss_mean_cls=0.0793, proj_loss=-0.424][2026-03-26 12:51:03] Step: 4818, Training Logs: loss_final: 0.505731, loss_mean: 0.827258, proj_loss: -0.401352, loss_mean_cls: 0.079824, grad_norm: 0.342017 +Steps: 0%| | 4819/1000000 [20:00<68:30:57, 4.03it/s, grad_norm=0.342, loss_final=0.506, loss_mean=0.827, loss_mean_cls=0.0798, proj_loss=-0.401][2026-03-26 12:51:03] Step: 4819, Training Logs: loss_final: 0.488249, loss_mean: 0.822751, proj_loss: -0.413062, loss_mean_cls: 0.078560, grad_norm: 0.643653 +Steps: 0%| | 4820/1000000 [20:00<68:30:52, 4.03it/s, grad_norm=0.644, loss_final=0.488, loss_mean=0.823, loss_mean_cls=0.0786, proj_loss=-0.413][2026-03-26 12:51:04] Step: 4820, Training Logs: loss_final: 0.493854, loss_mean: 0.824138, proj_loss: -0.411905, loss_mean_cls: 0.081621, grad_norm: 0.379400 +Steps: 0%| | 4821/1000000 [20:01<68:27:28, 4.04it/s, grad_norm=0.379, loss_final=0.494, loss_mean=0.824, loss_mean_cls=0.0816, proj_loss=-0.412][2026-03-26 12:51:04] Step: 4821, Training Logs: loss_final: 0.498191, loss_mean: 0.822243, proj_loss: -0.406512, loss_mean_cls: 0.082459, grad_norm: 0.801525 +Steps: 0%| | 4822/1000000 [20:01<68:24:51, 4.04it/s, grad_norm=0.802, loss_final=0.498, loss_mean=0.822, loss_mean_cls=0.0825, proj_loss=-0.407][2026-03-26 12:51:04] Step: 4822, Training Logs: loss_final: 0.504697, loss_mean: 0.825127, proj_loss: -0.402733, loss_mean_cls: 0.082304, grad_norm: 0.461565 +Steps: 0%| | 4823/1000000 [20:01<68:24:12, 4.04it/s, grad_norm=0.462, loss_final=0.505, loss_mean=0.825, loss_mean_cls=0.0823, proj_loss=-0.403][2026-03-26 12:51:04] Step: 4823, Training Logs: loss_final: 0.523415, loss_mean: 0.856471, proj_loss: -0.411447, loss_mean_cls: 0.078390, grad_norm: 0.777516 +Steps: 0%| | 4824/1000000 [20:01<68:23:12, 4.04it/s, grad_norm=0.778, loss_final=0.523, loss_mean=0.856, loss_mean_cls=0.0784, proj_loss=-0.411][2026-03-26 12:51:05] Step: 4824, Training Logs: loss_final: 0.504671, loss_mean: 0.848782, proj_loss: -0.421041, loss_mean_cls: 0.076929, grad_norm: 0.775162 +Steps: 0%| | 4825/1000000 [20:02<68:24:56, 4.04it/s, grad_norm=0.775, loss_final=0.505, loss_mean=0.849, loss_mean_cls=0.0769, proj_loss=-0.421][2026-03-26 12:51:05] Step: 4825, Training Logs: loss_final: 0.500456, loss_mean: 0.833767, proj_loss: -0.413798, loss_mean_cls: 0.080488, grad_norm: 0.337655 +Steps: 0%| | 4826/1000000 [20:02<68:22:41, 4.04it/s, grad_norm=0.338, loss_final=0.5, loss_mean=0.834, loss_mean_cls=0.0805, proj_loss=-0.414][2026-03-26 12:51:05] Step: 4826, Training Logs: loss_final: 0.512375, loss_mean: 0.840459, proj_loss: -0.408416, loss_mean_cls: 0.080332, grad_norm: 0.659959 +Steps: 0%| | 4827/1000000 [20:02<68:23:48, 4.04it/s, grad_norm=0.66, loss_final=0.512, loss_mean=0.84, loss_mean_cls=0.0803, proj_loss=-0.408][2026-03-26 12:51:05] Step: 4827, Training Logs: loss_final: 0.492371, loss_mean: 0.820664, proj_loss: -0.409887, loss_mean_cls: 0.081594, grad_norm: 0.260733 +Steps: 0%| | 4828/1000000 [20:02<68:26:10, 4.04it/s, grad_norm=0.261, loss_final=0.492, loss_mean=0.821, loss_mean_cls=0.0816, proj_loss=-0.41][2026-03-26 12:51:06] Step: 4828, Training Logs: loss_final: 0.503314, loss_mean: 0.841365, proj_loss: -0.416223, loss_mean_cls: 0.078171, grad_norm: 0.778881 +Steps: 0%| | 4829/1000000 [20:03<68:24:07, 4.04it/s, grad_norm=0.779, loss_final=0.503, loss_mean=0.841, loss_mean_cls=0.0782, proj_loss=-0.416][2026-03-26 12:51:06] Step: 4829, Training Logs: loss_final: 0.514949, loss_mean: 0.858721, proj_loss: -0.421016, loss_mean_cls: 0.077244, grad_norm: 0.516242 +Steps: 0%| | 4830/1000000 [20:03<68:25:37, 4.04it/s, grad_norm=0.516, loss_final=0.515, loss_mean=0.859, loss_mean_cls=0.0772, proj_loss=-0.421][2026-03-26 12:51:06] Step: 4830, Training Logs: loss_final: 0.494730, loss_mean: 0.822957, proj_loss: -0.409542, loss_mean_cls: 0.081315, grad_norm: 0.614488 +Steps: 0%| | 4831/1000000 [20:03<68:25:54, 4.04it/s, grad_norm=0.614, loss_final=0.495, loss_mean=0.823, loss_mean_cls=0.0813, proj_loss=-0.41][2026-03-26 12:51:06] Step: 4831, Training Logs: loss_final: 0.511175, loss_mean: 0.836545, proj_loss: -0.407517, loss_mean_cls: 0.082147, grad_norm: 0.663141 +Steps: 0%| | 4832/1000000 [20:03<68:29:37, 4.04it/s, grad_norm=0.663, loss_final=0.511, loss_mean=0.837, loss_mean_cls=0.0821, proj_loss=-0.408][2026-03-26 12:51:07] Step: 4832, Training Logs: loss_final: 0.493276, loss_mean: 0.820281, proj_loss: -0.408561, loss_mean_cls: 0.081556, grad_norm: 0.343508 +Steps: 0%| | 4833/1000000 [20:04<68:28:04, 4.04it/s, grad_norm=0.344, loss_final=0.493, loss_mean=0.82, loss_mean_cls=0.0816, proj_loss=-0.409][2026-03-26 12:51:07] Step: 4833, Training Logs: loss_final: 0.496946, loss_mean: 0.830783, proj_loss: -0.414189, loss_mean_cls: 0.080352, grad_norm: 0.818122 +Steps: 0%| | 4834/1000000 [20:04<68:26:39, 4.04it/s, grad_norm=0.818, loss_final=0.497, loss_mean=0.831, loss_mean_cls=0.0804, proj_loss=-0.414][2026-03-26 12:51:07] Step: 4834, Training Logs: loss_final: 0.489717, loss_mean: 0.802535, proj_loss: -0.396941, loss_mean_cls: 0.084123, grad_norm: 0.302676 +Steps: 0%| | 4835/1000000 [20:04<68:26:47, 4.04it/s, grad_norm=0.303, loss_final=0.49, loss_mean=0.803, loss_mean_cls=0.0841, proj_loss=-0.397][2026-03-26 12:51:07] Step: 4835, Training Logs: loss_final: 0.480190, loss_mean: 0.807595, proj_loss: -0.409110, loss_mean_cls: 0.081705, grad_norm: 0.481783 +Steps: 0%| | 4836/1000000 [20:04<68:29:44, 4.04it/s, grad_norm=0.482, loss_final=0.48, loss_mean=0.808, loss_mean_cls=0.0817, proj_loss=-0.409][2026-03-26 12:51:08] Step: 4836, Training Logs: loss_final: 0.479353, loss_mean: 0.819325, proj_loss: -0.419202, loss_mean_cls: 0.079231, grad_norm: 0.252532 +Steps: 0%| | 4837/1000000 [20:05<68:24:37, 4.04it/s, grad_norm=0.253, loss_final=0.479, loss_mean=0.819, loss_mean_cls=0.0792, proj_loss=-0.419][2026-03-26 12:51:08] Step: 4837, Training Logs: loss_final: 0.485365, loss_mean: 0.816356, proj_loss: -0.412077, loss_mean_cls: 0.081087, grad_norm: 0.269801 +Steps: 0%| | 4838/1000000 [20:05<68:25:10, 4.04it/s, grad_norm=0.27, loss_final=0.485, loss_mean=0.816, loss_mean_cls=0.0811, proj_loss=-0.412][2026-03-26 12:51:08] Step: 4838, Training Logs: loss_final: 0.487058, loss_mean: 0.830120, proj_loss: -0.419517, loss_mean_cls: 0.076455, grad_norm: 0.265502 +Steps: 0%| | 4839/1000000 [20:05<68:23:35, 4.04it/s, grad_norm=0.266, loss_final=0.487, loss_mean=0.83, loss_mean_cls=0.0765, proj_loss=-0.42][2026-03-26 12:51:08] Step: 4839, Training Logs: loss_final: 0.487349, loss_mean: 0.815409, proj_loss: -0.411194, loss_mean_cls: 0.083134, grad_norm: 0.282458 +Steps: 0%| | 4840/1000000 [20:05<68:24:33, 4.04it/s, grad_norm=0.282, loss_final=0.487, loss_mean=0.815, loss_mean_cls=0.0831, proj_loss=-0.411][2026-03-26 12:51:09] Step: 4840, Training Logs: loss_final: 0.486234, loss_mean: 0.819094, proj_loss: -0.413642, loss_mean_cls: 0.080781, grad_norm: 0.305074 +Steps: 0%| | 4841/1000000 [20:05<68:41:20, 4.02it/s, grad_norm=0.305, loss_final=0.486, loss_mean=0.819, loss_mean_cls=0.0808, proj_loss=-0.414][2026-03-26 12:51:09] Step: 4841, Training Logs: loss_final: 0.484287, loss_mean: 0.827843, proj_loss: -0.420599, loss_mean_cls: 0.077043, grad_norm: 0.297863 +Steps: 0%| | 4842/1000000 [20:06<68:39:00, 4.03it/s, grad_norm=0.298, loss_final=0.484, loss_mean=0.828, loss_mean_cls=0.077, proj_loss=-0.421][2026-03-26 12:51:09] Step: 4842, Training Logs: loss_final: 0.493482, loss_mean: 0.828820, proj_loss: -0.414812, loss_mean_cls: 0.079473, grad_norm: 0.248081 +Steps: 0%| | 4843/1000000 [20:06<68:33:58, 4.03it/s, grad_norm=0.248, loss_final=0.493, loss_mean=0.829, loss_mean_cls=0.0795, proj_loss=-0.415][2026-03-26 12:51:09] Step: 4843, Training Logs: loss_final: 0.497154, loss_mean: 0.822551, proj_loss: -0.405970, loss_mean_cls: 0.080573, grad_norm: 0.328005 +Steps: 0%| | 4844/1000000 [20:06<68:33:55, 4.03it/s, grad_norm=0.328, loss_final=0.497, loss_mean=0.823, loss_mean_cls=0.0806, proj_loss=-0.406][2026-03-26 12:51:10] Step: 4844, Training Logs: loss_final: 0.498994, loss_mean: 0.831061, proj_loss: -0.411960, loss_mean_cls: 0.079893, grad_norm: 0.256542 +Steps: 0%| | 4845/1000000 [20:06<68:27:52, 4.04it/s, grad_norm=0.257, loss_final=0.499, loss_mean=0.831, loss_mean_cls=0.0799, proj_loss=-0.412][2026-03-26 12:51:10] Step: 4845, Training Logs: loss_final: 0.504014, loss_mean: 0.841906, proj_loss: -0.415241, loss_mean_cls: 0.077350, grad_norm: 0.457540 +Steps: 0%| | 4846/1000000 [20:07<68:28:04, 4.04it/s, grad_norm=0.458, loss_final=0.504, loss_mean=0.842, loss_mean_cls=0.0773, proj_loss=-0.415][2026-03-26 12:51:10] Step: 4846, Training Logs: loss_final: 0.487478, loss_mean: 0.828026, proj_loss: -0.419905, loss_mean_cls: 0.079357, grad_norm: 0.601802 +Steps: 0%| | 4847/1000000 [20:07<68:28:15, 4.04it/s, grad_norm=0.602, loss_final=0.487, loss_mean=0.828, loss_mean_cls=0.0794, proj_loss=-0.42][2026-03-26 12:51:10] Step: 4847, Training Logs: loss_final: 0.488092, loss_mean: 0.831247, proj_loss: -0.420728, loss_mean_cls: 0.077573, grad_norm: 0.274720 +Steps: 0%| | 4848/1000000 [20:07<68:28:05, 4.04it/s, grad_norm=0.275, loss_final=0.488, loss_mean=0.831, loss_mean_cls=0.0776, proj_loss=-0.421][2026-03-26 12:51:11] Step: 4848, Training Logs: loss_final: 0.497448, loss_mean: 0.835593, proj_loss: -0.415803, loss_mean_cls: 0.077658, grad_norm: 0.473872 +Steps: 0%| | 4849/1000000 [20:07<68:28:07, 4.04it/s, grad_norm=0.474, loss_final=0.497, loss_mean=0.836, loss_mean_cls=0.0777, proj_loss=-0.416][2026-03-26 12:51:11] Step: 4849, Training Logs: loss_final: 0.497633, loss_mean: 0.829709, proj_loss: -0.412049, loss_mean_cls: 0.079972, grad_norm: 0.339098 +Steps: 0%| | 4850/1000000 [20:08<68:27:37, 4.04it/s, grad_norm=0.339, loss_final=0.498, loss_mean=0.83, loss_mean_cls=0.08, proj_loss=-0.412][2026-03-26 12:51:11] Step: 4850, Training Logs: loss_final: 0.493442, loss_mean: 0.820147, proj_loss: -0.408761, loss_mean_cls: 0.082056, grad_norm: 0.408149 +Steps: 0%| | 4851/1000000 [20:08<68:25:31, 4.04it/s, grad_norm=0.408, loss_final=0.493, loss_mean=0.82, loss_mean_cls=0.0821, proj_loss=-0.409][2026-03-26 12:51:11] Step: 4851, Training Logs: loss_final: 0.491825, loss_mean: 0.823205, proj_loss: -0.410454, loss_mean_cls: 0.079074, grad_norm: 0.343837 +Steps: 0%| | 4852/1000000 [20:08<68:24:59, 4.04it/s, grad_norm=0.344, loss_final=0.492, loss_mean=0.823, loss_mean_cls=0.0791, proj_loss=-0.41][2026-03-26 12:51:12] Step: 4852, Training Logs: loss_final: 0.503776, loss_mean: 0.838419, proj_loss: -0.413509, loss_mean_cls: 0.078866, grad_norm: 0.459964 +Steps: 0%| | 4853/1000000 [20:08<68:25:40, 4.04it/s, grad_norm=0.46, loss_final=0.504, loss_mean=0.838, loss_mean_cls=0.0789, proj_loss=-0.414][2026-03-26 12:51:12] Step: 4853, Training Logs: loss_final: 0.497491, loss_mean: 0.832251, proj_loss: -0.412839, loss_mean_cls: 0.078078, grad_norm: 0.501530 +Steps: 0%| | 4854/1000000 [20:09<68:26:49, 4.04it/s, grad_norm=0.502, loss_final=0.497, loss_mean=0.832, loss_mean_cls=0.0781, proj_loss=-0.413][2026-03-26 12:51:12] Step: 4854, Training Logs: loss_final: 0.485105, loss_mean: 0.830757, proj_loss: -0.422653, loss_mean_cls: 0.077001, grad_norm: 0.302801 +Steps: 0%| | 4855/1000000 [20:09<68:27:58, 4.04it/s, grad_norm=0.303, loss_final=0.485, loss_mean=0.831, loss_mean_cls=0.077, proj_loss=-0.423][2026-03-26 12:51:12] Step: 4855, Training Logs: loss_final: 0.499127, loss_mean: 0.835013, proj_loss: -0.414835, loss_mean_cls: 0.078949, grad_norm: 0.483353 +Steps: 0%| | 4856/1000000 [20:09<68:27:34, 4.04it/s, grad_norm=0.483, loss_final=0.499, loss_mean=0.835, loss_mean_cls=0.0789, proj_loss=-0.415][2026-03-26 12:51:13] Step: 4856, Training Logs: loss_final: 0.491689, loss_mean: 0.830220, proj_loss: -0.416409, loss_mean_cls: 0.077877, grad_norm: 0.411486 +Steps: 0%| | 4857/1000000 [20:09<68:27:10, 4.04it/s, grad_norm=0.411, loss_final=0.492, loss_mean=0.83, loss_mean_cls=0.0779, proj_loss=-0.416][2026-03-26 12:51:13] Step: 4857, Training Logs: loss_final: 0.505069, loss_mean: 0.824568, proj_loss: -0.402516, loss_mean_cls: 0.083017, grad_norm: 0.463741 +Steps: 0%| | 4858/1000000 [20:10<68:26:57, 4.04it/s, grad_norm=0.464, loss_final=0.505, loss_mean=0.825, loss_mean_cls=0.083, proj_loss=-0.403][2026-03-26 12:51:13] Step: 4858, Training Logs: loss_final: 0.496642, loss_mean: 0.828759, proj_loss: -0.412518, loss_mean_cls: 0.080401, grad_norm: 0.562434 +Steps: 0%| | 4859/1000000 [20:10<68:25:08, 4.04it/s, grad_norm=0.562, loss_final=0.497, loss_mean=0.829, loss_mean_cls=0.0804, proj_loss=-0.413][2026-03-26 12:51:13] Step: 4859, Training Logs: loss_final: 0.499065, loss_mean: 0.833799, proj_loss: -0.412599, loss_mean_cls: 0.077866, grad_norm: 0.383930 +Steps: 0%| | 4860/1000000 [20:10<69:09:28, 4.00it/s, grad_norm=0.384, loss_final=0.499, loss_mean=0.834, loss_mean_cls=0.0779, proj_loss=-0.413][2026-03-26 12:51:14] Step: 4860, Training Logs: loss_final: 0.484547, loss_mean: 0.817451, proj_loss: -0.412835, loss_mean_cls: 0.079932, grad_norm: 0.372877 +Steps: 0%| | 4861/1000000 [20:10<69:01:51, 4.00it/s, grad_norm=0.373, loss_final=0.485, loss_mean=0.817, loss_mean_cls=0.0799, proj_loss=-0.413][2026-03-26 12:51:14] Step: 4861, Training Logs: loss_final: 0.504666, loss_mean: 0.846599, proj_loss: -0.419100, loss_mean_cls: 0.077167, grad_norm: 0.399582 +Steps: 0%| | 4862/1000000 [20:11<68:51:41, 4.01it/s, grad_norm=0.4, loss_final=0.505, loss_mean=0.847, loss_mean_cls=0.0772, proj_loss=-0.419][2026-03-26 12:51:14] Step: 4862, Training Logs: loss_final: 0.491947, loss_mean: 0.826817, proj_loss: -0.413523, loss_mean_cls: 0.078653, grad_norm: 0.278395 +Steps: 0%| | 4863/1000000 [20:11<68:47:37, 4.02it/s, grad_norm=0.278, loss_final=0.492, loss_mean=0.827, loss_mean_cls=0.0787, proj_loss=-0.414][2026-03-26 12:51:14] Step: 4863, Training Logs: loss_final: 0.493350, loss_mean: 0.824317, proj_loss: -0.410607, loss_mean_cls: 0.079640, grad_norm: 0.419815 +Steps: 0%| | 4864/1000000 [20:11<68:39:02, 4.03it/s, grad_norm=0.42, loss_final=0.493, loss_mean=0.824, loss_mean_cls=0.0796, proj_loss=-0.411][2026-03-26 12:51:15] Step: 4864, Training Logs: loss_final: 0.495870, loss_mean: 0.826183, proj_loss: -0.408979, loss_mean_cls: 0.078666, grad_norm: 0.276255 +Steps: 0%| | 4865/1000000 [20:11<68:34:46, 4.03it/s, grad_norm=0.276, loss_final=0.496, loss_mean=0.826, loss_mean_cls=0.0787, proj_loss=-0.409][2026-03-26 12:51:15] Step: 4865, Training Logs: loss_final: 0.495059, loss_mean: 0.832806, proj_loss: -0.416377, loss_mean_cls: 0.078629, grad_norm: 0.434574 +Steps: 0%| | 4866/1000000 [20:12<68:29:34, 4.04it/s, grad_norm=0.435, loss_final=0.495, loss_mean=0.833, loss_mean_cls=0.0786, proj_loss=-0.416][2026-03-26 12:51:15] Step: 4866, Training Logs: loss_final: 0.497480, loss_mean: 0.831262, proj_loss: -0.413267, loss_mean_cls: 0.079484, grad_norm: 0.395610 +Steps: 0%| | 4867/1000000 [20:12<68:29:56, 4.04it/s, grad_norm=0.396, loss_final=0.497, loss_mean=0.831, loss_mean_cls=0.0795, proj_loss=-0.413][2026-03-26 12:51:15] Step: 4867, Training Logs: loss_final: 0.496590, loss_mean: 0.825619, proj_loss: -0.408786, loss_mean_cls: 0.079757, grad_norm: 0.269501 +Steps: 0%| | 4868/1000000 [20:12<68:27:00, 4.04it/s, grad_norm=0.27, loss_final=0.497, loss_mean=0.826, loss_mean_cls=0.0798, proj_loss=-0.409][2026-03-26 12:51:16] Step: 4868, Training Logs: loss_final: 0.486545, loss_mean: 0.822382, proj_loss: -0.415420, loss_mean_cls: 0.079583, grad_norm: 0.578863 +Steps: 0%| | 4869/1000000 [20:12<68:25:09, 4.04it/s, grad_norm=0.579, loss_final=0.487, loss_mean=0.822, loss_mean_cls=0.0796, proj_loss=-0.415][2026-03-26 12:51:16] Step: 4869, Training Logs: loss_final: 0.495102, loss_mean: 0.841818, proj_loss: -0.422463, loss_mean_cls: 0.075747, grad_norm: 0.269332 +Steps: 0%| | 4870/1000000 [20:13<68:22:14, 4.04it/s, grad_norm=0.269, loss_final=0.495, loss_mean=0.842, loss_mean_cls=0.0757, proj_loss=-0.422][2026-03-26 12:51:16] Step: 4870, Training Logs: loss_final: 0.498621, loss_mean: 0.837370, proj_loss: -0.417489, loss_mean_cls: 0.078740, grad_norm: 0.340032 +Steps: 0%| | 4871/1000000 [20:13<68:21:10, 4.04it/s, grad_norm=0.34, loss_final=0.499, loss_mean=0.837, loss_mean_cls=0.0787, proj_loss=-0.417][2026-03-26 12:51:16] Step: 4871, Training Logs: loss_final: 0.499824, loss_mean: 0.832908, proj_loss: -0.414141, loss_mean_cls: 0.081057, grad_norm: 0.432616 +Steps: 0%| | 4872/1000000 [20:13<68:20:59, 4.04it/s, grad_norm=0.433, loss_final=0.5, loss_mean=0.833, loss_mean_cls=0.0811, proj_loss=-0.414][2026-03-26 12:51:17] Step: 4872, Training Logs: loss_final: 0.487395, loss_mean: 0.829350, proj_loss: -0.419819, loss_mean_cls: 0.077865, grad_norm: 0.304738 +Steps: 0%| | 4873/1000000 [20:13<68:22:35, 4.04it/s, grad_norm=0.305, loss_final=0.487, loss_mean=0.829, loss_mean_cls=0.0779, proj_loss=-0.42][2026-03-26 12:51:17] Step: 4873, Training Logs: loss_final: 0.500719, loss_mean: 0.830636, proj_loss: -0.409492, loss_mean_cls: 0.079576, grad_norm: 0.311944 +Steps: 0%| | 4874/1000000 [20:14<68:23:16, 4.04it/s, grad_norm=0.312, loss_final=0.501, loss_mean=0.831, loss_mean_cls=0.0796, proj_loss=-0.409][2026-03-26 12:51:17] Step: 4874, Training Logs: loss_final: 0.501256, loss_mean: 0.830510, proj_loss: -0.409602, loss_mean_cls: 0.080349, grad_norm: 0.373036 +Steps: 0%| | 4875/1000000 [20:14<68:24:41, 4.04it/s, grad_norm=0.373, loss_final=0.501, loss_mean=0.831, loss_mean_cls=0.0803, proj_loss=-0.41][2026-03-26 12:51:17] Step: 4875, Training Logs: loss_final: 0.507364, loss_mean: 0.834439, proj_loss: -0.407466, loss_mean_cls: 0.080391, grad_norm: 0.278263 +Steps: 0%| | 4876/1000000 [20:14<68:24:10, 4.04it/s, grad_norm=0.278, loss_final=0.507, loss_mean=0.834, loss_mean_cls=0.0804, proj_loss=-0.407][2026-03-26 12:51:18] Step: 4876, Training Logs: loss_final: 0.486907, loss_mean: 0.802947, proj_loss: -0.399063, loss_mean_cls: 0.083023, grad_norm: 0.522493 +Steps: 0%| | 4877/1000000 [20:14<68:25:40, 4.04it/s, grad_norm=0.522, loss_final=0.487, loss_mean=0.803, loss_mean_cls=0.083, proj_loss=-0.399][2026-03-26 12:51:18] Step: 4877, Training Logs: loss_final: 0.498466, loss_mean: 0.841930, proj_loss: -0.420320, loss_mean_cls: 0.076856, grad_norm: 0.327618 +Steps: 0%| | 4878/1000000 [20:15<68:29:50, 4.04it/s, grad_norm=0.328, loss_final=0.498, loss_mean=0.842, loss_mean_cls=0.0769, proj_loss=-0.42][2026-03-26 12:51:18] Step: 4878, Training Logs: loss_final: 0.502125, loss_mean: 0.825895, proj_loss: -0.404972, loss_mean_cls: 0.081202, grad_norm: 0.471018 +Steps: 0%| | 4879/1000000 [20:15<68:32:36, 4.03it/s, grad_norm=0.471, loss_final=0.502, loss_mean=0.826, loss_mean_cls=0.0812, proj_loss=-0.405][2026-03-26 12:51:18] Step: 4879, Training Logs: loss_final: 0.480876, loss_mean: 0.811958, proj_loss: -0.412260, loss_mean_cls: 0.081179, grad_norm: 0.531406 +Steps: 0%| | 4880/1000000 [20:15<68:33:14, 4.03it/s, grad_norm=0.531, loss_final=0.481, loss_mean=0.812, loss_mean_cls=0.0812, proj_loss=-0.412][2026-03-26 12:51:19] Step: 4880, Training Logs: loss_final: 0.501340, loss_mean: 0.831313, proj_loss: -0.409377, loss_mean_cls: 0.079405, grad_norm: 0.389996 +Steps: 0%| | 4881/1000000 [20:15<68:30:26, 4.03it/s, grad_norm=0.39, loss_final=0.501, loss_mean=0.831, loss_mean_cls=0.0794, proj_loss=-0.409][2026-03-26 12:51:19] Step: 4881, Training Logs: loss_final: 0.498360, loss_mean: 0.830159, proj_loss: -0.411993, loss_mean_cls: 0.080194, grad_norm: 0.492975 +Steps: 0%| | 4882/1000000 [20:16<68:30:09, 4.04it/s, grad_norm=0.493, loss_final=0.498, loss_mean=0.83, loss_mean_cls=0.0802, proj_loss=-0.412][2026-03-26 12:51:19] Step: 4882, Training Logs: loss_final: 0.500286, loss_mean: 0.832224, proj_loss: -0.411109, loss_mean_cls: 0.079171, grad_norm: 0.320612 +Steps: 0%| | 4883/1000000 [20:16<68:28:26, 4.04it/s, grad_norm=0.321, loss_final=0.5, loss_mean=0.832, loss_mean_cls=0.0792, proj_loss=-0.411][2026-03-26 12:51:19] Step: 4883, Training Logs: loss_final: 0.509717, loss_mean: 0.839612, proj_loss: -0.410234, loss_mean_cls: 0.080339, grad_norm: 0.448837 +Steps: 0%| | 4884/1000000 [20:16<68:29:02, 4.04it/s, grad_norm=0.449, loss_final=0.51, loss_mean=0.84, loss_mean_cls=0.0803, proj_loss=-0.41][2026-03-26 12:51:20] Step: 4884, Training Logs: loss_final: 0.494533, loss_mean: 0.815617, proj_loss: -0.404121, loss_mean_cls: 0.083037, grad_norm: 0.372231 +Steps: 0%| | 4885/1000000 [20:16<68:26:04, 4.04it/s, grad_norm=0.372, loss_final=0.495, loss_mean=0.816, loss_mean_cls=0.083, proj_loss=-0.404][2026-03-26 12:51:20] Step: 4885, Training Logs: loss_final: 0.500577, loss_mean: 0.830126, proj_loss: -0.410123, loss_mean_cls: 0.080573, grad_norm: 0.344271 +Steps: 0%| | 4886/1000000 [20:17<68:23:21, 4.04it/s, grad_norm=0.344, loss_final=0.501, loss_mean=0.83, loss_mean_cls=0.0806, proj_loss=-0.41][2026-03-26 12:51:20] Step: 4886, Training Logs: loss_final: 0.489666, loss_mean: 0.826391, proj_loss: -0.415371, loss_mean_cls: 0.078646, grad_norm: 0.433602 +Steps: 0%| | 4887/1000000 [20:17<68:23:27, 4.04it/s, grad_norm=0.434, loss_final=0.49, loss_mean=0.826, loss_mean_cls=0.0786, proj_loss=-0.415][2026-03-26 12:51:20] Step: 4887, Training Logs: loss_final: 0.488915, loss_mean: 0.811084, proj_loss: -0.404799, loss_mean_cls: 0.082630, grad_norm: 0.318885 +Steps: 0%| | 4888/1000000 [20:17<68:25:07, 4.04it/s, grad_norm=0.319, loss_final=0.489, loss_mean=0.811, loss_mean_cls=0.0826, proj_loss=-0.405][2026-03-26 12:51:21] Step: 4888, Training Logs: loss_final: 0.495729, loss_mean: 0.842778, proj_loss: -0.423643, loss_mean_cls: 0.076594, grad_norm: 0.384798 +Steps: 0%| | 4889/1000000 [20:17<68:22:33, 4.04it/s, grad_norm=0.385, loss_final=0.496, loss_mean=0.843, loss_mean_cls=0.0766, proj_loss=-0.424][2026-03-26 12:51:21] Step: 4889, Training Logs: loss_final: 0.509043, loss_mean: 0.843368, proj_loss: -0.413275, loss_mean_cls: 0.078951, grad_norm: 0.595011 +Steps: 0%| | 4890/1000000 [20:18<68:23:50, 4.04it/s, grad_norm=0.595, loss_final=0.509, loss_mean=0.843, loss_mean_cls=0.079, proj_loss=-0.413][2026-03-26 12:51:21] Step: 4890, Training Logs: loss_final: 0.481478, loss_mean: 0.811684, proj_loss: -0.412314, loss_mean_cls: 0.082107, grad_norm: 0.404515 +Steps: 0%| | 4891/1000000 [20:18<68:25:38, 4.04it/s, grad_norm=0.405, loss_final=0.481, loss_mean=0.812, loss_mean_cls=0.0821, proj_loss=-0.412][2026-03-26 12:51:21] Step: 4891, Training Logs: loss_final: 0.488975, loss_mean: 0.819012, proj_loss: -0.410711, loss_mean_cls: 0.080675, grad_norm: 0.451161 +Steps: 0%| | 4892/1000000 [20:18<68:26:01, 4.04it/s, grad_norm=0.451, loss_final=0.489, loss_mean=0.819, loss_mean_cls=0.0807, proj_loss=-0.411][2026-03-26 12:51:22] Step: 4892, Training Logs: loss_final: 0.498840, loss_mean: 0.818477, proj_loss: -0.402381, loss_mean_cls: 0.082744, grad_norm: 0.357983 +Steps: 0%| | 4893/1000000 [20:18<68:31:01, 4.03it/s, grad_norm=0.358, loss_final=0.499, loss_mean=0.818, loss_mean_cls=0.0827, proj_loss=-0.402][2026-03-26 12:51:22] Step: 4893, Training Logs: loss_final: 0.492934, loss_mean: 0.826177, proj_loss: -0.413414, loss_mean_cls: 0.080171, grad_norm: 0.501925 +Steps: 0%| | 4894/1000000 [20:19<68:28:37, 4.04it/s, grad_norm=0.502, loss_final=0.493, loss_mean=0.826, loss_mean_cls=0.0802, proj_loss=-0.413][2026-03-26 12:51:22] Step: 4894, Training Logs: loss_final: 0.477309, loss_mean: 0.811850, proj_loss: -0.414126, loss_mean_cls: 0.079584, grad_norm: 0.334020 +Steps: 0%| | 4895/1000000 [20:19<68:30:05, 4.04it/s, grad_norm=0.334, loss_final=0.477, loss_mean=0.812, loss_mean_cls=0.0796, proj_loss=-0.414][2026-03-26 12:51:22] Step: 4895, Training Logs: loss_final: 0.507787, loss_mean: 0.842699, proj_loss: -0.412607, loss_mean_cls: 0.077695, grad_norm: 0.271489 +Steps: 0%| | 4896/1000000 [20:19<68:30:36, 4.03it/s, grad_norm=0.271, loss_final=0.508, loss_mean=0.843, loss_mean_cls=0.0777, proj_loss=-0.413][2026-03-26 12:51:23] Step: 4896, Training Logs: loss_final: 0.517251, loss_mean: 0.851994, proj_loss: -0.414770, loss_mean_cls: 0.080027, grad_norm: 0.508965 +Steps: 0%| | 4897/1000000 [20:19<68:29:07, 4.04it/s, grad_norm=0.509, loss_final=0.517, loss_mean=0.852, loss_mean_cls=0.08, proj_loss=-0.415][2026-03-26 12:51:23] Step: 4897, Training Logs: loss_final: 0.488348, loss_mean: 0.830157, proj_loss: -0.420552, loss_mean_cls: 0.078743, grad_norm: 0.272445 +Steps: 0%| | 4898/1000000 [20:20<68:27:53, 4.04it/s, grad_norm=0.272, loss_final=0.488, loss_mean=0.83, loss_mean_cls=0.0787, proj_loss=-0.421][2026-03-26 12:51:23] Step: 4898, Training Logs: loss_final: 0.498270, loss_mean: 0.838043, proj_loss: -0.417920, loss_mean_cls: 0.078147, grad_norm: 0.371656 +Steps: 0%| | 4899/1000000 [20:20<68:23:43, 4.04it/s, grad_norm=0.372, loss_final=0.498, loss_mean=0.838, loss_mean_cls=0.0781, proj_loss=-0.418][2026-03-26 12:51:23] Step: 4899, Training Logs: loss_final: 0.506346, loss_mean: 0.855342, proj_loss: -0.424643, loss_mean_cls: 0.075647, grad_norm: 0.299560 +Steps: 0%| | 4900/1000000 [20:20<68:22:47, 4.04it/s, grad_norm=0.3, loss_final=0.506, loss_mean=0.855, loss_mean_cls=0.0756, proj_loss=-0.425][2026-03-26 12:51:24] Step: 4900, Training Logs: loss_final: 0.499737, loss_mean: 0.835614, proj_loss: -0.413638, loss_mean_cls: 0.077761, grad_norm: 0.312870 +Steps: 0%| | 4901/1000000 [20:20<68:22:26, 4.04it/s, grad_norm=0.313, loss_final=0.5, loss_mean=0.836, loss_mean_cls=0.0778, proj_loss=-0.414][2026-03-26 12:51:24] Step: 4901, Training Logs: loss_final: 0.494636, loss_mean: 0.824957, proj_loss: -0.409631, loss_mean_cls: 0.079311, grad_norm: 0.367542 +Steps: 0%| | 4902/1000000 [20:21<68:23:27, 4.04it/s, grad_norm=0.368, loss_final=0.495, loss_mean=0.825, loss_mean_cls=0.0793, proj_loss=-0.41][2026-03-26 12:51:24] Step: 4902, Training Logs: loss_final: 0.486795, loss_mean: 0.819417, proj_loss: -0.412954, loss_mean_cls: 0.080333, grad_norm: 0.309196 +Steps: 0%| | 4903/1000000 [20:21<68:25:09, 4.04it/s, grad_norm=0.309, loss_final=0.487, loss_mean=0.819, loss_mean_cls=0.0803, proj_loss=-0.413][2026-03-26 12:51:24] Step: 4903, Training Logs: loss_final: 0.496694, loss_mean: 0.820935, proj_loss: -0.405529, loss_mean_cls: 0.081288, grad_norm: 0.370777 +Steps: 0%| | 4904/1000000 [20:21<68:25:25, 4.04it/s, grad_norm=0.371, loss_final=0.497, loss_mean=0.821, loss_mean_cls=0.0813, proj_loss=-0.406][2026-03-26 12:51:24] Step: 4904, Training Logs: loss_final: 0.498448, loss_mean: 0.842711, proj_loss: -0.421325, loss_mean_cls: 0.077061, grad_norm: 0.245691 +Steps: 0%| | 4905/1000000 [20:21<68:25:23, 4.04it/s, grad_norm=0.246, loss_final=0.498, loss_mean=0.843, loss_mean_cls=0.0771, proj_loss=-0.421][2026-03-26 12:51:25] Step: 4905, Training Logs: loss_final: 0.503991, loss_mean: 0.836625, proj_loss: -0.410643, loss_mean_cls: 0.078009, grad_norm: 0.227796 +Steps: 0%| | 4906/1000000 [20:22<68:23:57, 4.04it/s, grad_norm=0.228, loss_final=0.504, loss_mean=0.837, loss_mean_cls=0.078, proj_loss=-0.411][2026-03-26 12:51:25] Step: 4906, Training Logs: loss_final: 0.496390, loss_mean: 0.839472, proj_loss: -0.419445, loss_mean_cls: 0.076362, grad_norm: 0.304842 +Steps: 0%| | 4907/1000000 [20:22<68:25:04, 4.04it/s, grad_norm=0.305, loss_final=0.496, loss_mean=0.839, loss_mean_cls=0.0764, proj_loss=-0.419][2026-03-26 12:51:25] Step: 4907, Training Logs: loss_final: 0.499113, loss_mean: 0.836827, proj_loss: -0.415953, loss_mean_cls: 0.078239, grad_norm: 0.358404 +Steps: 0%| | 4908/1000000 [20:22<68:23:21, 4.04it/s, grad_norm=0.358, loss_final=0.499, loss_mean=0.837, loss_mean_cls=0.0782, proj_loss=-0.416][2026-03-26 12:51:25] Step: 4908, Training Logs: loss_final: 0.489074, loss_mean: 0.824050, proj_loss: -0.413811, loss_mean_cls: 0.078835, grad_norm: 0.408614 +Steps: 0%| | 4909/1000000 [20:22<68:24:36, 4.04it/s, grad_norm=0.409, loss_final=0.489, loss_mean=0.824, loss_mean_cls=0.0788, proj_loss=-0.414][2026-03-26 12:51:26] Step: 4909, Training Logs: loss_final: 0.496860, loss_mean: 0.829077, proj_loss: -0.412395, loss_mean_cls: 0.080178, grad_norm: 0.583009 +Steps: 0%| | 4910/1000000 [20:23<68:22:46, 4.04it/s, grad_norm=0.583, loss_final=0.497, loss_mean=0.829, loss_mean_cls=0.0802, proj_loss=-0.412][2026-03-26 12:51:26] Step: 4910, Training Logs: loss_final: 0.492965, loss_mean: 0.822509, proj_loss: -0.409774, loss_mean_cls: 0.080231, grad_norm: 0.262409 +Steps: 0%| | 4911/1000000 [20:23<68:23:48, 4.04it/s, grad_norm=0.262, loss_final=0.493, loss_mean=0.823, loss_mean_cls=0.0802, proj_loss=-0.41][2026-03-26 12:51:26] Step: 4911, Training Logs: loss_final: 0.495736, loss_mean: 0.829195, proj_loss: -0.413134, loss_mean_cls: 0.079675, grad_norm: 0.652846 +Steps: 0%| | 4912/1000000 [20:23<68:25:25, 4.04it/s, grad_norm=0.653, loss_final=0.496, loss_mean=0.829, loss_mean_cls=0.0797, proj_loss=-0.413][2026-03-26 12:51:26] Step: 4912, Training Logs: loss_final: 0.497965, loss_mean: 0.811405, proj_loss: -0.397822, loss_mean_cls: 0.084382, grad_norm: 0.638809 +Steps: 0%| | 4913/1000000 [20:23<68:26:10, 4.04it/s, grad_norm=0.639, loss_final=0.498, loss_mean=0.811, loss_mean_cls=0.0844, proj_loss=-0.398][2026-03-26 12:51:27] Step: 4913, Training Logs: loss_final: 0.504275, loss_mean: 0.820762, proj_loss: -0.397885, loss_mean_cls: 0.081398, grad_norm: 0.418238 +Steps: 0%| | 4914/1000000 [20:24<68:26:39, 4.04it/s, grad_norm=0.418, loss_final=0.504, loss_mean=0.821, loss_mean_cls=0.0814, proj_loss=-0.398][2026-03-26 12:51:27] Step: 4914, Training Logs: loss_final: 0.491855, loss_mean: 0.823519, proj_loss: -0.412111, loss_mean_cls: 0.080447, grad_norm: 0.640587 +Steps: 0%| | 4915/1000000 [20:24<68:27:43, 4.04it/s, grad_norm=0.641, loss_final=0.492, loss_mean=0.824, loss_mean_cls=0.0804, proj_loss=-0.412][2026-03-26 12:51:27] Step: 4915, Training Logs: loss_final: 0.486923, loss_mean: 0.821684, proj_loss: -0.414391, loss_mean_cls: 0.079630, grad_norm: 0.407002 +Steps: 0%| | 4916/1000000 [20:24<68:27:19, 4.04it/s, grad_norm=0.407, loss_final=0.487, loss_mean=0.822, loss_mean_cls=0.0796, proj_loss=-0.414][2026-03-26 12:51:27] Step: 4916, Training Logs: loss_final: 0.491526, loss_mean: 0.820320, proj_loss: -0.409253, loss_mean_cls: 0.080459, grad_norm: 0.819753 +Steps: 0%| | 4917/1000000 [20:24<68:26:36, 4.04it/s, grad_norm=0.82, loss_final=0.492, loss_mean=0.82, loss_mean_cls=0.0805, proj_loss=-0.409][2026-03-26 12:51:28] Step: 4917, Training Logs: loss_final: 0.491506, loss_mean: 0.817354, proj_loss: -0.406228, loss_mean_cls: 0.080380, grad_norm: 0.312672 +Steps: 0%| | 4918/1000000 [20:25<68:26:13, 4.04it/s, grad_norm=0.313, loss_final=0.492, loss_mean=0.817, loss_mean_cls=0.0804, proj_loss=-0.406][2026-03-26 12:51:28] Step: 4918, Training Logs: loss_final: 0.509810, loss_mean: 0.837829, proj_loss: -0.408433, loss_mean_cls: 0.080414, grad_norm: 0.733829 +Steps: 0%| | 4919/1000000 [20:25<68:26:52, 4.04it/s, grad_norm=0.734, loss_final=0.51, loss_mean=0.838, loss_mean_cls=0.0804, proj_loss=-0.408][2026-03-26 12:51:28] Step: 4919, Training Logs: loss_final: 0.497935, loss_mean: 0.833606, proj_loss: -0.414586, loss_mean_cls: 0.078915, grad_norm: 0.710876 +Steps: 0%| | 4920/1000000 [20:25<68:29:22, 4.04it/s, grad_norm=0.711, loss_final=0.498, loss_mean=0.834, loss_mean_cls=0.0789, proj_loss=-0.415][2026-03-26 12:51:28] Step: 4920, Training Logs: loss_final: 0.480616, loss_mean: 0.808073, proj_loss: -0.408267, loss_mean_cls: 0.080809, grad_norm: 0.397321 +Steps: 0%| | 4921/1000000 [20:25<68:26:38, 4.04it/s, grad_norm=0.397, loss_final=0.481, loss_mean=0.808, loss_mean_cls=0.0808, proj_loss=-0.408][2026-03-26 12:51:29] Step: 4921, Training Logs: loss_final: 0.487313, loss_mean: 0.832590, proj_loss: -0.422617, loss_mean_cls: 0.077339, grad_norm: 0.774141 +Steps: 0%| | 4922/1000000 [20:26<68:27:56, 4.04it/s, grad_norm=0.774, loss_final=0.487, loss_mean=0.833, loss_mean_cls=0.0773, proj_loss=-0.423][2026-03-26 12:51:29] Step: 4922, Training Logs: loss_final: 0.493147, loss_mean: 0.831388, proj_loss: -0.414991, loss_mean_cls: 0.076749, grad_norm: 0.450836 +Steps: 0%| | 4923/1000000 [20:26<68:26:48, 4.04it/s, grad_norm=0.451, loss_final=0.493, loss_mean=0.831, loss_mean_cls=0.0767, proj_loss=-0.415][2026-03-26 12:51:29] Step: 4923, Training Logs: loss_final: 0.501767, loss_mean: 0.851192, proj_loss: -0.425059, loss_mean_cls: 0.075635, grad_norm: 0.519146 +Steps: 0%| | 4924/1000000 [20:26<68:26:59, 4.04it/s, grad_norm=0.519, loss_final=0.502, loss_mean=0.851, loss_mean_cls=0.0756, proj_loss=-0.425][2026-03-26 12:51:29] Step: 4924, Training Logs: loss_final: 0.506465, loss_mean: 0.843098, proj_loss: -0.416313, loss_mean_cls: 0.079680, grad_norm: 0.389378 +Steps: 0%| | 4925/1000000 [20:26<68:23:59, 4.04it/s, grad_norm=0.389, loss_final=0.506, loss_mean=0.843, loss_mean_cls=0.0797, proj_loss=-0.416][2026-03-26 12:51:30] Step: 4925, Training Logs: loss_final: 0.507147, loss_mean: 0.845797, proj_loss: -0.416663, loss_mean_cls: 0.078012, grad_norm: 0.382526 +Steps: 0%| | 4926/1000000 [20:27<68:27:23, 4.04it/s, grad_norm=0.383, loss_final=0.507, loss_mean=0.846, loss_mean_cls=0.078, proj_loss=-0.417][2026-03-26 12:51:30] Step: 4926, Training Logs: loss_final: 0.506993, loss_mean: 0.832386, proj_loss: -0.405677, loss_mean_cls: 0.080285, grad_norm: 0.583622 +Steps: 0%| | 4927/1000000 [20:27<68:21:34, 4.04it/s, grad_norm=0.584, loss_final=0.507, loss_mean=0.832, loss_mean_cls=0.0803, proj_loss=-0.406][2026-03-26 12:51:30] Step: 4927, Training Logs: loss_final: 0.494497, loss_mean: 0.840890, proj_loss: -0.424013, loss_mean_cls: 0.077620, grad_norm: 0.314085 +Steps: 0%| | 4928/1000000 [20:27<68:23:20, 4.04it/s, grad_norm=0.314, loss_final=0.494, loss_mean=0.841, loss_mean_cls=0.0776, proj_loss=-0.424][2026-03-26 12:51:30] Step: 4928, Training Logs: loss_final: 0.518608, loss_mean: 0.857081, proj_loss: -0.416436, loss_mean_cls: 0.077963, grad_norm: 0.439097 +Steps: 0%| | 4929/1000000 [20:27<68:17:52, 4.05it/s, grad_norm=0.439, loss_final=0.519, loss_mean=0.857, loss_mean_cls=0.078, proj_loss=-0.416][2026-03-26 12:51:31] Step: 4929, Training Logs: loss_final: 0.497257, loss_mean: 0.826690, proj_loss: -0.409744, loss_mean_cls: 0.080311, grad_norm: 0.349069 +Steps: 0%| | 4930/1000000 [20:28<68:25:07, 4.04it/s, grad_norm=0.349, loss_final=0.497, loss_mean=0.827, loss_mean_cls=0.0803, proj_loss=-0.41][2026-03-26 12:51:31] Step: 4930, Training Logs: loss_final: 0.505640, loss_mean: 0.833001, proj_loss: -0.407409, loss_mean_cls: 0.080048, grad_norm: 0.352985 +Steps: 0%| | 4931/1000000 [20:28<68:51:27, 4.01it/s, grad_norm=0.353, loss_final=0.506, loss_mean=0.833, loss_mean_cls=0.08, proj_loss=-0.407][2026-03-26 12:51:31] Step: 4931, Training Logs: loss_final: 0.500176, loss_mean: 0.837571, proj_loss: -0.416140, loss_mean_cls: 0.078745, grad_norm: 0.701113 +Steps: 0%| | 4932/1000000 [20:28<68:45:41, 4.02it/s, grad_norm=0.701, loss_final=0.5, loss_mean=0.838, loss_mean_cls=0.0787, proj_loss=-0.416][2026-03-26 12:51:31] Step: 4932, Training Logs: loss_final: 0.490800, loss_mean: 0.815942, proj_loss: -0.406539, loss_mean_cls: 0.081398, grad_norm: 0.419064 +Steps: 0%| | 4933/1000000 [20:28<68:39:22, 4.03it/s, grad_norm=0.419, loss_final=0.491, loss_mean=0.816, loss_mean_cls=0.0814, proj_loss=-0.407][2026-03-26 12:51:32] Step: 4933, Training Logs: loss_final: 0.496937, loss_mean: 0.836886, proj_loss: -0.418011, loss_mean_cls: 0.078062, grad_norm: 0.515337 +Steps: 0%| | 4934/1000000 [20:29<68:34:21, 4.03it/s, grad_norm=0.515, loss_final=0.497, loss_mean=0.837, loss_mean_cls=0.0781, proj_loss=-0.418][2026-03-26 12:51:32] Step: 4934, Training Logs: loss_final: 0.509741, loss_mean: 0.847814, proj_loss: -0.415797, loss_mean_cls: 0.077724, grad_norm: 0.605864 +Steps: 0%| | 4935/1000000 [20:29<68:28:31, 4.04it/s, grad_norm=0.606, loss_final=0.51, loss_mean=0.848, loss_mean_cls=0.0777, proj_loss=-0.416][2026-03-26 12:51:32] Step: 4935, Training Logs: loss_final: 0.479718, loss_mean: 0.802077, proj_loss: -0.403050, loss_mean_cls: 0.080691, grad_norm: 0.340473 +Steps: 0%| | 4936/1000000 [20:29<68:27:20, 4.04it/s, grad_norm=0.34, loss_final=0.48, loss_mean=0.802, loss_mean_cls=0.0807, proj_loss=-0.403][2026-03-26 12:51:32] Step: 4936, Training Logs: loss_final: 0.492692, loss_mean: 0.821392, proj_loss: -0.409705, loss_mean_cls: 0.081006, grad_norm: 0.623351 +Steps: 0%| | 4937/1000000 [20:29<68:24:35, 4.04it/s, grad_norm=0.623, loss_final=0.493, loss_mean=0.821, loss_mean_cls=0.081, proj_loss=-0.41][2026-03-26 12:51:33] Step: 4937, Training Logs: loss_final: 0.481036, loss_mean: 0.807243, proj_loss: -0.407066, loss_mean_cls: 0.080859, grad_norm: 0.598924 +Steps: 0%| | 4938/1000000 [20:30<68:26:57, 4.04it/s, grad_norm=0.599, loss_final=0.481, loss_mean=0.807, loss_mean_cls=0.0809, proj_loss=-0.407][2026-03-26 12:51:33] Step: 4938, Training Logs: loss_final: 0.503417, loss_mean: 0.840349, proj_loss: -0.414342, loss_mean_cls: 0.077410, grad_norm: 0.437690 +Steps: 0%| | 4939/1000000 [20:30<68:25:26, 4.04it/s, grad_norm=0.438, loss_final=0.503, loss_mean=0.84, loss_mean_cls=0.0774, proj_loss=-0.414][2026-03-26 12:51:33] Step: 4939, Training Logs: loss_final: 0.498062, loss_mean: 0.838777, proj_loss: -0.418681, loss_mean_cls: 0.077966, grad_norm: 0.716257 +Steps: 0%| | 4940/1000000 [20:30<68:23:49, 4.04it/s, grad_norm=0.716, loss_final=0.498, loss_mean=0.839, loss_mean_cls=0.078, proj_loss=-0.419][2026-03-26 12:51:33] Step: 4940, Training Logs: loss_final: 0.501409, loss_mean: 0.842070, proj_loss: -0.417707, loss_mean_cls: 0.077047, grad_norm: 0.281552 +Steps: 0%| | 4941/1000000 [20:30<68:23:12, 4.04it/s, grad_norm=0.282, loss_final=0.501, loss_mean=0.842, loss_mean_cls=0.077, proj_loss=-0.418][2026-03-26 12:51:34] Step: 4941, Training Logs: loss_final: 0.480089, loss_mean: 0.814456, proj_loss: -0.413732, loss_mean_cls: 0.079366, grad_norm: 0.486699 +Steps: 0%| | 4942/1000000 [20:31<68:23:55, 4.04it/s, grad_norm=0.487, loss_final=0.48, loss_mean=0.814, loss_mean_cls=0.0794, proj_loss=-0.414][2026-03-26 12:51:34] Step: 4942, Training Logs: loss_final: 0.505887, loss_mean: 0.845789, proj_loss: -0.417804, loss_mean_cls: 0.077901, grad_norm: 0.335625 +Steps: 0%| | 4943/1000000 [20:31<68:24:24, 4.04it/s, grad_norm=0.336, loss_final=0.506, loss_mean=0.846, loss_mean_cls=0.0779, proj_loss=-0.418][2026-03-26 12:51:34] Step: 4943, Training Logs: loss_final: 0.499882, loss_mean: 0.837121, proj_loss: -0.415712, loss_mean_cls: 0.078474, grad_norm: 0.296287 +Steps: 0%| | 4944/1000000 [20:31<68:24:59, 4.04it/s, grad_norm=0.296, loss_final=0.5, loss_mean=0.837, loss_mean_cls=0.0785, proj_loss=-0.416][2026-03-26 12:51:34] Step: 4944, Training Logs: loss_final: 0.481837, loss_mean: 0.806431, proj_loss: -0.407317, loss_mean_cls: 0.082723, grad_norm: 0.355364 +Steps: 0%| | 4945/1000000 [20:31<69:05:15, 4.00it/s, grad_norm=0.355, loss_final=0.482, loss_mean=0.806, loss_mean_cls=0.0827, proj_loss=-0.407][2026-03-26 12:51:35] Step: 4945, Training Logs: loss_final: 0.491497, loss_mean: 0.832592, proj_loss: -0.420133, loss_mean_cls: 0.079038, grad_norm: 0.284903 +Steps: 0%| | 4946/1000000 [20:32<68:53:22, 4.01it/s, grad_norm=0.285, loss_final=0.491, loss_mean=0.833, loss_mean_cls=0.079, proj_loss=-0.42][2026-03-26 12:51:35] Step: 4946, Training Logs: loss_final: 0.500888, loss_mean: 0.839715, proj_loss: -0.416688, loss_mean_cls: 0.077862, grad_norm: 0.483826 +Steps: 0%| | 4947/1000000 [20:32<68:42:03, 4.02it/s, grad_norm=0.484, loss_final=0.501, loss_mean=0.84, loss_mean_cls=0.0779, proj_loss=-0.417][2026-03-26 12:51:35] Step: 4947, Training Logs: loss_final: 0.496454, loss_mean: 0.833190, proj_loss: -0.415902, loss_mean_cls: 0.079166, grad_norm: 0.487168 +Steps: 0%| | 4948/1000000 [20:32<68:38:24, 4.03it/s, grad_norm=0.487, loss_final=0.496, loss_mean=0.833, loss_mean_cls=0.0792, proj_loss=-0.416][2026-03-26 12:51:35] Step: 4948, Training Logs: loss_final: 0.489474, loss_mean: 0.823395, proj_loss: -0.412186, loss_mean_cls: 0.078265, grad_norm: 0.376055 +Steps: 0%| | 4949/1000000 [20:32<68:31:59, 4.03it/s, grad_norm=0.376, loss_final=0.489, loss_mean=0.823, loss_mean_cls=0.0783, proj_loss=-0.412][2026-03-26 12:51:36] Step: 4949, Training Logs: loss_final: 0.484272, loss_mean: 0.817658, proj_loss: -0.414913, loss_mean_cls: 0.081527, grad_norm: 0.599731 +Steps: 0%| | 4950/1000000 [20:33<68:29:20, 4.04it/s, grad_norm=0.6, loss_final=0.484, loss_mean=0.818, loss_mean_cls=0.0815, proj_loss=-0.415][2026-03-26 12:51:36] Step: 4950, Training Logs: loss_final: 0.492476, loss_mean: 0.829646, proj_loss: -0.415882, loss_mean_cls: 0.078713, grad_norm: 0.315955 +Steps: 0%| | 4951/1000000 [20:33<68:25:35, 4.04it/s, grad_norm=0.316, loss_final=0.492, loss_mean=0.83, loss_mean_cls=0.0787, proj_loss=-0.416][2026-03-26 12:51:36] Step: 4951, Training Logs: loss_final: 0.485562, loss_mean: 0.833333, proj_loss: -0.423838, loss_mean_cls: 0.076067, grad_norm: 0.302747 +Steps: 0%| | 4952/1000000 [20:33<68:24:12, 4.04it/s, grad_norm=0.303, loss_final=0.486, loss_mean=0.833, loss_mean_cls=0.0761, proj_loss=-0.424][2026-03-26 12:51:36] Step: 4952, Training Logs: loss_final: 0.493114, loss_mean: 0.826289, proj_loss: -0.412931, loss_mean_cls: 0.079756, grad_norm: 0.311911 +Steps: 0%| | 4953/1000000 [20:33<68:23:14, 4.04it/s, grad_norm=0.312, loss_final=0.493, loss_mean=0.826, loss_mean_cls=0.0798, proj_loss=-0.413][2026-03-26 12:51:37] Step: 4953, Training Logs: loss_final: 0.508921, loss_mean: 0.837861, proj_loss: -0.409566, loss_mean_cls: 0.080627, grad_norm: 0.451383 +Steps: 0%| | 4954/1000000 [20:33<68:23:10, 4.04it/s, grad_norm=0.451, loss_final=0.509, loss_mean=0.838, loss_mean_cls=0.0806, proj_loss=-0.41][2026-03-26 12:51:37] Step: 4954, Training Logs: loss_final: 0.509557, loss_mean: 0.843338, proj_loss: -0.412845, loss_mean_cls: 0.079065, grad_norm: 0.346785 +Steps: 0%| | 4955/1000000 [20:34<68:25:17, 4.04it/s, grad_norm=0.347, loss_final=0.51, loss_mean=0.843, loss_mean_cls=0.0791, proj_loss=-0.413][2026-03-26 12:51:37] Step: 4955, Training Logs: loss_final: 0.483554, loss_mean: 0.801287, proj_loss: -0.400529, loss_mean_cls: 0.082796, grad_norm: 0.360623 +Steps: 0%| | 4956/1000000 [20:34<68:27:19, 4.04it/s, grad_norm=0.361, loss_final=0.484, loss_mean=0.801, loss_mean_cls=0.0828, proj_loss=-0.401][2026-03-26 12:51:37] Step: 4956, Training Logs: loss_final: 0.477890, loss_mean: 0.815938, proj_loss: -0.417572, loss_mean_cls: 0.079525, grad_norm: 0.292267 +Steps: 0%| | 4957/1000000 [20:34<68:25:20, 4.04it/s, grad_norm=0.292, loss_final=0.478, loss_mean=0.816, loss_mean_cls=0.0795, proj_loss=-0.418][2026-03-26 12:51:38] Step: 4957, Training Logs: loss_final: 0.497484, loss_mean: 0.834052, proj_loss: -0.415474, loss_mean_cls: 0.078906, grad_norm: 0.371133 +Steps: 0%| | 4958/1000000 [20:34<68:26:25, 4.04it/s, grad_norm=0.371, loss_final=0.497, loss_mean=0.834, loss_mean_cls=0.0789, proj_loss=-0.415][2026-03-26 12:51:38] Step: 4958, Training Logs: loss_final: 0.490314, loss_mean: 0.823406, proj_loss: -0.413706, loss_mean_cls: 0.080614, grad_norm: 0.298842 +Steps: 0%| | 4959/1000000 [20:35<68:25:04, 4.04it/s, grad_norm=0.299, loss_final=0.49, loss_mean=0.823, loss_mean_cls=0.0806, proj_loss=-0.414][2026-03-26 12:51:38] Step: 4959, Training Logs: loss_final: 0.494104, loss_mean: 0.830862, proj_loss: -0.415004, loss_mean_cls: 0.078246, grad_norm: 0.301601 +Steps: 0%| | 4960/1000000 [20:35<68:25:25, 4.04it/s, grad_norm=0.302, loss_final=0.494, loss_mean=0.831, loss_mean_cls=0.0782, proj_loss=-0.415][2026-03-26 12:51:38] Step: 4960, Training Logs: loss_final: 0.503816, loss_mean: 0.830064, proj_loss: -0.405479, loss_mean_cls: 0.079232, grad_norm: 0.341782 +Steps: 0%| | 4961/1000000 [20:35<68:23:40, 4.04it/s, grad_norm=0.342, loss_final=0.504, loss_mean=0.83, loss_mean_cls=0.0792, proj_loss=-0.405][2026-03-26 12:51:39] Step: 4961, Training Logs: loss_final: 0.502088, loss_mean: 0.837727, proj_loss: -0.414183, loss_mean_cls: 0.078543, grad_norm: 0.303351 +Steps: 0%| | 4962/1000000 [20:35<68:22:21, 4.04it/s, grad_norm=0.303, loss_final=0.502, loss_mean=0.838, loss_mean_cls=0.0785, proj_loss=-0.414][2026-03-26 12:51:39] Step: 4962, Training Logs: loss_final: 0.491626, loss_mean: 0.821408, proj_loss: -0.411365, loss_mean_cls: 0.081583, grad_norm: 0.280376 +Steps: 0%| | 4963/1000000 [20:36<68:21:57, 4.04it/s, grad_norm=0.28, loss_final=0.492, loss_mean=0.821, loss_mean_cls=0.0816, proj_loss=-0.411][2026-03-26 12:51:39] Step: 4963, Training Logs: loss_final: 0.485121, loss_mean: 0.810256, proj_loss: -0.405735, loss_mean_cls: 0.080599, grad_norm: 0.226945 +Steps: 0%| | 4964/1000000 [20:36<68:20:38, 4.04it/s, grad_norm=0.227, loss_final=0.485, loss_mean=0.81, loss_mean_cls=0.0806, proj_loss=-0.406][2026-03-26 12:51:39] Step: 4964, Training Logs: loss_final: 0.476728, loss_mean: 0.814389, proj_loss: -0.416754, loss_mean_cls: 0.079093, grad_norm: 0.246612 +Steps: 0%| | 4965/1000000 [20:36<68:22:52, 4.04it/s, grad_norm=0.247, loss_final=0.477, loss_mean=0.814, loss_mean_cls=0.0791, proj_loss=-0.417][2026-03-26 12:51:40] Step: 4965, Training Logs: loss_final: 0.489101, loss_mean: 0.831391, proj_loss: -0.419378, loss_mean_cls: 0.077089, grad_norm: 0.314156 +Steps: 0%| | 4966/1000000 [20:36<68:23:22, 4.04it/s, grad_norm=0.314, loss_final=0.489, loss_mean=0.831, loss_mean_cls=0.0771, proj_loss=-0.419][2026-03-26 12:51:40] Step: 4966, Training Logs: loss_final: 0.498075, loss_mean: 0.837021, proj_loss: -0.417297, loss_mean_cls: 0.078351, grad_norm: 0.283027 +Steps: 0%| | 4967/1000000 [20:37<68:25:06, 4.04it/s, grad_norm=0.283, loss_final=0.498, loss_mean=0.837, loss_mean_cls=0.0784, proj_loss=-0.417][2026-03-26 12:51:40] Step: 4967, Training Logs: loss_final: 0.494126, loss_mean: 0.839685, proj_loss: -0.422171, loss_mean_cls: 0.076611, grad_norm: 0.319933 +Steps: 0%| | 4968/1000000 [20:37<68:25:33, 4.04it/s, grad_norm=0.32, loss_final=0.494, loss_mean=0.84, loss_mean_cls=0.0766, proj_loss=-0.422][2026-03-26 12:51:40] Step: 4968, Training Logs: loss_final: 0.489623, loss_mean: 0.821898, proj_loss: -0.410942, loss_mean_cls: 0.078666, grad_norm: 0.222499 +Steps: 0%| | 4969/1000000 [20:37<68:24:01, 4.04it/s, grad_norm=0.222, loss_final=0.49, loss_mean=0.822, loss_mean_cls=0.0787, proj_loss=-0.411][2026-03-26 12:51:41] Step: 4969, Training Logs: loss_final: 0.486775, loss_mean: 0.839776, proj_loss: -0.428254, loss_mean_cls: 0.075252, grad_norm: 0.277459 +Steps: 0%| | 4970/1000000 [20:37<68:24:08, 4.04it/s, grad_norm=0.277, loss_final=0.487, loss_mean=0.84, loss_mean_cls=0.0753, proj_loss=-0.428][2026-03-26 12:51:41] Step: 4970, Training Logs: loss_final: 0.508732, loss_mean: 0.841974, proj_loss: -0.410153, loss_mean_cls: 0.076910, grad_norm: 0.271794 +Steps: 0%| | 4971/1000000 [20:38<68:21:54, 4.04it/s, grad_norm=0.272, loss_final=0.509, loss_mean=0.842, loss_mean_cls=0.0769, proj_loss=-0.41][2026-03-26 12:51:41] Step: 4971, Training Logs: loss_final: 0.494532, loss_mean: 0.836691, proj_loss: -0.419042, loss_mean_cls: 0.076883, grad_norm: 0.234162 +Steps: 0%| | 4972/1000000 [20:38<68:26:45, 4.04it/s, grad_norm=0.234, loss_final=0.495, loss_mean=0.837, loss_mean_cls=0.0769, proj_loss=-0.419][2026-03-26 12:51:41] Step: 4972, Training Logs: loss_final: 0.496350, loss_mean: 0.830589, proj_loss: -0.412966, loss_mean_cls: 0.078728, grad_norm: 0.306223 +Steps: 0%| | 4973/1000000 [20:38<68:24:33, 4.04it/s, grad_norm=0.306, loss_final=0.496, loss_mean=0.831, loss_mean_cls=0.0787, proj_loss=-0.413][2026-03-26 12:51:42] Step: 4973, Training Logs: loss_final: 0.493900, loss_mean: 0.828321, proj_loss: -0.413687, loss_mean_cls: 0.079265, grad_norm: 0.290930 +Steps: 0%| | 4974/1000000 [20:38<68:24:45, 4.04it/s, grad_norm=0.291, loss_final=0.494, loss_mean=0.828, loss_mean_cls=0.0793, proj_loss=-0.414][2026-03-26 12:51:42] Step: 4974, Training Logs: loss_final: 0.479521, loss_mean: 0.824313, proj_loss: -0.422623, loss_mean_cls: 0.077831, grad_norm: 0.269556 +Steps: 0%| | 4975/1000000 [20:39<68:21:37, 4.04it/s, grad_norm=0.27, loss_final=0.48, loss_mean=0.824, loss_mean_cls=0.0778, proj_loss=-0.423][2026-03-26 12:51:42] Step: 4975, Training Logs: loss_final: 0.484361, loss_mean: 0.816035, proj_loss: -0.411688, loss_mean_cls: 0.080013, grad_norm: 0.225944 +Steps: 0%| | 4976/1000000 [20:39<68:25:33, 4.04it/s, grad_norm=0.226, loss_final=0.484, loss_mean=0.816, loss_mean_cls=0.08, proj_loss=-0.412][2026-03-26 12:51:42] Step: 4976, Training Logs: loss_final: 0.489126, loss_mean: 0.825100, proj_loss: -0.414856, loss_mean_cls: 0.078882, grad_norm: 0.329592 +Steps: 0%| | 4977/1000000 [20:39<68:24:36, 4.04it/s, grad_norm=0.33, loss_final=0.489, loss_mean=0.825, loss_mean_cls=0.0789, proj_loss=-0.415][2026-03-26 12:51:43] Step: 4977, Training Logs: loss_final: 0.502958, loss_mean: 0.833233, proj_loss: -0.410744, loss_mean_cls: 0.080468, grad_norm: 0.300265 +Steps: 0%| | 4978/1000000 [20:39<68:25:13, 4.04it/s, grad_norm=0.3, loss_final=0.503, loss_mean=0.833, loss_mean_cls=0.0805, proj_loss=-0.411][2026-03-26 12:51:43] Step: 4978, Training Logs: loss_final: 0.489824, loss_mean: 0.828349, proj_loss: -0.416581, loss_mean_cls: 0.078056, grad_norm: 0.247570 +Steps: 0%| | 4979/1000000 [20:40<68:24:56, 4.04it/s, grad_norm=0.248, loss_final=0.49, loss_mean=0.828, loss_mean_cls=0.0781, proj_loss=-0.417][2026-03-26 12:51:43] Step: 4979, Training Logs: loss_final: 0.481035, loss_mean: 0.817593, proj_loss: -0.414851, loss_mean_cls: 0.078293, grad_norm: 0.248438 +Steps: 0%| | 4980/1000000 [20:40<68:25:43, 4.04it/s, grad_norm=0.248, loss_final=0.481, loss_mean=0.818, loss_mean_cls=0.0783, proj_loss=-0.415][2026-03-26 12:51:43] Step: 4980, Training Logs: loss_final: 0.504369, loss_mean: 0.830826, proj_loss: -0.405584, loss_mean_cls: 0.079126, grad_norm: 0.406445 +Steps: 0%| | 4981/1000000 [20:40<68:27:40, 4.04it/s, grad_norm=0.406, loss_final=0.504, loss_mean=0.831, loss_mean_cls=0.0791, proj_loss=-0.406][2026-03-26 12:51:44] Step: 4981, Training Logs: loss_final: 0.495188, loss_mean: 0.834075, proj_loss: -0.416595, loss_mean_cls: 0.077708, grad_norm: 0.308124 +Steps: 0%| | 4982/1000000 [20:40<68:24:33, 4.04it/s, grad_norm=0.308, loss_final=0.495, loss_mean=0.834, loss_mean_cls=0.0777, proj_loss=-0.417][2026-03-26 12:51:44] Step: 4982, Training Logs: loss_final: 0.516563, loss_mean: 0.858967, proj_loss: -0.419319, loss_mean_cls: 0.076916, grad_norm: 0.378568 +Steps: 0%| | 4983/1000000 [20:41<68:23:43, 4.04it/s, grad_norm=0.379, loss_final=0.517, loss_mean=0.859, loss_mean_cls=0.0769, proj_loss=-0.419][2026-03-26 12:51:44] Step: 4983, Training Logs: loss_final: 0.486829, loss_mean: 0.820259, proj_loss: -0.413408, loss_mean_cls: 0.079978, grad_norm: 0.298364 +Steps: 0%| | 4984/1000000 [20:41<68:25:25, 4.04it/s, grad_norm=0.298, loss_final=0.487, loss_mean=0.82, loss_mean_cls=0.08, proj_loss=-0.413][2026-03-26 12:51:44] Step: 4984, Training Logs: loss_final: 0.500188, loss_mean: 0.825270, proj_loss: -0.406573, loss_mean_cls: 0.081492, grad_norm: 0.518371 +Steps: 0%| | 4985/1000000 [20:41<68:24:18, 4.04it/s, grad_norm=0.518, loss_final=0.5, loss_mean=0.825, loss_mean_cls=0.0815, proj_loss=-0.407][2026-03-26 12:51:45] Step: 4985, Training Logs: loss_final: 0.486052, loss_mean: 0.825729, proj_loss: -0.417434, loss_mean_cls: 0.077757, grad_norm: 0.373156 +Steps: 0%| | 4986/1000000 [20:41<68:25:11, 4.04it/s, grad_norm=0.373, loss_final=0.486, loss_mean=0.826, loss_mean_cls=0.0778, proj_loss=-0.417][2026-03-26 12:51:45] Step: 4986, Training Logs: loss_final: 0.485270, loss_mean: 0.814041, proj_loss: -0.409830, loss_mean_cls: 0.081059, grad_norm: 0.381918 +Steps: 0%| | 4987/1000000 [20:42<68:24:16, 4.04it/s, grad_norm=0.382, loss_final=0.485, loss_mean=0.814, loss_mean_cls=0.0811, proj_loss=-0.41][2026-03-26 12:51:45] Step: 4987, Training Logs: loss_final: 0.497597, loss_mean: 0.826747, proj_loss: -0.409282, loss_mean_cls: 0.080132, grad_norm: 0.425663 +Steps: 0%| | 4988/1000000 [20:42<68:24:36, 4.04it/s, grad_norm=0.426, loss_final=0.498, loss_mean=0.827, loss_mean_cls=0.0801, proj_loss=-0.409][2026-03-26 12:51:45] Step: 4988, Training Logs: loss_final: 0.490705, loss_mean: 0.821404, proj_loss: -0.410486, loss_mean_cls: 0.079786, grad_norm: 0.289394 +Steps: 0%| | 4989/1000000 [20:42<68:31:26, 4.03it/s, grad_norm=0.289, loss_final=0.491, loss_mean=0.821, loss_mean_cls=0.0798, proj_loss=-0.41][2026-03-26 12:51:46] Step: 4989, Training Logs: loss_final: 0.508765, loss_mean: 0.844723, proj_loss: -0.414513, loss_mean_cls: 0.078555, grad_norm: 0.455643 +Steps: 0%| | 4990/1000000 [20:42<68:26:31, 4.04it/s, grad_norm=0.456, loss_final=0.509, loss_mean=0.845, loss_mean_cls=0.0786, proj_loss=-0.415][2026-03-26 12:51:46] Step: 4990, Training Logs: loss_final: 0.494661, loss_mean: 0.817772, proj_loss: -0.403820, loss_mean_cls: 0.080709, grad_norm: 0.446271 +Steps: 0%| | 4991/1000000 [20:43<68:26:40, 4.04it/s, grad_norm=0.446, loss_final=0.495, loss_mean=0.818, loss_mean_cls=0.0807, proj_loss=-0.404][2026-03-26 12:51:46] Step: 4991, Training Logs: loss_final: 0.505830, loss_mean: 0.844199, proj_loss: -0.416240, loss_mean_cls: 0.077871, grad_norm: 0.446957 +Steps: 0%| | 4992/1000000 [20:43<68:24:47, 4.04it/s, grad_norm=0.447, loss_final=0.506, loss_mean=0.844, loss_mean_cls=0.0779, proj_loss=-0.416][2026-03-26 12:51:46] Step: 4992, Training Logs: loss_final: 0.493650, loss_mean: 0.827625, proj_loss: -0.412900, loss_mean_cls: 0.078925, grad_norm: 0.522932 +Steps: 0%| | 4993/1000000 [20:43<68:23:21, 4.04it/s, grad_norm=0.523, loss_final=0.494, loss_mean=0.828, loss_mean_cls=0.0789, proj_loss=-0.413][2026-03-26 12:51:47] Step: 4993, Training Logs: loss_final: 0.492305, loss_mean: 0.821847, proj_loss: -0.410006, loss_mean_cls: 0.080464, grad_norm: 0.444396 +Steps: 0%| | 4994/1000000 [20:43<68:22:01, 4.04it/s, grad_norm=0.444, loss_final=0.492, loss_mean=0.822, loss_mean_cls=0.0805, proj_loss=-0.41][2026-03-26 12:51:47] Step: 4994, Training Logs: loss_final: 0.479370, loss_mean: 0.830148, proj_loss: -0.424853, loss_mean_cls: 0.074076, grad_norm: 0.657379 +Steps: 0%| | 4995/1000000 [20:44<68:26:54, 4.04it/s, grad_norm=0.657, loss_final=0.479, loss_mean=0.83, loss_mean_cls=0.0741, proj_loss=-0.425][2026-03-26 12:51:47] Step: 4995, Training Logs: loss_final: 0.491368, loss_mean: 0.837625, proj_loss: -0.422250, loss_mean_cls: 0.075993, grad_norm: 0.515079 +Steps: 0%| | 4996/1000000 [20:44<72:52:46, 3.79it/s, grad_norm=0.515, loss_final=0.491, loss_mean=0.838, loss_mean_cls=0.076, proj_loss=-0.422][2026-03-26 12:51:47] Step: 4996, Training Logs: loss_final: 0.510527, loss_mean: 0.847257, proj_loss: -0.413862, loss_mean_cls: 0.077132, grad_norm: 0.419626 +Steps: 0%| | 4997/1000000 [20:44<71:27:15, 3.87it/s, grad_norm=0.42, loss_final=0.511, loss_mean=0.847, loss_mean_cls=0.0771, proj_loss=-0.414][2026-03-26 12:51:48] Step: 4997, Training Logs: loss_final: 0.501785, loss_mean: 0.849965, proj_loss: -0.424476, loss_mean_cls: 0.076296, grad_norm: 0.776419 +Steps: 0%| | 4998/1000000 [20:44<70:28:22, 3.92it/s, grad_norm=0.776, loss_final=0.502, loss_mean=0.85, loss_mean_cls=0.0763, proj_loss=-0.424][2026-03-26 12:51:48] Step: 4998, Training Logs: loss_final: 0.494037, loss_mean: 0.826376, proj_loss: -0.410589, loss_mean_cls: 0.078250, grad_norm: 0.578883 +Steps: 0%| | 4999/1000000 [20:45<69:43:36, 3.96it/s, grad_norm=0.579, loss_final=0.494, loss_mean=0.826, loss_mean_cls=0.0782, proj_loss=-0.411][2026-03-26 12:51:48] Step: 4999, Training Logs: loss_final: 0.494364, loss_mean: 0.829314, proj_loss: -0.414462, loss_mean_cls: 0.079512, grad_norm: 0.415275 +Steps: 0%| | 5000/1000000 [20:45<69:19:15, 3.99it/s, grad_norm=0.415, loss_final=0.494, loss_mean=0.829, loss_mean_cls=0.0795, proj_loss=-0.414][2026-03-26 12:51:48] Step: 5000, Training Logs: loss_final: 0.473378, loss_mean: 0.809760, proj_loss: -0.414870, loss_mean_cls: 0.078488, grad_norm: 0.835574 +Steps: 1%| | 5001/1000000 [20:45<69:01:18, 4.00it/s, grad_norm=0.836, loss_final=0.473, loss_mean=0.81, loss_mean_cls=0.0785, proj_loss=-0.415][2026-03-26 12:51:49] Step: 5001, Training Logs: loss_final: 0.500463, loss_mean: 0.837074, proj_loss: -0.415278, loss_mean_cls: 0.078667, grad_norm: 0.680689 +Steps: 1%| | 5002/1000000 [20:45<68:47:57, 4.02it/s, grad_norm=0.681, loss_final=0.5, loss_mean=0.837, loss_mean_cls=0.0787, proj_loss=-0.415][2026-03-26 12:51:49] Step: 5002, Training Logs: loss_final: 0.498667, loss_mean: 0.830947, proj_loss: -0.412544, loss_mean_cls: 0.080264, grad_norm: 0.390105 +Steps: 1%| | 5003/1000000 [20:46<68:41:20, 4.02it/s, grad_norm=0.39, loss_final=0.499, loss_mean=0.831, loss_mean_cls=0.0803, proj_loss=-0.413][2026-03-26 12:51:49] Step: 5003, Training Logs: loss_final: 0.478911, loss_mean: 0.810347, proj_loss: -0.411881, loss_mean_cls: 0.080445, grad_norm: 0.576389 +Steps: 1%| | 5004/1000000 [20:46<75:00:02, 3.69it/s, grad_norm=0.576, loss_final=0.479, loss_mean=0.81, loss_mean_cls=0.0804, proj_loss=-0.412][2026-03-26 12:51:49] Step: 5004, Training Logs: loss_final: 0.498213, loss_mean: 0.823911, proj_loss: -0.406178, loss_mean_cls: 0.080480, grad_norm: 0.435437 +Steps: 1%| | 5005/1000000 [20:47<105:33:20, 2.62it/s, grad_norm=0.435, loss_final=0.498, loss_mean=0.824, loss_mean_cls=0.0805, proj_loss=-0.406][2026-03-26 12:51:50] Step: 5005, Training Logs: loss_final: 0.491740, loss_mean: 0.814215, proj_loss: -0.404443, loss_mean_cls: 0.081968, grad_norm: 0.300299 +Steps: 1%| | 5006/1000000 [20:47<94:19:59, 2.93it/s, grad_norm=0.3, loss_final=0.492, loss_mean=0.814, loss_mean_cls=0.082, proj_loss=-0.404] [2026-03-26 12:51:50] Step: 5006, Training Logs: loss_final: 0.487856, loss_mean: 0.814912, proj_loss: -0.408472, loss_mean_cls: 0.081416, grad_norm: 0.332495 +Steps: 1%| | 5007/1000000 [20:47<86:32:59, 3.19it/s, grad_norm=0.332, loss_final=0.488, loss_mean=0.815, loss_mean_cls=0.0814, proj_loss=-0.408][2026-03-26 12:51:51] Step: 5007, Training Logs: loss_final: 0.496707, loss_mean: 0.830306, proj_loss: -0.412216, loss_mean_cls: 0.078618, grad_norm: 0.324884 +Steps: 1%| | 5008/1000000 [20:47<81:06:30, 3.41it/s, grad_norm=0.325, loss_final=0.497, loss_mean=0.83, loss_mean_cls=0.0786, proj_loss=-0.412][2026-03-26 12:51:51] Step: 5008, Training Logs: loss_final: 0.513989, loss_mean: 0.856759, proj_loss: -0.418226, loss_mean_cls: 0.075456, grad_norm: 0.251976 +Steps: 1%| | 5009/1000000 [20:48<77:19:30, 3.57it/s, grad_norm=0.252, loss_final=0.514, loss_mean=0.857, loss_mean_cls=0.0755, proj_loss=-0.418][2026-03-26 12:51:51] Step: 5009, Training Logs: loss_final: 0.482565, loss_mean: 0.814427, proj_loss: -0.411038, loss_mean_cls: 0.079176, grad_norm: 0.301064 +Steps: 1%| | 5010/1000000 [20:48<74:39:05, 3.70it/s, grad_norm=0.301, loss_final=0.483, loss_mean=0.814, loss_mean_cls=0.0792, proj_loss=-0.411][2026-03-26 12:51:51] Step: 5010, Training Logs: loss_final: 0.476258, loss_mean: 0.821582, proj_loss: -0.421553, loss_mean_cls: 0.076229, grad_norm: 0.353744 +Steps: 1%| | 5011/1000000 [20:48<72:46:36, 3.80it/s, grad_norm=0.354, loss_final=0.476, loss_mean=0.822, loss_mean_cls=0.0762, proj_loss=-0.422][2026-03-26 12:51:52] Step: 5011, Training Logs: loss_final: 0.489727, loss_mean: 0.817794, proj_loss: -0.410249, loss_mean_cls: 0.082182, grad_norm: 0.298701 +Steps: 1%| | 5012/1000000 [20:48<71:29:20, 3.87it/s, grad_norm=0.299, loss_final=0.49, loss_mean=0.818, loss_mean_cls=0.0822, proj_loss=-0.41][2026-03-26 12:51:52] Step: 5012, Training Logs: loss_final: 0.488079, loss_mean: 0.826461, proj_loss: -0.414860, loss_mean_cls: 0.076477, grad_norm: 0.381446 +Steps: 1%| | 5013/1000000 [20:49<70:35:45, 3.92it/s, grad_norm=0.381, loss_final=0.488, loss_mean=0.826, loss_mean_cls=0.0765, proj_loss=-0.415][2026-03-26 12:51:52] Step: 5013, Training Logs: loss_final: 0.490294, loss_mean: 0.824365, proj_loss: -0.413901, loss_mean_cls: 0.079830, grad_norm: 0.334735 +Steps: 1%| | 5014/1000000 [20:49<69:58:33, 3.95it/s, grad_norm=0.335, loss_final=0.49, loss_mean=0.824, loss_mean_cls=0.0798, proj_loss=-0.414][2026-03-26 12:51:52] Step: 5014, Training Logs: loss_final: 0.506713, loss_mean: 0.833724, proj_loss: -0.405673, loss_mean_cls: 0.078662, grad_norm: 0.448470 +Steps: 1%| | 5015/1000000 [20:49<69:27:53, 3.98it/s, grad_norm=0.448, loss_final=0.507, loss_mean=0.834, loss_mean_cls=0.0787, proj_loss=-0.406][2026-03-26 12:51:52] Step: 5015, Training Logs: loss_final: 0.481928, loss_mean: 0.796914, proj_loss: -0.398755, loss_mean_cls: 0.083769, grad_norm: 0.435366 +Steps: 1%| | 5016/1000000 [20:49<69:10:23, 4.00it/s, grad_norm=0.435, loss_final=0.482, loss_mean=0.797, loss_mean_cls=0.0838, proj_loss=-0.399][2026-03-26 12:51:53] Step: 5016, Training Logs: loss_final: 0.499205, loss_mean: 0.824404, proj_loss: -0.405301, loss_mean_cls: 0.080102, grad_norm: 0.506764 +Steps: 1%| | 5017/1000000 [20:50<68:54:33, 4.01it/s, grad_norm=0.507, loss_final=0.499, loss_mean=0.824, loss_mean_cls=0.0801, proj_loss=-0.405][2026-03-26 12:51:53] Step: 5017, Training Logs: loss_final: 0.490047, loss_mean: 0.815802, proj_loss: -0.406550, loss_mean_cls: 0.080796, grad_norm: 0.561995 +Steps: 1%| | 5018/1000000 [20:50<68:44:53, 4.02it/s, grad_norm=0.562, loss_final=0.49, loss_mean=0.816, loss_mean_cls=0.0808, proj_loss=-0.407][2026-03-26 12:51:53] Step: 5018, Training Logs: loss_final: 0.496295, loss_mean: 0.816980, proj_loss: -0.400602, loss_mean_cls: 0.079917, grad_norm: 0.320469 +Steps: 1%| | 5019/1000000 [20:50<68:36:55, 4.03it/s, grad_norm=0.32, loss_final=0.496, loss_mean=0.817, loss_mean_cls=0.0799, proj_loss=-0.401][2026-03-26 12:51:53] Step: 5019, Training Logs: loss_final: 0.510450, loss_mean: 0.844167, proj_loss: -0.411805, loss_mean_cls: 0.078089, grad_norm: 0.585621 +Steps: 1%| | 5020/1000000 [20:50<68:31:50, 4.03it/s, grad_norm=0.586, loss_final=0.51, loss_mean=0.844, loss_mean_cls=0.0781, proj_loss=-0.412][2026-03-26 12:51:54] Step: 5020, Training Logs: loss_final: 0.497675, loss_mean: 0.827947, proj_loss: -0.408436, loss_mean_cls: 0.078163, grad_norm: 0.272966 +Steps: 1%| | 5021/1000000 [20:51<68:27:00, 4.04it/s, grad_norm=0.273, loss_final=0.498, loss_mean=0.828, loss_mean_cls=0.0782, proj_loss=-0.408][2026-03-26 12:51:54] Step: 5021, Training Logs: loss_final: 0.484533, loss_mean: 0.824282, proj_loss: -0.417948, loss_mean_cls: 0.078200, grad_norm: 0.492581 +Steps: 1%| | 5022/1000000 [20:51<68:27:33, 4.04it/s, grad_norm=0.493, loss_final=0.485, loss_mean=0.824, loss_mean_cls=0.0782, proj_loss=-0.418][2026-03-26 12:51:54] Step: 5022, Training Logs: loss_final: 0.493835, loss_mean: 0.849203, proj_loss: -0.429590, loss_mean_cls: 0.074222, grad_norm: 0.291604 +Steps: 1%| | 5023/1000000 [20:51<68:26:36, 4.04it/s, grad_norm=0.292, loss_final=0.494, loss_mean=0.849, loss_mean_cls=0.0742, proj_loss=-0.43][2026-03-26 12:51:54] Step: 5023, Training Logs: loss_final: 0.503116, loss_mean: 0.841681, proj_loss: -0.415601, loss_mean_cls: 0.077035, grad_norm: 0.352692 +Steps: 1%| | 5024/1000000 [20:51<68:25:29, 4.04it/s, grad_norm=0.353, loss_final=0.503, loss_mean=0.842, loss_mean_cls=0.077, proj_loss=-0.416][2026-03-26 12:51:55] Step: 5024, Training Logs: loss_final: 0.476072, loss_mean: 0.806183, proj_loss: -0.411345, loss_mean_cls: 0.081234, grad_norm: 0.289531 +Steps: 1%| | 5025/1000000 [20:52<68:23:59, 4.04it/s, grad_norm=0.29, loss_final=0.476, loss_mean=0.806, loss_mean_cls=0.0812, proj_loss=-0.411][2026-03-26 12:51:55] Step: 5025, Training Logs: loss_final: 0.497095, loss_mean: 0.842246, proj_loss: -0.421470, loss_mean_cls: 0.076319, grad_norm: 0.308803 +Steps: 1%| | 5026/1000000 [20:52<68:24:50, 4.04it/s, grad_norm=0.309, loss_final=0.497, loss_mean=0.842, loss_mean_cls=0.0763, proj_loss=-0.421][2026-03-26 12:51:55] Step: 5026, Training Logs: loss_final: 0.496156, loss_mean: 0.830153, proj_loss: -0.411443, loss_mean_cls: 0.077445, grad_norm: 0.400327 +Steps: 1%| | 5027/1000000 [20:52<68:21:08, 4.04it/s, grad_norm=0.4, loss_final=0.496, loss_mean=0.83, loss_mean_cls=0.0774, proj_loss=-0.411][2026-03-26 12:51:55] Step: 5027, Training Logs: loss_final: 0.484620, loss_mean: 0.807269, proj_loss: -0.403813, loss_mean_cls: 0.081165, grad_norm: 0.301528 +Steps: 1%| | 5028/1000000 [20:52<68:22:24, 4.04it/s, grad_norm=0.302, loss_final=0.485, loss_mean=0.807, loss_mean_cls=0.0812, proj_loss=-0.404][2026-03-26 12:51:56] Step: 5028, Training Logs: loss_final: 0.475964, loss_mean: 0.799735, proj_loss: -0.405144, loss_mean_cls: 0.081373, grad_norm: 0.451774 +Steps: 1%| | 5029/1000000 [20:53<68:21:40, 4.04it/s, grad_norm=0.452, loss_final=0.476, loss_mean=0.8, loss_mean_cls=0.0814, proj_loss=-0.405][2026-03-26 12:51:56] Step: 5029, Training Logs: loss_final: 0.482422, loss_mean: 0.812854, proj_loss: -0.410217, loss_mean_cls: 0.079785, grad_norm: 0.419002 +Steps: 1%| | 5030/1000000 [20:53<68:21:31, 4.04it/s, grad_norm=0.419, loss_final=0.482, loss_mean=0.813, loss_mean_cls=0.0798, proj_loss=-0.41][2026-03-26 12:51:56] Step: 5030, Training Logs: loss_final: 0.495310, loss_mean: 0.819108, proj_loss: -0.403892, loss_mean_cls: 0.080094, grad_norm: 0.256946 +Steps: 1%| | 5031/1000000 [20:53<68:24:27, 4.04it/s, grad_norm=0.257, loss_final=0.495, loss_mean=0.819, loss_mean_cls=0.0801, proj_loss=-0.404][2026-03-26 12:51:56] Step: 5031, Training Logs: loss_final: 0.510447, loss_mean: 0.849837, proj_loss: -0.415798, loss_mean_cls: 0.076408, grad_norm: 0.289562 +Steps: 1%| | 5032/1000000 [20:53<68:24:14, 4.04it/s, grad_norm=0.29, loss_final=0.51, loss_mean=0.85, loss_mean_cls=0.0764, proj_loss=-0.416][2026-03-26 12:51:57] Step: 5032, Training Logs: loss_final: 0.489027, loss_mean: 0.833859, proj_loss: -0.420520, loss_mean_cls: 0.075688, grad_norm: 0.308464 +Steps: 1%| | 5033/1000000 [20:54<68:26:17, 4.04it/s, grad_norm=0.308, loss_final=0.489, loss_mean=0.834, loss_mean_cls=0.0757, proj_loss=-0.421][2026-03-26 12:51:57] Step: 5033, Training Logs: loss_final: 0.484599, loss_mean: 0.815807, proj_loss: -0.410155, loss_mean_cls: 0.078946, grad_norm: 0.433902 +Steps: 1%| | 5034/1000000 [20:54<68:26:26, 4.04it/s, grad_norm=0.434, loss_final=0.485, loss_mean=0.816, loss_mean_cls=0.0789, proj_loss=-0.41][2026-03-26 12:51:57] Step: 5034, Training Logs: loss_final: 0.483777, loss_mean: 0.820952, proj_loss: -0.415891, loss_mean_cls: 0.078716, grad_norm: 0.463726 +Steps: 1%| | 5035/1000000 [20:54<68:25:47, 4.04it/s, grad_norm=0.464, loss_final=0.484, loss_mean=0.821, loss_mean_cls=0.0787, proj_loss=-0.416][2026-03-26 12:51:57] Step: 5035, Training Logs: loss_final: 0.503053, loss_mean: 0.838081, proj_loss: -0.413815, loss_mean_cls: 0.078788, grad_norm: 0.370288 +Steps: 1%| | 5036/1000000 [20:54<68:22:25, 4.04it/s, grad_norm=0.37, loss_final=0.503, loss_mean=0.838, loss_mean_cls=0.0788, proj_loss=-0.414][2026-03-26 12:51:58] Step: 5036, Training Logs: loss_final: 0.499971, loss_mean: 0.836975, proj_loss: -0.414533, loss_mean_cls: 0.077529, grad_norm: 0.387746 +Steps: 1%| | 5037/1000000 [20:55<68:23:52, 4.04it/s, grad_norm=0.388, loss_final=0.5, loss_mean=0.837, loss_mean_cls=0.0775, proj_loss=-0.415][2026-03-26 12:51:58] Step: 5037, Training Logs: loss_final: 0.483437, loss_mean: 0.820971, proj_loss: -0.415579, loss_mean_cls: 0.078045, grad_norm: 0.343996 +Steps: 1%| | 5038/1000000 [20:55<68:22:05, 4.04it/s, grad_norm=0.344, loss_final=0.483, loss_mean=0.821, loss_mean_cls=0.078, proj_loss=-0.416][2026-03-26 12:51:58] Step: 5038, Training Logs: loss_final: 0.502649, loss_mean: 0.851404, proj_loss: -0.423297, loss_mean_cls: 0.074542, grad_norm: 0.396497 +Steps: 1%| | 5039/1000000 [20:55<68:22:48, 4.04it/s, grad_norm=0.396, loss_final=0.503, loss_mean=0.851, loss_mean_cls=0.0745, proj_loss=-0.423][2026-03-26 12:51:58] Step: 5039, Training Logs: loss_final: 0.494390, loss_mean: 0.820121, proj_loss: -0.406264, loss_mean_cls: 0.080533, grad_norm: 0.299259 +Steps: 1%| | 5040/1000000 [20:55<68:22:44, 4.04it/s, grad_norm=0.299, loss_final=0.494, loss_mean=0.82, loss_mean_cls=0.0805, proj_loss=-0.406][2026-03-26 12:51:59] Step: 5040, Training Logs: loss_final: 0.498626, loss_mean: 0.828226, proj_loss: -0.407243, loss_mean_cls: 0.077643, grad_norm: 0.370859 +Steps: 1%| | 5041/1000000 [20:56<68:25:15, 4.04it/s, grad_norm=0.371, loss_final=0.499, loss_mean=0.828, loss_mean_cls=0.0776, proj_loss=-0.407][2026-03-26 12:51:59] Step: 5041, Training Logs: loss_final: 0.496552, loss_mean: 0.831814, proj_loss: -0.413798, loss_mean_cls: 0.078536, grad_norm: 0.444523 +Steps: 1%| | 5042/1000000 [20:56<68:26:08, 4.04it/s, grad_norm=0.445, loss_final=0.497, loss_mean=0.832, loss_mean_cls=0.0785, proj_loss=-0.414][2026-03-26 12:51:59] Step: 5042, Training Logs: loss_final: 0.484404, loss_mean: 0.829513, proj_loss: -0.420777, loss_mean_cls: 0.075668, grad_norm: 0.254329 +Steps: 1%| | 5043/1000000 [20:56<68:22:56, 4.04it/s, grad_norm=0.254, loss_final=0.484, loss_mean=0.83, loss_mean_cls=0.0757, proj_loss=-0.421][2026-03-26 12:51:59] Step: 5043, Training Logs: loss_final: 0.481487, loss_mean: 0.816776, proj_loss: -0.413590, loss_mean_cls: 0.078300, grad_norm: 0.543518 +Steps: 1%| | 5044/1000000 [20:56<68:22:17, 4.04it/s, grad_norm=0.544, loss_final=0.481, loss_mean=0.817, loss_mean_cls=0.0783, proj_loss=-0.414][2026-03-26 12:52:00] Step: 5044, Training Logs: loss_final: 0.498015, loss_mean: 0.832155, proj_loss: -0.413164, loss_mean_cls: 0.079024, grad_norm: 0.421585 +Steps: 1%| | 5045/1000000 [20:57<68:25:57, 4.04it/s, grad_norm=0.422, loss_final=0.498, loss_mean=0.832, loss_mean_cls=0.079, proj_loss=-0.413][2026-03-26 12:52:00] Step: 5045, Training Logs: loss_final: 0.495508, loss_mean: 0.828987, proj_loss: -0.412967, loss_mean_cls: 0.079487, grad_norm: 0.290822 +Steps: 1%| | 5046/1000000 [20:57<68:26:42, 4.04it/s, grad_norm=0.291, loss_final=0.496, loss_mean=0.829, loss_mean_cls=0.0795, proj_loss=-0.413][2026-03-26 12:52:00] Step: 5046, Training Logs: loss_final: 0.496487, loss_mean: 0.831787, proj_loss: -0.414482, loss_mean_cls: 0.079182, grad_norm: 0.383462 +Steps: 1%| | 5047/1000000 [20:57<68:28:22, 4.04it/s, grad_norm=0.383, loss_final=0.496, loss_mean=0.832, loss_mean_cls=0.0792, proj_loss=-0.414][2026-03-26 12:52:00] Step: 5047, Training Logs: loss_final: 0.511108, loss_mean: 0.829126, proj_loss: -0.398695, loss_mean_cls: 0.080676, grad_norm: 0.363756 +Steps: 1%| | 5048/1000000 [20:57<68:27:39, 4.04it/s, grad_norm=0.364, loss_final=0.511, loss_mean=0.829, loss_mean_cls=0.0807, proj_loss=-0.399][2026-03-26 12:52:01] Step: 5048, Training Logs: loss_final: 0.496510, loss_mean: 0.847592, proj_loss: -0.425790, loss_mean_cls: 0.074709, grad_norm: 0.363361 +Steps: 1%| | 5049/1000000 [20:58<68:27:23, 4.04it/s, grad_norm=0.363, loss_final=0.497, loss_mean=0.848, loss_mean_cls=0.0747, proj_loss=-0.426][2026-03-26 12:52:01] Step: 5049, Training Logs: loss_final: 0.498172, loss_mean: 0.837362, proj_loss: -0.416360, loss_mean_cls: 0.077169, grad_norm: 0.388225 +Steps: 1%| | 5050/1000000 [20:58<69:20:04, 3.99it/s, grad_norm=0.388, loss_final=0.498, loss_mean=0.837, loss_mean_cls=0.0772, proj_loss=-0.416][2026-03-26 12:52:01] Step: 5050, Training Logs: loss_final: 0.498672, loss_mean: 0.831366, proj_loss: -0.411424, loss_mean_cls: 0.078729, grad_norm: 0.390316 +Steps: 1%| | 5051/1000000 [20:58<69:04:49, 4.00it/s, grad_norm=0.39, loss_final=0.499, loss_mean=0.831, loss_mean_cls=0.0787, proj_loss=-0.411][2026-03-26 12:52:01] Step: 5051, Training Logs: loss_final: 0.487651, loss_mean: 0.817029, proj_loss: -0.407551, loss_mean_cls: 0.078174, grad_norm: 0.283224 +Steps: 1%| | 5052/1000000 [20:58<69:09:27, 4.00it/s, grad_norm=0.283, loss_final=0.488, loss_mean=0.817, loss_mean_cls=0.0782, proj_loss=-0.408][2026-03-26 12:52:02] Step: 5052, Training Logs: loss_final: 0.488939, loss_mean: 0.825452, proj_loss: -0.413970, loss_mean_cls: 0.077457, grad_norm: 0.426311 +Steps: 1%| | 5053/1000000 [20:59<68:45:36, 4.02it/s, grad_norm=0.426, loss_final=0.489, loss_mean=0.825, loss_mean_cls=0.0775, proj_loss=-0.414][2026-03-26 12:52:02] Step: 5053, Training Logs: loss_final: 0.498269, loss_mean: 0.846575, proj_loss: -0.423645, loss_mean_cls: 0.075340, grad_norm: 0.382126 +Steps: 1%| | 5054/1000000 [20:59<68:39:22, 4.03it/s, grad_norm=0.382, loss_final=0.498, loss_mean=0.847, loss_mean_cls=0.0753, proj_loss=-0.424][2026-03-26 12:52:02] Step: 5054, Training Logs: loss_final: 0.488820, loss_mean: 0.816784, proj_loss: -0.407519, loss_mean_cls: 0.079555, grad_norm: 0.287261 +Steps: 1%| | 5055/1000000 [20:59<68:34:14, 4.03it/s, grad_norm=0.287, loss_final=0.489, loss_mean=0.817, loss_mean_cls=0.0796, proj_loss=-0.408][2026-03-26 12:52:02] Step: 5055, Training Logs: loss_final: 0.493420, loss_mean: 0.827322, proj_loss: -0.412010, loss_mean_cls: 0.078108, grad_norm: 0.323960 +Steps: 1%| | 5056/1000000 [20:59<68:34:31, 4.03it/s, grad_norm=0.324, loss_final=0.493, loss_mean=0.827, loss_mean_cls=0.0781, proj_loss=-0.412][2026-03-26 12:52:03] Step: 5056, Training Logs: loss_final: 0.496697, loss_mean: 0.844118, proj_loss: -0.422835, loss_mean_cls: 0.075414, grad_norm: 0.304875 +Steps: 1%| | 5057/1000000 [21:00<68:32:14, 4.03it/s, grad_norm=0.305, loss_final=0.497, loss_mean=0.844, loss_mean_cls=0.0754, proj_loss=-0.423][2026-03-26 12:52:03] Step: 5057, Training Logs: loss_final: 0.486397, loss_mean: 0.820553, proj_loss: -0.412109, loss_mean_cls: 0.077953, grad_norm: 0.363078 +Steps: 1%| | 5058/1000000 [21:00<68:26:48, 4.04it/s, grad_norm=0.363, loss_final=0.486, loss_mean=0.821, loss_mean_cls=0.078, proj_loss=-0.412][2026-03-26 12:52:03] Step: 5058, Training Logs: loss_final: 0.492098, loss_mean: 0.823512, proj_loss: -0.410231, loss_mean_cls: 0.078817, grad_norm: 0.256375 +Steps: 1%| | 5059/1000000 [21:00<68:24:00, 4.04it/s, grad_norm=0.256, loss_final=0.492, loss_mean=0.824, loss_mean_cls=0.0788, proj_loss=-0.41][2026-03-26 12:52:03] Step: 5059, Training Logs: loss_final: 0.483614, loss_mean: 0.824135, proj_loss: -0.418274, loss_mean_cls: 0.077753, grad_norm: 0.292286 +Steps: 1%| | 5060/1000000 [21:00<68:22:45, 4.04it/s, grad_norm=0.292, loss_final=0.484, loss_mean=0.824, loss_mean_cls=0.0778, proj_loss=-0.418][2026-03-26 12:52:04] Step: 5060, Training Logs: loss_final: 0.494852, loss_mean: 0.827812, proj_loss: -0.410570, loss_mean_cls: 0.077609, grad_norm: 0.468595 +Steps: 1%| | 5061/1000000 [21:01<68:21:03, 4.04it/s, grad_norm=0.469, loss_final=0.495, loss_mean=0.828, loss_mean_cls=0.0776, proj_loss=-0.411][2026-03-26 12:52:04] Step: 5061, Training Logs: loss_final: 0.496399, loss_mean: 0.814765, proj_loss: -0.399137, loss_mean_cls: 0.080771, grad_norm: 0.331287 +Steps: 1%| | 5062/1000000 [21:01<68:22:12, 4.04it/s, grad_norm=0.331, loss_final=0.496, loss_mean=0.815, loss_mean_cls=0.0808, proj_loss=-0.399][2026-03-26 12:52:04] Step: 5062, Training Logs: loss_final: 0.485687, loss_mean: 0.821404, proj_loss: -0.414754, loss_mean_cls: 0.079037, grad_norm: 0.265688 +Steps: 1%| | 5063/1000000 [21:01<68:20:50, 4.04it/s, grad_norm=0.266, loss_final=0.486, loss_mean=0.821, loss_mean_cls=0.079, proj_loss=-0.415][2026-03-26 12:52:04] Step: 5063, Training Logs: loss_final: 0.503998, loss_mean: 0.841395, proj_loss: -0.415410, loss_mean_cls: 0.078014, grad_norm: 0.387070 +Steps: 1%| | 5064/1000000 [21:01<68:22:37, 4.04it/s, grad_norm=0.387, loss_final=0.504, loss_mean=0.841, loss_mean_cls=0.078, proj_loss=-0.415][2026-03-26 12:52:05] Step: 5064, Training Logs: loss_final: 0.488173, loss_mean: 0.808382, proj_loss: -0.402238, loss_mean_cls: 0.082029, grad_norm: 0.403251 +Steps: 1%| | 5065/1000000 [21:01<68:22:44, 4.04it/s, grad_norm=0.403, loss_final=0.488, loss_mean=0.808, loss_mean_cls=0.082, proj_loss=-0.402][2026-03-26 12:52:05] Step: 5065, Training Logs: loss_final: 0.472918, loss_mean: 0.812211, proj_loss: -0.417586, loss_mean_cls: 0.078293, grad_norm: 0.261724 +Steps: 1%| | 5066/1000000 [21:02<68:23:55, 4.04it/s, grad_norm=0.262, loss_final=0.473, loss_mean=0.812, loss_mean_cls=0.0783, proj_loss=-0.418][2026-03-26 12:52:05] Step: 5066, Training Logs: loss_final: 0.493467, loss_mean: 0.830359, proj_loss: -0.415059, loss_mean_cls: 0.078167, grad_norm: 0.578912 +Steps: 1%| | 5067/1000000 [21:02<68:22:53, 4.04it/s, grad_norm=0.579, loss_final=0.493, loss_mean=0.83, loss_mean_cls=0.0782, proj_loss=-0.415][2026-03-26 12:52:05] Step: 5067, Training Logs: loss_final: 0.495922, loss_mean: 0.816875, proj_loss: -0.402017, loss_mean_cls: 0.081064, grad_norm: 0.503883 +Steps: 1%| | 5068/1000000 [21:02<68:24:00, 4.04it/s, grad_norm=0.504, loss_final=0.496, loss_mean=0.817, loss_mean_cls=0.0811, proj_loss=-0.402][2026-03-26 12:52:06] Step: 5068, Training Logs: loss_final: 0.498823, loss_mean: 0.832196, proj_loss: -0.411274, loss_mean_cls: 0.077902, grad_norm: 0.294078 +Steps: 1%| | 5069/1000000 [21:02<68:21:41, 4.04it/s, grad_norm=0.294, loss_final=0.499, loss_mean=0.832, loss_mean_cls=0.0779, proj_loss=-0.411][2026-03-26 12:52:06] Step: 5069, Training Logs: loss_final: 0.495312, loss_mean: 0.835997, proj_loss: -0.417963, loss_mean_cls: 0.077278, grad_norm: 0.636953 +Steps: 1%| | 5070/1000000 [21:03<68:22:14, 4.04it/s, grad_norm=0.637, loss_final=0.495, loss_mean=0.836, loss_mean_cls=0.0773, proj_loss=-0.418][2026-03-26 12:52:06] Step: 5070, Training Logs: loss_final: 0.489172, loss_mean: 0.826750, proj_loss: -0.414935, loss_mean_cls: 0.077357, grad_norm: 0.531343 +Steps: 1%| | 5071/1000000 [21:03<68:22:02, 4.04it/s, grad_norm=0.531, loss_final=0.489, loss_mean=0.827, loss_mean_cls=0.0774, proj_loss=-0.415][2026-03-26 12:52:06] Step: 5071, Training Logs: loss_final: 0.498178, loss_mean: 0.838227, proj_loss: -0.417580, loss_mean_cls: 0.077532, grad_norm: 0.318210 +Steps: 1%| | 5072/1000000 [21:03<68:24:34, 4.04it/s, grad_norm=0.318, loss_final=0.498, loss_mean=0.838, loss_mean_cls=0.0775, proj_loss=-0.418][2026-03-26 12:52:07] Step: 5072, Training Logs: loss_final: 0.493532, loss_mean: 0.842692, proj_loss: -0.424282, loss_mean_cls: 0.075122, grad_norm: 0.494903 +Steps: 1%| | 5073/1000000 [21:03<68:24:55, 4.04it/s, grad_norm=0.495, loss_final=0.494, loss_mean=0.843, loss_mean_cls=0.0751, proj_loss=-0.424][2026-03-26 12:52:07] Step: 5073, Training Logs: loss_final: 0.478699, loss_mean: 0.806437, proj_loss: -0.408495, loss_mean_cls: 0.080757, grad_norm: 0.427980 +Steps: 1%| | 5074/1000000 [21:04<68:26:11, 4.04it/s, grad_norm=0.428, loss_final=0.479, loss_mean=0.806, loss_mean_cls=0.0808, proj_loss=-0.408][2026-03-26 12:52:07] Step: 5074, Training Logs: loss_final: 0.493117, loss_mean: 0.830321, proj_loss: -0.414744, loss_mean_cls: 0.077540, grad_norm: 0.269552 +Steps: 1%| | 5075/1000000 [21:04<68:25:50, 4.04it/s, grad_norm=0.27, loss_final=0.493, loss_mean=0.83, loss_mean_cls=0.0775, proj_loss=-0.415][2026-03-26 12:52:07] Step: 5075, Training Logs: loss_final: 0.505205, loss_mean: 0.827253, proj_loss: -0.401735, loss_mean_cls: 0.079688, grad_norm: 0.385672 +Steps: 1%| | 5076/1000000 [21:04<68:25:52, 4.04it/s, grad_norm=0.386, loss_final=0.505, loss_mean=0.827, loss_mean_cls=0.0797, proj_loss=-0.402][2026-03-26 12:52:08] Step: 5076, Training Logs: loss_final: 0.493286, loss_mean: 0.834897, proj_loss: -0.418676, loss_mean_cls: 0.077065, grad_norm: 0.239518 +Steps: 1%| | 5077/1000000 [21:04<68:23:42, 4.04it/s, grad_norm=0.24, loss_final=0.493, loss_mean=0.835, loss_mean_cls=0.0771, proj_loss=-0.419][2026-03-26 12:52:08] Step: 5077, Training Logs: loss_final: 0.501027, loss_mean: 0.830673, proj_loss: -0.408359, loss_mean_cls: 0.078712, grad_norm: 0.347196 +Steps: 1%| | 5078/1000000 [21:05<68:22:54, 4.04it/s, grad_norm=0.347, loss_final=0.501, loss_mean=0.831, loss_mean_cls=0.0787, proj_loss=-0.408][2026-03-26 12:52:08] Step: 5078, Training Logs: loss_final: 0.491297, loss_mean: 0.825249, proj_loss: -0.411554, loss_mean_cls: 0.077602, grad_norm: 0.225111 +Steps: 1%| | 5079/1000000 [21:05<68:22:36, 4.04it/s, grad_norm=0.225, loss_final=0.491, loss_mean=0.825, loss_mean_cls=0.0776, proj_loss=-0.412][2026-03-26 12:52:08] Step: 5079, Training Logs: loss_final: 0.489019, loss_mean: 0.839976, proj_loss: -0.426396, loss_mean_cls: 0.075439, grad_norm: 0.300363 +Steps: 1%| | 5080/1000000 [21:05<68:22:34, 4.04it/s, grad_norm=0.3, loss_final=0.489, loss_mean=0.84, loss_mean_cls=0.0754, proj_loss=-0.426][2026-03-26 12:52:09] Step: 5080, Training Logs: loss_final: 0.469356, loss_mean: 0.813027, proj_loss: -0.421220, loss_mean_cls: 0.077549, grad_norm: 0.362788 +Steps: 1%| | 5081/1000000 [21:05<68:22:18, 4.04it/s, grad_norm=0.363, loss_final=0.469, loss_mean=0.813, loss_mean_cls=0.0775, proj_loss=-0.421][2026-03-26 12:52:09] Step: 5081, Training Logs: loss_final: 0.487997, loss_mean: 0.822971, proj_loss: -0.413820, loss_mean_cls: 0.078847, grad_norm: 0.316703 +Steps: 1%| | 5082/1000000 [21:06<68:22:06, 4.04it/s, grad_norm=0.317, loss_final=0.488, loss_mean=0.823, loss_mean_cls=0.0788, proj_loss=-0.414][2026-03-26 12:52:09] Step: 5082, Training Logs: loss_final: 0.489023, loss_mean: 0.822910, proj_loss: -0.413398, loss_mean_cls: 0.079511, grad_norm: 0.386493 +Steps: 1%| | 5083/1000000 [21:06<68:23:57, 4.04it/s, grad_norm=0.386, loss_final=0.489, loss_mean=0.823, loss_mean_cls=0.0795, proj_loss=-0.413][2026-03-26 12:52:09] Step: 5083, Training Logs: loss_final: 0.483012, loss_mean: 0.823574, proj_loss: -0.416791, loss_mean_cls: 0.076229, grad_norm: 0.385267 +Steps: 1%| | 5084/1000000 [21:06<68:24:07, 4.04it/s, grad_norm=0.385, loss_final=0.483, loss_mean=0.824, loss_mean_cls=0.0762, proj_loss=-0.417][2026-03-26 12:52:10] Step: 5084, Training Logs: loss_final: 0.492846, loss_mean: 0.812853, proj_loss: -0.401551, loss_mean_cls: 0.081543, grad_norm: 0.333342 +Steps: 1%| | 5085/1000000 [21:06<68:23:53, 4.04it/s, grad_norm=0.333, loss_final=0.493, loss_mean=0.813, loss_mean_cls=0.0815, proj_loss=-0.402][2026-03-26 12:52:10] Step: 5085, Training Logs: loss_final: 0.475833, loss_mean: 0.814287, proj_loss: -0.415790, loss_mean_cls: 0.077336, grad_norm: 0.428377 +Steps: 1%| | 5086/1000000 [21:07<68:21:43, 4.04it/s, grad_norm=0.428, loss_final=0.476, loss_mean=0.814, loss_mean_cls=0.0773, proj_loss=-0.416][2026-03-26 12:52:10] Step: 5086, Training Logs: loss_final: 0.499472, loss_mean: 0.835188, proj_loss: -0.413854, loss_mean_cls: 0.078139, grad_norm: 0.533060 +Steps: 1%| | 5087/1000000 [21:07<68:24:59, 4.04it/s, grad_norm=0.533, loss_final=0.499, loss_mean=0.835, loss_mean_cls=0.0781, proj_loss=-0.414][2026-03-26 12:52:10] Step: 5087, Training Logs: loss_final: 0.498498, loss_mean: 0.834951, proj_loss: -0.415145, loss_mean_cls: 0.078692, grad_norm: 0.290656 +Steps: 1%| | 5088/1000000 [21:07<68:25:57, 4.04it/s, grad_norm=0.291, loss_final=0.498, loss_mean=0.835, loss_mean_cls=0.0787, proj_loss=-0.415][2026-03-26 12:52:11] Step: 5088, Training Logs: loss_final: 0.501930, loss_mean: 0.838067, proj_loss: -0.413837, loss_mean_cls: 0.077700, grad_norm: 0.334573 +Steps: 1%| | 5089/1000000 [21:07<68:24:33, 4.04it/s, grad_norm=0.335, loss_final=0.502, loss_mean=0.838, loss_mean_cls=0.0777, proj_loss=-0.414][2026-03-26 12:52:11] Step: 5089, Training Logs: loss_final: 0.492298, loss_mean: 0.819558, proj_loss: -0.407821, loss_mean_cls: 0.080561, grad_norm: 0.312963 +Steps: 1%| | 5090/1000000 [21:08<68:26:47, 4.04it/s, grad_norm=0.313, loss_final=0.492, loss_mean=0.82, loss_mean_cls=0.0806, proj_loss=-0.408][2026-03-26 12:52:11] Step: 5090, Training Logs: loss_final: 0.490667, loss_mean: 0.838586, proj_loss: -0.423195, loss_mean_cls: 0.075277, grad_norm: 0.297507 +Steps: 1%| | 5091/1000000 [21:08<68:26:41, 4.04it/s, grad_norm=0.298, loss_final=0.491, loss_mean=0.839, loss_mean_cls=0.0753, proj_loss=-0.423][2026-03-26 12:52:11] Step: 5091, Training Logs: loss_final: 0.498895, loss_mean: 0.834725, proj_loss: -0.413702, loss_mean_cls: 0.077871, grad_norm: 0.326451 +Steps: 1%| | 5092/1000000 [21:08<68:28:29, 4.04it/s, grad_norm=0.326, loss_final=0.499, loss_mean=0.835, loss_mean_cls=0.0779, proj_loss=-0.414][2026-03-26 12:52:12] Step: 5092, Training Logs: loss_final: 0.482059, loss_mean: 0.812064, proj_loss: -0.409478, loss_mean_cls: 0.079472, grad_norm: 0.304633 +Steps: 1%| | 5093/1000000 [21:08<68:27:26, 4.04it/s, grad_norm=0.305, loss_final=0.482, loss_mean=0.812, loss_mean_cls=0.0795, proj_loss=-0.409][2026-03-26 12:52:12] Step: 5093, Training Logs: loss_final: 0.494017, loss_mean: 0.819587, proj_loss: -0.404354, loss_mean_cls: 0.078784, grad_norm: 0.458980 +Steps: 1%| | 5094/1000000 [21:09<68:26:26, 4.04it/s, grad_norm=0.459, loss_final=0.494, loss_mean=0.82, loss_mean_cls=0.0788, proj_loss=-0.404][2026-03-26 12:52:12] Step: 5094, Training Logs: loss_final: 0.479937, loss_mean: 0.799986, proj_loss: -0.401810, loss_mean_cls: 0.081760, grad_norm: 0.361699 +Steps: 1%| | 5095/1000000 [21:09<68:27:08, 4.04it/s, grad_norm=0.362, loss_final=0.48, loss_mean=0.8, loss_mean_cls=0.0818, proj_loss=-0.402][2026-03-26 12:52:12] Step: 5095, Training Logs: loss_final: 0.491767, loss_mean: 0.835932, proj_loss: -0.420978, loss_mean_cls: 0.076814, grad_norm: 0.453930 +Steps: 1%| | 5096/1000000 [21:09<68:26:39, 4.04it/s, grad_norm=0.454, loss_final=0.492, loss_mean=0.836, loss_mean_cls=0.0768, proj_loss=-0.421][2026-03-26 12:52:13] Step: 5096, Training Logs: loss_final: 0.504490, loss_mean: 0.837500, proj_loss: -0.411223, loss_mean_cls: 0.078213, grad_norm: 0.485245 +Steps: 1%| | 5097/1000000 [21:09<68:24:07, 4.04it/s, grad_norm=0.485, loss_final=0.504, loss_mean=0.838, loss_mean_cls=0.0782, proj_loss=-0.411][2026-03-26 12:52:13] Step: 5097, Training Logs: loss_final: 0.478749, loss_mean: 0.821658, proj_loss: -0.420068, loss_mean_cls: 0.077159, grad_norm: 0.325263 +Steps: 1%| | 5098/1000000 [21:10<68:23:51, 4.04it/s, grad_norm=0.325, loss_final=0.479, loss_mean=0.822, loss_mean_cls=0.0772, proj_loss=-0.42][2026-03-26 12:52:13] Step: 5098, Training Logs: loss_final: 0.488660, loss_mean: 0.819176, proj_loss: -0.409127, loss_mean_cls: 0.078611, grad_norm: 0.506765 +Steps: 1%| | 5099/1000000 [21:10<68:23:02, 4.04it/s, grad_norm=0.507, loss_final=0.489, loss_mean=0.819, loss_mean_cls=0.0786, proj_loss=-0.409][2026-03-26 12:52:13] Step: 5099, Training Logs: loss_final: 0.472832, loss_mean: 0.803486, proj_loss: -0.410770, loss_mean_cls: 0.080115, grad_norm: 0.317551 +Steps: 1%| | 5100/1000000 [21:10<69:13:50, 3.99it/s, grad_norm=0.318, loss_final=0.473, loss_mean=0.803, loss_mean_cls=0.0801, proj_loss=-0.411][2026-03-26 12:52:14] Step: 5100, Training Logs: loss_final: 0.498226, loss_mean: 0.827188, proj_loss: -0.408080, loss_mean_cls: 0.079119, grad_norm: 0.500645 +Steps: 1%| | 5101/1000000 [21:10<69:03:54, 4.00it/s, grad_norm=0.501, loss_final=0.498, loss_mean=0.827, loss_mean_cls=0.0791, proj_loss=-0.408][2026-03-26 12:52:14] Step: 5101, Training Logs: loss_final: 0.491063, loss_mean: 0.821489, proj_loss: -0.409538, loss_mean_cls: 0.079112, grad_norm: 0.515079 +Steps: 1%| | 5102/1000000 [21:11<68:51:54, 4.01it/s, grad_norm=0.515, loss_final=0.491, loss_mean=0.821, loss_mean_cls=0.0791, proj_loss=-0.41][2026-03-26 12:52:14] Step: 5102, Training Logs: loss_final: 0.508741, loss_mean: 0.845560, proj_loss: -0.413324, loss_mean_cls: 0.076505, grad_norm: 0.302039 +Steps: 1%| | 5103/1000000 [21:11<68:42:41, 4.02it/s, grad_norm=0.302, loss_final=0.509, loss_mean=0.846, loss_mean_cls=0.0765, proj_loss=-0.413][2026-03-26 12:52:14] Step: 5103, Training Logs: loss_final: 0.484854, loss_mean: 0.824316, proj_loss: -0.415468, loss_mean_cls: 0.076006, grad_norm: 0.451744 +Steps: 1%| | 5104/1000000 [21:11<68:36:50, 4.03it/s, grad_norm=0.452, loss_final=0.485, loss_mean=0.824, loss_mean_cls=0.076, proj_loss=-0.415][2026-03-26 12:52:15] Step: 5104, Training Logs: loss_final: 0.504014, loss_mean: 0.839804, proj_loss: -0.413447, loss_mean_cls: 0.077657, grad_norm: 0.331141 +Steps: 1%| | 5105/1000000 [21:11<68:35:50, 4.03it/s, grad_norm=0.331, loss_final=0.504, loss_mean=0.84, loss_mean_cls=0.0777, proj_loss=-0.413][2026-03-26 12:52:15] Step: 5105, Training Logs: loss_final: 0.502753, loss_mean: 0.836856, proj_loss: -0.412759, loss_mean_cls: 0.078656, grad_norm: 0.526564 +Steps: 1%| | 5106/1000000 [21:12<68:32:19, 4.03it/s, grad_norm=0.527, loss_final=0.503, loss_mean=0.837, loss_mean_cls=0.0787, proj_loss=-0.413][2026-03-26 12:52:15] Step: 5106, Training Logs: loss_final: 0.497823, loss_mean: 0.836172, proj_loss: -0.414926, loss_mean_cls: 0.076576, grad_norm: 0.649893 +Steps: 1%| | 5107/1000000 [21:12<68:28:35, 4.04it/s, grad_norm=0.65, loss_final=0.498, loss_mean=0.836, loss_mean_cls=0.0766, proj_loss=-0.415][2026-03-26 12:52:15] Step: 5107, Training Logs: loss_final: 0.493637, loss_mean: 0.837958, proj_loss: -0.420335, loss_mean_cls: 0.076014, grad_norm: 0.363048 +Steps: 1%| | 5108/1000000 [21:12<68:29:45, 4.03it/s, grad_norm=0.363, loss_final=0.494, loss_mean=0.838, loss_mean_cls=0.076, proj_loss=-0.42][2026-03-26 12:52:16] Step: 5108, Training Logs: loss_final: 0.510869, loss_mean: 0.844836, proj_loss: -0.411963, loss_mean_cls: 0.077996, grad_norm: 0.617991 +Steps: 1%| | 5109/1000000 [21:12<68:27:17, 4.04it/s, grad_norm=0.618, loss_final=0.511, loss_mean=0.845, loss_mean_cls=0.078, proj_loss=-0.412][2026-03-26 12:52:16] Step: 5109, Training Logs: loss_final: 0.488022, loss_mean: 0.809714, proj_loss: -0.402923, loss_mean_cls: 0.081232, grad_norm: 0.380407 +Steps: 1%| | 5110/1000000 [21:13<68:26:24, 4.04it/s, grad_norm=0.38, loss_final=0.488, loss_mean=0.81, loss_mean_cls=0.0812, proj_loss=-0.403][2026-03-26 12:52:16] Step: 5110, Training Logs: loss_final: 0.485471, loss_mean: 0.817068, proj_loss: -0.410347, loss_mean_cls: 0.078750, grad_norm: 0.544761 +Steps: 1%| | 5111/1000000 [21:13<68:24:21, 4.04it/s, grad_norm=0.545, loss_final=0.485, loss_mean=0.817, loss_mean_cls=0.0787, proj_loss=-0.41][2026-03-26 12:52:16] Step: 5111, Training Logs: loss_final: 0.493329, loss_mean: 0.821554, proj_loss: -0.408497, loss_mean_cls: 0.080272, grad_norm: 0.473605 +Steps: 1%| | 5112/1000000 [21:13<68:26:58, 4.04it/s, grad_norm=0.474, loss_final=0.493, loss_mean=0.822, loss_mean_cls=0.0803, proj_loss=-0.408][2026-03-26 12:52:17] Step: 5112, Training Logs: loss_final: 0.497054, loss_mean: 0.833156, proj_loss: -0.413514, loss_mean_cls: 0.077412, grad_norm: 0.307081 +Steps: 1%| | 5113/1000000 [21:13<68:27:32, 4.04it/s, grad_norm=0.307, loss_final=0.497, loss_mean=0.833, loss_mean_cls=0.0774, proj_loss=-0.414][2026-03-26 12:52:17] Step: 5113, Training Logs: loss_final: 0.499197, loss_mean: 0.822384, proj_loss: -0.403731, loss_mean_cls: 0.080544, grad_norm: 0.597294 +Steps: 1%| | 5114/1000000 [21:14<68:25:41, 4.04it/s, grad_norm=0.597, loss_final=0.499, loss_mean=0.822, loss_mean_cls=0.0805, proj_loss=-0.404][2026-03-26 12:52:17] Step: 5114, Training Logs: loss_final: 0.500986, loss_mean: 0.834348, proj_loss: -0.411327, loss_mean_cls: 0.077965, grad_norm: 0.482430 +Steps: 1%| | 5115/1000000 [21:14<68:24:07, 4.04it/s, grad_norm=0.482, loss_final=0.501, loss_mean=0.834, loss_mean_cls=0.078, proj_loss=-0.411][2026-03-26 12:52:17] Step: 5115, Training Logs: loss_final: 0.504227, loss_mean: 0.847334, proj_loss: -0.419586, loss_mean_cls: 0.076479, grad_norm: 0.382934 +Steps: 1%| | 5116/1000000 [21:14<68:25:32, 4.04it/s, grad_norm=0.383, loss_final=0.504, loss_mean=0.847, loss_mean_cls=0.0765, proj_loss=-0.42][2026-03-26 12:52:18] Step: 5116, Training Logs: loss_final: 0.498404, loss_mean: 0.821199, proj_loss: -0.404175, loss_mean_cls: 0.081380, grad_norm: 0.393105 +Steps: 1%| | 5117/1000000 [21:14<68:25:13, 4.04it/s, grad_norm=0.393, loss_final=0.498, loss_mean=0.821, loss_mean_cls=0.0814, proj_loss=-0.404][2026-03-26 12:52:18] Step: 5117, Training Logs: loss_final: 0.493788, loss_mean: 0.822889, proj_loss: -0.408366, loss_mean_cls: 0.079264, grad_norm: 0.432789 +Steps: 1%| | 5118/1000000 [21:15<68:25:21, 4.04it/s, grad_norm=0.433, loss_final=0.494, loss_mean=0.823, loss_mean_cls=0.0793, proj_loss=-0.408][2026-03-26 12:52:18] Step: 5118, Training Logs: loss_final: 0.491872, loss_mean: 0.831740, proj_loss: -0.416153, loss_mean_cls: 0.076285, grad_norm: 0.342363 +Steps: 1%| | 5119/1000000 [21:15<68:23:27, 4.04it/s, grad_norm=0.342, loss_final=0.492, loss_mean=0.832, loss_mean_cls=0.0763, proj_loss=-0.416][2026-03-26 12:52:18] Step: 5119, Training Logs: loss_final: 0.476925, loss_mean: 0.821690, proj_loss: -0.420662, loss_mean_cls: 0.075896, grad_norm: 0.323044 +Steps: 1%| | 5120/1000000 [21:15<68:23:51, 4.04it/s, grad_norm=0.323, loss_final=0.477, loss_mean=0.822, loss_mean_cls=0.0759, proj_loss=-0.421][2026-03-26 12:52:19] Step: 5120, Training Logs: loss_final: 0.493217, loss_mean: 0.834263, proj_loss: -0.417065, loss_mean_cls: 0.076019, grad_norm: 0.280467 +Steps: 1%| | 5121/1000000 [21:15<68:26:44, 4.04it/s, grad_norm=0.28, loss_final=0.493, loss_mean=0.834, loss_mean_cls=0.076, proj_loss=-0.417][2026-03-26 12:52:19] Step: 5121, Training Logs: loss_final: 0.473955, loss_mean: 0.813383, proj_loss: -0.416738, loss_mean_cls: 0.077310, grad_norm: 0.447552 +Steps: 1%| | 5122/1000000 [21:16<68:25:09, 4.04it/s, grad_norm=0.448, loss_final=0.474, loss_mean=0.813, loss_mean_cls=0.0773, proj_loss=-0.417][2026-03-26 12:52:19] Step: 5122, Training Logs: loss_final: 0.479464, loss_mean: 0.815639, proj_loss: -0.413859, loss_mean_cls: 0.077684, grad_norm: 0.239380 +Steps: 1%| | 5123/1000000 [21:16<68:25:04, 4.04it/s, grad_norm=0.239, loss_final=0.479, loss_mean=0.816, loss_mean_cls=0.0777, proj_loss=-0.414][2026-03-26 12:52:19] Step: 5123, Training Logs: loss_final: 0.468872, loss_mean: 0.809683, proj_loss: -0.418591, loss_mean_cls: 0.077780, grad_norm: 0.608039 +Steps: 1%| | 5124/1000000 [21:16<68:25:56, 4.04it/s, grad_norm=0.608, loss_final=0.469, loss_mean=0.81, loss_mean_cls=0.0778, proj_loss=-0.419][2026-03-26 12:52:20] Step: 5124, Training Logs: loss_final: 0.484323, loss_mean: 0.817744, proj_loss: -0.412250, loss_mean_cls: 0.078830, grad_norm: 0.508908 +Steps: 1%| | 5125/1000000 [21:16<68:26:17, 4.04it/s, grad_norm=0.509, loss_final=0.484, loss_mean=0.818, loss_mean_cls=0.0788, proj_loss=-0.412][2026-03-26 12:52:20] Step: 5125, Training Logs: loss_final: 0.487110, loss_mean: 0.821982, proj_loss: -0.413608, loss_mean_cls: 0.078736, grad_norm: 0.231991 +Steps: 1%| | 5126/1000000 [21:17<68:26:44, 4.04it/s, grad_norm=0.232, loss_final=0.487, loss_mean=0.822, loss_mean_cls=0.0787, proj_loss=-0.414][2026-03-26 12:52:20] Step: 5126, Training Logs: loss_final: 0.483877, loss_mean: 0.829695, proj_loss: -0.420669, loss_mean_cls: 0.074851, grad_norm: 0.667103 +Steps: 1%| | 5127/1000000 [21:17<68:25:32, 4.04it/s, grad_norm=0.667, loss_final=0.484, loss_mean=0.83, loss_mean_cls=0.0749, proj_loss=-0.421][2026-03-26 12:52:20] Step: 5127, Training Logs: loss_final: 0.498269, loss_mean: 0.835929, proj_loss: -0.414011, loss_mean_cls: 0.076351, grad_norm: 0.537539 +Steps: 1%| | 5128/1000000 [21:17<68:22:49, 4.04it/s, grad_norm=0.538, loss_final=0.498, loss_mean=0.836, loss_mean_cls=0.0764, proj_loss=-0.414][2026-03-26 12:52:20] Step: 5128, Training Logs: loss_final: 0.486732, loss_mean: 0.833433, proj_loss: -0.422467, loss_mean_cls: 0.075766, grad_norm: 0.460496 +Steps: 1%| | 5129/1000000 [21:17<68:27:00, 4.04it/s, grad_norm=0.46, loss_final=0.487, loss_mean=0.833, loss_mean_cls=0.0758, proj_loss=-0.422][2026-03-26 12:52:21] Step: 5129, Training Logs: loss_final: 0.488254, loss_mean: 0.814088, proj_loss: -0.405424, loss_mean_cls: 0.079590, grad_norm: 0.880795 +Steps: 1%| | 5130/1000000 [21:18<68:24:30, 4.04it/s, grad_norm=0.881, loss_final=0.488, loss_mean=0.814, loss_mean_cls=0.0796, proj_loss=-0.405][2026-03-26 12:52:21] Step: 5130, Training Logs: loss_final: 0.485702, loss_mean: 0.811602, proj_loss: -0.405924, loss_mean_cls: 0.080024, grad_norm: 0.339782 +Steps: 1%| | 5131/1000000 [21:18<68:22:12, 4.04it/s, grad_norm=0.34, loss_final=0.486, loss_mean=0.812, loss_mean_cls=0.08, proj_loss=-0.406][2026-03-26 12:52:21] Step: 5131, Training Logs: loss_final: 0.490874, loss_mean: 0.836431, proj_loss: -0.421875, loss_mean_cls: 0.076318, grad_norm: 0.553380 +Steps: 1%| | 5132/1000000 [21:18<68:22:53, 4.04it/s, grad_norm=0.553, loss_final=0.491, loss_mean=0.836, loss_mean_cls=0.0763, proj_loss=-0.422][2026-03-26 12:52:21] Step: 5132, Training Logs: loss_final: 0.488738, loss_mean: 0.822350, proj_loss: -0.411742, loss_mean_cls: 0.078131, grad_norm: 0.577077 +Steps: 1%| | 5133/1000000 [21:18<68:24:54, 4.04it/s, grad_norm=0.577, loss_final=0.489, loss_mean=0.822, loss_mean_cls=0.0781, proj_loss=-0.412][2026-03-26 12:52:22] Step: 5133, Training Logs: loss_final: 0.490665, loss_mean: 0.830080, proj_loss: -0.417162, loss_mean_cls: 0.077747, grad_norm: 0.291140 +Steps: 1%| | 5134/1000000 [21:19<68:25:30, 4.04it/s, grad_norm=0.291, loss_final=0.491, loss_mean=0.83, loss_mean_cls=0.0777, proj_loss=-0.417][2026-03-26 12:52:22] Step: 5134, Training Logs: loss_final: 0.485714, loss_mean: 0.815508, proj_loss: -0.410310, loss_mean_cls: 0.080515, grad_norm: 0.803766 +Steps: 1%| | 5135/1000000 [21:19<68:26:39, 4.04it/s, grad_norm=0.804, loss_final=0.486, loss_mean=0.816, loss_mean_cls=0.0805, proj_loss=-0.41][2026-03-26 12:52:22] Step: 5135, Training Logs: loss_final: 0.510197, loss_mean: 0.846621, proj_loss: -0.413205, loss_mean_cls: 0.076781, grad_norm: 0.637990 +Steps: 1%| | 5136/1000000 [21:19<68:24:08, 4.04it/s, grad_norm=0.638, loss_final=0.51, loss_mean=0.847, loss_mean_cls=0.0768, proj_loss=-0.413][2026-03-26 12:52:22] Step: 5136, Training Logs: loss_final: 0.514224, loss_mean: 0.840350, proj_loss: -0.403731, loss_mean_cls: 0.077606, grad_norm: 0.292271 +Steps: 1%| | 5137/1000000 [21:19<68:23:53, 4.04it/s, grad_norm=0.292, loss_final=0.514, loss_mean=0.84, loss_mean_cls=0.0776, proj_loss=-0.404][2026-03-26 12:52:23] Step: 5137, Training Logs: loss_final: 0.496931, loss_mean: 0.837249, proj_loss: -0.416125, loss_mean_cls: 0.075807, grad_norm: 0.471155 +Steps: 1%| | 5138/1000000 [21:20<68:23:19, 4.04it/s, grad_norm=0.471, loss_final=0.497, loss_mean=0.837, loss_mean_cls=0.0758, proj_loss=-0.416][2026-03-26 12:52:23] Step: 5138, Training Logs: loss_final: 0.476934, loss_mean: 0.818465, proj_loss: -0.418945, loss_mean_cls: 0.077414, grad_norm: 0.365299 +Steps: 1%| | 5139/1000000 [21:20<68:23:05, 4.04it/s, grad_norm=0.365, loss_final=0.477, loss_mean=0.818, loss_mean_cls=0.0774, proj_loss=-0.419][2026-03-26 12:52:23] Step: 5139, Training Logs: loss_final: 0.505064, loss_mean: 0.844529, proj_loss: -0.416379, loss_mean_cls: 0.076915, grad_norm: 0.384802 +Steps: 1%| | 5140/1000000 [21:20<68:23:46, 4.04it/s, grad_norm=0.385, loss_final=0.505, loss_mean=0.845, loss_mean_cls=0.0769, proj_loss=-0.416][2026-03-26 12:52:23] Step: 5140, Training Logs: loss_final: 0.478496, loss_mean: 0.810729, proj_loss: -0.410797, loss_mean_cls: 0.078565, grad_norm: 0.416362 +Steps: 1%| | 5141/1000000 [21:20<68:22:49, 4.04it/s, grad_norm=0.416, loss_final=0.478, loss_mean=0.811, loss_mean_cls=0.0786, proj_loss=-0.411][2026-03-26 12:52:24] Step: 5141, Training Logs: loss_final: 0.487794, loss_mean: 0.825971, proj_loss: -0.414913, loss_mean_cls: 0.076736, grad_norm: 0.417090 +Steps: 1%| | 5142/1000000 [21:21<68:25:38, 4.04it/s, grad_norm=0.417, loss_final=0.488, loss_mean=0.826, loss_mean_cls=0.0767, proj_loss=-0.415][2026-03-26 12:52:24] Step: 5142, Training Logs: loss_final: 0.488766, loss_mean: 0.825008, proj_loss: -0.413865, loss_mean_cls: 0.077623, grad_norm: 0.396247 +Steps: 1%| | 5143/1000000 [21:21<68:25:32, 4.04it/s, grad_norm=0.396, loss_final=0.489, loss_mean=0.825, loss_mean_cls=0.0776, proj_loss=-0.414][2026-03-26 12:52:24] Step: 5143, Training Logs: loss_final: 0.494505, loss_mean: 0.827013, proj_loss: -0.409722, loss_mean_cls: 0.077214, grad_norm: 0.539914 +Steps: 1%| | 5144/1000000 [21:21<68:25:14, 4.04it/s, grad_norm=0.54, loss_final=0.495, loss_mean=0.827, loss_mean_cls=0.0772, proj_loss=-0.41][2026-03-26 12:52:24] Step: 5144, Training Logs: loss_final: 0.487526, loss_mean: 0.810963, proj_loss: -0.403238, loss_mean_cls: 0.079801, grad_norm: 0.478064 +Steps: 1%| | 5145/1000000 [21:21<68:27:44, 4.04it/s, grad_norm=0.478, loss_final=0.488, loss_mean=0.811, loss_mean_cls=0.0798, proj_loss=-0.403][2026-03-26 12:52:25] Step: 5145, Training Logs: loss_final: 0.506266, loss_mean: 0.841358, proj_loss: -0.411800, loss_mean_cls: 0.076707, grad_norm: 0.370076 +Steps: 1%| | 5146/1000000 [21:22<68:26:38, 4.04it/s, grad_norm=0.37, loss_final=0.506, loss_mean=0.841, loss_mean_cls=0.0767, proj_loss=-0.412][2026-03-26 12:52:25] Step: 5146, Training Logs: loss_final: 0.470374, loss_mean: 0.803186, proj_loss: -0.410766, loss_mean_cls: 0.077953, grad_norm: 0.400002 +Steps: 1%| | 5147/1000000 [21:22<68:25:43, 4.04it/s, grad_norm=0.4, loss_final=0.47, loss_mean=0.803, loss_mean_cls=0.078, proj_loss=-0.411][2026-03-26 12:52:25] Step: 5147, Training Logs: loss_final: 0.502450, loss_mean: 0.831870, proj_loss: -0.407752, loss_mean_cls: 0.078331, grad_norm: 0.369812 +Steps: 1%| | 5148/1000000 [21:22<68:24:18, 4.04it/s, grad_norm=0.37, loss_final=0.502, loss_mean=0.832, loss_mean_cls=0.0783, proj_loss=-0.408][2026-03-26 12:52:25] Step: 5148, Training Logs: loss_final: 0.479709, loss_mean: 0.815593, proj_loss: -0.413043, loss_mean_cls: 0.077159, grad_norm: 0.608720 +Steps: 1%| | 5149/1000000 [21:22<68:23:15, 4.04it/s, grad_norm=0.609, loss_final=0.48, loss_mean=0.816, loss_mean_cls=0.0772, proj_loss=-0.413][2026-03-26 12:52:26] Step: 5149, Training Logs: loss_final: 0.501014, loss_mean: 0.821381, proj_loss: -0.400417, loss_mean_cls: 0.080050, grad_norm: 0.344044 +Steps: 1%| | 5150/1000000 [21:23<68:24:16, 4.04it/s, grad_norm=0.344, loss_final=0.501, loss_mean=0.821, loss_mean_cls=0.0801, proj_loss=-0.4][2026-03-26 12:52:26] Step: 5150, Training Logs: loss_final: 0.493467, loss_mean: 0.825859, proj_loss: -0.410611, loss_mean_cls: 0.078218, grad_norm: 0.644617 +Steps: 1%| | 5151/1000000 [21:23<68:22:48, 4.04it/s, grad_norm=0.645, loss_final=0.493, loss_mean=0.826, loss_mean_cls=0.0782, proj_loss=-0.411][2026-03-26 12:52:26] Step: 5151, Training Logs: loss_final: 0.489791, loss_mean: 0.818103, proj_loss: -0.406894, loss_mean_cls: 0.078582, grad_norm: 0.487475 +Steps: 1%| | 5152/1000000 [21:23<68:24:23, 4.04it/s, grad_norm=0.487, loss_final=0.49, loss_mean=0.818, loss_mean_cls=0.0786, proj_loss=-0.407][2026-03-26 12:52:26] Step: 5152, Training Logs: loss_final: 0.504674, loss_mean: 0.829299, proj_loss: -0.404539, loss_mean_cls: 0.079915, grad_norm: 0.379603 +Steps: 1%| | 5153/1000000 [21:23<68:24:27, 4.04it/s, grad_norm=0.38, loss_final=0.505, loss_mean=0.829, loss_mean_cls=0.0799, proj_loss=-0.405][2026-03-26 12:52:27] Step: 5153, Training Logs: loss_final: 0.486855, loss_mean: 0.819307, proj_loss: -0.410536, loss_mean_cls: 0.078085, grad_norm: 0.493962 +Steps: 1%| | 5154/1000000 [21:24<68:25:46, 4.04it/s, grad_norm=0.494, loss_final=0.487, loss_mean=0.819, loss_mean_cls=0.0781, proj_loss=-0.411][2026-03-26 12:52:27] Step: 5154, Training Logs: loss_final: 0.492426, loss_mean: 0.832200, proj_loss: -0.415859, loss_mean_cls: 0.076085, grad_norm: 0.236948 +Steps: 1%| | 5155/1000000 [21:24<68:22:08, 4.04it/s, grad_norm=0.237, loss_final=0.492, loss_mean=0.832, loss_mean_cls=0.0761, proj_loss=-0.416][2026-03-26 12:52:27] Step: 5155, Training Logs: loss_final: 0.509710, loss_mean: 0.850311, proj_loss: -0.417281, loss_mean_cls: 0.076680, grad_norm: 0.339183 +Steps: 1%| | 5156/1000000 [21:24<68:25:31, 4.04it/s, grad_norm=0.339, loss_final=0.51, loss_mean=0.85, loss_mean_cls=0.0767, proj_loss=-0.417][2026-03-26 12:52:27] Step: 5156, Training Logs: loss_final: 0.493571, loss_mean: 0.833966, proj_loss: -0.415513, loss_mean_cls: 0.075119, grad_norm: 0.319436 +Steps: 1%| | 5157/1000000 [21:24<68:26:48, 4.04it/s, grad_norm=0.319, loss_final=0.494, loss_mean=0.834, loss_mean_cls=0.0751, proj_loss=-0.416][2026-03-26 12:52:28] Step: 5157, Training Logs: loss_final: 0.496532, loss_mean: 0.820082, proj_loss: -0.403579, loss_mean_cls: 0.080029, grad_norm: 0.429605 +Steps: 1%| | 5158/1000000 [21:25<68:26:46, 4.04it/s, grad_norm=0.43, loss_final=0.497, loss_mean=0.82, loss_mean_cls=0.08, proj_loss=-0.404][2026-03-26 12:52:28] Step: 5158, Training Logs: loss_final: 0.493336, loss_mean: 0.847253, proj_loss: -0.426253, loss_mean_cls: 0.072335, grad_norm: 0.230225 +Steps: 1%| | 5159/1000000 [21:25<68:24:29, 4.04it/s, grad_norm=0.23, loss_final=0.493, loss_mean=0.847, loss_mean_cls=0.0723, proj_loss=-0.426][2026-03-26 12:52:28] Step: 5159, Training Logs: loss_final: 0.499129, loss_mean: 0.838639, proj_loss: -0.416892, loss_mean_cls: 0.077382, grad_norm: 0.360342 +Steps: 1%| | 5160/1000000 [21:25<68:25:12, 4.04it/s, grad_norm=0.36, loss_final=0.499, loss_mean=0.839, loss_mean_cls=0.0774, proj_loss=-0.417][2026-03-26 12:52:28] Step: 5160, Training Logs: loss_final: 0.488516, loss_mean: 0.813546, proj_loss: -0.404434, loss_mean_cls: 0.079404, grad_norm: 0.279742 +Steps: 1%| | 5161/1000000 [21:25<68:26:41, 4.04it/s, grad_norm=0.28, loss_final=0.489, loss_mean=0.814, loss_mean_cls=0.0794, proj_loss=-0.404][2026-03-26 12:52:29] Step: 5161, Training Logs: loss_final: 0.477996, loss_mean: 0.827741, proj_loss: -0.424839, loss_mean_cls: 0.075094, grad_norm: 0.336644 +Steps: 1%| | 5162/1000000 [21:26<68:25:11, 4.04it/s, grad_norm=0.337, loss_final=0.478, loss_mean=0.828, loss_mean_cls=0.0751, proj_loss=-0.425][2026-03-26 12:52:29] Step: 5162, Training Logs: loss_final: 0.493904, loss_mean: 0.831683, proj_loss: -0.415269, loss_mean_cls: 0.077490, grad_norm: 0.371791 +Steps: 1%| | 5163/1000000 [21:26<68:25:11, 4.04it/s, grad_norm=0.372, loss_final=0.494, loss_mean=0.832, loss_mean_cls=0.0775, proj_loss=-0.415][2026-03-26 12:52:29] Step: 5163, Training Logs: loss_final: 0.469613, loss_mean: 0.805883, proj_loss: -0.414223, loss_mean_cls: 0.077953, grad_norm: 0.258045 +Steps: 1%| | 5164/1000000 [21:26<68:25:11, 4.04it/s, grad_norm=0.258, loss_final=0.47, loss_mean=0.806, loss_mean_cls=0.078, proj_loss=-0.414][2026-03-26 12:52:29] Step: 5164, Training Logs: loss_final: 0.489418, loss_mean: 0.815231, proj_loss: -0.404129, loss_mean_cls: 0.078317, grad_norm: 0.413163 +Steps: 1%| | 5165/1000000 [21:26<68:26:25, 4.04it/s, grad_norm=0.413, loss_final=0.489, loss_mean=0.815, loss_mean_cls=0.0783, proj_loss=-0.404][2026-03-26 12:52:30] Step: 5165, Training Logs: loss_final: 0.482751, loss_mean: 0.821997, proj_loss: -0.414774, loss_mean_cls: 0.075528, grad_norm: 0.246466 +Steps: 1%| | 5166/1000000 [21:27<68:25:53, 4.04it/s, grad_norm=0.246, loss_final=0.483, loss_mean=0.822, loss_mean_cls=0.0755, proj_loss=-0.415][2026-03-26 12:52:30] Step: 5166, Training Logs: loss_final: 0.491981, loss_mean: 0.839505, proj_loss: -0.421368, loss_mean_cls: 0.073845, grad_norm: 0.465770 +Steps: 1%| | 5167/1000000 [21:27<68:26:29, 4.04it/s, grad_norm=0.466, loss_final=0.492, loss_mean=0.84, loss_mean_cls=0.0738, proj_loss=-0.421][2026-03-26 12:52:30] Step: 5167, Training Logs: loss_final: 0.497847, loss_mean: 0.825740, proj_loss: -0.407524, loss_mean_cls: 0.079632, grad_norm: 0.551540 +Steps: 1%| | 5168/1000000 [21:27<68:22:37, 4.04it/s, grad_norm=0.552, loss_final=0.498, loss_mean=0.826, loss_mean_cls=0.0796, proj_loss=-0.408][2026-03-26 12:52:30] Step: 5168, Training Logs: loss_final: 0.482992, loss_mean: 0.814415, proj_loss: -0.410470, loss_mean_cls: 0.079047, grad_norm: 0.224463 +Steps: 1%| | 5169/1000000 [21:27<68:25:29, 4.04it/s, grad_norm=0.224, loss_final=0.483, loss_mean=0.814, loss_mean_cls=0.079, proj_loss=-0.41][2026-03-26 12:52:31] Step: 5169, Training Logs: loss_final: 0.498367, loss_mean: 0.836628, proj_loss: -0.413772, loss_mean_cls: 0.075511, grad_norm: 0.368581 +Steps: 1%| | 5170/1000000 [21:27<68:25:01, 4.04it/s, grad_norm=0.369, loss_final=0.498, loss_mean=0.837, loss_mean_cls=0.0755, proj_loss=-0.414][2026-03-26 12:52:31] Step: 5170, Training Logs: loss_final: 0.503212, loss_mean: 0.839366, proj_loss: -0.412429, loss_mean_cls: 0.076275, grad_norm: 0.333072 +Steps: 1%| | 5171/1000000 [21:28<68:25:08, 4.04it/s, grad_norm=0.333, loss_final=0.503, loss_mean=0.839, loss_mean_cls=0.0763, proj_loss=-0.412][2026-03-26 12:52:31] Step: 5171, Training Logs: loss_final: 0.488329, loss_mean: 0.824753, proj_loss: -0.411205, loss_mean_cls: 0.074781, grad_norm: 0.308257 +Steps: 1%| | 5172/1000000 [21:28<68:23:08, 4.04it/s, grad_norm=0.308, loss_final=0.488, loss_mean=0.825, loss_mean_cls=0.0748, proj_loss=-0.411][2026-03-26 12:52:31] Step: 5172, Training Logs: loss_final: 0.494040, loss_mean: 0.829026, proj_loss: -0.411505, loss_mean_cls: 0.076520, grad_norm: 0.264473 +Steps: 1%| | 5173/1000000 [21:28<68:27:07, 4.04it/s, grad_norm=0.264, loss_final=0.494, loss_mean=0.829, loss_mean_cls=0.0765, proj_loss=-0.412][2026-03-26 12:52:32] Step: 5173, Training Logs: loss_final: 0.502672, loss_mean: 0.853888, proj_loss: -0.423655, loss_mean_cls: 0.072439, grad_norm: 0.406569 +Steps: 1%| | 5174/1000000 [21:28<68:24:10, 4.04it/s, grad_norm=0.407, loss_final=0.503, loss_mean=0.854, loss_mean_cls=0.0724, proj_loss=-0.424][2026-03-26 12:52:32] Step: 5174, Training Logs: loss_final: 0.490185, loss_mean: 0.825872, proj_loss: -0.412310, loss_mean_cls: 0.076623, grad_norm: 0.382960 +Steps: 1%| | 5175/1000000 [21:29<68:23:55, 4.04it/s, grad_norm=0.383, loss_final=0.49, loss_mean=0.826, loss_mean_cls=0.0766, proj_loss=-0.412][2026-03-26 12:52:32] Step: 5175, Training Logs: loss_final: 0.492161, loss_mean: 0.835059, proj_loss: -0.418728, loss_mean_cls: 0.075830, grad_norm: 0.249081 +Steps: 1%| | 5176/1000000 [21:29<68:23:59, 4.04it/s, grad_norm=0.249, loss_final=0.492, loss_mean=0.835, loss_mean_cls=0.0758, proj_loss=-0.419][2026-03-26 12:52:32] Step: 5176, Training Logs: loss_final: 0.485784, loss_mean: 0.814340, proj_loss: -0.406797, loss_mean_cls: 0.078241, grad_norm: 0.430924 +Steps: 1%| | 5177/1000000 [21:29<68:24:58, 4.04it/s, grad_norm=0.431, loss_final=0.486, loss_mean=0.814, loss_mean_cls=0.0782, proj_loss=-0.407][2026-03-26 12:52:33] Step: 5177, Training Logs: loss_final: 0.472881, loss_mean: 0.811683, proj_loss: -0.416501, loss_mean_cls: 0.077699, grad_norm: 0.412872 +Steps: 1%| | 5178/1000000 [21:29<68:26:05, 4.04it/s, grad_norm=0.413, loss_final=0.473, loss_mean=0.812, loss_mean_cls=0.0777, proj_loss=-0.417][2026-03-26 12:52:33] Step: 5178, Training Logs: loss_final: 0.512764, loss_mean: 0.843622, proj_loss: -0.408090, loss_mean_cls: 0.077232, grad_norm: 0.286546 +Steps: 1%| | 5179/1000000 [21:30<68:24:51, 4.04it/s, grad_norm=0.287, loss_final=0.513, loss_mean=0.844, loss_mean_cls=0.0772, proj_loss=-0.408][2026-03-26 12:52:33] Step: 5179, Training Logs: loss_final: 0.501832, loss_mean: 0.830490, proj_loss: -0.407985, loss_mean_cls: 0.079327, grad_norm: 0.503552 +Steps: 1%| | 5180/1000000 [21:30<68:24:10, 4.04it/s, grad_norm=0.504, loss_final=0.502, loss_mean=0.83, loss_mean_cls=0.0793, proj_loss=-0.408][2026-03-26 12:52:33] Step: 5180, Training Logs: loss_final: 0.479500, loss_mean: 0.823503, proj_loss: -0.420089, loss_mean_cls: 0.076086, grad_norm: 0.403813 +Steps: 1%| | 5181/1000000 [21:30<68:24:39, 4.04it/s, grad_norm=0.404, loss_final=0.48, loss_mean=0.824, loss_mean_cls=0.0761, proj_loss=-0.42][2026-03-26 12:52:34] Step: 5181, Training Logs: loss_final: 0.487836, loss_mean: 0.834655, proj_loss: -0.421731, loss_mean_cls: 0.074912, grad_norm: 0.389455 +Steps: 1%| | 5182/1000000 [21:30<68:23:44, 4.04it/s, grad_norm=0.389, loss_final=0.488, loss_mean=0.835, loss_mean_cls=0.0749, proj_loss=-0.422][2026-03-26 12:52:34] Step: 5182, Training Logs: loss_final: 0.488115, loss_mean: 0.817474, proj_loss: -0.408086, loss_mean_cls: 0.078728, grad_norm: 0.499583 +Steps: 1%| | 5183/1000000 [21:31<68:24:35, 4.04it/s, grad_norm=0.5, loss_final=0.488, loss_mean=0.817, loss_mean_cls=0.0787, proj_loss=-0.408][2026-03-26 12:52:34] Step: 5183, Training Logs: loss_final: 0.483066, loss_mean: 0.809741, proj_loss: -0.404804, loss_mean_cls: 0.078129, grad_norm: 0.327854 +Steps: 1%| | 5184/1000000 [21:31<68:23:55, 4.04it/s, grad_norm=0.328, loss_final=0.483, loss_mean=0.81, loss_mean_cls=0.0781, proj_loss=-0.405][2026-03-26 12:52:34] Step: 5184, Training Logs: loss_final: 0.484871, loss_mean: 0.812074, proj_loss: -0.406745, loss_mean_cls: 0.079543, grad_norm: 0.542899 +Steps: 1%| | 5185/1000000 [21:31<68:25:02, 4.04it/s, grad_norm=0.543, loss_final=0.485, loss_mean=0.812, loss_mean_cls=0.0795, proj_loss=-0.407][2026-03-26 12:52:35] Step: 5185, Training Logs: loss_final: 0.473642, loss_mean: 0.816438, proj_loss: -0.418522, loss_mean_cls: 0.075725, grad_norm: 0.263178 +Steps: 1%| | 5186/1000000 [21:31<68:26:48, 4.04it/s, grad_norm=0.263, loss_final=0.474, loss_mean=0.816, loss_mean_cls=0.0757, proj_loss=-0.419][2026-03-26 12:52:35] Step: 5186, Training Logs: loss_final: 0.479994, loss_mean: 0.814062, proj_loss: -0.410537, loss_mean_cls: 0.076469, grad_norm: 0.448684 +Steps: 1%| | 5187/1000000 [21:32<68:25:49, 4.04it/s, grad_norm=0.449, loss_final=0.48, loss_mean=0.814, loss_mean_cls=0.0765, proj_loss=-0.411][2026-03-26 12:52:35] Step: 5187, Training Logs: loss_final: 0.480420, loss_mean: 0.829888, proj_loss: -0.423115, loss_mean_cls: 0.073647, grad_norm: 0.315819 +Steps: 1%| | 5188/1000000 [21:32<68:25:21, 4.04it/s, grad_norm=0.316, loss_final=0.48, loss_mean=0.83, loss_mean_cls=0.0736, proj_loss=-0.423][2026-03-26 12:52:35] Step: 5188, Training Logs: loss_final: 0.492506, loss_mean: 0.838886, proj_loss: -0.421151, loss_mean_cls: 0.074771, grad_norm: 0.383219 +Steps: 1%| | 5189/1000000 [21:32<68:25:22, 4.04it/s, grad_norm=0.383, loss_final=0.493, loss_mean=0.839, loss_mean_cls=0.0748, proj_loss=-0.421][2026-03-26 12:52:36] Step: 5189, Training Logs: loss_final: 0.489969, loss_mean: 0.831689, proj_loss: -0.417092, loss_mean_cls: 0.075372, grad_norm: 0.309835 +Steps: 1%| | 5190/1000000 [21:32<68:26:38, 4.04it/s, grad_norm=0.31, loss_final=0.49, loss_mean=0.832, loss_mean_cls=0.0754, proj_loss=-0.417][2026-03-26 12:52:36] Step: 5190, Training Logs: loss_final: 0.501747, loss_mean: 0.837414, proj_loss: -0.412293, loss_mean_cls: 0.076626, grad_norm: 0.466408 +Steps: 1%| | 5191/1000000 [21:33<68:25:59, 4.04it/s, grad_norm=0.466, loss_final=0.502, loss_mean=0.837, loss_mean_cls=0.0766, proj_loss=-0.412][2026-03-26 12:52:36] Step: 5191, Training Logs: loss_final: 0.492077, loss_mean: 0.831774, proj_loss: -0.415514, loss_mean_cls: 0.075817, grad_norm: 0.285176 +Steps: 1%| | 5192/1000000 [21:33<68:28:07, 4.04it/s, grad_norm=0.285, loss_final=0.492, loss_mean=0.832, loss_mean_cls=0.0758, proj_loss=-0.416][2026-03-26 12:52:36] Step: 5192, Training Logs: loss_final: 0.499829, loss_mean: 0.838203, proj_loss: -0.415092, loss_mean_cls: 0.076717, grad_norm: 0.272275 +Steps: 1%| | 5193/1000000 [21:33<68:25:59, 4.04it/s, grad_norm=0.272, loss_final=0.5, loss_mean=0.838, loss_mean_cls=0.0767, proj_loss=-0.415][2026-03-26 12:52:37] Step: 5193, Training Logs: loss_final: 0.485763, loss_mean: 0.812870, proj_loss: -0.405566, loss_mean_cls: 0.078458, grad_norm: 0.241681 +Steps: 1%| | 5194/1000000 [21:33<68:26:15, 4.04it/s, grad_norm=0.242, loss_final=0.486, loss_mean=0.813, loss_mean_cls=0.0785, proj_loss=-0.406][2026-03-26 12:52:37] Step: 5194, Training Logs: loss_final: 0.493732, loss_mean: 0.828509, proj_loss: -0.411688, loss_mean_cls: 0.076911, grad_norm: 0.405860 +Steps: 1%| | 5195/1000000 [21:34<68:22:26, 4.04it/s, grad_norm=0.406, loss_final=0.494, loss_mean=0.829, loss_mean_cls=0.0769, proj_loss=-0.412][2026-03-26 12:52:37] Step: 5195, Training Logs: loss_final: 0.482261, loss_mean: 0.810338, proj_loss: -0.407353, loss_mean_cls: 0.079276, grad_norm: 0.349530 +Steps: 1%| | 5196/1000000 [21:34<68:23:12, 4.04it/s, grad_norm=0.35, loss_final=0.482, loss_mean=0.81, loss_mean_cls=0.0793, proj_loss=-0.407][2026-03-26 12:52:37] Step: 5196, Training Logs: loss_final: 0.492514, loss_mean: 0.829412, proj_loss: -0.413230, loss_mean_cls: 0.076332, grad_norm: 0.245597 +Steps: 1%| | 5197/1000000 [21:34<69:04:15, 4.00it/s, grad_norm=0.246, loss_final=0.493, loss_mean=0.829, loss_mean_cls=0.0763, proj_loss=-0.413][2026-03-26 12:52:38] Step: 5197, Training Logs: loss_final: 0.496090, loss_mean: 0.836119, proj_loss: -0.416039, loss_mean_cls: 0.076009, grad_norm: 0.337521 +Steps: 1%| | 5198/1000000 [21:34<68:52:36, 4.01it/s, grad_norm=0.338, loss_final=0.496, loss_mean=0.836, loss_mean_cls=0.076, proj_loss=-0.416][2026-03-26 12:52:38] Step: 5198, Training Logs: loss_final: 0.491959, loss_mean: 0.833112, proj_loss: -0.417008, loss_mean_cls: 0.075855, grad_norm: 0.367649 +Steps: 1%| | 5199/1000000 [21:35<68:43:54, 4.02it/s, grad_norm=0.368, loss_final=0.492, loss_mean=0.833, loss_mean_cls=0.0759, proj_loss=-0.417][2026-03-26 12:52:38] Step: 5199, Training Logs: loss_final: 0.503644, loss_mean: 0.831672, proj_loss: -0.407040, loss_mean_cls: 0.079012, grad_norm: 0.315648 +Steps: 1%| | 5200/1000000 [21:35<68:37:41, 4.03it/s, grad_norm=0.316, loss_final=0.504, loss_mean=0.832, loss_mean_cls=0.079, proj_loss=-0.407][2026-03-26 12:52:38] Step: 5200, Training Logs: loss_final: 0.485078, loss_mean: 0.831489, proj_loss: -0.419826, loss_mean_cls: 0.073416, grad_norm: 0.429179 +Steps: 1%| | 5201/1000000 [21:35<68:32:28, 4.03it/s, grad_norm=0.429, loss_final=0.485, loss_mean=0.831, loss_mean_cls=0.0734, proj_loss=-0.42][2026-03-26 12:52:39] Step: 5201, Training Logs: loss_final: 0.487707, loss_mean: 0.825834, proj_loss: -0.413716, loss_mean_cls: 0.075590, grad_norm: 0.315640 +Steps: 1%| | 5202/1000000 [21:35<68:30:55, 4.03it/s, grad_norm=0.316, loss_final=0.488, loss_mean=0.826, loss_mean_cls=0.0756, proj_loss=-0.414][2026-03-26 12:52:39] Step: 5202, Training Logs: loss_final: 0.491759, loss_mean: 0.837073, proj_loss: -0.419634, loss_mean_cls: 0.074319, grad_norm: 0.243691 +Steps: 1%| | 5203/1000000 [21:36<68:29:46, 4.03it/s, grad_norm=0.244, loss_final=0.492, loss_mean=0.837, loss_mean_cls=0.0743, proj_loss=-0.42][2026-03-26 12:52:39] Step: 5203, Training Logs: loss_final: 0.498364, loss_mean: 0.841389, proj_loss: -0.416979, loss_mean_cls: 0.073954, grad_norm: 0.326736 +Steps: 1%| | 5204/1000000 [21:36<68:33:12, 4.03it/s, grad_norm=0.327, loss_final=0.498, loss_mean=0.841, loss_mean_cls=0.074, proj_loss=-0.417][2026-03-26 12:52:39] Step: 5204, Training Logs: loss_final: 0.481018, loss_mean: 0.815668, proj_loss: -0.412865, loss_mean_cls: 0.078215, grad_norm: 0.356962 +Steps: 1%| | 5205/1000000 [21:36<68:29:47, 4.03it/s, grad_norm=0.357, loss_final=0.481, loss_mean=0.816, loss_mean_cls=0.0782, proj_loss=-0.413][2026-03-26 12:52:40] Step: 5205, Training Logs: loss_final: 0.483154, loss_mean: 0.819355, proj_loss: -0.413055, loss_mean_cls: 0.076853, grad_norm: 0.259255 +Steps: 1%| | 5206/1000000 [21:36<68:27:32, 4.04it/s, grad_norm=0.259, loss_final=0.483, loss_mean=0.819, loss_mean_cls=0.0769, proj_loss=-0.413][2026-03-26 12:52:40] Step: 5206, Training Logs: loss_final: 0.493763, loss_mean: 0.839951, proj_loss: -0.418175, loss_mean_cls: 0.071986, grad_norm: 0.318448 +Steps: 1%| | 5207/1000000 [21:37<68:25:58, 4.04it/s, grad_norm=0.318, loss_final=0.494, loss_mean=0.84, loss_mean_cls=0.072, proj_loss=-0.418][2026-03-26 12:52:40] Step: 5207, Training Logs: loss_final: 0.496256, loss_mean: 0.840783, proj_loss: -0.419155, loss_mean_cls: 0.074629, grad_norm: 0.477852 +Steps: 1%| | 5208/1000000 [21:37<68:26:11, 4.04it/s, grad_norm=0.478, loss_final=0.496, loss_mean=0.841, loss_mean_cls=0.0746, proj_loss=-0.419][2026-03-26 12:52:40] Step: 5208, Training Logs: loss_final: 0.498938, loss_mean: 0.826585, proj_loss: -0.405204, loss_mean_cls: 0.077557, grad_norm: 0.349114 +Steps: 1%| | 5209/1000000 [21:37<68:24:05, 4.04it/s, grad_norm=0.349, loss_final=0.499, loss_mean=0.827, loss_mean_cls=0.0776, proj_loss=-0.405][2026-03-26 12:52:41] Step: 5209, Training Logs: loss_final: 0.492351, loss_mean: 0.817193, proj_loss: -0.402853, loss_mean_cls: 0.078011, grad_norm: 0.465967 +Steps: 1%| | 5210/1000000 [21:37<68:26:37, 4.04it/s, grad_norm=0.466, loss_final=0.492, loss_mean=0.817, loss_mean_cls=0.078, proj_loss=-0.403][2026-03-26 12:52:41] Step: 5210, Training Logs: loss_final: 0.487714, loss_mean: 0.829280, proj_loss: -0.417951, loss_mean_cls: 0.076385, grad_norm: 0.329145 +Steps: 1%| | 5211/1000000 [21:38<68:33:38, 4.03it/s, grad_norm=0.329, loss_final=0.488, loss_mean=0.829, loss_mean_cls=0.0764, proj_loss=-0.418][2026-03-26 12:52:41] Step: 5211, Training Logs: loss_final: 0.502151, loss_mean: 0.832483, proj_loss: -0.407158, loss_mean_cls: 0.076827, grad_norm: 0.400056 +Steps: 1%| | 5212/1000000 [21:38<68:30:55, 4.03it/s, grad_norm=0.4, loss_final=0.502, loss_mean=0.832, loss_mean_cls=0.0768, proj_loss=-0.407][2026-03-26 12:52:41] Step: 5212, Training Logs: loss_final: 0.493754, loss_mean: 0.838427, proj_loss: -0.419695, loss_mean_cls: 0.075022, grad_norm: 0.398510 +Steps: 1%| | 5213/1000000 [21:38<68:32:34, 4.03it/s, grad_norm=0.399, loss_final=0.494, loss_mean=0.838, loss_mean_cls=0.075, proj_loss=-0.42][2026-03-26 12:52:42] Step: 5213, Training Logs: loss_final: 0.485630, loss_mean: 0.811198, proj_loss: -0.404476, loss_mean_cls: 0.078907, grad_norm: 0.516344 +Steps: 1%| | 5214/1000000 [21:38<68:41:27, 4.02it/s, grad_norm=0.516, loss_final=0.486, loss_mean=0.811, loss_mean_cls=0.0789, proj_loss=-0.404][2026-03-26 12:52:42] Step: 5214, Training Logs: loss_final: 0.498382, loss_mean: 0.838833, proj_loss: -0.415481, loss_mean_cls: 0.075030, grad_norm: 0.382227 +Steps: 1%| | 5215/1000000 [21:39<68:36:06, 4.03it/s, grad_norm=0.382, loss_final=0.498, loss_mean=0.839, loss_mean_cls=0.075, proj_loss=-0.415][2026-03-26 12:52:42] Step: 5215, Training Logs: loss_final: 0.485385, loss_mean: 0.823371, proj_loss: -0.413502, loss_mean_cls: 0.075516, grad_norm: 0.396604 +Steps: 1%| | 5216/1000000 [21:39<68:33:02, 4.03it/s, grad_norm=0.397, loss_final=0.485, loss_mean=0.823, loss_mean_cls=0.0755, proj_loss=-0.414][2026-03-26 12:52:42] Step: 5216, Training Logs: loss_final: 0.479409, loss_mean: 0.809650, proj_loss: -0.407582, loss_mean_cls: 0.077341, grad_norm: 0.535244 +Steps: 1%| | 5217/1000000 [21:39<68:31:42, 4.03it/s, grad_norm=0.535, loss_final=0.479, loss_mean=0.81, loss_mean_cls=0.0773, proj_loss=-0.408][2026-03-26 12:52:43] Step: 5217, Training Logs: loss_final: 0.492389, loss_mean: 0.818565, proj_loss: -0.405049, loss_mean_cls: 0.078873, grad_norm: 0.522623 +Steps: 1%| | 5218/1000000 [21:39<68:29:43, 4.03it/s, grad_norm=0.523, loss_final=0.492, loss_mean=0.819, loss_mean_cls=0.0789, proj_loss=-0.405][2026-03-26 12:52:43] Step: 5218, Training Logs: loss_final: 0.503902, loss_mean: 0.834759, proj_loss: -0.408558, loss_mean_cls: 0.077701, grad_norm: 0.445473 +Steps: 1%| | 5219/1000000 [21:40<68:30:28, 4.03it/s, grad_norm=0.445, loss_final=0.504, loss_mean=0.835, loss_mean_cls=0.0777, proj_loss=-0.409][2026-03-26 12:52:43] Step: 5219, Training Logs: loss_final: 0.493043, loss_mean: 0.830810, proj_loss: -0.413924, loss_mean_cls: 0.076157, grad_norm: 0.464277 +Steps: 1%| | 5220/1000000 [21:40<68:26:15, 4.04it/s, grad_norm=0.464, loss_final=0.493, loss_mean=0.831, loss_mean_cls=0.0762, proj_loss=-0.414][2026-03-26 12:52:43] Step: 5220, Training Logs: loss_final: 0.497881, loss_mean: 0.833397, proj_loss: -0.411936, loss_mean_cls: 0.076420, grad_norm: 0.630757 +Steps: 1%| | 5221/1000000 [21:40<68:22:31, 4.04it/s, grad_norm=0.631, loss_final=0.498, loss_mean=0.833, loss_mean_cls=0.0764, proj_loss=-0.412][2026-03-26 12:52:44] Step: 5221, Training Logs: loss_final: 0.480470, loss_mean: 0.823994, proj_loss: -0.417766, loss_mean_cls: 0.074242, grad_norm: 0.350268 +Steps: 1%| | 5222/1000000 [21:40<68:20:50, 4.04it/s, grad_norm=0.35, loss_final=0.48, loss_mean=0.824, loss_mean_cls=0.0742, proj_loss=-0.418][2026-03-26 12:52:44] Step: 5222, Training Logs: loss_final: 0.500762, loss_mean: 0.847585, proj_loss: -0.421252, loss_mean_cls: 0.074428, grad_norm: 0.441098 +Steps: 1%| | 5223/1000000 [21:41<68:21:52, 4.04it/s, grad_norm=0.441, loss_final=0.501, loss_mean=0.848, loss_mean_cls=0.0744, proj_loss=-0.421][2026-03-26 12:52:44] Step: 5223, Training Logs: loss_final: 0.496375, loss_mean: 0.817648, proj_loss: -0.401403, loss_mean_cls: 0.080130, grad_norm: 0.325739 +Steps: 1%| | 5224/1000000 [21:41<68:23:08, 4.04it/s, grad_norm=0.326, loss_final=0.496, loss_mean=0.818, loss_mean_cls=0.0801, proj_loss=-0.401][2026-03-26 12:52:44] Step: 5224, Training Logs: loss_final: 0.488402, loss_mean: 0.819505, proj_loss: -0.407944, loss_mean_cls: 0.076841, grad_norm: 0.384894 +Steps: 1%| | 5225/1000000 [21:41<68:23:57, 4.04it/s, grad_norm=0.385, loss_final=0.488, loss_mean=0.82, loss_mean_cls=0.0768, proj_loss=-0.408][2026-03-26 12:52:45] Step: 5225, Training Logs: loss_final: 0.481671, loss_mean: 0.817905, proj_loss: -0.414187, loss_mean_cls: 0.077953, grad_norm: 0.613461 +Steps: 1%| | 5226/1000000 [21:41<68:23:04, 4.04it/s, grad_norm=0.613, loss_final=0.482, loss_mean=0.818, loss_mean_cls=0.078, proj_loss=-0.414][2026-03-26 12:52:45] Step: 5226, Training Logs: loss_final: 0.484783, loss_mean: 0.825405, proj_loss: -0.417268, loss_mean_cls: 0.076646, grad_norm: 0.294069 +Steps: 1%| | 5227/1000000 [21:42<68:23:36, 4.04it/s, grad_norm=0.294, loss_final=0.485, loss_mean=0.825, loss_mean_cls=0.0766, proj_loss=-0.417][2026-03-26 12:52:45] Step: 5227, Training Logs: loss_final: 0.498524, loss_mean: 0.834772, proj_loss: -0.412329, loss_mean_cls: 0.076081, grad_norm: 0.321696 +Steps: 1%| | 5228/1000000 [21:42<68:23:27, 4.04it/s, grad_norm=0.322, loss_final=0.499, loss_mean=0.835, loss_mean_cls=0.0761, proj_loss=-0.412][2026-03-26 12:52:45] Step: 5228, Training Logs: loss_final: 0.487272, loss_mean: 0.827587, proj_loss: -0.414933, loss_mean_cls: 0.074619, grad_norm: 0.321923 +Steps: 1%| | 5229/1000000 [21:42<68:25:17, 4.04it/s, grad_norm=0.322, loss_final=0.487, loss_mean=0.828, loss_mean_cls=0.0746, proj_loss=-0.415][2026-03-26 12:52:46] Step: 5229, Training Logs: loss_final: 0.496550, loss_mean: 0.856627, proj_loss: -0.430951, loss_mean_cls: 0.070874, grad_norm: 0.358227 +Steps: 1%| | 5230/1000000 [21:42<68:26:15, 4.04it/s, grad_norm=0.358, loss_final=0.497, loss_mean=0.857, loss_mean_cls=0.0709, proj_loss=-0.431][2026-03-26 12:52:46] Step: 5230, Training Logs: loss_final: 0.506963, loss_mean: 0.846352, proj_loss: -0.414058, loss_mean_cls: 0.074669, grad_norm: 0.246680 +Steps: 1%| | 5231/1000000 [21:43<68:24:06, 4.04it/s, grad_norm=0.247, loss_final=0.507, loss_mean=0.846, loss_mean_cls=0.0747, proj_loss=-0.414][2026-03-26 12:52:46] Step: 5231, Training Logs: loss_final: 0.487038, loss_mean: 0.829700, proj_loss: -0.416931, loss_mean_cls: 0.074270, grad_norm: 0.331828 +Steps: 1%| | 5232/1000000 [21:43<68:26:04, 4.04it/s, grad_norm=0.332, loss_final=0.487, loss_mean=0.83, loss_mean_cls=0.0743, proj_loss=-0.417][2026-03-26 12:52:46] Step: 5232, Training Logs: loss_final: 0.486554, loss_mean: 0.828246, proj_loss: -0.416563, loss_mean_cls: 0.074870, grad_norm: 0.283572 +Steps: 1%| | 5233/1000000 [21:43<68:25:53, 4.04it/s, grad_norm=0.284, loss_final=0.487, loss_mean=0.828, loss_mean_cls=0.0749, proj_loss=-0.417][2026-03-26 12:52:46] Step: 5233, Training Logs: loss_final: 0.492034, loss_mean: 0.831270, proj_loss: -0.414811, loss_mean_cls: 0.075575, grad_norm: 0.318812 +Steps: 1%| | 5234/1000000 [21:43<68:27:22, 4.04it/s, grad_norm=0.319, loss_final=0.492, loss_mean=0.831, loss_mean_cls=0.0756, proj_loss=-0.415][2026-03-26 12:52:47] Step: 5234, Training Logs: loss_final: 0.487060, loss_mean: 0.825487, proj_loss: -0.414312, loss_mean_cls: 0.075885, grad_norm: 0.267457 +Steps: 1%| | 5235/1000000 [21:44<68:26:08, 4.04it/s, grad_norm=0.267, loss_final=0.487, loss_mean=0.825, loss_mean_cls=0.0759, proj_loss=-0.414][2026-03-26 12:52:47] Step: 5235, Training Logs: loss_final: 0.511805, loss_mean: 0.846211, proj_loss: -0.410724, loss_mean_cls: 0.076318, grad_norm: 0.391820 +Steps: 1%| | 5236/1000000 [21:44<68:25:38, 4.04it/s, grad_norm=0.392, loss_final=0.512, loss_mean=0.846, loss_mean_cls=0.0763, proj_loss=-0.411][2026-03-26 12:52:47] Step: 5236, Training Logs: loss_final: 0.502796, loss_mean: 0.841787, proj_loss: -0.414777, loss_mean_cls: 0.075786, grad_norm: 0.653698 +Steps: 1%| | 5237/1000000 [21:44<68:23:51, 4.04it/s, grad_norm=0.654, loss_final=0.503, loss_mean=0.842, loss_mean_cls=0.0758, proj_loss=-0.415][2026-03-26 12:52:47] Step: 5237, Training Logs: loss_final: 0.481955, loss_mean: 0.816850, proj_loss: -0.410968, loss_mean_cls: 0.076073, grad_norm: 0.269345 +Steps: 1%| | 5238/1000000 [21:44<68:26:37, 4.04it/s, grad_norm=0.269, loss_final=0.482, loss_mean=0.817, loss_mean_cls=0.0761, proj_loss=-0.411][2026-03-26 12:52:48] Step: 5238, Training Logs: loss_final: 0.483716, loss_mean: 0.828157, proj_loss: -0.419182, loss_mean_cls: 0.074741, grad_norm: 0.293140 +Steps: 1%| | 5239/1000000 [21:45<68:26:18, 4.04it/s, grad_norm=0.293, loss_final=0.484, loss_mean=0.828, loss_mean_cls=0.0747, proj_loss=-0.419][2026-03-26 12:52:48] Step: 5239, Training Logs: loss_final: 0.490745, loss_mean: 0.827711, proj_loss: -0.412454, loss_mean_cls: 0.075489, grad_norm: 0.274922 +Steps: 1%| | 5240/1000000 [21:45<68:27:46, 4.04it/s, grad_norm=0.275, loss_final=0.491, loss_mean=0.828, loss_mean_cls=0.0755, proj_loss=-0.412][2026-03-26 12:52:48] Step: 5240, Training Logs: loss_final: 0.488482, loss_mean: 0.825377, proj_loss: -0.412215, loss_mean_cls: 0.075320, grad_norm: 0.275821 +Steps: 1%| | 5241/1000000 [21:45<68:23:57, 4.04it/s, grad_norm=0.276, loss_final=0.488, loss_mean=0.825, loss_mean_cls=0.0753, proj_loss=-0.412][2026-03-26 12:52:48] Step: 5241, Training Logs: loss_final: 0.491450, loss_mean: 0.824375, proj_loss: -0.410112, loss_mean_cls: 0.077187, grad_norm: 0.368601 +Steps: 1%| | 5242/1000000 [21:45<68:25:15, 4.04it/s, grad_norm=0.369, loss_final=0.491, loss_mean=0.824, loss_mean_cls=0.0772, proj_loss=-0.41][2026-03-26 12:52:49] Step: 5242, Training Logs: loss_final: 0.496250, loss_mean: 0.832257, proj_loss: -0.412312, loss_mean_cls: 0.076305, grad_norm: 0.287198 +Steps: 1%| | 5243/1000000 [21:46<68:22:50, 4.04it/s, grad_norm=0.287, loss_final=0.496, loss_mean=0.832, loss_mean_cls=0.0763, proj_loss=-0.412][2026-03-26 12:52:49] Step: 5243, Training Logs: loss_final: 0.486841, loss_mean: 0.829514, proj_loss: -0.416973, loss_mean_cls: 0.074300, grad_norm: 0.488562 +Steps: 1%| | 5244/1000000 [21:46<68:25:08, 4.04it/s, grad_norm=0.489, loss_final=0.487, loss_mean=0.83, loss_mean_cls=0.0743, proj_loss=-0.417][2026-03-26 12:52:49] Step: 5244, Training Logs: loss_final: 0.499910, loss_mean: 0.850331, proj_loss: -0.422514, loss_mean_cls: 0.072093, grad_norm: 0.568548 +Steps: 1%| | 5245/1000000 [21:46<68:23:05, 4.04it/s, grad_norm=0.569, loss_final=0.5, loss_mean=0.85, loss_mean_cls=0.0721, proj_loss=-0.423][2026-03-26 12:52:49] Step: 5245, Training Logs: loss_final: 0.494661, loss_mean: 0.827950, proj_loss: -0.409074, loss_mean_cls: 0.075784, grad_norm: 0.452322 +Steps: 1%| | 5246/1000000 [21:46<68:28:06, 4.04it/s, grad_norm=0.452, loss_final=0.495, loss_mean=0.828, loss_mean_cls=0.0758, proj_loss=-0.409][2026-03-26 12:52:50] Step: 5246, Training Logs: loss_final: 0.493229, loss_mean: 0.822985, proj_loss: -0.406160, loss_mean_cls: 0.076404, grad_norm: 0.416027 +Steps: 1%| | 5247/1000000 [21:47<68:33:14, 4.03it/s, grad_norm=0.416, loss_final=0.493, loss_mean=0.823, loss_mean_cls=0.0764, proj_loss=-0.406][2026-03-26 12:52:50] Step: 5247, Training Logs: loss_final: 0.476912, loss_mean: 0.810615, proj_loss: -0.410258, loss_mean_cls: 0.076555, grad_norm: 0.537341 +Steps: 1%| | 5248/1000000 [21:47<69:40:59, 3.97it/s, grad_norm=0.537, loss_final=0.477, loss_mean=0.811, loss_mean_cls=0.0766, proj_loss=-0.41][2026-03-26 12:52:50] Step: 5248, Training Logs: loss_final: 0.473425, loss_mean: 0.816366, proj_loss: -0.417900, loss_mean_cls: 0.074959, grad_norm: 0.384350 +Steps: 1%| | 5249/1000000 [21:47<69:18:56, 3.99it/s, grad_norm=0.384, loss_final=0.473, loss_mean=0.816, loss_mean_cls=0.075, proj_loss=-0.418][2026-03-26 12:52:50] Step: 5249, Training Logs: loss_final: 0.477275, loss_mean: 0.826051, proj_loss: -0.423176, loss_mean_cls: 0.074400, grad_norm: 0.419080 +Steps: 1%| | 5250/1000000 [21:47<69:02:18, 4.00it/s, grad_norm=0.419, loss_final=0.477, loss_mean=0.826, loss_mean_cls=0.0744, proj_loss=-0.423][2026-03-26 12:52:51] Step: 5250, Training Logs: loss_final: 0.486696, loss_mean: 0.817004, proj_loss: -0.406892, loss_mean_cls: 0.076584, grad_norm: 0.347964 +Steps: 1%| | 5251/1000000 [21:48<68:49:51, 4.01it/s, grad_norm=0.348, loss_final=0.487, loss_mean=0.817, loss_mean_cls=0.0766, proj_loss=-0.407][2026-03-26 12:52:51] Step: 5251, Training Logs: loss_final: 0.483657, loss_mean: 0.816911, proj_loss: -0.409950, loss_mean_cls: 0.076696, grad_norm: 0.390592 +Steps: 1%| | 5252/1000000 [21:48<68:45:26, 4.02it/s, grad_norm=0.391, loss_final=0.484, loss_mean=0.817, loss_mean_cls=0.0767, proj_loss=-0.41][2026-03-26 12:52:51] Step: 5252, Training Logs: loss_final: 0.494389, loss_mean: 0.839689, proj_loss: -0.419546, loss_mean_cls: 0.074246, grad_norm: 0.266731 +Steps: 1%| | 5253/1000000 [21:48<68:42:14, 4.02it/s, grad_norm=0.267, loss_final=0.494, loss_mean=0.84, loss_mean_cls=0.0742, proj_loss=-0.42][2026-03-26 12:52:51] Step: 5253, Training Logs: loss_final: 0.493504, loss_mean: 0.841742, proj_loss: -0.422860, loss_mean_cls: 0.074621, grad_norm: 0.311018 +Steps: 1%| | 5254/1000000 [21:48<68:37:53, 4.03it/s, grad_norm=0.311, loss_final=0.494, loss_mean=0.842, loss_mean_cls=0.0746, proj_loss=-0.423][2026-03-26 12:52:52] Step: 5254, Training Logs: loss_final: 0.486999, loss_mean: 0.828228, proj_loss: -0.415951, loss_mean_cls: 0.074722, grad_norm: 0.300318 +Steps: 1%| | 5255/1000000 [21:49<68:32:33, 4.03it/s, grad_norm=0.3, loss_final=0.487, loss_mean=0.828, loss_mean_cls=0.0747, proj_loss=-0.416][2026-03-26 12:52:52] Step: 5255, Training Logs: loss_final: 0.511633, loss_mean: 0.853909, proj_loss: -0.415672, loss_mean_cls: 0.073395, grad_norm: 0.323561 +Steps: 1%| | 5256/1000000 [21:49<68:28:54, 4.03it/s, grad_norm=0.324, loss_final=0.512, loss_mean=0.854, loss_mean_cls=0.0734, proj_loss=-0.416][2026-03-26 12:52:52] Step: 5256, Training Logs: loss_final: 0.478976, loss_mean: 0.816576, proj_loss: -0.412863, loss_mean_cls: 0.075263, grad_norm: 0.451699 +Steps: 1%| | 5257/1000000 [21:49<68:29:33, 4.03it/s, grad_norm=0.452, loss_final=0.479, loss_mean=0.817, loss_mean_cls=0.0753, proj_loss=-0.413][2026-03-26 12:52:52] Step: 5257, Training Logs: loss_final: 0.514625, loss_mean: 0.852603, proj_loss: -0.412387, loss_mean_cls: 0.074409, grad_norm: 0.348026 +Steps: 1%| | 5258/1000000 [21:49<68:28:12, 4.04it/s, grad_norm=0.348, loss_final=0.515, loss_mean=0.853, loss_mean_cls=0.0744, proj_loss=-0.412][2026-03-26 12:52:53] Step: 5258, Training Logs: loss_final: 0.502604, loss_mean: 0.843997, proj_loss: -0.415025, loss_mean_cls: 0.073632, grad_norm: 0.310607 +Steps: 1%| | 5259/1000000 [21:50<68:27:59, 4.04it/s, grad_norm=0.311, loss_final=0.503, loss_mean=0.844, loss_mean_cls=0.0736, proj_loss=-0.415][2026-03-26 12:52:53] Step: 5259, Training Logs: loss_final: 0.479294, loss_mean: 0.809126, proj_loss: -0.406694, loss_mean_cls: 0.076862, grad_norm: 0.477278 +Steps: 1%| | 5260/1000000 [21:50<68:27:52, 4.04it/s, grad_norm=0.477, loss_final=0.479, loss_mean=0.809, loss_mean_cls=0.0769, proj_loss=-0.407][2026-03-26 12:52:53] Step: 5260, Training Logs: loss_final: 0.496125, loss_mean: 0.830417, proj_loss: -0.411037, loss_mean_cls: 0.076745, grad_norm: 0.429867 +Steps: 1%| | 5261/1000000 [21:50<68:27:07, 4.04it/s, grad_norm=0.43, loss_final=0.496, loss_mean=0.83, loss_mean_cls=0.0767, proj_loss=-0.411][2026-03-26 12:52:53] Step: 5261, Training Logs: loss_final: 0.498723, loss_mean: 0.839472, proj_loss: -0.415897, loss_mean_cls: 0.075148, grad_norm: 0.481481 +Steps: 1%| | 5262/1000000 [21:50<68:26:00, 4.04it/s, grad_norm=0.481, loss_final=0.499, loss_mean=0.839, loss_mean_cls=0.0751, proj_loss=-0.416][2026-03-26 12:52:54] Step: 5262, Training Logs: loss_final: 0.496458, loss_mean: 0.844205, proj_loss: -0.421260, loss_mean_cls: 0.073513, grad_norm: 0.292284 +Steps: 1%| | 5263/1000000 [21:51<68:26:30, 4.04it/s, grad_norm=0.292, loss_final=0.496, loss_mean=0.844, loss_mean_cls=0.0735, proj_loss=-0.421][2026-03-26 12:52:54] Step: 5263, Training Logs: loss_final: 0.501989, loss_mean: 0.845362, proj_loss: -0.418046, loss_mean_cls: 0.074674, grad_norm: 0.382425 +Steps: 1%| | 5264/1000000 [21:51<68:28:38, 4.04it/s, grad_norm=0.382, loss_final=0.502, loss_mean=0.845, loss_mean_cls=0.0747, proj_loss=-0.418][2026-03-26 12:52:54] Step: 5264, Training Logs: loss_final: 0.486656, loss_mean: 0.816895, proj_loss: -0.407219, loss_mean_cls: 0.076980, grad_norm: 0.332204 +Steps: 1%| | 5265/1000000 [21:51<68:26:46, 4.04it/s, grad_norm=0.332, loss_final=0.487, loss_mean=0.817, loss_mean_cls=0.077, proj_loss=-0.407][2026-03-26 12:52:54] Step: 5265, Training Logs: loss_final: 0.496445, loss_mean: 0.835573, proj_loss: -0.414224, loss_mean_cls: 0.075096, grad_norm: 0.363344 +Steps: 1%| | 5266/1000000 [21:51<68:23:39, 4.04it/s, grad_norm=0.363, loss_final=0.496, loss_mean=0.836, loss_mean_cls=0.0751, proj_loss=-0.414][2026-03-26 12:52:55] Step: 5266, Training Logs: loss_final: 0.484983, loss_mean: 0.818581, proj_loss: -0.409975, loss_mean_cls: 0.076378, grad_norm: 0.322166 +Steps: 1%| | 5267/1000000 [21:52<68:23:01, 4.04it/s, grad_norm=0.322, loss_final=0.485, loss_mean=0.819, loss_mean_cls=0.0764, proj_loss=-0.41][2026-03-26 12:52:55] Step: 5267, Training Logs: loss_final: 0.485822, loss_mean: 0.820548, proj_loss: -0.410388, loss_mean_cls: 0.075662, grad_norm: 0.291597 +Steps: 1%| | 5268/1000000 [21:52<68:23:24, 4.04it/s, grad_norm=0.292, loss_final=0.486, loss_mean=0.821, loss_mean_cls=0.0757, proj_loss=-0.41][2026-03-26 12:52:55] Step: 5268, Training Logs: loss_final: 0.494645, loss_mean: 0.829789, proj_loss: -0.411982, loss_mean_cls: 0.076838, grad_norm: 0.492962 +Steps: 1%| | 5269/1000000 [21:52<68:24:37, 4.04it/s, grad_norm=0.493, loss_final=0.495, loss_mean=0.83, loss_mean_cls=0.0768, proj_loss=-0.412][2026-03-26 12:52:55] Step: 5269, Training Logs: loss_final: 0.496839, loss_mean: 0.841777, proj_loss: -0.417406, loss_mean_cls: 0.072469, grad_norm: 0.321816 +Steps: 1%| | 5270/1000000 [21:52<68:23:56, 4.04it/s, grad_norm=0.322, loss_final=0.497, loss_mean=0.842, loss_mean_cls=0.0725, proj_loss=-0.417][2026-03-26 12:52:56] Step: 5270, Training Logs: loss_final: 0.493825, loss_mean: 0.833129, proj_loss: -0.414499, loss_mean_cls: 0.075194, grad_norm: 0.379819 +Steps: 1%| | 5271/1000000 [21:53<68:23:28, 4.04it/s, grad_norm=0.38, loss_final=0.494, loss_mean=0.833, loss_mean_cls=0.0752, proj_loss=-0.414][2026-03-26 12:52:56] Step: 5271, Training Logs: loss_final: 0.492944, loss_mean: 0.843450, proj_loss: -0.424650, loss_mean_cls: 0.074145, grad_norm: 0.403675 +Steps: 1%| | 5272/1000000 [21:53<68:22:47, 4.04it/s, grad_norm=0.404, loss_final=0.493, loss_mean=0.843, loss_mean_cls=0.0741, proj_loss=-0.425][2026-03-26 12:52:56] Step: 5272, Training Logs: loss_final: 0.499777, loss_mean: 0.855491, proj_loss: -0.427927, loss_mean_cls: 0.072213, grad_norm: 0.249538 +Steps: 1%| | 5273/1000000 [21:53<68:23:56, 4.04it/s, grad_norm=0.25, loss_final=0.5, loss_mean=0.855, loss_mean_cls=0.0722, proj_loss=-0.428][2026-03-26 12:52:56] Step: 5273, Training Logs: loss_final: 0.495154, loss_mean: 0.833022, proj_loss: -0.413112, loss_mean_cls: 0.075244, grad_norm: 0.426551 +Steps: 1%| | 5274/1000000 [21:53<68:24:14, 4.04it/s, grad_norm=0.427, loss_final=0.495, loss_mean=0.833, loss_mean_cls=0.0752, proj_loss=-0.413][2026-03-26 12:52:57] Step: 5274, Training Logs: loss_final: 0.504637, loss_mean: 0.855736, proj_loss: -0.423221, loss_mean_cls: 0.072121, grad_norm: 0.385444 +Steps: 1%| | 5275/1000000 [21:54<68:26:31, 4.04it/s, grad_norm=0.385, loss_final=0.505, loss_mean=0.856, loss_mean_cls=0.0721, proj_loss=-0.423][2026-03-26 12:52:57] Step: 5275, Training Logs: loss_final: 0.484887, loss_mean: 0.819697, proj_loss: -0.410797, loss_mean_cls: 0.075986, grad_norm: 0.479503 +Steps: 1%| | 5276/1000000 [21:54<68:25:25, 4.04it/s, grad_norm=0.48, loss_final=0.485, loss_mean=0.82, loss_mean_cls=0.076, proj_loss=-0.411][2026-03-26 12:52:57] Step: 5276, Training Logs: loss_final: 0.493776, loss_mean: 0.843164, proj_loss: -0.421687, loss_mean_cls: 0.072300, grad_norm: 0.298878 +Steps: 1%| | 5277/1000000 [21:54<68:25:43, 4.04it/s, grad_norm=0.299, loss_final=0.494, loss_mean=0.843, loss_mean_cls=0.0723, proj_loss=-0.422][2026-03-26 12:52:57] Step: 5277, Training Logs: loss_final: 0.489214, loss_mean: 0.831778, proj_loss: -0.417561, loss_mean_cls: 0.074997, grad_norm: 0.246448 +Steps: 1%| | 5278/1000000 [21:54<68:27:13, 4.04it/s, grad_norm=0.246, loss_final=0.489, loss_mean=0.832, loss_mean_cls=0.075, proj_loss=-0.418][2026-03-26 12:52:58] Step: 5278, Training Logs: loss_final: 0.475227, loss_mean: 0.808394, proj_loss: -0.409864, loss_mean_cls: 0.076698, grad_norm: 0.313587 +Steps: 1%| | 5279/1000000 [21:55<68:27:15, 4.04it/s, grad_norm=0.314, loss_final=0.475, loss_mean=0.808, loss_mean_cls=0.0767, proj_loss=-0.41][2026-03-26 12:52:58] Step: 5279, Training Logs: loss_final: 0.489189, loss_mean: 0.829327, proj_loss: -0.414406, loss_mean_cls: 0.074268, grad_norm: 0.363068 +Steps: 1%| | 5280/1000000 [21:55<68:29:32, 4.03it/s, grad_norm=0.363, loss_final=0.489, loss_mean=0.829, loss_mean_cls=0.0743, proj_loss=-0.414][2026-03-26 12:52:58] Step: 5280, Training Logs: loss_final: 0.506886, loss_mean: 0.842562, proj_loss: -0.410751, loss_mean_cls: 0.075075, grad_norm: 0.364961 +Steps: 1%| | 5281/1000000 [21:55<68:27:10, 4.04it/s, grad_norm=0.365, loss_final=0.507, loss_mean=0.843, loss_mean_cls=0.0751, proj_loss=-0.411][2026-03-26 12:52:58] Step: 5281, Training Logs: loss_final: 0.490437, loss_mean: 0.822859, proj_loss: -0.408586, loss_mean_cls: 0.076164, grad_norm: 0.309217 +Steps: 1%| | 5282/1000000 [21:55<68:26:27, 4.04it/s, grad_norm=0.309, loss_final=0.49, loss_mean=0.823, loss_mean_cls=0.0762, proj_loss=-0.409][2026-03-26 12:52:59] Step: 5282, Training Logs: loss_final: 0.495043, loss_mean: 0.825347, proj_loss: -0.407071, loss_mean_cls: 0.076766, grad_norm: 0.240466 +Steps: 1%| | 5283/1000000 [21:56<68:26:31, 4.04it/s, grad_norm=0.24, loss_final=0.495, loss_mean=0.825, loss_mean_cls=0.0768, proj_loss=-0.407][2026-03-26 12:52:59] Step: 5283, Training Logs: loss_final: 0.490358, loss_mean: 0.826302, proj_loss: -0.411285, loss_mean_cls: 0.075341, grad_norm: 0.370494 +Steps: 1%| | 5284/1000000 [21:56<68:28:52, 4.03it/s, grad_norm=0.37, loss_final=0.49, loss_mean=0.826, loss_mean_cls=0.0753, proj_loss=-0.411][2026-03-26 12:52:59] Step: 5284, Training Logs: loss_final: 0.476809, loss_mean: 0.807634, proj_loss: -0.408217, loss_mean_cls: 0.077391, grad_norm: 0.326036 +Steps: 1%| | 5285/1000000 [21:56<68:24:29, 4.04it/s, grad_norm=0.326, loss_final=0.477, loss_mean=0.808, loss_mean_cls=0.0774, proj_loss=-0.408][2026-03-26 12:52:59] Step: 5285, Training Logs: loss_final: 0.490334, loss_mean: 0.831981, proj_loss: -0.415769, loss_mean_cls: 0.074121, grad_norm: 0.296666 +Steps: 1%| | 5286/1000000 [21:56<68:26:49, 4.04it/s, grad_norm=0.297, loss_final=0.49, loss_mean=0.832, loss_mean_cls=0.0741, proj_loss=-0.416][2026-03-26 12:53:00] Step: 5286, Training Logs: loss_final: 0.489234, loss_mean: 0.832436, proj_loss: -0.417183, loss_mean_cls: 0.073981, grad_norm: 0.364889 +Steps: 1%| | 5287/1000000 [21:56<68:24:40, 4.04it/s, grad_norm=0.365, loss_final=0.489, loss_mean=0.832, loss_mean_cls=0.074, proj_loss=-0.417][2026-03-26 12:53:00] Step: 5287, Training Logs: loss_final: 0.493300, loss_mean: 0.832247, proj_loss: -0.414581, loss_mean_cls: 0.075635, grad_norm: 0.334142 +Steps: 1%| | 5288/1000000 [21:57<68:24:47, 4.04it/s, grad_norm=0.334, loss_final=0.493, loss_mean=0.832, loss_mean_cls=0.0756, proj_loss=-0.415][2026-03-26 12:53:00] Step: 5288, Training Logs: loss_final: 0.463369, loss_mean: 0.814938, proj_loss: -0.424848, loss_mean_cls: 0.073279, grad_norm: 0.286351 +Steps: 1%| | 5289/1000000 [21:57<68:20:26, 4.04it/s, grad_norm=0.286, loss_final=0.463, loss_mean=0.815, loss_mean_cls=0.0733, proj_loss=-0.425][2026-03-26 12:53:00] Step: 5289, Training Logs: loss_final: 0.479620, loss_mean: 0.824950, proj_loss: -0.419972, loss_mean_cls: 0.074642, grad_norm: 0.222562 +Steps: 1%| | 5290/1000000 [21:57<68:21:39, 4.04it/s, grad_norm=0.223, loss_final=0.48, loss_mean=0.825, loss_mean_cls=0.0746, proj_loss=-0.42][2026-03-26 12:53:01] Step: 5290, Training Logs: loss_final: 0.493366, loss_mean: 0.824275, proj_loss: -0.407070, loss_mean_cls: 0.076160, grad_norm: 0.357551 +Steps: 1%| | 5291/1000000 [21:58<72:12:59, 3.83it/s, grad_norm=0.358, loss_final=0.493, loss_mean=0.824, loss_mean_cls=0.0762, proj_loss=-0.407][2026-03-26 12:53:01] Step: 5291, Training Logs: loss_final: 0.482720, loss_mean: 0.817911, proj_loss: -0.411151, loss_mean_cls: 0.075960, grad_norm: 0.337594 +Steps: 1%| | 5292/1000000 [21:58<71:14:10, 3.88it/s, grad_norm=0.338, loss_final=0.483, loss_mean=0.818, loss_mean_cls=0.076, proj_loss=-0.411][2026-03-26 12:53:01] Step: 5292, Training Logs: loss_final: 0.500748, loss_mean: 0.837499, proj_loss: -0.411803, loss_mean_cls: 0.075052, grad_norm: 0.225748 +Steps: 1%| | 5293/1000000 [21:58<70:25:52, 3.92it/s, grad_norm=0.226, loss_final=0.501, loss_mean=0.837, loss_mean_cls=0.0751, proj_loss=-0.412][2026-03-26 12:53:01] Step: 5293, Training Logs: loss_final: 0.486338, loss_mean: 0.820910, proj_loss: -0.411259, loss_mean_cls: 0.076688, grad_norm: 0.272572 +Steps: 1%| | 5294/1000000 [21:58<69:52:03, 3.95it/s, grad_norm=0.273, loss_final=0.486, loss_mean=0.821, loss_mean_cls=0.0767, proj_loss=-0.411][2026-03-26 12:53:02] Step: 5294, Training Logs: loss_final: 0.484645, loss_mean: 0.816998, proj_loss: -0.409007, loss_mean_cls: 0.076655, grad_norm: 0.300993 +Steps: 1%| | 5295/1000000 [21:59<70:06:58, 3.94it/s, grad_norm=0.301, loss_final=0.485, loss_mean=0.817, loss_mean_cls=0.0767, proj_loss=-0.409][2026-03-26 12:53:02] Step: 5295, Training Logs: loss_final: 0.502230, loss_mean: 0.826940, proj_loss: -0.401254, loss_mean_cls: 0.076544, grad_norm: 0.450414 +Steps: 1%| | 5296/1000000 [21:59<69:36:15, 3.97it/s, grad_norm=0.45, loss_final=0.502, loss_mean=0.827, loss_mean_cls=0.0765, proj_loss=-0.401][2026-03-26 12:53:02] Step: 5296, Training Logs: loss_final: 0.472908, loss_mean: 0.815091, proj_loss: -0.416533, loss_mean_cls: 0.074350, grad_norm: 0.406062 +Steps: 1%| | 5297/1000000 [21:59<69:13:56, 3.99it/s, grad_norm=0.406, loss_final=0.473, loss_mean=0.815, loss_mean_cls=0.0744, proj_loss=-0.417][2026-03-26 12:53:02] Step: 5297, Training Logs: loss_final: 0.483865, loss_mean: 0.817066, proj_loss: -0.410252, loss_mean_cls: 0.077052, grad_norm: 0.390610 +Steps: 1%| | 5298/1000000 [21:59<68:59:13, 4.01it/s, grad_norm=0.391, loss_final=0.484, loss_mean=0.817, loss_mean_cls=0.0771, proj_loss=-0.41][2026-03-26 12:53:03] Step: 5298, Training Logs: loss_final: 0.486682, loss_mean: 0.826952, proj_loss: -0.413928, loss_mean_cls: 0.073658, grad_norm: 0.630037 +Steps: 1%| | 5299/1000000 [22:00<68:50:46, 4.01it/s, grad_norm=0.63, loss_final=0.487, loss_mean=0.827, loss_mean_cls=0.0737, proj_loss=-0.414][2026-03-26 12:53:03] Step: 5299, Training Logs: loss_final: 0.498268, loss_mean: 0.836492, proj_loss: -0.412333, loss_mean_cls: 0.074109, grad_norm: 0.481176 +Steps: 1%| | 5300/1000000 [22:00<68:42:42, 4.02it/s, grad_norm=0.481, loss_final=0.498, loss_mean=0.836, loss_mean_cls=0.0741, proj_loss=-0.412][2026-03-26 12:53:03] Step: 5300, Training Logs: loss_final: 0.479778, loss_mean: 0.816117, proj_loss: -0.410833, loss_mean_cls: 0.074495, grad_norm: 0.427537 +Steps: 1%| | 5301/1000000 [22:00<68:52:55, 4.01it/s, grad_norm=0.428, loss_final=0.48, loss_mean=0.816, loss_mean_cls=0.0745, proj_loss=-0.411][2026-03-26 12:53:03] Step: 5301, Training Logs: loss_final: 0.476016, loss_mean: 0.814875, proj_loss: -0.413650, loss_mean_cls: 0.074791, grad_norm: 0.627430 +Steps: 1%| | 5302/1000000 [22:00<68:44:32, 4.02it/s, grad_norm=0.627, loss_final=0.476, loss_mean=0.815, loss_mean_cls=0.0748, proj_loss=-0.414][2026-03-26 12:53:04] Step: 5302, Training Logs: loss_final: 0.474064, loss_mean: 0.828505, proj_loss: -0.425117, loss_mean_cls: 0.070676, grad_norm: 0.396681 +Steps: 1%| | 5303/1000000 [22:01<68:41:02, 4.02it/s, grad_norm=0.397, loss_final=0.474, loss_mean=0.829, loss_mean_cls=0.0707, proj_loss=-0.425][2026-03-26 12:53:04] Step: 5303, Training Logs: loss_final: 0.485654, loss_mean: 0.827517, proj_loss: -0.414327, loss_mean_cls: 0.072464, grad_norm: 0.493493 +Steps: 1%| | 5304/1000000 [22:01<68:36:23, 4.03it/s, grad_norm=0.493, loss_final=0.486, loss_mean=0.828, loss_mean_cls=0.0725, proj_loss=-0.414][2026-03-26 12:53:04] Step: 5304, Training Logs: loss_final: 0.481122, loss_mean: 0.840186, proj_loss: -0.429709, loss_mean_cls: 0.070646, grad_norm: 0.385080 +Steps: 1%| | 5305/1000000 [22:01<68:35:38, 4.03it/s, grad_norm=0.385, loss_final=0.481, loss_mean=0.84, loss_mean_cls=0.0706, proj_loss=-0.43][2026-03-26 12:53:04] Step: 5305, Training Logs: loss_final: 0.474645, loss_mean: 0.809476, proj_loss: -0.410641, loss_mean_cls: 0.075809, grad_norm: 0.515799 +Steps: 1%| | 5306/1000000 [22:01<68:31:13, 4.03it/s, grad_norm=0.516, loss_final=0.475, loss_mean=0.809, loss_mean_cls=0.0758, proj_loss=-0.411][2026-03-26 12:53:05] Step: 5306, Training Logs: loss_final: 0.511325, loss_mean: 0.856277, proj_loss: -0.417727, loss_mean_cls: 0.072775, grad_norm: 0.596905 +Steps: 1%| | 5307/1000000 [22:02<68:28:19, 4.04it/s, grad_norm=0.597, loss_final=0.511, loss_mean=0.856, loss_mean_cls=0.0728, proj_loss=-0.418][2026-03-26 12:53:05] Step: 5307, Training Logs: loss_final: 0.489346, loss_mean: 0.823781, proj_loss: -0.410301, loss_mean_cls: 0.075866, grad_norm: 0.391206 +Steps: 1%| | 5308/1000000 [22:02<68:29:01, 4.03it/s, grad_norm=0.391, loss_final=0.489, loss_mean=0.824, loss_mean_cls=0.0759, proj_loss=-0.41][2026-03-26 12:53:05] Step: 5308, Training Logs: loss_final: 0.486850, loss_mean: 0.819345, proj_loss: -0.407493, loss_mean_cls: 0.074999, grad_norm: 0.547758 +Steps: 1%| | 5309/1000000 [22:02<68:27:32, 4.04it/s, grad_norm=0.548, loss_final=0.487, loss_mean=0.819, loss_mean_cls=0.075, proj_loss=-0.407][2026-03-26 12:53:05] Step: 5309, Training Logs: loss_final: 0.490844, loss_mean: 0.834119, proj_loss: -0.416428, loss_mean_cls: 0.073153, grad_norm: 0.546635 +Steps: 1%| | 5310/1000000 [22:02<68:28:23, 4.04it/s, grad_norm=0.547, loss_final=0.491, loss_mean=0.834, loss_mean_cls=0.0732, proj_loss=-0.416][2026-03-26 12:53:06] Step: 5310, Training Logs: loss_final: 0.491667, loss_mean: 0.830125, proj_loss: -0.413062, loss_mean_cls: 0.074604, grad_norm: 0.403790 +Steps: 1%| | 5311/1000000 [22:03<68:26:50, 4.04it/s, grad_norm=0.404, loss_final=0.492, loss_mean=0.83, loss_mean_cls=0.0746, proj_loss=-0.413][2026-03-26 12:53:06] Step: 5311, Training Logs: loss_final: 0.493457, loss_mean: 0.836220, proj_loss: -0.416054, loss_mean_cls: 0.073291, grad_norm: 0.269968 +Steps: 1%| | 5312/1000000 [22:03<68:28:14, 4.04it/s, grad_norm=0.27, loss_final=0.493, loss_mean=0.836, loss_mean_cls=0.0733, proj_loss=-0.416][2026-03-26 12:53:06] Step: 5312, Training Logs: loss_final: 0.480862, loss_mean: 0.818500, proj_loss: -0.412264, loss_mean_cls: 0.074626, grad_norm: 0.289984 +Steps: 1%| | 5313/1000000 [22:03<68:26:17, 4.04it/s, grad_norm=0.29, loss_final=0.481, loss_mean=0.818, loss_mean_cls=0.0746, proj_loss=-0.412][2026-03-26 12:53:06] Step: 5313, Training Logs: loss_final: 0.468435, loss_mean: 0.810125, proj_loss: -0.416123, loss_mean_cls: 0.074433, grad_norm: 0.431532 +Steps: 1%| | 5314/1000000 [22:03<68:28:41, 4.03it/s, grad_norm=0.432, loss_final=0.468, loss_mean=0.81, loss_mean_cls=0.0744, proj_loss=-0.416][2026-03-26 12:53:07] Step: 5314, Training Logs: loss_final: 0.473663, loss_mean: 0.808098, proj_loss: -0.411001, loss_mean_cls: 0.076567, grad_norm: 0.280135 +Steps: 1%| | 5315/1000000 [22:03<68:26:54, 4.04it/s, grad_norm=0.28, loss_final=0.474, loss_mean=0.808, loss_mean_cls=0.0766, proj_loss=-0.411][2026-03-26 12:53:07] Step: 5315, Training Logs: loss_final: 0.489141, loss_mean: 0.826874, proj_loss: -0.412729, loss_mean_cls: 0.074996, grad_norm: 0.271048 +Steps: 1%| | 5316/1000000 [22:04<68:27:44, 4.04it/s, grad_norm=0.271, loss_final=0.489, loss_mean=0.827, loss_mean_cls=0.075, proj_loss=-0.413][2026-03-26 12:53:07] Step: 5316, Training Logs: loss_final: 0.491470, loss_mean: 0.841010, proj_loss: -0.421303, loss_mean_cls: 0.071763, grad_norm: 0.378135 +Steps: 1%| | 5317/1000000 [22:04<68:27:04, 4.04it/s, grad_norm=0.378, loss_final=0.491, loss_mean=0.841, loss_mean_cls=0.0718, proj_loss=-0.421][2026-03-26 12:53:07] Step: 5317, Training Logs: loss_final: 0.488541, loss_mean: 0.831900, proj_loss: -0.417105, loss_mean_cls: 0.073746, grad_norm: 0.212048 +Steps: 1%| | 5318/1000000 [22:04<68:26:37, 4.04it/s, grad_norm=0.212, loss_final=0.489, loss_mean=0.832, loss_mean_cls=0.0737, proj_loss=-0.417][2026-03-26 12:53:08] Step: 5318, Training Logs: loss_final: 0.473313, loss_mean: 0.816685, proj_loss: -0.417252, loss_mean_cls: 0.073880, grad_norm: 0.378609 +Steps: 1%| | 5319/1000000 [22:04<68:22:37, 4.04it/s, grad_norm=0.379, loss_final=0.473, loss_mean=0.817, loss_mean_cls=0.0739, proj_loss=-0.417][2026-03-26 12:53:08] Step: 5319, Training Logs: loss_final: 0.476263, loss_mean: 0.825046, proj_loss: -0.422541, loss_mean_cls: 0.073757, grad_norm: 0.393975 +Steps: 1%| | 5320/1000000 [22:05<68:28:10, 4.04it/s, grad_norm=0.394, loss_final=0.476, loss_mean=0.825, loss_mean_cls=0.0738, proj_loss=-0.423][2026-03-26 12:53:08] Step: 5320, Training Logs: loss_final: 0.483060, loss_mean: 0.816815, proj_loss: -0.409818, loss_mean_cls: 0.076062, grad_norm: 0.216050 +Steps: 1%| | 5321/1000000 [22:05<68:25:24, 4.04it/s, grad_norm=0.216, loss_final=0.483, loss_mean=0.817, loss_mean_cls=0.0761, proj_loss=-0.41][2026-03-26 12:53:08] Step: 5321, Training Logs: loss_final: 0.477250, loss_mean: 0.830927, proj_loss: -0.425898, loss_mean_cls: 0.072220, grad_norm: 0.339558 +Steps: 1%| | 5322/1000000 [22:05<68:31:17, 4.03it/s, grad_norm=0.34, loss_final=0.477, loss_mean=0.831, loss_mean_cls=0.0722, proj_loss=-0.426][2026-03-26 12:53:09] Step: 5322, Training Logs: loss_final: 0.492261, loss_mean: 0.833303, proj_loss: -0.414831, loss_mean_cls: 0.073789, grad_norm: 0.250666 +Steps: 1%| | 5323/1000000 [22:05<68:28:35, 4.03it/s, grad_norm=0.251, loss_final=0.492, loss_mean=0.833, loss_mean_cls=0.0738, proj_loss=-0.415][2026-03-26 12:53:09] Step: 5323, Training Logs: loss_final: 0.494029, loss_mean: 0.835358, proj_loss: -0.414294, loss_mean_cls: 0.072966, grad_norm: 0.462980 +Steps: 1%| | 5324/1000000 [22:06<68:28:09, 4.04it/s, grad_norm=0.463, loss_final=0.494, loss_mean=0.835, loss_mean_cls=0.073, proj_loss=-0.414][2026-03-26 12:53:09] Step: 5324, Training Logs: loss_final: 0.491072, loss_mean: 0.840183, proj_loss: -0.421047, loss_mean_cls: 0.071936, grad_norm: 0.424679 +Steps: 1%| | 5325/1000000 [22:06<68:26:46, 4.04it/s, grad_norm=0.425, loss_final=0.491, loss_mean=0.84, loss_mean_cls=0.0719, proj_loss=-0.421][2026-03-26 12:53:09] Step: 5325, Training Logs: loss_final: 0.493234, loss_mean: 0.838657, proj_loss: -0.419610, loss_mean_cls: 0.074186, grad_norm: 0.366354 +Steps: 1%| | 5326/1000000 [22:06<68:25:19, 4.04it/s, grad_norm=0.366, loss_final=0.493, loss_mean=0.839, loss_mean_cls=0.0742, proj_loss=-0.42][2026-03-26 12:53:10] Step: 5326, Training Logs: loss_final: 0.499020, loss_mean: 0.846987, proj_loss: -0.419508, loss_mean_cls: 0.071541, grad_norm: 0.545260 +Steps: 1%| | 5327/1000000 [22:06<68:26:42, 4.04it/s, grad_norm=0.545, loss_final=0.499, loss_mean=0.847, loss_mean_cls=0.0715, proj_loss=-0.42][2026-03-26 12:53:10] Step: 5327, Training Logs: loss_final: 0.484816, loss_mean: 0.824894, proj_loss: -0.413836, loss_mean_cls: 0.073757, grad_norm: 0.426366 +Steps: 1%| | 5328/1000000 [22:07<68:26:50, 4.04it/s, grad_norm=0.426, loss_final=0.485, loss_mean=0.825, loss_mean_cls=0.0738, proj_loss=-0.414][2026-03-26 12:53:10] Step: 5328, Training Logs: loss_final: 0.482177, loss_mean: 0.820388, proj_loss: -0.412341, loss_mean_cls: 0.074131, grad_norm: 0.380506 +Steps: 1%| | 5329/1000000 [22:07<68:27:12, 4.04it/s, grad_norm=0.381, loss_final=0.482, loss_mean=0.82, loss_mean_cls=0.0741, proj_loss=-0.412][2026-03-26 12:53:10] Step: 5329, Training Logs: loss_final: 0.480593, loss_mean: 0.813584, proj_loss: -0.409387, loss_mean_cls: 0.076396, grad_norm: 0.465146 +Steps: 1%| | 5330/1000000 [22:07<68:28:21, 4.04it/s, grad_norm=0.465, loss_final=0.481, loss_mean=0.814, loss_mean_cls=0.0764, proj_loss=-0.409][2026-03-26 12:53:11] Step: 5330, Training Logs: loss_final: 0.495393, loss_mean: 0.838571, proj_loss: -0.416908, loss_mean_cls: 0.073731, grad_norm: 0.273546 +Steps: 1%| | 5331/1000000 [22:07<68:26:42, 4.04it/s, grad_norm=0.274, loss_final=0.495, loss_mean=0.839, loss_mean_cls=0.0737, proj_loss=-0.417][2026-03-26 12:53:11] Step: 5331, Training Logs: loss_final: 0.490472, loss_mean: 0.823427, proj_loss: -0.408883, loss_mean_cls: 0.075929, grad_norm: 0.372041 +Steps: 1%| | 5332/1000000 [22:08<68:24:31, 4.04it/s, grad_norm=0.372, loss_final=0.49, loss_mean=0.823, loss_mean_cls=0.0759, proj_loss=-0.409][2026-03-26 12:53:11] Step: 5332, Training Logs: loss_final: 0.504817, loss_mean: 0.837572, proj_loss: -0.407813, loss_mean_cls: 0.075058, grad_norm: 0.296678 +Steps: 1%| | 5333/1000000 [22:08<68:25:36, 4.04it/s, grad_norm=0.297, loss_final=0.505, loss_mean=0.838, loss_mean_cls=0.0751, proj_loss=-0.408][2026-03-26 12:53:11] Step: 5333, Training Logs: loss_final: 0.483824, loss_mean: 0.829455, proj_loss: -0.418927, loss_mean_cls: 0.073297, grad_norm: 0.342113 +Steps: 1%| | 5334/1000000 [22:08<68:24:04, 4.04it/s, grad_norm=0.342, loss_final=0.484, loss_mean=0.829, loss_mean_cls=0.0733, proj_loss=-0.419][2026-03-26 12:53:12] Step: 5334, Training Logs: loss_final: 0.481537, loss_mean: 0.815062, proj_loss: -0.409227, loss_mean_cls: 0.075702, grad_norm: 0.312683 +Steps: 1%| | 5335/1000000 [22:08<68:24:09, 4.04it/s, grad_norm=0.313, loss_final=0.482, loss_mean=0.815, loss_mean_cls=0.0757, proj_loss=-0.409][2026-03-26 12:53:12] Step: 5335, Training Logs: loss_final: 0.478267, loss_mean: 0.807744, proj_loss: -0.406165, loss_mean_cls: 0.076688, grad_norm: 0.477412 +Steps: 1%| | 5336/1000000 [22:09<68:24:37, 4.04it/s, grad_norm=0.477, loss_final=0.478, loss_mean=0.808, loss_mean_cls=0.0767, proj_loss=-0.406][2026-03-26 12:53:12] Step: 5336, Training Logs: loss_final: 0.482964, loss_mean: 0.823422, proj_loss: -0.415597, loss_mean_cls: 0.075138, grad_norm: 0.608673 +Steps: 1%| | 5337/1000000 [22:09<68:26:29, 4.04it/s, grad_norm=0.609, loss_final=0.483, loss_mean=0.823, loss_mean_cls=0.0751, proj_loss=-0.416][2026-03-26 12:53:12] Step: 5337, Training Logs: loss_final: 0.480299, loss_mean: 0.824177, proj_loss: -0.417271, loss_mean_cls: 0.073393, grad_norm: 0.372888 +Steps: 1%| | 5338/1000000 [22:09<68:25:07, 4.04it/s, grad_norm=0.373, loss_final=0.48, loss_mean=0.824, loss_mean_cls=0.0734, proj_loss=-0.417][2026-03-26 12:53:13] Step: 5338, Training Logs: loss_final: 0.492009, loss_mean: 0.818403, proj_loss: -0.402932, loss_mean_cls: 0.076538, grad_norm: 0.326720 +Steps: 1%| | 5339/1000000 [22:09<68:26:27, 4.04it/s, grad_norm=0.327, loss_final=0.492, loss_mean=0.818, loss_mean_cls=0.0765, proj_loss=-0.403][2026-03-26 12:53:13] Step: 5339, Training Logs: loss_final: 0.473254, loss_mean: 0.815947, proj_loss: -0.416369, loss_mean_cls: 0.073676, grad_norm: 0.454024 +Steps: 1%| | 5340/1000000 [22:10<71:40:59, 3.85it/s, grad_norm=0.454, loss_final=0.473, loss_mean=0.816, loss_mean_cls=0.0737, proj_loss=-0.416][2026-03-26 12:53:13] Step: 5340, Training Logs: loss_final: 0.492984, loss_mean: 0.838511, proj_loss: -0.418814, loss_mean_cls: 0.073287, grad_norm: 0.362998 +Steps: 1%| | 5341/1000000 [22:10<70:41:24, 3.91it/s, grad_norm=0.363, loss_final=0.493, loss_mean=0.839, loss_mean_cls=0.0733, proj_loss=-0.419][2026-03-26 12:53:13] Step: 5341, Training Logs: loss_final: 0.476381, loss_mean: 0.809695, proj_loss: -0.408203, loss_mean_cls: 0.074889, grad_norm: 0.291637 +Steps: 1%| | 5342/1000000 [22:10<70:00:05, 3.95it/s, grad_norm=0.292, loss_final=0.476, loss_mean=0.81, loss_mean_cls=0.0749, proj_loss=-0.408][2026-03-26 12:53:14] Step: 5342, Training Logs: loss_final: 0.498850, loss_mean: 0.829391, proj_loss: -0.406871, loss_mean_cls: 0.076330, grad_norm: 0.440545 +Steps: 1%| | 5343/1000000 [22:10<69:32:28, 3.97it/s, grad_norm=0.441, loss_final=0.499, loss_mean=0.829, loss_mean_cls=0.0763, proj_loss=-0.407][2026-03-26 12:53:14] Step: 5343, Training Logs: loss_final: 0.492764, loss_mean: 0.834677, proj_loss: -0.416578, loss_mean_cls: 0.074664, grad_norm: 0.314051 +Steps: 1%| | 5344/1000000 [22:11<69:13:20, 3.99it/s, grad_norm=0.314, loss_final=0.493, loss_mean=0.835, loss_mean_cls=0.0747, proj_loss=-0.417][2026-03-26 12:53:14] Step: 5344, Training Logs: loss_final: 0.497386, loss_mean: 0.836282, proj_loss: -0.412976, loss_mean_cls: 0.074080, grad_norm: 0.275726 +Steps: 1%| | 5345/1000000 [22:11<68:56:30, 4.01it/s, grad_norm=0.276, loss_final=0.497, loss_mean=0.836, loss_mean_cls=0.0741, proj_loss=-0.413][2026-03-26 12:53:14] Step: 5345, Training Logs: loss_final: 0.480398, loss_mean: 0.819340, proj_loss: -0.412775, loss_mean_cls: 0.073833, grad_norm: 0.399693 +Steps: 1%| | 5346/1000000 [22:11<68:47:44, 4.02it/s, grad_norm=0.4, loss_final=0.48, loss_mean=0.819, loss_mean_cls=0.0738, proj_loss=-0.413][2026-03-26 12:53:15] Step: 5346, Training Logs: loss_final: 0.480919, loss_mean: 0.834311, proj_loss: -0.424305, loss_mean_cls: 0.070913, grad_norm: 0.434603 +Steps: 1%| | 5347/1000000 [22:11<68:41:13, 4.02it/s, grad_norm=0.435, loss_final=0.481, loss_mean=0.834, loss_mean_cls=0.0709, proj_loss=-0.424][2026-03-26 12:53:15] Step: 5347, Training Logs: loss_final: 0.484442, loss_mean: 0.829557, proj_loss: -0.418252, loss_mean_cls: 0.073137, grad_norm: 0.268806 +Steps: 1%| | 5348/1000000 [22:12<68:35:58, 4.03it/s, grad_norm=0.269, loss_final=0.484, loss_mean=0.83, loss_mean_cls=0.0731, proj_loss=-0.418][2026-03-26 12:53:15] Step: 5348, Training Logs: loss_final: 0.489539, loss_mean: 0.832533, proj_loss: -0.415419, loss_mean_cls: 0.072425, grad_norm: 0.478907 +Steps: 1%| | 5349/1000000 [22:12<68:34:16, 4.03it/s, grad_norm=0.479, loss_final=0.49, loss_mean=0.833, loss_mean_cls=0.0724, proj_loss=-0.415][2026-03-26 12:53:15] Step: 5349, Training Logs: loss_final: 0.495684, loss_mean: 0.827174, proj_loss: -0.407075, loss_mean_cls: 0.075585, grad_norm: 0.457247 +Steps: 1%| | 5350/1000000 [22:12<68:29:15, 4.03it/s, grad_norm=0.457, loss_final=0.496, loss_mean=0.827, loss_mean_cls=0.0756, proj_loss=-0.407][2026-03-26 12:53:16] Step: 5350, Training Logs: loss_final: 0.479053, loss_mean: 0.825125, proj_loss: -0.418998, loss_mean_cls: 0.072926, grad_norm: 0.260104 +Steps: 1%| | 5351/1000000 [22:12<68:29:32, 4.03it/s, grad_norm=0.26, loss_final=0.479, loss_mean=0.825, loss_mean_cls=0.0729, proj_loss=-0.419][2026-03-26 12:53:16] Step: 5351, Training Logs: loss_final: 0.473873, loss_mean: 0.816366, proj_loss: -0.416885, loss_mean_cls: 0.074393, grad_norm: 0.540417 +Steps: 1%| | 5352/1000000 [22:13<68:26:19, 4.04it/s, grad_norm=0.54, loss_final=0.474, loss_mean=0.816, loss_mean_cls=0.0744, proj_loss=-0.417][2026-03-26 12:53:16] Step: 5352, Training Logs: loss_final: 0.465010, loss_mean: 0.790797, proj_loss: -0.402941, loss_mean_cls: 0.077155, grad_norm: 0.213698 +Steps: 1%| | 5353/1000000 [22:13<68:25:29, 4.04it/s, grad_norm=0.214, loss_final=0.465, loss_mean=0.791, loss_mean_cls=0.0772, proj_loss=-0.403][2026-03-26 12:53:16] Step: 5353, Training Logs: loss_final: 0.491873, loss_mean: 0.822617, proj_loss: -0.406801, loss_mean_cls: 0.076057, grad_norm: 0.562644 +Steps: 1%| | 5354/1000000 [22:13<68:24:23, 4.04it/s, grad_norm=0.563, loss_final=0.492, loss_mean=0.823, loss_mean_cls=0.0761, proj_loss=-0.407][2026-03-26 12:53:17] Step: 5354, Training Logs: loss_final: 0.489607, loss_mean: 0.828578, proj_loss: -0.413303, loss_mean_cls: 0.074332, grad_norm: 0.581908 +Steps: 1%| | 5355/1000000 [22:13<68:23:57, 4.04it/s, grad_norm=0.582, loss_final=0.49, loss_mean=0.829, loss_mean_cls=0.0743, proj_loss=-0.413][2026-03-26 12:53:17] Step: 5355, Training Logs: loss_final: 0.493336, loss_mean: 0.837197, proj_loss: -0.418029, loss_mean_cls: 0.074167, grad_norm: 0.248644 +Steps: 1%| | 5356/1000000 [22:14<68:24:22, 4.04it/s, grad_norm=0.249, loss_final=0.493, loss_mean=0.837, loss_mean_cls=0.0742, proj_loss=-0.418][2026-03-26 12:53:17] Step: 5356, Training Logs: loss_final: 0.475170, loss_mean: 0.813291, proj_loss: -0.411984, loss_mean_cls: 0.073863, grad_norm: 0.526414 +Steps: 1%| | 5357/1000000 [22:14<68:24:01, 4.04it/s, grad_norm=0.526, loss_final=0.475, loss_mean=0.813, loss_mean_cls=0.0739, proj_loss=-0.412][2026-03-26 12:53:17] Step: 5357, Training Logs: loss_final: 0.478761, loss_mean: 0.824136, proj_loss: -0.417704, loss_mean_cls: 0.072329, grad_norm: 0.320699 +Steps: 1%| | 5358/1000000 [22:14<68:26:17, 4.04it/s, grad_norm=0.321, loss_final=0.479, loss_mean=0.824, loss_mean_cls=0.0723, proj_loss=-0.418][2026-03-26 12:53:18] Step: 5358, Training Logs: loss_final: 0.498460, loss_mean: 0.838030, proj_loss: -0.413207, loss_mean_cls: 0.073637, grad_norm: 0.324558 +Steps: 1%| | 5359/1000000 [22:14<68:26:31, 4.04it/s, grad_norm=0.325, loss_final=0.498, loss_mean=0.838, loss_mean_cls=0.0736, proj_loss=-0.413][2026-03-26 12:53:18] Step: 5359, Training Logs: loss_final: 0.503033, loss_mean: 0.845562, proj_loss: -0.415714, loss_mean_cls: 0.073184, grad_norm: 0.312767 +Steps: 1%| | 5360/1000000 [22:15<68:28:09, 4.04it/s, grad_norm=0.313, loss_final=0.503, loss_mean=0.846, loss_mean_cls=0.0732, proj_loss=-0.416][2026-03-26 12:53:18] Step: 5360, Training Logs: loss_final: 0.471413, loss_mean: 0.816698, proj_loss: -0.418782, loss_mean_cls: 0.073497, grad_norm: 0.320287 +Steps: 1%| | 5361/1000000 [22:15<68:26:53, 4.04it/s, grad_norm=0.32, loss_final=0.471, loss_mean=0.817, loss_mean_cls=0.0735, proj_loss=-0.419][2026-03-26 12:53:18] Step: 5361, Training Logs: loss_final: 0.471335, loss_mean: 0.807260, proj_loss: -0.410875, loss_mean_cls: 0.074950, grad_norm: 0.427048 +Steps: 1%| | 5362/1000000 [22:15<68:26:39, 4.04it/s, grad_norm=0.427, loss_final=0.471, loss_mean=0.807, loss_mean_cls=0.075, proj_loss=-0.411][2026-03-26 12:53:19] Step: 5362, Training Logs: loss_final: 0.498043, loss_mean: 0.834121, proj_loss: -0.409998, loss_mean_cls: 0.073920, grad_norm: 0.290577 +Steps: 1%| | 5363/1000000 [22:15<68:26:05, 4.04it/s, grad_norm=0.291, loss_final=0.498, loss_mean=0.834, loss_mean_cls=0.0739, proj_loss=-0.41][2026-03-26 12:53:19] Step: 5363, Training Logs: loss_final: 0.476705, loss_mean: 0.817091, proj_loss: -0.415065, loss_mean_cls: 0.074679, grad_norm: 0.224172 +Steps: 1%| | 5364/1000000 [22:16<68:31:07, 4.03it/s, grad_norm=0.224, loss_final=0.477, loss_mean=0.817, loss_mean_cls=0.0747, proj_loss=-0.415][2026-03-26 12:53:19] Step: 5364, Training Logs: loss_final: 0.484337, loss_mean: 0.821733, proj_loss: -0.412130, loss_mean_cls: 0.074734, grad_norm: 0.543618 +Steps: 1%| | 5365/1000000 [22:16<68:28:24, 4.03it/s, grad_norm=0.544, loss_final=0.484, loss_mean=0.822, loss_mean_cls=0.0747, proj_loss=-0.412][2026-03-26 12:53:19] Step: 5365, Training Logs: loss_final: 0.495404, loss_mean: 0.842083, proj_loss: -0.419739, loss_mean_cls: 0.073060, grad_norm: 0.311964 +Steps: 1%| | 5366/1000000 [22:16<68:28:55, 4.03it/s, grad_norm=0.312, loss_final=0.495, loss_mean=0.842, loss_mean_cls=0.0731, proj_loss=-0.42][2026-03-26 12:53:20] Step: 5366, Training Logs: loss_final: 0.498060, loss_mean: 0.833599, proj_loss: -0.409716, loss_mean_cls: 0.074177, grad_norm: 0.303511 +Steps: 1%| | 5367/1000000 [22:16<68:26:02, 4.04it/s, grad_norm=0.304, loss_final=0.498, loss_mean=0.834, loss_mean_cls=0.0742, proj_loss=-0.41][2026-03-26 12:53:20] Step: 5367, Training Logs: loss_final: 0.484332, loss_mean: 0.828919, proj_loss: -0.417466, loss_mean_cls: 0.072880, grad_norm: 0.375343 +Steps: 1%| | 5368/1000000 [22:17<68:24:14, 4.04it/s, grad_norm=0.375, loss_final=0.484, loss_mean=0.829, loss_mean_cls=0.0729, proj_loss=-0.417][2026-03-26 12:53:20] Step: 5368, Training Logs: loss_final: 0.491216, loss_mean: 0.820016, proj_loss: -0.404723, loss_mean_cls: 0.075924, grad_norm: 0.403599 +Steps: 1%| | 5369/1000000 [22:17<68:23:58, 4.04it/s, grad_norm=0.404, loss_final=0.491, loss_mean=0.82, loss_mean_cls=0.0759, proj_loss=-0.405][2026-03-26 12:53:20] Step: 5369, Training Logs: loss_final: 0.491341, loss_mean: 0.821999, proj_loss: -0.405175, loss_mean_cls: 0.074516, grad_norm: 0.404962 +Steps: 1%| | 5370/1000000 [22:17<68:24:28, 4.04it/s, grad_norm=0.405, loss_final=0.491, loss_mean=0.822, loss_mean_cls=0.0745, proj_loss=-0.405][2026-03-26 12:53:21] Step: 5370, Training Logs: loss_final: 0.491721, loss_mean: 0.828053, proj_loss: -0.410833, loss_mean_cls: 0.074502, grad_norm: 0.365112 +Steps: 1%| | 5371/1000000 [22:17<68:21:41, 4.04it/s, grad_norm=0.365, loss_final=0.492, loss_mean=0.828, loss_mean_cls=0.0745, proj_loss=-0.411][2026-03-26 12:53:21] Step: 5371, Training Logs: loss_final: 0.489808, loss_mean: 0.821617, proj_loss: -0.407299, loss_mean_cls: 0.075490, grad_norm: 0.367725 +Steps: 1%| | 5372/1000000 [22:18<68:25:45, 4.04it/s, grad_norm=0.368, loss_final=0.49, loss_mean=0.822, loss_mean_cls=0.0755, proj_loss=-0.407][2026-03-26 12:53:21] Step: 5372, Training Logs: loss_final: 0.507118, loss_mean: 0.850102, proj_loss: -0.415346, loss_mean_cls: 0.072362, grad_norm: 0.325063 +Steps: 1%| | 5373/1000000 [22:18<68:23:48, 4.04it/s, grad_norm=0.325, loss_final=0.507, loss_mean=0.85, loss_mean_cls=0.0724, proj_loss=-0.415][2026-03-26 12:53:21] Step: 5373, Training Logs: loss_final: 0.488003, loss_mean: 0.830978, proj_loss: -0.416774, loss_mean_cls: 0.073800, grad_norm: 0.327091 +Steps: 1%| | 5374/1000000 [22:18<68:25:10, 4.04it/s, grad_norm=0.327, loss_final=0.488, loss_mean=0.831, loss_mean_cls=0.0738, proj_loss=-0.417][2026-03-26 12:53:22] Step: 5374, Training Logs: loss_final: 0.481015, loss_mean: 0.821083, proj_loss: -0.414731, loss_mean_cls: 0.074663, grad_norm: 0.255636 +Steps: 1%| | 5375/1000000 [22:18<68:22:41, 4.04it/s, grad_norm=0.256, loss_final=0.481, loss_mean=0.821, loss_mean_cls=0.0747, proj_loss=-0.415][2026-03-26 12:53:22] Step: 5375, Training Logs: loss_final: 0.491567, loss_mean: 0.818399, proj_loss: -0.402749, loss_mean_cls: 0.075916, grad_norm: 0.351804 +Steps: 1%| | 5376/1000000 [22:19<68:24:59, 4.04it/s, grad_norm=0.352, loss_final=0.492, loss_mean=0.818, loss_mean_cls=0.0759, proj_loss=-0.403][2026-03-26 12:53:22] Step: 5376, Training Logs: loss_final: 0.484320, loss_mean: 0.818566, proj_loss: -0.409344, loss_mean_cls: 0.075099, grad_norm: 0.296640 +Steps: 1%| | 5377/1000000 [22:19<68:22:51, 4.04it/s, grad_norm=0.297, loss_final=0.484, loss_mean=0.819, loss_mean_cls=0.0751, proj_loss=-0.409][2026-03-26 12:53:22] Step: 5377, Training Logs: loss_final: 0.491189, loss_mean: 0.832853, proj_loss: -0.415207, loss_mean_cls: 0.073543, grad_norm: 0.279807 +Steps: 1%| | 5378/1000000 [22:19<68:23:34, 4.04it/s, grad_norm=0.28, loss_final=0.491, loss_mean=0.833, loss_mean_cls=0.0735, proj_loss=-0.415][2026-03-26 12:53:23] Step: 5378, Training Logs: loss_final: 0.482732, loss_mean: 0.813209, proj_loss: -0.406186, loss_mean_cls: 0.075710, grad_norm: 0.476947 +Steps: 1%| | 5379/1000000 [22:19<68:23:30, 4.04it/s, grad_norm=0.477, loss_final=0.483, loss_mean=0.813, loss_mean_cls=0.0757, proj_loss=-0.406][2026-03-26 12:53:23] Step: 5379, Training Logs: loss_final: 0.473137, loss_mean: 0.807511, proj_loss: -0.409289, loss_mean_cls: 0.074914, grad_norm: 0.359717 +Steps: 1%| | 5380/1000000 [22:20<68:26:35, 4.04it/s, grad_norm=0.36, loss_final=0.473, loss_mean=0.808, loss_mean_cls=0.0749, proj_loss=-0.409][2026-03-26 12:53:23] Step: 5380, Training Logs: loss_final: 0.477743, loss_mean: 0.818839, proj_loss: -0.415052, loss_mean_cls: 0.073956, grad_norm: 0.382642 +Steps: 1%| | 5381/1000000 [22:20<68:24:16, 4.04it/s, grad_norm=0.383, loss_final=0.478, loss_mean=0.819, loss_mean_cls=0.074, proj_loss=-0.415][2026-03-26 12:53:23] Step: 5381, Training Logs: loss_final: 0.491698, loss_mean: 0.831737, proj_loss: -0.413739, loss_mean_cls: 0.073700, grad_norm: 0.701449 +Steps: 1%| | 5382/1000000 [22:20<68:29:06, 4.03it/s, grad_norm=0.701, loss_final=0.492, loss_mean=0.832, loss_mean_cls=0.0737, proj_loss=-0.414][2026-03-26 12:53:24] Step: 5382, Training Logs: loss_final: 0.494844, loss_mean: 0.835839, proj_loss: -0.413594, loss_mean_cls: 0.072599, grad_norm: 0.408763 +Steps: 1%| | 5383/1000000 [22:20<68:27:07, 4.04it/s, grad_norm=0.409, loss_final=0.495, loss_mean=0.836, loss_mean_cls=0.0726, proj_loss=-0.414][2026-03-26 12:53:24] Step: 5383, Training Logs: loss_final: 0.485142, loss_mean: 0.819752, proj_loss: -0.410616, loss_mean_cls: 0.076007, grad_norm: 0.397871 +Steps: 1%| | 5384/1000000 [22:21<68:26:29, 4.04it/s, grad_norm=0.398, loss_final=0.485, loss_mean=0.82, loss_mean_cls=0.076, proj_loss=-0.411][2026-03-26 12:53:24] Step: 5384, Training Logs: loss_final: 0.476686, loss_mean: 0.823716, proj_loss: -0.419120, loss_mean_cls: 0.072089, grad_norm: 0.404209 +Steps: 1%| | 5385/1000000 [22:21<68:27:59, 4.04it/s, grad_norm=0.404, loss_final=0.477, loss_mean=0.824, loss_mean_cls=0.0721, proj_loss=-0.419][2026-03-26 12:53:24] Step: 5385, Training Logs: loss_final: 0.469795, loss_mean: 0.808049, proj_loss: -0.412944, loss_mean_cls: 0.074690, grad_norm: 0.334681 +Steps: 1%| | 5386/1000000 [22:21<68:26:42, 4.04it/s, grad_norm=0.335, loss_final=0.47, loss_mean=0.808, loss_mean_cls=0.0747, proj_loss=-0.413][2026-03-26 12:53:25] Step: 5386, Training Logs: loss_final: 0.512225, loss_mean: 0.854169, proj_loss: -0.416831, loss_mean_cls: 0.074887, grad_norm: 0.377533 +Steps: 1%| | 5387/1000000 [22:21<68:29:00, 4.03it/s, grad_norm=0.378, loss_final=0.512, loss_mean=0.854, loss_mean_cls=0.0749, proj_loss=-0.417][2026-03-26 12:53:25] Step: 5387, Training Logs: loss_final: 0.475287, loss_mean: 0.814475, proj_loss: -0.413615, loss_mean_cls: 0.074426, grad_norm: 0.273339 +Steps: 1%| | 5388/1000000 [22:22<68:27:42, 4.04it/s, grad_norm=0.273, loss_final=0.475, loss_mean=0.814, loss_mean_cls=0.0744, proj_loss=-0.414][2026-03-26 12:53:25] Step: 5388, Training Logs: loss_final: 0.497490, loss_mean: 0.844840, proj_loss: -0.418569, loss_mean_cls: 0.071218, grad_norm: 0.320653 +Steps: 1%| | 5389/1000000 [22:22<68:24:24, 4.04it/s, grad_norm=0.321, loss_final=0.497, loss_mean=0.845, loss_mean_cls=0.0712, proj_loss=-0.419][2026-03-26 12:53:25] Step: 5389, Training Logs: loss_final: 0.485052, loss_mean: 0.827307, proj_loss: -0.415417, loss_mean_cls: 0.073162, grad_norm: 0.454186 +Steps: 1%| | 5390/1000000 [22:22<68:22:33, 4.04it/s, grad_norm=0.454, loss_final=0.485, loss_mean=0.827, loss_mean_cls=0.0732, proj_loss=-0.415][2026-03-26 12:53:25] Step: 5390, Training Logs: loss_final: 0.495218, loss_mean: 0.840291, proj_loss: -0.416905, loss_mean_cls: 0.071832, grad_norm: 0.589842 +Steps: 1%| | 5391/1000000 [22:22<68:23:22, 4.04it/s, grad_norm=0.59, loss_final=0.495, loss_mean=0.84, loss_mean_cls=0.0718, proj_loss=-0.417][2026-03-26 12:53:26] Step: 5391, Training Logs: loss_final: 0.481340, loss_mean: 0.821709, proj_loss: -0.414158, loss_mean_cls: 0.073788, grad_norm: 0.244105 +Steps: 1%| | 5392/1000000 [22:23<68:23:57, 4.04it/s, grad_norm=0.244, loss_final=0.481, loss_mean=0.822, loss_mean_cls=0.0738, proj_loss=-0.414][2026-03-26 12:53:26] Step: 5392, Training Logs: loss_final: 0.482106, loss_mean: 0.816771, proj_loss: -0.409647, loss_mean_cls: 0.074982, grad_norm: 0.552361 +Steps: 1%| | 5393/1000000 [22:23<68:24:11, 4.04it/s, grad_norm=0.552, loss_final=0.482, loss_mean=0.817, loss_mean_cls=0.075, proj_loss=-0.41][2026-03-26 12:53:26] Step: 5393, Training Logs: loss_final: 0.485192, loss_mean: 0.828632, proj_loss: -0.417838, loss_mean_cls: 0.074399, grad_norm: 0.486346 +Steps: 1%| | 5394/1000000 [22:23<68:25:02, 4.04it/s, grad_norm=0.486, loss_final=0.485, loss_mean=0.829, loss_mean_cls=0.0744, proj_loss=-0.418][2026-03-26 12:53:26] Step: 5394, Training Logs: loss_final: 0.475235, loss_mean: 0.805798, proj_loss: -0.405155, loss_mean_cls: 0.074592, grad_norm: 0.434309 +Steps: 1%| | 5395/1000000 [22:23<68:24:45, 4.04it/s, grad_norm=0.434, loss_final=0.475, loss_mean=0.806, loss_mean_cls=0.0746, proj_loss=-0.405][2026-03-26 12:53:27] Step: 5395, Training Logs: loss_final: 0.479709, loss_mean: 0.820778, proj_loss: -0.414766, loss_mean_cls: 0.073697, grad_norm: 0.559859 +Steps: 1%| | 5396/1000000 [22:24<68:21:59, 4.04it/s, grad_norm=0.56, loss_final=0.48, loss_mean=0.821, loss_mean_cls=0.0737, proj_loss=-0.415][2026-03-26 12:53:27] Step: 5396, Training Logs: loss_final: 0.484922, loss_mean: 0.828767, proj_loss: -0.416322, loss_mean_cls: 0.072477, grad_norm: 0.345826 +Steps: 1%| | 5397/1000000 [22:24<68:23:33, 4.04it/s, grad_norm=0.346, loss_final=0.485, loss_mean=0.829, loss_mean_cls=0.0725, proj_loss=-0.416][2026-03-26 12:53:27] Step: 5397, Training Logs: loss_final: 0.474859, loss_mean: 0.826777, proj_loss: -0.423300, loss_mean_cls: 0.071382, grad_norm: 0.629598 +Steps: 1%| | 5398/1000000 [22:24<68:24:04, 4.04it/s, grad_norm=0.63, loss_final=0.475, loss_mean=0.827, loss_mean_cls=0.0714, proj_loss=-0.423][2026-03-26 12:53:27] Step: 5398, Training Logs: loss_final: 0.472210, loss_mean: 0.809906, proj_loss: -0.412238, loss_mean_cls: 0.074541, grad_norm: 0.414956 +Steps: 1%| | 5399/1000000 [22:24<68:24:34, 4.04it/s, grad_norm=0.415, loss_final=0.472, loss_mean=0.81, loss_mean_cls=0.0745, proj_loss=-0.412][2026-03-26 12:53:28] Step: 5399, Training Logs: loss_final: 0.493120, loss_mean: 0.827373, proj_loss: -0.408075, loss_mean_cls: 0.073823, grad_norm: 0.461748 +Steps: 1%| | 5400/1000000 [22:25<68:25:22, 4.04it/s, grad_norm=0.462, loss_final=0.493, loss_mean=0.827, loss_mean_cls=0.0738, proj_loss=-0.408][2026-03-26 12:53:28] Step: 5400, Training Logs: loss_final: 0.491302, loss_mean: 0.822409, proj_loss: -0.405990, loss_mean_cls: 0.074882, grad_norm: 0.564211 +Steps: 1%| | 5401/1000000 [22:25<68:26:10, 4.04it/s, grad_norm=0.564, loss_final=0.491, loss_mean=0.822, loss_mean_cls=0.0749, proj_loss=-0.406][2026-03-26 12:53:28] Step: 5401, Training Logs: loss_final: 0.488766, loss_mean: 0.812931, proj_loss: -0.400564, loss_mean_cls: 0.076400, grad_norm: 0.300009 +Steps: 1%| | 5402/1000000 [22:25<68:24:54, 4.04it/s, grad_norm=0.3, loss_final=0.489, loss_mean=0.813, loss_mean_cls=0.0764, proj_loss=-0.401][2026-03-26 12:53:28] Step: 5402, Training Logs: loss_final: 0.495056, loss_mean: 0.830165, proj_loss: -0.409592, loss_mean_cls: 0.074484, grad_norm: 0.406215 +Steps: 1%| | 5403/1000000 [22:25<68:25:58, 4.04it/s, grad_norm=0.406, loss_final=0.495, loss_mean=0.83, loss_mean_cls=0.0745, proj_loss=-0.41][2026-03-26 12:53:29] Step: 5403, Training Logs: loss_final: 0.478137, loss_mean: 0.822390, proj_loss: -0.417451, loss_mean_cls: 0.073197, grad_norm: 0.305461 +Steps: 1%| | 5404/1000000 [22:26<68:25:33, 4.04it/s, grad_norm=0.305, loss_final=0.478, loss_mean=0.822, loss_mean_cls=0.0732, proj_loss=-0.417][2026-03-26 12:53:29] Step: 5404, Training Logs: loss_final: 0.491504, loss_mean: 0.831473, proj_loss: -0.414180, loss_mean_cls: 0.074211, grad_norm: 0.302695 +Steps: 1%| | 5405/1000000 [22:26<68:26:09, 4.04it/s, grad_norm=0.303, loss_final=0.492, loss_mean=0.831, loss_mean_cls=0.0742, proj_loss=-0.414][2026-03-26 12:53:29] Step: 5405, Training Logs: loss_final: 0.485783, loss_mean: 0.830705, proj_loss: -0.417565, loss_mean_cls: 0.072643, grad_norm: 0.313412 +Steps: 1%| | 5406/1000000 [22:26<68:25:45, 4.04it/s, grad_norm=0.313, loss_final=0.486, loss_mean=0.831, loss_mean_cls=0.0726, proj_loss=-0.418][2026-03-26 12:53:29] Step: 5406, Training Logs: loss_final: 0.475547, loss_mean: 0.814354, proj_loss: -0.411867, loss_mean_cls: 0.073060, grad_norm: 0.341607 +Steps: 1%| | 5407/1000000 [22:26<68:26:09, 4.04it/s, grad_norm=0.342, loss_final=0.476, loss_mean=0.814, loss_mean_cls=0.0731, proj_loss=-0.412][2026-03-26 12:53:30] Step: 5407, Training Logs: loss_final: 0.480587, loss_mean: 0.816600, proj_loss: -0.410139, loss_mean_cls: 0.074126, grad_norm: 0.383986 +Steps: 1%| | 5408/1000000 [22:27<68:24:56, 4.04it/s, grad_norm=0.384, loss_final=0.481, loss_mean=0.817, loss_mean_cls=0.0741, proj_loss=-0.41][2026-03-26 12:53:30] Step: 5408, Training Logs: loss_final: 0.475920, loss_mean: 0.818583, proj_loss: -0.416290, loss_mean_cls: 0.073626, grad_norm: 0.317283 +Steps: 1%| | 5409/1000000 [22:27<68:23:38, 4.04it/s, grad_norm=0.317, loss_final=0.476, loss_mean=0.819, loss_mean_cls=0.0736, proj_loss=-0.416][2026-03-26 12:53:30] Step: 5409, Training Logs: loss_final: 0.489810, loss_mean: 0.813541, proj_loss: -0.399439, loss_mean_cls: 0.075708, grad_norm: 0.594139 +Steps: 1%| | 5410/1000000 [22:27<68:26:40, 4.04it/s, grad_norm=0.594, loss_final=0.49, loss_mean=0.814, loss_mean_cls=0.0757, proj_loss=-0.399][2026-03-26 12:53:30] Step: 5410, Training Logs: loss_final: 0.483815, loss_mean: 0.830798, proj_loss: -0.418313, loss_mean_cls: 0.071330, grad_norm: 0.376003 +Steps: 1%| | 5411/1000000 [22:27<68:24:17, 4.04it/s, grad_norm=0.376, loss_final=0.484, loss_mean=0.831, loss_mean_cls=0.0713, proj_loss=-0.418][2026-03-26 12:53:31] Step: 5411, Training Logs: loss_final: 0.479422, loss_mean: 0.820775, proj_loss: -0.415519, loss_mean_cls: 0.074165, grad_norm: 0.291407 +Steps: 1%| | 5412/1000000 [22:28<68:23:55, 4.04it/s, grad_norm=0.291, loss_final=0.479, loss_mean=0.821, loss_mean_cls=0.0742, proj_loss=-0.416][2026-03-26 12:53:31] Step: 5412, Training Logs: loss_final: 0.489130, loss_mean: 0.830035, proj_loss: -0.413496, loss_mean_cls: 0.072591, grad_norm: 0.369084 +Steps: 1%| | 5413/1000000 [22:28<68:24:04, 4.04it/s, grad_norm=0.369, loss_final=0.489, loss_mean=0.83, loss_mean_cls=0.0726, proj_loss=-0.413][2026-03-26 12:53:31] Step: 5413, Training Logs: loss_final: 0.482349, loss_mean: 0.827518, proj_loss: -0.418047, loss_mean_cls: 0.072878, grad_norm: 0.291737 +Steps: 1%| | 5414/1000000 [22:28<68:24:25, 4.04it/s, grad_norm=0.292, loss_final=0.482, loss_mean=0.828, loss_mean_cls=0.0729, proj_loss=-0.418][2026-03-26 12:53:31] Step: 5414, Training Logs: loss_final: 0.504338, loss_mean: 0.841525, proj_loss: -0.410094, loss_mean_cls: 0.072907, grad_norm: 0.709191 +Steps: 1%| | 5415/1000000 [22:28<68:23:05, 4.04it/s, grad_norm=0.709, loss_final=0.504, loss_mean=0.842, loss_mean_cls=0.0729, proj_loss=-0.41][2026-03-26 12:53:32] Step: 5415, Training Logs: loss_final: 0.511718, loss_mean: 0.854862, proj_loss: -0.414141, loss_mean_cls: 0.070997, grad_norm: 0.609549 +Steps: 1%| | 5416/1000000 [22:29<68:23:58, 4.04it/s, grad_norm=0.61, loss_final=0.512, loss_mean=0.855, loss_mean_cls=0.071, proj_loss=-0.414][2026-03-26 12:53:32] Step: 5416, Training Logs: loss_final: 0.488537, loss_mean: 0.820878, proj_loss: -0.406652, loss_mean_cls: 0.074311, grad_norm: 0.308907 +Steps: 1%| | 5417/1000000 [22:29<68:22:05, 4.04it/s, grad_norm=0.309, loss_final=0.489, loss_mean=0.821, loss_mean_cls=0.0743, proj_loss=-0.407][2026-03-26 12:53:32] Step: 5417, Training Logs: loss_final: 0.489388, loss_mean: 0.835180, proj_loss: -0.418476, loss_mean_cls: 0.072684, grad_norm: 0.413920 +Steps: 1%| | 5418/1000000 [22:29<68:28:20, 4.03it/s, grad_norm=0.414, loss_final=0.489, loss_mean=0.835, loss_mean_cls=0.0727, proj_loss=-0.418][2026-03-26 12:53:32] Step: 5418, Training Logs: loss_final: 0.498843, loss_mean: 0.840960, proj_loss: -0.414020, loss_mean_cls: 0.071903, grad_norm: 0.428061 +Steps: 1%| | 5419/1000000 [22:29<68:24:43, 4.04it/s, grad_norm=0.428, loss_final=0.499, loss_mean=0.841, loss_mean_cls=0.0719, proj_loss=-0.414][2026-03-26 12:53:33] Step: 5419, Training Logs: loss_final: 0.506575, loss_mean: 0.856490, proj_loss: -0.420283, loss_mean_cls: 0.070369, grad_norm: 0.331840 +Steps: 1%| | 5420/1000000 [22:30<68:24:22, 4.04it/s, grad_norm=0.332, loss_final=0.507, loss_mean=0.856, loss_mean_cls=0.0704, proj_loss=-0.42][2026-03-26 12:53:33] Step: 5420, Training Logs: loss_final: 0.499453, loss_mean: 0.839380, proj_loss: -0.411942, loss_mean_cls: 0.072015, grad_norm: 0.403447 +Steps: 1%| | 5421/1000000 [22:30<68:20:38, 4.04it/s, grad_norm=0.403, loss_final=0.499, loss_mean=0.839, loss_mean_cls=0.072, proj_loss=-0.412][2026-03-26 12:53:33] Step: 5421, Training Logs: loss_final: 0.469410, loss_mean: 0.809687, proj_loss: -0.414374, loss_mean_cls: 0.074097, grad_norm: 0.263533 +Steps: 1%| | 5422/1000000 [22:30<68:21:13, 4.04it/s, grad_norm=0.264, loss_final=0.469, loss_mean=0.81, loss_mean_cls=0.0741, proj_loss=-0.414][2026-03-26 12:53:33] Step: 5422, Training Logs: loss_final: 0.467484, loss_mean: 0.803473, proj_loss: -0.411196, loss_mean_cls: 0.075206, grad_norm: 0.290681 +Steps: 1%| | 5423/1000000 [22:30<68:19:59, 4.04it/s, grad_norm=0.291, loss_final=0.467, loss_mean=0.803, loss_mean_cls=0.0752, proj_loss=-0.411][2026-03-26 12:53:34] Step: 5423, Training Logs: loss_final: 0.482708, loss_mean: 0.826398, proj_loss: -0.416611, loss_mean_cls: 0.072921, grad_norm: 0.438543 +Steps: 1%| | 5424/1000000 [22:31<68:20:37, 4.04it/s, grad_norm=0.439, loss_final=0.483, loss_mean=0.826, loss_mean_cls=0.0729, proj_loss=-0.417][2026-03-26 12:53:34] Step: 5424, Training Logs: loss_final: 0.473091, loss_mean: 0.825004, proj_loss: -0.422643, loss_mean_cls: 0.070730, grad_norm: 0.332679 +Steps: 1%| | 5425/1000000 [22:31<68:21:13, 4.04it/s, grad_norm=0.333, loss_final=0.473, loss_mean=0.825, loss_mean_cls=0.0707, proj_loss=-0.423][2026-03-26 12:53:34] Step: 5425, Training Logs: loss_final: 0.488100, loss_mean: 0.832690, proj_loss: -0.416558, loss_mean_cls: 0.071968, grad_norm: 0.419512 +Steps: 1%| | 5426/1000000 [22:31<68:20:16, 4.04it/s, grad_norm=0.42, loss_final=0.488, loss_mean=0.833, loss_mean_cls=0.072, proj_loss=-0.417][2026-03-26 12:53:34] Step: 5426, Training Logs: loss_final: 0.496287, loss_mean: 0.839290, proj_loss: -0.415670, loss_mean_cls: 0.072667, grad_norm: 0.306995 +Steps: 1%| | 5427/1000000 [22:31<68:20:47, 4.04it/s, grad_norm=0.307, loss_final=0.496, loss_mean=0.839, loss_mean_cls=0.0727, proj_loss=-0.416][2026-03-26 12:53:35] Step: 5427, Training Logs: loss_final: 0.480621, loss_mean: 0.803921, proj_loss: -0.400885, loss_mean_cls: 0.077585, grad_norm: 0.333133 +Steps: 1%| | 5428/1000000 [22:32<68:23:08, 4.04it/s, grad_norm=0.333, loss_final=0.481, loss_mean=0.804, loss_mean_cls=0.0776, proj_loss=-0.401][2026-03-26 12:53:35] Step: 5428, Training Logs: loss_final: 0.495407, loss_mean: 0.838964, proj_loss: -0.415421, loss_mean_cls: 0.071864, grad_norm: 0.441583 +Steps: 1%| | 5429/1000000 [22:32<68:23:45, 4.04it/s, grad_norm=0.442, loss_final=0.495, loss_mean=0.839, loss_mean_cls=0.0719, proj_loss=-0.415][2026-03-26 12:53:35] Step: 5429, Training Logs: loss_final: 0.469383, loss_mean: 0.801577, proj_loss: -0.407795, loss_mean_cls: 0.075600, grad_norm: 0.425376 +Steps: 1%| | 5430/1000000 [22:32<68:22:52, 4.04it/s, grad_norm=0.425, loss_final=0.469, loss_mean=0.802, loss_mean_cls=0.0756, proj_loss=-0.408][2026-03-26 12:53:35] Step: 5430, Training Logs: loss_final: 0.492187, loss_mean: 0.843885, proj_loss: -0.424137, loss_mean_cls: 0.072440, grad_norm: 0.307789 +Steps: 1%| | 5431/1000000 [22:32<68:22:54, 4.04it/s, grad_norm=0.308, loss_final=0.492, loss_mean=0.844, loss_mean_cls=0.0724, proj_loss=-0.424][2026-03-26 12:53:36] Step: 5431, Training Logs: loss_final: 0.487813, loss_mean: 0.824205, proj_loss: -0.410537, loss_mean_cls: 0.074145, grad_norm: 0.466605 +Steps: 1%| | 5432/1000000 [22:33<68:22:32, 4.04it/s, grad_norm=0.467, loss_final=0.488, loss_mean=0.824, loss_mean_cls=0.0741, proj_loss=-0.411][2026-03-26 12:53:36] Step: 5432, Training Logs: loss_final: 0.493824, loss_mean: 0.848322, proj_loss: -0.424439, loss_mean_cls: 0.069941, grad_norm: 0.365662 +Steps: 1%| | 5433/1000000 [22:33<68:25:58, 4.04it/s, grad_norm=0.366, loss_final=0.494, loss_mean=0.848, loss_mean_cls=0.0699, proj_loss=-0.424][2026-03-26 12:53:36] Step: 5433, Training Logs: loss_final: 0.486999, loss_mean: 0.836704, proj_loss: -0.421695, loss_mean_cls: 0.071990, grad_norm: 0.260398 +Steps: 1%| | 5434/1000000 [22:33<68:24:00, 4.04it/s, grad_norm=0.26, loss_final=0.487, loss_mean=0.837, loss_mean_cls=0.072, proj_loss=-0.422][2026-03-26 12:53:36] Step: 5434, Training Logs: loss_final: 0.479881, loss_mean: 0.817697, proj_loss: -0.411321, loss_mean_cls: 0.073505, grad_norm: 0.382756 +Steps: 1%| | 5435/1000000 [22:33<68:24:56, 4.04it/s, grad_norm=0.383, loss_final=0.48, loss_mean=0.818, loss_mean_cls=0.0735, proj_loss=-0.411][2026-03-26 12:53:37] Step: 5435, Training Logs: loss_final: 0.471172, loss_mean: 0.825224, proj_loss: -0.424453, loss_mean_cls: 0.070400, grad_norm: 0.528227 +Steps: 1%| | 5436/1000000 [22:33<68:22:48, 4.04it/s, grad_norm=0.528, loss_final=0.471, loss_mean=0.825, loss_mean_cls=0.0704, proj_loss=-0.424][2026-03-26 12:53:37] Step: 5436, Training Logs: loss_final: 0.508341, loss_mean: 0.862379, proj_loss: -0.424107, loss_mean_cls: 0.070069, grad_norm: 0.229238 +Steps: 1%| | 5437/1000000 [22:34<68:22:14, 4.04it/s, grad_norm=0.229, loss_final=0.508, loss_mean=0.862, loss_mean_cls=0.0701, proj_loss=-0.424][2026-03-26 12:53:37] Step: 5437, Training Logs: loss_final: 0.495675, loss_mean: 0.837516, proj_loss: -0.414264, loss_mean_cls: 0.072423, grad_norm: 0.411614 +Steps: 1%| | 5438/1000000 [22:34<68:24:20, 4.04it/s, grad_norm=0.412, loss_final=0.496, loss_mean=0.838, loss_mean_cls=0.0724, proj_loss=-0.414][2026-03-26 12:53:37] Step: 5438, Training Logs: loss_final: 0.485778, loss_mean: 0.820210, proj_loss: -0.409221, loss_mean_cls: 0.074789, grad_norm: 0.481675 +Steps: 1%| | 5439/1000000 [22:34<68:27:47, 4.04it/s, grad_norm=0.482, loss_final=0.486, loss_mean=0.82, loss_mean_cls=0.0748, proj_loss=-0.409][2026-03-26 12:53:38] Step: 5439, Training Logs: loss_final: 0.471395, loss_mean: 0.809083, proj_loss: -0.412289, loss_mean_cls: 0.074602, grad_norm: 0.347239 +Steps: 1%| | 5440/1000000 [22:34<68:25:18, 4.04it/s, grad_norm=0.347, loss_final=0.471, loss_mean=0.809, loss_mean_cls=0.0746, proj_loss=-0.412][2026-03-26 12:53:38] Step: 5440, Training Logs: loss_final: 0.489019, loss_mean: 0.829959, proj_loss: -0.414231, loss_mean_cls: 0.073291, grad_norm: 0.387981 +Steps: 1%| | 5441/1000000 [22:35<68:25:35, 4.04it/s, grad_norm=0.388, loss_final=0.489, loss_mean=0.83, loss_mean_cls=0.0733, proj_loss=-0.414][2026-03-26 12:53:38] Step: 5441, Training Logs: loss_final: 0.476854, loss_mean: 0.818488, proj_loss: -0.415336, loss_mean_cls: 0.073701, grad_norm: 0.537483 +Steps: 1%| | 5442/1000000 [22:35<68:27:45, 4.04it/s, grad_norm=0.537, loss_final=0.477, loss_mean=0.818, loss_mean_cls=0.0737, proj_loss=-0.415][2026-03-26 12:53:38] Step: 5442, Training Logs: loss_final: 0.484214, loss_mean: 0.823765, proj_loss: -0.413473, loss_mean_cls: 0.073922, grad_norm: 0.293383 +Steps: 1%| | 5443/1000000 [22:36<101:26:36, 2.72it/s, grad_norm=0.293, loss_final=0.484, loss_mean=0.824, loss_mean_cls=0.0739, proj_loss=-0.413][2026-03-26 12:53:39] Step: 5443, Training Logs: loss_final: 0.484156, loss_mean: 0.827005, proj_loss: -0.415388, loss_mean_cls: 0.072538, grad_norm: 0.336753 +Steps: 1%| | 5444/1000000 [22:36<91:31:51, 3.02it/s, grad_norm=0.337, loss_final=0.484, loss_mean=0.827, loss_mean_cls=0.0725, proj_loss=-0.415] [2026-03-26 12:53:39] Step: 5444, Training Logs: loss_final: 0.487855, loss_mean: 0.826054, proj_loss: -0.411357, loss_mean_cls: 0.073158, grad_norm: 0.510850 +Steps: 1%| | 5445/1000000 [22:36<85:16:26, 3.24it/s, grad_norm=0.511, loss_final=0.488, loss_mean=0.826, loss_mean_cls=0.0732, proj_loss=-0.411][2026-03-26 12:53:40] Step: 5445, Training Logs: loss_final: 0.500837, loss_mean: 0.838608, proj_loss: -0.411207, loss_mean_cls: 0.073437, grad_norm: 0.409736 +Steps: 1%| | 5446/1000000 [22:36<80:12:57, 3.44it/s, grad_norm=0.41, loss_final=0.501, loss_mean=0.839, loss_mean_cls=0.0734, proj_loss=-0.411][2026-03-26 12:53:40] Step: 5446, Training Logs: loss_final: 0.487825, loss_mean: 0.816480, proj_loss: -0.404800, loss_mean_cls: 0.076145, grad_norm: 0.426522 +Steps: 1%| | 5447/1000000 [22:37<76:39:38, 3.60it/s, grad_norm=0.427, loss_final=0.488, loss_mean=0.816, loss_mean_cls=0.0761, proj_loss=-0.405][2026-03-26 12:53:40] Step: 5447, Training Logs: loss_final: 0.480343, loss_mean: 0.818384, proj_loss: -0.411315, loss_mean_cls: 0.073274, grad_norm: 0.498946 +Steps: 1%| | 5448/1000000 [22:37<74:10:37, 3.72it/s, grad_norm=0.499, loss_final=0.48, loss_mean=0.818, loss_mean_cls=0.0733, proj_loss=-0.411][2026-03-26 12:53:40] Step: 5448, Training Logs: loss_final: 0.477921, loss_mean: 0.821323, proj_loss: -0.415967, loss_mean_cls: 0.072564, grad_norm: 0.387213 +Steps: 1%| | 5449/1000000 [22:37<72:28:24, 3.81it/s, grad_norm=0.387, loss_final=0.478, loss_mean=0.821, loss_mean_cls=0.0726, proj_loss=-0.416][2026-03-26 12:53:41] Step: 5449, Training Logs: loss_final: 0.488781, loss_mean: 0.816425, proj_loss: -0.402910, loss_mean_cls: 0.075265, grad_norm: 0.367904 +Steps: 1%| | 5450/1000000 [22:37<71:15:00, 3.88it/s, grad_norm=0.368, loss_final=0.489, loss_mean=0.816, loss_mean_cls=0.0753, proj_loss=-0.403][2026-03-26 12:53:41] Step: 5450, Training Logs: loss_final: 0.478466, loss_mean: 0.815789, proj_loss: -0.412334, loss_mean_cls: 0.075011, grad_norm: 0.535471 +Steps: 1%| | 5451/1000000 [22:38<70:21:15, 3.93it/s, grad_norm=0.535, loss_final=0.478, loss_mean=0.816, loss_mean_cls=0.075, proj_loss=-0.412][2026-03-26 12:53:41] Step: 5451, Training Logs: loss_final: 0.508222, loss_mean: 0.838894, proj_loss: -0.404542, loss_mean_cls: 0.073869, grad_norm: 0.544776 +Steps: 1%| | 5452/1000000 [22:38<69:44:49, 3.96it/s, grad_norm=0.545, loss_final=0.508, loss_mean=0.839, loss_mean_cls=0.0739, proj_loss=-0.405][2026-03-26 12:53:41] Step: 5452, Training Logs: loss_final: 0.470003, loss_mean: 0.801976, proj_loss: -0.407334, loss_mean_cls: 0.075361, grad_norm: 0.339995 +Steps: 1%| | 5453/1000000 [22:38<69:16:56, 3.99it/s, grad_norm=0.34, loss_final=0.47, loss_mean=0.802, loss_mean_cls=0.0754, proj_loss=-0.407][2026-03-26 12:53:42] Step: 5453, Training Logs: loss_final: 0.493256, loss_mean: 0.836907, proj_loss: -0.415257, loss_mean_cls: 0.071607, grad_norm: 0.494043 +Steps: 1%| | 5454/1000000 [22:38<69:04:42, 4.00it/s, grad_norm=0.494, loss_final=0.493, loss_mean=0.837, loss_mean_cls=0.0716, proj_loss=-0.415][2026-03-26 12:53:42] Step: 5454, Training Logs: loss_final: 0.478075, loss_mean: 0.813941, proj_loss: -0.410313, loss_mean_cls: 0.074447, grad_norm: 0.422396 +Steps: 1%| | 5455/1000000 [22:39<68:48:48, 4.01it/s, grad_norm=0.422, loss_final=0.478, loss_mean=0.814, loss_mean_cls=0.0744, proj_loss=-0.41][2026-03-26 12:53:42] Step: 5455, Training Logs: loss_final: 0.469702, loss_mean: 0.813371, proj_loss: -0.415541, loss_mean_cls: 0.071872, grad_norm: 0.324988 +Steps: 1%| | 5456/1000000 [22:39<68:42:28, 4.02it/s, grad_norm=0.325, loss_final=0.47, loss_mean=0.813, loss_mean_cls=0.0719, proj_loss=-0.416][2026-03-26 12:53:42] Step: 5456, Training Logs: loss_final: 0.484890, loss_mean: 0.821535, proj_loss: -0.409846, loss_mean_cls: 0.073201, grad_norm: 0.409463 +Steps: 1%| | 5457/1000000 [22:39<68:31:46, 4.03it/s, grad_norm=0.409, loss_final=0.485, loss_mean=0.822, loss_mean_cls=0.0732, proj_loss=-0.41][2026-03-26 12:53:42] Step: 5457, Training Logs: loss_final: 0.479961, loss_mean: 0.820831, proj_loss: -0.413858, loss_mean_cls: 0.072988, grad_norm: 0.275599 +Steps: 1%| | 5458/1000000 [22:39<68:27:21, 4.04it/s, grad_norm=0.276, loss_final=0.48, loss_mean=0.821, loss_mean_cls=0.073, proj_loss=-0.414][2026-03-26 12:53:43] Step: 5458, Training Logs: loss_final: 0.483735, loss_mean: 0.819297, proj_loss: -0.408832, loss_mean_cls: 0.073270, grad_norm: 0.595284 +Steps: 1%| | 5459/1000000 [22:40<68:22:20, 4.04it/s, grad_norm=0.595, loss_final=0.484, loss_mean=0.819, loss_mean_cls=0.0733, proj_loss=-0.409][2026-03-26 12:53:43] Step: 5459, Training Logs: loss_final: 0.485863, loss_mean: 0.822087, proj_loss: -0.409695, loss_mean_cls: 0.073470, grad_norm: 0.527180 +Steps: 1%| | 5460/1000000 [22:40<68:22:27, 4.04it/s, grad_norm=0.527, loss_final=0.486, loss_mean=0.822, loss_mean_cls=0.0735, proj_loss=-0.41][2026-03-26 12:53:43] Step: 5460, Training Logs: loss_final: 0.468492, loss_mean: 0.809633, proj_loss: -0.414451, loss_mean_cls: 0.073310, grad_norm: 0.368145 +Steps: 1%| | 5461/1000000 [22:40<68:21:51, 4.04it/s, grad_norm=0.368, loss_final=0.468, loss_mean=0.81, loss_mean_cls=0.0733, proj_loss=-0.414][2026-03-26 12:53:43] Step: 5461, Training Logs: loss_final: 0.485146, loss_mean: 0.829198, proj_loss: -0.417693, loss_mean_cls: 0.073641, grad_norm: 0.578718 +Steps: 1%| | 5462/1000000 [22:40<68:24:45, 4.04it/s, grad_norm=0.579, loss_final=0.485, loss_mean=0.829, loss_mean_cls=0.0736, proj_loss=-0.418][2026-03-26 12:53:44] Step: 5462, Training Logs: loss_final: 0.460356, loss_mean: 0.809798, proj_loss: -0.420745, loss_mean_cls: 0.071303, grad_norm: 0.353042 +Steps: 1%| | 5463/1000000 [22:41<68:21:59, 4.04it/s, grad_norm=0.353, loss_final=0.46, loss_mean=0.81, loss_mean_cls=0.0713, proj_loss=-0.421][2026-03-26 12:53:44] Step: 5463, Training Logs: loss_final: 0.470165, loss_mean: 0.814726, proj_loss: -0.416760, loss_mean_cls: 0.072199, grad_norm: 0.453721 +Steps: 1%| | 5464/1000000 [22:41<68:21:49, 4.04it/s, grad_norm=0.454, loss_final=0.47, loss_mean=0.815, loss_mean_cls=0.0722, proj_loss=-0.417][2026-03-26 12:53:44] Step: 5464, Training Logs: loss_final: 0.484923, loss_mean: 0.808724, proj_loss: -0.399767, loss_mean_cls: 0.075967, grad_norm: 0.440338 +Steps: 1%| | 5465/1000000 [22:41<68:21:05, 4.04it/s, grad_norm=0.44, loss_final=0.485, loss_mean=0.809, loss_mean_cls=0.076, proj_loss=-0.4][2026-03-26 12:53:44] Step: 5465, Training Logs: loss_final: 0.485251, loss_mean: 0.822719, proj_loss: -0.410903, loss_mean_cls: 0.073434, grad_norm: 0.471270 +Steps: 1%| | 5466/1000000 [22:41<68:22:10, 4.04it/s, grad_norm=0.471, loss_final=0.485, loss_mean=0.823, loss_mean_cls=0.0734, proj_loss=-0.411][2026-03-26 12:53:45] Step: 5466, Training Logs: loss_final: 0.467892, loss_mean: 0.807121, proj_loss: -0.413582, loss_mean_cls: 0.074353, grad_norm: 0.656077 +Steps: 1%| | 5467/1000000 [22:42<68:20:53, 4.04it/s, grad_norm=0.656, loss_final=0.468, loss_mean=0.807, loss_mean_cls=0.0744, proj_loss=-0.414][2026-03-26 12:53:45] Step: 5467, Training Logs: loss_final: 0.480161, loss_mean: 0.827436, proj_loss: -0.419596, loss_mean_cls: 0.072321, grad_norm: 0.407040 +Steps: 1%| | 5468/1000000 [22:42<68:20:48, 4.04it/s, grad_norm=0.407, loss_final=0.48, loss_mean=0.827, loss_mean_cls=0.0723, proj_loss=-0.42][2026-03-26 12:53:45] Step: 5468, Training Logs: loss_final: 0.475206, loss_mean: 0.814127, proj_loss: -0.412297, loss_mean_cls: 0.073376, grad_norm: 0.433755 +Steps: 1%| | 5469/1000000 [22:42<68:20:04, 4.04it/s, grad_norm=0.434, loss_final=0.475, loss_mean=0.814, loss_mean_cls=0.0734, proj_loss=-0.412][2026-03-26 12:53:45] Step: 5469, Training Logs: loss_final: 0.502721, loss_mean: 0.848248, proj_loss: -0.416737, loss_mean_cls: 0.071210, grad_norm: 0.550965 +Steps: 1%| | 5470/1000000 [22:42<68:19:53, 4.04it/s, grad_norm=0.551, loss_final=0.503, loss_mean=0.848, loss_mean_cls=0.0712, proj_loss=-0.417][2026-03-26 12:53:46] Step: 5470, Training Logs: loss_final: 0.492553, loss_mean: 0.831269, proj_loss: -0.411828, loss_mean_cls: 0.073111, grad_norm: 0.394662 +Steps: 1%| | 5471/1000000 [22:43<68:21:00, 4.04it/s, grad_norm=0.395, loss_final=0.493, loss_mean=0.831, loss_mean_cls=0.0731, proj_loss=-0.412][2026-03-26 12:53:46] Step: 5471, Training Logs: loss_final: 0.480339, loss_mean: 0.818250, proj_loss: -0.410917, loss_mean_cls: 0.073005, grad_norm: 0.506705 +Steps: 1%| | 5472/1000000 [22:43<68:20:31, 4.04it/s, grad_norm=0.507, loss_final=0.48, loss_mean=0.818, loss_mean_cls=0.073, proj_loss=-0.411][2026-03-26 12:53:46] Step: 5472, Training Logs: loss_final: 0.496824, loss_mean: 0.836519, proj_loss: -0.412734, loss_mean_cls: 0.073039, grad_norm: 0.609124 +Steps: 1%| | 5473/1000000 [22:43<68:26:24, 4.04it/s, grad_norm=0.609, loss_final=0.497, loss_mean=0.837, loss_mean_cls=0.073, proj_loss=-0.413][2026-03-26 12:53:46] Step: 5473, Training Logs: loss_final: 0.493350, loss_mean: 0.837377, proj_loss: -0.416612, loss_mean_cls: 0.072584, grad_norm: 0.306543 +Steps: 1%| | 5474/1000000 [22:43<68:28:00, 4.03it/s, grad_norm=0.307, loss_final=0.493, loss_mean=0.837, loss_mean_cls=0.0726, proj_loss=-0.417][2026-03-26 12:53:47] Step: 5474, Training Logs: loss_final: 0.484725, loss_mean: 0.834190, proj_loss: -0.420061, loss_mean_cls: 0.070596, grad_norm: 0.578144 +Steps: 1%| | 5475/1000000 [22:44<68:29:12, 4.03it/s, grad_norm=0.578, loss_final=0.485, loss_mean=0.834, loss_mean_cls=0.0706, proj_loss=-0.42][2026-03-26 12:53:47] Step: 5475, Training Logs: loss_final: 0.505009, loss_mean: 0.843496, proj_loss: -0.411284, loss_mean_cls: 0.072797, grad_norm: 0.504449 +Steps: 1%| | 5476/1000000 [22:44<68:26:14, 4.04it/s, grad_norm=0.504, loss_final=0.505, loss_mean=0.843, loss_mean_cls=0.0728, proj_loss=-0.411][2026-03-26 12:53:47] Step: 5476, Training Logs: loss_final: 0.472723, loss_mean: 0.813597, proj_loss: -0.413835, loss_mean_cls: 0.072960, grad_norm: 0.356036 +Steps: 1%| | 5477/1000000 [22:44<68:29:20, 4.03it/s, grad_norm=0.356, loss_final=0.473, loss_mean=0.814, loss_mean_cls=0.073, proj_loss=-0.414][2026-03-26 12:53:47] Step: 5477, Training Logs: loss_final: 0.487566, loss_mean: 0.827863, proj_loss: -0.413577, loss_mean_cls: 0.073279, grad_norm: 0.657100 +Steps: 1%| | 5478/1000000 [22:44<68:25:45, 4.04it/s, grad_norm=0.657, loss_final=0.488, loss_mean=0.828, loss_mean_cls=0.0733, proj_loss=-0.414][2026-03-26 12:53:48] Step: 5478, Training Logs: loss_final: 0.496601, loss_mean: 0.839803, proj_loss: -0.415719, loss_mean_cls: 0.072517, grad_norm: 0.491462 +Steps: 1%| | 5479/1000000 [22:45<68:24:19, 4.04it/s, grad_norm=0.491, loss_final=0.497, loss_mean=0.84, loss_mean_cls=0.0725, proj_loss=-0.416][2026-03-26 12:53:48] Step: 5479, Training Logs: loss_final: 0.477088, loss_mean: 0.806333, proj_loss: -0.404940, loss_mean_cls: 0.075695, grad_norm: 0.401992 +Steps: 1%| | 5480/1000000 [22:45<68:22:36, 4.04it/s, grad_norm=0.402, loss_final=0.477, loss_mean=0.806, loss_mean_cls=0.0757, proj_loss=-0.405][2026-03-26 12:53:48] Step: 5480, Training Logs: loss_final: 0.483834, loss_mean: 0.830332, proj_loss: -0.418540, loss_mean_cls: 0.072042, grad_norm: 0.322593 +Steps: 1%| | 5481/1000000 [22:45<68:22:01, 4.04it/s, grad_norm=0.323, loss_final=0.484, loss_mean=0.83, loss_mean_cls=0.072, proj_loss=-0.419][2026-03-26 12:53:48] Step: 5481, Training Logs: loss_final: 0.484399, loss_mean: 0.830720, proj_loss: -0.418497, loss_mean_cls: 0.072176, grad_norm: 0.560365 +Steps: 1%| | 5482/1000000 [22:45<68:22:55, 4.04it/s, grad_norm=0.56, loss_final=0.484, loss_mean=0.831, loss_mean_cls=0.0722, proj_loss=-0.418][2026-03-26 12:53:49] Step: 5482, Training Logs: loss_final: 0.475808, loss_mean: 0.813694, proj_loss: -0.411009, loss_mean_cls: 0.073123, grad_norm: 0.696938 +Steps: 1%| | 5483/1000000 [22:46<68:22:33, 4.04it/s, grad_norm=0.697, loss_final=0.476, loss_mean=0.814, loss_mean_cls=0.0731, proj_loss=-0.411][2026-03-26 12:53:49] Step: 5483, Training Logs: loss_final: 0.465377, loss_mean: 0.808408, proj_loss: -0.415773, loss_mean_cls: 0.072742, grad_norm: 0.317354 +Steps: 1%| | 5484/1000000 [22:46<68:22:33, 4.04it/s, grad_norm=0.317, loss_final=0.465, loss_mean=0.808, loss_mean_cls=0.0727, proj_loss=-0.416][2026-03-26 12:53:49] Step: 5484, Training Logs: loss_final: 0.487277, loss_mean: 0.826510, proj_loss: -0.412411, loss_mean_cls: 0.073178, grad_norm: 0.556455 +Steps: 1%| | 5485/1000000 [22:46<68:23:39, 4.04it/s, grad_norm=0.556, loss_final=0.487, loss_mean=0.827, loss_mean_cls=0.0732, proj_loss=-0.412][2026-03-26 12:53:49] Step: 5485, Training Logs: loss_final: 0.498189, loss_mean: 0.843168, proj_loss: -0.417112, loss_mean_cls: 0.072133, grad_norm: 0.326322 +Steps: 1%| | 5486/1000000 [22:46<68:26:21, 4.04it/s, grad_norm=0.326, loss_final=0.498, loss_mean=0.843, loss_mean_cls=0.0721, proj_loss=-0.417][2026-03-26 12:53:50] Step: 5486, Training Logs: loss_final: 0.491390, loss_mean: 0.834330, proj_loss: -0.415816, loss_mean_cls: 0.072876, grad_norm: 0.461411 +Steps: 1%| | 5487/1000000 [22:47<68:35:36, 4.03it/s, grad_norm=0.461, loss_final=0.491, loss_mean=0.834, loss_mean_cls=0.0729, proj_loss=-0.416][2026-03-26 12:53:50] Step: 5487, Training Logs: loss_final: 0.481011, loss_mean: 0.820734, proj_loss: -0.411969, loss_mean_cls: 0.072246, grad_norm: 0.487605 +Steps: 1%| | 5488/1000000 [22:47<68:33:12, 4.03it/s, grad_norm=0.488, loss_final=0.481, loss_mean=0.821, loss_mean_cls=0.0722, proj_loss=-0.412][2026-03-26 12:53:50] Step: 5488, Training Logs: loss_final: 0.499374, loss_mean: 0.833509, proj_loss: -0.407433, loss_mean_cls: 0.073298, grad_norm: 0.331937 +Steps: 1%| | 5489/1000000 [22:47<68:30:18, 4.03it/s, grad_norm=0.332, loss_final=0.499, loss_mean=0.834, loss_mean_cls=0.0733, proj_loss=-0.407][2026-03-26 12:53:50] Step: 5489, Training Logs: loss_final: 0.476425, loss_mean: 0.810511, proj_loss: -0.408795, loss_mean_cls: 0.074709, grad_norm: 0.767037 +Steps: 1%| | 5490/1000000 [22:47<68:29:49, 4.03it/s, grad_norm=0.767, loss_final=0.476, loss_mean=0.811, loss_mean_cls=0.0747, proj_loss=-0.409][2026-03-26 12:53:51] Step: 5490, Training Logs: loss_final: 0.469712, loss_mean: 0.811311, proj_loss: -0.413673, loss_mean_cls: 0.072073, grad_norm: 0.318295 +Steps: 1%| | 5491/1000000 [22:48<68:27:47, 4.04it/s, grad_norm=0.318, loss_final=0.47, loss_mean=0.811, loss_mean_cls=0.0721, proj_loss=-0.414][2026-03-26 12:53:51] Step: 5491, Training Logs: loss_final: 0.489034, loss_mean: 0.840580, proj_loss: -0.421377, loss_mean_cls: 0.069831, grad_norm: 0.639228 +Steps: 1%| | 5492/1000000 [22:48<68:29:32, 4.03it/s, grad_norm=0.639, loss_final=0.489, loss_mean=0.841, loss_mean_cls=0.0698, proj_loss=-0.421][2026-03-26 12:53:51] Step: 5492, Training Logs: loss_final: 0.478669, loss_mean: 0.821932, proj_loss: -0.415702, loss_mean_cls: 0.072438, grad_norm: 0.477510 +Steps: 1%| | 5493/1000000 [22:48<68:26:15, 4.04it/s, grad_norm=0.478, loss_final=0.479, loss_mean=0.822, loss_mean_cls=0.0724, proj_loss=-0.416][2026-03-26 12:53:51] Step: 5493, Training Logs: loss_final: 0.491285, loss_mean: 0.823445, proj_loss: -0.407449, loss_mean_cls: 0.075290, grad_norm: 0.463635 +Steps: 1%| | 5494/1000000 [22:48<68:27:30, 4.04it/s, grad_norm=0.464, loss_final=0.491, loss_mean=0.823, loss_mean_cls=0.0753, proj_loss=-0.407][2026-03-26 12:53:52] Step: 5494, Training Logs: loss_final: 0.482308, loss_mean: 0.822093, proj_loss: -0.413202, loss_mean_cls: 0.073418, grad_norm: 0.532247 +Steps: 1%| | 5495/1000000 [22:49<68:28:30, 4.03it/s, grad_norm=0.532, loss_final=0.482, loss_mean=0.822, loss_mean_cls=0.0734, proj_loss=-0.413][2026-03-26 12:53:52] Step: 5495, Training Logs: loss_final: 0.491547, loss_mean: 0.828780, proj_loss: -0.411458, loss_mean_cls: 0.074224, grad_norm: 0.321996 +Steps: 1%| | 5496/1000000 [22:49<69:50:46, 3.96it/s, grad_norm=0.322, loss_final=0.492, loss_mean=0.829, loss_mean_cls=0.0742, proj_loss=-0.411][2026-03-26 12:53:52] Step: 5496, Training Logs: loss_final: 0.500790, loss_mean: 0.835242, proj_loss: -0.408322, loss_mean_cls: 0.073870, grad_norm: 0.452129 +Steps: 1%| | 5497/1000000 [22:49<69:24:22, 3.98it/s, grad_norm=0.452, loss_final=0.501, loss_mean=0.835, loss_mean_cls=0.0739, proj_loss=-0.408][2026-03-26 12:53:52] Step: 5497, Training Logs: loss_final: 0.475923, loss_mean: 0.816526, proj_loss: -0.413792, loss_mean_cls: 0.073189, grad_norm: 0.301786 +Steps: 1%| | 5498/1000000 [22:49<69:06:03, 4.00it/s, grad_norm=0.302, loss_final=0.476, loss_mean=0.817, loss_mean_cls=0.0732, proj_loss=-0.414][2026-03-26 12:53:53] Step: 5498, Training Logs: loss_final: 0.479306, loss_mean: 0.817408, proj_loss: -0.412532, loss_mean_cls: 0.074431, grad_norm: 0.337063 +Steps: 1%| | 5499/1000000 [22:50<68:52:43, 4.01it/s, grad_norm=0.337, loss_final=0.479, loss_mean=0.817, loss_mean_cls=0.0744, proj_loss=-0.413][2026-03-26 12:53:53] Step: 5499, Training Logs: loss_final: 0.479240, loss_mean: 0.826472, proj_loss: -0.419082, loss_mean_cls: 0.071849, grad_norm: 0.306720 +Steps: 1%| | 5500/1000000 [22:50<68:44:47, 4.02it/s, grad_norm=0.307, loss_final=0.479, loss_mean=0.826, loss_mean_cls=0.0718, proj_loss=-0.419][2026-03-26 12:53:53] Step: 5500, Training Logs: loss_final: 0.483821, loss_mean: 0.821007, proj_loss: -0.410863, loss_mean_cls: 0.073676, grad_norm: 0.308341 +Steps: 1%| | 5501/1000000 [22:50<68:35:36, 4.03it/s, grad_norm=0.308, loss_final=0.484, loss_mean=0.821, loss_mean_cls=0.0737, proj_loss=-0.411][2026-03-26 12:53:53] Step: 5501, Training Logs: loss_final: 0.495166, loss_mean: 0.837203, proj_loss: -0.414634, loss_mean_cls: 0.072597, grad_norm: 0.456061 +Steps: 1%| | 5502/1000000 [22:50<68:30:14, 4.03it/s, grad_norm=0.456, loss_final=0.495, loss_mean=0.837, loss_mean_cls=0.0726, proj_loss=-0.415][2026-03-26 12:53:54] Step: 5502, Training Logs: loss_final: 0.492909, loss_mean: 0.831210, proj_loss: -0.411068, loss_mean_cls: 0.072767, grad_norm: 0.719747 +Steps: 1%| | 5503/1000000 [22:51<68:24:15, 4.04it/s, grad_norm=0.72, loss_final=0.493, loss_mean=0.831, loss_mean_cls=0.0728, proj_loss=-0.411][2026-03-26 12:53:54] Step: 5503, Training Logs: loss_final: 0.471095, loss_mean: 0.825423, proj_loss: -0.424057, loss_mean_cls: 0.069729, grad_norm: 0.322117 +Steps: 1%| | 5504/1000000 [22:51<68:22:30, 4.04it/s, grad_norm=0.322, loss_final=0.471, loss_mean=0.825, loss_mean_cls=0.0697, proj_loss=-0.424][2026-03-26 12:53:54] Step: 5504, Training Logs: loss_final: 0.471480, loss_mean: 0.820005, proj_loss: -0.419037, loss_mean_cls: 0.070511, grad_norm: 0.349503 +Steps: 1%| | 5505/1000000 [22:51<68:17:37, 4.05it/s, grad_norm=0.35, loss_final=0.471, loss_mean=0.82, loss_mean_cls=0.0705, proj_loss=-0.419][2026-03-26 12:53:54] Step: 5505, Training Logs: loss_final: 0.493931, loss_mean: 0.829878, proj_loss: -0.410236, loss_mean_cls: 0.074289, grad_norm: 0.355812 +Steps: 1%| | 5506/1000000 [22:51<68:23:50, 4.04it/s, grad_norm=0.356, loss_final=0.494, loss_mean=0.83, loss_mean_cls=0.0743, proj_loss=-0.41][2026-03-26 12:53:55] Step: 5506, Training Logs: loss_final: 0.503168, loss_mean: 0.849631, proj_loss: -0.417813, loss_mean_cls: 0.071350, grad_norm: 0.295678 +Steps: 1%| | 5507/1000000 [22:51<68:23:14, 4.04it/s, grad_norm=0.296, loss_final=0.503, loss_mean=0.85, loss_mean_cls=0.0713, proj_loss=-0.418][2026-03-26 12:53:55] Step: 5507, Training Logs: loss_final: 0.464384, loss_mean: 0.805459, proj_loss: -0.413866, loss_mean_cls: 0.072791, grad_norm: 0.522471 +Steps: 1%| | 5508/1000000 [22:52<68:25:39, 4.04it/s, grad_norm=0.522, loss_final=0.464, loss_mean=0.805, loss_mean_cls=0.0728, proj_loss=-0.414][2026-03-26 12:53:55] Step: 5508, Training Logs: loss_final: 0.480475, loss_mean: 0.812358, proj_loss: -0.406302, loss_mean_cls: 0.074419, grad_norm: 0.358659 +Steps: 1%| | 5509/1000000 [22:52<68:22:44, 4.04it/s, grad_norm=0.359, loss_final=0.48, loss_mean=0.812, loss_mean_cls=0.0744, proj_loss=-0.406][2026-03-26 12:53:55] Step: 5509, Training Logs: loss_final: 0.477645, loss_mean: 0.825658, proj_loss: -0.419725, loss_mean_cls: 0.071712, grad_norm: 0.605002 +Steps: 1%| | 5510/1000000 [22:52<68:22:38, 4.04it/s, grad_norm=0.605, loss_final=0.478, loss_mean=0.826, loss_mean_cls=0.0717, proj_loss=-0.42][2026-03-26 12:53:56] Step: 5510, Training Logs: loss_final: 0.490086, loss_mean: 0.827720, proj_loss: -0.410826, loss_mean_cls: 0.073193, grad_norm: 0.688249 +Steps: 1%| | 5511/1000000 [22:52<68:20:29, 4.04it/s, grad_norm=0.688, loss_final=0.49, loss_mean=0.828, loss_mean_cls=0.0732, proj_loss=-0.411][2026-03-26 12:53:56] Step: 5511, Training Logs: loss_final: 0.486033, loss_mean: 0.816122, proj_loss: -0.404770, loss_mean_cls: 0.074682, grad_norm: 0.316594 +Steps: 1%| | 5512/1000000 [22:53<68:21:49, 4.04it/s, grad_norm=0.317, loss_final=0.486, loss_mean=0.816, loss_mean_cls=0.0747, proj_loss=-0.405][2026-03-26 12:53:56] Step: 5512, Training Logs: loss_final: 0.488256, loss_mean: 0.836944, proj_loss: -0.419507, loss_mean_cls: 0.070819, grad_norm: 0.760750 +Steps: 1%| | 5513/1000000 [22:53<68:21:51, 4.04it/s, grad_norm=0.761, loss_final=0.488, loss_mean=0.837, loss_mean_cls=0.0708, proj_loss=-0.42][2026-03-26 12:53:56] Step: 5513, Training Logs: loss_final: 0.485537, loss_mean: 0.823448, proj_loss: -0.410674, loss_mean_cls: 0.072763, grad_norm: 0.779486 +Steps: 1%| | 5514/1000000 [22:53<68:22:37, 4.04it/s, grad_norm=0.779, loss_final=0.486, loss_mean=0.823, loss_mean_cls=0.0728, proj_loss=-0.411][2026-03-26 12:53:57] Step: 5514, Training Logs: loss_final: 0.492800, loss_mean: 0.846806, proj_loss: -0.423462, loss_mean_cls: 0.069457, grad_norm: 0.330397 +Steps: 1%| | 5515/1000000 [22:53<68:24:58, 4.04it/s, grad_norm=0.33, loss_final=0.493, loss_mean=0.847, loss_mean_cls=0.0695, proj_loss=-0.423][2026-03-26 12:53:57] Step: 5515, Training Logs: loss_final: 0.482856, loss_mean: 0.816476, proj_loss: -0.407093, loss_mean_cls: 0.073473, grad_norm: 0.545178 +Steps: 1%| | 5516/1000000 [22:54<68:22:48, 4.04it/s, grad_norm=0.545, loss_final=0.483, loss_mean=0.816, loss_mean_cls=0.0735, proj_loss=-0.407][2026-03-26 12:53:57] Step: 5516, Training Logs: loss_final: 0.473314, loss_mean: 0.807920, proj_loss: -0.408450, loss_mean_cls: 0.073844, grad_norm: 0.299745 +Steps: 1%| | 5517/1000000 [22:54<68:21:07, 4.04it/s, grad_norm=0.3, loss_final=0.473, loss_mean=0.808, loss_mean_cls=0.0738, proj_loss=-0.408][2026-03-26 12:53:57] Step: 5517, Training Logs: loss_final: 0.496036, loss_mean: 0.844847, proj_loss: -0.418913, loss_mean_cls: 0.070102, grad_norm: 0.338166 +Steps: 1%| | 5518/1000000 [22:54<68:22:22, 4.04it/s, grad_norm=0.338, loss_final=0.496, loss_mean=0.845, loss_mean_cls=0.0701, proj_loss=-0.419][2026-03-26 12:53:58] Step: 5518, Training Logs: loss_final: 0.489462, loss_mean: 0.818516, proj_loss: -0.403910, loss_mean_cls: 0.074855, grad_norm: 0.394794 +Steps: 1%| | 5519/1000000 [22:54<68:22:38, 4.04it/s, grad_norm=0.395, loss_final=0.489, loss_mean=0.819, loss_mean_cls=0.0749, proj_loss=-0.404][2026-03-26 12:53:58] Step: 5519, Training Logs: loss_final: 0.487067, loss_mean: 0.840731, proj_loss: -0.423678, loss_mean_cls: 0.070013, grad_norm: 0.297095 +Steps: 1%| | 5520/1000000 [22:55<68:22:50, 4.04it/s, grad_norm=0.297, loss_final=0.487, loss_mean=0.841, loss_mean_cls=0.07, proj_loss=-0.424][2026-03-26 12:53:58] Step: 5520, Training Logs: loss_final: 0.487966, loss_mean: 0.824160, proj_loss: -0.409662, loss_mean_cls: 0.073468, grad_norm: 0.466886 +Steps: 1%| | 5521/1000000 [22:55<68:24:03, 4.04it/s, grad_norm=0.467, loss_final=0.488, loss_mean=0.824, loss_mean_cls=0.0735, proj_loss=-0.41][2026-03-26 12:53:58] Step: 5521, Training Logs: loss_final: 0.481122, loss_mean: 0.823294, proj_loss: -0.413753, loss_mean_cls: 0.071581, grad_norm: 0.309239 +Steps: 1%| | 5522/1000000 [22:55<68:22:53, 4.04it/s, grad_norm=0.309, loss_final=0.481, loss_mean=0.823, loss_mean_cls=0.0716, proj_loss=-0.414][2026-03-26 12:53:59] Step: 5522, Training Logs: loss_final: 0.488337, loss_mean: 0.823033, proj_loss: -0.407772, loss_mean_cls: 0.073076, grad_norm: 0.342273 +Steps: 1%| | 5523/1000000 [22:55<68:22:02, 4.04it/s, grad_norm=0.342, loss_final=0.488, loss_mean=0.823, loss_mean_cls=0.0731, proj_loss=-0.408][2026-03-26 12:53:59] Step: 5523, Training Logs: loss_final: 0.505454, loss_mean: 0.838867, proj_loss: -0.407249, loss_mean_cls: 0.073836, grad_norm: 0.342636 +Steps: 1%| | 5524/1000000 [22:56<68:20:48, 4.04it/s, grad_norm=0.343, loss_final=0.505, loss_mean=0.839, loss_mean_cls=0.0738, proj_loss=-0.407][2026-03-26 12:53:59] Step: 5524, Training Logs: loss_final: 0.486959, loss_mean: 0.831702, proj_loss: -0.416613, loss_mean_cls: 0.071870, grad_norm: 0.272726 +Steps: 1%| | 5525/1000000 [22:56<68:20:57, 4.04it/s, grad_norm=0.273, loss_final=0.487, loss_mean=0.832, loss_mean_cls=0.0719, proj_loss=-0.417][2026-03-26 12:53:59] Step: 5525, Training Logs: loss_final: 0.467426, loss_mean: 0.798735, proj_loss: -0.406554, loss_mean_cls: 0.075245, grad_norm: 0.516882 +Steps: 1%| | 5526/1000000 [22:56<68:19:15, 4.04it/s, grad_norm=0.517, loss_final=0.467, loss_mean=0.799, loss_mean_cls=0.0752, proj_loss=-0.407][2026-03-26 12:54:00] Step: 5526, Training Logs: loss_final: 0.498845, loss_mean: 0.842716, proj_loss: -0.414481, loss_mean_cls: 0.070610, grad_norm: 0.328278 +Steps: 1%| | 5527/1000000 [22:56<68:20:54, 4.04it/s, grad_norm=0.328, loss_final=0.499, loss_mean=0.843, loss_mean_cls=0.0706, proj_loss=-0.414][2026-03-26 12:54:00] Step: 5527, Training Logs: loss_final: 0.496366, loss_mean: 0.841880, proj_loss: -0.416113, loss_mean_cls: 0.070599, grad_norm: 0.269632 +Steps: 1%| | 5528/1000000 [22:57<68:18:52, 4.04it/s, grad_norm=0.27, loss_final=0.496, loss_mean=0.842, loss_mean_cls=0.0706, proj_loss=-0.416][2026-03-26 12:54:00] Step: 5528, Training Logs: loss_final: 0.489079, loss_mean: 0.830796, proj_loss: -0.414342, loss_mean_cls: 0.072624, grad_norm: 0.358863 +Steps: 1%| | 5529/1000000 [22:57<68:20:13, 4.04it/s, grad_norm=0.359, loss_final=0.489, loss_mean=0.831, loss_mean_cls=0.0726, proj_loss=-0.414][2026-03-26 12:54:00] Step: 5529, Training Logs: loss_final: 0.504850, loss_mean: 0.847684, proj_loss: -0.414443, loss_mean_cls: 0.071608, grad_norm: 0.351840 +Steps: 1%| | 5530/1000000 [22:57<68:22:57, 4.04it/s, grad_norm=0.352, loss_final=0.505, loss_mean=0.848, loss_mean_cls=0.0716, proj_loss=-0.414][2026-03-26 12:54:01] Step: 5530, Training Logs: loss_final: 0.476112, loss_mean: 0.825790, proj_loss: -0.420636, loss_mean_cls: 0.070958, grad_norm: 0.271819 +Steps: 1%| | 5531/1000000 [22:57<68:23:19, 4.04it/s, grad_norm=0.272, loss_final=0.476, loss_mean=0.826, loss_mean_cls=0.071, proj_loss=-0.421][2026-03-26 12:54:01] Step: 5531, Training Logs: loss_final: 0.476794, loss_mean: 0.811443, proj_loss: -0.409035, loss_mean_cls: 0.074386, grad_norm: 0.472185 +Steps: 1%| | 5532/1000000 [22:58<68:24:32, 4.04it/s, grad_norm=0.472, loss_final=0.477, loss_mean=0.811, loss_mean_cls=0.0744, proj_loss=-0.409][2026-03-26 12:54:01] Step: 5532, Training Logs: loss_final: 0.478085, loss_mean: 0.803336, proj_loss: -0.400515, loss_mean_cls: 0.075264, grad_norm: 0.422782 +Steps: 1%| | 5533/1000000 [22:58<68:23:26, 4.04it/s, grad_norm=0.423, loss_final=0.478, loss_mean=0.803, loss_mean_cls=0.0753, proj_loss=-0.401][2026-03-26 12:54:01] Step: 5533, Training Logs: loss_final: 0.491721, loss_mean: 0.825169, proj_loss: -0.407834, loss_mean_cls: 0.074386, grad_norm: 0.405453 +Steps: 1%| | 5534/1000000 [22:58<68:23:53, 4.04it/s, grad_norm=0.405, loss_final=0.492, loss_mean=0.825, loss_mean_cls=0.0744, proj_loss=-0.408][2026-03-26 12:54:02] Step: 5534, Training Logs: loss_final: 0.476410, loss_mean: 0.818536, proj_loss: -0.414867, loss_mean_cls: 0.072740, grad_norm: 0.723591 +Steps: 1%| | 5535/1000000 [22:58<68:59:49, 4.00it/s, grad_norm=0.724, loss_final=0.476, loss_mean=0.819, loss_mean_cls=0.0727, proj_loss=-0.415][2026-03-26 12:54:02] Step: 5535, Training Logs: loss_final: 0.491923, loss_mean: 0.842783, proj_loss: -0.420147, loss_mean_cls: 0.069287, grad_norm: 0.414324 +Steps: 1%| | 5536/1000000 [22:59<68:49:32, 4.01it/s, grad_norm=0.414, loss_final=0.492, loss_mean=0.843, loss_mean_cls=0.0693, proj_loss=-0.42][2026-03-26 12:54:02] Step: 5536, Training Logs: loss_final: 0.475848, loss_mean: 0.809655, proj_loss: -0.407551, loss_mean_cls: 0.073744, grad_norm: 0.582899 +Steps: 1%| | 5537/1000000 [22:59<68:43:21, 4.02it/s, grad_norm=0.583, loss_final=0.476, loss_mean=0.81, loss_mean_cls=0.0737, proj_loss=-0.408][2026-03-26 12:54:02] Step: 5537, Training Logs: loss_final: 0.484103, loss_mean: 0.831598, proj_loss: -0.417399, loss_mean_cls: 0.069904, grad_norm: 0.526881 +Steps: 1%| | 5538/1000000 [22:59<68:37:51, 4.02it/s, grad_norm=0.527, loss_final=0.484, loss_mean=0.832, loss_mean_cls=0.0699, proj_loss=-0.417][2026-03-26 12:54:03] Step: 5538, Training Logs: loss_final: 0.486530, loss_mean: 0.832919, proj_loss: -0.418265, loss_mean_cls: 0.071876, grad_norm: 0.306765 +Steps: 1%| | 5539/1000000 [22:59<68:34:31, 4.03it/s, grad_norm=0.307, loss_final=0.487, loss_mean=0.833, loss_mean_cls=0.0719, proj_loss=-0.418][2026-03-26 12:54:03] Step: 5539, Training Logs: loss_final: 0.470031, loss_mean: 0.818420, proj_loss: -0.419506, loss_mean_cls: 0.071117, grad_norm: 0.435920 +Steps: 1%| | 5540/1000000 [23:00<68:32:52, 4.03it/s, grad_norm=0.436, loss_final=0.47, loss_mean=0.818, loss_mean_cls=0.0711, proj_loss=-0.42][2026-03-26 12:54:03] Step: 5540, Training Logs: loss_final: 0.485888, loss_mean: 0.826226, proj_loss: -0.412749, loss_mean_cls: 0.072410, grad_norm: 0.457522 +Steps: 1%| | 5541/1000000 [23:00<68:28:49, 4.03it/s, grad_norm=0.458, loss_final=0.486, loss_mean=0.826, loss_mean_cls=0.0724, proj_loss=-0.413][2026-03-26 12:54:03] Step: 5541, Training Logs: loss_final: 0.500130, loss_mean: 0.837032, proj_loss: -0.409133, loss_mean_cls: 0.072231, grad_norm: 0.314067 +Steps: 1%| | 5542/1000000 [23:00<69:09:31, 3.99it/s, grad_norm=0.314, loss_final=0.5, loss_mean=0.837, loss_mean_cls=0.0722, proj_loss=-0.409][2026-03-26 12:54:04] Step: 5542, Training Logs: loss_final: 0.495033, loss_mean: 0.833962, proj_loss: -0.411004, loss_mean_cls: 0.072075, grad_norm: 0.469382 +Steps: 1%| | 5543/1000000 [23:00<69:20:25, 3.98it/s, grad_norm=0.469, loss_final=0.495, loss_mean=0.834, loss_mean_cls=0.0721, proj_loss=-0.411][2026-03-26 12:54:04] Step: 5543, Training Logs: loss_final: 0.477085, loss_mean: 0.823012, proj_loss: -0.417185, loss_mean_cls: 0.071259, grad_norm: 0.328258 +Steps: 1%| | 5544/1000000 [23:01<69:04:42, 4.00it/s, grad_norm=0.328, loss_final=0.477, loss_mean=0.823, loss_mean_cls=0.0713, proj_loss=-0.417][2026-03-26 12:54:04] Step: 5544, Training Logs: loss_final: 0.472094, loss_mean: 0.812401, proj_loss: -0.412603, loss_mean_cls: 0.072297, grad_norm: 0.503940 +Steps: 1%| | 5545/1000000 [23:01<68:53:33, 4.01it/s, grad_norm=0.504, loss_final=0.472, loss_mean=0.812, loss_mean_cls=0.0723, proj_loss=-0.413][2026-03-26 12:54:04] Step: 5545, Training Logs: loss_final: 0.483031, loss_mean: 0.821640, proj_loss: -0.411861, loss_mean_cls: 0.073252, grad_norm: 0.393893 +Steps: 1%| | 5546/1000000 [23:01<68:45:52, 4.02it/s, grad_norm=0.394, loss_final=0.483, loss_mean=0.822, loss_mean_cls=0.0733, proj_loss=-0.412][2026-03-26 12:54:05] Step: 5546, Training Logs: loss_final: 0.470991, loss_mean: 0.806123, proj_loss: -0.408389, loss_mean_cls: 0.073257, grad_norm: 0.268305 +Steps: 1%| | 5547/1000000 [23:01<68:37:26, 4.03it/s, grad_norm=0.268, loss_final=0.471, loss_mean=0.806, loss_mean_cls=0.0733, proj_loss=-0.408][2026-03-26 12:54:05] Step: 5547, Training Logs: loss_final: 0.471507, loss_mean: 0.813904, proj_loss: -0.415739, loss_mean_cls: 0.073342, grad_norm: 0.499553 +Steps: 1%| | 5548/1000000 [23:02<68:33:02, 4.03it/s, grad_norm=0.5, loss_final=0.472, loss_mean=0.814, loss_mean_cls=0.0733, proj_loss=-0.416][2026-03-26 12:54:05] Step: 5548, Training Logs: loss_final: 0.470617, loss_mean: 0.811663, proj_loss: -0.413741, loss_mean_cls: 0.072695, grad_norm: 0.287439 +Steps: 1%| | 5549/1000000 [23:02<68:26:09, 4.04it/s, grad_norm=0.287, loss_final=0.471, loss_mean=0.812, loss_mean_cls=0.0727, proj_loss=-0.414][2026-03-26 12:54:05] Step: 5549, Training Logs: loss_final: 0.487386, loss_mean: 0.819952, proj_loss: -0.405564, loss_mean_cls: 0.072998, grad_norm: 0.451791 +Steps: 1%| | 5550/1000000 [23:02<68:24:24, 4.04it/s, grad_norm=0.452, loss_final=0.487, loss_mean=0.82, loss_mean_cls=0.073, proj_loss=-0.406][2026-03-26 12:54:06] Step: 5550, Training Logs: loss_final: 0.491470, loss_mean: 0.829405, proj_loss: -0.411025, loss_mean_cls: 0.073090, grad_norm: 0.515206 +Steps: 1%| | 5551/1000000 [23:02<68:20:33, 4.04it/s, grad_norm=0.515, loss_final=0.491, loss_mean=0.829, loss_mean_cls=0.0731, proj_loss=-0.411][2026-03-26 12:54:06] Step: 5551, Training Logs: loss_final: 0.490324, loss_mean: 0.837231, proj_loss: -0.417963, loss_mean_cls: 0.071056, grad_norm: 0.268093 +Steps: 1%| | 5552/1000000 [23:03<68:21:49, 4.04it/s, grad_norm=0.268, loss_final=0.49, loss_mean=0.837, loss_mean_cls=0.0711, proj_loss=-0.418][2026-03-26 12:54:06] Step: 5552, Training Logs: loss_final: 0.469624, loss_mean: 0.807458, proj_loss: -0.409987, loss_mean_cls: 0.072153, grad_norm: 0.422992 +Steps: 1%| | 5553/1000000 [23:03<68:17:30, 4.04it/s, grad_norm=0.423, loss_final=0.47, loss_mean=0.807, loss_mean_cls=0.0722, proj_loss=-0.41][2026-03-26 12:54:06] Step: 5553, Training Logs: loss_final: 0.485371, loss_mean: 0.828934, proj_loss: -0.415974, loss_mean_cls: 0.072411, grad_norm: 0.315783 +Steps: 1%| | 5554/1000000 [23:03<68:23:28, 4.04it/s, grad_norm=0.316, loss_final=0.485, loss_mean=0.829, loss_mean_cls=0.0724, proj_loss=-0.416][2026-03-26 12:54:07] Step: 5554, Training Logs: loss_final: 0.485423, loss_mean: 0.826365, proj_loss: -0.414371, loss_mean_cls: 0.073429, grad_norm: 0.362384 +Steps: 1%| | 5555/1000000 [23:03<68:19:52, 4.04it/s, grad_norm=0.362, loss_final=0.485, loss_mean=0.826, loss_mean_cls=0.0734, proj_loss=-0.414][2026-03-26 12:54:07] Step: 5555, Training Logs: loss_final: 0.488179, loss_mean: 0.831631, proj_loss: -0.413618, loss_mean_cls: 0.070166, grad_norm: 0.409289 +Steps: 1%| | 5556/1000000 [23:04<68:20:17, 4.04it/s, grad_norm=0.409, loss_final=0.488, loss_mean=0.832, loss_mean_cls=0.0702, proj_loss=-0.414][2026-03-26 12:54:07] Step: 5556, Training Logs: loss_final: 0.467038, loss_mean: 0.812012, proj_loss: -0.416202, loss_mean_cls: 0.071228, grad_norm: 0.381373 +Steps: 1%| | 5557/1000000 [23:04<68:19:09, 4.04it/s, grad_norm=0.381, loss_final=0.467, loss_mean=0.812, loss_mean_cls=0.0712, proj_loss=-0.416][2026-03-26 12:54:07] Step: 5557, Training Logs: loss_final: 0.480748, loss_mean: 0.826635, proj_loss: -0.417679, loss_mean_cls: 0.071792, grad_norm: 0.388153 +Steps: 1%| | 5558/1000000 [23:04<68:22:19, 4.04it/s, grad_norm=0.388, loss_final=0.481, loss_mean=0.827, loss_mean_cls=0.0718, proj_loss=-0.418][2026-03-26 12:54:08] Step: 5558, Training Logs: loss_final: 0.485807, loss_mean: 0.834296, proj_loss: -0.418163, loss_mean_cls: 0.069674, grad_norm: 0.846831 +Steps: 1%| | 5559/1000000 [23:04<68:23:25, 4.04it/s, grad_norm=0.847, loss_final=0.486, loss_mean=0.834, loss_mean_cls=0.0697, proj_loss=-0.418][2026-03-26 12:54:08] Step: 5559, Training Logs: loss_final: 0.486977, loss_mean: 0.833557, proj_loss: -0.418429, loss_mean_cls: 0.071849, grad_norm: 0.310487 +Steps: 1%| | 5560/1000000 [23:05<68:23:37, 4.04it/s, grad_norm=0.31, loss_final=0.487, loss_mean=0.834, loss_mean_cls=0.0718, proj_loss=-0.418][2026-03-26 12:54:08] Step: 5560, Training Logs: loss_final: 0.482797, loss_mean: 0.825289, proj_loss: -0.414443, loss_mean_cls: 0.071950, grad_norm: 0.686325 +Steps: 1%| | 5561/1000000 [23:05<68:20:30, 4.04it/s, grad_norm=0.686, loss_final=0.483, loss_mean=0.825, loss_mean_cls=0.072, proj_loss=-0.414][2026-03-26 12:54:08] Step: 5561, Training Logs: loss_final: 0.475261, loss_mean: 0.809433, proj_loss: -0.407841, loss_mean_cls: 0.073669, grad_norm: 0.558880 +Steps: 1%| | 5562/1000000 [23:05<68:19:15, 4.04it/s, grad_norm=0.559, loss_final=0.475, loss_mean=0.809, loss_mean_cls=0.0737, proj_loss=-0.408][2026-03-26 12:54:09] Step: 5562, Training Logs: loss_final: 0.468216, loss_mean: 0.804183, proj_loss: -0.408924, loss_mean_cls: 0.072957, grad_norm: 0.336187 +Steps: 1%| | 5563/1000000 [23:05<68:20:15, 4.04it/s, grad_norm=0.336, loss_final=0.468, loss_mean=0.804, loss_mean_cls=0.073, proj_loss=-0.409][2026-03-26 12:54:09] Step: 5563, Training Logs: loss_final: 0.491220, loss_mean: 0.838653, proj_loss: -0.417785, loss_mean_cls: 0.070353, grad_norm: 0.682054 +Steps: 1%| | 5564/1000000 [23:06<68:18:42, 4.04it/s, grad_norm=0.682, loss_final=0.491, loss_mean=0.839, loss_mean_cls=0.0704, proj_loss=-0.418][2026-03-26 12:54:09] Step: 5564, Training Logs: loss_final: 0.479211, loss_mean: 0.837454, proj_loss: -0.427209, loss_mean_cls: 0.068967, grad_norm: 0.330029 +Steps: 1%| | 5565/1000000 [23:06<68:18:45, 4.04it/s, grad_norm=0.33, loss_final=0.479, loss_mean=0.837, loss_mean_cls=0.069, proj_loss=-0.427][2026-03-26 12:54:09] Step: 5565, Training Logs: loss_final: 0.485009, loss_mean: 0.821557, proj_loss: -0.409615, loss_mean_cls: 0.073067, grad_norm: 0.377500 +Steps: 1%| | 5566/1000000 [23:06<68:20:39, 4.04it/s, grad_norm=0.378, loss_final=0.485, loss_mean=0.822, loss_mean_cls=0.0731, proj_loss=-0.41][2026-03-26 12:54:10] Step: 5566, Training Logs: loss_final: 0.485991, loss_mean: 0.827402, proj_loss: -0.413565, loss_mean_cls: 0.072154, grad_norm: 0.389222 +Steps: 1%| | 5567/1000000 [23:06<68:21:33, 4.04it/s, grad_norm=0.389, loss_final=0.486, loss_mean=0.827, loss_mean_cls=0.0722, proj_loss=-0.414][2026-03-26 12:54:10] Step: 5567, Training Logs: loss_final: 0.475802, loss_mean: 0.822728, proj_loss: -0.419284, loss_mean_cls: 0.072358, grad_norm: 0.307608 +Steps: 1%| | 5568/1000000 [23:07<68:19:30, 4.04it/s, grad_norm=0.308, loss_final=0.476, loss_mean=0.823, loss_mean_cls=0.0724, proj_loss=-0.419][2026-03-26 12:54:10] Step: 5568, Training Logs: loss_final: 0.489618, loss_mean: 0.835974, proj_loss: -0.418197, loss_mean_cls: 0.071841, grad_norm: 0.349141 +Steps: 1%| | 5569/1000000 [23:07<68:20:00, 4.04it/s, grad_norm=0.349, loss_final=0.49, loss_mean=0.836, loss_mean_cls=0.0718, proj_loss=-0.418][2026-03-26 12:54:10] Step: 5569, Training Logs: loss_final: 0.491323, loss_mean: 0.830092, proj_loss: -0.410574, loss_mean_cls: 0.071804, grad_norm: 0.285330 +Steps: 1%| | 5570/1000000 [23:07<68:21:52, 4.04it/s, grad_norm=0.285, loss_final=0.491, loss_mean=0.83, loss_mean_cls=0.0718, proj_loss=-0.411][2026-03-26 12:54:10] Step: 5570, Training Logs: loss_final: 0.477714, loss_mean: 0.829556, proj_loss: -0.421058, loss_mean_cls: 0.069215, grad_norm: 0.281688 +Steps: 1%| | 5571/1000000 [23:07<68:24:26, 4.04it/s, grad_norm=0.282, loss_final=0.478, loss_mean=0.83, loss_mean_cls=0.0692, proj_loss=-0.421][2026-03-26 12:54:11] Step: 5571, Training Logs: loss_final: 0.490915, loss_mean: 0.826718, proj_loss: -0.408391, loss_mean_cls: 0.072588, grad_norm: 0.284347 +Steps: 1%| | 5572/1000000 [23:08<68:22:07, 4.04it/s, grad_norm=0.284, loss_final=0.491, loss_mean=0.827, loss_mean_cls=0.0726, proj_loss=-0.408][2026-03-26 12:54:11] Step: 5572, Training Logs: loss_final: 0.485200, loss_mean: 0.824770, proj_loss: -0.412323, loss_mean_cls: 0.072753, grad_norm: 0.239982 +Steps: 1%| | 5573/1000000 [23:08<68:24:02, 4.04it/s, grad_norm=0.24, loss_final=0.485, loss_mean=0.825, loss_mean_cls=0.0728, proj_loss=-0.412][2026-03-26 12:54:11] Step: 5573, Training Logs: loss_final: 0.477474, loss_mean: 0.830357, proj_loss: -0.422759, loss_mean_cls: 0.069876, grad_norm: 0.252096 +Steps: 1%| | 5574/1000000 [23:08<68:21:13, 4.04it/s, grad_norm=0.252, loss_final=0.477, loss_mean=0.83, loss_mean_cls=0.0699, proj_loss=-0.423][2026-03-26 12:54:11] Step: 5574, Training Logs: loss_final: 0.480258, loss_mean: 0.822233, proj_loss: -0.414091, loss_mean_cls: 0.072116, grad_norm: 0.294535 +Steps: 1%| | 5575/1000000 [23:08<68:22:10, 4.04it/s, grad_norm=0.295, loss_final=0.48, loss_mean=0.822, loss_mean_cls=0.0721, proj_loss=-0.414][2026-03-26 12:54:12] Step: 5575, Training Logs: loss_final: 0.487102, loss_mean: 0.830968, proj_loss: -0.415597, loss_mean_cls: 0.071731, grad_norm: 0.393709 +Steps: 1%| | 5576/1000000 [23:09<68:21:35, 4.04it/s, grad_norm=0.394, loss_final=0.487, loss_mean=0.831, loss_mean_cls=0.0717, proj_loss=-0.416][2026-03-26 12:54:12] Step: 5576, Training Logs: loss_final: 0.480928, loss_mean: 0.827485, proj_loss: -0.417360, loss_mean_cls: 0.070804, grad_norm: 0.466145 +Steps: 1%| | 5577/1000000 [23:09<68:22:21, 4.04it/s, grad_norm=0.466, loss_final=0.481, loss_mean=0.827, loss_mean_cls=0.0708, proj_loss=-0.417][2026-03-26 12:54:12] Step: 5577, Training Logs: loss_final: 0.479585, loss_mean: 0.831342, proj_loss: -0.422060, loss_mean_cls: 0.070303, grad_norm: 0.250671 +Steps: 1%| | 5578/1000000 [23:09<68:18:20, 4.04it/s, grad_norm=0.251, loss_final=0.48, loss_mean=0.831, loss_mean_cls=0.0703, proj_loss=-0.422][2026-03-26 12:54:12] Step: 5578, Training Logs: loss_final: 0.471284, loss_mean: 0.815952, proj_loss: -0.416299, loss_mean_cls: 0.071631, grad_norm: 0.297881 +Steps: 1%| | 5579/1000000 [23:09<68:16:54, 4.05it/s, grad_norm=0.298, loss_final=0.471, loss_mean=0.816, loss_mean_cls=0.0716, proj_loss=-0.416][2026-03-26 12:54:13] Step: 5579, Training Logs: loss_final: 0.507275, loss_mean: 0.847382, proj_loss: -0.411693, loss_mean_cls: 0.071586, grad_norm: 0.393622 +Steps: 1%| | 5580/1000000 [23:10<68:14:43, 4.05it/s, grad_norm=0.394, loss_final=0.507, loss_mean=0.847, loss_mean_cls=0.0716, proj_loss=-0.412][2026-03-26 12:54:13] Step: 5580, Training Logs: loss_final: 0.482120, loss_mean: 0.828647, proj_loss: -0.416993, loss_mean_cls: 0.070466, grad_norm: 0.261206 +Steps: 1%| | 5581/1000000 [23:10<68:17:54, 4.04it/s, grad_norm=0.261, loss_final=0.482, loss_mean=0.829, loss_mean_cls=0.0705, proj_loss=-0.417][2026-03-26 12:54:13] Step: 5581, Training Logs: loss_final: 0.492871, loss_mean: 0.834816, proj_loss: -0.413659, loss_mean_cls: 0.071714, grad_norm: 0.244951 +Steps: 1%| | 5582/1000000 [23:10<68:19:33, 4.04it/s, grad_norm=0.245, loss_final=0.493, loss_mean=0.835, loss_mean_cls=0.0717, proj_loss=-0.414][2026-03-26 12:54:13] Step: 5582, Training Logs: loss_final: 0.487270, loss_mean: 0.824767, proj_loss: -0.410017, loss_mean_cls: 0.072520, grad_norm: 0.292229 +Steps: 1%| | 5583/1000000 [23:10<68:18:24, 4.04it/s, grad_norm=0.292, loss_final=0.487, loss_mean=0.825, loss_mean_cls=0.0725, proj_loss=-0.41][2026-03-26 12:54:14] Step: 5583, Training Logs: loss_final: 0.488304, loss_mean: 0.846225, proj_loss: -0.427194, loss_mean_cls: 0.069274, grad_norm: 0.343356 +Steps: 1%| | 5584/1000000 [23:11<68:18:19, 4.04it/s, grad_norm=0.343, loss_final=0.488, loss_mean=0.846, loss_mean_cls=0.0693, proj_loss=-0.427][2026-03-26 12:54:14] Step: 5584, Training Logs: loss_final: 0.484505, loss_mean: 0.821603, proj_loss: -0.409721, loss_mean_cls: 0.072624, grad_norm: 0.269951 +Steps: 1%| | 5585/1000000 [23:11<68:17:52, 4.04it/s, grad_norm=0.27, loss_final=0.485, loss_mean=0.822, loss_mean_cls=0.0726, proj_loss=-0.41][2026-03-26 12:54:14] Step: 5585, Training Logs: loss_final: 0.477354, loss_mean: 0.815077, proj_loss: -0.411568, loss_mean_cls: 0.073844, grad_norm: 0.272538 +Steps: 1%| | 5586/1000000 [23:11<68:19:35, 4.04it/s, grad_norm=0.273, loss_final=0.477, loss_mean=0.815, loss_mean_cls=0.0738, proj_loss=-0.412][2026-03-26 12:54:14] Step: 5586, Training Logs: loss_final: 0.487264, loss_mean: 0.840760, proj_loss: -0.423079, loss_mean_cls: 0.069583, grad_norm: 0.348563 +Steps: 1%| | 5587/1000000 [23:11<68:20:15, 4.04it/s, grad_norm=0.349, loss_final=0.487, loss_mean=0.841, loss_mean_cls=0.0696, proj_loss=-0.423][2026-03-26 12:54:15] Step: 5587, Training Logs: loss_final: 0.471434, loss_mean: 0.820148, proj_loss: -0.418964, loss_mean_cls: 0.070250, grad_norm: 0.298083 +Steps: 1%| | 5588/1000000 [23:12<68:19:18, 4.04it/s, grad_norm=0.298, loss_final=0.471, loss_mean=0.82, loss_mean_cls=0.0703, proj_loss=-0.419][2026-03-26 12:54:15] Step: 5588, Training Logs: loss_final: 0.492792, loss_mean: 0.849367, proj_loss: -0.425270, loss_mean_cls: 0.068696, grad_norm: 0.292645 +Steps: 1%| | 5589/1000000 [23:12<68:18:52, 4.04it/s, grad_norm=0.293, loss_final=0.493, loss_mean=0.849, loss_mean_cls=0.0687, proj_loss=-0.425][2026-03-26 12:54:15] Step: 5589, Training Logs: loss_final: 0.489341, loss_mean: 0.834792, proj_loss: -0.416731, loss_mean_cls: 0.071279, grad_norm: 0.235571 +Steps: 1%| | 5590/1000000 [23:12<68:20:44, 4.04it/s, grad_norm=0.236, loss_final=0.489, loss_mean=0.835, loss_mean_cls=0.0713, proj_loss=-0.417][2026-03-26 12:54:15] Step: 5590, Training Logs: loss_final: 0.486814, loss_mean: 0.817778, proj_loss: -0.404830, loss_mean_cls: 0.073866, grad_norm: 0.250848 +Steps: 1%| | 5591/1000000 [23:12<68:20:08, 4.04it/s, grad_norm=0.251, loss_final=0.487, loss_mean=0.818, loss_mean_cls=0.0739, proj_loss=-0.405][2026-03-26 12:54:16] Step: 5591, Training Logs: loss_final: 0.473145, loss_mean: 0.819897, proj_loss: -0.417456, loss_mean_cls: 0.070704, grad_norm: 0.293857 +Steps: 1%| | 5592/1000000 [23:13<68:19:37, 4.04it/s, grad_norm=0.294, loss_final=0.473, loss_mean=0.82, loss_mean_cls=0.0707, proj_loss=-0.417][2026-03-26 12:54:16] Step: 5592, Training Logs: loss_final: 0.482893, loss_mean: 0.819295, proj_loss: -0.410193, loss_mean_cls: 0.073791, grad_norm: 0.324556 +Steps: 1%| | 5593/1000000 [23:13<68:20:05, 4.04it/s, grad_norm=0.325, loss_final=0.483, loss_mean=0.819, loss_mean_cls=0.0738, proj_loss=-0.41][2026-03-26 12:54:16] Step: 5593, Training Logs: loss_final: 0.483311, loss_mean: 0.819280, proj_loss: -0.408595, loss_mean_cls: 0.072626, grad_norm: 0.267880 +Steps: 1%| | 5594/1000000 [23:13<68:20:06, 4.04it/s, grad_norm=0.268, loss_final=0.483, loss_mean=0.819, loss_mean_cls=0.0726, proj_loss=-0.409][2026-03-26 12:54:16] Step: 5594, Training Logs: loss_final: 0.476279, loss_mean: 0.825150, proj_loss: -0.419023, loss_mean_cls: 0.070152, grad_norm: 0.280272 +Steps: 1%| | 5595/1000000 [23:13<68:18:03, 4.04it/s, grad_norm=0.28, loss_final=0.476, loss_mean=0.825, loss_mean_cls=0.0702, proj_loss=-0.419][2026-03-26 12:54:17] Step: 5595, Training Logs: loss_final: 0.492418, loss_mean: 0.832810, proj_loss: -0.412220, loss_mean_cls: 0.071827, grad_norm: 0.245131 +Steps: 1%| | 5596/1000000 [23:14<68:17:44, 4.04it/s, grad_norm=0.245, loss_final=0.492, loss_mean=0.833, loss_mean_cls=0.0718, proj_loss=-0.412][2026-03-26 12:54:17] Step: 5596, Training Logs: loss_final: 0.494636, loss_mean: 0.834788, proj_loss: -0.412658, loss_mean_cls: 0.072506, grad_norm: 0.266096 +Steps: 1%| | 5597/1000000 [23:14<68:19:02, 4.04it/s, grad_norm=0.266, loss_final=0.495, loss_mean=0.835, loss_mean_cls=0.0725, proj_loss=-0.413][2026-03-26 12:54:17] Step: 5597, Training Logs: loss_final: 0.478929, loss_mean: 0.828520, proj_loss: -0.419260, loss_mean_cls: 0.069669, grad_norm: 0.276351 +Steps: 1%| | 5598/1000000 [23:14<68:20:05, 4.04it/s, grad_norm=0.276, loss_final=0.479, loss_mean=0.829, loss_mean_cls=0.0697, proj_loss=-0.419][2026-03-26 12:54:17] Step: 5598, Training Logs: loss_final: 0.462216, loss_mean: 0.803448, proj_loss: -0.414300, loss_mean_cls: 0.073069, grad_norm: 0.268728 +Steps: 1%| | 5599/1000000 [23:14<68:17:56, 4.04it/s, grad_norm=0.269, loss_final=0.462, loss_mean=0.803, loss_mean_cls=0.0731, proj_loss=-0.414][2026-03-26 12:54:18] Step: 5599, Training Logs: loss_final: 0.478068, loss_mean: 0.806883, proj_loss: -0.403603, loss_mean_cls: 0.074788, grad_norm: 0.431865 +Steps: 1%| | 5600/1000000 [23:15<68:20:35, 4.04it/s, grad_norm=0.432, loss_final=0.478, loss_mean=0.807, loss_mean_cls=0.0748, proj_loss=-0.404][2026-03-26 12:54:18] Step: 5600, Training Logs: loss_final: 0.484108, loss_mean: 0.820921, proj_loss: -0.409049, loss_mean_cls: 0.072235, grad_norm: 0.459652 +Steps: 1%| | 5601/1000000 [23:15<68:19:30, 4.04it/s, grad_norm=0.46, loss_final=0.484, loss_mean=0.821, loss_mean_cls=0.0722, proj_loss=-0.409][2026-03-26 12:54:18] Step: 5601, Training Logs: loss_final: 0.475036, loss_mean: 0.824303, proj_loss: -0.419115, loss_mean_cls: 0.069848, grad_norm: 0.285253 +Steps: 1%| | 5602/1000000 [23:15<68:20:24, 4.04it/s, grad_norm=0.285, loss_final=0.475, loss_mean=0.824, loss_mean_cls=0.0698, proj_loss=-0.419][2026-03-26 12:54:18] Step: 5602, Training Logs: loss_final: 0.481057, loss_mean: 0.821749, proj_loss: -0.412532, loss_mean_cls: 0.071840, grad_norm: 0.710053 +Steps: 1%| | 5603/1000000 [23:15<68:21:17, 4.04it/s, grad_norm=0.71, loss_final=0.481, loss_mean=0.822, loss_mean_cls=0.0718, proj_loss=-0.413][2026-03-26 12:54:19] Step: 5603, Training Logs: loss_final: 0.475701, loss_mean: 0.818381, proj_loss: -0.414388, loss_mean_cls: 0.071708, grad_norm: 0.614204 +Steps: 1%| | 5604/1000000 [23:16<68:22:37, 4.04it/s, grad_norm=0.614, loss_final=0.476, loss_mean=0.818, loss_mean_cls=0.0717, proj_loss=-0.414][2026-03-26 12:54:19] Step: 5604, Training Logs: loss_final: 0.469415, loss_mean: 0.801596, proj_loss: -0.406336, loss_mean_cls: 0.074156, grad_norm: 0.233822 +Steps: 1%| | 5605/1000000 [23:16<68:21:22, 4.04it/s, grad_norm=0.234, loss_final=0.469, loss_mean=0.802, loss_mean_cls=0.0742, proj_loss=-0.406][2026-03-26 12:54:19] Step: 5605, Training Logs: loss_final: 0.480167, loss_mean: 0.831297, proj_loss: -0.420983, loss_mean_cls: 0.069853, grad_norm: 0.532052 +Steps: 1%| | 5606/1000000 [23:16<68:20:24, 4.04it/s, grad_norm=0.532, loss_final=0.48, loss_mean=0.831, loss_mean_cls=0.0699, proj_loss=-0.421][2026-03-26 12:54:19] Step: 5606, Training Logs: loss_final: 0.474086, loss_mean: 0.815457, proj_loss: -0.412773, loss_mean_cls: 0.071402, grad_norm: 0.319239 +Steps: 1%| | 5607/1000000 [23:16<68:20:40, 4.04it/s, grad_norm=0.319, loss_final=0.474, loss_mean=0.815, loss_mean_cls=0.0714, proj_loss=-0.413][2026-03-26 12:54:20] Step: 5607, Training Logs: loss_final: 0.477818, loss_mean: 0.819375, proj_loss: -0.413694, loss_mean_cls: 0.072137, grad_norm: 0.487654 +Steps: 1%| | 5608/1000000 [23:17<68:22:02, 4.04it/s, grad_norm=0.488, loss_final=0.478, loss_mean=0.819, loss_mean_cls=0.0721, proj_loss=-0.414][2026-03-26 12:54:20] Step: 5608, Training Logs: loss_final: 0.486758, loss_mean: 0.831761, proj_loss: -0.416505, loss_mean_cls: 0.071502, grad_norm: 0.719883 +Steps: 1%| | 5609/1000000 [23:17<68:24:17, 4.04it/s, grad_norm=0.72, loss_final=0.487, loss_mean=0.832, loss_mean_cls=0.0715, proj_loss=-0.417][2026-03-26 12:54:20] Step: 5609, Training Logs: loss_final: 0.485333, loss_mean: 0.828839, proj_loss: -0.415137, loss_mean_cls: 0.071632, grad_norm: 0.333156 +Steps: 1%| | 5610/1000000 [23:17<68:21:35, 4.04it/s, grad_norm=0.333, loss_final=0.485, loss_mean=0.829, loss_mean_cls=0.0716, proj_loss=-0.415][2026-03-26 12:54:20] Step: 5610, Training Logs: loss_final: 0.472759, loss_mean: 0.822773, proj_loss: -0.420448, loss_mean_cls: 0.070434, grad_norm: 0.601215 +Steps: 1%| | 5611/1000000 [23:17<68:21:17, 4.04it/s, grad_norm=0.601, loss_final=0.473, loss_mean=0.823, loss_mean_cls=0.0704, proj_loss=-0.42][2026-03-26 12:54:21] Step: 5611, Training Logs: loss_final: 0.494431, loss_mean: 0.836051, proj_loss: -0.413423, loss_mean_cls: 0.071803, grad_norm: 0.522903 +Steps: 1%| | 5612/1000000 [23:17<68:19:07, 4.04it/s, grad_norm=0.523, loss_final=0.494, loss_mean=0.836, loss_mean_cls=0.0718, proj_loss=-0.413][2026-03-26 12:54:21] Step: 5612, Training Logs: loss_final: 0.505983, loss_mean: 0.839022, proj_loss: -0.406415, loss_mean_cls: 0.073376, grad_norm: 0.444320 +Steps: 1%| | 5613/1000000 [23:18<68:20:47, 4.04it/s, grad_norm=0.444, loss_final=0.506, loss_mean=0.839, loss_mean_cls=0.0734, proj_loss=-0.406][2026-03-26 12:54:21] Step: 5613, Training Logs: loss_final: 0.482202, loss_mean: 0.817718, proj_loss: -0.408973, loss_mean_cls: 0.073457, grad_norm: 0.864164 +Steps: 1%| | 5614/1000000 [23:18<68:23:40, 4.04it/s, grad_norm=0.864, loss_final=0.482, loss_mean=0.818, loss_mean_cls=0.0735, proj_loss=-0.409][2026-03-26 12:54:21] Step: 5614, Training Logs: loss_final: 0.466201, loss_mean: 0.815259, proj_loss: -0.420672, loss_mean_cls: 0.071613, grad_norm: 0.317511 +Steps: 1%| | 5615/1000000 [23:18<68:22:35, 4.04it/s, grad_norm=0.318, loss_final=0.466, loss_mean=0.815, loss_mean_cls=0.0716, proj_loss=-0.421][2026-03-26 12:54:22] Step: 5615, Training Logs: loss_final: 0.475988, loss_mean: 0.811720, proj_loss: -0.408226, loss_mean_cls: 0.072494, grad_norm: 0.866398 +Steps: 1%| | 5616/1000000 [23:18<68:30:04, 4.03it/s, grad_norm=0.866, loss_final=0.476, loss_mean=0.812, loss_mean_cls=0.0725, proj_loss=-0.408][2026-03-26 12:54:22] Step: 5616, Training Logs: loss_final: 0.490623, loss_mean: 0.827406, proj_loss: -0.409379, loss_mean_cls: 0.072596, grad_norm: 0.585956 +Steps: 1%| | 5617/1000000 [23:19<68:30:12, 4.03it/s, grad_norm=0.586, loss_final=0.491, loss_mean=0.827, loss_mean_cls=0.0726, proj_loss=-0.409][2026-03-26 12:54:22] Step: 5617, Training Logs: loss_final: 0.483602, loss_mean: 0.820668, proj_loss: -0.409807, loss_mean_cls: 0.072741, grad_norm: 0.349579 +Steps: 1%| | 5618/1000000 [23:19<68:25:21, 4.04it/s, grad_norm=0.35, loss_final=0.484, loss_mean=0.821, loss_mean_cls=0.0727, proj_loss=-0.41][2026-03-26 12:54:22] Step: 5618, Training Logs: loss_final: 0.487109, loss_mean: 0.831831, proj_loss: -0.416176, loss_mean_cls: 0.071454, grad_norm: 0.629208 +Steps: 1%| | 5619/1000000 [23:19<68:22:06, 4.04it/s, grad_norm=0.629, loss_final=0.487, loss_mean=0.832, loss_mean_cls=0.0715, proj_loss=-0.416][2026-03-26 12:54:23] Step: 5619, Training Logs: loss_final: 0.468016, loss_mean: 0.819575, proj_loss: -0.421833, loss_mean_cls: 0.070274, grad_norm: 0.371580 +Steps: 1%| | 5620/1000000 [23:19<68:26:06, 4.04it/s, grad_norm=0.372, loss_final=0.468, loss_mean=0.82, loss_mean_cls=0.0703, proj_loss=-0.422][2026-03-26 12:54:23] Step: 5620, Training Logs: loss_final: 0.499035, loss_mean: 0.842800, proj_loss: -0.413575, loss_mean_cls: 0.069811, grad_norm: 0.484330 +Steps: 1%| | 5621/1000000 [23:20<68:24:49, 4.04it/s, grad_norm=0.484, loss_final=0.499, loss_mean=0.843, loss_mean_cls=0.0698, proj_loss=-0.414][2026-03-26 12:54:23] Step: 5621, Training Logs: loss_final: 0.506436, loss_mean: 0.852442, proj_loss: -0.417014, loss_mean_cls: 0.071007, grad_norm: 0.433042 +Steps: 1%| | 5622/1000000 [23:20<68:22:26, 4.04it/s, grad_norm=0.433, loss_final=0.506, loss_mean=0.852, loss_mean_cls=0.071, proj_loss=-0.417][2026-03-26 12:54:23] Step: 5622, Training Logs: loss_final: 0.500325, loss_mean: 0.829280, proj_loss: -0.401888, loss_mean_cls: 0.072933, grad_norm: 0.227781 +Steps: 1%| | 5623/1000000 [23:20<68:22:41, 4.04it/s, grad_norm=0.228, loss_final=0.5, loss_mean=0.829, loss_mean_cls=0.0729, proj_loss=-0.402][2026-03-26 12:54:24] Step: 5623, Training Logs: loss_final: 0.479671, loss_mean: 0.819569, proj_loss: -0.411775, loss_mean_cls: 0.071876, grad_norm: 0.468037 +Steps: 1%| | 5624/1000000 [23:20<68:20:05, 4.04it/s, grad_norm=0.468, loss_final=0.48, loss_mean=0.82, loss_mean_cls=0.0719, proj_loss=-0.412][2026-03-26 12:54:24] Step: 5624, Training Logs: loss_final: 0.477139, loss_mean: 0.821852, proj_loss: -0.415567, loss_mean_cls: 0.070854, grad_norm: 0.475709 +Steps: 1%| | 5625/1000000 [23:21<68:20:05, 4.04it/s, grad_norm=0.476, loss_final=0.477, loss_mean=0.822, loss_mean_cls=0.0709, proj_loss=-0.416][2026-03-26 12:54:24] Step: 5625, Training Logs: loss_final: 0.501752, loss_mean: 0.839603, proj_loss: -0.410459, loss_mean_cls: 0.072608, grad_norm: 0.466006 +Steps: 1%| | 5626/1000000 [23:21<68:23:22, 4.04it/s, grad_norm=0.466, loss_final=0.502, loss_mean=0.84, loss_mean_cls=0.0726, proj_loss=-0.41][2026-03-26 12:54:24] Step: 5626, Training Logs: loss_final: 0.475291, loss_mean: 0.814889, proj_loss: -0.412129, loss_mean_cls: 0.072530, grad_norm: 0.458860 +Steps: 1%| | 5627/1000000 [23:21<68:18:17, 4.04it/s, grad_norm=0.459, loss_final=0.475, loss_mean=0.815, loss_mean_cls=0.0725, proj_loss=-0.412][2026-03-26 12:54:25] Step: 5627, Training Logs: loss_final: 0.480681, loss_mean: 0.816302, proj_loss: -0.409090, loss_mean_cls: 0.073469, grad_norm: 0.395385 +Steps: 1%| | 5628/1000000 [23:22<88:27:30, 3.12it/s, grad_norm=0.395, loss_final=0.481, loss_mean=0.816, loss_mean_cls=0.0735, proj_loss=-0.409][2026-03-26 12:54:25] Step: 5628, Training Logs: loss_final: 0.487443, loss_mean: 0.819876, proj_loss: -0.405134, loss_mean_cls: 0.072702, grad_norm: 0.580052 +Steps: 1%| | 5629/1000000 [23:22<82:24:36, 3.35it/s, grad_norm=0.58, loss_final=0.487, loss_mean=0.82, loss_mean_cls=0.0727, proj_loss=-0.405][2026-03-26 12:54:25] Step: 5629, Training Logs: loss_final: 0.462562, loss_mean: 0.795347, proj_loss: -0.406008, loss_mean_cls: 0.073223, grad_norm: 0.286596 +Steps: 1%| | 5630/1000000 [23:22<78:08:00, 3.54it/s, grad_norm=0.287, loss_final=0.463, loss_mean=0.795, loss_mean_cls=0.0732, proj_loss=-0.406][2026-03-26 12:54:26] Step: 5630, Training Logs: loss_final: 0.500563, loss_mean: 0.848106, proj_loss: -0.418429, loss_mean_cls: 0.070886, grad_norm: 0.457651 +Steps: 1%| | 5631/1000000 [23:22<75:11:28, 3.67it/s, grad_norm=0.458, loss_final=0.501, loss_mean=0.848, loss_mean_cls=0.0709, proj_loss=-0.418][2026-03-26 12:54:26] Step: 5631, Training Logs: loss_final: 0.479067, loss_mean: 0.817604, proj_loss: -0.409820, loss_mean_cls: 0.071283, grad_norm: 0.448107 +Steps: 1%| | 5632/1000000 [23:23<73:06:41, 3.78it/s, grad_norm=0.448, loss_final=0.479, loss_mean=0.818, loss_mean_cls=0.0713, proj_loss=-0.41][2026-03-26 12:54:26] Step: 5632, Training Logs: loss_final: 0.481096, loss_mean: 0.826879, proj_loss: -0.416070, loss_mean_cls: 0.070287, grad_norm: 0.407955 +Steps: 1%| | 5633/1000000 [23:23<71:41:08, 3.85it/s, grad_norm=0.408, loss_final=0.481, loss_mean=0.827, loss_mean_cls=0.0703, proj_loss=-0.416][2026-03-26 12:54:26] Step: 5633, Training Logs: loss_final: 0.489657, loss_mean: 0.829757, proj_loss: -0.411528, loss_mean_cls: 0.071428, grad_norm: 0.602484 +Steps: 1%| | 5634/1000000 [23:23<70:39:51, 3.91it/s, grad_norm=0.602, loss_final=0.49, loss_mean=0.83, loss_mean_cls=0.0714, proj_loss=-0.412][2026-03-26 12:54:27] Step: 5634, Training Logs: loss_final: 0.489297, loss_mean: 0.837023, proj_loss: -0.417880, loss_mean_cls: 0.070154, grad_norm: 0.385578 +Steps: 1%| | 5635/1000000 [23:23<69:56:44, 3.95it/s, grad_norm=0.386, loss_final=0.489, loss_mean=0.837, loss_mean_cls=0.0702, proj_loss=-0.418][2026-03-26 12:54:27] Step: 5635, Training Logs: loss_final: 0.484509, loss_mean: 0.824983, proj_loss: -0.413332, loss_mean_cls: 0.072858, grad_norm: 0.525481 +Steps: 1%| | 5636/1000000 [23:24<69:25:42, 3.98it/s, grad_norm=0.525, loss_final=0.485, loss_mean=0.825, loss_mean_cls=0.0729, proj_loss=-0.413][2026-03-26 12:54:27] Step: 5636, Training Logs: loss_final: 0.475871, loss_mean: 0.815217, proj_loss: -0.412242, loss_mean_cls: 0.072897, grad_norm: 0.348476 +Steps: 1%| | 5637/1000000 [23:24<69:07:10, 4.00it/s, grad_norm=0.348, loss_final=0.476, loss_mean=0.815, loss_mean_cls=0.0729, proj_loss=-0.412][2026-03-26 12:54:27] Step: 5637, Training Logs: loss_final: 0.479171, loss_mean: 0.826598, proj_loss: -0.418019, loss_mean_cls: 0.070591, grad_norm: 0.655047 +Steps: 1%| | 5638/1000000 [23:24<68:53:27, 4.01it/s, grad_norm=0.655, loss_final=0.479, loss_mean=0.827, loss_mean_cls=0.0706, proj_loss=-0.418][2026-03-26 12:54:28] Step: 5638, Training Logs: loss_final: 0.477319, loss_mean: 0.811220, proj_loss: -0.406761, loss_mean_cls: 0.072859, grad_norm: 0.651449 +Steps: 1%| | 5639/1000000 [23:24<68:46:46, 4.02it/s, grad_norm=0.651, loss_final=0.477, loss_mean=0.811, loss_mean_cls=0.0729, proj_loss=-0.407][2026-03-26 12:54:28] Step: 5639, Training Logs: loss_final: 0.472128, loss_mean: 0.820874, proj_loss: -0.419583, loss_mean_cls: 0.070838, grad_norm: 0.381399 +Steps: 1%| | 5640/1000000 [23:25<68:34:22, 4.03it/s, grad_norm=0.381, loss_final=0.472, loss_mean=0.821, loss_mean_cls=0.0708, proj_loss=-0.42][2026-03-26 12:54:28] Step: 5640, Training Logs: loss_final: 0.488870, loss_mean: 0.834929, proj_loss: -0.416416, loss_mean_cls: 0.070358, grad_norm: 0.801523 +Steps: 1%| | 5641/1000000 [23:25<88:43:06, 3.11it/s, grad_norm=0.802, loss_final=0.489, loss_mean=0.835, loss_mean_cls=0.0704, proj_loss=-0.416][2026-03-26 12:54:29] Step: 5641, Training Logs: loss_final: 0.473121, loss_mean: 0.811602, proj_loss: -0.410383, loss_mean_cls: 0.071902, grad_norm: 0.458374 +Steps: 1%| | 5642/1000000 [23:25<82:37:24, 3.34it/s, grad_norm=0.458, loss_final=0.473, loss_mean=0.812, loss_mean_cls=0.0719, proj_loss=-0.41][2026-03-26 12:54:29] Step: 5642, Training Logs: loss_final: 0.496880, loss_mean: 0.834770, proj_loss: -0.410172, loss_mean_cls: 0.072281, grad_norm: 0.516230 +Steps: 1%| | 5643/1000000 [23:26<78:19:10, 3.53it/s, grad_norm=0.516, loss_final=0.497, loss_mean=0.835, loss_mean_cls=0.0723, proj_loss=-0.41][2026-03-26 12:54:29] Step: 5643, Training Logs: loss_final: 0.497957, loss_mean: 0.846321, proj_loss: -0.418046, loss_mean_cls: 0.069682, grad_norm: 0.641791 +Steps: 1%| | 5644/1000000 [23:26<75:19:38, 3.67it/s, grad_norm=0.642, loss_final=0.498, loss_mean=0.846, loss_mean_cls=0.0697, proj_loss=-0.418][2026-03-26 12:54:29] Step: 5644, Training Logs: loss_final: 0.488854, loss_mean: 0.832502, proj_loss: -0.414376, loss_mean_cls: 0.070728, grad_norm: 0.510532 +Steps: 1%| | 5645/1000000 [23:26<73:13:51, 3.77it/s, grad_norm=0.511, loss_final=0.489, loss_mean=0.833, loss_mean_cls=0.0707, proj_loss=-0.414][2026-03-26 12:54:30] Step: 5645, Training Logs: loss_final: 0.493155, loss_mean: 0.834407, proj_loss: -0.413111, loss_mean_cls: 0.071859, grad_norm: 0.592630 +Steps: 1%| | 5646/1000000 [23:26<71:46:08, 3.85it/s, grad_norm=0.593, loss_final=0.493, loss_mean=0.834, loss_mean_cls=0.0719, proj_loss=-0.413][2026-03-26 12:54:30] Step: 5646, Training Logs: loss_final: 0.478161, loss_mean: 0.820189, proj_loss: -0.413168, loss_mean_cls: 0.071141, grad_norm: 0.365869 +Steps: 1%| | 5647/1000000 [23:27<70:46:07, 3.90it/s, grad_norm=0.366, loss_final=0.478, loss_mean=0.82, loss_mean_cls=0.0711, proj_loss=-0.413][2026-03-26 12:54:30] Step: 5647, Training Logs: loss_final: 0.502049, loss_mean: 0.834382, proj_loss: -0.405152, loss_mean_cls: 0.072819, grad_norm: 0.386053 +Steps: 1%| | 5648/1000000 [23:27<70:01:45, 3.94it/s, grad_norm=0.386, loss_final=0.502, loss_mean=0.834, loss_mean_cls=0.0728, proj_loss=-0.405][2026-03-26 12:54:30] Step: 5648, Training Logs: loss_final: 0.490946, loss_mean: 0.831155, proj_loss: -0.412435, loss_mean_cls: 0.072226, grad_norm: 0.431889 +Steps: 1%| | 5649/1000000 [23:27<69:30:50, 3.97it/s, grad_norm=0.432, loss_final=0.491, loss_mean=0.831, loss_mean_cls=0.0722, proj_loss=-0.412][2026-03-26 12:54:31] Step: 5649, Training Logs: loss_final: 0.488880, loss_mean: 0.832552, proj_loss: -0.414194, loss_mean_cls: 0.070523, grad_norm: 0.289689 +Steps: 1%| | 5650/1000000 [23:27<69:09:37, 3.99it/s, grad_norm=0.29, loss_final=0.489, loss_mean=0.833, loss_mean_cls=0.0705, proj_loss=-0.414][2026-03-26 12:54:31] Step: 5650, Training Logs: loss_final: 0.461895, loss_mean: 0.807704, proj_loss: -0.416521, loss_mean_cls: 0.070712, grad_norm: 0.375244 +Steps: 1%| | 5651/1000000 [23:28<68:54:02, 4.01it/s, grad_norm=0.375, loss_final=0.462, loss_mean=0.808, loss_mean_cls=0.0707, proj_loss=-0.417][2026-03-26 12:54:31] Step: 5651, Training Logs: loss_final: 0.488277, loss_mean: 0.846745, proj_loss: -0.427272, loss_mean_cls: 0.068803, grad_norm: 0.383295 +Steps: 1%| | 5652/1000000 [23:28<68:45:00, 4.02it/s, grad_norm=0.383, loss_final=0.488, loss_mean=0.847, loss_mean_cls=0.0688, proj_loss=-0.427][2026-03-26 12:54:31] Step: 5652, Training Logs: loss_final: 0.481759, loss_mean: 0.819427, proj_loss: -0.410459, loss_mean_cls: 0.072791, grad_norm: 0.524297 +Steps: 1%| | 5653/1000000 [23:28<89:10:44, 3.10it/s, grad_norm=0.524, loss_final=0.482, loss_mean=0.819, loss_mean_cls=0.0728, proj_loss=-0.41][2026-03-26 12:54:32] Step: 5653, Training Logs: loss_final: 0.482733, loss_mean: 0.823257, proj_loss: -0.412205, loss_mean_cls: 0.071680, grad_norm: 0.259908 +Steps: 1%| | 5654/1000000 [23:29<83:04:56, 3.32it/s, grad_norm=0.26, loss_final=0.483, loss_mean=0.823, loss_mean_cls=0.0717, proj_loss=-0.412][2026-03-26 12:54:32] Step: 5654, Training Logs: loss_final: 0.489955, loss_mean: 0.821703, proj_loss: -0.403963, loss_mean_cls: 0.072215, grad_norm: 0.440607 +Steps: 1%| | 5655/1000000 [23:29<78:29:51, 3.52it/s, grad_norm=0.441, loss_final=0.49, loss_mean=0.822, loss_mean_cls=0.0722, proj_loss=-0.404][2026-03-26 12:54:32] Step: 5655, Training Logs: loss_final: 0.486909, loss_mean: 0.822880, proj_loss: -0.408096, loss_mean_cls: 0.072125, grad_norm: 0.478392 +Steps: 1%| | 5656/1000000 [23:29<75:25:26, 3.66it/s, grad_norm=0.478, loss_final=0.487, loss_mean=0.823, loss_mean_cls=0.0721, proj_loss=-0.408][2026-03-26 12:54:33] Step: 5656, Training Logs: loss_final: 0.480485, loss_mean: 0.836641, proj_loss: -0.425424, loss_mean_cls: 0.069267, grad_norm: 0.287485 +Steps: 1%| | 5657/1000000 [23:29<73:17:40, 3.77it/s, grad_norm=0.287, loss_final=0.48, loss_mean=0.837, loss_mean_cls=0.0693, proj_loss=-0.425][2026-03-26 12:54:33] Step: 5657, Training Logs: loss_final: 0.479046, loss_mean: 0.816818, proj_loss: -0.409619, loss_mean_cls: 0.071848, grad_norm: 0.375870 +Steps: 1%| | 5658/1000000 [23:30<71:47:47, 3.85it/s, grad_norm=0.376, loss_final=0.479, loss_mean=0.817, loss_mean_cls=0.0718, proj_loss=-0.41][2026-03-26 12:54:33] Step: 5658, Training Logs: loss_final: 0.478051, loss_mean: 0.838022, proj_loss: -0.427776, loss_mean_cls: 0.067805, grad_norm: 0.335585 +Steps: 1%| | 5659/1000000 [23:30<70:46:06, 3.90it/s, grad_norm=0.336, loss_final=0.478, loss_mean=0.838, loss_mean_cls=0.0678, proj_loss=-0.428][2026-03-26 12:54:33] Step: 5659, Training Logs: loss_final: 0.479340, loss_mean: 0.801615, proj_loss: -0.397549, loss_mean_cls: 0.075274, grad_norm: 0.371221 +Steps: 1%| | 5660/1000000 [23:30<70:02:53, 3.94it/s, grad_norm=0.371, loss_final=0.479, loss_mean=0.802, loss_mean_cls=0.0753, proj_loss=-0.398][2026-03-26 12:54:34] Step: 5660, Training Logs: loss_final: 0.484714, loss_mean: 0.829433, proj_loss: -0.414750, loss_mean_cls: 0.070031, grad_norm: 0.390547 +Steps: 1%| | 5661/1000000 [23:30<69:33:42, 3.97it/s, grad_norm=0.391, loss_final=0.485, loss_mean=0.829, loss_mean_cls=0.07, proj_loss=-0.415][2026-03-26 12:54:34] Step: 5661, Training Logs: loss_final: 0.493111, loss_mean: 0.839723, proj_loss: -0.417373, loss_mean_cls: 0.070761, grad_norm: 0.410269 +Steps: 1%| | 5662/1000000 [23:31<69:16:01, 3.99it/s, grad_norm=0.41, loss_final=0.493, loss_mean=0.84, loss_mean_cls=0.0708, proj_loss=-0.417][2026-03-26 12:54:34] Step: 5662, Training Logs: loss_final: 0.485120, loss_mean: 0.813965, proj_loss: -0.402976, loss_mean_cls: 0.074131, grad_norm: 0.298478 +Steps: 1%| | 5663/1000000 [23:31<68:57:39, 4.01it/s, grad_norm=0.298, loss_final=0.485, loss_mean=0.814, loss_mean_cls=0.0741, proj_loss=-0.403][2026-03-26 12:54:34] Step: 5663, Training Logs: loss_final: 0.474693, loss_mean: 0.826056, proj_loss: -0.419985, loss_mean_cls: 0.068623, grad_norm: 0.364686 +Steps: 1%| | 5664/1000000 [23:31<68:43:50, 4.02it/s, grad_norm=0.365, loss_final=0.475, loss_mean=0.826, loss_mean_cls=0.0686, proj_loss=-0.42][2026-03-26 12:54:34] Step: 5664, Training Logs: loss_final: 0.475773, loss_mean: 0.809158, proj_loss: -0.406499, loss_mean_cls: 0.073114, grad_norm: 0.364805 +Steps: 1%| | 5665/1000000 [23:31<68:37:33, 4.02it/s, grad_norm=0.365, loss_final=0.476, loss_mean=0.809, loss_mean_cls=0.0731, proj_loss=-0.406][2026-03-26 12:54:35] Step: 5665, Training Logs: loss_final: 0.482925, loss_mean: 0.827956, proj_loss: -0.416070, loss_mean_cls: 0.071038, grad_norm: 0.252400 +Steps: 1%| | 5666/1000000 [23:32<68:27:00, 4.04it/s, grad_norm=0.252, loss_final=0.483, loss_mean=0.828, loss_mean_cls=0.071, proj_loss=-0.416][2026-03-26 12:54:35] Step: 5666, Training Logs: loss_final: 0.485134, loss_mean: 0.829876, proj_loss: -0.415113, loss_mean_cls: 0.070371, grad_norm: 0.365976 +Steps: 1%| | 5667/1000000 [23:32<68:27:06, 4.04it/s, grad_norm=0.366, loss_final=0.485, loss_mean=0.83, loss_mean_cls=0.0704, proj_loss=-0.415][2026-03-26 12:54:35] Step: 5667, Training Logs: loss_final: 0.482327, loss_mean: 0.819816, proj_loss: -0.409300, loss_mean_cls: 0.071810, grad_norm: 0.368993 +Steps: 1%| | 5668/1000000 [23:32<68:21:54, 4.04it/s, grad_norm=0.369, loss_final=0.482, loss_mean=0.82, loss_mean_cls=0.0718, proj_loss=-0.409][2026-03-26 12:54:35] Step: 5668, Training Logs: loss_final: 0.455605, loss_mean: 0.804573, proj_loss: -0.419722, loss_mean_cls: 0.070754, grad_norm: 0.266921 +Steps: 1%| | 5669/1000000 [23:32<68:20:43, 4.04it/s, grad_norm=0.267, loss_final=0.456, loss_mean=0.805, loss_mean_cls=0.0708, proj_loss=-0.42][2026-03-26 12:54:36] Step: 5669, Training Logs: loss_final: 0.481981, loss_mean: 0.819481, proj_loss: -0.409765, loss_mean_cls: 0.072265, grad_norm: 0.316950 +Steps: 1%| | 5670/1000000 [23:33<68:20:43, 4.04it/s, grad_norm=0.317, loss_final=0.482, loss_mean=0.819, loss_mean_cls=0.0723, proj_loss=-0.41][2026-03-26 12:54:36] Step: 5670, Training Logs: loss_final: 0.478422, loss_mean: 0.836876, proj_loss: -0.427297, loss_mean_cls: 0.068844, grad_norm: 0.412979 +Steps: 1%| | 5671/1000000 [23:33<68:20:10, 4.04it/s, grad_norm=0.413, loss_final=0.478, loss_mean=0.837, loss_mean_cls=0.0688, proj_loss=-0.427][2026-03-26 12:54:36] Step: 5671, Training Logs: loss_final: 0.498124, loss_mean: 0.842658, proj_loss: -0.416022, loss_mean_cls: 0.071488, grad_norm: 0.474997 +Steps: 1%| | 5672/1000000 [23:33<68:21:22, 4.04it/s, grad_norm=0.475, loss_final=0.498, loss_mean=0.843, loss_mean_cls=0.0715, proj_loss=-0.416][2026-03-26 12:54:36] Step: 5672, Training Logs: loss_final: 0.476766, loss_mean: 0.826846, proj_loss: -0.419840, loss_mean_cls: 0.069761, grad_norm: 0.342616 +Steps: 1%| | 5673/1000000 [23:33<68:21:53, 4.04it/s, grad_norm=0.343, loss_final=0.477, loss_mean=0.827, loss_mean_cls=0.0698, proj_loss=-0.42][2026-03-26 12:54:37] Step: 5673, Training Logs: loss_final: 0.484770, loss_mean: 0.837374, proj_loss: -0.422085, loss_mean_cls: 0.069481, grad_norm: 0.364196 +Steps: 1%| | 5674/1000000 [23:34<68:22:39, 4.04it/s, grad_norm=0.364, loss_final=0.485, loss_mean=0.837, loss_mean_cls=0.0695, proj_loss=-0.422][2026-03-26 12:54:37] Step: 5674, Training Logs: loss_final: 0.488780, loss_mean: 0.829634, proj_loss: -0.412008, loss_mean_cls: 0.071153, grad_norm: 0.359094 +Steps: 1%| | 5675/1000000 [23:34<68:23:06, 4.04it/s, grad_norm=0.359, loss_final=0.489, loss_mean=0.83, loss_mean_cls=0.0712, proj_loss=-0.412][2026-03-26 12:54:37] Step: 5675, Training Logs: loss_final: 0.469632, loss_mean: 0.811104, proj_loss: -0.412489, loss_mean_cls: 0.071018, grad_norm: 0.226384 +Steps: 1%| | 5676/1000000 [23:34<68:24:32, 4.04it/s, grad_norm=0.226, loss_final=0.47, loss_mean=0.811, loss_mean_cls=0.071, proj_loss=-0.412][2026-03-26 12:54:37] Step: 5676, Training Logs: loss_final: 0.488786, loss_mean: 0.830149, proj_loss: -0.412917, loss_mean_cls: 0.071553, grad_norm: 0.419206 +Steps: 1%| | 5677/1000000 [23:34<68:22:39, 4.04it/s, grad_norm=0.419, loss_final=0.489, loss_mean=0.83, loss_mean_cls=0.0716, proj_loss=-0.413][2026-03-26 12:54:38] Step: 5677, Training Logs: loss_final: 0.499768, loss_mean: 0.849997, proj_loss: -0.419714, loss_mean_cls: 0.069485, grad_norm: 0.363317 +Steps: 1%| | 5678/1000000 [23:35<68:25:48, 4.04it/s, grad_norm=0.363, loss_final=0.5, loss_mean=0.85, loss_mean_cls=0.0695, proj_loss=-0.42][2026-03-26 12:54:38] Step: 5678, Training Logs: loss_final: 0.505194, loss_mean: 0.845020, proj_loss: -0.410846, loss_mean_cls: 0.071020, grad_norm: 0.369515 +Steps: 1%| | 5679/1000000 [23:35<68:24:59, 4.04it/s, grad_norm=0.37, loss_final=0.505, loss_mean=0.845, loss_mean_cls=0.071, proj_loss=-0.411][2026-03-26 12:54:38] Step: 5679, Training Logs: loss_final: 0.470880, loss_mean: 0.805702, proj_loss: -0.406625, loss_mean_cls: 0.071804, grad_norm: 0.294451 +Steps: 1%| | 5680/1000000 [23:35<68:24:54, 4.04it/s, grad_norm=0.294, loss_final=0.471, loss_mean=0.806, loss_mean_cls=0.0718, proj_loss=-0.407][2026-03-26 12:54:38] Step: 5680, Training Logs: loss_final: 0.468458, loss_mean: 0.802266, proj_loss: -0.406735, loss_mean_cls: 0.072928, grad_norm: 0.312077 +Steps: 1%| | 5681/1000000 [23:35<68:22:48, 4.04it/s, grad_norm=0.312, loss_final=0.468, loss_mean=0.802, loss_mean_cls=0.0729, proj_loss=-0.407][2026-03-26 12:54:39] Step: 5681, Training Logs: loss_final: 0.472034, loss_mean: 0.817048, proj_loss: -0.415636, loss_mean_cls: 0.070622, grad_norm: 0.690445 +Steps: 1%| | 5682/1000000 [23:36<68:22:54, 4.04it/s, grad_norm=0.69, loss_final=0.472, loss_mean=0.817, loss_mean_cls=0.0706, proj_loss=-0.416][2026-03-26 12:54:39] Step: 5682, Training Logs: loss_final: 0.477638, loss_mean: 0.827004, proj_loss: -0.418649, loss_mean_cls: 0.069284, grad_norm: 0.496407 +Steps: 1%| | 5683/1000000 [23:36<68:20:55, 4.04it/s, grad_norm=0.496, loss_final=0.478, loss_mean=0.827, loss_mean_cls=0.0693, proj_loss=-0.419][2026-03-26 12:54:39] Step: 5683, Training Logs: loss_final: 0.483941, loss_mean: 0.813677, proj_loss: -0.403061, loss_mean_cls: 0.073325, grad_norm: 0.511004 +Steps: 1%| | 5684/1000000 [23:36<68:21:16, 4.04it/s, grad_norm=0.511, loss_final=0.484, loss_mean=0.814, loss_mean_cls=0.0733, proj_loss=-0.403][2026-03-26 12:54:39] Step: 5684, Training Logs: loss_final: 0.490826, loss_mean: 0.825760, proj_loss: -0.407806, loss_mean_cls: 0.072872, grad_norm: 0.679608 +Steps: 1%| | 5685/1000000 [23:36<68:18:43, 4.04it/s, grad_norm=0.68, loss_final=0.491, loss_mean=0.826, loss_mean_cls=0.0729, proj_loss=-0.408][2026-03-26 12:54:40] Step: 5685, Training Logs: loss_final: 0.461219, loss_mean: 0.805295, proj_loss: -0.415748, loss_mean_cls: 0.071672, grad_norm: 0.399453 +Steps: 1%| | 5686/1000000 [23:37<68:17:55, 4.04it/s, grad_norm=0.399, loss_final=0.461, loss_mean=0.805, loss_mean_cls=0.0717, proj_loss=-0.416][2026-03-26 12:54:40] Step: 5686, Training Logs: loss_final: 0.481557, loss_mean: 0.825630, proj_loss: -0.415169, loss_mean_cls: 0.071096, grad_norm: 0.544175 +Steps: 1%| | 5687/1000000 [23:37<68:16:45, 4.05it/s, grad_norm=0.544, loss_final=0.482, loss_mean=0.826, loss_mean_cls=0.0711, proj_loss=-0.415][2026-03-26 12:54:40] Step: 5687, Training Logs: loss_final: 0.461826, loss_mean: 0.797909, proj_loss: -0.408794, loss_mean_cls: 0.072710, grad_norm: 0.618224 +Steps: 1%| | 5688/1000000 [23:37<68:22:43, 4.04it/s, grad_norm=0.618, loss_final=0.462, loss_mean=0.798, loss_mean_cls=0.0727, proj_loss=-0.409][2026-03-26 12:54:40] Step: 5688, Training Logs: loss_final: 0.486867, loss_mean: 0.839660, proj_loss: -0.421399, loss_mean_cls: 0.068606, grad_norm: 0.308318 +Steps: 1%| | 5689/1000000 [23:37<68:18:56, 4.04it/s, grad_norm=0.308, loss_final=0.487, loss_mean=0.84, loss_mean_cls=0.0686, proj_loss=-0.421][2026-03-26 12:54:41] Step: 5689, Training Logs: loss_final: 0.490622, loss_mean: 0.835201, proj_loss: -0.415764, loss_mean_cls: 0.071185, grad_norm: 0.587674 +Steps: 1%| | 5690/1000000 [23:38<68:20:13, 4.04it/s, grad_norm=0.588, loss_final=0.491, loss_mean=0.835, loss_mean_cls=0.0712, proj_loss=-0.416][2026-03-26 12:54:41] Step: 5690, Training Logs: loss_final: 0.463169, loss_mean: 0.807446, proj_loss: -0.415559, loss_mean_cls: 0.071282, grad_norm: 0.329265 +Steps: 1%| | 5691/1000000 [23:38<68:18:15, 4.04it/s, grad_norm=0.329, loss_final=0.463, loss_mean=0.807, loss_mean_cls=0.0713, proj_loss=-0.416][2026-03-26 12:54:41] Step: 5691, Training Logs: loss_final: 0.483811, loss_mean: 0.829244, proj_loss: -0.415148, loss_mean_cls: 0.069715, grad_norm: 0.676509 +Steps: 1%| | 5692/1000000 [23:38<68:58:55, 4.00it/s, grad_norm=0.677, loss_final=0.484, loss_mean=0.829, loss_mean_cls=0.0697, proj_loss=-0.415][2026-03-26 12:54:41] Step: 5692, Training Logs: loss_final: 0.476540, loss_mean: 0.812057, proj_loss: -0.407751, loss_mean_cls: 0.072234, grad_norm: 0.552171 +Steps: 1%| | 5693/1000000 [23:38<68:46:21, 4.02it/s, grad_norm=0.552, loss_final=0.477, loss_mean=0.812, loss_mean_cls=0.0722, proj_loss=-0.408][2026-03-26 12:54:42] Step: 5693, Training Logs: loss_final: 0.483851, loss_mean: 0.825910, proj_loss: -0.413776, loss_mean_cls: 0.071717, grad_norm: 0.486020 +Steps: 1%| | 5694/1000000 [23:39<68:38:21, 4.02it/s, grad_norm=0.486, loss_final=0.484, loss_mean=0.826, loss_mean_cls=0.0717, proj_loss=-0.414][2026-03-26 12:54:42] Step: 5694, Training Logs: loss_final: 0.470640, loss_mean: 0.816904, proj_loss: -0.416650, loss_mean_cls: 0.070386, grad_norm: 0.781494 +Steps: 1%| | 5695/1000000 [23:39<68:32:08, 4.03it/s, grad_norm=0.781, loss_final=0.471, loss_mean=0.817, loss_mean_cls=0.0704, proj_loss=-0.417][2026-03-26 12:54:42] Step: 5695, Training Logs: loss_final: 0.482982, loss_mean: 0.821088, proj_loss: -0.410376, loss_mean_cls: 0.072270, grad_norm: 0.243756 +Steps: 1%| | 5696/1000000 [23:39<68:31:00, 4.03it/s, grad_norm=0.244, loss_final=0.483, loss_mean=0.821, loss_mean_cls=0.0723, proj_loss=-0.41][2026-03-26 12:54:42] Step: 5696, Training Logs: loss_final: 0.469136, loss_mean: 0.816850, proj_loss: -0.418292, loss_mean_cls: 0.070578, grad_norm: 0.893220 +Steps: 1%| | 5697/1000000 [23:39<68:27:31, 4.03it/s, grad_norm=0.893, loss_final=0.469, loss_mean=0.817, loss_mean_cls=0.0706, proj_loss=-0.418][2026-03-26 12:54:43] Step: 5697, Training Logs: loss_final: 0.485879, loss_mean: 0.838267, proj_loss: -0.421648, loss_mean_cls: 0.069260, grad_norm: 0.483818 +Steps: 1%| | 5698/1000000 [23:40<68:26:24, 4.04it/s, grad_norm=0.484, loss_final=0.486, loss_mean=0.838, loss_mean_cls=0.0693, proj_loss=-0.422][2026-03-26 12:54:43] Step: 5698, Training Logs: loss_final: 0.489994, loss_mean: 0.846450, proj_loss: -0.424380, loss_mean_cls: 0.067924, grad_norm: 0.669521 +Steps: 1%| | 5699/1000000 [23:40<68:24:07, 4.04it/s, grad_norm=0.67, loss_final=0.49, loss_mean=0.846, loss_mean_cls=0.0679, proj_loss=-0.424][2026-03-26 12:54:43] Step: 5699, Training Logs: loss_final: 0.483417, loss_mean: 0.832953, proj_loss: -0.419995, loss_mean_cls: 0.070460, grad_norm: 0.441765 +Steps: 1%| | 5700/1000000 [23:40<68:45:16, 4.02it/s, grad_norm=0.442, loss_final=0.483, loss_mean=0.833, loss_mean_cls=0.0705, proj_loss=-0.42][2026-03-26 12:54:43] Step: 5700, Training Logs: loss_final: 0.471542, loss_mean: 0.819302, proj_loss: -0.418090, loss_mean_cls: 0.070331, grad_norm: 0.642485 +Steps: 1%| | 5701/1000000 [23:40<68:36:01, 4.03it/s, grad_norm=0.642, loss_final=0.472, loss_mean=0.819, loss_mean_cls=0.0703, proj_loss=-0.418][2026-03-26 12:54:44] Step: 5701, Training Logs: loss_final: 0.494518, loss_mean: 0.834943, proj_loss: -0.412432, loss_mean_cls: 0.072006, grad_norm: 0.568615 +Steps: 1%| | 5702/1000000 [23:41<68:34:20, 4.03it/s, grad_norm=0.569, loss_final=0.495, loss_mean=0.835, loss_mean_cls=0.072, proj_loss=-0.412][2026-03-26 12:54:44] Step: 5702, Training Logs: loss_final: 0.488342, loss_mean: 0.840012, proj_loss: -0.420873, loss_mean_cls: 0.069203, grad_norm: 0.466592 +Steps: 1%| | 5703/1000000 [23:41<68:29:25, 4.03it/s, grad_norm=0.467, loss_final=0.488, loss_mean=0.84, loss_mean_cls=0.0692, proj_loss=-0.421][2026-03-26 12:54:44] Step: 5703, Training Logs: loss_final: 0.467128, loss_mean: 0.822792, proj_loss: -0.424636, loss_mean_cls: 0.068971, grad_norm: 0.725040 +Steps: 1%| | 5704/1000000 [23:41<68:26:31, 4.04it/s, grad_norm=0.725, loss_final=0.467, loss_mean=0.823, loss_mean_cls=0.069, proj_loss=-0.425][2026-03-26 12:54:44] Step: 5704, Training Logs: loss_final: 0.485353, loss_mean: 0.834564, proj_loss: -0.419220, loss_mean_cls: 0.070009, grad_norm: 0.552551 +Steps: 1%| | 5705/1000000 [23:41<68:26:22, 4.04it/s, grad_norm=0.553, loss_final=0.485, loss_mean=0.835, loss_mean_cls=0.07, proj_loss=-0.419][2026-03-26 12:54:45] Step: 5705, Training Logs: loss_final: 0.478294, loss_mean: 0.822940, proj_loss: -0.415538, loss_mean_cls: 0.070892, grad_norm: 1.145937 +Steps: 1%| | 5706/1000000 [23:42<68:22:22, 4.04it/s, grad_norm=1.15, loss_final=0.478, loss_mean=0.823, loss_mean_cls=0.0709, proj_loss=-0.416][2026-03-26 12:54:45] Step: 5706, Training Logs: loss_final: 0.480016, loss_mean: 0.826814, proj_loss: -0.417428, loss_mean_cls: 0.070630, grad_norm: 0.397932 +Steps: 1%| | 5707/1000000 [23:42<68:19:44, 4.04it/s, grad_norm=0.398, loss_final=0.48, loss_mean=0.827, loss_mean_cls=0.0706, proj_loss=-0.417][2026-03-26 12:54:45] Step: 5707, Training Logs: loss_final: 0.487012, loss_mean: 0.822285, proj_loss: -0.408723, loss_mean_cls: 0.073450, grad_norm: 1.034395 +Steps: 1%| | 5708/1000000 [23:42<68:19:16, 4.04it/s, grad_norm=1.03, loss_final=0.487, loss_mean=0.822, loss_mean_cls=0.0735, proj_loss=-0.409][2026-03-26 12:54:45] Step: 5708, Training Logs: loss_final: 0.477682, loss_mean: 0.818480, proj_loss: -0.412712, loss_mean_cls: 0.071913, grad_norm: 0.293504 +Steps: 1%| | 5709/1000000 [23:42<68:18:41, 4.04it/s, grad_norm=0.294, loss_final=0.478, loss_mean=0.818, loss_mean_cls=0.0719, proj_loss=-0.413][2026-03-26 12:54:46] Step: 5709, Training Logs: loss_final: 0.485171, loss_mean: 0.834539, proj_loss: -0.418486, loss_mean_cls: 0.069118, grad_norm: 0.718172 +Steps: 1%| | 5710/1000000 [23:42<68:18:15, 4.04it/s, grad_norm=0.718, loss_final=0.485, loss_mean=0.835, loss_mean_cls=0.0691, proj_loss=-0.418][2026-03-26 12:54:46] Step: 5710, Training Logs: loss_final: 0.481656, loss_mean: 0.834403, proj_loss: -0.422030, loss_mean_cls: 0.069282, grad_norm: 0.453316 +Steps: 1%| | 5711/1000000 [23:43<68:19:15, 4.04it/s, grad_norm=0.453, loss_final=0.482, loss_mean=0.834, loss_mean_cls=0.0693, proj_loss=-0.422][2026-03-26 12:54:46] Step: 5711, Training Logs: loss_final: 0.475802, loss_mean: 0.823556, proj_loss: -0.417571, loss_mean_cls: 0.069816, grad_norm: 0.629161 +Steps: 1%| | 5712/1000000 [23:43<68:21:18, 4.04it/s, grad_norm=0.629, loss_final=0.476, loss_mean=0.824, loss_mean_cls=0.0698, proj_loss=-0.418][2026-03-26 12:54:46] Step: 5712, Training Logs: loss_final: 0.480422, loss_mean: 0.816079, proj_loss: -0.408141, loss_mean_cls: 0.072485, grad_norm: 0.685854 +Steps: 1%| | 5713/1000000 [23:43<68:21:44, 4.04it/s, grad_norm=0.686, loss_final=0.48, loss_mean=0.816, loss_mean_cls=0.0725, proj_loss=-0.408][2026-03-26 12:54:47] Step: 5713, Training Logs: loss_final: 0.483767, loss_mean: 0.826090, proj_loss: -0.413619, loss_mean_cls: 0.071297, grad_norm: 0.452415 +Steps: 1%| | 5714/1000000 [23:43<68:17:57, 4.04it/s, grad_norm=0.452, loss_final=0.484, loss_mean=0.826, loss_mean_cls=0.0713, proj_loss=-0.414][2026-03-26 12:54:47] Step: 5714, Training Logs: loss_final: 0.484023, loss_mean: 0.822495, proj_loss: -0.410966, loss_mean_cls: 0.072494, grad_norm: 0.631076 +Steps: 1%| | 5715/1000000 [23:44<68:19:07, 4.04it/s, grad_norm=0.631, loss_final=0.484, loss_mean=0.822, loss_mean_cls=0.0725, proj_loss=-0.411][2026-03-26 12:54:47] Step: 5715, Training Logs: loss_final: 0.486061, loss_mean: 0.828245, proj_loss: -0.412592, loss_mean_cls: 0.070408, grad_norm: 0.339356 +Steps: 1%| | 5716/1000000 [23:44<68:22:15, 4.04it/s, grad_norm=0.339, loss_final=0.486, loss_mean=0.828, loss_mean_cls=0.0704, proj_loss=-0.413][2026-03-26 12:54:47] Step: 5716, Training Logs: loss_final: 0.485869, loss_mean: 0.818711, proj_loss: -0.405196, loss_mean_cls: 0.072354, grad_norm: 0.826314 +Steps: 1%| | 5717/1000000 [23:44<68:25:28, 4.04it/s, grad_norm=0.826, loss_final=0.486, loss_mean=0.819, loss_mean_cls=0.0724, proj_loss=-0.405][2026-03-26 12:54:48] Step: 5717, Training Logs: loss_final: 0.495994, loss_mean: 0.841443, proj_loss: -0.415532, loss_mean_cls: 0.070082, grad_norm: 0.300902 +Steps: 1%| | 5718/1000000 [23:44<68:20:00, 4.04it/s, grad_norm=0.301, loss_final=0.496, loss_mean=0.841, loss_mean_cls=0.0701, proj_loss=-0.416][2026-03-26 12:54:48] Step: 5718, Training Logs: loss_final: 0.492978, loss_mean: 0.844819, proj_loss: -0.421647, loss_mean_cls: 0.069807, grad_norm: 0.760496 +Steps: 1%| | 5719/1000000 [23:45<68:20:06, 4.04it/s, grad_norm=0.76, loss_final=0.493, loss_mean=0.845, loss_mean_cls=0.0698, proj_loss=-0.422][2026-03-26 12:54:48] Step: 5719, Training Logs: loss_final: 0.491788, loss_mean: 0.819035, proj_loss: -0.400648, loss_mean_cls: 0.073400, grad_norm: 0.516656 +Steps: 1%| | 5720/1000000 [23:45<68:16:09, 4.05it/s, grad_norm=0.517, loss_final=0.492, loss_mean=0.819, loss_mean_cls=0.0734, proj_loss=-0.401][2026-03-26 12:54:48] Step: 5720, Training Logs: loss_final: 0.475192, loss_mean: 0.814398, proj_loss: -0.410376, loss_mean_cls: 0.071170, grad_norm: 0.630066 +Steps: 1%| | 5721/1000000 [23:45<68:15:14, 4.05it/s, grad_norm=0.63, loss_final=0.475, loss_mean=0.814, loss_mean_cls=0.0712, proj_loss=-0.41][2026-03-26 12:54:49] Step: 5721, Training Logs: loss_final: 0.482772, loss_mean: 0.838143, proj_loss: -0.424004, loss_mean_cls: 0.068634, grad_norm: 0.546367 +Steps: 1%| | 5722/1000000 [23:45<68:17:33, 4.04it/s, grad_norm=0.546, loss_final=0.483, loss_mean=0.838, loss_mean_cls=0.0686, proj_loss=-0.424][2026-03-26 12:54:49] Step: 5722, Training Logs: loss_final: 0.480947, loss_mean: 0.823616, proj_loss: -0.414850, loss_mean_cls: 0.072181, grad_norm: 0.439675 +Steps: 1%| | 5723/1000000 [23:46<68:17:50, 4.04it/s, grad_norm=0.44, loss_final=0.481, loss_mean=0.824, loss_mean_cls=0.0722, proj_loss=-0.415][2026-03-26 12:54:49] Step: 5723, Training Logs: loss_final: 0.483878, loss_mean: 0.838161, proj_loss: -0.423354, loss_mean_cls: 0.069072, grad_norm: 0.464429 +Steps: 1%| | 5724/1000000 [23:46<68:18:42, 4.04it/s, grad_norm=0.464, loss_final=0.484, loss_mean=0.838, loss_mean_cls=0.0691, proj_loss=-0.423][2026-03-26 12:54:49] Step: 5724, Training Logs: loss_final: 0.486628, loss_mean: 0.830990, proj_loss: -0.414952, loss_mean_cls: 0.070590, grad_norm: 0.321835 +Steps: 1%| | 5725/1000000 [23:46<68:17:00, 4.04it/s, grad_norm=0.322, loss_final=0.487, loss_mean=0.831, loss_mean_cls=0.0706, proj_loss=-0.415][2026-03-26 12:54:50] Step: 5725, Training Logs: loss_final: 0.485696, loss_mean: 0.833662, proj_loss: -0.418149, loss_mean_cls: 0.070183, grad_norm: 0.409452 +Steps: 1%| | 5726/1000000 [23:46<68:20:12, 4.04it/s, grad_norm=0.409, loss_final=0.486, loss_mean=0.834, loss_mean_cls=0.0702, proj_loss=-0.418][2026-03-26 12:54:50] Step: 5726, Training Logs: loss_final: 0.487194, loss_mean: 0.828582, proj_loss: -0.412521, loss_mean_cls: 0.071133, grad_norm: 0.323764 +Steps: 1%| | 5727/1000000 [23:47<68:20:05, 4.04it/s, grad_norm=0.324, loss_final=0.487, loss_mean=0.829, loss_mean_cls=0.0711, proj_loss=-0.413][2026-03-26 12:54:50] Step: 5727, Training Logs: loss_final: 0.480565, loss_mean: 0.827203, proj_loss: -0.416814, loss_mean_cls: 0.070176, grad_norm: 0.357432 +Steps: 1%| | 5728/1000000 [23:47<68:22:06, 4.04it/s, grad_norm=0.357, loss_final=0.481, loss_mean=0.827, loss_mean_cls=0.0702, proj_loss=-0.417][2026-03-26 12:54:50] Step: 5728, Training Logs: loss_final: 0.460976, loss_mean: 0.799124, proj_loss: -0.410198, loss_mean_cls: 0.072051, grad_norm: 0.347991 +Steps: 1%| | 5729/1000000 [23:47<68:19:54, 4.04it/s, grad_norm=0.348, loss_final=0.461, loss_mean=0.799, loss_mean_cls=0.0721, proj_loss=-0.41][2026-03-26 12:54:51] Step: 5729, Training Logs: loss_final: 0.491649, loss_mean: 0.833102, proj_loss: -0.413186, loss_mean_cls: 0.071733, grad_norm: 0.263058 +Steps: 1%| | 5730/1000000 [23:47<68:21:46, 4.04it/s, grad_norm=0.263, loss_final=0.492, loss_mean=0.833, loss_mean_cls=0.0717, proj_loss=-0.413][2026-03-26 12:54:51] Step: 5730, Training Logs: loss_final: 0.501147, loss_mean: 0.835927, proj_loss: -0.406544, loss_mean_cls: 0.071764, grad_norm: 0.471688 +Steps: 1%| | 5731/1000000 [23:48<68:18:41, 4.04it/s, grad_norm=0.472, loss_final=0.501, loss_mean=0.836, loss_mean_cls=0.0718, proj_loss=-0.407][2026-03-26 12:54:51] Step: 5731, Training Logs: loss_final: 0.497625, loss_mean: 0.836351, proj_loss: -0.409932, loss_mean_cls: 0.071206, grad_norm: 0.324366 +Steps: 1%| | 5732/1000000 [23:48<68:21:58, 4.04it/s, grad_norm=0.324, loss_final=0.498, loss_mean=0.836, loss_mean_cls=0.0712, proj_loss=-0.41][2026-03-26 12:54:51] Step: 5732, Training Logs: loss_final: 0.486560, loss_mean: 0.832305, proj_loss: -0.417220, loss_mean_cls: 0.071476, grad_norm: 0.507184 +Steps: 1%| | 5733/1000000 [23:48<68:21:21, 4.04it/s, grad_norm=0.507, loss_final=0.487, loss_mean=0.832, loss_mean_cls=0.0715, proj_loss=-0.417][2026-03-26 12:54:52] Step: 5733, Training Logs: loss_final: 0.506524, loss_mean: 0.849541, proj_loss: -0.411810, loss_mean_cls: 0.068793, grad_norm: 0.542158 +Steps: 1%| | 5734/1000000 [23:48<68:21:13, 4.04it/s, grad_norm=0.542, loss_final=0.507, loss_mean=0.85, loss_mean_cls=0.0688, proj_loss=-0.412][2026-03-26 12:54:52] Step: 5734, Training Logs: loss_final: 0.466262, loss_mean: 0.823399, proj_loss: -0.425159, loss_mean_cls: 0.068022, grad_norm: 0.240993 +Steps: 1%| | 5735/1000000 [23:49<68:21:37, 4.04it/s, grad_norm=0.241, loss_final=0.466, loss_mean=0.823, loss_mean_cls=0.068, proj_loss=-0.425][2026-03-26 12:54:52] Step: 5735, Training Logs: loss_final: 0.475286, loss_mean: 0.826605, proj_loss: -0.420599, loss_mean_cls: 0.069280, grad_norm: 0.422626 +Steps: 1%| | 5736/1000000 [23:49<68:23:58, 4.04it/s, grad_norm=0.423, loss_final=0.475, loss_mean=0.827, loss_mean_cls=0.0693, proj_loss=-0.421][2026-03-26 12:54:52] Step: 5736, Training Logs: loss_final: 0.479239, loss_mean: 0.829912, proj_loss: -0.420142, loss_mean_cls: 0.069469, grad_norm: 0.366884 +Steps: 1%| | 5737/1000000 [23:49<68:22:42, 4.04it/s, grad_norm=0.367, loss_final=0.479, loss_mean=0.83, loss_mean_cls=0.0695, proj_loss=-0.42][2026-03-26 12:54:53] Step: 5737, Training Logs: loss_final: 0.470085, loss_mean: 0.816463, proj_loss: -0.416933, loss_mean_cls: 0.070555, grad_norm: 0.363543 +Steps: 1%| | 5738/1000000 [23:49<68:23:45, 4.04it/s, grad_norm=0.364, loss_final=0.47, loss_mean=0.816, loss_mean_cls=0.0706, proj_loss=-0.417][2026-03-26 12:54:53] Step: 5738, Training Logs: loss_final: 0.468979, loss_mean: 0.806105, proj_loss: -0.409090, loss_mean_cls: 0.071964, grad_norm: 0.430227 +Steps: 1%| | 5739/1000000 [23:50<68:24:58, 4.04it/s, grad_norm=0.43, loss_final=0.469, loss_mean=0.806, loss_mean_cls=0.072, proj_loss=-0.409][2026-03-26 12:54:53] Step: 5739, Training Logs: loss_final: 0.472727, loss_mean: 0.807982, proj_loss: -0.407804, loss_mean_cls: 0.072549, grad_norm: 0.357148 +Steps: 1%| | 5740/1000000 [23:50<68:27:16, 4.03it/s, grad_norm=0.357, loss_final=0.473, loss_mean=0.808, loss_mean_cls=0.0725, proj_loss=-0.408][2026-03-26 12:54:53] Step: 5740, Training Logs: loss_final: 0.485067, loss_mean: 0.825363, proj_loss: -0.411368, loss_mean_cls: 0.071073, grad_norm: 0.414918 +Steps: 1%| | 5741/1000000 [23:50<68:27:38, 4.03it/s, grad_norm=0.415, loss_final=0.485, loss_mean=0.825, loss_mean_cls=0.0711, proj_loss=-0.411][2026-03-26 12:54:54] Step: 5741, Training Logs: loss_final: 0.479657, loss_mean: 0.828667, proj_loss: -0.418445, loss_mean_cls: 0.069435, grad_norm: 0.258319 +Steps: 1%| | 5742/1000000 [23:50<68:26:31, 4.04it/s, grad_norm=0.258, loss_final=0.48, loss_mean=0.829, loss_mean_cls=0.0694, proj_loss=-0.418][2026-03-26 12:54:54] Step: 5742, Training Logs: loss_final: 0.479218, loss_mean: 0.824347, proj_loss: -0.415713, loss_mean_cls: 0.070584, grad_norm: 0.553326 +Steps: 1%| | 5743/1000000 [23:51<68:24:45, 4.04it/s, grad_norm=0.553, loss_final=0.479, loss_mean=0.824, loss_mean_cls=0.0706, proj_loss=-0.416][2026-03-26 12:54:54] Step: 5743, Training Logs: loss_final: 0.498146, loss_mean: 0.841990, proj_loss: -0.415145, loss_mean_cls: 0.071301, grad_norm: 0.409777 +Steps: 1%| | 5744/1000000 [23:51<68:21:37, 4.04it/s, grad_norm=0.41, loss_final=0.498, loss_mean=0.842, loss_mean_cls=0.0713, proj_loss=-0.415][2026-03-26 12:54:54] Step: 5744, Training Logs: loss_final: 0.462190, loss_mean: 0.809551, proj_loss: -0.417753, loss_mean_cls: 0.070392, grad_norm: 0.343720 +Steps: 1%| | 5745/1000000 [23:51<69:27:19, 3.98it/s, grad_norm=0.344, loss_final=0.462, loss_mean=0.81, loss_mean_cls=0.0704, proj_loss=-0.418][2026-03-26 12:54:55] Step: 5745, Training Logs: loss_final: 0.488628, loss_mean: 0.846181, proj_loss: -0.425836, loss_mean_cls: 0.068283, grad_norm: 0.399926 +Steps: 1%| | 5746/1000000 [23:51<69:09:01, 3.99it/s, grad_norm=0.4, loss_final=0.489, loss_mean=0.846, loss_mean_cls=0.0683, proj_loss=-0.426][2026-03-26 12:54:55] Step: 5746, Training Logs: loss_final: 0.484612, loss_mean: 0.816231, proj_loss: -0.404607, loss_mean_cls: 0.072989, grad_norm: 0.439869 +Steps: 1%| | 5747/1000000 [23:52<68:54:20, 4.01it/s, grad_norm=0.44, loss_final=0.485, loss_mean=0.816, loss_mean_cls=0.073, proj_loss=-0.405][2026-03-26 12:54:55] Step: 5747, Training Logs: loss_final: 0.480238, loss_mean: 0.811991, proj_loss: -0.405273, loss_mean_cls: 0.073520, grad_norm: 0.262421 +Steps: 1%| | 5748/1000000 [23:52<68:44:39, 4.02it/s, grad_norm=0.262, loss_final=0.48, loss_mean=0.812, loss_mean_cls=0.0735, proj_loss=-0.405][2026-03-26 12:54:55] Step: 5748, Training Logs: loss_final: 0.491480, loss_mean: 0.834803, proj_loss: -0.414940, loss_mean_cls: 0.071616, grad_norm: 0.368180 +Steps: 1%| | 5749/1000000 [23:52<68:39:50, 4.02it/s, grad_norm=0.368, loss_final=0.491, loss_mean=0.835, loss_mean_cls=0.0716, proj_loss=-0.415][2026-03-26 12:54:56] Step: 5749, Training Logs: loss_final: 0.478779, loss_mean: 0.824398, proj_loss: -0.417222, loss_mean_cls: 0.071602, grad_norm: 0.381998 +Steps: 1%| | 5750/1000000 [23:52<68:35:17, 4.03it/s, grad_norm=0.382, loss_final=0.479, loss_mean=0.824, loss_mean_cls=0.0716, proj_loss=-0.417][2026-03-26 12:54:56] Step: 5750, Training Logs: loss_final: 0.479799, loss_mean: 0.824605, proj_loss: -0.415505, loss_mean_cls: 0.070700, grad_norm: 0.246189 +Steps: 1%| | 5751/1000000 [23:53<68:30:49, 4.03it/s, grad_norm=0.246, loss_final=0.48, loss_mean=0.825, loss_mean_cls=0.0707, proj_loss=-0.416][2026-03-26 12:54:56] Step: 5751, Training Logs: loss_final: 0.485477, loss_mean: 0.835284, proj_loss: -0.419438, loss_mean_cls: 0.069631, grad_norm: 0.248394 +Steps: 1%| | 5752/1000000 [23:53<68:27:54, 4.03it/s, grad_norm=0.248, loss_final=0.485, loss_mean=0.835, loss_mean_cls=0.0696, proj_loss=-0.419][2026-03-26 12:54:56] Step: 5752, Training Logs: loss_final: 0.496536, loss_mean: 0.834825, proj_loss: -0.408900, loss_mean_cls: 0.070611, grad_norm: 0.227691 +Steps: 1%| | 5753/1000000 [23:53<68:29:08, 4.03it/s, grad_norm=0.228, loss_final=0.497, loss_mean=0.835, loss_mean_cls=0.0706, proj_loss=-0.409][2026-03-26 12:54:57] Step: 5753, Training Logs: loss_final: 0.484462, loss_mean: 0.836777, proj_loss: -0.422263, loss_mean_cls: 0.069948, grad_norm: 0.380650 +Steps: 1%| | 5754/1000000 [23:53<68:23:18, 4.04it/s, grad_norm=0.381, loss_final=0.484, loss_mean=0.837, loss_mean_cls=0.0699, proj_loss=-0.422][2026-03-26 12:54:57] Step: 5754, Training Logs: loss_final: 0.484495, loss_mean: 0.835079, proj_loss: -0.419597, loss_mean_cls: 0.069013, grad_norm: 0.304675 +Steps: 1%| | 5755/1000000 [23:54<68:23:43, 4.04it/s, grad_norm=0.305, loss_final=0.484, loss_mean=0.835, loss_mean_cls=0.069, proj_loss=-0.42][2026-03-26 12:54:57] Step: 5755, Training Logs: loss_final: 0.477324, loss_mean: 0.817996, proj_loss: -0.412386, loss_mean_cls: 0.071715, grad_norm: 0.248607 +Steps: 1%| | 5756/1000000 [23:54<68:20:29, 4.04it/s, grad_norm=0.249, loss_final=0.477, loss_mean=0.818, loss_mean_cls=0.0717, proj_loss=-0.412][2026-03-26 12:54:57] Step: 5756, Training Logs: loss_final: 0.466483, loss_mean: 0.811788, proj_loss: -0.415698, loss_mean_cls: 0.070393, grad_norm: 0.479749 +Steps: 1%| | 5757/1000000 [23:54<68:21:16, 4.04it/s, grad_norm=0.48, loss_final=0.466, loss_mean=0.812, loss_mean_cls=0.0704, proj_loss=-0.416][2026-03-26 12:54:58] Step: 5757, Training Logs: loss_final: 0.467168, loss_mean: 0.810568, proj_loss: -0.414520, loss_mean_cls: 0.071120, grad_norm: 0.353494 +Steps: 1%| | 5758/1000000 [23:54<68:21:00, 4.04it/s, grad_norm=0.353, loss_final=0.467, loss_mean=0.811, loss_mean_cls=0.0711, proj_loss=-0.415][2026-03-26 12:54:58] Step: 5758, Training Logs: loss_final: 0.483147, loss_mean: 0.829637, proj_loss: -0.416231, loss_mean_cls: 0.069741, grad_norm: 0.371536 +Steps: 1%| | 5759/1000000 [23:55<68:20:16, 4.04it/s, grad_norm=0.372, loss_final=0.483, loss_mean=0.83, loss_mean_cls=0.0697, proj_loss=-0.416][2026-03-26 12:54:58] Step: 5759, Training Logs: loss_final: 0.476800, loss_mean: 0.814227, proj_loss: -0.410133, loss_mean_cls: 0.072706, grad_norm: 0.315803 +Steps: 1%| | 5760/1000000 [23:55<68:22:48, 4.04it/s, grad_norm=0.316, loss_final=0.477, loss_mean=0.814, loss_mean_cls=0.0727, proj_loss=-0.41][2026-03-26 12:54:58] Step: 5760, Training Logs: loss_final: 0.477576, loss_mean: 0.810262, proj_loss: -0.405889, loss_mean_cls: 0.073203, grad_norm: 0.370755 +Steps: 1%| | 5761/1000000 [23:55<68:21:28, 4.04it/s, grad_norm=0.371, loss_final=0.478, loss_mean=0.81, loss_mean_cls=0.0732, proj_loss=-0.406][2026-03-26 12:54:59] Step: 5761, Training Logs: loss_final: 0.497688, loss_mean: 0.856077, proj_loss: -0.426474, loss_mean_cls: 0.068085, grad_norm: 0.261993 +Steps: 1%| | 5762/1000000 [23:55<68:22:45, 4.04it/s, grad_norm=0.262, loss_final=0.498, loss_mean=0.856, loss_mean_cls=0.0681, proj_loss=-0.426][2026-03-26 12:54:59] Step: 5762, Training Logs: loss_final: 0.481716, loss_mean: 0.835870, proj_loss: -0.422050, loss_mean_cls: 0.067895, grad_norm: 0.518301 +Steps: 1%| | 5763/1000000 [23:56<68:23:35, 4.04it/s, grad_norm=0.518, loss_final=0.482, loss_mean=0.836, loss_mean_cls=0.0679, proj_loss=-0.422][2026-03-26 12:54:59] Step: 5763, Training Logs: loss_final: 0.504293, loss_mean: 0.837958, proj_loss: -0.406579, loss_mean_cls: 0.072915, grad_norm: 0.449317 +Steps: 1%| | 5764/1000000 [23:56<68:20:23, 4.04it/s, grad_norm=0.449, loss_final=0.504, loss_mean=0.838, loss_mean_cls=0.0729, proj_loss=-0.407][2026-03-26 12:54:59] Step: 5764, Training Logs: loss_final: 0.485518, loss_mean: 0.834780, proj_loss: -0.418628, loss_mean_cls: 0.069367, grad_norm: 0.274923 +Steps: 1%| | 5765/1000000 [23:56<68:23:29, 4.04it/s, grad_norm=0.275, loss_final=0.486, loss_mean=0.835, loss_mean_cls=0.0694, proj_loss=-0.419][2026-03-26 12:55:00] Step: 5765, Training Logs: loss_final: 0.466920, loss_mean: 0.812914, proj_loss: -0.416201, loss_mean_cls: 0.070208, grad_norm: 0.343705 +Steps: 1%| | 5766/1000000 [23:56<68:22:31, 4.04it/s, grad_norm=0.344, loss_final=0.467, loss_mean=0.813, loss_mean_cls=0.0702, proj_loss=-0.416][2026-03-26 12:55:00] Step: 5766, Training Logs: loss_final: 0.474617, loss_mean: 0.820539, proj_loss: -0.416953, loss_mean_cls: 0.071030, grad_norm: 0.357188 +Steps: 1%| | 5767/1000000 [23:57<68:25:17, 4.04it/s, grad_norm=0.357, loss_final=0.475, loss_mean=0.821, loss_mean_cls=0.071, proj_loss=-0.417][2026-03-26 12:55:00] Step: 5767, Training Logs: loss_final: 0.482433, loss_mean: 0.811391, proj_loss: -0.401986, loss_mean_cls: 0.073029, grad_norm: 0.450110 +Steps: 1%| | 5768/1000000 [23:57<68:28:52, 4.03it/s, grad_norm=0.45, loss_final=0.482, loss_mean=0.811, loss_mean_cls=0.073, proj_loss=-0.402][2026-03-26 12:55:00] Step: 5768, Training Logs: loss_final: 0.476562, loss_mean: 0.820655, proj_loss: -0.414460, loss_mean_cls: 0.070367, grad_norm: 0.299869 +Steps: 1%| | 5769/1000000 [23:57<68:24:03, 4.04it/s, grad_norm=0.3, loss_final=0.477, loss_mean=0.821, loss_mean_cls=0.0704, proj_loss=-0.414][2026-03-26 12:55:01] Step: 5769, Training Logs: loss_final: 0.478953, loss_mean: 0.828504, proj_loss: -0.417788, loss_mean_cls: 0.068237, grad_norm: 0.260435 +Steps: 1%| | 5770/1000000 [23:57<68:30:09, 4.03it/s, grad_norm=0.26, loss_final=0.479, loss_mean=0.829, loss_mean_cls=0.0682, proj_loss=-0.418][2026-03-26 12:55:01] Step: 5770, Training Logs: loss_final: 0.471482, loss_mean: 0.817203, proj_loss: -0.415684, loss_mean_cls: 0.069963, grad_norm: 0.339139 +Steps: 1%| | 5771/1000000 [23:58<68:23:09, 4.04it/s, grad_norm=0.339, loss_final=0.471, loss_mean=0.817, loss_mean_cls=0.07, proj_loss=-0.416][2026-03-26 12:55:01] Step: 5771, Training Logs: loss_final: 0.493454, loss_mean: 0.833655, proj_loss: -0.411679, loss_mean_cls: 0.071478, grad_norm: 0.223285 +Steps: 1%| | 5772/1000000 [23:58<68:25:25, 4.04it/s, grad_norm=0.223, loss_final=0.493, loss_mean=0.834, loss_mean_cls=0.0715, proj_loss=-0.412][2026-03-26 12:55:01] Step: 5772, Training Logs: loss_final: 0.485689, loss_mean: 0.830247, proj_loss: -0.415568, loss_mean_cls: 0.071010, grad_norm: 0.299663 +Steps: 1%| | 5773/1000000 [23:58<68:25:02, 4.04it/s, grad_norm=0.3, loss_final=0.486, loss_mean=0.83, loss_mean_cls=0.071, proj_loss=-0.416][2026-03-26 12:55:01] Step: 5773, Training Logs: loss_final: 0.475227, loss_mean: 0.822768, proj_loss: -0.417564, loss_mean_cls: 0.070022, grad_norm: 0.540680 +Steps: 1%| | 5774/1000000 [23:58<68:25:51, 4.04it/s, grad_norm=0.541, loss_final=0.475, loss_mean=0.823, loss_mean_cls=0.07, proj_loss=-0.418][2026-03-26 12:55:02] Step: 5774, Training Logs: loss_final: 0.465130, loss_mean: 0.802621, proj_loss: -0.408995, loss_mean_cls: 0.071503, grad_norm: 0.284268 +Steps: 1%| | 5775/1000000 [23:59<68:23:04, 4.04it/s, grad_norm=0.284, loss_final=0.465, loss_mean=0.803, loss_mean_cls=0.0715, proj_loss=-0.409][2026-03-26 12:55:02] Step: 5775, Training Logs: loss_final: 0.475276, loss_mean: 0.820563, proj_loss: -0.415666, loss_mean_cls: 0.070380, grad_norm: 0.402171 +Steps: 1%| | 5776/1000000 [23:59<69:12:07, 3.99it/s, grad_norm=0.402, loss_final=0.475, loss_mean=0.821, loss_mean_cls=0.0704, proj_loss=-0.416][2026-03-26 12:55:02] Step: 5776, Training Logs: loss_final: 0.489614, loss_mean: 0.846619, proj_loss: -0.424678, loss_mean_cls: 0.067674, grad_norm: 0.474661 +Steps: 1%| | 5777/1000000 [23:59<68:56:06, 4.01it/s, grad_norm=0.475, loss_final=0.49, loss_mean=0.847, loss_mean_cls=0.0677, proj_loss=-0.425][2026-03-26 12:55:02] Step: 5777, Training Logs: loss_final: 0.465466, loss_mean: 0.822523, proj_loss: -0.425240, loss_mean_cls: 0.068183, grad_norm: 0.306616 +Steps: 1%| | 5778/1000000 [23:59<68:45:00, 4.02it/s, grad_norm=0.307, loss_final=0.465, loss_mean=0.823, loss_mean_cls=0.0682, proj_loss=-0.425][2026-03-26 12:55:03] Step: 5778, Training Logs: loss_final: 0.485971, loss_mean: 0.836411, proj_loss: -0.418570, loss_mean_cls: 0.068130, grad_norm: 0.443596 +Steps: 1%| | 5779/1000000 [24:00<68:36:34, 4.03it/s, grad_norm=0.444, loss_final=0.486, loss_mean=0.836, loss_mean_cls=0.0681, proj_loss=-0.419][2026-03-26 12:55:03] Step: 5779, Training Logs: loss_final: 0.493758, loss_mean: 0.840022, proj_loss: -0.415887, loss_mean_cls: 0.069622, grad_norm: 0.308148 +Steps: 1%| | 5780/1000000 [24:00<68:34:05, 4.03it/s, grad_norm=0.308, loss_final=0.494, loss_mean=0.84, loss_mean_cls=0.0696, proj_loss=-0.416][2026-03-26 12:55:03] Step: 5780, Training Logs: loss_final: 0.491748, loss_mean: 0.832749, proj_loss: -0.412435, loss_mean_cls: 0.071434, grad_norm: 0.523049 +Steps: 1%| | 5781/1000000 [24:00<68:27:21, 4.03it/s, grad_norm=0.523, loss_final=0.492, loss_mean=0.833, loss_mean_cls=0.0714, proj_loss=-0.412][2026-03-26 12:55:03] Step: 5781, Training Logs: loss_final: 0.497158, loss_mean: 0.839164, proj_loss: -0.413138, loss_mean_cls: 0.071132, grad_norm: 0.699142 +Steps: 1%| | 5782/1000000 [24:00<68:25:59, 4.04it/s, grad_norm=0.699, loss_final=0.497, loss_mean=0.839, loss_mean_cls=0.0711, proj_loss=-0.413][2026-03-26 12:55:04] Step: 5782, Training Logs: loss_final: 0.491756, loss_mean: 0.845573, proj_loss: -0.422280, loss_mean_cls: 0.068464, grad_norm: 0.293507 +Steps: 1%| | 5783/1000000 [24:01<68:24:01, 4.04it/s, grad_norm=0.294, loss_final=0.492, loss_mean=0.846, loss_mean_cls=0.0685, proj_loss=-0.422][2026-03-26 12:55:04] Step: 5783, Training Logs: loss_final: 0.482783, loss_mean: 0.824072, proj_loss: -0.412209, loss_mean_cls: 0.070920, grad_norm: 0.539441 +Steps: 1%| | 5784/1000000 [24:01<68:23:42, 4.04it/s, grad_norm=0.539, loss_final=0.483, loss_mean=0.824, loss_mean_cls=0.0709, proj_loss=-0.412][2026-03-26 12:55:04] Step: 5784, Training Logs: loss_final: 0.496964, loss_mean: 0.836172, proj_loss: -0.410505, loss_mean_cls: 0.071298, grad_norm: 0.492532 +Steps: 1%| | 5785/1000000 [24:01<68:22:26, 4.04it/s, grad_norm=0.493, loss_final=0.497, loss_mean=0.836, loss_mean_cls=0.0713, proj_loss=-0.411][2026-03-26 12:55:04] Step: 5785, Training Logs: loss_final: 0.463094, loss_mean: 0.792291, proj_loss: -0.402536, loss_mean_cls: 0.073339, grad_norm: 0.433171 +Steps: 1%| | 5786/1000000 [24:01<68:23:01, 4.04it/s, grad_norm=0.433, loss_final=0.463, loss_mean=0.792, loss_mean_cls=0.0733, proj_loss=-0.403][2026-03-26 12:55:05] Step: 5786, Training Logs: loss_final: 0.489851, loss_mean: 0.831317, proj_loss: -0.412710, loss_mean_cls: 0.071244, grad_norm: 0.461285 +Steps: 1%| | 5787/1000000 [24:02<68:23:47, 4.04it/s, grad_norm=0.461, loss_final=0.49, loss_mean=0.831, loss_mean_cls=0.0712, proj_loss=-0.413][2026-03-26 12:55:05] Step: 5787, Training Logs: loss_final: 0.507802, loss_mean: 0.845903, proj_loss: -0.409300, loss_mean_cls: 0.071199, grad_norm: 0.331389 +Steps: 1%| | 5788/1000000 [24:02<68:21:10, 4.04it/s, grad_norm=0.331, loss_final=0.508, loss_mean=0.846, loss_mean_cls=0.0712, proj_loss=-0.409][2026-03-26 12:55:05] Step: 5788, Training Logs: loss_final: 0.490010, loss_mean: 0.843074, proj_loss: -0.422271, loss_mean_cls: 0.069207, grad_norm: 0.393834 +Steps: 1%| | 5789/1000000 [24:02<68:18:15, 4.04it/s, grad_norm=0.394, loss_final=0.49, loss_mean=0.843, loss_mean_cls=0.0692, proj_loss=-0.422][2026-03-26 12:55:05] Step: 5789, Training Logs: loss_final: 0.478368, loss_mean: 0.824427, proj_loss: -0.416806, loss_mean_cls: 0.070747, grad_norm: 0.428799 +Steps: 1%| | 5790/1000000 [24:02<68:21:37, 4.04it/s, grad_norm=0.429, loss_final=0.478, loss_mean=0.824, loss_mean_cls=0.0707, proj_loss=-0.417][2026-03-26 12:55:06] Step: 5790, Training Logs: loss_final: 0.502732, loss_mean: 0.857689, proj_loss: -0.422496, loss_mean_cls: 0.067539, grad_norm: 0.337323 +Steps: 1%| | 5791/1000000 [24:03<68:19:24, 4.04it/s, grad_norm=0.337, loss_final=0.503, loss_mean=0.858, loss_mean_cls=0.0675, proj_loss=-0.422][2026-03-26 12:55:06] Step: 5791, Training Logs: loss_final: 0.479295, loss_mean: 0.825595, proj_loss: -0.416903, loss_mean_cls: 0.070604, grad_norm: 0.366825 +Steps: 1%| | 5792/1000000 [24:03<68:22:06, 4.04it/s, grad_norm=0.367, loss_final=0.479, loss_mean=0.826, loss_mean_cls=0.0706, proj_loss=-0.417][2026-03-26 12:55:06] Step: 5792, Training Logs: loss_final: 0.480112, loss_mean: 0.823721, proj_loss: -0.414944, loss_mean_cls: 0.071335, grad_norm: 0.330372 +Steps: 1%| | 5793/1000000 [24:03<68:21:40, 4.04it/s, grad_norm=0.33, loss_final=0.48, loss_mean=0.824, loss_mean_cls=0.0713, proj_loss=-0.415][2026-03-26 12:55:06] Step: 5793, Training Logs: loss_final: 0.472701, loss_mean: 0.811051, proj_loss: -0.409978, loss_mean_cls: 0.071629, grad_norm: 0.359043 +Steps: 1%| | 5794/1000000 [24:03<68:22:03, 4.04it/s, grad_norm=0.359, loss_final=0.473, loss_mean=0.811, loss_mean_cls=0.0716, proj_loss=-0.41][2026-03-26 12:55:07] Step: 5794, Training Logs: loss_final: 0.476248, loss_mean: 0.819883, proj_loss: -0.415255, loss_mean_cls: 0.071620, grad_norm: 0.310993 +Steps: 1%| | 5795/1000000 [24:04<68:24:20, 4.04it/s, grad_norm=0.311, loss_final=0.476, loss_mean=0.82, loss_mean_cls=0.0716, proj_loss=-0.415][2026-03-26 12:55:07] Step: 5795, Training Logs: loss_final: 0.468706, loss_mean: 0.817904, proj_loss: -0.418520, loss_mean_cls: 0.069322, grad_norm: 0.432833 +Steps: 1%| | 5796/1000000 [24:04<68:25:46, 4.04it/s, grad_norm=0.433, loss_final=0.469, loss_mean=0.818, loss_mean_cls=0.0693, proj_loss=-0.419][2026-03-26 12:55:07] Step: 5796, Training Logs: loss_final: 0.501674, loss_mean: 0.861715, proj_loss: -0.427208, loss_mean_cls: 0.067166, grad_norm: 0.497516 +Steps: 1%| | 5797/1000000 [24:04<68:28:31, 4.03it/s, grad_norm=0.498, loss_final=0.502, loss_mean=0.862, loss_mean_cls=0.0672, proj_loss=-0.427][2026-03-26 12:55:07] Step: 5797, Training Logs: loss_final: 0.486451, loss_mean: 0.824984, proj_loss: -0.410707, loss_mean_cls: 0.072174, grad_norm: 0.295747 +Steps: 1%| | 5798/1000000 [24:04<68:25:47, 4.04it/s, grad_norm=0.296, loss_final=0.486, loss_mean=0.825, loss_mean_cls=0.0722, proj_loss=-0.411][2026-03-26 12:55:08] Step: 5798, Training Logs: loss_final: 0.469006, loss_mean: 0.828066, proj_loss: -0.426211, loss_mean_cls: 0.067152, grad_norm: 0.331856 +Steps: 1%| | 5799/1000000 [24:05<68:23:38, 4.04it/s, grad_norm=0.332, loss_final=0.469, loss_mean=0.828, loss_mean_cls=0.0672, proj_loss=-0.426][2026-03-26 12:55:08] Step: 5799, Training Logs: loss_final: 0.480384, loss_mean: 0.821596, proj_loss: -0.412405, loss_mean_cls: 0.071193, grad_norm: 0.302897 +Steps: 1%| | 5800/1000000 [24:05<68:21:22, 4.04it/s, grad_norm=0.303, loss_final=0.48, loss_mean=0.822, loss_mean_cls=0.0712, proj_loss=-0.412][2026-03-26 12:55:08] Step: 5800, Training Logs: loss_final: 0.508393, loss_mean: 0.847536, proj_loss: -0.410323, loss_mean_cls: 0.071179, grad_norm: 0.332180 +Steps: 1%| | 5801/1000000 [24:05<68:20:39, 4.04it/s, grad_norm=0.332, loss_final=0.508, loss_mean=0.848, loss_mean_cls=0.0712, proj_loss=-0.41][2026-03-26 12:55:08] Step: 5801, Training Logs: loss_final: 0.480841, loss_mean: 0.814221, proj_loss: -0.406598, loss_mean_cls: 0.073218, grad_norm: 0.311441 +Steps: 1%| | 5802/1000000 [24:05<68:22:30, 4.04it/s, grad_norm=0.311, loss_final=0.481, loss_mean=0.814, loss_mean_cls=0.0732, proj_loss=-0.407][2026-03-26 12:55:09] Step: 5802, Training Logs: loss_final: 0.479204, loss_mean: 0.810777, proj_loss: -0.403993, loss_mean_cls: 0.072420, grad_norm: 0.371049 +Steps: 1%| | 5803/1000000 [24:06<68:23:45, 4.04it/s, grad_norm=0.371, loss_final=0.479, loss_mean=0.811, loss_mean_cls=0.0724, proj_loss=-0.404][2026-03-26 12:55:09] Step: 5803, Training Logs: loss_final: 0.505772, loss_mean: 0.847280, proj_loss: -0.412040, loss_mean_cls: 0.070531, grad_norm: 0.422044 +Steps: 1%| | 5804/1000000 [24:06<68:24:26, 4.04it/s, grad_norm=0.422, loss_final=0.506, loss_mean=0.847, loss_mean_cls=0.0705, proj_loss=-0.412][2026-03-26 12:55:09] Step: 5804, Training Logs: loss_final: 0.490836, loss_mean: 0.843295, proj_loss: -0.420756, loss_mean_cls: 0.068297, grad_norm: 0.342099 +Steps: 1%| | 5805/1000000 [24:06<68:19:13, 4.04it/s, grad_norm=0.342, loss_final=0.491, loss_mean=0.843, loss_mean_cls=0.0683, proj_loss=-0.421][2026-03-26 12:55:09] Step: 5805, Training Logs: loss_final: 0.496425, loss_mean: 0.838818, proj_loss: -0.414301, loss_mean_cls: 0.071908, grad_norm: 0.330756 +Steps: 1%| | 5806/1000000 [24:06<68:17:49, 4.04it/s, grad_norm=0.331, loss_final=0.496, loss_mean=0.839, loss_mean_cls=0.0719, proj_loss=-0.414][2026-03-26 12:55:10] Step: 5806, Training Logs: loss_final: 0.471182, loss_mean: 0.820716, proj_loss: -0.418262, loss_mean_cls: 0.068728, grad_norm: 0.333185 +Steps: 1%| | 5807/1000000 [24:07<68:16:32, 4.04it/s, grad_norm=0.333, loss_final=0.471, loss_mean=0.821, loss_mean_cls=0.0687, proj_loss=-0.418][2026-03-26 12:55:10] Step: 5807, Training Logs: loss_final: 0.478285, loss_mean: 0.824460, proj_loss: -0.416602, loss_mean_cls: 0.070426, grad_norm: 0.482872 +Steps: 1%| | 5808/1000000 [24:07<68:18:43, 4.04it/s, grad_norm=0.483, loss_final=0.478, loss_mean=0.824, loss_mean_cls=0.0704, proj_loss=-0.417][2026-03-26 12:55:10] Step: 5808, Training Logs: loss_final: 0.459156, loss_mean: 0.802710, proj_loss: -0.414506, loss_mean_cls: 0.070953, grad_norm: 0.252715 +Steps: 1%| | 5809/1000000 [24:07<68:20:05, 4.04it/s, grad_norm=0.253, loss_final=0.459, loss_mean=0.803, loss_mean_cls=0.071, proj_loss=-0.415][2026-03-26 12:55:10] Step: 5809, Training Logs: loss_final: 0.470860, loss_mean: 0.798488, proj_loss: -0.401456, loss_mean_cls: 0.073828, grad_norm: 0.350981 +Steps: 1%| | 5810/1000000 [24:07<68:23:00, 4.04it/s, grad_norm=0.351, loss_final=0.471, loss_mean=0.798, loss_mean_cls=0.0738, proj_loss=-0.401][2026-03-26 12:55:11] Step: 5810, Training Logs: loss_final: 0.469790, loss_mean: 0.806515, proj_loss: -0.408637, loss_mean_cls: 0.071911, grad_norm: 0.284011 +Steps: 1%| | 5811/1000000 [24:08<68:21:41, 4.04it/s, grad_norm=0.284, loss_final=0.47, loss_mean=0.807, loss_mean_cls=0.0719, proj_loss=-0.409][2026-03-26 12:55:11] Step: 5811, Training Logs: loss_final: 0.477948, loss_mean: 0.834647, proj_loss: -0.424313, loss_mean_cls: 0.067615, grad_norm: 0.722012 +Steps: 1%| | 5812/1000000 [24:08<68:22:55, 4.04it/s, grad_norm=0.722, loss_final=0.478, loss_mean=0.835, loss_mean_cls=0.0676, proj_loss=-0.424][2026-03-26 12:55:11] Step: 5812, Training Logs: loss_final: 0.492642, loss_mean: 0.837417, proj_loss: -0.414688, loss_mean_cls: 0.069914, grad_norm: 0.443043 +Steps: 1%| | 5813/1000000 [24:08<68:23:42, 4.04it/s, grad_norm=0.443, loss_final=0.493, loss_mean=0.837, loss_mean_cls=0.0699, proj_loss=-0.415][2026-03-26 12:55:11] Step: 5813, Training Logs: loss_final: 0.480360, loss_mean: 0.827735, proj_loss: -0.417820, loss_mean_cls: 0.070445, grad_norm: 0.402171 +Steps: 1%| | 5814/1000000 [24:08<68:53:04, 4.01it/s, grad_norm=0.402, loss_final=0.48, loss_mean=0.828, loss_mean_cls=0.0704, proj_loss=-0.418][2026-03-26 12:55:12] Step: 5814, Training Logs: loss_final: 0.477394, loss_mean: 0.827365, proj_loss: -0.419665, loss_mean_cls: 0.069695, grad_norm: 0.451015 +Steps: 1%| | 5815/1000000 [24:09<68:28:10, 4.03it/s, grad_norm=0.451, loss_final=0.477, loss_mean=0.827, loss_mean_cls=0.0697, proj_loss=-0.42][2026-03-26 12:55:12] Step: 5815, Training Logs: loss_final: 0.487294, loss_mean: 0.825770, proj_loss: -0.410211, loss_mean_cls: 0.071735, grad_norm: 0.285665 +Steps: 1%| | 5816/1000000 [24:09<68:29:03, 4.03it/s, grad_norm=0.286, loss_final=0.487, loss_mean=0.826, loss_mean_cls=0.0717, proj_loss=-0.41][2026-03-26 12:55:12] Step: 5816, Training Logs: loss_final: 0.475865, loss_mean: 0.801877, proj_loss: -0.400013, loss_mean_cls: 0.074001, grad_norm: 0.612811 +Steps: 1%| | 5817/1000000 [24:09<68:29:06, 4.03it/s, grad_norm=0.613, loss_final=0.476, loss_mean=0.802, loss_mean_cls=0.074, proj_loss=-0.4][2026-03-26 12:55:12] Step: 5817, Training Logs: loss_final: 0.493465, loss_mean: 0.823847, proj_loss: -0.402881, loss_mean_cls: 0.072500, grad_norm: 0.405165 +Steps: 1%| | 5818/1000000 [24:09<68:30:25, 4.03it/s, grad_norm=0.405, loss_final=0.493, loss_mean=0.824, loss_mean_cls=0.0725, proj_loss=-0.403][2026-03-26 12:55:13] Step: 5818, Training Logs: loss_final: 0.482175, loss_mean: 0.818863, proj_loss: -0.408540, loss_mean_cls: 0.071852, grad_norm: 0.375028 +Steps: 1%| | 5819/1000000 [24:10<68:26:42, 4.03it/s, grad_norm=0.375, loss_final=0.482, loss_mean=0.819, loss_mean_cls=0.0719, proj_loss=-0.409][2026-03-26 12:55:13] Step: 5819, Training Logs: loss_final: 0.476647, loss_mean: 0.829580, proj_loss: -0.420741, loss_mean_cls: 0.067809, grad_norm: 0.221113 +Steps: 1%| | 5820/1000000 [24:10<68:20:11, 4.04it/s, grad_norm=0.221, loss_final=0.477, loss_mean=0.83, loss_mean_cls=0.0678, proj_loss=-0.421][2026-03-26 12:55:13] Step: 5820, Training Logs: loss_final: 0.482816, loss_mean: 0.831008, proj_loss: -0.417883, loss_mean_cls: 0.069690, grad_norm: 0.494284 +Steps: 1%| | 5821/1000000 [24:10<68:17:14, 4.04it/s, grad_norm=0.494, loss_final=0.483, loss_mean=0.831, loss_mean_cls=0.0697, proj_loss=-0.418][2026-03-26 12:55:13] Step: 5821, Training Logs: loss_final: 0.476117, loss_mean: 0.818031, proj_loss: -0.412680, loss_mean_cls: 0.070766, grad_norm: 0.301519 +Steps: 1%| | 5822/1000000 [24:10<68:19:36, 4.04it/s, grad_norm=0.302, loss_final=0.476, loss_mean=0.818, loss_mean_cls=0.0708, proj_loss=-0.413][2026-03-26 12:55:14] Step: 5822, Training Logs: loss_final: 0.482022, loss_mean: 0.816653, proj_loss: -0.406879, loss_mean_cls: 0.072248, grad_norm: 0.323795 +Steps: 1%| | 5823/1000000 [24:10<68:20:16, 4.04it/s, grad_norm=0.324, loss_final=0.482, loss_mean=0.817, loss_mean_cls=0.0722, proj_loss=-0.407][2026-03-26 12:55:14] Step: 5823, Training Logs: loss_final: 0.475415, loss_mean: 0.805818, proj_loss: -0.403003, loss_mean_cls: 0.072600, grad_norm: 0.458986 +Steps: 1%| | 5824/1000000 [24:11<68:21:42, 4.04it/s, grad_norm=0.459, loss_final=0.475, loss_mean=0.806, loss_mean_cls=0.0726, proj_loss=-0.403][2026-03-26 12:55:14] Step: 5824, Training Logs: loss_final: 0.475677, loss_mean: 0.813892, proj_loss: -0.410403, loss_mean_cls: 0.072189, grad_norm: 0.265886 +Steps: 1%| | 5825/1000000 [24:11<68:21:14, 4.04it/s, grad_norm=0.266, loss_final=0.476, loss_mean=0.814, loss_mean_cls=0.0722, proj_loss=-0.41][2026-03-26 12:55:14] Step: 5825, Training Logs: loss_final: 0.489216, loss_mean: 0.831205, proj_loss: -0.412269, loss_mean_cls: 0.070281, grad_norm: 0.394489 +Steps: 1%| | 5826/1000000 [24:11<68:21:01, 4.04it/s, grad_norm=0.394, loss_final=0.489, loss_mean=0.831, loss_mean_cls=0.0703, proj_loss=-0.412][2026-03-26 12:55:15] Step: 5826, Training Logs: loss_final: 0.497052, loss_mean: 0.841359, proj_loss: -0.415041, loss_mean_cls: 0.070735, grad_norm: 0.308182 +Steps: 1%| | 5827/1000000 [24:11<68:18:27, 4.04it/s, grad_norm=0.308, loss_final=0.497, loss_mean=0.841, loss_mean_cls=0.0707, proj_loss=-0.415][2026-03-26 12:55:15] Step: 5827, Training Logs: loss_final: 0.476355, loss_mean: 0.816256, proj_loss: -0.411620, loss_mean_cls: 0.071719, grad_norm: 0.570379 +Steps: 1%| | 5828/1000000 [24:12<68:19:59, 4.04it/s, grad_norm=0.57, loss_final=0.476, loss_mean=0.816, loss_mean_cls=0.0717, proj_loss=-0.412][2026-03-26 12:55:15] Step: 5828, Training Logs: loss_final: 0.484049, loss_mean: 0.827561, proj_loss: -0.412539, loss_mean_cls: 0.069028, grad_norm: 0.304770 +Steps: 1%| | 5829/1000000 [24:12<68:20:21, 4.04it/s, grad_norm=0.305, loss_final=0.484, loss_mean=0.828, loss_mean_cls=0.069, proj_loss=-0.413][2026-03-26 12:55:15] Step: 5829, Training Logs: loss_final: 0.483986, loss_mean: 0.825874, proj_loss: -0.413042, loss_mean_cls: 0.071154, grad_norm: 0.316537 +Steps: 1%| | 5830/1000000 [24:12<68:20:05, 4.04it/s, grad_norm=0.317, loss_final=0.484, loss_mean=0.826, loss_mean_cls=0.0712, proj_loss=-0.413][2026-03-26 12:55:16] Step: 5830, Training Logs: loss_final: 0.465042, loss_mean: 0.795562, proj_loss: -0.403141, loss_mean_cls: 0.072622, grad_norm: 0.475834 +Steps: 1%| | 5831/1000000 [24:12<68:19:30, 4.04it/s, grad_norm=0.476, loss_final=0.465, loss_mean=0.796, loss_mean_cls=0.0726, proj_loss=-0.403][2026-03-26 12:55:16] Step: 5831, Training Logs: loss_final: 0.475947, loss_mean: 0.829718, proj_loss: -0.422113, loss_mean_cls: 0.068342, grad_norm: 0.290193 +Steps: 1%| | 5832/1000000 [24:13<68:26:22, 4.04it/s, grad_norm=0.29, loss_final=0.476, loss_mean=0.83, loss_mean_cls=0.0683, proj_loss=-0.422][2026-03-26 12:55:16] Step: 5832, Training Logs: loss_final: 0.491442, loss_mean: 0.829153, proj_loss: -0.409513, loss_mean_cls: 0.071801, grad_norm: 0.501129 +Steps: 1%| | 5833/1000000 [24:13<68:22:58, 4.04it/s, grad_norm=0.501, loss_final=0.491, loss_mean=0.829, loss_mean_cls=0.0718, proj_loss=-0.41][2026-03-26 12:55:16] Step: 5833, Training Logs: loss_final: 0.482785, loss_mean: 0.831254, proj_loss: -0.417405, loss_mean_cls: 0.068936, grad_norm: 0.266619 +Steps: 1%| | 5834/1000000 [24:13<68:23:44, 4.04it/s, grad_norm=0.267, loss_final=0.483, loss_mean=0.831, loss_mean_cls=0.0689, proj_loss=-0.417][2026-03-26 12:55:17] Step: 5834, Training Logs: loss_final: 0.484700, loss_mean: 0.831971, proj_loss: -0.416900, loss_mean_cls: 0.069629, grad_norm: 0.400282 +Steps: 1%| | 5835/1000000 [24:13<68:26:23, 4.04it/s, grad_norm=0.4, loss_final=0.485, loss_mean=0.832, loss_mean_cls=0.0696, proj_loss=-0.417][2026-03-26 12:55:17] Step: 5835, Training Logs: loss_final: 0.499822, loss_mean: 0.845335, proj_loss: -0.415892, loss_mean_cls: 0.070379, grad_norm: 0.289539 +Steps: 1%| | 5836/1000000 [24:14<68:18:31, 4.04it/s, grad_norm=0.29, loss_final=0.5, loss_mean=0.845, loss_mean_cls=0.0704, proj_loss=-0.416][2026-03-26 12:55:17] Step: 5836, Training Logs: loss_final: 0.475685, loss_mean: 0.823126, proj_loss: -0.416757, loss_mean_cls: 0.069315, grad_norm: 0.402660 +Steps: 1%| | 5837/1000000 [24:14<68:17:39, 4.04it/s, grad_norm=0.403, loss_final=0.476, loss_mean=0.823, loss_mean_cls=0.0693, proj_loss=-0.417][2026-03-26 12:55:17] Step: 5837, Training Logs: loss_final: 0.484782, loss_mean: 0.827126, proj_loss: -0.412852, loss_mean_cls: 0.070508, grad_norm: 0.416340 +Steps: 1%| | 5838/1000000 [24:14<68:18:52, 4.04it/s, grad_norm=0.416, loss_final=0.485, loss_mean=0.827, loss_mean_cls=0.0705, proj_loss=-0.413][2026-03-26 12:55:18] Step: 5838, Training Logs: loss_final: 0.476428, loss_mean: 0.818936, proj_loss: -0.412844, loss_mean_cls: 0.070336, grad_norm: 0.224884 +Steps: 1%| | 5839/1000000 [24:14<68:20:04, 4.04it/s, grad_norm=0.225, loss_final=0.476, loss_mean=0.819, loss_mean_cls=0.0703, proj_loss=-0.413][2026-03-26 12:55:18] Step: 5839, Training Logs: loss_final: 0.475449, loss_mean: 0.819377, proj_loss: -0.414548, loss_mean_cls: 0.070620, grad_norm: 0.265617 +Steps: 1%| | 5840/1000000 [24:15<68:20:39, 4.04it/s, grad_norm=0.266, loss_final=0.475, loss_mean=0.819, loss_mean_cls=0.0706, proj_loss=-0.415][2026-03-26 12:55:18] Step: 5840, Training Logs: loss_final: 0.470082, loss_mean: 0.808214, proj_loss: -0.409970, loss_mean_cls: 0.071839, grad_norm: 0.239106 +Steps: 1%| | 5841/1000000 [24:15<68:23:38, 4.04it/s, grad_norm=0.239, loss_final=0.47, loss_mean=0.808, loss_mean_cls=0.0718, proj_loss=-0.41][2026-03-26 12:55:18] Step: 5841, Training Logs: loss_final: 0.478765, loss_mean: 0.827149, proj_loss: -0.417145, loss_mean_cls: 0.068761, grad_norm: 0.278838 +Steps: 1%| | 5842/1000000 [24:15<68:20:49, 4.04it/s, grad_norm=0.279, loss_final=0.479, loss_mean=0.827, loss_mean_cls=0.0688, proj_loss=-0.417][2026-03-26 12:55:19] Step: 5842, Training Logs: loss_final: 0.456465, loss_mean: 0.808179, proj_loss: -0.420826, loss_mean_cls: 0.069111, grad_norm: 0.274480 +Steps: 1%| | 5843/1000000 [24:15<68:22:01, 4.04it/s, grad_norm=0.274, loss_final=0.456, loss_mean=0.808, loss_mean_cls=0.0691, proj_loss=-0.421][2026-03-26 12:55:19] Step: 5843, Training Logs: loss_final: 0.491867, loss_mean: 0.827810, proj_loss: -0.407435, loss_mean_cls: 0.071492, grad_norm: 0.323008 +Steps: 1%| | 5844/1000000 [24:16<68:17:48, 4.04it/s, grad_norm=0.323, loss_final=0.492, loss_mean=0.828, loss_mean_cls=0.0715, proj_loss=-0.407][2026-03-26 12:55:19] Step: 5844, Training Logs: loss_final: 0.477295, loss_mean: 0.829076, proj_loss: -0.421162, loss_mean_cls: 0.069381, grad_norm: 0.255845 +Steps: 1%| | 5845/1000000 [24:16<68:18:56, 4.04it/s, grad_norm=0.256, loss_final=0.477, loss_mean=0.829, loss_mean_cls=0.0694, proj_loss=-0.421][2026-03-26 12:55:19] Step: 5845, Training Logs: loss_final: 0.477256, loss_mean: 0.826445, proj_loss: -0.418529, loss_mean_cls: 0.069340, grad_norm: 0.476532 +Steps: 1%| | 5846/1000000 [24:16<68:15:26, 4.05it/s, grad_norm=0.477, loss_final=0.477, loss_mean=0.826, loss_mean_cls=0.0693, proj_loss=-0.419][2026-03-26 12:55:20] Step: 5846, Training Logs: loss_final: 0.491272, loss_mean: 0.835977, proj_loss: -0.415089, loss_mean_cls: 0.070384, grad_norm: 0.269249 +Steps: 1%| | 5847/1000000 [24:16<68:16:28, 4.04it/s, grad_norm=0.269, loss_final=0.491, loss_mean=0.836, loss_mean_cls=0.0704, proj_loss=-0.415][2026-03-26 12:55:20] Step: 5847, Training Logs: loss_final: 0.470416, loss_mean: 0.814268, proj_loss: -0.415015, loss_mean_cls: 0.071164, grad_norm: 0.331194 +Steps: 1%| | 5848/1000000 [24:17<68:18:52, 4.04it/s, grad_norm=0.331, loss_final=0.47, loss_mean=0.814, loss_mean_cls=0.0712, proj_loss=-0.415][2026-03-26 12:55:20] Step: 5848, Training Logs: loss_final: 0.500482, loss_mean: 0.838273, proj_loss: -0.408944, loss_mean_cls: 0.071152, grad_norm: 0.295002 +Steps: 1%| | 5849/1000000 [24:17<68:20:59, 4.04it/s, grad_norm=0.295, loss_final=0.5, loss_mean=0.838, loss_mean_cls=0.0712, proj_loss=-0.409][2026-03-26 12:55:20] Step: 5849, Training Logs: loss_final: 0.483396, loss_mean: 0.826190, proj_loss: -0.414001, loss_mean_cls: 0.071207, grad_norm: 0.377985 +Steps: 1%| | 5850/1000000 [24:17<72:23:10, 3.81it/s, grad_norm=0.378, loss_final=0.483, loss_mean=0.826, loss_mean_cls=0.0712, proj_loss=-0.414][2026-03-26 12:55:21] Step: 5850, Training Logs: loss_final: 0.476032, loss_mean: 0.812190, proj_loss: -0.407830, loss_mean_cls: 0.071672, grad_norm: 0.507662 +Steps: 1%| | 5851/1000000 [24:17<71:08:55, 3.88it/s, grad_norm=0.508, loss_final=0.476, loss_mean=0.812, loss_mean_cls=0.0717, proj_loss=-0.408][2026-03-26 12:55:21] Step: 5851, Training Logs: loss_final: 0.493794, loss_mean: 0.831643, proj_loss: -0.408781, loss_mean_cls: 0.070932, grad_norm: 0.310047 +Steps: 1%| | 5852/1000000 [24:18<70:19:57, 3.93it/s, grad_norm=0.31, loss_final=0.494, loss_mean=0.832, loss_mean_cls=0.0709, proj_loss=-0.409][2026-03-26 12:55:21] Step: 5852, Training Logs: loss_final: 0.468183, loss_mean: 0.827114, proj_loss: -0.426445, loss_mean_cls: 0.067515, grad_norm: 0.386860 +Steps: 1%| | 5853/1000000 [24:18<69:44:00, 3.96it/s, grad_norm=0.387, loss_final=0.468, loss_mean=0.827, loss_mean_cls=0.0675, proj_loss=-0.426][2026-03-26 12:55:21] Step: 5853, Training Logs: loss_final: 0.486197, loss_mean: 0.822900, proj_loss: -0.408861, loss_mean_cls: 0.072157, grad_norm: 0.296525 +Steps: 1%| | 5854/1000000 [24:18<69:20:46, 3.98it/s, grad_norm=0.297, loss_final=0.486, loss_mean=0.823, loss_mean_cls=0.0722, proj_loss=-0.409][2026-03-26 12:55:22] Step: 5854, Training Logs: loss_final: 0.474613, loss_mean: 0.825612, proj_loss: -0.420409, loss_mean_cls: 0.069411, grad_norm: 0.321070 +Steps: 1%| | 5855/1000000 [24:18<69:02:58, 4.00it/s, grad_norm=0.321, loss_final=0.475, loss_mean=0.826, loss_mean_cls=0.0694, proj_loss=-0.42][2026-03-26 12:55:22] Step: 5855, Training Logs: loss_final: 0.478154, loss_mean: 0.827744, proj_loss: -0.418717, loss_mean_cls: 0.069126, grad_norm: 0.449136 +Steps: 1%| | 5856/1000000 [24:19<68:51:15, 4.01it/s, grad_norm=0.449, loss_final=0.478, loss_mean=0.828, loss_mean_cls=0.0691, proj_loss=-0.419][2026-03-26 12:55:22] Step: 5856, Training Logs: loss_final: 0.490753, loss_mean: 0.848092, proj_loss: -0.425155, loss_mean_cls: 0.067815, grad_norm: 0.324753 +Steps: 1%| | 5857/1000000 [24:19<68:42:02, 4.02it/s, grad_norm=0.325, loss_final=0.491, loss_mean=0.848, loss_mean_cls=0.0678, proj_loss=-0.425][2026-03-26 12:55:22] Step: 5857, Training Logs: loss_final: 0.477597, loss_mean: 0.819465, proj_loss: -0.412981, loss_mean_cls: 0.071113, grad_norm: 0.430860 +Steps: 1%| | 5858/1000000 [24:19<68:33:53, 4.03it/s, grad_norm=0.431, loss_final=0.478, loss_mean=0.819, loss_mean_cls=0.0711, proj_loss=-0.413][2026-03-26 12:55:23] Step: 5858, Training Logs: loss_final: 0.466198, loss_mean: 0.813015, proj_loss: -0.417427, loss_mean_cls: 0.070611, grad_norm: 0.274449 +Steps: 1%| | 5859/1000000 [24:19<68:28:21, 4.03it/s, grad_norm=0.274, loss_final=0.466, loss_mean=0.813, loss_mean_cls=0.0706, proj_loss=-0.417][2026-03-26 12:55:23] Step: 5859, Training Logs: loss_final: 0.475724, loss_mean: 0.804533, proj_loss: -0.402046, loss_mean_cls: 0.073236, grad_norm: 0.306587 +Steps: 1%| | 5860/1000000 [24:20<68:25:26, 4.04it/s, grad_norm=0.307, loss_final=0.476, loss_mean=0.805, loss_mean_cls=0.0732, proj_loss=-0.402][2026-03-26 12:55:23] Step: 5860, Training Logs: loss_final: 0.476222, loss_mean: 0.821631, proj_loss: -0.415408, loss_mean_cls: 0.069999, grad_norm: 0.338695 +Steps: 1%| | 5861/1000000 [24:20<68:23:52, 4.04it/s, grad_norm=0.339, loss_final=0.476, loss_mean=0.822, loss_mean_cls=0.07, proj_loss=-0.415][2026-03-26 12:55:23] Step: 5861, Training Logs: loss_final: 0.472702, loss_mean: 0.813781, proj_loss: -0.412322, loss_mean_cls: 0.071243, grad_norm: 0.408317 +Steps: 1%| | 5862/1000000 [24:20<68:21:34, 4.04it/s, grad_norm=0.408, loss_final=0.473, loss_mean=0.814, loss_mean_cls=0.0712, proj_loss=-0.412][2026-03-26 12:55:24] Step: 5862, Training Logs: loss_final: 0.492064, loss_mean: 0.829533, proj_loss: -0.408986, loss_mean_cls: 0.071517, grad_norm: 0.304152 +Steps: 1%| | 5863/1000000 [24:20<68:21:39, 4.04it/s, grad_norm=0.304, loss_final=0.492, loss_mean=0.83, loss_mean_cls=0.0715, proj_loss=-0.409][2026-03-26 12:55:24] Step: 5863, Training Logs: loss_final: 0.475062, loss_mean: 0.815472, proj_loss: -0.411540, loss_mean_cls: 0.071130, grad_norm: 0.444155 +Steps: 1%| | 5864/1000000 [24:21<68:24:04, 4.04it/s, grad_norm=0.444, loss_final=0.475, loss_mean=0.815, loss_mean_cls=0.0711, proj_loss=-0.412][2026-03-26 12:55:24] Step: 5864, Training Logs: loss_final: 0.480576, loss_mean: 0.827938, proj_loss: -0.416865, loss_mean_cls: 0.069502, grad_norm: 0.438790 +Steps: 1%| | 5865/1000000 [24:21<68:24:43, 4.04it/s, grad_norm=0.439, loss_final=0.481, loss_mean=0.828, loss_mean_cls=0.0695, proj_loss=-0.417][2026-03-26 12:55:24] Step: 5865, Training Logs: loss_final: 0.472834, loss_mean: 0.825320, proj_loss: -0.420895, loss_mean_cls: 0.068409, grad_norm: 0.509464 +Steps: 1%| | 5866/1000000 [24:21<68:21:24, 4.04it/s, grad_norm=0.509, loss_final=0.473, loss_mean=0.825, loss_mean_cls=0.0684, proj_loss=-0.421][2026-03-26 12:55:25] Step: 5866, Training Logs: loss_final: 0.473830, loss_mean: 0.826449, proj_loss: -0.420879, loss_mean_cls: 0.068260, grad_norm: 0.466931 +Steps: 1%| | 5867/1000000 [24:21<68:20:24, 4.04it/s, grad_norm=0.467, loss_final=0.474, loss_mean=0.826, loss_mean_cls=0.0683, proj_loss=-0.421][2026-03-26 12:55:25] Step: 5867, Training Logs: loss_final: 0.483445, loss_mean: 0.820508, proj_loss: -0.409720, loss_mean_cls: 0.072656, grad_norm: 0.434090 +Steps: 1%| | 5868/1000000 [24:22<68:18:52, 4.04it/s, grad_norm=0.434, loss_final=0.483, loss_mean=0.821, loss_mean_cls=0.0727, proj_loss=-0.41][2026-03-26 12:55:25] Step: 5868, Training Logs: loss_final: 0.484640, loss_mean: 0.837902, proj_loss: -0.421204, loss_mean_cls: 0.067942, grad_norm: 0.330559 +Steps: 1%| | 5869/1000000 [24:22<68:17:14, 4.04it/s, grad_norm=0.331, loss_final=0.485, loss_mean=0.838, loss_mean_cls=0.0679, proj_loss=-0.421][2026-03-26 12:55:25] Step: 5869, Training Logs: loss_final: 0.475277, loss_mean: 0.818926, proj_loss: -0.414668, loss_mean_cls: 0.071019, grad_norm: 0.380283 +Steps: 1%| | 5870/1000000 [24:22<68:17:06, 4.04it/s, grad_norm=0.38, loss_final=0.475, loss_mean=0.819, loss_mean_cls=0.071, proj_loss=-0.415][2026-03-26 12:55:26] Step: 5870, Training Logs: loss_final: 0.507289, loss_mean: 0.862774, proj_loss: -0.422124, loss_mean_cls: 0.066638, grad_norm: 0.243959 +Steps: 1%| | 5871/1000000 [24:22<68:17:44, 4.04it/s, grad_norm=0.244, loss_final=0.507, loss_mean=0.863, loss_mean_cls=0.0666, proj_loss=-0.422][2026-03-26 12:55:26] Step: 5871, Training Logs: loss_final: 0.482838, loss_mean: 0.829475, proj_loss: -0.416545, loss_mean_cls: 0.069908, grad_norm: 0.435543 +Steps: 1%| | 5872/1000000 [24:23<68:20:09, 4.04it/s, grad_norm=0.436, loss_final=0.483, loss_mean=0.829, loss_mean_cls=0.0699, proj_loss=-0.417][2026-03-26 12:55:26] Step: 5872, Training Logs: loss_final: 0.464936, loss_mean: 0.816880, proj_loss: -0.420356, loss_mean_cls: 0.068413, grad_norm: 0.362848 +Steps: 1%| | 5873/1000000 [24:23<68:20:10, 4.04it/s, grad_norm=0.363, loss_final=0.465, loss_mean=0.817, loss_mean_cls=0.0684, proj_loss=-0.42][2026-03-26 12:55:26] Step: 5873, Training Logs: loss_final: 0.482261, loss_mean: 0.838532, proj_loss: -0.423783, loss_mean_cls: 0.067512, grad_norm: 0.246713 +Steps: 1%| | 5874/1000000 [24:23<68:20:41, 4.04it/s, grad_norm=0.247, loss_final=0.482, loss_mean=0.839, loss_mean_cls=0.0675, proj_loss=-0.424][2026-03-26 12:55:27] Step: 5874, Training Logs: loss_final: 0.484999, loss_mean: 0.838905, proj_loss: -0.422171, loss_mean_cls: 0.068265, grad_norm: 0.341991 +Steps: 1%| | 5875/1000000 [24:23<68:21:42, 4.04it/s, grad_norm=0.342, loss_final=0.485, loss_mean=0.839, loss_mean_cls=0.0683, proj_loss=-0.422][2026-03-26 12:55:27] Step: 5875, Training Logs: loss_final: 0.484129, loss_mean: 0.825442, proj_loss: -0.410861, loss_mean_cls: 0.069548, grad_norm: 0.324525 +Steps: 1%| | 5876/1000000 [24:24<68:23:47, 4.04it/s, grad_norm=0.325, loss_final=0.484, loss_mean=0.825, loss_mean_cls=0.0695, proj_loss=-0.411][2026-03-26 12:55:27] Step: 5876, Training Logs: loss_final: 0.489200, loss_mean: 0.828187, proj_loss: -0.411030, loss_mean_cls: 0.072044, grad_norm: 0.256401 +Steps: 1%| | 5877/1000000 [24:24<68:24:08, 4.04it/s, grad_norm=0.256, loss_final=0.489, loss_mean=0.828, loss_mean_cls=0.072, proj_loss=-0.411][2026-03-26 12:55:27] Step: 5877, Training Logs: loss_final: 0.484471, loss_mean: 0.835874, proj_loss: -0.420303, loss_mean_cls: 0.068900, grad_norm: 0.279654 +Steps: 1%| | 5878/1000000 [24:24<68:24:09, 4.04it/s, grad_norm=0.28, loss_final=0.484, loss_mean=0.836, loss_mean_cls=0.0689, proj_loss=-0.42][2026-03-26 12:55:28] Step: 5878, Training Logs: loss_final: 0.467850, loss_mean: 0.808683, proj_loss: -0.412080, loss_mean_cls: 0.071248, grad_norm: 0.362424 +Steps: 1%| | 5879/1000000 [24:24<68:21:10, 4.04it/s, grad_norm=0.362, loss_final=0.468, loss_mean=0.809, loss_mean_cls=0.0712, proj_loss=-0.412][2026-03-26 12:55:28] Step: 5879, Training Logs: loss_final: 0.468386, loss_mean: 0.809008, proj_loss: -0.411146, loss_mean_cls: 0.070524, grad_norm: 0.271128 +Steps: 1%| | 5880/1000000 [24:25<68:21:36, 4.04it/s, grad_norm=0.271, loss_final=0.468, loss_mean=0.809, loss_mean_cls=0.0705, proj_loss=-0.411][2026-03-26 12:55:28] Step: 5880, Training Logs: loss_final: 0.498003, loss_mean: 0.839118, proj_loss: -0.411619, loss_mean_cls: 0.070504, grad_norm: 0.407848 +Steps: 1%| | 5881/1000000 [24:25<68:20:43, 4.04it/s, grad_norm=0.408, loss_final=0.498, loss_mean=0.839, loss_mean_cls=0.0705, proj_loss=-0.412][2026-03-26 12:55:28] Step: 5881, Training Logs: loss_final: 0.487812, loss_mean: 0.832542, proj_loss: -0.414512, loss_mean_cls: 0.069783, grad_norm: 0.296891 +Steps: 1%| | 5882/1000000 [24:25<68:21:46, 4.04it/s, grad_norm=0.297, loss_final=0.488, loss_mean=0.833, loss_mean_cls=0.0698, proj_loss=-0.415][2026-03-26 12:55:29] Step: 5882, Training Logs: loss_final: 0.475901, loss_mean: 0.820882, proj_loss: -0.415364, loss_mean_cls: 0.070383, grad_norm: 0.363041 +Steps: 1%| | 5883/1000000 [24:25<68:19:54, 4.04it/s, grad_norm=0.363, loss_final=0.476, loss_mean=0.821, loss_mean_cls=0.0704, proj_loss=-0.415][2026-03-26 12:55:29] Step: 5883, Training Logs: loss_final: 0.477792, loss_mean: 0.817706, proj_loss: -0.411866, loss_mean_cls: 0.071952, grad_norm: 0.628500 +Steps: 1%| | 5884/1000000 [24:26<68:19:06, 4.04it/s, grad_norm=0.629, loss_final=0.478, loss_mean=0.818, loss_mean_cls=0.072, proj_loss=-0.412][2026-03-26 12:55:29] Step: 5884, Training Logs: loss_final: 0.460728, loss_mean: 0.804610, proj_loss: -0.414569, loss_mean_cls: 0.070686, grad_norm: 0.279562 +Steps: 1%| | 5885/1000000 [24:26<68:20:26, 4.04it/s, grad_norm=0.28, loss_final=0.461, loss_mean=0.805, loss_mean_cls=0.0707, proj_loss=-0.415][2026-03-26 12:55:29] Step: 5885, Training Logs: loss_final: 0.486267, loss_mean: 0.821356, proj_loss: -0.407255, loss_mean_cls: 0.072167, grad_norm: 0.428034 +Steps: 1%| | 5886/1000000 [24:26<68:21:51, 4.04it/s, grad_norm=0.428, loss_final=0.486, loss_mean=0.821, loss_mean_cls=0.0722, proj_loss=-0.407][2026-03-26 12:55:30] Step: 5886, Training Logs: loss_final: 0.453172, loss_mean: 0.796314, proj_loss: -0.413328, loss_mean_cls: 0.070187, grad_norm: 0.314446 +Steps: 1%| | 5887/1000000 [24:26<68:20:03, 4.04it/s, grad_norm=0.314, loss_final=0.453, loss_mean=0.796, loss_mean_cls=0.0702, proj_loss=-0.413][2026-03-26 12:55:30] Step: 5887, Training Logs: loss_final: 0.471456, loss_mean: 0.810856, proj_loss: -0.410529, loss_mean_cls: 0.071129, grad_norm: 0.517511 +Steps: 1%| | 5888/1000000 [24:27<68:23:26, 4.04it/s, grad_norm=0.518, loss_final=0.471, loss_mean=0.811, loss_mean_cls=0.0711, proj_loss=-0.411][2026-03-26 12:55:30] Step: 5888, Training Logs: loss_final: 0.485232, loss_mean: 0.825340, proj_loss: -0.410726, loss_mean_cls: 0.070618, grad_norm: 0.217400 +Steps: 1%| | 5889/1000000 [24:27<68:22:10, 4.04it/s, grad_norm=0.217, loss_final=0.485, loss_mean=0.825, loss_mean_cls=0.0706, proj_loss=-0.411][2026-03-26 12:55:30] Step: 5889, Training Logs: loss_final: 0.483678, loss_mean: 0.810220, proj_loss: -0.399623, loss_mean_cls: 0.073081, grad_norm: 0.528474 +Steps: 1%| | 5890/1000000 [24:27<68:23:22, 4.04it/s, grad_norm=0.528, loss_final=0.484, loss_mean=0.81, loss_mean_cls=0.0731, proj_loss=-0.4][2026-03-26 12:55:31] Step: 5890, Training Logs: loss_final: 0.478367, loss_mean: 0.836314, proj_loss: -0.424677, loss_mean_cls: 0.066730, grad_norm: 0.334151 +Steps: 1%| | 5891/1000000 [24:27<68:22:52, 4.04it/s, grad_norm=0.334, loss_final=0.478, loss_mean=0.836, loss_mean_cls=0.0667, proj_loss=-0.425][2026-03-26 12:55:31] Step: 5891, Training Logs: loss_final: 0.466430, loss_mean: 0.816071, proj_loss: -0.419069, loss_mean_cls: 0.069428, grad_norm: 0.323723 +Steps: 1%| | 5892/1000000 [24:28<68:24:14, 4.04it/s, grad_norm=0.324, loss_final=0.466, loss_mean=0.816, loss_mean_cls=0.0694, proj_loss=-0.419][2026-03-26 12:55:31] Step: 5892, Training Logs: loss_final: 0.465139, loss_mean: 0.800165, proj_loss: -0.406464, loss_mean_cls: 0.071439, grad_norm: 0.326997 +Steps: 1%| | 5893/1000000 [24:28<68:27:47, 4.03it/s, grad_norm=0.327, loss_final=0.465, loss_mean=0.8, loss_mean_cls=0.0714, proj_loss=-0.406][2026-03-26 12:55:31] Step: 5893, Training Logs: loss_final: 0.483153, loss_mean: 0.821764, proj_loss: -0.409963, loss_mean_cls: 0.071352, grad_norm: 0.295146 +Steps: 1%| | 5894/1000000 [24:28<68:25:06, 4.04it/s, grad_norm=0.295, loss_final=0.483, loss_mean=0.822, loss_mean_cls=0.0714, proj_loss=-0.41][2026-03-26 12:55:32] Step: 5894, Training Logs: loss_final: 0.477608, loss_mean: 0.817658, proj_loss: -0.410250, loss_mean_cls: 0.070200, grad_norm: 0.251261 +Steps: 1%| | 5895/1000000 [24:28<68:26:32, 4.03it/s, grad_norm=0.251, loss_final=0.478, loss_mean=0.818, loss_mean_cls=0.0702, proj_loss=-0.41][2026-03-26 12:55:32] Step: 5895, Training Logs: loss_final: 0.488022, loss_mean: 0.829398, proj_loss: -0.412913, loss_mean_cls: 0.071537, grad_norm: 0.272502 +Steps: 1%| | 5896/1000000 [24:29<68:24:01, 4.04it/s, grad_norm=0.273, loss_final=0.488, loss_mean=0.829, loss_mean_cls=0.0715, proj_loss=-0.413][2026-03-26 12:55:32] Step: 5896, Training Logs: loss_final: 0.470902, loss_mean: 0.817251, proj_loss: -0.416235, loss_mean_cls: 0.069886, grad_norm: 0.265495 +Steps: 1%| | 5897/1000000 [24:29<68:26:46, 4.03it/s, grad_norm=0.265, loss_final=0.471, loss_mean=0.817, loss_mean_cls=0.0699, proj_loss=-0.416][2026-03-26 12:55:32] Step: 5897, Training Logs: loss_final: 0.471888, loss_mean: 0.819543, proj_loss: -0.417584, loss_mean_cls: 0.069929, grad_norm: 0.265054 +Steps: 1%| | 5898/1000000 [24:29<68:25:13, 4.04it/s, grad_norm=0.265, loss_final=0.472, loss_mean=0.82, loss_mean_cls=0.0699, proj_loss=-0.418][2026-03-26 12:55:32] Step: 5898, Training Logs: loss_final: 0.461143, loss_mean: 0.804574, proj_loss: -0.414119, loss_mean_cls: 0.070687, grad_norm: 0.272285 +Steps: 1%| | 5899/1000000 [24:29<68:23:53, 4.04it/s, grad_norm=0.272, loss_final=0.461, loss_mean=0.805, loss_mean_cls=0.0707, proj_loss=-0.414][2026-03-26 12:55:33] Step: 5899, Training Logs: loss_final: 0.474204, loss_mean: 0.820253, proj_loss: -0.416209, loss_mean_cls: 0.070159, grad_norm: 0.424356 +Steps: 1%| | 5900/1000000 [24:30<68:22:52, 4.04it/s, grad_norm=0.424, loss_final=0.474, loss_mean=0.82, loss_mean_cls=0.0702, proj_loss=-0.416][2026-03-26 12:55:33] Step: 5900, Training Logs: loss_final: 0.494350, loss_mean: 0.839400, proj_loss: -0.413965, loss_mean_cls: 0.068915, grad_norm: 0.377648 +Steps: 1%| | 5901/1000000 [24:30<68:22:57, 4.04it/s, grad_norm=0.378, loss_final=0.494, loss_mean=0.839, loss_mean_cls=0.0689, proj_loss=-0.414][2026-03-26 12:55:33] Step: 5901, Training Logs: loss_final: 0.481760, loss_mean: 0.829231, proj_loss: -0.416330, loss_mean_cls: 0.068859, grad_norm: 0.383954 +Steps: 1%| | 5902/1000000 [24:30<68:23:45, 4.04it/s, grad_norm=0.384, loss_final=0.482, loss_mean=0.829, loss_mean_cls=0.0689, proj_loss=-0.416][2026-03-26 12:55:33] Step: 5902, Training Logs: loss_final: 0.474539, loss_mean: 0.817229, proj_loss: -0.412819, loss_mean_cls: 0.070129, grad_norm: 0.404381 +Steps: 1%| | 5903/1000000 [24:30<68:24:15, 4.04it/s, grad_norm=0.404, loss_final=0.475, loss_mean=0.817, loss_mean_cls=0.0701, proj_loss=-0.413][2026-03-26 12:55:34] Step: 5903, Training Logs: loss_final: 0.474308, loss_mean: 0.817371, proj_loss: -0.413970, loss_mean_cls: 0.070907, grad_norm: 0.424937 +Steps: 1%| | 5904/1000000 [24:31<68:23:40, 4.04it/s, grad_norm=0.425, loss_final=0.474, loss_mean=0.817, loss_mean_cls=0.0709, proj_loss=-0.414][2026-03-26 12:55:34] Step: 5904, Training Logs: loss_final: 0.489989, loss_mean: 0.822114, proj_loss: -0.404321, loss_mean_cls: 0.072196, grad_norm: 0.381733 +Steps: 1%| | 5905/1000000 [24:31<68:22:34, 4.04it/s, grad_norm=0.382, loss_final=0.49, loss_mean=0.822, loss_mean_cls=0.0722, proj_loss=-0.404][2026-03-26 12:55:34] Step: 5905, Training Logs: loss_final: 0.476378, loss_mean: 0.818492, proj_loss: -0.412934, loss_mean_cls: 0.070820, grad_norm: 0.375143 +Steps: 1%| | 5906/1000000 [24:31<68:20:44, 4.04it/s, grad_norm=0.375, loss_final=0.476, loss_mean=0.818, loss_mean_cls=0.0708, proj_loss=-0.413][2026-03-26 12:55:34] Step: 5906, Training Logs: loss_final: 0.484283, loss_mean: 0.824848, proj_loss: -0.411450, loss_mean_cls: 0.070885, grad_norm: 0.411341 +Steps: 1%| | 5907/1000000 [24:31<68:23:16, 4.04it/s, grad_norm=0.411, loss_final=0.484, loss_mean=0.825, loss_mean_cls=0.0709, proj_loss=-0.411][2026-03-26 12:55:35] Step: 5907, Training Logs: loss_final: 0.482390, loss_mean: 0.836545, proj_loss: -0.421950, loss_mean_cls: 0.067794, grad_norm: 0.325906 +Steps: 1%| | 5908/1000000 [24:32<68:21:18, 4.04it/s, grad_norm=0.326, loss_final=0.482, loss_mean=0.837, loss_mean_cls=0.0678, proj_loss=-0.422][2026-03-26 12:55:35] Step: 5908, Training Logs: loss_final: 0.473514, loss_mean: 0.815392, proj_loss: -0.412788, loss_mean_cls: 0.070910, grad_norm: 0.305333 +Steps: 1%| | 5909/1000000 [24:32<68:21:40, 4.04it/s, grad_norm=0.305, loss_final=0.474, loss_mean=0.815, loss_mean_cls=0.0709, proj_loss=-0.413][2026-03-26 12:55:35] Step: 5909, Training Logs: loss_final: 0.469139, loss_mean: 0.808488, proj_loss: -0.410976, loss_mean_cls: 0.071627, grad_norm: 0.288116 +Steps: 1%| | 5910/1000000 [24:32<68:22:51, 4.04it/s, grad_norm=0.288, loss_final=0.469, loss_mean=0.808, loss_mean_cls=0.0716, proj_loss=-0.411][2026-03-26 12:55:35] Step: 5910, Training Logs: loss_final: 0.477377, loss_mean: 0.827122, proj_loss: -0.418942, loss_mean_cls: 0.069197, grad_norm: 0.479087 +Steps: 1%| | 5911/1000000 [24:32<68:20:31, 4.04it/s, grad_norm=0.479, loss_final=0.477, loss_mean=0.827, loss_mean_cls=0.0692, proj_loss=-0.419][2026-03-26 12:55:36] Step: 5911, Training Logs: loss_final: 0.479311, loss_mean: 0.817480, proj_loss: -0.410028, loss_mean_cls: 0.071859, grad_norm: 0.478520 +Steps: 1%| | 5912/1000000 [24:33<68:26:51, 4.03it/s, grad_norm=0.479, loss_final=0.479, loss_mean=0.817, loss_mean_cls=0.0719, proj_loss=-0.41][2026-03-26 12:55:36] Step: 5912, Training Logs: loss_final: 0.478698, loss_mean: 0.837107, proj_loss: -0.425317, loss_mean_cls: 0.066909, grad_norm: 0.335708 +Steps: 1%| | 5913/1000000 [24:33<68:20:26, 4.04it/s, grad_norm=0.336, loss_final=0.479, loss_mean=0.837, loss_mean_cls=0.0669, proj_loss=-0.425][2026-03-26 12:55:36] Step: 5913, Training Logs: loss_final: 0.482730, loss_mean: 0.824030, proj_loss: -0.412224, loss_mean_cls: 0.070924, grad_norm: 0.426357 +Steps: 1%| | 5914/1000000 [24:33<68:20:17, 4.04it/s, grad_norm=0.426, loss_final=0.483, loss_mean=0.824, loss_mean_cls=0.0709, proj_loss=-0.412][2026-03-26 12:55:36] Step: 5914, Training Logs: loss_final: 0.469551, loss_mean: 0.818744, proj_loss: -0.418972, loss_mean_cls: 0.069778, grad_norm: 0.371328 +Steps: 1%| | 5915/1000000 [24:33<68:19:19, 4.04it/s, grad_norm=0.371, loss_final=0.47, loss_mean=0.819, loss_mean_cls=0.0698, proj_loss=-0.419][2026-03-26 12:55:37] Step: 5915, Training Logs: loss_final: 0.483218, loss_mean: 0.822603, proj_loss: -0.410019, loss_mean_cls: 0.070634, grad_norm: 0.723992 +Steps: 1%| | 5916/1000000 [24:34<68:20:56, 4.04it/s, grad_norm=0.724, loss_final=0.483, loss_mean=0.823, loss_mean_cls=0.0706, proj_loss=-0.41][2026-03-26 12:55:37] Step: 5916, Training Logs: loss_final: 0.499556, loss_mean: 0.841108, proj_loss: -0.411712, loss_mean_cls: 0.070159, grad_norm: 0.423467 +Steps: 1%| | 5917/1000000 [24:34<70:52:17, 3.90it/s, grad_norm=0.423, loss_final=0.5, loss_mean=0.841, loss_mean_cls=0.0702, proj_loss=-0.412][2026-03-26 12:55:37] Step: 5917, Training Logs: loss_final: 0.490163, loss_mean: 0.835223, proj_loss: -0.414635, loss_mean_cls: 0.069575, grad_norm: 0.351034 +Steps: 1%| | 5918/1000000 [24:34<70:06:19, 3.94it/s, grad_norm=0.351, loss_final=0.49, loss_mean=0.835, loss_mean_cls=0.0696, proj_loss=-0.415][2026-03-26 12:55:37] Step: 5918, Training Logs: loss_final: 0.487689, loss_mean: 0.832758, proj_loss: -0.414930, loss_mean_cls: 0.069861, grad_norm: 0.295647 +Steps: 1%| | 5919/1000000 [24:34<69:34:28, 3.97it/s, grad_norm=0.296, loss_final=0.488, loss_mean=0.833, loss_mean_cls=0.0699, proj_loss=-0.415][2026-03-26 12:55:38] Step: 5919, Training Logs: loss_final: 0.470348, loss_mean: 0.826633, proj_loss: -0.423944, loss_mean_cls: 0.067659, grad_norm: 0.559271 +Steps: 1%| | 5920/1000000 [24:35<69:12:13, 3.99it/s, grad_norm=0.559, loss_final=0.47, loss_mean=0.827, loss_mean_cls=0.0677, proj_loss=-0.424][2026-03-26 12:55:38] Step: 5920, Training Logs: loss_final: 0.480203, loss_mean: 0.829566, proj_loss: -0.418682, loss_mean_cls: 0.069318, grad_norm: 0.597266 +Steps: 1%| | 5921/1000000 [24:35<68:56:56, 4.00it/s, grad_norm=0.597, loss_final=0.48, loss_mean=0.83, loss_mean_cls=0.0693, proj_loss=-0.419][2026-03-26 12:55:38] Step: 5921, Training Logs: loss_final: 0.465860, loss_mean: 0.800735, proj_loss: -0.406782, loss_mean_cls: 0.071907, grad_norm: 0.223087 +Steps: 1%| | 5922/1000000 [24:35<68:48:02, 4.01it/s, grad_norm=0.223, loss_final=0.466, loss_mean=0.801, loss_mean_cls=0.0719, proj_loss=-0.407][2026-03-26 12:55:38] Step: 5922, Training Logs: loss_final: 0.479389, loss_mean: 0.834968, proj_loss: -0.423104, loss_mean_cls: 0.067525, grad_norm: 0.647451 +Steps: 1%| | 5923/1000000 [24:35<68:34:36, 4.03it/s, grad_norm=0.647, loss_final=0.479, loss_mean=0.835, loss_mean_cls=0.0675, proj_loss=-0.423][2026-03-26 12:55:39] Step: 5923, Training Logs: loss_final: 0.475834, loss_mean: 0.811137, proj_loss: -0.407221, loss_mean_cls: 0.071919, grad_norm: 0.427683 +Steps: 1%| | 5924/1000000 [24:36<68:31:31, 4.03it/s, grad_norm=0.428, loss_final=0.476, loss_mean=0.811, loss_mean_cls=0.0719, proj_loss=-0.407][2026-03-26 12:55:39] Step: 5924, Training Logs: loss_final: 0.471797, loss_mean: 0.823505, proj_loss: -0.421018, loss_mean_cls: 0.069310, grad_norm: 0.570713 +Steps: 1%| | 5925/1000000 [24:36<68:28:43, 4.03it/s, grad_norm=0.571, loss_final=0.472, loss_mean=0.824, loss_mean_cls=0.0693, proj_loss=-0.421][2026-03-26 12:55:39] Step: 5925, Training Logs: loss_final: 0.460731, loss_mean: 0.807094, proj_loss: -0.417011, loss_mean_cls: 0.070648, grad_norm: 0.527392 +Steps: 1%| | 5926/1000000 [24:36<68:22:32, 4.04it/s, grad_norm=0.527, loss_final=0.461, loss_mean=0.807, loss_mean_cls=0.0706, proj_loss=-0.417][2026-03-26 12:55:39] Step: 5926, Training Logs: loss_final: 0.463690, loss_mean: 0.800931, proj_loss: -0.410466, loss_mean_cls: 0.073226, grad_norm: 0.308398 +Steps: 1%| | 5927/1000000 [24:36<68:24:32, 4.04it/s, grad_norm=0.308, loss_final=0.464, loss_mean=0.801, loss_mean_cls=0.0732, proj_loss=-0.41][2026-03-26 12:55:40] Step: 5927, Training Logs: loss_final: 0.492360, loss_mean: 0.828955, proj_loss: -0.408268, loss_mean_cls: 0.071674, grad_norm: 0.418746 +Steps: 1%| | 5928/1000000 [24:37<68:23:49, 4.04it/s, grad_norm=0.419, loss_final=0.492, loss_mean=0.829, loss_mean_cls=0.0717, proj_loss=-0.408][2026-03-26 12:55:40] Step: 5928, Training Logs: loss_final: 0.476962, loss_mean: 0.821965, proj_loss: -0.415259, loss_mean_cls: 0.070256, grad_norm: 0.336509 +Steps: 1%| | 5929/1000000 [24:37<68:28:22, 4.03it/s, grad_norm=0.337, loss_final=0.477, loss_mean=0.822, loss_mean_cls=0.0703, proj_loss=-0.415][2026-03-26 12:55:40] Step: 5929, Training Logs: loss_final: 0.488269, loss_mean: 0.840599, proj_loss: -0.422073, loss_mean_cls: 0.069743, grad_norm: 0.370478 +Steps: 1%| | 5930/1000000 [24:37<68:21:43, 4.04it/s, grad_norm=0.37, loss_final=0.488, loss_mean=0.841, loss_mean_cls=0.0697, proj_loss=-0.422][2026-03-26 12:55:40] Step: 5930, Training Logs: loss_final: 0.474763, loss_mean: 0.826176, proj_loss: -0.420762, loss_mean_cls: 0.069350, grad_norm: 0.246512 +Steps: 1%| | 5931/1000000 [24:37<68:19:51, 4.04it/s, grad_norm=0.247, loss_final=0.475, loss_mean=0.826, loss_mean_cls=0.0693, proj_loss=-0.421][2026-03-26 12:55:41] Step: 5931, Training Logs: loss_final: 0.464112, loss_mean: 0.794391, proj_loss: -0.403501, loss_mean_cls: 0.073221, grad_norm: 0.310155 +Steps: 1%| | 5932/1000000 [24:38<68:20:40, 4.04it/s, grad_norm=0.31, loss_final=0.464, loss_mean=0.794, loss_mean_cls=0.0732, proj_loss=-0.404][2026-03-26 12:55:41] Step: 5932, Training Logs: loss_final: 0.498044, loss_mean: 0.840773, proj_loss: -0.412995, loss_mean_cls: 0.070266, grad_norm: 0.438349 +Steps: 1%| | 5933/1000000 [24:38<68:21:08, 4.04it/s, grad_norm=0.438, loss_final=0.498, loss_mean=0.841, loss_mean_cls=0.0703, proj_loss=-0.413][2026-03-26 12:55:41] Step: 5933, Training Logs: loss_final: 0.486956, loss_mean: 0.833724, proj_loss: -0.416920, loss_mean_cls: 0.070152, grad_norm: 0.285819 +Steps: 1%| | 5934/1000000 [24:38<68:24:39, 4.04it/s, grad_norm=0.286, loss_final=0.487, loss_mean=0.834, loss_mean_cls=0.0702, proj_loss=-0.417][2026-03-26 12:55:41] Step: 5934, Training Logs: loss_final: 0.477050, loss_mean: 0.825830, proj_loss: -0.418212, loss_mean_cls: 0.069432, grad_norm: 0.330938 +Steps: 1%| | 5935/1000000 [24:38<69:03:03, 4.00it/s, grad_norm=0.331, loss_final=0.477, loss_mean=0.826, loss_mean_cls=0.0694, proj_loss=-0.418][2026-03-26 12:55:42] Step: 5935, Training Logs: loss_final: 0.477677, loss_mean: 0.816557, proj_loss: -0.409411, loss_mean_cls: 0.070531, grad_norm: 0.329348 +Steps: 1%| | 5936/1000000 [24:39<68:52:40, 4.01it/s, grad_norm=0.329, loss_final=0.478, loss_mean=0.817, loss_mean_cls=0.0705, proj_loss=-0.409][2026-03-26 12:55:42] Step: 5936, Training Logs: loss_final: 0.470012, loss_mean: 0.817189, proj_loss: -0.417111, loss_mean_cls: 0.069934, grad_norm: 0.522797 +Steps: 1%| | 5937/1000000 [24:39<68:40:30, 4.02it/s, grad_norm=0.523, loss_final=0.47, loss_mean=0.817, loss_mean_cls=0.0699, proj_loss=-0.417][2026-03-26 12:55:42] Step: 5937, Training Logs: loss_final: 0.455463, loss_mean: 0.804396, proj_loss: -0.417238, loss_mean_cls: 0.068305, grad_norm: 0.437944 +Steps: 1%| | 5938/1000000 [24:39<68:36:21, 4.02it/s, grad_norm=0.438, loss_final=0.455, loss_mean=0.804, loss_mean_cls=0.0683, proj_loss=-0.417][2026-03-26 12:55:42] Step: 5938, Training Logs: loss_final: 0.483736, loss_mean: 0.828401, proj_loss: -0.414434, loss_mean_cls: 0.069769, grad_norm: 0.297672 +Steps: 1%| | 5939/1000000 [24:39<68:32:53, 4.03it/s, grad_norm=0.298, loss_final=0.484, loss_mean=0.828, loss_mean_cls=0.0698, proj_loss=-0.414][2026-03-26 12:55:43] Step: 5939, Training Logs: loss_final: 0.476034, loss_mean: 0.827677, proj_loss: -0.420138, loss_mean_cls: 0.068495, grad_norm: 0.448219 +Steps: 1%| | 5940/1000000 [24:40<68:25:13, 4.04it/s, grad_norm=0.448, loss_final=0.476, loss_mean=0.828, loss_mean_cls=0.0685, proj_loss=-0.42][2026-03-26 12:55:43] Step: 5940, Training Logs: loss_final: 0.477918, loss_mean: 0.821743, proj_loss: -0.414066, loss_mean_cls: 0.070241, grad_norm: 0.246364 +Steps: 1%| | 5941/1000000 [24:40<68:22:43, 4.04it/s, grad_norm=0.246, loss_final=0.478, loss_mean=0.822, loss_mean_cls=0.0702, proj_loss=-0.414][2026-03-26 12:55:43] Step: 5941, Training Logs: loss_final: 0.472435, loss_mean: 0.819880, proj_loss: -0.417066, loss_mean_cls: 0.069622, grad_norm: 0.266903 +Steps: 1%| | 5942/1000000 [24:40<68:22:48, 4.04it/s, grad_norm=0.267, loss_final=0.472, loss_mean=0.82, loss_mean_cls=0.0696, proj_loss=-0.417][2026-03-26 12:55:43] Step: 5942, Training Logs: loss_final: 0.465421, loss_mean: 0.803837, proj_loss: -0.409286, loss_mean_cls: 0.070870, grad_norm: 0.272090 +Steps: 1%| | 5943/1000000 [24:40<68:22:44, 4.04it/s, grad_norm=0.272, loss_final=0.465, loss_mean=0.804, loss_mean_cls=0.0709, proj_loss=-0.409][2026-03-26 12:55:44] Step: 5943, Training Logs: loss_final: 0.449993, loss_mean: 0.790159, proj_loss: -0.412222, loss_mean_cls: 0.072056, grad_norm: 0.340075 +Steps: 1%| | 5944/1000000 [24:41<68:22:30, 4.04it/s, grad_norm=0.34, loss_final=0.45, loss_mean=0.79, loss_mean_cls=0.0721, proj_loss=-0.412][2026-03-26 12:55:44] Step: 5944, Training Logs: loss_final: 0.472881, loss_mean: 0.814086, proj_loss: -0.412502, loss_mean_cls: 0.071296, grad_norm: 0.389098 +Steps: 1%| | 5945/1000000 [24:41<68:20:54, 4.04it/s, grad_norm=0.389, loss_final=0.473, loss_mean=0.814, loss_mean_cls=0.0713, proj_loss=-0.413][2026-03-26 12:55:44] Step: 5945, Training Logs: loss_final: 0.475367, loss_mean: 0.823651, proj_loss: -0.417422, loss_mean_cls: 0.069139, grad_norm: 0.221504 +Steps: 1%| | 5946/1000000 [24:41<68:23:49, 4.04it/s, grad_norm=0.222, loss_final=0.475, loss_mean=0.824, loss_mean_cls=0.0691, proj_loss=-0.417][2026-03-26 12:55:44] Step: 5946, Training Logs: loss_final: 0.487375, loss_mean: 0.825338, proj_loss: -0.409147, loss_mean_cls: 0.071185, grad_norm: 0.300901 +Steps: 1%| | 5947/1000000 [24:41<68:20:50, 4.04it/s, grad_norm=0.301, loss_final=0.487, loss_mean=0.825, loss_mean_cls=0.0712, proj_loss=-0.409][2026-03-26 12:55:45] Step: 5947, Training Logs: loss_final: 0.479746, loss_mean: 0.811323, proj_loss: -0.405104, loss_mean_cls: 0.073527, grad_norm: 0.359362 +Steps: 1%| | 5948/1000000 [24:42<68:22:20, 4.04it/s, grad_norm=0.359, loss_final=0.48, loss_mean=0.811, loss_mean_cls=0.0735, proj_loss=-0.405][2026-03-26 12:55:45] Step: 5948, Training Logs: loss_final: 0.472927, loss_mean: 0.822634, proj_loss: -0.418679, loss_mean_cls: 0.068972, grad_norm: 0.295922 +Steps: 1%| | 5949/1000000 [24:42<68:21:20, 4.04it/s, grad_norm=0.296, loss_final=0.473, loss_mean=0.823, loss_mean_cls=0.069, proj_loss=-0.419][2026-03-26 12:55:45] Step: 5949, Training Logs: loss_final: 0.473336, loss_mean: 0.820751, proj_loss: -0.417134, loss_mean_cls: 0.069719, grad_norm: 0.273391 +Steps: 1%| | 5950/1000000 [24:42<68:19:49, 4.04it/s, grad_norm=0.273, loss_final=0.473, loss_mean=0.821, loss_mean_cls=0.0697, proj_loss=-0.417][2026-03-26 12:55:45] Step: 5950, Training Logs: loss_final: 0.474426, loss_mean: 0.815863, proj_loss: -0.412141, loss_mean_cls: 0.070705, grad_norm: 0.381299 +Steps: 1%| | 5951/1000000 [24:42<68:21:50, 4.04it/s, grad_norm=0.381, loss_final=0.474, loss_mean=0.816, loss_mean_cls=0.0707, proj_loss=-0.412][2026-03-26 12:55:46] Step: 5951, Training Logs: loss_final: 0.485327, loss_mean: 0.833169, proj_loss: -0.417261, loss_mean_cls: 0.069418, grad_norm: 0.334818 +Steps: 1%| | 5952/1000000 [24:43<68:22:19, 4.04it/s, grad_norm=0.335, loss_final=0.485, loss_mean=0.833, loss_mean_cls=0.0694, proj_loss=-0.417][2026-03-26 12:55:46] Step: 5952, Training Logs: loss_final: 0.474532, loss_mean: 0.815308, proj_loss: -0.411098, loss_mean_cls: 0.070323, grad_norm: 0.317920 +Steps: 1%| | 5953/1000000 [24:43<68:22:51, 4.04it/s, grad_norm=0.318, loss_final=0.475, loss_mean=0.815, loss_mean_cls=0.0703, proj_loss=-0.411][2026-03-26 12:55:46] Step: 5953, Training Logs: loss_final: 0.466877, loss_mean: 0.806806, proj_loss: -0.410946, loss_mean_cls: 0.071017, grad_norm: 0.312279 +Steps: 1%| | 5954/1000000 [24:43<68:28:38, 4.03it/s, grad_norm=0.312, loss_final=0.467, loss_mean=0.807, loss_mean_cls=0.071, proj_loss=-0.411][2026-03-26 12:55:46] Step: 5954, Training Logs: loss_final: 0.469817, loss_mean: 0.822805, proj_loss: -0.421180, loss_mean_cls: 0.068193, grad_norm: 0.290961 +Steps: 1%| | 5955/1000000 [24:43<68:29:43, 4.03it/s, grad_norm=0.291, loss_final=0.47, loss_mean=0.823, loss_mean_cls=0.0682, proj_loss=-0.421][2026-03-26 12:55:47] Step: 5955, Training Logs: loss_final: 0.473121, loss_mean: 0.824195, proj_loss: -0.419805, loss_mean_cls: 0.068731, grad_norm: 0.260980 +Steps: 1%| | 5956/1000000 [24:44<68:25:19, 4.04it/s, grad_norm=0.261, loss_final=0.473, loss_mean=0.824, loss_mean_cls=0.0687, proj_loss=-0.42][2026-03-26 12:55:47] Step: 5956, Training Logs: loss_final: 0.491022, loss_mean: 0.849229, proj_loss: -0.425421, loss_mean_cls: 0.067215, grad_norm: 0.281892 +Steps: 1%| | 5957/1000000 [24:44<68:24:35, 4.04it/s, grad_norm=0.282, loss_final=0.491, loss_mean=0.849, loss_mean_cls=0.0672, proj_loss=-0.425][2026-03-26 12:55:47] Step: 5957, Training Logs: loss_final: 0.484141, loss_mean: 0.828963, proj_loss: -0.414556, loss_mean_cls: 0.069735, grad_norm: 0.417175 +Steps: 1%| | 5958/1000000 [24:44<68:26:11, 4.03it/s, grad_norm=0.417, loss_final=0.484, loss_mean=0.829, loss_mean_cls=0.0697, proj_loss=-0.415][2026-03-26 12:55:47] Step: 5958, Training Logs: loss_final: 0.477594, loss_mean: 0.816282, proj_loss: -0.410581, loss_mean_cls: 0.071892, grad_norm: 0.296527 +Steps: 1%| | 5959/1000000 [24:44<68:24:18, 4.04it/s, grad_norm=0.297, loss_final=0.478, loss_mean=0.816, loss_mean_cls=0.0719, proj_loss=-0.411][2026-03-26 12:55:48] Step: 5959, Training Logs: loss_final: 0.470945, loss_mean: 0.822863, proj_loss: -0.419823, loss_mean_cls: 0.067905, grad_norm: 0.289185 +Steps: 1%| | 5960/1000000 [24:44<68:23:32, 4.04it/s, grad_norm=0.289, loss_final=0.471, loss_mean=0.823, loss_mean_cls=0.0679, proj_loss=-0.42][2026-03-26 12:55:48] Step: 5960, Training Logs: loss_final: 0.478423, loss_mean: 0.822641, proj_loss: -0.414771, loss_mean_cls: 0.070552, grad_norm: 0.311786 +Steps: 1%| | 5961/1000000 [24:45<68:23:31, 4.04it/s, grad_norm=0.312, loss_final=0.478, loss_mean=0.823, loss_mean_cls=0.0706, proj_loss=-0.415][2026-03-26 12:55:48] Step: 5961, Training Logs: loss_final: 0.484294, loss_mean: 0.840489, proj_loss: -0.424458, loss_mean_cls: 0.068263, grad_norm: 0.307390 +Steps: 1%| | 5962/1000000 [24:45<68:22:12, 4.04it/s, grad_norm=0.307, loss_final=0.484, loss_mean=0.84, loss_mean_cls=0.0683, proj_loss=-0.424][2026-03-26 12:55:48] Step: 5962, Training Logs: loss_final: 0.487611, loss_mean: 0.832149, proj_loss: -0.413863, loss_mean_cls: 0.069325, grad_norm: 0.407283 +Steps: 1%| | 5963/1000000 [24:45<68:22:12, 4.04it/s, grad_norm=0.407, loss_final=0.488, loss_mean=0.832, loss_mean_cls=0.0693, proj_loss=-0.414][2026-03-26 12:55:49] Step: 5963, Training Logs: loss_final: 0.492538, loss_mean: 0.845716, proj_loss: -0.421250, loss_mean_cls: 0.068071, grad_norm: 0.240778 +Steps: 1%| | 5964/1000000 [24:45<68:22:53, 4.04it/s, grad_norm=0.241, loss_final=0.493, loss_mean=0.846, loss_mean_cls=0.0681, proj_loss=-0.421][2026-03-26 12:55:49] Step: 5964, Training Logs: loss_final: 0.466017, loss_mean: 0.803373, proj_loss: -0.408396, loss_mean_cls: 0.071040, grad_norm: 0.319191 +Steps: 1%| | 5965/1000000 [24:46<68:27:29, 4.03it/s, grad_norm=0.319, loss_final=0.466, loss_mean=0.803, loss_mean_cls=0.071, proj_loss=-0.408][2026-03-26 12:55:49] Step: 5965, Training Logs: loss_final: 0.475569, loss_mean: 0.821440, proj_loss: -0.415871, loss_mean_cls: 0.070000, grad_norm: 0.280435 +Steps: 1%| | 5966/1000000 [24:46<68:27:34, 4.03it/s, grad_norm=0.28, loss_final=0.476, loss_mean=0.821, loss_mean_cls=0.07, proj_loss=-0.416][2026-03-26 12:55:49] Step: 5966, Training Logs: loss_final: 0.473325, loss_mean: 0.821957, proj_loss: -0.418007, loss_mean_cls: 0.069375, grad_norm: 0.331438 +Steps: 1%| | 5967/1000000 [24:46<68:25:50, 4.04it/s, grad_norm=0.331, loss_final=0.473, loss_mean=0.822, loss_mean_cls=0.0694, proj_loss=-0.418][2026-03-26 12:55:50] Step: 5967, Training Logs: loss_final: 0.478934, loss_mean: 0.830422, proj_loss: -0.420218, loss_mean_cls: 0.068730, grad_norm: 0.314142 +Steps: 1%| | 5968/1000000 [24:46<68:24:07, 4.04it/s, grad_norm=0.314, loss_final=0.479, loss_mean=0.83, loss_mean_cls=0.0687, proj_loss=-0.42][2026-03-26 12:55:50] Step: 5968, Training Logs: loss_final: 0.488374, loss_mean: 0.828178, proj_loss: -0.410653, loss_mean_cls: 0.070848, grad_norm: 0.276085 +Steps: 1%| | 5969/1000000 [24:47<68:25:21, 4.04it/s, grad_norm=0.276, loss_final=0.488, loss_mean=0.828, loss_mean_cls=0.0708, proj_loss=-0.411][2026-03-26 12:55:50] Step: 5969, Training Logs: loss_final: 0.474128, loss_mean: 0.810500, proj_loss: -0.407591, loss_mean_cls: 0.071220, grad_norm: 0.271610 +Steps: 1%| | 5970/1000000 [24:47<68:24:13, 4.04it/s, grad_norm=0.272, loss_final=0.474, loss_mean=0.811, loss_mean_cls=0.0712, proj_loss=-0.408][2026-03-26 12:55:50] Step: 5970, Training Logs: loss_final: 0.464930, loss_mean: 0.797720, proj_loss: -0.405504, loss_mean_cls: 0.072715, grad_norm: 0.325287 +Steps: 1%| | 5971/1000000 [24:47<68:21:46, 4.04it/s, grad_norm=0.325, loss_final=0.465, loss_mean=0.798, loss_mean_cls=0.0727, proj_loss=-0.406][2026-03-26 12:55:51] Step: 5971, Training Logs: loss_final: 0.486978, loss_mean: 0.828198, proj_loss: -0.412220, loss_mean_cls: 0.071000, grad_norm: 0.382849 +Steps: 1%| | 5972/1000000 [24:47<68:23:31, 4.04it/s, grad_norm=0.383, loss_final=0.487, loss_mean=0.828, loss_mean_cls=0.071, proj_loss=-0.412][2026-03-26 12:55:51] Step: 5972, Training Logs: loss_final: 0.478083, loss_mean: 0.819345, proj_loss: -0.411610, loss_mean_cls: 0.070349, grad_norm: 0.206543 +Steps: 1%| | 5973/1000000 [24:48<68:22:27, 4.04it/s, grad_norm=0.207, loss_final=0.478, loss_mean=0.819, loss_mean_cls=0.0703, proj_loss=-0.412][2026-03-26 12:55:51] Step: 5973, Training Logs: loss_final: 0.465841, loss_mean: 0.808014, proj_loss: -0.412603, loss_mean_cls: 0.070430, grad_norm: 0.591474 +Steps: 1%| | 5974/1000000 [24:48<68:18:23, 4.04it/s, grad_norm=0.591, loss_final=0.466, loss_mean=0.808, loss_mean_cls=0.0704, proj_loss=-0.413][2026-03-26 12:55:51] Step: 5974, Training Logs: loss_final: 0.478651, loss_mean: 0.828158, proj_loss: -0.417934, loss_mean_cls: 0.068426, grad_norm: 0.478489 +Steps: 1%| | 5975/1000000 [24:48<68:34:56, 4.03it/s, grad_norm=0.478, loss_final=0.479, loss_mean=0.828, loss_mean_cls=0.0684, proj_loss=-0.418][2026-03-26 12:55:52] Step: 5975, Training Logs: loss_final: 0.482156, loss_mean: 0.822189, proj_loss: -0.410937, loss_mean_cls: 0.070904, grad_norm: 0.278550 +Steps: 1%| | 5976/1000000 [24:48<68:25:46, 4.04it/s, grad_norm=0.279, loss_final=0.482, loss_mean=0.822, loss_mean_cls=0.0709, proj_loss=-0.411][2026-03-26 12:55:52] Step: 5976, Training Logs: loss_final: 0.457226, loss_mean: 0.804585, proj_loss: -0.416833, loss_mean_cls: 0.069474, grad_norm: 0.355626 +Steps: 1%| | 5977/1000000 [24:49<68:23:23, 4.04it/s, grad_norm=0.356, loss_final=0.457, loss_mean=0.805, loss_mean_cls=0.0695, proj_loss=-0.417][2026-03-26 12:55:52] Step: 5977, Training Logs: loss_final: 0.481565, loss_mean: 0.823892, proj_loss: -0.413485, loss_mean_cls: 0.071158, grad_norm: 0.303405 +Steps: 1%| | 5978/1000000 [24:49<68:22:51, 4.04it/s, grad_norm=0.303, loss_final=0.482, loss_mean=0.824, loss_mean_cls=0.0712, proj_loss=-0.413][2026-03-26 12:55:52] Step: 5978, Training Logs: loss_final: 0.479655, loss_mean: 0.827212, proj_loss: -0.417385, loss_mean_cls: 0.069828, grad_norm: 0.519901 +Steps: 1%| | 5979/1000000 [24:49<68:20:39, 4.04it/s, grad_norm=0.52, loss_final=0.48, loss_mean=0.827, loss_mean_cls=0.0698, proj_loss=-0.417][2026-03-26 12:55:53] Step: 5979, Training Logs: loss_final: 0.501942, loss_mean: 0.840542, proj_loss: -0.409687, loss_mean_cls: 0.071087, grad_norm: 0.313727 +Steps: 1%| | 5980/1000000 [24:49<68:23:50, 4.04it/s, grad_norm=0.314, loss_final=0.502, loss_mean=0.841, loss_mean_cls=0.0711, proj_loss=-0.41][2026-03-26 12:55:53] Step: 5980, Training Logs: loss_final: 0.461824, loss_mean: 0.809698, proj_loss: -0.417368, loss_mean_cls: 0.069493, grad_norm: 0.779720 +Steps: 1%| | 5981/1000000 [24:50<68:24:34, 4.04it/s, grad_norm=0.78, loss_final=0.462, loss_mean=0.81, loss_mean_cls=0.0695, proj_loss=-0.417][2026-03-26 12:55:53] Step: 5981, Training Logs: loss_final: 0.483196, loss_mean: 0.837479, proj_loss: -0.422403, loss_mean_cls: 0.068121, grad_norm: 0.687428 +Steps: 1%| | 5982/1000000 [24:50<68:21:24, 4.04it/s, grad_norm=0.687, loss_final=0.483, loss_mean=0.837, loss_mean_cls=0.0681, proj_loss=-0.422][2026-03-26 12:55:53] Step: 5982, Training Logs: loss_final: 0.475129, loss_mean: 0.825246, proj_loss: -0.418594, loss_mean_cls: 0.068478, grad_norm: 0.274933 +Steps: 1%| | 5983/1000000 [24:50<68:20:12, 4.04it/s, grad_norm=0.275, loss_final=0.475, loss_mean=0.825, loss_mean_cls=0.0685, proj_loss=-0.419][2026-03-26 12:55:54] Step: 5983, Training Logs: loss_final: 0.502844, loss_mean: 0.838938, proj_loss: -0.406974, loss_mean_cls: 0.070880, grad_norm: 0.539555 +Steps: 1%| | 5984/1000000 [24:50<68:20:19, 4.04it/s, grad_norm=0.54, loss_final=0.503, loss_mean=0.839, loss_mean_cls=0.0709, proj_loss=-0.407][2026-03-26 12:55:54] Step: 5984, Training Logs: loss_final: 0.459009, loss_mean: 0.789560, proj_loss: -0.402703, loss_mean_cls: 0.072152, grad_norm: 0.299851 +Steps: 1%| | 5985/1000000 [24:51<68:19:24, 4.04it/s, grad_norm=0.3, loss_final=0.459, loss_mean=0.79, loss_mean_cls=0.0722, proj_loss=-0.403][2026-03-26 12:55:54] Step: 5985, Training Logs: loss_final: 0.498415, loss_mean: 0.848050, proj_loss: -0.418655, loss_mean_cls: 0.069021, grad_norm: 0.975780 +Steps: 1%| | 5986/1000000 [24:51<68:21:36, 4.04it/s, grad_norm=0.976, loss_final=0.498, loss_mean=0.848, loss_mean_cls=0.069, proj_loss=-0.419][2026-03-26 12:55:54] Step: 5986, Training Logs: loss_final: 0.476902, loss_mean: 0.808131, proj_loss: -0.403537, loss_mean_cls: 0.072308, grad_norm: 0.534989 +Steps: 1%| | 5987/1000000 [24:51<69:30:48, 3.97it/s, grad_norm=0.535, loss_final=0.477, loss_mean=0.808, loss_mean_cls=0.0723, proj_loss=-0.404][2026-03-26 12:55:55] Step: 5987, Training Logs: loss_final: 0.481902, loss_mean: 0.827748, proj_loss: -0.415562, loss_mean_cls: 0.069716, grad_norm: 0.769548 +Steps: 1%| | 5988/1000000 [24:51<69:15:46, 3.99it/s, grad_norm=0.77, loss_final=0.482, loss_mean=0.828, loss_mean_cls=0.0697, proj_loss=-0.416][2026-03-26 12:55:55] Step: 5988, Training Logs: loss_final: 0.469847, loss_mean: 0.813928, proj_loss: -0.413737, loss_mean_cls: 0.069656, grad_norm: 0.824023 +Steps: 1%| | 5989/1000000 [24:52<68:58:17, 4.00it/s, grad_norm=0.824, loss_final=0.47, loss_mean=0.814, loss_mean_cls=0.0697, proj_loss=-0.414][2026-03-26 12:55:55] Step: 5989, Training Logs: loss_final: 0.476740, loss_mean: 0.830466, proj_loss: -0.421575, loss_mean_cls: 0.067849, grad_norm: 0.427480 +Steps: 1%| | 5990/1000000 [24:52<68:48:42, 4.01it/s, grad_norm=0.427, loss_final=0.477, loss_mean=0.83, loss_mean_cls=0.0678, proj_loss=-0.422][2026-03-26 12:55:55] Step: 5990, Training Logs: loss_final: 0.485927, loss_mean: 0.836265, proj_loss: -0.419140, loss_mean_cls: 0.068801, grad_norm: 0.938310 +Steps: 1%| | 5991/1000000 [24:52<68:39:00, 4.02it/s, grad_norm=0.938, loss_final=0.486, loss_mean=0.836, loss_mean_cls=0.0688, proj_loss=-0.419][2026-03-26 12:55:56] Step: 5991, Training Logs: loss_final: 0.469946, loss_mean: 0.813340, proj_loss: -0.413533, loss_mean_cls: 0.070139, grad_norm: 0.237525 +Steps: 1%| | 5992/1000000 [24:52<68:35:57, 4.03it/s, grad_norm=0.238, loss_final=0.47, loss_mean=0.813, loss_mean_cls=0.0701, proj_loss=-0.414][2026-03-26 12:55:56] Step: 5992, Training Logs: loss_final: 0.476187, loss_mean: 0.820340, proj_loss: -0.414747, loss_mean_cls: 0.070594, grad_norm: 0.751127 +Steps: 1%| | 5993/1000000 [24:53<68:31:14, 4.03it/s, grad_norm=0.751, loss_final=0.476, loss_mean=0.82, loss_mean_cls=0.0706, proj_loss=-0.415][2026-03-26 12:55:56] Step: 5993, Training Logs: loss_final: 0.498170, loss_mean: 0.844704, proj_loss: -0.415775, loss_mean_cls: 0.069242, grad_norm: 0.318155 +Steps: 1%| | 5994/1000000 [24:53<68:28:09, 4.03it/s, grad_norm=0.318, loss_final=0.498, loss_mean=0.845, loss_mean_cls=0.0692, proj_loss=-0.416][2026-03-26 12:55:56] Step: 5994, Training Logs: loss_final: 0.477477, loss_mean: 0.815076, proj_loss: -0.408733, loss_mean_cls: 0.071133, grad_norm: 0.711255 +Steps: 1%| | 5995/1000000 [24:53<73:14:40, 3.77it/s, grad_norm=0.711, loss_final=0.477, loss_mean=0.815, loss_mean_cls=0.0711, proj_loss=-0.409][2026-03-26 12:55:57] Step: 5995, Training Logs: loss_final: 0.464979, loss_mean: 0.818219, proj_loss: -0.421640, loss_mean_cls: 0.068400, grad_norm: 0.433912 +Steps: 1%| | 5996/1000000 [24:53<72:00:34, 3.83it/s, grad_norm=0.434, loss_final=0.465, loss_mean=0.818, loss_mean_cls=0.0684, proj_loss=-0.422][2026-03-26 12:55:57] Step: 5996, Training Logs: loss_final: 0.484394, loss_mean: 0.836790, proj_loss: -0.421544, loss_mean_cls: 0.069148, grad_norm: 0.639829 +Steps: 1%| | 5997/1000000 [24:54<70:52:18, 3.90it/s, grad_norm=0.64, loss_final=0.484, loss_mean=0.837, loss_mean_cls=0.0691, proj_loss=-0.422][2026-03-26 12:55:57] Step: 5997, Training Logs: loss_final: 0.493552, loss_mean: 0.838212, proj_loss: -0.414775, loss_mean_cls: 0.070115, grad_norm: 0.383268 +Steps: 1%| | 5998/1000000 [24:54<70:08:49, 3.94it/s, grad_norm=0.383, loss_final=0.494, loss_mean=0.838, loss_mean_cls=0.0701, proj_loss=-0.415][2026-03-26 12:55:57] Step: 5998, Training Logs: loss_final: 0.467661, loss_mean: 0.816027, proj_loss: -0.418177, loss_mean_cls: 0.069811, grad_norm: 0.466414 +Steps: 1%| | 5999/1000000 [24:54<69:36:31, 3.97it/s, grad_norm=0.466, loss_final=0.468, loss_mean=0.816, loss_mean_cls=0.0698, proj_loss=-0.418][2026-03-26 12:55:58] Step: 5999, Training Logs: loss_final: 0.471882, loss_mean: 0.817712, proj_loss: -0.415104, loss_mean_cls: 0.069273, grad_norm: 0.321858 +Steps: 1%| | 6000/1000000 [24:54<69:18:10, 3.98it/s, grad_norm=0.322, loss_final=0.472, loss_mean=0.818, loss_mean_cls=0.0693, proj_loss=-0.415][2026-03-26 12:55:58] Step: 6000, Training Logs: loss_final: 0.465306, loss_mean: 0.814331, proj_loss: -0.417921, loss_mean_cls: 0.068897, grad_norm: 0.392304 +Steps: 1%| | 6001/1000000 [24:55<69:00:39, 4.00it/s, grad_norm=0.392, loss_final=0.465, loss_mean=0.814, loss_mean_cls=0.0689, proj_loss=-0.418][2026-03-26 12:55:58] Step: 6001, Training Logs: loss_final: 0.499113, loss_mean: 0.843274, proj_loss: -0.413622, loss_mean_cls: 0.069461, grad_norm: 0.380500 +Steps: 1%| | 6002/1000000 [24:55<68:44:00, 4.02it/s, grad_norm=0.38, loss_final=0.499, loss_mean=0.843, loss_mean_cls=0.0695, proj_loss=-0.414][2026-03-26 12:55:58] Step: 6002, Training Logs: loss_final: 0.466448, loss_mean: 0.811093, proj_loss: -0.415067, loss_mean_cls: 0.070422, grad_norm: 0.490553 +Steps: 1%| | 6003/1000000 [24:55<68:35:37, 4.03it/s, grad_norm=0.491, loss_final=0.466, loss_mean=0.811, loss_mean_cls=0.0704, proj_loss=-0.415][2026-03-26 12:55:59] Step: 6003, Training Logs: loss_final: 0.482683, loss_mean: 0.822945, proj_loss: -0.411188, loss_mean_cls: 0.070926, grad_norm: 0.672142 +Steps: 1%| | 6004/1000000 [24:55<68:28:50, 4.03it/s, grad_norm=0.672, loss_final=0.483, loss_mean=0.823, loss_mean_cls=0.0709, proj_loss=-0.411][2026-03-26 12:55:59] Step: 6004, Training Logs: loss_final: 0.469819, loss_mean: 0.816881, proj_loss: -0.416608, loss_mean_cls: 0.069546, grad_norm: 0.493175 +Steps: 1%| | 6005/1000000 [24:56<68:30:40, 4.03it/s, grad_norm=0.493, loss_final=0.47, loss_mean=0.817, loss_mean_cls=0.0695, proj_loss=-0.417][2026-03-26 12:55:59] Step: 6005, Training Logs: loss_final: 0.478756, loss_mean: 0.805014, proj_loss: -0.399480, loss_mean_cls: 0.073223, grad_norm: 0.765564 +Steps: 1%| | 6006/1000000 [24:56<68:27:45, 4.03it/s, grad_norm=0.766, loss_final=0.479, loss_mean=0.805, loss_mean_cls=0.0732, proj_loss=-0.399][2026-03-26 12:55:59] Step: 6006, Training Logs: loss_final: 0.477717, loss_mean: 0.826615, proj_loss: -0.418606, loss_mean_cls: 0.069709, grad_norm: 0.275213 +Steps: 1%| | 6007/1000000 [24:56<68:24:44, 4.04it/s, grad_norm=0.275, loss_final=0.478, loss_mean=0.827, loss_mean_cls=0.0697, proj_loss=-0.419][2026-03-26 12:56:00] Step: 6007, Training Logs: loss_final: 0.474503, loss_mean: 0.828979, proj_loss: -0.423070, loss_mean_cls: 0.068594, grad_norm: 0.505664 +Steps: 1%| | 6008/1000000 [24:56<68:25:53, 4.03it/s, grad_norm=0.506, loss_final=0.475, loss_mean=0.829, loss_mean_cls=0.0686, proj_loss=-0.423][2026-03-26 12:56:00] Step: 6008, Training Logs: loss_final: 0.490777, loss_mean: 0.842756, proj_loss: -0.419239, loss_mean_cls: 0.067260, grad_norm: 0.263617 +Steps: 1%| | 6009/1000000 [24:57<68:24:19, 4.04it/s, grad_norm=0.264, loss_final=0.491, loss_mean=0.843, loss_mean_cls=0.0673, proj_loss=-0.419][2026-03-26 12:56:00] Step: 6009, Training Logs: loss_final: 0.486278, loss_mean: 0.829519, proj_loss: -0.413115, loss_mean_cls: 0.069874, grad_norm: 0.654906 +Steps: 1%| | 6010/1000000 [24:57<69:14:38, 3.99it/s, grad_norm=0.655, loss_final=0.486, loss_mean=0.83, loss_mean_cls=0.0699, proj_loss=-0.413][2026-03-26 12:56:00] Step: 6010, Training Logs: loss_final: 0.471989, loss_mean: 0.814583, proj_loss: -0.413096, loss_mean_cls: 0.070501, grad_norm: 0.659017 +Steps: 1%| | 6011/1000000 [24:57<68:55:45, 4.01it/s, grad_norm=0.659, loss_final=0.472, loss_mean=0.815, loss_mean_cls=0.0705, proj_loss=-0.413][2026-03-26 12:56:01] Step: 6011, Training Logs: loss_final: 0.481903, loss_mean: 0.817661, proj_loss: -0.407815, loss_mean_cls: 0.072057, grad_norm: 0.531299 +Steps: 1%| | 6012/1000000 [24:57<68:48:16, 4.01it/s, grad_norm=0.531, loss_final=0.482, loss_mean=0.818, loss_mean_cls=0.0721, proj_loss=-0.408][2026-03-26 12:56:01] Step: 6012, Training Logs: loss_final: 0.476723, loss_mean: 0.828073, proj_loss: -0.420114, loss_mean_cls: 0.068764, grad_norm: 0.599925 +Steps: 1%| | 6013/1000000 [24:58<68:40:10, 4.02it/s, grad_norm=0.6, loss_final=0.477, loss_mean=0.828, loss_mean_cls=0.0688, proj_loss=-0.42][2026-03-26 12:56:01] Step: 6013, Training Logs: loss_final: 0.477017, loss_mean: 0.820909, proj_loss: -0.413122, loss_mean_cls: 0.069230, grad_norm: 0.319733 +Steps: 1%| | 6014/1000000 [24:58<68:40:27, 4.02it/s, grad_norm=0.32, loss_final=0.477, loss_mean=0.821, loss_mean_cls=0.0692, proj_loss=-0.413][2026-03-26 12:56:01] Step: 6014, Training Logs: loss_final: 0.476810, loss_mean: 0.823611, proj_loss: -0.416487, loss_mean_cls: 0.069686, grad_norm: 0.528253 +Steps: 1%| | 6015/1000000 [24:58<68:32:51, 4.03it/s, grad_norm=0.528, loss_final=0.477, loss_mean=0.824, loss_mean_cls=0.0697, proj_loss=-0.416][2026-03-26 12:56:02] Step: 6015, Training Logs: loss_final: 0.479658, loss_mean: 0.827678, proj_loss: -0.418250, loss_mean_cls: 0.070231, grad_norm: 0.376185 +Steps: 1%| | 6016/1000000 [24:58<68:31:43, 4.03it/s, grad_norm=0.376, loss_final=0.48, loss_mean=0.828, loss_mean_cls=0.0702, proj_loss=-0.418][2026-03-26 12:56:02] Step: 6016, Training Logs: loss_final: 0.461098, loss_mean: 0.801807, proj_loss: -0.411794, loss_mean_cls: 0.071085, grad_norm: 0.367429 +Steps: 1%| | 6017/1000000 [24:59<68:26:50, 4.03it/s, grad_norm=0.367, loss_final=0.461, loss_mean=0.802, loss_mean_cls=0.0711, proj_loss=-0.412][2026-03-26 12:56:02] Step: 6017, Training Logs: loss_final: 0.468888, loss_mean: 0.814929, proj_loss: -0.416148, loss_mean_cls: 0.070106, grad_norm: 0.266247 +Steps: 1%| | 6018/1000000 [24:59<68:27:37, 4.03it/s, grad_norm=0.266, loss_final=0.469, loss_mean=0.815, loss_mean_cls=0.0701, proj_loss=-0.416][2026-03-26 12:56:02] Step: 6018, Training Logs: loss_final: 0.477417, loss_mean: 0.827291, proj_loss: -0.418080, loss_mean_cls: 0.068206, grad_norm: 0.514617 +Steps: 1%| | 6019/1000000 [24:59<68:23:28, 4.04it/s, grad_norm=0.515, loss_final=0.477, loss_mean=0.827, loss_mean_cls=0.0682, proj_loss=-0.418][2026-03-26 12:56:03] Step: 6019, Training Logs: loss_final: 0.473085, loss_mean: 0.815322, proj_loss: -0.412535, loss_mean_cls: 0.070299, grad_norm: 0.386534 +Steps: 1%| | 6020/1000000 [24:59<68:25:18, 4.04it/s, grad_norm=0.387, loss_final=0.473, loss_mean=0.815, loss_mean_cls=0.0703, proj_loss=-0.413][2026-03-26 12:56:03] Step: 6020, Training Logs: loss_final: 0.471156, loss_mean: 0.817300, proj_loss: -0.415589, loss_mean_cls: 0.069445, grad_norm: 0.490543 +Steps: 1%| | 6021/1000000 [25:00<68:24:30, 4.04it/s, grad_norm=0.491, loss_final=0.471, loss_mean=0.817, loss_mean_cls=0.0694, proj_loss=-0.416][2026-03-26 12:56:03] Step: 6021, Training Logs: loss_final: 0.453703, loss_mean: 0.796383, proj_loss: -0.413297, loss_mean_cls: 0.070617, grad_norm: 0.504261 +Steps: 1%| | 6022/1000000 [25:00<68:21:45, 4.04it/s, grad_norm=0.504, loss_final=0.454, loss_mean=0.796, loss_mean_cls=0.0706, proj_loss=-0.413][2026-03-26 12:56:03] Step: 6022, Training Logs: loss_final: 0.492572, loss_mean: 0.841446, proj_loss: -0.417649, loss_mean_cls: 0.068775, grad_norm: 0.492304 +Steps: 1%| | 6023/1000000 [25:00<68:20:38, 4.04it/s, grad_norm=0.492, loss_final=0.493, loss_mean=0.841, loss_mean_cls=0.0688, proj_loss=-0.418][2026-03-26 12:56:04] Step: 6023, Training Logs: loss_final: 0.475347, loss_mean: 0.823437, proj_loss: -0.417652, loss_mean_cls: 0.069562, grad_norm: 0.682851 +Steps: 1%| | 6024/1000000 [25:00<68:18:41, 4.04it/s, grad_norm=0.683, loss_final=0.475, loss_mean=0.823, loss_mean_cls=0.0696, proj_loss=-0.418][2026-03-26 12:56:04] Step: 6024, Training Logs: loss_final: 0.492082, loss_mean: 0.847996, proj_loss: -0.421975, loss_mean_cls: 0.066061, grad_norm: 0.321993 +Steps: 1%| | 6025/1000000 [25:01<68:18:51, 4.04it/s, grad_norm=0.322, loss_final=0.492, loss_mean=0.848, loss_mean_cls=0.0661, proj_loss=-0.422][2026-03-26 12:56:04] Step: 6025, Training Logs: loss_final: 0.477292, loss_mean: 0.829657, proj_loss: -0.420501, loss_mean_cls: 0.068136, grad_norm: 0.666266 +Steps: 1%| | 6026/1000000 [25:01<68:18:18, 4.04it/s, grad_norm=0.666, loss_final=0.477, loss_mean=0.83, loss_mean_cls=0.0681, proj_loss=-0.421][2026-03-26 12:56:04] Step: 6026, Training Logs: loss_final: 0.489264, loss_mean: 0.813925, proj_loss: -0.398083, loss_mean_cls: 0.073422, grad_norm: 0.274599 +Steps: 1%| | 6027/1000000 [25:01<68:20:44, 4.04it/s, grad_norm=0.275, loss_final=0.489, loss_mean=0.814, loss_mean_cls=0.0734, proj_loss=-0.398][2026-03-26 12:56:05] Step: 6027, Training Logs: loss_final: 0.477612, loss_mean: 0.834699, proj_loss: -0.425122, loss_mean_cls: 0.068035, grad_norm: 0.658846 +Steps: 1%| | 6028/1000000 [25:01<68:24:52, 4.04it/s, grad_norm=0.659, loss_final=0.478, loss_mean=0.835, loss_mean_cls=0.068, proj_loss=-0.425][2026-03-26 12:56:05] Step: 6028, Training Logs: loss_final: 0.470919, loss_mean: 0.815373, proj_loss: -0.414109, loss_mean_cls: 0.069656, grad_norm: 0.459882 +Steps: 1%| | 6029/1000000 [25:02<68:24:46, 4.04it/s, grad_norm=0.46, loss_final=0.471, loss_mean=0.815, loss_mean_cls=0.0697, proj_loss=-0.414][2026-03-26 12:56:05] Step: 6029, Training Logs: loss_final: 0.480411, loss_mean: 0.830066, proj_loss: -0.417440, loss_mean_cls: 0.067785, grad_norm: 0.447375 +Steps: 1%| | 6030/1000000 [25:02<68:22:12, 4.04it/s, grad_norm=0.447, loss_final=0.48, loss_mean=0.83, loss_mean_cls=0.0678, proj_loss=-0.417][2026-03-26 12:56:05] Step: 6030, Training Logs: loss_final: 0.478544, loss_mean: 0.835081, proj_loss: -0.424562, loss_mean_cls: 0.068025, grad_norm: 0.422109 +Steps: 1%| | 6031/1000000 [25:02<68:23:37, 4.04it/s, grad_norm=0.422, loss_final=0.479, loss_mean=0.835, loss_mean_cls=0.068, proj_loss=-0.425][2026-03-26 12:56:06] Step: 6031, Training Logs: loss_final: 0.482722, loss_mean: 0.813687, proj_loss: -0.404065, loss_mean_cls: 0.073099, grad_norm: 0.413684 +Steps: 1%| | 6032/1000000 [25:02<68:22:07, 4.04it/s, grad_norm=0.414, loss_final=0.483, loss_mean=0.814, loss_mean_cls=0.0731, proj_loss=-0.404][2026-03-26 12:56:06] Step: 6032, Training Logs: loss_final: 0.485973, loss_mean: 0.831941, proj_loss: -0.415438, loss_mean_cls: 0.069469, grad_norm: 0.478266 +Steps: 1%| | 6033/1000000 [25:03<68:22:00, 4.04it/s, grad_norm=0.478, loss_final=0.486, loss_mean=0.832, loss_mean_cls=0.0695, proj_loss=-0.415][2026-03-26 12:56:06] Step: 6033, Training Logs: loss_final: 0.469729, loss_mean: 0.824131, proj_loss: -0.421013, loss_mean_cls: 0.066610, grad_norm: 0.331783 +Steps: 1%| | 6034/1000000 [25:03<68:23:45, 4.04it/s, grad_norm=0.332, loss_final=0.47, loss_mean=0.824, loss_mean_cls=0.0666, proj_loss=-0.421][2026-03-26 12:56:06] Step: 6034, Training Logs: loss_final: 0.471175, loss_mean: 0.807143, proj_loss: -0.406997, loss_mean_cls: 0.071028, grad_norm: 0.294369 +Steps: 1%| | 6035/1000000 [25:03<68:20:08, 4.04it/s, grad_norm=0.294, loss_final=0.471, loss_mean=0.807, loss_mean_cls=0.071, proj_loss=-0.407][2026-03-26 12:56:07] Step: 6035, Training Logs: loss_final: 0.479306, loss_mean: 0.824473, proj_loss: -0.414160, loss_mean_cls: 0.068992, grad_norm: 0.293427 +Steps: 1%| | 6036/1000000 [25:03<68:19:22, 4.04it/s, grad_norm=0.293, loss_final=0.479, loss_mean=0.824, loss_mean_cls=0.069, proj_loss=-0.414][2026-03-26 12:56:07] Step: 6036, Training Logs: loss_final: 0.461275, loss_mean: 0.815025, proj_loss: -0.421562, loss_mean_cls: 0.067812, grad_norm: 0.328101 +Steps: 1%| | 6037/1000000 [25:04<68:23:07, 4.04it/s, grad_norm=0.328, loss_final=0.461, loss_mean=0.815, loss_mean_cls=0.0678, proj_loss=-0.422][2026-03-26 12:56:07] Step: 6037, Training Logs: loss_final: 0.498397, loss_mean: 0.849750, proj_loss: -0.419902, loss_mean_cls: 0.068549, grad_norm: 0.315083 +Steps: 1%| | 6038/1000000 [25:04<68:19:48, 4.04it/s, grad_norm=0.315, loss_final=0.498, loss_mean=0.85, loss_mean_cls=0.0685, proj_loss=-0.42][2026-03-26 12:56:07] Step: 6038, Training Logs: loss_final: 0.483900, loss_mean: 0.840021, proj_loss: -0.423867, loss_mean_cls: 0.067745, grad_norm: 0.371322 +Steps: 1%| | 6039/1000000 [25:04<68:20:33, 4.04it/s, grad_norm=0.371, loss_final=0.484, loss_mean=0.84, loss_mean_cls=0.0677, proj_loss=-0.424][2026-03-26 12:56:08] Step: 6039, Training Logs: loss_final: 0.483098, loss_mean: 0.824652, proj_loss: -0.412167, loss_mean_cls: 0.070613, grad_norm: 0.443458 +Steps: 1%| | 6040/1000000 [25:04<68:19:32, 4.04it/s, grad_norm=0.443, loss_final=0.483, loss_mean=0.825, loss_mean_cls=0.0706, proj_loss=-0.412][2026-03-26 12:56:08] Step: 6040, Training Logs: loss_final: 0.485354, loss_mean: 0.842732, proj_loss: -0.424940, loss_mean_cls: 0.067562, grad_norm: 0.357912 +Steps: 1%| | 6041/1000000 [25:05<68:21:33, 4.04it/s, grad_norm=0.358, loss_final=0.485, loss_mean=0.843, loss_mean_cls=0.0676, proj_loss=-0.425][2026-03-26 12:56:08] Step: 6041, Training Logs: loss_final: 0.487005, loss_mean: 0.828590, proj_loss: -0.411262, loss_mean_cls: 0.069677, grad_norm: 0.403216 +Steps: 1%| | 6042/1000000 [25:05<68:22:02, 4.04it/s, grad_norm=0.403, loss_final=0.487, loss_mean=0.829, loss_mean_cls=0.0697, proj_loss=-0.411][2026-03-26 12:56:08] Step: 6042, Training Logs: loss_final: 0.480890, loss_mean: 0.829309, proj_loss: -0.417991, loss_mean_cls: 0.069571, grad_norm: 0.237157 +Steps: 1%| | 6043/1000000 [25:05<68:22:39, 4.04it/s, grad_norm=0.237, loss_final=0.481, loss_mean=0.829, loss_mean_cls=0.0696, proj_loss=-0.418][2026-03-26 12:56:09] Step: 6043, Training Logs: loss_final: 0.476810, loss_mean: 0.829234, proj_loss: -0.421448, loss_mean_cls: 0.069024, grad_norm: 0.301666 +Steps: 1%| | 6044/1000000 [25:05<68:19:43, 4.04it/s, grad_norm=0.302, loss_final=0.477, loss_mean=0.829, loss_mean_cls=0.069, proj_loss=-0.421][2026-03-26 12:56:09] Step: 6044, Training Logs: loss_final: 0.497080, loss_mean: 0.829163, proj_loss: -0.403073, loss_mean_cls: 0.070990, grad_norm: 0.285391 +Steps: 1%| | 6045/1000000 [25:06<68:21:39, 4.04it/s, grad_norm=0.285, loss_final=0.497, loss_mean=0.829, loss_mean_cls=0.071, proj_loss=-0.403][2026-03-26 12:56:09] Step: 6045, Training Logs: loss_final: 0.472128, loss_mean: 0.818750, proj_loss: -0.416398, loss_mean_cls: 0.069776, grad_norm: 0.284918 +Steps: 1%| | 6046/1000000 [25:06<68:21:44, 4.04it/s, grad_norm=0.285, loss_final=0.472, loss_mean=0.819, loss_mean_cls=0.0698, proj_loss=-0.416][2026-03-26 12:56:09] Step: 6046, Training Logs: loss_final: 0.471397, loss_mean: 0.823361, proj_loss: -0.420003, loss_mean_cls: 0.068039, grad_norm: 0.263272 +Steps: 1%| | 6047/1000000 [25:06<68:37:15, 4.02it/s, grad_norm=0.263, loss_final=0.471, loss_mean=0.823, loss_mean_cls=0.068, proj_loss=-0.42][2026-03-26 12:56:10] Step: 6047, Training Logs: loss_final: 0.483170, loss_mean: 0.832092, proj_loss: -0.417407, loss_mean_cls: 0.068485, grad_norm: 0.323804 +Steps: 1%| | 6048/1000000 [25:06<68:33:10, 4.03it/s, grad_norm=0.324, loss_final=0.483, loss_mean=0.832, loss_mean_cls=0.0685, proj_loss=-0.417][2026-03-26 12:56:10] Step: 6048, Training Logs: loss_final: 0.477109, loss_mean: 0.816768, proj_loss: -0.411654, loss_mean_cls: 0.071995, grad_norm: 0.302959 +Steps: 1%| | 6049/1000000 [25:07<68:27:26, 4.03it/s, grad_norm=0.303, loss_final=0.477, loss_mean=0.817, loss_mean_cls=0.072, proj_loss=-0.412][2026-03-26 12:56:10] Step: 6049, Training Logs: loss_final: 0.483299, loss_mean: 0.817698, proj_loss: -0.406172, loss_mean_cls: 0.071773, grad_norm: 0.284437 +Steps: 1%| | 6050/1000000 [25:07<68:26:22, 4.03it/s, grad_norm=0.284, loss_final=0.483, loss_mean=0.818, loss_mean_cls=0.0718, proj_loss=-0.406][2026-03-26 12:56:10] Step: 6050, Training Logs: loss_final: 0.474070, loss_mean: 0.811670, proj_loss: -0.409264, loss_mean_cls: 0.071664, grad_norm: 0.379465 +Steps: 1%| | 6051/1000000 [25:07<68:24:44, 4.04it/s, grad_norm=0.379, loss_final=0.474, loss_mean=0.812, loss_mean_cls=0.0717, proj_loss=-0.409][2026-03-26 12:56:11] Step: 6051, Training Logs: loss_final: 0.485168, loss_mean: 0.833189, proj_loss: -0.417326, loss_mean_cls: 0.069305, grad_norm: 0.291689 +Steps: 1%| | 6052/1000000 [25:07<68:24:12, 4.04it/s, grad_norm=0.292, loss_final=0.485, loss_mean=0.833, loss_mean_cls=0.0693, proj_loss=-0.417][2026-03-26 12:56:11] Step: 6052, Training Logs: loss_final: 0.483351, loss_mean: 0.829100, proj_loss: -0.415543, loss_mean_cls: 0.069794, grad_norm: 0.272916 +Steps: 1%| | 6053/1000000 [25:08<68:23:03, 4.04it/s, grad_norm=0.273, loss_final=0.483, loss_mean=0.829, loss_mean_cls=0.0698, proj_loss=-0.416][2026-03-26 12:56:11] Step: 6053, Training Logs: loss_final: 0.478542, loss_mean: 0.817167, proj_loss: -0.409785, loss_mean_cls: 0.071160, grad_norm: 0.478009 +Steps: 1%| | 6054/1000000 [25:08<68:23:06, 4.04it/s, grad_norm=0.478, loss_final=0.479, loss_mean=0.817, loss_mean_cls=0.0712, proj_loss=-0.41][2026-03-26 12:56:11] Step: 6054, Training Logs: loss_final: 0.455015, loss_mean: 0.805784, proj_loss: -0.419988, loss_mean_cls: 0.069219, grad_norm: 0.225318 +Steps: 1%| | 6055/1000000 [25:08<68:21:07, 4.04it/s, grad_norm=0.225, loss_final=0.455, loss_mean=0.806, loss_mean_cls=0.0692, proj_loss=-0.42][2026-03-26 12:56:11] Step: 6055, Training Logs: loss_final: 0.491609, loss_mean: 0.828414, proj_loss: -0.407745, loss_mean_cls: 0.070940, grad_norm: 0.459855 +Steps: 1%| | 6056/1000000 [25:08<68:37:23, 4.02it/s, grad_norm=0.46, loss_final=0.492, loss_mean=0.828, loss_mean_cls=0.0709, proj_loss=-0.408][2026-03-26 12:56:12] Step: 6056, Training Logs: loss_final: 0.481725, loss_mean: 0.831243, proj_loss: -0.417853, loss_mean_cls: 0.068335, grad_norm: 0.417710 +Steps: 1%| | 6057/1000000 [25:09<68:32:11, 4.03it/s, grad_norm=0.418, loss_final=0.482, loss_mean=0.831, loss_mean_cls=0.0683, proj_loss=-0.418][2026-03-26 12:56:12] Step: 6057, Training Logs: loss_final: 0.473358, loss_mean: 0.808108, proj_loss: -0.406116, loss_mean_cls: 0.071366, grad_norm: 0.279708 +Steps: 1%| | 6058/1000000 [25:09<68:30:04, 4.03it/s, grad_norm=0.28, loss_final=0.473, loss_mean=0.808, loss_mean_cls=0.0714, proj_loss=-0.406][2026-03-26 12:56:12] Step: 6058, Training Logs: loss_final: 0.487815, loss_mean: 0.834116, proj_loss: -0.415904, loss_mean_cls: 0.069603, grad_norm: 0.438856 +Steps: 1%| | 6059/1000000 [25:09<68:23:03, 4.04it/s, grad_norm=0.439, loss_final=0.488, loss_mean=0.834, loss_mean_cls=0.0696, proj_loss=-0.416][2026-03-26 12:56:12] Step: 6059, Training Logs: loss_final: 0.481513, loss_mean: 0.829385, proj_loss: -0.416370, loss_mean_cls: 0.068498, grad_norm: 0.347892 +Steps: 1%| | 6060/1000000 [25:09<68:24:20, 4.04it/s, grad_norm=0.348, loss_final=0.482, loss_mean=0.829, loss_mean_cls=0.0685, proj_loss=-0.416][2026-03-26 12:56:13] Step: 6060, Training Logs: loss_final: 0.465653, loss_mean: 0.806693, proj_loss: -0.411590, loss_mean_cls: 0.070549, grad_norm: 0.423734 +Steps: 1%| | 6061/1000000 [25:10<68:17:26, 4.04it/s, grad_norm=0.424, loss_final=0.466, loss_mean=0.807, loss_mean_cls=0.0705, proj_loss=-0.412][2026-03-26 12:56:13] Step: 6061, Training Logs: loss_final: 0.485076, loss_mean: 0.832184, proj_loss: -0.416207, loss_mean_cls: 0.069099, grad_norm: 0.358157 +Steps: 1%| | 6062/1000000 [25:10<68:20:26, 4.04it/s, grad_norm=0.358, loss_final=0.485, loss_mean=0.832, loss_mean_cls=0.0691, proj_loss=-0.416][2026-03-26 12:56:13] Step: 6062, Training Logs: loss_final: 0.471956, loss_mean: 0.813435, proj_loss: -0.412259, loss_mean_cls: 0.070780, grad_norm: 0.252626 +Steps: 1%| | 6063/1000000 [25:10<68:16:04, 4.04it/s, grad_norm=0.253, loss_final=0.472, loss_mean=0.813, loss_mean_cls=0.0708, proj_loss=-0.412][2026-03-26 12:56:13] Step: 6063, Training Logs: loss_final: 0.485960, loss_mean: 0.836499, proj_loss: -0.418800, loss_mean_cls: 0.068261, grad_norm: 0.336800 +Steps: 1%| | 6064/1000000 [25:10<68:24:33, 4.04it/s, grad_norm=0.337, loss_final=0.486, loss_mean=0.836, loss_mean_cls=0.0683, proj_loss=-0.419][2026-03-26 12:56:14] Step: 6064, Training Logs: loss_final: 0.504739, loss_mean: 0.850250, proj_loss: -0.415013, loss_mean_cls: 0.069502, grad_norm: 0.261526 +Steps: 1%| | 6065/1000000 [25:11<68:24:06, 4.04it/s, grad_norm=0.262, loss_final=0.505, loss_mean=0.85, loss_mean_cls=0.0695, proj_loss=-0.415][2026-03-26 12:56:14] Step: 6065, Training Logs: loss_final: 0.469134, loss_mean: 0.821892, proj_loss: -0.421080, loss_mean_cls: 0.068323, grad_norm: 0.453944 +Steps: 1%| | 6066/1000000 [25:11<68:27:03, 4.03it/s, grad_norm=0.454, loss_final=0.469, loss_mean=0.822, loss_mean_cls=0.0683, proj_loss=-0.421][2026-03-26 12:56:14] Step: 6066, Training Logs: loss_final: 0.480165, loss_mean: 0.808462, proj_loss: -0.401228, loss_mean_cls: 0.072931, grad_norm: 0.446223 +Steps: 1%| | 6067/1000000 [25:11<68:28:07, 4.03it/s, grad_norm=0.446, loss_final=0.48, loss_mean=0.808, loss_mean_cls=0.0729, proj_loss=-0.401][2026-03-26 12:56:14] Step: 6067, Training Logs: loss_final: 0.475818, loss_mean: 0.811032, proj_loss: -0.406179, loss_mean_cls: 0.070965, grad_norm: 0.343141 +Steps: 1%| | 6068/1000000 [25:11<68:28:55, 4.03it/s, grad_norm=0.343, loss_final=0.476, loss_mean=0.811, loss_mean_cls=0.071, proj_loss=-0.406][2026-03-26 12:56:15] Step: 6068, Training Logs: loss_final: 0.476730, loss_mean: 0.822333, proj_loss: -0.415130, loss_mean_cls: 0.069528, grad_norm: 0.414320 +Steps: 1%| | 6069/1000000 [25:12<68:27:01, 4.03it/s, grad_norm=0.414, loss_final=0.477, loss_mean=0.822, loss_mean_cls=0.0695, proj_loss=-0.415][2026-03-26 12:56:15] Step: 6069, Training Logs: loss_final: 0.479195, loss_mean: 0.824962, proj_loss: -0.416114, loss_mean_cls: 0.070348, grad_norm: 0.270736 +Steps: 1%| | 6070/1000000 [25:12<68:28:30, 4.03it/s, grad_norm=0.271, loss_final=0.479, loss_mean=0.825, loss_mean_cls=0.0703, proj_loss=-0.416][2026-03-26 12:56:15] Step: 6070, Training Logs: loss_final: 0.475354, loss_mean: 0.827169, proj_loss: -0.419674, loss_mean_cls: 0.067858, grad_norm: 0.442010 +Steps: 1%| | 6071/1000000 [25:12<68:28:04, 4.03it/s, grad_norm=0.442, loss_final=0.475, loss_mean=0.827, loss_mean_cls=0.0679, proj_loss=-0.42][2026-03-26 12:56:15] Step: 6071, Training Logs: loss_final: 0.462878, loss_mean: 0.812022, proj_loss: -0.418364, loss_mean_cls: 0.069219, grad_norm: 0.203200 +Steps: 1%| | 6072/1000000 [25:12<68:27:54, 4.03it/s, grad_norm=0.203, loss_final=0.463, loss_mean=0.812, loss_mean_cls=0.0692, proj_loss=-0.418][2026-03-26 12:56:16] Step: 6072, Training Logs: loss_final: 0.474419, loss_mean: 0.814955, proj_loss: -0.410330, loss_mean_cls: 0.069795, grad_norm: 0.651016 +Steps: 1%| | 6073/1000000 [25:13<68:24:04, 4.04it/s, grad_norm=0.651, loss_final=0.474, loss_mean=0.815, loss_mean_cls=0.0698, proj_loss=-0.41][2026-03-26 12:56:16] Step: 6073, Training Logs: loss_final: 0.464879, loss_mean: 0.816181, proj_loss: -0.419763, loss_mean_cls: 0.068462, grad_norm: 0.388272 +Steps: 1%| | 6074/1000000 [25:13<68:26:00, 4.03it/s, grad_norm=0.388, loss_final=0.465, loss_mean=0.816, loss_mean_cls=0.0685, proj_loss=-0.42][2026-03-26 12:56:16] Step: 6074, Training Logs: loss_final: 0.483417, loss_mean: 0.824023, proj_loss: -0.410456, loss_mean_cls: 0.069850, grad_norm: 0.537035 +Steps: 1%| | 6075/1000000 [25:13<68:24:55, 4.04it/s, grad_norm=0.537, loss_final=0.483, loss_mean=0.824, loss_mean_cls=0.0698, proj_loss=-0.41][2026-03-26 12:56:16] Step: 6075, Training Logs: loss_final: 0.495647, loss_mean: 0.843431, proj_loss: -0.416489, loss_mean_cls: 0.068705, grad_norm: 0.561725 +Steps: 1%| | 6076/1000000 [25:13<68:23:30, 4.04it/s, grad_norm=0.562, loss_final=0.496, loss_mean=0.843, loss_mean_cls=0.0687, proj_loss=-0.416][2026-03-26 12:56:17] Step: 6076, Training Logs: loss_final: 0.484218, loss_mean: 0.831193, proj_loss: -0.416234, loss_mean_cls: 0.069259, grad_norm: 0.450102 +Steps: 1%| | 6077/1000000 [25:14<68:24:02, 4.04it/s, grad_norm=0.45, loss_final=0.484, loss_mean=0.831, loss_mean_cls=0.0693, proj_loss=-0.416][2026-03-26 12:56:17] Step: 6077, Training Logs: loss_final: 0.469837, loss_mean: 0.816762, proj_loss: -0.416500, loss_mean_cls: 0.069575, grad_norm: 0.766478 +Steps: 1%| | 6078/1000000 [25:14<68:23:05, 4.04it/s, grad_norm=0.766, loss_final=0.47, loss_mean=0.817, loss_mean_cls=0.0696, proj_loss=-0.416][2026-03-26 12:56:17] Step: 6078, Training Logs: loss_final: 0.485463, loss_mean: 0.827409, proj_loss: -0.410933, loss_mean_cls: 0.068987, grad_norm: 0.313060 +Steps: 1%| | 6079/1000000 [25:14<68:23:32, 4.04it/s, grad_norm=0.313, loss_final=0.485, loss_mean=0.827, loss_mean_cls=0.069, proj_loss=-0.411][2026-03-26 12:56:17] Step: 6079, Training Logs: loss_final: 0.499154, loss_mean: 0.829139, proj_loss: -0.403354, loss_mean_cls: 0.073369, grad_norm: 0.881841 +Steps: 1%| | 6080/1000000 [25:14<68:22:02, 4.04it/s, grad_norm=0.882, loss_final=0.499, loss_mean=0.829, loss_mean_cls=0.0734, proj_loss=-0.403][2026-03-26 12:56:18] Step: 6080, Training Logs: loss_final: 0.488090, loss_mean: 0.844142, proj_loss: -0.422466, loss_mean_cls: 0.066414, grad_norm: 0.362527 +Steps: 1%| | 6081/1000000 [25:15<68:22:46, 4.04it/s, grad_norm=0.363, loss_final=0.488, loss_mean=0.844, loss_mean_cls=0.0664, proj_loss=-0.422][2026-03-26 12:56:18] Step: 6081, Training Logs: loss_final: 0.466770, loss_mean: 0.808652, proj_loss: -0.412226, loss_mean_cls: 0.070344, grad_norm: 0.808832 +Steps: 1%| | 6082/1000000 [25:15<68:20:19, 4.04it/s, grad_norm=0.809, loss_final=0.467, loss_mean=0.809, loss_mean_cls=0.0703, proj_loss=-0.412][2026-03-26 12:56:18] Step: 6082, Training Logs: loss_final: 0.489888, loss_mean: 0.834291, proj_loss: -0.414120, loss_mean_cls: 0.069718, grad_norm: 0.697459 +Steps: 1%| | 6083/1000000 [25:15<68:21:26, 4.04it/s, grad_norm=0.697, loss_final=0.49, loss_mean=0.834, loss_mean_cls=0.0697, proj_loss=-0.414][2026-03-26 12:56:18] Step: 6083, Training Logs: loss_final: 0.474682, loss_mean: 0.832069, proj_loss: -0.424063, loss_mean_cls: 0.066676, grad_norm: 0.443468 +Steps: 1%| | 6084/1000000 [25:15<68:20:47, 4.04it/s, grad_norm=0.443, loss_final=0.475, loss_mean=0.832, loss_mean_cls=0.0667, proj_loss=-0.424][2026-03-26 12:56:19] Step: 6084, Training Logs: loss_final: 0.478130, loss_mean: 0.821940, proj_loss: -0.413505, loss_mean_cls: 0.069694, grad_norm: 0.875159 +Steps: 1%| | 6085/1000000 [25:16<68:19:17, 4.04it/s, grad_norm=0.875, loss_final=0.478, loss_mean=0.822, loss_mean_cls=0.0697, proj_loss=-0.414][2026-03-26 12:56:19] Step: 6085, Training Logs: loss_final: 0.480618, loss_mean: 0.827381, proj_loss: -0.415909, loss_mean_cls: 0.069145, grad_norm: 0.469856 +Steps: 1%| | 6086/1000000 [25:16<68:18:27, 4.04it/s, grad_norm=0.47, loss_final=0.481, loss_mean=0.827, loss_mean_cls=0.0691, proj_loss=-0.416][2026-03-26 12:56:19] Step: 6086, Training Logs: loss_final: 0.477967, loss_mean: 0.828371, proj_loss: -0.419340, loss_mean_cls: 0.068936, grad_norm: 0.910953 +Steps: 1%| | 6087/1000000 [25:16<68:18:25, 4.04it/s, grad_norm=0.911, loss_final=0.478, loss_mean=0.828, loss_mean_cls=0.0689, proj_loss=-0.419][2026-03-26 12:56:19] Step: 6087, Training Logs: loss_final: 0.483577, loss_mean: 0.835122, proj_loss: -0.419690, loss_mean_cls: 0.068146, grad_norm: 0.316853 +Steps: 1%| | 6088/1000000 [25:16<68:19:54, 4.04it/s, grad_norm=0.317, loss_final=0.484, loss_mean=0.835, loss_mean_cls=0.0681, proj_loss=-0.42][2026-03-26 12:56:20] Step: 6088, Training Logs: loss_final: 0.490036, loss_mean: 0.835574, proj_loss: -0.414959, loss_mean_cls: 0.069421, grad_norm: 0.713504 +Steps: 1%| | 6089/1000000 [25:17<68:19:07, 4.04it/s, grad_norm=0.714, loss_final=0.49, loss_mean=0.836, loss_mean_cls=0.0694, proj_loss=-0.415][2026-03-26 12:56:20] Step: 6089, Training Logs: loss_final: 0.493546, loss_mean: 0.836962, proj_loss: -0.413594, loss_mean_cls: 0.070177, grad_norm: 0.272748 +Steps: 1%| | 6090/1000000 [25:17<68:23:56, 4.04it/s, grad_norm=0.273, loss_final=0.494, loss_mean=0.837, loss_mean_cls=0.0702, proj_loss=-0.414][2026-03-26 12:56:20] Step: 6090, Training Logs: loss_final: 0.477292, loss_mean: 0.824708, proj_loss: -0.416772, loss_mean_cls: 0.069357, grad_norm: 0.881851 +Steps: 1%| | 6091/1000000 [25:17<68:31:01, 4.03it/s, grad_norm=0.882, loss_final=0.477, loss_mean=0.825, loss_mean_cls=0.0694, proj_loss=-0.417][2026-03-26 12:56:20] Step: 6091, Training Logs: loss_final: 0.495524, loss_mean: 0.838743, proj_loss: -0.413155, loss_mean_cls: 0.069936, grad_norm: 0.414756 +Steps: 1%| | 6092/1000000 [25:17<68:27:27, 4.03it/s, grad_norm=0.415, loss_final=0.496, loss_mean=0.839, loss_mean_cls=0.0699, proj_loss=-0.413][2026-03-26 12:56:21] Step: 6092, Training Logs: loss_final: 0.475965, loss_mean: 0.823893, proj_loss: -0.416733, loss_mean_cls: 0.068805, grad_norm: 0.743184 +Steps: 1%| | 6093/1000000 [25:18<68:25:17, 4.04it/s, grad_norm=0.743, loss_final=0.476, loss_mean=0.824, loss_mean_cls=0.0688, proj_loss=-0.417][2026-03-26 12:56:21] Step: 6093, Training Logs: loss_final: 0.478756, loss_mean: 0.823363, proj_loss: -0.414755, loss_mean_cls: 0.070148, grad_norm: 0.445849 +Steps: 1%| | 6094/1000000 [25:18<68:26:27, 4.03it/s, grad_norm=0.446, loss_final=0.479, loss_mean=0.823, loss_mean_cls=0.0701, proj_loss=-0.415][2026-03-26 12:56:21] Step: 6094, Training Logs: loss_final: 0.479451, loss_mean: 0.834086, proj_loss: -0.422342, loss_mean_cls: 0.067706, grad_norm: 0.619598 +Steps: 1%| | 6095/1000000 [25:18<68:25:16, 4.04it/s, grad_norm=0.62, loss_final=0.479, loss_mean=0.834, loss_mean_cls=0.0677, proj_loss=-0.422][2026-03-26 12:56:21] Step: 6095, Training Logs: loss_final: 0.474894, loss_mean: 0.824950, proj_loss: -0.418607, loss_mean_cls: 0.068551, grad_norm: 0.417996 +Steps: 1%| | 6096/1000000 [25:18<68:26:33, 4.03it/s, grad_norm=0.418, loss_final=0.475, loss_mean=0.825, loss_mean_cls=0.0686, proj_loss=-0.419][2026-03-26 12:56:22] Step: 6096, Training Logs: loss_final: 0.475012, loss_mean: 0.816805, proj_loss: -0.411693, loss_mean_cls: 0.069900, grad_norm: 0.642070 +Steps: 1%| | 6097/1000000 [25:19<68:28:23, 4.03it/s, grad_norm=0.642, loss_final=0.475, loss_mean=0.817, loss_mean_cls=0.0699, proj_loss=-0.412][2026-03-26 12:56:22] Step: 6097, Training Logs: loss_final: 0.470804, loss_mean: 0.820009, proj_loss: -0.417825, loss_mean_cls: 0.068621, grad_norm: 0.423951 +Steps: 1%| | 6098/1000000 [25:19<68:25:17, 4.04it/s, grad_norm=0.424, loss_final=0.471, loss_mean=0.82, loss_mean_cls=0.0686, proj_loss=-0.418][2026-03-26 12:56:22] Step: 6098, Training Logs: loss_final: 0.468615, loss_mean: 0.806767, proj_loss: -0.409610, loss_mean_cls: 0.071458, grad_norm: 0.754200 +Steps: 1%| | 6099/1000000 [25:19<68:23:42, 4.04it/s, grad_norm=0.754, loss_final=0.469, loss_mean=0.807, loss_mean_cls=0.0715, proj_loss=-0.41][2026-03-26 12:56:22] Step: 6099, Training Logs: loss_final: 0.501820, loss_mean: 0.857317, proj_loss: -0.422981, loss_mean_cls: 0.067484, grad_norm: 0.481143 +Steps: 1%| | 6100/1000000 [25:19<68:25:59, 4.03it/s, grad_norm=0.481, loss_final=0.502, loss_mean=0.857, loss_mean_cls=0.0675, proj_loss=-0.423][2026-03-26 12:56:23] Step: 6100, Training Logs: loss_final: 0.484433, loss_mean: 0.827686, proj_loss: -0.413048, loss_mean_cls: 0.069795, grad_norm: 0.537418 +Steps: 1%| | 6101/1000000 [25:20<68:29:10, 4.03it/s, grad_norm=0.537, loss_final=0.484, loss_mean=0.828, loss_mean_cls=0.0698, proj_loss=-0.413][2026-03-26 12:56:23] Step: 6101, Training Logs: loss_final: 0.481097, loss_mean: 0.835906, proj_loss: -0.422109, loss_mean_cls: 0.067300, grad_norm: 0.274348 +Steps: 1%| | 6102/1000000 [25:20<68:23:51, 4.04it/s, grad_norm=0.274, loss_final=0.481, loss_mean=0.836, loss_mean_cls=0.0673, proj_loss=-0.422][2026-03-26 12:56:23] Step: 6102, Training Logs: loss_final: 0.476791, loss_mean: 0.824774, proj_loss: -0.417068, loss_mean_cls: 0.069085, grad_norm: 0.351249 +Steps: 1%| | 6103/1000000 [25:20<68:22:38, 4.04it/s, grad_norm=0.351, loss_final=0.477, loss_mean=0.825, loss_mean_cls=0.0691, proj_loss=-0.417][2026-03-26 12:56:23] Step: 6103, Training Logs: loss_final: 0.474033, loss_mean: 0.816594, proj_loss: -0.413581, loss_mean_cls: 0.071020, grad_norm: 0.414406 +Steps: 1%| | 6104/1000000 [25:20<68:24:44, 4.04it/s, grad_norm=0.414, loss_final=0.474, loss_mean=0.817, loss_mean_cls=0.071, proj_loss=-0.414][2026-03-26 12:56:24] Step: 6104, Training Logs: loss_final: 0.461549, loss_mean: 0.811449, proj_loss: -0.418458, loss_mean_cls: 0.068558, grad_norm: 0.289312 +Steps: 1%| | 6105/1000000 [25:20<68:25:09, 4.04it/s, grad_norm=0.289, loss_final=0.462, loss_mean=0.811, loss_mean_cls=0.0686, proj_loss=-0.418][2026-03-26 12:56:24] Step: 6105, Training Logs: loss_final: 0.459105, loss_mean: 0.800937, proj_loss: -0.412377, loss_mean_cls: 0.070544, grad_norm: 0.385143 +Steps: 1%| | 6106/1000000 [25:21<68:24:19, 4.04it/s, grad_norm=0.385, loss_final=0.459, loss_mean=0.801, loss_mean_cls=0.0705, proj_loss=-0.412][2026-03-26 12:56:24] Step: 6106, Training Logs: loss_final: 0.485792, loss_mean: 0.836997, proj_loss: -0.419491, loss_mean_cls: 0.068286, grad_norm: 0.220331 +Steps: 1%| | 6107/1000000 [25:21<68:21:21, 4.04it/s, grad_norm=0.22, loss_final=0.486, loss_mean=0.837, loss_mean_cls=0.0683, proj_loss=-0.419][2026-03-26 12:56:24] Step: 6107, Training Logs: loss_final: 0.462544, loss_mean: 0.813082, proj_loss: -0.419346, loss_mean_cls: 0.068808, grad_norm: 0.314819 +Steps: 1%| | 6108/1000000 [25:21<68:26:07, 4.03it/s, grad_norm=0.315, loss_final=0.463, loss_mean=0.813, loss_mean_cls=0.0688, proj_loss=-0.419][2026-03-26 12:56:25] Step: 6108, Training Logs: loss_final: 0.468599, loss_mean: 0.824074, proj_loss: -0.422394, loss_mean_cls: 0.066920, grad_norm: 0.343160 +Steps: 1%| | 6109/1000000 [25:21<68:22:21, 4.04it/s, grad_norm=0.343, loss_final=0.469, loss_mean=0.824, loss_mean_cls=0.0669, proj_loss=-0.422][2026-03-26 12:56:25] Step: 6109, Training Logs: loss_final: 0.486126, loss_mean: 0.828085, proj_loss: -0.411663, loss_mean_cls: 0.069703, grad_norm: 0.382604 +Steps: 1%| | 6110/1000000 [25:22<68:21:21, 4.04it/s, grad_norm=0.383, loss_final=0.486, loss_mean=0.828, loss_mean_cls=0.0697, proj_loss=-0.412][2026-03-26 12:56:25] Step: 6110, Training Logs: loss_final: 0.489194, loss_mean: 0.831928, proj_loss: -0.412631, loss_mean_cls: 0.069897, grad_norm: 0.285835 +Steps: 1%| | 6111/1000000 [25:22<68:21:06, 4.04it/s, grad_norm=0.286, loss_final=0.489, loss_mean=0.832, loss_mean_cls=0.0699, proj_loss=-0.413][2026-03-26 12:56:25] Step: 6111, Training Logs: loss_final: 0.468147, loss_mean: 0.814845, proj_loss: -0.415381, loss_mean_cls: 0.068683, grad_norm: 0.255980 +Steps: 1%| | 6112/1000000 [25:22<68:23:21, 4.04it/s, grad_norm=0.256, loss_final=0.468, loss_mean=0.815, loss_mean_cls=0.0687, proj_loss=-0.415][2026-03-26 12:56:26] Step: 6112, Training Logs: loss_final: 0.485218, loss_mean: 0.832454, proj_loss: -0.416545, loss_mean_cls: 0.069309, grad_norm: 0.322236 +Steps: 1%| | 6113/1000000 [25:22<68:25:52, 4.03it/s, grad_norm=0.322, loss_final=0.485, loss_mean=0.832, loss_mean_cls=0.0693, proj_loss=-0.417][2026-03-26 12:56:26] Step: 6113, Training Logs: loss_final: 0.474209, loss_mean: 0.832442, proj_loss: -0.424479, loss_mean_cls: 0.066247, grad_norm: 0.291265 +Steps: 1%| | 6114/1000000 [25:23<68:27:12, 4.03it/s, grad_norm=0.291, loss_final=0.474, loss_mean=0.832, loss_mean_cls=0.0662, proj_loss=-0.424][2026-03-26 12:56:26] Step: 6114, Training Logs: loss_final: 0.478024, loss_mean: 0.816501, proj_loss: -0.409615, loss_mean_cls: 0.071138, grad_norm: 0.249346 +Steps: 1%| | 6115/1000000 [25:23<68:24:50, 4.04it/s, grad_norm=0.249, loss_final=0.478, loss_mean=0.817, loss_mean_cls=0.0711, proj_loss=-0.41][2026-03-26 12:56:26] Step: 6115, Training Logs: loss_final: 0.472803, loss_mean: 0.820236, proj_loss: -0.417311, loss_mean_cls: 0.069879, grad_norm: 0.334531 +Steps: 1%| | 6116/1000000 [25:23<68:24:22, 4.04it/s, grad_norm=0.335, loss_final=0.473, loss_mean=0.82, loss_mean_cls=0.0699, proj_loss=-0.417][2026-03-26 12:56:27] Step: 6116, Training Logs: loss_final: 0.470481, loss_mean: 0.817943, proj_loss: -0.415458, loss_mean_cls: 0.067996, grad_norm: 0.285073 +Steps: 1%| | 6117/1000000 [25:23<68:08:28, 4.05it/s, grad_norm=0.285, loss_final=0.47, loss_mean=0.818, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 12:56:27] Step: 6117, Training Logs: loss_final: 0.466990, loss_mean: 0.810179, proj_loss: -0.413311, loss_mean_cls: 0.070122, grad_norm: inf +Steps: 1%| | 6118/1000000 [25:24<67:03:06, 4.12it/s, grad_norm=inf, loss_final=0.467, loss_mean=0.81, loss_mean_cls=0.0701, proj_loss=-0.413][2026-03-26 12:56:27] Step: 6118, Training Logs: loss_final: 0.483121, loss_mean: 0.837496, proj_loss: -0.421945, loss_mean_cls: 0.067569, grad_norm: 0.470578 +Steps: 1%| | 6119/1000000 [25:24<67:28:17, 4.09it/s, grad_norm=0.471, loss_final=0.483, loss_mean=0.837, loss_mean_cls=0.0676, proj_loss=-0.422][2026-03-26 12:56:27] Step: 6119, Training Logs: loss_final: 0.481712, loss_mean: 0.828830, proj_loss: -0.416308, loss_mean_cls: 0.069190, grad_norm: 0.470532 +Steps: 1%| | 6120/1000000 [25:24<67:39:50, 4.08it/s, grad_norm=0.471, loss_final=0.482, loss_mean=0.829, loss_mean_cls=0.0692, proj_loss=-0.416][2026-03-26 12:56:28] Step: 6120, Training Logs: loss_final: 0.472921, loss_mean: 0.824518, proj_loss: -0.419874, loss_mean_cls: 0.068276, grad_norm: 0.503389 +Steps: 1%| | 6121/1000000 [25:24<67:51:16, 4.07it/s, grad_norm=0.503, loss_final=0.473, loss_mean=0.825, loss_mean_cls=0.0683, proj_loss=-0.42][2026-03-26 12:56:28] Step: 6121, Training Logs: loss_final: 0.464559, loss_mean: 0.802491, proj_loss: -0.408944, loss_mean_cls: 0.071012, grad_norm: 0.312354 +Steps: 1%| | 6122/1000000 [25:25<67:59:42, 4.06it/s, grad_norm=0.312, loss_final=0.465, loss_mean=0.802, loss_mean_cls=0.071, proj_loss=-0.409][2026-03-26 12:56:28] Step: 6122, Training Logs: loss_final: 0.460348, loss_mean: 0.786621, proj_loss: -0.399513, loss_mean_cls: 0.073239, grad_norm: 0.652337 +Steps: 1%| | 6123/1000000 [25:25<68:04:31, 4.06it/s, grad_norm=0.652, loss_final=0.46, loss_mean=0.787, loss_mean_cls=0.0732, proj_loss=-0.4][2026-03-26 12:56:28] Step: 6123, Training Logs: loss_final: 0.494971, loss_mean: 0.821715, proj_loss: -0.399845, loss_mean_cls: 0.073100, grad_norm: 0.300261 +Steps: 1%| | 6124/1000000 [25:25<68:10:33, 4.05it/s, grad_norm=0.3, loss_final=0.495, loss_mean=0.822, loss_mean_cls=0.0731, proj_loss=-0.4][2026-03-26 12:56:29] Step: 6124, Training Logs: loss_final: 0.473218, loss_mean: 0.823162, proj_loss: -0.417685, loss_mean_cls: 0.067741, grad_norm: 0.796268 +Steps: 1%| | 6125/1000000 [25:25<68:11:51, 4.05it/s, grad_norm=0.796, loss_final=0.473, loss_mean=0.823, loss_mean_cls=0.0677, proj_loss=-0.418][2026-03-26 12:56:29] Step: 6125, Training Logs: loss_final: 0.505745, loss_mean: 0.849706, proj_loss: -0.412108, loss_mean_cls: 0.068147, grad_norm: 0.786584 +Steps: 1%| | 6126/1000000 [25:26<68:14:40, 4.05it/s, grad_norm=0.787, loss_final=0.506, loss_mean=0.85, loss_mean_cls=0.0681, proj_loss=-0.412][2026-03-26 12:56:29] Step: 6126, Training Logs: loss_final: 0.487098, loss_mean: 0.823045, proj_loss: -0.407291, loss_mean_cls: 0.071344, grad_norm: 0.358785 +Steps: 1%| | 6127/1000000 [25:26<68:16:58, 4.04it/s, grad_norm=0.359, loss_final=0.487, loss_mean=0.823, loss_mean_cls=0.0713, proj_loss=-0.407][2026-03-26 12:56:29] Step: 6127, Training Logs: loss_final: 0.490320, loss_mean: 0.831197, proj_loss: -0.410947, loss_mean_cls: 0.070070, grad_norm: 0.635549 +Steps: 1%| | 6128/1000000 [25:26<68:19:23, 4.04it/s, grad_norm=0.636, loss_final=0.49, loss_mean=0.831, loss_mean_cls=0.0701, proj_loss=-0.411][2026-03-26 12:56:30] Step: 6128, Training Logs: loss_final: 0.474354, loss_mean: 0.814118, proj_loss: -0.410292, loss_mean_cls: 0.070528, grad_norm: 0.314927 +Steps: 1%| | 6129/1000000 [25:26<68:20:58, 4.04it/s, grad_norm=0.315, loss_final=0.474, loss_mean=0.814, loss_mean_cls=0.0705, proj_loss=-0.41][2026-03-26 12:56:30] Step: 6129, Training Logs: loss_final: 0.481966, loss_mean: 0.820313, proj_loss: -0.409232, loss_mean_cls: 0.070885, grad_norm: 0.836118 +Steps: 1%| | 6130/1000000 [25:27<68:21:08, 4.04it/s, grad_norm=0.836, loss_final=0.482, loss_mean=0.82, loss_mean_cls=0.0709, proj_loss=-0.409][2026-03-26 12:56:30] Step: 6130, Training Logs: loss_final: 0.470631, loss_mean: 0.816462, proj_loss: -0.414288, loss_mean_cls: 0.068457, grad_norm: 0.333167 +Steps: 1%| | 6131/1000000 [25:27<68:23:15, 4.04it/s, grad_norm=0.333, loss_final=0.471, loss_mean=0.816, loss_mean_cls=0.0685, proj_loss=-0.414][2026-03-26 12:56:30] Step: 6131, Training Logs: loss_final: 0.475604, loss_mean: 0.817426, proj_loss: -0.411717, loss_mean_cls: 0.069895, grad_norm: 0.509161 +Steps: 1%| | 6132/1000000 [25:27<68:22:54, 4.04it/s, grad_norm=0.509, loss_final=0.476, loss_mean=0.817, loss_mean_cls=0.0699, proj_loss=-0.412][2026-03-26 12:56:31] Step: 6132, Training Logs: loss_final: 0.480938, loss_mean: 0.822511, proj_loss: -0.412261, loss_mean_cls: 0.070688, grad_norm: 0.258537 +Steps: 1%| | 6133/1000000 [25:27<68:21:20, 4.04it/s, grad_norm=0.259, loss_final=0.481, loss_mean=0.823, loss_mean_cls=0.0707, proj_loss=-0.412][2026-03-26 12:56:31] Step: 6133, Training Logs: loss_final: 0.488545, loss_mean: 0.835273, proj_loss: -0.414715, loss_mean_cls: 0.067986, grad_norm: 0.372812 +Steps: 1%| | 6134/1000000 [25:28<68:21:31, 4.04it/s, grad_norm=0.373, loss_final=0.489, loss_mean=0.835, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 12:56:31] Step: 6134, Training Logs: loss_final: 0.480451, loss_mean: 0.829326, proj_loss: -0.416878, loss_mean_cls: 0.068004, grad_norm: 0.401812 +Steps: 1%| | 6135/1000000 [25:28<68:24:41, 4.04it/s, grad_norm=0.402, loss_final=0.48, loss_mean=0.829, loss_mean_cls=0.068, proj_loss=-0.417][2026-03-26 12:56:31] Step: 6135, Training Logs: loss_final: 0.486095, loss_mean: 0.834686, proj_loss: -0.416783, loss_mean_cls: 0.068192, grad_norm: 0.459884 +Steps: 1%| | 6136/1000000 [25:28<68:25:36, 4.03it/s, grad_norm=0.46, loss_final=0.486, loss_mean=0.835, loss_mean_cls=0.0682, proj_loss=-0.417][2026-03-26 12:56:32] Step: 6136, Training Logs: loss_final: 0.479566, loss_mean: 0.824712, proj_loss: -0.414200, loss_mean_cls: 0.069055, grad_norm: 0.525151 +Steps: 1%| | 6137/1000000 [25:28<68:22:29, 4.04it/s, grad_norm=0.525, loss_final=0.48, loss_mean=0.825, loss_mean_cls=0.0691, proj_loss=-0.414][2026-03-26 12:56:32] Step: 6137, Training Logs: loss_final: 0.472645, loss_mean: 0.812392, proj_loss: -0.410269, loss_mean_cls: 0.070523, grad_norm: 0.479470 +Steps: 1%| | 6138/1000000 [25:29<68:23:44, 4.04it/s, grad_norm=0.479, loss_final=0.473, loss_mean=0.812, loss_mean_cls=0.0705, proj_loss=-0.41][2026-03-26 12:56:32] Step: 6138, Training Logs: loss_final: 0.484045, loss_mean: 0.839041, proj_loss: -0.422020, loss_mean_cls: 0.067023, grad_norm: 0.647127 +Steps: 1%| | 6139/1000000 [25:29<68:21:10, 4.04it/s, grad_norm=0.647, loss_final=0.484, loss_mean=0.839, loss_mean_cls=0.067, proj_loss=-0.422][2026-03-26 12:56:32] Step: 6139, Training Logs: loss_final: 0.485054, loss_mean: 0.828114, proj_loss: -0.413473, loss_mean_cls: 0.070413, grad_norm: 0.469780 +Steps: 1%| | 6140/1000000 [25:29<68:21:21, 4.04it/s, grad_norm=0.47, loss_final=0.485, loss_mean=0.828, loss_mean_cls=0.0704, proj_loss=-0.413][2026-03-26 12:56:33] Step: 6140, Training Logs: loss_final: 0.469230, loss_mean: 0.824570, proj_loss: -0.423083, loss_mean_cls: 0.067744, grad_norm: 0.383040 +Steps: 1%| | 6141/1000000 [25:29<68:18:06, 4.04it/s, grad_norm=0.383, loss_final=0.469, loss_mean=0.825, loss_mean_cls=0.0677, proj_loss=-0.423][2026-03-26 12:56:33] Step: 6141, Training Logs: loss_final: 0.465692, loss_mean: 0.805070, proj_loss: -0.410278, loss_mean_cls: 0.070900, grad_norm: 0.357980 +Steps: 1%| | 6142/1000000 [25:30<68:22:25, 4.04it/s, grad_norm=0.358, loss_final=0.466, loss_mean=0.805, loss_mean_cls=0.0709, proj_loss=-0.41][2026-03-26 12:56:33] Step: 6142, Training Logs: loss_final: 0.472712, loss_mean: 0.819500, proj_loss: -0.417671, loss_mean_cls: 0.070883, grad_norm: 0.301706 +Steps: 1%| | 6143/1000000 [25:30<68:19:16, 4.04it/s, grad_norm=0.302, loss_final=0.473, loss_mean=0.819, loss_mean_cls=0.0709, proj_loss=-0.418][2026-03-26 12:56:33] Step: 6143, Training Logs: loss_final: 0.481629, loss_mean: 0.821569, proj_loss: -0.410453, loss_mean_cls: 0.070513, grad_norm: 0.387732 +Steps: 1%| | 6144/1000000 [25:30<68:19:18, 4.04it/s, grad_norm=0.388, loss_final=0.482, loss_mean=0.822, loss_mean_cls=0.0705, proj_loss=-0.41][2026-03-26 12:56:34] Step: 6144, Training Logs: loss_final: 0.477730, loss_mean: 0.820038, proj_loss: -0.413017, loss_mean_cls: 0.070709, grad_norm: 0.307262 +Steps: 1%| | 6145/1000000 [25:30<68:18:42, 4.04it/s, grad_norm=0.307, loss_final=0.478, loss_mean=0.82, loss_mean_cls=0.0707, proj_loss=-0.413][2026-03-26 12:56:34] Step: 6145, Training Logs: loss_final: 0.491251, loss_mean: 0.828013, proj_loss: -0.407760, loss_mean_cls: 0.070998, grad_norm: 0.246292 +Steps: 1%| | 6146/1000000 [25:31<68:20:12, 4.04it/s, grad_norm=0.246, loss_final=0.491, loss_mean=0.828, loss_mean_cls=0.071, proj_loss=-0.408][2026-03-26 12:56:34] Step: 6146, Training Logs: loss_final: 0.476805, loss_mean: 0.825923, proj_loss: -0.418388, loss_mean_cls: 0.069269, grad_norm: 0.298383 +Steps: 1%| | 6147/1000000 [25:31<68:20:18, 4.04it/s, grad_norm=0.298, loss_final=0.477, loss_mean=0.826, loss_mean_cls=0.0693, proj_loss=-0.418][2026-03-26 12:56:34] Step: 6147, Training Logs: loss_final: 0.481772, loss_mean: 0.829531, proj_loss: -0.416537, loss_mean_cls: 0.068778, grad_norm: 0.288405 +Steps: 1%| | 6148/1000000 [25:31<68:22:18, 4.04it/s, grad_norm=0.288, loss_final=0.482, loss_mean=0.83, loss_mean_cls=0.0688, proj_loss=-0.417][2026-03-26 12:56:35] Step: 6148, Training Logs: loss_final: 0.469229, loss_mean: 0.818184, proj_loss: -0.418393, loss_mean_cls: 0.069438, grad_norm: 0.254274 +Steps: 1%| | 6149/1000000 [25:31<68:20:57, 4.04it/s, grad_norm=0.254, loss_final=0.469, loss_mean=0.818, loss_mean_cls=0.0694, proj_loss=-0.418][2026-03-26 12:56:35] Step: 6149, Training Logs: loss_final: 0.470536, loss_mean: 0.823481, proj_loss: -0.420836, loss_mean_cls: 0.067891, grad_norm: 0.265171 +Steps: 1%| | 6150/1000000 [25:32<68:21:24, 4.04it/s, grad_norm=0.265, loss_final=0.471, loss_mean=0.823, loss_mean_cls=0.0679, proj_loss=-0.421][2026-03-26 12:56:35] Step: 6150, Training Logs: loss_final: 0.481033, loss_mean: 0.831082, proj_loss: -0.418086, loss_mean_cls: 0.068037, grad_norm: 0.379500 +Steps: 1%| | 6151/1000000 [25:32<68:22:04, 4.04it/s, grad_norm=0.38, loss_final=0.481, loss_mean=0.831, loss_mean_cls=0.068, proj_loss=-0.418][2026-03-26 12:56:35] Step: 6151, Training Logs: loss_final: 0.474979, loss_mean: 0.817144, proj_loss: -0.412368, loss_mean_cls: 0.070202, grad_norm: 0.338853 +Steps: 1%| | 6152/1000000 [25:32<68:22:42, 4.04it/s, grad_norm=0.339, loss_final=0.475, loss_mean=0.817, loss_mean_cls=0.0702, proj_loss=-0.412][2026-03-26 12:56:36] Step: 6152, Training Logs: loss_final: 0.463264, loss_mean: 0.818945, proj_loss: -0.422873, loss_mean_cls: 0.067193, grad_norm: 0.266827 +Steps: 1%| | 6153/1000000 [25:32<68:24:35, 4.04it/s, grad_norm=0.267, loss_final=0.463, loss_mean=0.819, loss_mean_cls=0.0672, proj_loss=-0.423][2026-03-26 12:56:36] Step: 6153, Training Logs: loss_final: 0.476662, loss_mean: 0.828364, proj_loss: -0.420156, loss_mean_cls: 0.068455, grad_norm: 0.327499 +Steps: 1%| | 6154/1000000 [25:33<68:23:18, 4.04it/s, grad_norm=0.327, loss_final=0.477, loss_mean=0.828, loss_mean_cls=0.0685, proj_loss=-0.42][2026-03-26 12:56:36] Step: 6154, Training Logs: loss_final: 0.475159, loss_mean: 0.805651, proj_loss: -0.402770, loss_mean_cls: 0.072278, grad_norm: 0.329476 +Steps: 1%| | 6155/1000000 [25:33<68:56:01, 4.00it/s, grad_norm=0.329, loss_final=0.475, loss_mean=0.806, loss_mean_cls=0.0723, proj_loss=-0.403][2026-03-26 12:56:36] Step: 6155, Training Logs: loss_final: 0.483522, loss_mean: 0.830492, proj_loss: -0.416963, loss_mean_cls: 0.069993, grad_norm: 0.400647 +Steps: 1%| | 6156/1000000 [25:33<68:48:38, 4.01it/s, grad_norm=0.401, loss_final=0.484, loss_mean=0.83, loss_mean_cls=0.07, proj_loss=-0.417][2026-03-26 12:56:37] Step: 6156, Training Logs: loss_final: 0.478398, loss_mean: 0.833392, proj_loss: -0.422017, loss_mean_cls: 0.067023, grad_norm: 0.280933 +Steps: 1%| | 6157/1000000 [25:33<68:42:51, 4.02it/s, grad_norm=0.281, loss_final=0.478, loss_mean=0.833, loss_mean_cls=0.067, proj_loss=-0.422][2026-03-26 12:56:37] Step: 6157, Training Logs: loss_final: 0.443485, loss_mean: 0.797663, proj_loss: -0.421219, loss_mean_cls: 0.067041, grad_norm: 0.443552 +Steps: 1%| | 6158/1000000 [25:34<68:35:59, 4.02it/s, grad_norm=0.444, loss_final=0.443, loss_mean=0.798, loss_mean_cls=0.067, proj_loss=-0.421][2026-03-26 12:56:37] Step: 6158, Training Logs: loss_final: 0.483689, loss_mean: 0.808677, proj_loss: -0.397652, loss_mean_cls: 0.072664, grad_norm: 0.245708 +Steps: 1%| | 6159/1000000 [25:34<68:33:20, 4.03it/s, grad_norm=0.246, loss_final=0.484, loss_mean=0.809, loss_mean_cls=0.0727, proj_loss=-0.398][2026-03-26 12:56:37] Step: 6159, Training Logs: loss_final: 0.474393, loss_mean: 0.816627, proj_loss: -0.412640, loss_mean_cls: 0.070406, grad_norm: 0.464686 +Steps: 1%| | 6160/1000000 [25:34<68:31:49, 4.03it/s, grad_norm=0.465, loss_final=0.474, loss_mean=0.817, loss_mean_cls=0.0704, proj_loss=-0.413][2026-03-26 12:56:37] Step: 6160, Training Logs: loss_final: 0.486243, loss_mean: 0.827578, proj_loss: -0.410637, loss_mean_cls: 0.069302, grad_norm: 0.319754 +Steps: 1%| | 6161/1000000 [25:34<68:30:53, 4.03it/s, grad_norm=0.32, loss_final=0.486, loss_mean=0.828, loss_mean_cls=0.0693, proj_loss=-0.411][2026-03-26 12:56:38] Step: 6161, Training Logs: loss_final: 0.471288, loss_mean: 0.815625, proj_loss: -0.414386, loss_mean_cls: 0.070048, grad_norm: 0.386389 +Steps: 1%| | 6162/1000000 [25:35<68:26:44, 4.03it/s, grad_norm=0.386, loss_final=0.471, loss_mean=0.816, loss_mean_cls=0.07, proj_loss=-0.414][2026-03-26 12:56:38] Step: 6162, Training Logs: loss_final: 0.489934, loss_mean: 0.842809, proj_loss: -0.419471, loss_mean_cls: 0.066595, grad_norm: 0.436553 +Steps: 1%| | 6163/1000000 [25:35<68:24:08, 4.04it/s, grad_norm=0.437, loss_final=0.49, loss_mean=0.843, loss_mean_cls=0.0666, proj_loss=-0.419][2026-03-26 12:56:38] Step: 6163, Training Logs: loss_final: 0.473573, loss_mean: 0.825184, proj_loss: -0.420466, loss_mean_cls: 0.068855, grad_norm: 0.286009 +Steps: 1%| | 6164/1000000 [25:35<68:22:12, 4.04it/s, grad_norm=0.286, loss_final=0.474, loss_mean=0.825, loss_mean_cls=0.0689, proj_loss=-0.42][2026-03-26 12:56:38] Step: 6164, Training Logs: loss_final: 0.476364, loss_mean: 0.827780, proj_loss: -0.420174, loss_mean_cls: 0.068758, grad_norm: 0.567134 +Steps: 1%| | 6165/1000000 [25:35<68:20:59, 4.04it/s, grad_norm=0.567, loss_final=0.476, loss_mean=0.828, loss_mean_cls=0.0688, proj_loss=-0.42][2026-03-26 12:56:39] Step: 6165, Training Logs: loss_final: 0.474214, loss_mean: 0.812926, proj_loss: -0.408872, loss_mean_cls: 0.070159, grad_norm: 0.259177 +Steps: 1%| | 6166/1000000 [25:36<68:18:00, 4.04it/s, grad_norm=0.259, loss_final=0.474, loss_mean=0.813, loss_mean_cls=0.0702, proj_loss=-0.409][2026-03-26 12:56:39] Step: 6166, Training Logs: loss_final: 0.472188, loss_mean: 0.824518, proj_loss: -0.420016, loss_mean_cls: 0.067686, grad_norm: 0.497743 +Steps: 1%| | 6167/1000000 [25:36<68:20:23, 4.04it/s, grad_norm=0.498, loss_final=0.472, loss_mean=0.825, loss_mean_cls=0.0677, proj_loss=-0.42][2026-03-26 12:56:39] Step: 6167, Training Logs: loss_final: 0.482939, loss_mean: 0.840146, proj_loss: -0.424387, loss_mean_cls: 0.067179, grad_norm: 0.380474 +Steps: 1%| | 6168/1000000 [25:36<68:18:06, 4.04it/s, grad_norm=0.38, loss_final=0.483, loss_mean=0.84, loss_mean_cls=0.0672, proj_loss=-0.424][2026-03-26 12:56:39] Step: 6168, Training Logs: loss_final: 0.485255, loss_mean: 0.825869, proj_loss: -0.410617, loss_mean_cls: 0.070003, grad_norm: 0.544984 +Steps: 1%| | 6169/1000000 [25:36<68:18:17, 4.04it/s, grad_norm=0.545, loss_final=0.485, loss_mean=0.826, loss_mean_cls=0.07, proj_loss=-0.411][2026-03-26 12:56:40] Step: 6169, Training Logs: loss_final: 0.478129, loss_mean: 0.823874, proj_loss: -0.414971, loss_mean_cls: 0.069227, grad_norm: 0.407136 +Steps: 1%| | 6170/1000000 [25:37<68:18:35, 4.04it/s, grad_norm=0.407, loss_final=0.478, loss_mean=0.824, loss_mean_cls=0.0692, proj_loss=-0.415][2026-03-26 12:56:40] Step: 6170, Training Logs: loss_final: 0.472548, loss_mean: 0.817889, proj_loss: -0.414596, loss_mean_cls: 0.069256, grad_norm: 0.381595 +Steps: 1%| | 6171/1000000 [25:37<68:19:00, 4.04it/s, grad_norm=0.382, loss_final=0.473, loss_mean=0.818, loss_mean_cls=0.0693, proj_loss=-0.415][2026-03-26 12:56:40] Step: 6171, Training Logs: loss_final: 0.479251, loss_mean: 0.821329, proj_loss: -0.411843, loss_mean_cls: 0.069765, grad_norm: 0.529733 +Steps: 1%| | 6172/1000000 [25:37<68:18:36, 4.04it/s, grad_norm=0.53, loss_final=0.479, loss_mean=0.821, loss_mean_cls=0.0698, proj_loss=-0.412][2026-03-26 12:56:40] Step: 6172, Training Logs: loss_final: 0.479571, loss_mean: 0.821080, proj_loss: -0.411875, loss_mean_cls: 0.070367, grad_norm: 0.309872 +Steps: 1%| | 6173/1000000 [25:37<68:22:42, 4.04it/s, grad_norm=0.31, loss_final=0.48, loss_mean=0.821, loss_mean_cls=0.0704, proj_loss=-0.412][2026-03-26 12:56:41] Step: 6173, Training Logs: loss_final: 0.491889, loss_mean: 0.833185, proj_loss: -0.411594, loss_mean_cls: 0.070298, grad_norm: 0.506244 +Steps: 1%| | 6174/1000000 [25:38<68:21:14, 4.04it/s, grad_norm=0.506, loss_final=0.492, loss_mean=0.833, loss_mean_cls=0.0703, proj_loss=-0.412][2026-03-26 12:56:41] Step: 6174, Training Logs: loss_final: 0.471498, loss_mean: 0.811830, proj_loss: -0.410661, loss_mean_cls: 0.070328, grad_norm: 0.263556 +Steps: 1%| | 6175/1000000 [25:38<68:20:41, 4.04it/s, grad_norm=0.264, loss_final=0.471, loss_mean=0.812, loss_mean_cls=0.0703, proj_loss=-0.411][2026-03-26 12:56:41] Step: 6175, Training Logs: loss_final: 0.480636, loss_mean: 0.829286, proj_loss: -0.417428, loss_mean_cls: 0.068779, grad_norm: 0.489538 +Steps: 1%| | 6176/1000000 [25:38<68:22:45, 4.04it/s, grad_norm=0.49, loss_final=0.481, loss_mean=0.829, loss_mean_cls=0.0688, proj_loss=-0.417][2026-03-26 12:56:41] Step: 6176, Training Logs: loss_final: 0.468437, loss_mean: 0.805487, proj_loss: -0.408353, loss_mean_cls: 0.071303, grad_norm: 0.323109 +Steps: 1%| | 6177/1000000 [25:38<68:22:02, 4.04it/s, grad_norm=0.323, loss_final=0.468, loss_mean=0.805, loss_mean_cls=0.0713, proj_loss=-0.408][2026-03-26 12:56:42] Step: 6177, Training Logs: loss_final: 0.475978, loss_mean: 0.812442, proj_loss: -0.407205, loss_mean_cls: 0.070740, grad_norm: 0.467849 +Steps: 1%| | 6178/1000000 [25:39<68:35:09, 4.03it/s, grad_norm=0.468, loss_final=0.476, loss_mean=0.812, loss_mean_cls=0.0707, proj_loss=-0.407][2026-03-26 12:56:42] Step: 6178, Training Logs: loss_final: 0.481343, loss_mean: 0.817732, proj_loss: -0.407111, loss_mean_cls: 0.070722, grad_norm: 0.443574 +Steps: 1%| | 6179/1000000 [25:39<68:56:20, 4.00it/s, grad_norm=0.444, loss_final=0.481, loss_mean=0.818, loss_mean_cls=0.0707, proj_loss=-0.407][2026-03-26 12:56:42] Step: 6179, Training Logs: loss_final: 0.465047, loss_mean: 0.804530, proj_loss: -0.409537, loss_mean_cls: 0.070054, grad_norm: 0.468750 +Steps: 1%| | 6180/1000000 [25:39<68:45:12, 4.02it/s, grad_norm=0.469, loss_final=0.465, loss_mean=0.805, loss_mean_cls=0.0701, proj_loss=-0.41][2026-03-26 12:56:42] Step: 6180, Training Logs: loss_final: 0.462953, loss_mean: 0.806146, proj_loss: -0.413565, loss_mean_cls: 0.070372, grad_norm: 0.332253 +Steps: 1%| | 6181/1000000 [25:39<68:36:50, 4.02it/s, grad_norm=0.332, loss_final=0.463, loss_mean=0.806, loss_mean_cls=0.0704, proj_loss=-0.414][2026-03-26 12:56:43] Step: 6181, Training Logs: loss_final: 0.490691, loss_mean: 0.849858, proj_loss: -0.425839, loss_mean_cls: 0.066672, grad_norm: 0.399345 +Steps: 1%| | 6182/1000000 [25:40<68:34:13, 4.03it/s, grad_norm=0.399, loss_final=0.491, loss_mean=0.85, loss_mean_cls=0.0667, proj_loss=-0.426][2026-03-26 12:56:43] Step: 6182, Training Logs: loss_final: 0.475910, loss_mean: 0.826510, proj_loss: -0.418481, loss_mean_cls: 0.067880, grad_norm: 0.369681 +Steps: 1%| | 6183/1000000 [25:40<68:28:00, 4.03it/s, grad_norm=0.37, loss_final=0.476, loss_mean=0.827, loss_mean_cls=0.0679, proj_loss=-0.418][2026-03-26 12:56:43] Step: 6183, Training Logs: loss_final: 0.460275, loss_mean: 0.813376, proj_loss: -0.421071, loss_mean_cls: 0.067970, grad_norm: 0.357883 +Steps: 1%| | 6184/1000000 [25:40<68:25:30, 4.03it/s, grad_norm=0.358, loss_final=0.46, loss_mean=0.813, loss_mean_cls=0.068, proj_loss=-0.421][2026-03-26 12:56:43] Step: 6184, Training Logs: loss_final: 0.473535, loss_mean: 0.831186, proj_loss: -0.424803, loss_mean_cls: 0.067152, grad_norm: 0.456944 +Steps: 1%| | 6185/1000000 [25:40<68:20:47, 4.04it/s, grad_norm=0.457, loss_final=0.474, loss_mean=0.831, loss_mean_cls=0.0672, proj_loss=-0.425][2026-03-26 12:56:44] Step: 6185, Training Logs: loss_final: 0.477486, loss_mean: 0.829226, proj_loss: -0.419525, loss_mean_cls: 0.067784, grad_norm: 0.215152 +Steps: 1%| | 6186/1000000 [25:41<68:20:16, 4.04it/s, grad_norm=0.215, loss_final=0.477, loss_mean=0.829, loss_mean_cls=0.0678, proj_loss=-0.42][2026-03-26 12:56:44] Step: 6186, Training Logs: loss_final: 0.466250, loss_mean: 0.823408, proj_loss: -0.424195, loss_mean_cls: 0.067036, grad_norm: 0.815593 +Steps: 1%| | 6187/1000000 [25:41<68:18:24, 4.04it/s, grad_norm=0.816, loss_final=0.466, loss_mean=0.823, loss_mean_cls=0.067, proj_loss=-0.424][2026-03-26 12:56:44] Step: 6187, Training Logs: loss_final: 0.470133, loss_mean: 0.818684, proj_loss: -0.417723, loss_mean_cls: 0.069172, grad_norm: 0.472675 +Steps: 1%| | 6188/1000000 [25:41<68:19:43, 4.04it/s, grad_norm=0.473, loss_final=0.47, loss_mean=0.819, loss_mean_cls=0.0692, proj_loss=-0.418][2026-03-26 12:56:44] Step: 6188, Training Logs: loss_final: 0.481989, loss_mean: 0.827083, proj_loss: -0.414317, loss_mean_cls: 0.069223, grad_norm: 0.631785 +Steps: 1%| | 6189/1000000 [25:41<68:17:32, 4.04it/s, grad_norm=0.632, loss_final=0.482, loss_mean=0.827, loss_mean_cls=0.0692, proj_loss=-0.414][2026-03-26 12:56:45] Step: 6189, Training Logs: loss_final: 0.475243, loss_mean: 0.818594, proj_loss: -0.413346, loss_mean_cls: 0.069995, grad_norm: 0.711206 +Steps: 1%| | 6190/1000000 [25:42<68:20:47, 4.04it/s, grad_norm=0.711, loss_final=0.475, loss_mean=0.819, loss_mean_cls=0.07, proj_loss=-0.413][2026-03-26 12:56:45] Step: 6190, Training Logs: loss_final: 0.463791, loss_mean: 0.822168, proj_loss: -0.424886, loss_mean_cls: 0.066510, grad_norm: 0.373257 +Steps: 1%| | 6191/1000000 [25:42<68:14:53, 4.04it/s, grad_norm=0.373, loss_final=0.464, loss_mean=0.822, loss_mean_cls=0.0665, proj_loss=-0.425][2026-03-26 12:56:45] Step: 6191, Training Logs: loss_final: 0.478484, loss_mean: 0.822782, proj_loss: -0.414145, loss_mean_cls: 0.069847, grad_norm: 0.535177 +Steps: 1%| | 6192/1000000 [25:42<68:15:46, 4.04it/s, grad_norm=0.535, loss_final=0.478, loss_mean=0.823, loss_mean_cls=0.0698, proj_loss=-0.414][2026-03-26 12:56:45] Step: 6192, Training Logs: loss_final: 0.488839, loss_mean: 0.828887, proj_loss: -0.410623, loss_mean_cls: 0.070575, grad_norm: 0.451453 +Steps: 1%| | 6193/1000000 [25:42<68:17:51, 4.04it/s, grad_norm=0.451, loss_final=0.489, loss_mean=0.829, loss_mean_cls=0.0706, proj_loss=-0.411][2026-03-26 12:56:46] Step: 6193, Training Logs: loss_final: 0.494866, loss_mean: 0.846671, proj_loss: -0.419547, loss_mean_cls: 0.067742, grad_norm: 0.535438 +Steps: 1%| | 6194/1000000 [25:43<68:20:58, 4.04it/s, grad_norm=0.535, loss_final=0.495, loss_mean=0.847, loss_mean_cls=0.0677, proj_loss=-0.42][2026-03-26 12:56:46] Step: 6194, Training Logs: loss_final: 0.479969, loss_mean: 0.832178, proj_loss: -0.419647, loss_mean_cls: 0.067438, grad_norm: 0.429442 +Steps: 1%| | 6195/1000000 [25:43<68:19:37, 4.04it/s, grad_norm=0.429, loss_final=0.48, loss_mean=0.832, loss_mean_cls=0.0674, proj_loss=-0.42][2026-03-26 12:56:46] Step: 6195, Training Logs: loss_final: 0.489996, loss_mean: 0.823502, proj_loss: -0.404770, loss_mean_cls: 0.071263, grad_norm: 0.533873 +Steps: 1%| | 6196/1000000 [25:43<68:21:15, 4.04it/s, grad_norm=0.534, loss_final=0.49, loss_mean=0.824, loss_mean_cls=0.0713, proj_loss=-0.405][2026-03-26 12:56:46] Step: 6196, Training Logs: loss_final: 0.481684, loss_mean: 0.835124, proj_loss: -0.420402, loss_mean_cls: 0.066961, grad_norm: 0.420880 +Steps: 1%| | 6197/1000000 [25:43<68:18:41, 4.04it/s, grad_norm=0.421, loss_final=0.482, loss_mean=0.835, loss_mean_cls=0.067, proj_loss=-0.42][2026-03-26 12:56:47] Step: 6197, Training Logs: loss_final: 0.474151, loss_mean: 0.838817, proj_loss: -0.430594, loss_mean_cls: 0.065928, grad_norm: 0.575090 +Steps: 1%| | 6198/1000000 [25:44<68:17:47, 4.04it/s, grad_norm=0.575, loss_final=0.474, loss_mean=0.839, loss_mean_cls=0.0659, proj_loss=-0.431][2026-03-26 12:56:47] Step: 6198, Training Logs: loss_final: 0.464551, loss_mean: 0.803894, proj_loss: -0.410267, loss_mean_cls: 0.070924, grad_norm: 0.367677 +Steps: 1%| | 6199/1000000 [25:44<68:16:25, 4.04it/s, grad_norm=0.368, loss_final=0.465, loss_mean=0.804, loss_mean_cls=0.0709, proj_loss=-0.41][2026-03-26 12:56:47] Step: 6199, Training Logs: loss_final: 0.473794, loss_mean: 0.810741, proj_loss: -0.407982, loss_mean_cls: 0.071035, grad_norm: 0.353704 +Steps: 1%| | 6200/1000000 [25:44<68:18:47, 4.04it/s, grad_norm=0.354, loss_final=0.474, loss_mean=0.811, loss_mean_cls=0.071, proj_loss=-0.408][2026-03-26 12:56:47] Step: 6200, Training Logs: loss_final: 0.473654, loss_mean: 0.807916, proj_loss: -0.405402, loss_mean_cls: 0.071140, grad_norm: 0.307509 +Steps: 1%| | 6201/1000000 [25:44<68:18:38, 4.04it/s, grad_norm=0.308, loss_final=0.474, loss_mean=0.808, loss_mean_cls=0.0711, proj_loss=-0.405][2026-03-26 12:56:48] Step: 6201, Training Logs: loss_final: 0.470522, loss_mean: 0.816114, proj_loss: -0.414597, loss_mean_cls: 0.069005, grad_norm: 0.252568 +Steps: 1%| | 6202/1000000 [25:45<68:20:02, 4.04it/s, grad_norm=0.253, loss_final=0.471, loss_mean=0.816, loss_mean_cls=0.069, proj_loss=-0.415][2026-03-26 12:56:48] Step: 6202, Training Logs: loss_final: 0.477142, loss_mean: 0.827607, proj_loss: -0.418725, loss_mean_cls: 0.068259, grad_norm: 0.346425 +Steps: 1%| | 6203/1000000 [25:45<68:22:36, 4.04it/s, grad_norm=0.346, loss_final=0.477, loss_mean=0.828, loss_mean_cls=0.0683, proj_loss=-0.419][2026-03-26 12:56:48] Step: 6203, Training Logs: loss_final: 0.474777, loss_mean: 0.822734, proj_loss: -0.416837, loss_mean_cls: 0.068880, grad_norm: 0.414333 +Steps: 1%| | 6204/1000000 [25:45<68:21:30, 4.04it/s, grad_norm=0.414, loss_final=0.475, loss_mean=0.823, loss_mean_cls=0.0689, proj_loss=-0.417][2026-03-26 12:56:48] Step: 6204, Training Logs: loss_final: 0.476687, loss_mean: 0.813336, proj_loss: -0.408340, loss_mean_cls: 0.071691, grad_norm: 0.251386 +Steps: 1%| | 6205/1000000 [25:45<68:20:16, 4.04it/s, grad_norm=0.251, loss_final=0.477, loss_mean=0.813, loss_mean_cls=0.0717, proj_loss=-0.408][2026-03-26 12:56:49] Step: 6205, Training Logs: loss_final: 0.470622, loss_mean: 0.829257, proj_loss: -0.425544, loss_mean_cls: 0.066910, grad_norm: 0.464669 +Steps: 1%| | 6206/1000000 [25:46<68:20:19, 4.04it/s, grad_norm=0.465, loss_final=0.471, loss_mean=0.829, loss_mean_cls=0.0669, proj_loss=-0.426][2026-03-26 12:56:49] Step: 6206, Training Logs: loss_final: 0.471888, loss_mean: 0.823860, proj_loss: -0.419765, loss_mean_cls: 0.067794, grad_norm: 0.182558 +Steps: 1%| | 6207/1000000 [25:46<68:18:26, 4.04it/s, grad_norm=0.183, loss_final=0.472, loss_mean=0.824, loss_mean_cls=0.0678, proj_loss=-0.42][2026-03-26 12:56:49] Step: 6207, Training Logs: loss_final: 0.470975, loss_mean: 0.814987, proj_loss: -0.413761, loss_mean_cls: 0.069749, grad_norm: 0.654296 +Steps: 1%| | 6208/1000000 [25:46<68:13:58, 4.05it/s, grad_norm=0.654, loss_final=0.471, loss_mean=0.815, loss_mean_cls=0.0697, proj_loss=-0.414][2026-03-26 12:56:49] Step: 6208, Training Logs: loss_final: 0.461490, loss_mean: 0.814633, proj_loss: -0.421170, loss_mean_cls: 0.068027, grad_norm: 0.540706 +Steps: 1%| | 6209/1000000 [25:46<68:15:42, 4.04it/s, grad_norm=0.541, loss_final=0.461, loss_mean=0.815, loss_mean_cls=0.068, proj_loss=-0.421][2026-03-26 12:56:50] Step: 6209, Training Logs: loss_final: 0.478522, loss_mean: 0.821461, proj_loss: -0.412795, loss_mean_cls: 0.069856, grad_norm: 0.284126 +Steps: 1%| | 6210/1000000 [25:46<68:15:41, 4.04it/s, grad_norm=0.284, loss_final=0.479, loss_mean=0.821, loss_mean_cls=0.0699, proj_loss=-0.413][2026-03-26 12:56:50] Step: 6210, Training Logs: loss_final: 0.475625, loss_mean: 0.825820, proj_loss: -0.418952, loss_mean_cls: 0.068757, grad_norm: 0.492315 +Steps: 1%| | 6211/1000000 [25:47<68:16:19, 4.04it/s, grad_norm=0.492, loss_final=0.476, loss_mean=0.826, loss_mean_cls=0.0688, proj_loss=-0.419][2026-03-26 12:56:50] Step: 6211, Training Logs: loss_final: 0.477622, loss_mean: 0.823124, proj_loss: -0.414840, loss_mean_cls: 0.069337, grad_norm: 0.306611 +Steps: 1%| | 6212/1000000 [25:47<68:15:03, 4.04it/s, grad_norm=0.307, loss_final=0.478, loss_mean=0.823, loss_mean_cls=0.0693, proj_loss=-0.415][2026-03-26 12:56:50] Step: 6212, Training Logs: loss_final: 0.480230, loss_mean: 0.826597, proj_loss: -0.415664, loss_mean_cls: 0.069296, grad_norm: 0.551131 +Steps: 1%| | 6213/1000000 [25:47<68:15:15, 4.04it/s, grad_norm=0.551, loss_final=0.48, loss_mean=0.827, loss_mean_cls=0.0693, proj_loss=-0.416][2026-03-26 12:56:51] Step: 6213, Training Logs: loss_final: 0.476560, loss_mean: 0.826981, proj_loss: -0.419066, loss_mean_cls: 0.068645, grad_norm: 0.259763 +Steps: 1%| | 6214/1000000 [25:47<68:15:38, 4.04it/s, grad_norm=0.26, loss_final=0.477, loss_mean=0.827, loss_mean_cls=0.0686, proj_loss=-0.419][2026-03-26 12:56:51] Step: 6214, Training Logs: loss_final: 0.465747, loss_mean: 0.817154, proj_loss: -0.419903, loss_mean_cls: 0.068497, grad_norm: 0.400320 +Steps: 1%| | 6215/1000000 [25:48<68:14:41, 4.05it/s, grad_norm=0.4, loss_final=0.466, loss_mean=0.817, loss_mean_cls=0.0685, proj_loss=-0.42][2026-03-26 12:56:51] Step: 6215, Training Logs: loss_final: 0.472616, loss_mean: 0.822872, proj_loss: -0.417897, loss_mean_cls: 0.067642, grad_norm: 0.304240 +Steps: 1%| | 6216/1000000 [25:48<68:16:07, 4.04it/s, grad_norm=0.304, loss_final=0.473, loss_mean=0.823, loss_mean_cls=0.0676, proj_loss=-0.418][2026-03-26 12:56:51] Step: 6216, Training Logs: loss_final: 0.490611, loss_mean: 0.838892, proj_loss: -0.417200, loss_mean_cls: 0.068918, grad_norm: 0.257912 +Steps: 1%| | 6217/1000000 [25:48<68:17:41, 4.04it/s, grad_norm=0.258, loss_final=0.491, loss_mean=0.839, loss_mean_cls=0.0689, proj_loss=-0.417][2026-03-26 12:56:52] Step: 6217, Training Logs: loss_final: 0.468610, loss_mean: 0.817506, proj_loss: -0.416908, loss_mean_cls: 0.068012, grad_norm: 0.270049 +Steps: 1%| | 6218/1000000 [25:48<68:19:36, 4.04it/s, grad_norm=0.27, loss_final=0.469, loss_mean=0.818, loss_mean_cls=0.068, proj_loss=-0.417][2026-03-26 12:56:52] Step: 6218, Training Logs: loss_final: 0.481230, loss_mean: 0.824193, proj_loss: -0.412944, loss_mean_cls: 0.069980, grad_norm: 0.348314 +Steps: 1%| | 6219/1000000 [25:49<68:18:43, 4.04it/s, grad_norm=0.348, loss_final=0.481, loss_mean=0.824, loss_mean_cls=0.07, proj_loss=-0.413][2026-03-26 12:56:52] Step: 6219, Training Logs: loss_final: 0.492441, loss_mean: 0.825559, proj_loss: -0.404741, loss_mean_cls: 0.071623, grad_norm: 0.403986 +Steps: 1%| | 6220/1000000 [25:49<68:17:50, 4.04it/s, grad_norm=0.404, loss_final=0.492, loss_mean=0.826, loss_mean_cls=0.0716, proj_loss=-0.405][2026-03-26 12:56:52] Step: 6220, Training Logs: loss_final: 0.485984, loss_mean: 0.822376, proj_loss: -0.407917, loss_mean_cls: 0.071525, grad_norm: 0.481534 +Steps: 1%| | 6221/1000000 [25:49<68:17:12, 4.04it/s, grad_norm=0.482, loss_final=0.486, loss_mean=0.822, loss_mean_cls=0.0715, proj_loss=-0.408][2026-03-26 12:56:53] Step: 6221, Training Logs: loss_final: 0.466207, loss_mean: 0.802408, proj_loss: -0.407686, loss_mean_cls: 0.071486, grad_norm: 0.292572 +Steps: 1%| | 6222/1000000 [25:49<68:20:27, 4.04it/s, grad_norm=0.293, loss_final=0.466, loss_mean=0.802, loss_mean_cls=0.0715, proj_loss=-0.408][2026-03-26 12:56:53] Step: 6222, Training Logs: loss_final: 0.460757, loss_mean: 0.805202, proj_loss: -0.414219, loss_mean_cls: 0.069774, grad_norm: 0.469437 +Steps: 1%| | 6223/1000000 [25:50<68:19:33, 4.04it/s, grad_norm=0.469, loss_final=0.461, loss_mean=0.805, loss_mean_cls=0.0698, proj_loss=-0.414][2026-03-26 12:56:53] Step: 6223, Training Logs: loss_final: 0.474695, loss_mean: 0.817402, proj_loss: -0.412093, loss_mean_cls: 0.069386, grad_norm: 0.458203 +Steps: 1%| | 6224/1000000 [25:50<68:20:39, 4.04it/s, grad_norm=0.458, loss_final=0.475, loss_mean=0.817, loss_mean_cls=0.0694, proj_loss=-0.412][2026-03-26 12:56:53] Step: 6224, Training Logs: loss_final: 0.465407, loss_mean: 0.818442, proj_loss: -0.420762, loss_mean_cls: 0.067727, grad_norm: 0.347630 +Steps: 1%| | 6225/1000000 [25:50<68:14:47, 4.04it/s, grad_norm=0.348, loss_final=0.465, loss_mean=0.818, loss_mean_cls=0.0677, proj_loss=-0.421][2026-03-26 12:56:54] Step: 6225, Training Logs: loss_final: 0.482237, loss_mean: 0.834706, proj_loss: -0.421161, loss_mean_cls: 0.068693, grad_norm: 0.243859 +Steps: 1%| | 6226/1000000 [25:50<68:36:28, 4.02it/s, grad_norm=0.244, loss_final=0.482, loss_mean=0.835, loss_mean_cls=0.0687, proj_loss=-0.421][2026-03-26 12:56:54] Step: 6226, Training Logs: loss_final: 0.459792, loss_mean: 0.804489, proj_loss: -0.414759, loss_mean_cls: 0.070062, grad_norm: 0.438929 +Steps: 1%| | 6227/1000000 [25:51<68:29:58, 4.03it/s, grad_norm=0.439, loss_final=0.46, loss_mean=0.804, loss_mean_cls=0.0701, proj_loss=-0.415][2026-03-26 12:56:54] Step: 6227, Training Logs: loss_final: 0.454899, loss_mean: 0.794410, proj_loss: -0.409832, loss_mean_cls: 0.070320, grad_norm: 0.376475 +Steps: 1%| | 6228/1000000 [25:51<68:28:02, 4.03it/s, grad_norm=0.376, loss_final=0.455, loss_mean=0.794, loss_mean_cls=0.0703, proj_loss=-0.41][2026-03-26 12:56:54] Step: 6228, Training Logs: loss_final: 0.448341, loss_mean: 0.789406, proj_loss: -0.410880, loss_mean_cls: 0.069815, grad_norm: 0.318016 +Steps: 1%| | 6229/1000000 [25:51<68:23:04, 4.04it/s, grad_norm=0.318, loss_final=0.448, loss_mean=0.789, loss_mean_cls=0.0698, proj_loss=-0.411][2026-03-26 12:56:55] Step: 6229, Training Logs: loss_final: 0.475427, loss_mean: 0.825077, proj_loss: -0.419255, loss_mean_cls: 0.069606, grad_norm: 0.239384 +Steps: 1%| | 6230/1000000 [25:51<68:24:49, 4.03it/s, grad_norm=0.239, loss_final=0.475, loss_mean=0.825, loss_mean_cls=0.0696, proj_loss=-0.419][2026-03-26 12:56:55] Step: 6230, Training Logs: loss_final: 0.478043, loss_mean: 0.820224, proj_loss: -0.411701, loss_mean_cls: 0.069521, grad_norm: 0.342603 +Steps: 1%| | 6231/1000000 [25:52<68:22:34, 4.04it/s, grad_norm=0.343, loss_final=0.478, loss_mean=0.82, loss_mean_cls=0.0695, proj_loss=-0.412][2026-03-26 12:56:55] Step: 6231, Training Logs: loss_final: 0.456313, loss_mean: 0.806961, proj_loss: -0.419528, loss_mean_cls: 0.068880, grad_norm: 0.459657 +Steps: 1%| | 6232/1000000 [25:52<68:24:35, 4.04it/s, grad_norm=0.46, loss_final=0.456, loss_mean=0.807, loss_mean_cls=0.0689, proj_loss=-0.42][2026-03-26 12:56:55] Step: 6232, Training Logs: loss_final: 0.464990, loss_mean: 0.814986, proj_loss: -0.419071, loss_mean_cls: 0.069075, grad_norm: 0.316967 +Steps: 1%| | 6233/1000000 [25:52<68:20:44, 4.04it/s, grad_norm=0.317, loss_final=0.465, loss_mean=0.815, loss_mean_cls=0.0691, proj_loss=-0.419][2026-03-26 12:56:56] Step: 6233, Training Logs: loss_final: 0.484661, loss_mean: 0.833215, proj_loss: -0.416577, loss_mean_cls: 0.068023, grad_norm: 0.296525 +Steps: 1%| | 6234/1000000 [25:52<68:20:43, 4.04it/s, grad_norm=0.297, loss_final=0.485, loss_mean=0.833, loss_mean_cls=0.068, proj_loss=-0.417][2026-03-26 12:56:56] Step: 6234, Training Logs: loss_final: 0.464884, loss_mean: 0.807203, proj_loss: -0.412073, loss_mean_cls: 0.069753, grad_norm: 0.356818 +Steps: 1%| | 6235/1000000 [25:53<68:19:02, 4.04it/s, grad_norm=0.357, loss_final=0.465, loss_mean=0.807, loss_mean_cls=0.0698, proj_loss=-0.412][2026-03-26 12:56:56] Step: 6235, Training Logs: loss_final: 0.490067, loss_mean: 0.836981, proj_loss: -0.416197, loss_mean_cls: 0.069284, grad_norm: 0.257464 +Steps: 1%| | 6236/1000000 [25:53<68:18:25, 4.04it/s, grad_norm=0.257, loss_final=0.49, loss_mean=0.837, loss_mean_cls=0.0693, proj_loss=-0.416][2026-03-26 12:56:56] Step: 6236, Training Logs: loss_final: 0.482920, loss_mean: 0.823781, proj_loss: -0.410343, loss_mean_cls: 0.069482, grad_norm: 0.414623 +Steps: 1%| | 6237/1000000 [25:53<68:18:14, 4.04it/s, grad_norm=0.415, loss_final=0.483, loss_mean=0.824, loss_mean_cls=0.0695, proj_loss=-0.41][2026-03-26 12:56:57] Step: 6237, Training Logs: loss_final: 0.479359, loss_mean: 0.834523, proj_loss: -0.422006, loss_mean_cls: 0.066842, grad_norm: 0.399928 +Steps: 1%| | 6238/1000000 [25:53<69:30:51, 3.97it/s, grad_norm=0.4, loss_final=0.479, loss_mean=0.835, loss_mean_cls=0.0668, proj_loss=-0.422][2026-03-26 12:56:57] Step: 6238, Training Logs: loss_final: 0.496005, loss_mean: 0.840196, proj_loss: -0.413020, loss_mean_cls: 0.068828, grad_norm: 0.332147 +Steps: 1%| | 6239/1000000 [25:54<69:12:13, 3.99it/s, grad_norm=0.332, loss_final=0.496, loss_mean=0.84, loss_mean_cls=0.0688, proj_loss=-0.413][2026-03-26 12:56:57] Step: 6239, Training Logs: loss_final: 0.472680, loss_mean: 0.811133, proj_loss: -0.408882, loss_mean_cls: 0.070429, grad_norm: 0.371289 +Steps: 1%| | 6240/1000000 [25:54<68:58:04, 4.00it/s, grad_norm=0.371, loss_final=0.473, loss_mean=0.811, loss_mean_cls=0.0704, proj_loss=-0.409][2026-03-26 12:56:57] Step: 6240, Training Logs: loss_final: 0.493711, loss_mean: 0.840413, proj_loss: -0.415895, loss_mean_cls: 0.069193, grad_norm: 0.287656 +Steps: 1%| | 6241/1000000 [25:54<68:44:46, 4.02it/s, grad_norm=0.288, loss_final=0.494, loss_mean=0.84, loss_mean_cls=0.0692, proj_loss=-0.416][2026-03-26 12:56:58] Step: 6241, Training Logs: loss_final: 0.473784, loss_mean: 0.828887, proj_loss: -0.422625, loss_mean_cls: 0.067522, grad_norm: 0.473419 +Steps: 1%| | 6242/1000000 [25:54<68:37:14, 4.02it/s, grad_norm=0.473, loss_final=0.474, loss_mean=0.829, loss_mean_cls=0.0675, proj_loss=-0.423][2026-03-26 12:56:58] Step: 6242, Training Logs: loss_final: 0.474762, loss_mean: 0.832173, proj_loss: -0.424815, loss_mean_cls: 0.067405, grad_norm: 0.595151 +Steps: 1%| | 6243/1000000 [25:55<68:30:26, 4.03it/s, grad_norm=0.595, loss_final=0.475, loss_mean=0.832, loss_mean_cls=0.0674, proj_loss=-0.425][2026-03-26 12:56:58] Step: 6243, Training Logs: loss_final: 0.484698, loss_mean: 0.823557, proj_loss: -0.409757, loss_mean_cls: 0.070898, grad_norm: 0.268542 +Steps: 1%| | 6244/1000000 [25:55<68:26:38, 4.03it/s, grad_norm=0.269, loss_final=0.485, loss_mean=0.824, loss_mean_cls=0.0709, proj_loss=-0.41][2026-03-26 12:56:58] Step: 6244, Training Logs: loss_final: 0.475677, loss_mean: 0.825703, proj_loss: -0.417599, loss_mean_cls: 0.067574, grad_norm: 0.717622 +Steps: 1%| | 6245/1000000 [25:55<68:22:29, 4.04it/s, grad_norm=0.718, loss_final=0.476, loss_mean=0.826, loss_mean_cls=0.0676, proj_loss=-0.418][2026-03-26 12:56:59] Step: 6245, Training Logs: loss_final: 0.489050, loss_mean: 0.834881, proj_loss: -0.415250, loss_mean_cls: 0.069420, grad_norm: 0.494486 +Steps: 1%| | 6246/1000000 [25:55<68:21:21, 4.04it/s, grad_norm=0.494, loss_final=0.489, loss_mean=0.835, loss_mean_cls=0.0694, proj_loss=-0.415][2026-03-26 12:56:59] Step: 6246, Training Logs: loss_final: 0.481431, loss_mean: 0.829813, proj_loss: -0.416709, loss_mean_cls: 0.068326, grad_norm: 0.457095 +Steps: 1%| | 6247/1000000 [25:56<68:19:23, 4.04it/s, grad_norm=0.457, loss_final=0.481, loss_mean=0.83, loss_mean_cls=0.0683, proj_loss=-0.417][2026-03-26 12:56:59] Step: 6247, Training Logs: loss_final: 0.474806, loss_mean: 0.818452, proj_loss: -0.413196, loss_mean_cls: 0.069549, grad_norm: 0.304479 +Steps: 1%| | 6248/1000000 [25:56<68:21:59, 4.04it/s, grad_norm=0.304, loss_final=0.475, loss_mean=0.818, loss_mean_cls=0.0695, proj_loss=-0.413][2026-03-26 12:56:59] Step: 6248, Training Logs: loss_final: 0.472958, loss_mean: 0.819991, proj_loss: -0.417171, loss_mean_cls: 0.070138, grad_norm: 0.513567 +Steps: 1%| | 6249/1000000 [25:56<68:21:32, 4.04it/s, grad_norm=0.514, loss_final=0.473, loss_mean=0.82, loss_mean_cls=0.0701, proj_loss=-0.417][2026-03-26 12:57:00] Step: 6249, Training Logs: loss_final: 0.476269, loss_mean: 0.828062, proj_loss: -0.419223, loss_mean_cls: 0.067429, grad_norm: 0.353620 +Steps: 1%| | 6250/1000000 [25:56<68:19:44, 4.04it/s, grad_norm=0.354, loss_final=0.476, loss_mean=0.828, loss_mean_cls=0.0674, proj_loss=-0.419][2026-03-26 12:57:00] Step: 6250, Training Logs: loss_final: 0.477708, loss_mean: 0.825875, proj_loss: -0.416240, loss_mean_cls: 0.068074, grad_norm: 0.403297 +Steps: 1%| | 6251/1000000 [25:57<68:21:16, 4.04it/s, grad_norm=0.403, loss_final=0.478, loss_mean=0.826, loss_mean_cls=0.0681, proj_loss=-0.416][2026-03-26 12:57:00] Step: 6251, Training Logs: loss_final: 0.478394, loss_mean: 0.824599, proj_loss: -0.414095, loss_mean_cls: 0.067890, grad_norm: 0.479657 +Steps: 1%| | 6252/1000000 [25:57<68:19:22, 4.04it/s, grad_norm=0.48, loss_final=0.478, loss_mean=0.825, loss_mean_cls=0.0679, proj_loss=-0.414][2026-03-26 12:57:00] Step: 6252, Training Logs: loss_final: 0.472728, loss_mean: 0.810203, proj_loss: -0.408784, loss_mean_cls: 0.071309, grad_norm: 0.262269 +Steps: 1%| | 6253/1000000 [25:57<68:20:24, 4.04it/s, grad_norm=0.262, loss_final=0.473, loss_mean=0.81, loss_mean_cls=0.0713, proj_loss=-0.409][2026-03-26 12:57:01] Step: 6253, Training Logs: loss_final: 0.472550, loss_mean: 0.809447, proj_loss: -0.407243, loss_mean_cls: 0.070346, grad_norm: 0.369458 +Steps: 1%| | 6254/1000000 [25:57<70:19:14, 3.93it/s, grad_norm=0.369, loss_final=0.473, loss_mean=0.809, loss_mean_cls=0.0703, proj_loss=-0.407][2026-03-26 12:57:01] Step: 6254, Training Logs: loss_final: 0.457236, loss_mean: 0.801450, proj_loss: -0.414142, loss_mean_cls: 0.069928, grad_norm: 0.273094 +Steps: 1%| | 6255/1000000 [25:58<69:47:28, 3.96it/s, grad_norm=0.273, loss_final=0.457, loss_mean=0.801, loss_mean_cls=0.0699, proj_loss=-0.414][2026-03-26 12:57:01] Step: 6255, Training Logs: loss_final: 0.471234, loss_mean: 0.826343, proj_loss: -0.422767, loss_mean_cls: 0.067659, grad_norm: 0.443212 +Steps: 1%| | 6256/1000000 [25:58<69:21:47, 3.98it/s, grad_norm=0.443, loss_final=0.471, loss_mean=0.826, loss_mean_cls=0.0677, proj_loss=-0.423][2026-03-26 12:57:01] Step: 6256, Training Logs: loss_final: 0.473044, loss_mean: 0.814122, proj_loss: -0.411834, loss_mean_cls: 0.070756, grad_norm: 0.242556 +Steps: 1%| | 6257/1000000 [25:58<69:06:17, 3.99it/s, grad_norm=0.243, loss_final=0.473, loss_mean=0.814, loss_mean_cls=0.0708, proj_loss=-0.412][2026-03-26 12:57:02] Step: 6257, Training Logs: loss_final: 0.473926, loss_mean: 0.824133, proj_loss: -0.418782, loss_mean_cls: 0.068575, grad_norm: 0.465157 +Steps: 1%| | 6258/1000000 [25:58<68:58:38, 4.00it/s, grad_norm=0.465, loss_final=0.474, loss_mean=0.824, loss_mean_cls=0.0686, proj_loss=-0.419][2026-03-26 12:57:02] Step: 6258, Training Logs: loss_final: 0.471646, loss_mean: 0.820227, proj_loss: -0.416679, loss_mean_cls: 0.068098, grad_norm: 0.398954 +Steps: 1%| | 6259/1000000 [25:59<68:49:54, 4.01it/s, grad_norm=0.399, loss_final=0.472, loss_mean=0.82, loss_mean_cls=0.0681, proj_loss=-0.417][2026-03-26 12:57:02] Step: 6259, Training Logs: loss_final: 0.469072, loss_mean: 0.807901, proj_loss: -0.408483, loss_mean_cls: 0.069653, grad_norm: 0.417880 +Steps: 1%| | 6260/1000000 [25:59<68:42:06, 4.02it/s, grad_norm=0.418, loss_final=0.469, loss_mean=0.808, loss_mean_cls=0.0697, proj_loss=-0.408][2026-03-26 12:57:02] Step: 6260, Training Logs: loss_final: 0.496028, loss_mean: 0.841164, proj_loss: -0.413624, loss_mean_cls: 0.068488, grad_norm: 0.612729 +Steps: 1%| | 6261/1000000 [25:59<68:36:38, 4.02it/s, grad_norm=0.613, loss_final=0.496, loss_mean=0.841, loss_mean_cls=0.0685, proj_loss=-0.414][2026-03-26 12:57:03] Step: 6261, Training Logs: loss_final: 0.475689, loss_mean: 0.815957, proj_loss: -0.411025, loss_mean_cls: 0.070757, grad_norm: 0.273532 +Steps: 1%| | 6262/1000000 [25:59<68:33:05, 4.03it/s, grad_norm=0.274, loss_final=0.476, loss_mean=0.816, loss_mean_cls=0.0708, proj_loss=-0.411][2026-03-26 12:57:03] Step: 6262, Training Logs: loss_final: 0.488460, loss_mean: 0.838715, proj_loss: -0.417949, loss_mean_cls: 0.067694, grad_norm: 0.452364 +Steps: 1%| | 6263/1000000 [26:00<68:27:36, 4.03it/s, grad_norm=0.452, loss_final=0.488, loss_mean=0.839, loss_mean_cls=0.0677, proj_loss=-0.418][2026-03-26 12:57:03] Step: 6263, Training Logs: loss_final: 0.476442, loss_mean: 0.827780, proj_loss: -0.419189, loss_mean_cls: 0.067852, grad_norm: 0.304478 +Steps: 1%| | 6264/1000000 [26:00<68:28:11, 4.03it/s, grad_norm=0.304, loss_final=0.476, loss_mean=0.828, loss_mean_cls=0.0679, proj_loss=-0.419][2026-03-26 12:57:03] Step: 6264, Training Logs: loss_final: 0.478446, loss_mean: 0.818576, proj_loss: -0.409848, loss_mean_cls: 0.069717, grad_norm: 0.540101 +Steps: 1%| | 6265/1000000 [26:00<68:26:43, 4.03it/s, grad_norm=0.54, loss_final=0.478, loss_mean=0.819, loss_mean_cls=0.0697, proj_loss=-0.41][2026-03-26 12:57:04] Step: 6265, Training Logs: loss_final: 0.456445, loss_mean: 0.802902, proj_loss: -0.415730, loss_mean_cls: 0.069273, grad_norm: 0.307907 +Steps: 1%| | 6266/1000000 [26:00<68:22:52, 4.04it/s, grad_norm=0.308, loss_final=0.456, loss_mean=0.803, loss_mean_cls=0.0693, proj_loss=-0.416][2026-03-26 12:57:04] Step: 6266, Training Logs: loss_final: 0.475564, loss_mean: 0.823009, proj_loss: -0.415799, loss_mean_cls: 0.068354, grad_norm: 0.464513 +Steps: 1%| | 6267/1000000 [26:01<68:22:42, 4.04it/s, grad_norm=0.465, loss_final=0.476, loss_mean=0.823, loss_mean_cls=0.0684, proj_loss=-0.416][2026-03-26 12:57:04] Step: 6267, Training Logs: loss_final: 0.449086, loss_mean: 0.800217, proj_loss: -0.419820, loss_mean_cls: 0.068689, grad_norm: 0.378448 +Steps: 1%| | 6268/1000000 [26:01<68:18:48, 4.04it/s, grad_norm=0.378, loss_final=0.449, loss_mean=0.8, loss_mean_cls=0.0687, proj_loss=-0.42][2026-03-26 12:57:04] Step: 6268, Training Logs: loss_final: 0.485680, loss_mean: 0.831309, proj_loss: -0.413835, loss_mean_cls: 0.068207, grad_norm: 0.391016 +Steps: 1%| | 6269/1000000 [26:01<68:21:30, 4.04it/s, grad_norm=0.391, loss_final=0.486, loss_mean=0.831, loss_mean_cls=0.0682, proj_loss=-0.414][2026-03-26 12:57:05] Step: 6269, Training Logs: loss_final: 0.482053, loss_mean: 0.840371, proj_loss: -0.425144, loss_mean_cls: 0.066825, grad_norm: 0.372498 +Steps: 1%| | 6270/1000000 [26:01<68:18:22, 4.04it/s, grad_norm=0.372, loss_final=0.482, loss_mean=0.84, loss_mean_cls=0.0668, proj_loss=-0.425][2026-03-26 12:57:05] Step: 6270, Training Logs: loss_final: 0.476081, loss_mean: 0.816305, proj_loss: -0.410141, loss_mean_cls: 0.069917, grad_norm: 0.254502 +Steps: 1%| | 6271/1000000 [26:02<68:19:50, 4.04it/s, grad_norm=0.255, loss_final=0.476, loss_mean=0.816, loss_mean_cls=0.0699, proj_loss=-0.41][2026-03-26 12:57:05] Step: 6271, Training Logs: loss_final: 0.500627, loss_mean: 0.842426, proj_loss: -0.411470, loss_mean_cls: 0.069671, grad_norm: 0.358442 +Steps: 1%| | 6272/1000000 [26:02<68:19:48, 4.04it/s, grad_norm=0.358, loss_final=0.501, loss_mean=0.842, loss_mean_cls=0.0697, proj_loss=-0.411][2026-03-26 12:57:05] Step: 6272, Training Logs: loss_final: 0.479301, loss_mean: 0.828773, proj_loss: -0.417930, loss_mean_cls: 0.068458, grad_norm: 0.277392 +Steps: 1%| | 6273/1000000 [26:02<68:21:22, 4.04it/s, grad_norm=0.277, loss_final=0.479, loss_mean=0.829, loss_mean_cls=0.0685, proj_loss=-0.418][2026-03-26 12:57:06] Step: 6273, Training Logs: loss_final: 0.482133, loss_mean: 0.821836, proj_loss: -0.410538, loss_mean_cls: 0.070835, grad_norm: 0.389885 +Steps: 1%| | 6274/1000000 [26:02<68:21:47, 4.04it/s, grad_norm=0.39, loss_final=0.482, loss_mean=0.822, loss_mean_cls=0.0708, proj_loss=-0.411][2026-03-26 12:57:06] Step: 6274, Training Logs: loss_final: 0.498732, loss_mean: 0.848294, proj_loss: -0.417058, loss_mean_cls: 0.067496, grad_norm: 0.333507 +Steps: 1%| | 6275/1000000 [26:03<68:22:45, 4.04it/s, grad_norm=0.334, loss_final=0.499, loss_mean=0.848, loss_mean_cls=0.0675, proj_loss=-0.417][2026-03-26 12:57:06] Step: 6275, Training Logs: loss_final: 0.477666, loss_mean: 0.824698, proj_loss: -0.416482, loss_mean_cls: 0.069449, grad_norm: 0.311973 +Steps: 1%| | 6276/1000000 [26:03<68:23:46, 4.04it/s, grad_norm=0.312, loss_final=0.478, loss_mean=0.825, loss_mean_cls=0.0694, proj_loss=-0.416][2026-03-26 12:57:06] Step: 6276, Training Logs: loss_final: 0.472309, loss_mean: 0.819390, proj_loss: -0.415675, loss_mean_cls: 0.068595, grad_norm: 0.425026 +Steps: 1%| | 6277/1000000 [26:03<68:22:35, 4.04it/s, grad_norm=0.425, loss_final=0.472, loss_mean=0.819, loss_mean_cls=0.0686, proj_loss=-0.416][2026-03-26 12:57:07] Step: 6277, Training Logs: loss_final: 0.490221, loss_mean: 0.841009, proj_loss: -0.418718, loss_mean_cls: 0.067930, grad_norm: 0.220260 +Steps: 1%| | 6278/1000000 [26:03<68:23:01, 4.04it/s, grad_norm=0.22, loss_final=0.49, loss_mean=0.841, loss_mean_cls=0.0679, proj_loss=-0.419][2026-03-26 12:57:07] Step: 6278, Training Logs: loss_final: 0.470971, loss_mean: 0.824339, proj_loss: -0.421249, loss_mean_cls: 0.067881, grad_norm: 0.295434 +Steps: 1%| | 6279/1000000 [26:04<68:21:18, 4.04it/s, grad_norm=0.295, loss_final=0.471, loss_mean=0.824, loss_mean_cls=0.0679, proj_loss=-0.421][2026-03-26 12:57:07] Step: 6279, Training Logs: loss_final: 0.487089, loss_mean: 0.823539, proj_loss: -0.405961, loss_mean_cls: 0.069511, grad_norm: 0.306801 +Steps: 1%| | 6280/1000000 [26:04<68:23:01, 4.04it/s, grad_norm=0.307, loss_final=0.487, loss_mean=0.824, loss_mean_cls=0.0695, proj_loss=-0.406][2026-03-26 12:57:07] Step: 6280, Training Logs: loss_final: 0.479602, loss_mean: 0.825856, proj_loss: -0.415337, loss_mean_cls: 0.069083, grad_norm: 0.225295 +Steps: 1%| | 6281/1000000 [26:04<68:20:36, 4.04it/s, grad_norm=0.225, loss_final=0.48, loss_mean=0.826, loss_mean_cls=0.0691, proj_loss=-0.415][2026-03-26 12:57:08] Step: 6281, Training Logs: loss_final: 0.462775, loss_mean: 0.813056, proj_loss: -0.418721, loss_mean_cls: 0.068441, grad_norm: 0.254162 +Steps: 1%| | 6282/1000000 [26:04<68:22:36, 4.04it/s, grad_norm=0.254, loss_final=0.463, loss_mean=0.813, loss_mean_cls=0.0684, proj_loss=-0.419][2026-03-26 12:57:08] Step: 6282, Training Logs: loss_final: 0.490221, loss_mean: 0.842427, proj_loss: -0.420044, loss_mean_cls: 0.067838, grad_norm: 0.320378 +Steps: 1%| | 6283/1000000 [26:05<68:20:32, 4.04it/s, grad_norm=0.32, loss_final=0.49, loss_mean=0.842, loss_mean_cls=0.0678, proj_loss=-0.42][2026-03-26 12:57:08] Step: 6283, Training Logs: loss_final: 0.473683, loss_mean: 0.820508, proj_loss: -0.415834, loss_mean_cls: 0.069009, grad_norm: 0.400703 +Steps: 1%| | 6284/1000000 [26:05<68:22:34, 4.04it/s, grad_norm=0.401, loss_final=0.474, loss_mean=0.821, loss_mean_cls=0.069, proj_loss=-0.416][2026-03-26 12:57:08] Step: 6284, Training Logs: loss_final: 0.470695, loss_mean: 0.815303, proj_loss: -0.413958, loss_mean_cls: 0.069351, grad_norm: 0.266577 +Steps: 1%| | 6285/1000000 [26:05<68:20:12, 4.04it/s, grad_norm=0.267, loss_final=0.471, loss_mean=0.815, loss_mean_cls=0.0694, proj_loss=-0.414][2026-03-26 12:57:08] Step: 6285, Training Logs: loss_final: 0.468451, loss_mean: 0.813075, proj_loss: -0.415157, loss_mean_cls: 0.070533, grad_norm: 0.428609 +Steps: 1%| | 6286/1000000 [26:05<68:21:51, 4.04it/s, grad_norm=0.429, loss_final=0.468, loss_mean=0.813, loss_mean_cls=0.0705, proj_loss=-0.415][2026-03-26 12:57:09] Step: 6286, Training Logs: loss_final: 0.467985, loss_mean: 0.813920, proj_loss: -0.415756, loss_mean_cls: 0.069821, grad_norm: 0.578115 +Steps: 1%| | 6287/1000000 [26:06<68:19:13, 4.04it/s, grad_norm=0.578, loss_final=0.468, loss_mean=0.814, loss_mean_cls=0.0698, proj_loss=-0.416][2026-03-26 12:57:09] Step: 6287, Training Logs: loss_final: 0.476378, loss_mean: 0.820632, proj_loss: -0.412499, loss_mean_cls: 0.068244, grad_norm: 0.235694 +Steps: 1%| | 6288/1000000 [26:06<68:21:23, 4.04it/s, grad_norm=0.236, loss_final=0.476, loss_mean=0.821, loss_mean_cls=0.0682, proj_loss=-0.412][2026-03-26 12:57:09] Step: 6288, Training Logs: loss_final: 0.467553, loss_mean: 0.808716, proj_loss: -0.410978, loss_mean_cls: 0.069815, grad_norm: 0.502296 +Steps: 1%| | 6289/1000000 [26:06<68:19:02, 4.04it/s, grad_norm=0.502, loss_final=0.468, loss_mean=0.809, loss_mean_cls=0.0698, proj_loss=-0.411][2026-03-26 12:57:09] Step: 6289, Training Logs: loss_final: 0.479846, loss_mean: 0.824444, proj_loss: -0.413389, loss_mean_cls: 0.068790, grad_norm: 0.561294 +Steps: 1%| | 6290/1000000 [26:06<68:23:47, 4.04it/s, grad_norm=0.561, loss_final=0.48, loss_mean=0.824, loss_mean_cls=0.0688, proj_loss=-0.413][2026-03-26 12:57:10] Step: 6290, Training Logs: loss_final: 0.475578, loss_mean: 0.816049, proj_loss: -0.411001, loss_mean_cls: 0.070530, grad_norm: 0.246982 +Steps: 1%| | 6291/1000000 [26:07<68:23:28, 4.04it/s, grad_norm=0.247, loss_final=0.476, loss_mean=0.816, loss_mean_cls=0.0705, proj_loss=-0.411][2026-03-26 12:57:10] Step: 6291, Training Logs: loss_final: 0.487453, loss_mean: 0.841568, proj_loss: -0.422118, loss_mean_cls: 0.068003, grad_norm: 0.431835 +Steps: 1%| | 6292/1000000 [26:07<68:22:09, 4.04it/s, grad_norm=0.432, loss_final=0.487, loss_mean=0.842, loss_mean_cls=0.068, proj_loss=-0.422][2026-03-26 12:57:10] Step: 6292, Training Logs: loss_final: 0.477859, loss_mean: 0.827386, proj_loss: -0.417695, loss_mean_cls: 0.068168, grad_norm: 0.368500 +Steps: 1%| | 6293/1000000 [26:07<68:18:26, 4.04it/s, grad_norm=0.369, loss_final=0.478, loss_mean=0.827, loss_mean_cls=0.0682, proj_loss=-0.418][2026-03-26 12:57:10] Step: 6293, Training Logs: loss_final: 0.491087, loss_mean: 0.846330, proj_loss: -0.422620, loss_mean_cls: 0.067377, grad_norm: 0.363165 +Steps: 1%| | 6294/1000000 [26:07<68:19:59, 4.04it/s, grad_norm=0.363, loss_final=0.491, loss_mean=0.846, loss_mean_cls=0.0674, proj_loss=-0.423][2026-03-26 12:57:11] Step: 6294, Training Logs: loss_final: 0.467587, loss_mean: 0.814475, proj_loss: -0.416086, loss_mean_cls: 0.069198, grad_norm: 0.318614 +Steps: 1%| | 6295/1000000 [26:08<68:19:23, 4.04it/s, grad_norm=0.319, loss_final=0.468, loss_mean=0.814, loss_mean_cls=0.0692, proj_loss=-0.416][2026-03-26 12:57:11] Step: 6295, Training Logs: loss_final: 0.473143, loss_mean: 0.808956, proj_loss: -0.407700, loss_mean_cls: 0.071887, grad_norm: 0.352216 +Steps: 1%| | 6296/1000000 [26:08<68:18:43, 4.04it/s, grad_norm=0.352, loss_final=0.473, loss_mean=0.809, loss_mean_cls=0.0719, proj_loss=-0.408][2026-03-26 12:57:11] Step: 6296, Training Logs: loss_final: 0.473158, loss_mean: 0.813578, proj_loss: -0.410814, loss_mean_cls: 0.070394, grad_norm: 0.360231 +Steps: 1%| | 6297/1000000 [26:08<68:19:50, 4.04it/s, grad_norm=0.36, loss_final=0.473, loss_mean=0.814, loss_mean_cls=0.0704, proj_loss=-0.411][2026-03-26 12:57:11] Step: 6297, Training Logs: loss_final: 0.470046, loss_mean: 0.816188, proj_loss: -0.415521, loss_mean_cls: 0.069379, grad_norm: 0.360213 +Steps: 1%| | 6298/1000000 [26:08<68:19:54, 4.04it/s, grad_norm=0.36, loss_final=0.47, loss_mean=0.816, loss_mean_cls=0.0694, proj_loss=-0.416][2026-03-26 12:57:12] Step: 6298, Training Logs: loss_final: 0.480235, loss_mean: 0.816199, proj_loss: -0.407017, loss_mean_cls: 0.071054, grad_norm: 0.355039 +Steps: 1%| | 6299/1000000 [26:09<68:20:07, 4.04it/s, grad_norm=0.355, loss_final=0.48, loss_mean=0.816, loss_mean_cls=0.0711, proj_loss=-0.407][2026-03-26 12:57:12] Step: 6299, Training Logs: loss_final: 0.465755, loss_mean: 0.805942, proj_loss: -0.410672, loss_mean_cls: 0.070485, grad_norm: 0.445208 +Steps: 1%| | 6300/1000000 [26:09<68:21:10, 4.04it/s, grad_norm=0.445, loss_final=0.466, loss_mean=0.806, loss_mean_cls=0.0705, proj_loss=-0.411][2026-03-26 12:57:12] Step: 6300, Training Logs: loss_final: 0.493902, loss_mean: 0.837777, proj_loss: -0.412592, loss_mean_cls: 0.068717, grad_norm: 0.466550 +Steps: 1%| | 6301/1000000 [26:09<68:23:57, 4.04it/s, grad_norm=0.467, loss_final=0.494, loss_mean=0.838, loss_mean_cls=0.0687, proj_loss=-0.413][2026-03-26 12:57:12] Step: 6301, Training Logs: loss_final: 0.479707, loss_mean: 0.821650, proj_loss: -0.412494, loss_mean_cls: 0.070551, grad_norm: 0.405014 +Steps: 1%| | 6302/1000000 [26:09<68:21:41, 4.04it/s, grad_norm=0.405, loss_final=0.48, loss_mean=0.822, loss_mean_cls=0.0706, proj_loss=-0.412][2026-03-26 12:57:13] Step: 6302, Training Logs: loss_final: 0.468788, loss_mean: 0.811759, proj_loss: -0.412146, loss_mean_cls: 0.069175, grad_norm: 0.431604 +Steps: 1%| | 6303/1000000 [26:10<68:21:08, 4.04it/s, grad_norm=0.432, loss_final=0.469, loss_mean=0.812, loss_mean_cls=0.0692, proj_loss=-0.412][2026-03-26 12:57:13] Step: 6303, Training Logs: loss_final: 0.477426, loss_mean: 0.822538, proj_loss: -0.413643, loss_mean_cls: 0.068530, grad_norm: 0.445855 +Steps: 1%| | 6304/1000000 [26:10<68:18:45, 4.04it/s, grad_norm=0.446, loss_final=0.477, loss_mean=0.823, loss_mean_cls=0.0685, proj_loss=-0.414][2026-03-26 12:57:13] Step: 6304, Training Logs: loss_final: 0.464401, loss_mean: 0.820750, proj_loss: -0.423913, loss_mean_cls: 0.067564, grad_norm: 0.466950 +Steps: 1%| | 6305/1000000 [26:10<68:17:33, 4.04it/s, grad_norm=0.467, loss_final=0.464, loss_mean=0.821, loss_mean_cls=0.0676, proj_loss=-0.424][2026-03-26 12:57:13] Step: 6305, Training Logs: loss_final: 0.463707, loss_mean: 0.810884, proj_loss: -0.416352, loss_mean_cls: 0.069175, grad_norm: 0.288972 +Steps: 1%| | 6306/1000000 [26:10<68:25:05, 4.03it/s, grad_norm=0.289, loss_final=0.464, loss_mean=0.811, loss_mean_cls=0.0692, proj_loss=-0.416][2026-03-26 12:57:14] Step: 6306, Training Logs: loss_final: 0.473339, loss_mean: 0.823227, proj_loss: -0.418684, loss_mean_cls: 0.068796, grad_norm: 0.490658 +Steps: 1%| | 6307/1000000 [26:11<68:43:11, 4.02it/s, grad_norm=0.491, loss_final=0.473, loss_mean=0.823, loss_mean_cls=0.0688, proj_loss=-0.419][2026-03-26 12:57:14] Step: 6307, Training Logs: loss_final: 0.479102, loss_mean: 0.810872, proj_loss: -0.403189, loss_mean_cls: 0.071419, grad_norm: 0.271794 +Steps: 1%| | 6308/1000000 [26:11<68:34:07, 4.03it/s, grad_norm=0.272, loss_final=0.479, loss_mean=0.811, loss_mean_cls=0.0714, proj_loss=-0.403][2026-03-26 12:57:14] Step: 6308, Training Logs: loss_final: 0.457532, loss_mean: 0.808704, proj_loss: -0.418635, loss_mean_cls: 0.067462, grad_norm: 0.266705 +Steps: 1%| | 6309/1000000 [26:11<68:28:09, 4.03it/s, grad_norm=0.267, loss_final=0.458, loss_mean=0.809, loss_mean_cls=0.0675, proj_loss=-0.419][2026-03-26 12:57:14] Step: 6309, Training Logs: loss_final: 0.459391, loss_mean: 0.817301, proj_loss: -0.423677, loss_mean_cls: 0.065767, grad_norm: 0.257024 +Steps: 1%| | 6310/1000000 [26:11<68:23:33, 4.04it/s, grad_norm=0.257, loss_final=0.459, loss_mean=0.817, loss_mean_cls=0.0658, proj_loss=-0.424][2026-03-26 12:57:15] Step: 6310, Training Logs: loss_final: 0.473598, loss_mean: 0.810049, proj_loss: -0.406662, loss_mean_cls: 0.070211, grad_norm: 0.282732 +Steps: 1%| | 6311/1000000 [26:12<68:22:56, 4.04it/s, grad_norm=0.283, loss_final=0.474, loss_mean=0.81, loss_mean_cls=0.0702, proj_loss=-0.407][2026-03-26 12:57:15] Step: 6311, Training Logs: loss_final: 0.486898, loss_mean: 0.844366, proj_loss: -0.424217, loss_mean_cls: 0.066749, grad_norm: 0.267192 +Steps: 1%| | 6312/1000000 [26:12<68:20:50, 4.04it/s, grad_norm=0.267, loss_final=0.487, loss_mean=0.844, loss_mean_cls=0.0667, proj_loss=-0.424][2026-03-26 12:57:15] Step: 6312, Training Logs: loss_final: 0.481560, loss_mean: 0.832251, proj_loss: -0.419638, loss_mean_cls: 0.068947, grad_norm: 0.346434 +Steps: 1%| | 6313/1000000 [26:12<68:18:49, 4.04it/s, grad_norm=0.346, loss_final=0.482, loss_mean=0.832, loss_mean_cls=0.0689, proj_loss=-0.42][2026-03-26 12:57:15] Step: 6313, Training Logs: loss_final: 0.485278, loss_mean: 0.815800, proj_loss: -0.402040, loss_mean_cls: 0.071518, grad_norm: 0.299831 +Steps: 1%| | 6314/1000000 [26:12<68:17:03, 4.04it/s, grad_norm=0.3, loss_final=0.485, loss_mean=0.816, loss_mean_cls=0.0715, proj_loss=-0.402][2026-03-26 12:57:16] Step: 6314, Training Logs: loss_final: 0.483208, loss_mean: 0.829472, proj_loss: -0.415523, loss_mean_cls: 0.069259, grad_norm: 0.387857 +Steps: 1%| | 6315/1000000 [26:13<68:17:28, 4.04it/s, grad_norm=0.388, loss_final=0.483, loss_mean=0.829, loss_mean_cls=0.0693, proj_loss=-0.416][2026-03-26 12:57:16] Step: 6315, Training Logs: loss_final: 0.467639, loss_mean: 0.817529, proj_loss: -0.417776, loss_mean_cls: 0.067886, grad_norm: 0.385889 +Steps: 1%| | 6316/1000000 [26:13<68:19:13, 4.04it/s, grad_norm=0.386, loss_final=0.468, loss_mean=0.818, loss_mean_cls=0.0679, proj_loss=-0.418][2026-03-26 12:57:16] Step: 6316, Training Logs: loss_final: 0.481397, loss_mean: 0.826774, proj_loss: -0.414808, loss_mean_cls: 0.069432, grad_norm: 0.237555 +Steps: 1%| | 6317/1000000 [26:13<68:18:23, 4.04it/s, grad_norm=0.238, loss_final=0.481, loss_mean=0.827, loss_mean_cls=0.0694, proj_loss=-0.415][2026-03-26 12:57:16] Step: 6317, Training Logs: loss_final: 0.459456, loss_mean: 0.812283, proj_loss: -0.420569, loss_mean_cls: 0.067742, grad_norm: 0.374506 +Steps: 1%| | 6318/1000000 [26:13<68:18:56, 4.04it/s, grad_norm=0.375, loss_final=0.459, loss_mean=0.812, loss_mean_cls=0.0677, proj_loss=-0.421][2026-03-26 12:57:17] Step: 6318, Training Logs: loss_final: 0.477104, loss_mean: 0.827044, proj_loss: -0.417947, loss_mean_cls: 0.068007, grad_norm: 0.389207 +Steps: 1%| | 6319/1000000 [26:14<68:18:56, 4.04it/s, grad_norm=0.389, loss_final=0.477, loss_mean=0.827, loss_mean_cls=0.068, proj_loss=-0.418][2026-03-26 12:57:17] Step: 6319, Training Logs: loss_final: 0.489267, loss_mean: 0.834849, proj_loss: -0.415001, loss_mean_cls: 0.069419, grad_norm: 0.434753 +Steps: 1%| | 6320/1000000 [26:14<68:20:42, 4.04it/s, grad_norm=0.435, loss_final=0.489, loss_mean=0.835, loss_mean_cls=0.0694, proj_loss=-0.415][2026-03-26 12:57:17] Step: 6320, Training Logs: loss_final: 0.481140, loss_mean: 0.830782, proj_loss: -0.417819, loss_mean_cls: 0.068178, grad_norm: 0.263292 +Steps: 1%| | 6321/1000000 [26:14<68:21:28, 4.04it/s, grad_norm=0.263, loss_final=0.481, loss_mean=0.831, loss_mean_cls=0.0682, proj_loss=-0.418][2026-03-26 12:57:17] Step: 6321, Training Logs: loss_final: 0.483018, loss_mean: 0.835347, proj_loss: -0.419897, loss_mean_cls: 0.067569, grad_norm: 0.461535 +Steps: 1%| | 6322/1000000 [26:14<68:21:16, 4.04it/s, grad_norm=0.462, loss_final=0.483, loss_mean=0.835, loss_mean_cls=0.0676, proj_loss=-0.42][2026-03-26 12:57:18] Step: 6322, Training Logs: loss_final: 0.461008, loss_mean: 0.822682, proj_loss: -0.427270, loss_mean_cls: 0.065596, grad_norm: 0.325070 +Steps: 1%| | 6323/1000000 [26:15<68:20:17, 4.04it/s, grad_norm=0.325, loss_final=0.461, loss_mean=0.823, loss_mean_cls=0.0656, proj_loss=-0.427][2026-03-26 12:57:18] Step: 6323, Training Logs: loss_final: 0.467966, loss_mean: 0.807143, proj_loss: -0.410671, loss_mean_cls: 0.071494, grad_norm: 0.477272 +Steps: 1%| | 6324/1000000 [26:15<68:21:16, 4.04it/s, grad_norm=0.477, loss_final=0.468, loss_mean=0.807, loss_mean_cls=0.0715, proj_loss=-0.411][2026-03-26 12:57:18] Step: 6324, Training Logs: loss_final: 0.481599, loss_mean: 0.834426, proj_loss: -0.420733, loss_mean_cls: 0.067907, grad_norm: 0.548589 +Steps: 1%| | 6325/1000000 [26:15<68:17:57, 4.04it/s, grad_norm=0.549, loss_final=0.482, loss_mean=0.834, loss_mean_cls=0.0679, proj_loss=-0.421][2026-03-26 12:57:18] Step: 6325, Training Logs: loss_final: 0.486677, loss_mean: 0.833368, proj_loss: -0.414911, loss_mean_cls: 0.068220, grad_norm: 0.309686 +Steps: 1%| | 6326/1000000 [26:15<68:19:46, 4.04it/s, grad_norm=0.31, loss_final=0.487, loss_mean=0.833, loss_mean_cls=0.0682, proj_loss=-0.415][2026-03-26 12:57:19] Step: 6326, Training Logs: loss_final: 0.487322, loss_mean: 0.837597, proj_loss: -0.417982, loss_mean_cls: 0.067706, grad_norm: 0.533280 +Steps: 1%| | 6327/1000000 [26:16<68:16:19, 4.04it/s, grad_norm=0.533, loss_final=0.487, loss_mean=0.838, loss_mean_cls=0.0677, proj_loss=-0.418][2026-03-26 12:57:19] Step: 6327, Training Logs: loss_final: 0.493753, loss_mean: 0.845983, proj_loss: -0.419683, loss_mean_cls: 0.067453, grad_norm: 0.222994 +Steps: 1%| | 6328/1000000 [26:16<68:18:48, 4.04it/s, grad_norm=0.223, loss_final=0.494, loss_mean=0.846, loss_mean_cls=0.0675, proj_loss=-0.42][2026-03-26 12:57:19] Step: 6328, Training Logs: loss_final: 0.460337, loss_mean: 0.809083, proj_loss: -0.417342, loss_mean_cls: 0.068596, grad_norm: 0.590141 +Steps: 1%| | 6329/1000000 [26:16<68:15:32, 4.04it/s, grad_norm=0.59, loss_final=0.46, loss_mean=0.809, loss_mean_cls=0.0686, proj_loss=-0.417][2026-03-26 12:57:19] Step: 6329, Training Logs: loss_final: 0.477682, loss_mean: 0.828743, proj_loss: -0.418829, loss_mean_cls: 0.067768, grad_norm: 0.481043 +Steps: 1%| | 6330/1000000 [26:16<68:16:26, 4.04it/s, grad_norm=0.481, loss_final=0.478, loss_mean=0.829, loss_mean_cls=0.0678, proj_loss=-0.419][2026-03-26 12:57:20] Step: 6330, Training Logs: loss_final: 0.488207, loss_mean: 0.831997, proj_loss: -0.412408, loss_mean_cls: 0.068618, grad_norm: 0.534369 +Steps: 1%| | 6331/1000000 [26:16<68:18:21, 4.04it/s, grad_norm=0.534, loss_final=0.488, loss_mean=0.832, loss_mean_cls=0.0686, proj_loss=-0.412][2026-03-26 12:57:20] Step: 6331, Training Logs: loss_final: 0.472861, loss_mean: 0.830072, proj_loss: -0.424557, loss_mean_cls: 0.067346, grad_norm: 0.802262 +Steps: 1%| | 6332/1000000 [26:17<68:16:19, 4.04it/s, grad_norm=0.802, loss_final=0.473, loss_mean=0.83, loss_mean_cls=0.0673, proj_loss=-0.425][2026-03-26 12:57:20] Step: 6332, Training Logs: loss_final: 0.470325, loss_mean: 0.828191, proj_loss: -0.424216, loss_mean_cls: 0.066351, grad_norm: 0.332345 +Steps: 1%| | 6333/1000000 [26:17<68:13:45, 4.05it/s, grad_norm=0.332, loss_final=0.47, loss_mean=0.828, loss_mean_cls=0.0664, proj_loss=-0.424][2026-03-26 12:57:20] Step: 6333, Training Logs: loss_final: 0.457881, loss_mean: 0.812871, proj_loss: -0.422375, loss_mean_cls: 0.067386, grad_norm: 0.419573 +Steps: 1%| | 6334/1000000 [26:17<68:17:26, 4.04it/s, grad_norm=0.42, loss_final=0.458, loss_mean=0.813, loss_mean_cls=0.0674, proj_loss=-0.422][2026-03-26 12:57:21] Step: 6334, Training Logs: loss_final: 0.473821, loss_mean: 0.815766, proj_loss: -0.411300, loss_mean_cls: 0.069355, grad_norm: 0.319274 +Steps: 1%| | 6335/1000000 [26:17<68:17:11, 4.04it/s, grad_norm=0.319, loss_final=0.474, loss_mean=0.816, loss_mean_cls=0.0694, proj_loss=-0.411][2026-03-26 12:57:21] Step: 6335, Training Logs: loss_final: 0.479874, loss_mean: 0.823200, proj_loss: -0.413417, loss_mean_cls: 0.070090, grad_norm: 0.617378 +Steps: 1%| | 6336/1000000 [26:18<68:19:29, 4.04it/s, grad_norm=0.617, loss_final=0.48, loss_mean=0.823, loss_mean_cls=0.0701, proj_loss=-0.413][2026-03-26 12:57:21] Step: 6336, Training Logs: loss_final: 0.465734, loss_mean: 0.801180, proj_loss: -0.407015, loss_mean_cls: 0.071569, grad_norm: 0.384655 +Steps: 1%| | 6337/1000000 [26:18<68:18:31, 4.04it/s, grad_norm=0.385, loss_final=0.466, loss_mean=0.801, loss_mean_cls=0.0716, proj_loss=-0.407][2026-03-26 12:57:21] Step: 6337, Training Logs: loss_final: 0.483286, loss_mean: 0.817407, proj_loss: -0.405454, loss_mean_cls: 0.071333, grad_norm: 0.438873 +Steps: 1%| | 6338/1000000 [26:18<68:19:36, 4.04it/s, grad_norm=0.439, loss_final=0.483, loss_mean=0.817, loss_mean_cls=0.0713, proj_loss=-0.405][2026-03-26 12:57:22] Step: 6338, Training Logs: loss_final: 0.475802, loss_mean: 0.820342, proj_loss: -0.413984, loss_mean_cls: 0.069444, grad_norm: 0.469107 +Steps: 1%| | 6339/1000000 [26:18<68:20:27, 4.04it/s, grad_norm=0.469, loss_final=0.476, loss_mean=0.82, loss_mean_cls=0.0694, proj_loss=-0.414][2026-03-26 12:57:22] Step: 6339, Training Logs: loss_final: 0.488786, loss_mean: 0.827527, proj_loss: -0.408724, loss_mean_cls: 0.069983, grad_norm: 0.468322 +Steps: 1%| | 6340/1000000 [26:19<68:18:21, 4.04it/s, grad_norm=0.468, loss_final=0.489, loss_mean=0.828, loss_mean_cls=0.07, proj_loss=-0.409][2026-03-26 12:57:22] Step: 6340, Training Logs: loss_final: 0.482792, loss_mean: 0.839757, proj_loss: -0.423888, loss_mean_cls: 0.066923, grad_norm: 0.364222 +Steps: 1%| | 6341/1000000 [26:19<68:18:46, 4.04it/s, grad_norm=0.364, loss_final=0.483, loss_mean=0.84, loss_mean_cls=0.0669, proj_loss=-0.424][2026-03-26 12:57:22] Step: 6341, Training Logs: loss_final: 0.482122, loss_mean: 0.819963, proj_loss: -0.408557, loss_mean_cls: 0.070716, grad_norm: 0.468340 +Steps: 1%| | 6342/1000000 [26:19<68:38:29, 4.02it/s, grad_norm=0.468, loss_final=0.482, loss_mean=0.82, loss_mean_cls=0.0707, proj_loss=-0.409][2026-03-26 12:57:23] Step: 6342, Training Logs: loss_final: 0.472641, loss_mean: 0.823893, proj_loss: -0.418461, loss_mean_cls: 0.067209, grad_norm: 0.530227 +Steps: 1%| | 6343/1000000 [26:19<68:34:56, 4.02it/s, grad_norm=0.53, loss_final=0.473, loss_mean=0.824, loss_mean_cls=0.0672, proj_loss=-0.418][2026-03-26 12:57:23] Step: 6343, Training Logs: loss_final: 0.466059, loss_mean: 0.816039, proj_loss: -0.418490, loss_mean_cls: 0.068510, grad_norm: 0.324843 +Steps: 1%| | 6344/1000000 [26:20<68:30:41, 4.03it/s, grad_norm=0.325, loss_final=0.466, loss_mean=0.816, loss_mean_cls=0.0685, proj_loss=-0.418][2026-03-26 12:57:23] Step: 6344, Training Logs: loss_final: 0.476377, loss_mean: 0.816916, proj_loss: -0.411862, loss_mean_cls: 0.071323, grad_norm: 0.605909 +Steps: 1%| | 6345/1000000 [26:20<68:26:09, 4.03it/s, grad_norm=0.606, loss_final=0.476, loss_mean=0.817, loss_mean_cls=0.0713, proj_loss=-0.412][2026-03-26 12:57:23] Step: 6345, Training Logs: loss_final: 0.455308, loss_mean: 0.801681, proj_loss: -0.415965, loss_mean_cls: 0.069592, grad_norm: 0.364710 +Steps: 1%| | 6346/1000000 [26:20<68:23:23, 4.04it/s, grad_norm=0.365, loss_final=0.455, loss_mean=0.802, loss_mean_cls=0.0696, proj_loss=-0.416][2026-03-26 12:57:24] Step: 6346, Training Logs: loss_final: 0.488353, loss_mean: 0.848198, proj_loss: -0.425430, loss_mean_cls: 0.065585, grad_norm: 0.737346 +Steps: 1%| | 6347/1000000 [26:20<68:21:13, 4.04it/s, grad_norm=0.737, loss_final=0.488, loss_mean=0.848, loss_mean_cls=0.0656, proj_loss=-0.425][2026-03-26 12:57:24] Step: 6347, Training Logs: loss_final: 0.468768, loss_mean: 0.820772, proj_loss: -0.419160, loss_mean_cls: 0.067156, grad_norm: 0.311733 +Steps: 1%| | 6348/1000000 [26:21<68:19:10, 4.04it/s, grad_norm=0.312, loss_final=0.469, loss_mean=0.821, loss_mean_cls=0.0672, proj_loss=-0.419][2026-03-26 12:57:24] Step: 6348, Training Logs: loss_final: 0.470461, loss_mean: 0.828363, proj_loss: -0.424664, loss_mean_cls: 0.066762, grad_norm: 0.609961 +Steps: 1%| | 6349/1000000 [26:21<68:21:01, 4.04it/s, grad_norm=0.61, loss_final=0.47, loss_mean=0.828, loss_mean_cls=0.0668, proj_loss=-0.425][2026-03-26 12:57:24] Step: 6349, Training Logs: loss_final: 0.490831, loss_mean: 0.830679, proj_loss: -0.410891, loss_mean_cls: 0.071042, grad_norm: 0.628333 +Steps: 1%| | 6350/1000000 [26:21<68:19:32, 4.04it/s, grad_norm=0.628, loss_final=0.491, loss_mean=0.831, loss_mean_cls=0.071, proj_loss=-0.411][2026-03-26 12:57:25] Step: 6350, Training Logs: loss_final: 0.479613, loss_mean: 0.830100, proj_loss: -0.419224, loss_mean_cls: 0.068737, grad_norm: 0.262218 +Steps: 1%| | 6351/1000000 [26:21<68:19:58, 4.04it/s, grad_norm=0.262, loss_final=0.48, loss_mean=0.83, loss_mean_cls=0.0687, proj_loss=-0.419][2026-03-26 12:57:25] Step: 6351, Training Logs: loss_final: 0.474159, loss_mean: 0.824310, proj_loss: -0.419214, loss_mean_cls: 0.069062, grad_norm: 0.435093 +Steps: 1%| | 6352/1000000 [26:22<68:17:31, 4.04it/s, grad_norm=0.435, loss_final=0.474, loss_mean=0.824, loss_mean_cls=0.0691, proj_loss=-0.419][2026-03-26 12:57:25] Step: 6352, Training Logs: loss_final: 0.476626, loss_mean: 0.816170, proj_loss: -0.409866, loss_mean_cls: 0.070322, grad_norm: 0.318027 +Steps: 1%| | 6353/1000000 [26:22<68:18:33, 4.04it/s, grad_norm=0.318, loss_final=0.477, loss_mean=0.816, loss_mean_cls=0.0703, proj_loss=-0.41][2026-03-26 12:57:25] Step: 6353, Training Logs: loss_final: 0.478456, loss_mean: 0.837120, proj_loss: -0.425250, loss_mean_cls: 0.066586, grad_norm: 0.417140 +Steps: 1%| | 6354/1000000 [26:22<68:18:40, 4.04it/s, grad_norm=0.417, loss_final=0.478, loss_mean=0.837, loss_mean_cls=0.0666, proj_loss=-0.425][2026-03-26 12:57:26] Step: 6354, Training Logs: loss_final: 0.473008, loss_mean: 0.817136, proj_loss: -0.414425, loss_mean_cls: 0.070297, grad_norm: 0.462235 +Steps: 1%| | 6355/1000000 [26:22<68:18:35, 4.04it/s, grad_norm=0.462, loss_final=0.473, loss_mean=0.817, loss_mean_cls=0.0703, proj_loss=-0.414][2026-03-26 12:57:26] Step: 6355, Training Logs: loss_final: 0.470324, loss_mean: 0.808104, proj_loss: -0.408146, loss_mean_cls: 0.070365, grad_norm: 0.347542 +Steps: 1%| | 6356/1000000 [26:23<68:22:05, 4.04it/s, grad_norm=0.348, loss_final=0.47, loss_mean=0.808, loss_mean_cls=0.0704, proj_loss=-0.408][2026-03-26 12:57:26] Step: 6356, Training Logs: loss_final: 0.484766, loss_mean: 0.828978, proj_loss: -0.413068, loss_mean_cls: 0.068857, grad_norm: 0.358270 +Steps: 1%| | 6357/1000000 [26:23<68:20:44, 4.04it/s, grad_norm=0.358, loss_final=0.485, loss_mean=0.829, loss_mean_cls=0.0689, proj_loss=-0.413][2026-03-26 12:57:26] Step: 6357, Training Logs: loss_final: 0.468072, loss_mean: 0.811516, proj_loss: -0.413650, loss_mean_cls: 0.070206, grad_norm: 0.324913 +Steps: 1%| | 6358/1000000 [26:23<68:20:04, 4.04it/s, grad_norm=0.325, loss_final=0.468, loss_mean=0.812, loss_mean_cls=0.0702, proj_loss=-0.414][2026-03-26 12:57:27] Step: 6358, Training Logs: loss_final: 0.474201, loss_mean: 0.813774, proj_loss: -0.410936, loss_mean_cls: 0.071363, grad_norm: 0.337495 +Steps: 1%| | 6359/1000000 [26:23<68:20:51, 4.04it/s, grad_norm=0.337, loss_final=0.474, loss_mean=0.814, loss_mean_cls=0.0714, proj_loss=-0.411][2026-03-26 12:57:27] Step: 6359, Training Logs: loss_final: 0.479335, loss_mean: 0.814705, proj_loss: -0.406124, loss_mean_cls: 0.070754, grad_norm: 0.359982 +Steps: 1%| | 6360/1000000 [26:24<68:20:36, 4.04it/s, grad_norm=0.36, loss_final=0.479, loss_mean=0.815, loss_mean_cls=0.0708, proj_loss=-0.406][2026-03-26 12:57:27] Step: 6360, Training Logs: loss_final: 0.462481, loss_mean: 0.818093, proj_loss: -0.422412, loss_mean_cls: 0.066799, grad_norm: 0.293535 +Steps: 1%| | 6361/1000000 [26:24<68:21:58, 4.04it/s, grad_norm=0.294, loss_final=0.462, loss_mean=0.818, loss_mean_cls=0.0668, proj_loss=-0.422][2026-03-26 12:57:27] Step: 6361, Training Logs: loss_final: 0.467256, loss_mean: 0.814056, proj_loss: -0.416525, loss_mean_cls: 0.069725, grad_norm: 0.498681 +Steps: 1%| | 6362/1000000 [26:24<68:22:48, 4.04it/s, grad_norm=0.499, loss_final=0.467, loss_mean=0.814, loss_mean_cls=0.0697, proj_loss=-0.417][2026-03-26 12:57:28] Step: 6362, Training Logs: loss_final: 0.474762, loss_mean: 0.816319, proj_loss: -0.411658, loss_mean_cls: 0.070102, grad_norm: 0.324004 +Steps: 1%| | 6363/1000000 [26:24<68:22:26, 4.04it/s, grad_norm=0.324, loss_final=0.475, loss_mean=0.816, loss_mean_cls=0.0701, proj_loss=-0.412][2026-03-26 12:57:28] Step: 6363, Training Logs: loss_final: 0.475158, loss_mean: 0.818079, proj_loss: -0.412400, loss_mean_cls: 0.069480, grad_norm: 0.379310 +Steps: 1%| | 6364/1000000 [26:25<68:23:20, 4.04it/s, grad_norm=0.379, loss_final=0.475, loss_mean=0.818, loss_mean_cls=0.0695, proj_loss=-0.412][2026-03-26 12:57:28] Step: 6364, Training Logs: loss_final: 0.462460, loss_mean: 0.812671, proj_loss: -0.418078, loss_mean_cls: 0.067866, grad_norm: 0.349721 +Steps: 1%| | 6365/1000000 [26:25<68:22:24, 4.04it/s, grad_norm=0.35, loss_final=0.462, loss_mean=0.813, loss_mean_cls=0.0679, proj_loss=-0.418][2026-03-26 12:57:28] Step: 6365, Training Logs: loss_final: 0.468162, loss_mean: 0.820761, proj_loss: -0.419930, loss_mean_cls: 0.067330, grad_norm: 0.345646 +Steps: 1%| | 6366/1000000 [26:25<68:23:14, 4.04it/s, grad_norm=0.346, loss_final=0.468, loss_mean=0.821, loss_mean_cls=0.0673, proj_loss=-0.42][2026-03-26 12:57:29] Step: 6366, Training Logs: loss_final: 0.466506, loss_mean: 0.816354, proj_loss: -0.417868, loss_mean_cls: 0.068019, grad_norm: 0.339342 +Steps: 1%| | 6367/1000000 [26:25<68:22:12, 4.04it/s, grad_norm=0.339, loss_final=0.467, loss_mean=0.816, loss_mean_cls=0.068, proj_loss=-0.418][2026-03-26 12:57:29] Step: 6367, Training Logs: loss_final: 0.483316, loss_mean: 0.826620, proj_loss: -0.413484, loss_mean_cls: 0.070180, grad_norm: 0.482725 +Steps: 1%| | 6368/1000000 [26:26<68:24:10, 4.04it/s, grad_norm=0.483, loss_final=0.483, loss_mean=0.827, loss_mean_cls=0.0702, proj_loss=-0.413][2026-03-26 12:57:29] Step: 6368, Training Logs: loss_final: 0.485588, loss_mean: 0.821960, proj_loss: -0.406167, loss_mean_cls: 0.069796, grad_norm: 0.411969 +Steps: 1%| | 6369/1000000 [26:26<68:20:15, 4.04it/s, grad_norm=0.412, loss_final=0.486, loss_mean=0.822, loss_mean_cls=0.0698, proj_loss=-0.406][2026-03-26 12:57:29] Step: 6369, Training Logs: loss_final: 0.473178, loss_mean: 0.813717, proj_loss: -0.409073, loss_mean_cls: 0.068535, grad_norm: 0.365287 +Steps: 1%| | 6370/1000000 [26:26<68:21:19, 4.04it/s, grad_norm=0.365, loss_final=0.473, loss_mean=0.814, loss_mean_cls=0.0685, proj_loss=-0.409][2026-03-26 12:57:30] Step: 6370, Training Logs: loss_final: 0.481225, loss_mean: 0.806448, proj_loss: -0.397859, loss_mean_cls: 0.072637, grad_norm: 0.359824 +Steps: 1%| | 6371/1000000 [26:26<68:16:52, 4.04it/s, grad_norm=0.36, loss_final=0.481, loss_mean=0.806, loss_mean_cls=0.0726, proj_loss=-0.398][2026-03-26 12:57:30] Step: 6371, Training Logs: loss_final: 0.472286, loss_mean: 0.814226, proj_loss: -0.411854, loss_mean_cls: 0.069913, grad_norm: 0.443975 +Steps: 1%| | 6372/1000000 [26:27<68:17:22, 4.04it/s, grad_norm=0.444, loss_final=0.472, loss_mean=0.814, loss_mean_cls=0.0699, proj_loss=-0.412][2026-03-26 12:57:30] Step: 6372, Training Logs: loss_final: 0.471653, loss_mean: 0.829223, proj_loss: -0.424537, loss_mean_cls: 0.066967, grad_norm: 0.358409 +Steps: 1%| | 6373/1000000 [26:27<68:15:50, 4.04it/s, grad_norm=0.358, loss_final=0.472, loss_mean=0.829, loss_mean_cls=0.067, proj_loss=-0.425][2026-03-26 12:57:30] Step: 6373, Training Logs: loss_final: 0.470834, loss_mean: 0.804499, proj_loss: -0.405157, loss_mean_cls: 0.071492, grad_norm: 0.462504 +Steps: 1%| | 6374/1000000 [26:27<68:16:47, 4.04it/s, grad_norm=0.463, loss_final=0.471, loss_mean=0.804, loss_mean_cls=0.0715, proj_loss=-0.405][2026-03-26 12:57:31] Step: 6374, Training Logs: loss_final: 0.475987, loss_mean: 0.826328, proj_loss: -0.418330, loss_mean_cls: 0.067990, grad_norm: 0.294454 +Steps: 1%| | 6375/1000000 [26:27<68:14:33, 4.04it/s, grad_norm=0.294, loss_final=0.476, loss_mean=0.826, loss_mean_cls=0.068, proj_loss=-0.418][2026-03-26 12:57:31] Step: 6375, Training Logs: loss_final: 0.473524, loss_mean: 0.832420, proj_loss: -0.424656, loss_mean_cls: 0.065761, grad_norm: 0.562534 +Steps: 1%| | 6376/1000000 [26:28<68:16:14, 4.04it/s, grad_norm=0.563, loss_final=0.474, loss_mean=0.832, loss_mean_cls=0.0658, proj_loss=-0.425][2026-03-26 12:57:31] Step: 6376, Training Logs: loss_final: 0.471383, loss_mean: 0.823433, proj_loss: -0.420349, loss_mean_cls: 0.068299, grad_norm: 0.307189 +Steps: 1%| | 6377/1000000 [26:28<68:16:51, 4.04it/s, grad_norm=0.307, loss_final=0.471, loss_mean=0.823, loss_mean_cls=0.0683, proj_loss=-0.42][2026-03-26 12:57:31] Step: 6377, Training Logs: loss_final: 0.458664, loss_mean: 0.805367, proj_loss: -0.415628, loss_mean_cls: 0.068925, grad_norm: 0.274026 +Steps: 1%| | 6378/1000000 [26:28<70:35:12, 3.91it/s, grad_norm=0.274, loss_final=0.459, loss_mean=0.805, loss_mean_cls=0.0689, proj_loss=-0.416][2026-03-26 12:57:32] Step: 6378, Training Logs: loss_final: 0.467613, loss_mean: 0.817588, proj_loss: -0.418102, loss_mean_cls: 0.068127, grad_norm: 0.291492 +Steps: 1%| | 6379/1000000 [26:28<69:56:49, 3.95it/s, grad_norm=0.291, loss_final=0.468, loss_mean=0.818, loss_mean_cls=0.0681, proj_loss=-0.418][2026-03-26 12:57:32] Step: 6379, Training Logs: loss_final: 0.481280, loss_mean: 0.826628, proj_loss: -0.414280, loss_mean_cls: 0.068932, grad_norm: 0.300864 +Steps: 1%| | 6380/1000000 [26:29<69:29:35, 3.97it/s, grad_norm=0.301, loss_final=0.481, loss_mean=0.827, loss_mean_cls=0.0689, proj_loss=-0.414][2026-03-26 12:57:32] Step: 6380, Training Logs: loss_final: 0.476531, loss_mean: 0.820033, proj_loss: -0.413010, loss_mean_cls: 0.069508, grad_norm: 0.251007 +Steps: 1%| | 6381/1000000 [26:29<69:07:30, 3.99it/s, grad_norm=0.251, loss_final=0.477, loss_mean=0.82, loss_mean_cls=0.0695, proj_loss=-0.413][2026-03-26 12:57:32] Step: 6381, Training Logs: loss_final: 0.457693, loss_mean: 0.804272, proj_loss: -0.415546, loss_mean_cls: 0.068968, grad_norm: 0.283509 +Steps: 1%| | 6382/1000000 [26:29<68:54:53, 4.01it/s, grad_norm=0.284, loss_final=0.458, loss_mean=0.804, loss_mean_cls=0.069, proj_loss=-0.416][2026-03-26 12:57:33] Step: 6382, Training Logs: loss_final: 0.481465, loss_mean: 0.823143, proj_loss: -0.411291, loss_mean_cls: 0.069612, grad_norm: 0.226234 +Steps: 1%| | 6383/1000000 [26:29<68:41:00, 4.02it/s, grad_norm=0.226, loss_final=0.481, loss_mean=0.823, loss_mean_cls=0.0696, proj_loss=-0.411][2026-03-26 12:57:33] Step: 6383, Training Logs: loss_final: 0.474968, loss_mean: 0.812763, proj_loss: -0.407287, loss_mean_cls: 0.069493, grad_norm: 0.298353 +Steps: 1%| | 6384/1000000 [26:30<68:34:53, 4.02it/s, grad_norm=0.298, loss_final=0.475, loss_mean=0.813, loss_mean_cls=0.0695, proj_loss=-0.407][2026-03-26 12:57:33] Step: 6384, Training Logs: loss_final: 0.460435, loss_mean: 0.804510, proj_loss: -0.413349, loss_mean_cls: 0.069274, grad_norm: 0.220782 +Steps: 1%| | 6385/1000000 [26:30<68:29:33, 4.03it/s, grad_norm=0.221, loss_final=0.46, loss_mean=0.805, loss_mean_cls=0.0693, proj_loss=-0.413][2026-03-26 12:57:33] Step: 6385, Training Logs: loss_final: 0.453440, loss_mean: 0.808178, proj_loss: -0.422589, loss_mean_cls: 0.067850, grad_norm: 0.208403 +Steps: 1%| | 6386/1000000 [26:30<68:26:48, 4.03it/s, grad_norm=0.208, loss_final=0.453, loss_mean=0.808, loss_mean_cls=0.0679, proj_loss=-0.423][2026-03-26 12:57:34] Step: 6386, Training Logs: loss_final: 0.479322, loss_mean: 0.830480, proj_loss: -0.419736, loss_mean_cls: 0.068579, grad_norm: 0.260576 +Steps: 1%| | 6387/1000000 [26:30<68:24:12, 4.03it/s, grad_norm=0.261, loss_final=0.479, loss_mean=0.83, loss_mean_cls=0.0686, proj_loss=-0.42][2026-03-26 12:57:34] Step: 6387, Training Logs: loss_final: 0.479451, loss_mean: 0.830781, proj_loss: -0.418909, loss_mean_cls: 0.067578, grad_norm: 0.239095 +Steps: 1%| | 6388/1000000 [26:31<68:25:10, 4.03it/s, grad_norm=0.239, loss_final=0.479, loss_mean=0.831, loss_mean_cls=0.0676, proj_loss=-0.419][2026-03-26 12:57:34] Step: 6388, Training Logs: loss_final: 0.462938, loss_mean: 0.793745, proj_loss: -0.402924, loss_mean_cls: 0.072118, grad_norm: 0.357997 +Steps: 1%| | 6389/1000000 [26:31<68:23:55, 4.04it/s, grad_norm=0.358, loss_final=0.463, loss_mean=0.794, loss_mean_cls=0.0721, proj_loss=-0.403][2026-03-26 12:57:34] Step: 6389, Training Logs: loss_final: 0.502016, loss_mean: 0.850890, proj_loss: -0.417451, loss_mean_cls: 0.068577, grad_norm: 0.315322 +Steps: 1%| | 6390/1000000 [26:31<68:23:43, 4.04it/s, grad_norm=0.315, loss_final=0.502, loss_mean=0.851, loss_mean_cls=0.0686, proj_loss=-0.417][2026-03-26 12:57:35] Step: 6390, Training Logs: loss_final: 0.470887, loss_mean: 0.808728, proj_loss: -0.407582, loss_mean_cls: 0.069741, grad_norm: 0.324827 +Steps: 1%| | 6391/1000000 [26:31<68:36:18, 4.02it/s, grad_norm=0.325, loss_final=0.471, loss_mean=0.809, loss_mean_cls=0.0697, proj_loss=-0.408][2026-03-26 12:57:35] Step: 6391, Training Logs: loss_final: 0.477508, loss_mean: 0.827334, proj_loss: -0.417540, loss_mean_cls: 0.067714, grad_norm: 0.322107 +Steps: 1%| | 6392/1000000 [26:32<68:29:21, 4.03it/s, grad_norm=0.322, loss_final=0.478, loss_mean=0.827, loss_mean_cls=0.0677, proj_loss=-0.418][2026-03-26 12:57:35] Step: 6392, Training Logs: loss_final: 0.473692, loss_mean: 0.827609, proj_loss: -0.421847, loss_mean_cls: 0.067931, grad_norm: 0.596171 +Steps: 1%| | 6393/1000000 [26:32<68:24:46, 4.03it/s, grad_norm=0.596, loss_final=0.474, loss_mean=0.828, loss_mean_cls=0.0679, proj_loss=-0.422][2026-03-26 12:57:35] Step: 6393, Training Logs: loss_final: 0.460674, loss_mean: 0.812043, proj_loss: -0.420186, loss_mean_cls: 0.068817, grad_norm: 0.373492 +Steps: 1%| | 6394/1000000 [26:32<68:22:42, 4.04it/s, grad_norm=0.373, loss_final=0.461, loss_mean=0.812, loss_mean_cls=0.0688, proj_loss=-0.42][2026-03-26 12:57:36] Step: 6394, Training Logs: loss_final: 0.493349, loss_mean: 0.832861, proj_loss: -0.409250, loss_mean_cls: 0.069738, grad_norm: 0.442418 +Steps: 1%| | 6395/1000000 [26:32<68:22:32, 4.04it/s, grad_norm=0.442, loss_final=0.493, loss_mean=0.833, loss_mean_cls=0.0697, proj_loss=-0.409][2026-03-26 12:57:36] Step: 6395, Training Logs: loss_final: 0.489640, loss_mean: 0.844947, proj_loss: -0.421997, loss_mean_cls: 0.066691, grad_norm: 0.340488 +Steps: 1%| | 6396/1000000 [26:33<68:22:43, 4.04it/s, grad_norm=0.34, loss_final=0.49, loss_mean=0.845, loss_mean_cls=0.0667, proj_loss=-0.422][2026-03-26 12:57:36] Step: 6396, Training Logs: loss_final: 0.464092, loss_mean: 0.806208, proj_loss: -0.412435, loss_mean_cls: 0.070319, grad_norm: 0.360259 +Steps: 1%| | 6397/1000000 [26:33<68:22:25, 4.04it/s, grad_norm=0.36, loss_final=0.464, loss_mean=0.806, loss_mean_cls=0.0703, proj_loss=-0.412][2026-03-26 12:57:36] Step: 6397, Training Logs: loss_final: 0.474043, loss_mean: 0.822303, proj_loss: -0.417034, loss_mean_cls: 0.068775, grad_norm: 0.396031 +Steps: 1%| | 6398/1000000 [26:33<68:23:57, 4.04it/s, grad_norm=0.396, loss_final=0.474, loss_mean=0.822, loss_mean_cls=0.0688, proj_loss=-0.417][2026-03-26 12:57:37] Step: 6398, Training Logs: loss_final: 0.464120, loss_mean: 0.809661, proj_loss: -0.414822, loss_mean_cls: 0.069282, grad_norm: 0.320602 +Steps: 1%| | 6399/1000000 [26:33<68:26:22, 4.03it/s, grad_norm=0.321, loss_final=0.464, loss_mean=0.81, loss_mean_cls=0.0693, proj_loss=-0.415][2026-03-26 12:57:37] Step: 6399, Training Logs: loss_final: 0.471368, loss_mean: 0.812173, proj_loss: -0.410910, loss_mean_cls: 0.070106, grad_norm: 0.300061 +Steps: 1%| | 6400/1000000 [26:34<68:25:20, 4.03it/s, grad_norm=0.3, loss_final=0.471, loss_mean=0.812, loss_mean_cls=0.0701, proj_loss=-0.411][2026-03-26 12:57:37] Step: 6400, Training Logs: loss_final: 0.494469, loss_mean: 0.830770, proj_loss: -0.407676, loss_mean_cls: 0.071374, grad_norm: 0.509705 +Steps: 1%| | 6401/1000000 [26:34<68:37:40, 4.02it/s, grad_norm=0.51, loss_final=0.494, loss_mean=0.831, loss_mean_cls=0.0714, proj_loss=-0.408][2026-03-26 12:57:37] Step: 6401, Training Logs: loss_final: 0.475828, loss_mean: 0.813765, proj_loss: -0.407767, loss_mean_cls: 0.069829, grad_norm: 0.322984 +Steps: 1%| | 6402/1000000 [26:34<68:30:32, 4.03it/s, grad_norm=0.323, loss_final=0.476, loss_mean=0.814, loss_mean_cls=0.0698, proj_loss=-0.408][2026-03-26 12:57:37] Step: 6402, Training Logs: loss_final: 0.467247, loss_mean: 0.820101, proj_loss: -0.420051, loss_mean_cls: 0.067197, grad_norm: 0.476108 +Steps: 1%| | 6403/1000000 [26:34<68:29:38, 4.03it/s, grad_norm=0.476, loss_final=0.467, loss_mean=0.82, loss_mean_cls=0.0672, proj_loss=-0.42][2026-03-26 12:57:38] Step: 6403, Training Logs: loss_final: 0.465581, loss_mean: 0.811698, proj_loss: -0.415157, loss_mean_cls: 0.069041, grad_norm: 0.569312 +Steps: 1%| | 6404/1000000 [26:35<68:22:41, 4.04it/s, grad_norm=0.569, loss_final=0.466, loss_mean=0.812, loss_mean_cls=0.069, proj_loss=-0.415][2026-03-26 12:57:38] Step: 6404, Training Logs: loss_final: 0.477263, loss_mean: 0.816623, proj_loss: -0.409103, loss_mean_cls: 0.069743, grad_norm: 0.383639 +Steps: 1%| | 6405/1000000 [26:35<68:21:58, 4.04it/s, grad_norm=0.384, loss_final=0.477, loss_mean=0.817, loss_mean_cls=0.0697, proj_loss=-0.409][2026-03-26 12:57:38] Step: 6405, Training Logs: loss_final: 0.460724, loss_mean: 0.806567, proj_loss: -0.415463, loss_mean_cls: 0.069620, grad_norm: 0.872056 +Steps: 1%| | 6406/1000000 [26:35<68:18:58, 4.04it/s, grad_norm=0.872, loss_final=0.461, loss_mean=0.807, loss_mean_cls=0.0696, proj_loss=-0.415][2026-03-26 12:57:38] Step: 6406, Training Logs: loss_final: 0.480010, loss_mean: 0.829976, proj_loss: -0.418175, loss_mean_cls: 0.068209, grad_norm: 0.327714 +Steps: 1%| | 6407/1000000 [26:35<68:16:15, 4.04it/s, grad_norm=0.328, loss_final=0.48, loss_mean=0.83, loss_mean_cls=0.0682, proj_loss=-0.418][2026-03-26 12:57:39] Step: 6407, Training Logs: loss_final: 0.490723, loss_mean: 0.861046, proj_loss: -0.434472, loss_mean_cls: 0.064149, grad_norm: 0.963176 +Steps: 1%| | 6408/1000000 [26:36<68:15:42, 4.04it/s, grad_norm=0.963, loss_final=0.491, loss_mean=0.861, loss_mean_cls=0.0641, proj_loss=-0.434][2026-03-26 12:57:39] Step: 6408, Training Logs: loss_final: 0.468708, loss_mean: 0.812566, proj_loss: -0.412187, loss_mean_cls: 0.068329, grad_norm: 0.524021 +Steps: 1%| | 6409/1000000 [26:36<68:15:52, 4.04it/s, grad_norm=0.524, loss_final=0.469, loss_mean=0.813, loss_mean_cls=0.0683, proj_loss=-0.412][2026-03-26 12:57:39] Step: 6409, Training Logs: loss_final: 0.484187, loss_mean: 0.830810, proj_loss: -0.415908, loss_mean_cls: 0.069285, grad_norm: 0.519223 +Steps: 1%| | 6410/1000000 [26:36<68:19:44, 4.04it/s, grad_norm=0.519, loss_final=0.484, loss_mean=0.831, loss_mean_cls=0.0693, proj_loss=-0.416][2026-03-26 12:57:39] Step: 6410, Training Logs: loss_final: 0.483054, loss_mean: 0.838747, proj_loss: -0.422879, loss_mean_cls: 0.067187, grad_norm: 0.577389 +Steps: 1%| | 6411/1000000 [26:36<68:18:15, 4.04it/s, grad_norm=0.577, loss_final=0.483, loss_mean=0.839, loss_mean_cls=0.0672, proj_loss=-0.423][2026-03-26 12:57:40] Step: 6411, Training Logs: loss_final: 0.477169, loss_mean: 0.825760, proj_loss: -0.416754, loss_mean_cls: 0.068163, grad_norm: 0.331155 +Steps: 1%| | 6412/1000000 [26:37<68:19:52, 4.04it/s, grad_norm=0.331, loss_final=0.477, loss_mean=0.826, loss_mean_cls=0.0682, proj_loss=-0.417][2026-03-26 12:57:40] Step: 6412, Training Logs: loss_final: 0.463261, loss_mean: 0.809635, proj_loss: -0.414486, loss_mean_cls: 0.068112, grad_norm: 0.535298 +Steps: 1%| | 6413/1000000 [26:37<68:20:34, 4.04it/s, grad_norm=0.535, loss_final=0.463, loss_mean=0.81, loss_mean_cls=0.0681, proj_loss=-0.414][2026-03-26 12:57:40] Step: 6413, Training Logs: loss_final: 0.467397, loss_mean: 0.810983, proj_loss: -0.412211, loss_mean_cls: 0.068625, grad_norm: 0.355681 +Steps: 1%| | 6414/1000000 [26:37<68:20:09, 4.04it/s, grad_norm=0.356, loss_final=0.467, loss_mean=0.811, loss_mean_cls=0.0686, proj_loss=-0.412][2026-03-26 12:57:40] Step: 6414, Training Logs: loss_final: 0.474402, loss_mean: 0.821865, proj_loss: -0.416649, loss_mean_cls: 0.069186, grad_norm: 0.586078 +Steps: 1%| | 6415/1000000 [26:37<68:19:13, 4.04it/s, grad_norm=0.586, loss_final=0.474, loss_mean=0.822, loss_mean_cls=0.0692, proj_loss=-0.417][2026-03-26 12:57:41] Step: 6415, Training Logs: loss_final: 0.483369, loss_mean: 0.829947, proj_loss: -0.415189, loss_mean_cls: 0.068611, grad_norm: 0.367711 +Steps: 1%| | 6416/1000000 [26:38<68:18:45, 4.04it/s, grad_norm=0.368, loss_final=0.483, loss_mean=0.83, loss_mean_cls=0.0686, proj_loss=-0.415][2026-03-26 12:57:41] Step: 6416, Training Logs: loss_final: 0.478377, loss_mean: 0.829539, proj_loss: -0.419171, loss_mean_cls: 0.068009, grad_norm: 0.400298 +Steps: 1%| | 6417/1000000 [26:38<68:18:35, 4.04it/s, grad_norm=0.4, loss_final=0.478, loss_mean=0.83, loss_mean_cls=0.068, proj_loss=-0.419][2026-03-26 12:57:41] Step: 6417, Training Logs: loss_final: 0.483530, loss_mean: 0.832810, proj_loss: -0.417804, loss_mean_cls: 0.068525, grad_norm: 0.412407 +Steps: 1%| | 6418/1000000 [26:38<68:21:05, 4.04it/s, grad_norm=0.412, loss_final=0.484, loss_mean=0.833, loss_mean_cls=0.0685, proj_loss=-0.418][2026-03-26 12:57:41] Step: 6418, Training Logs: loss_final: 0.477072, loss_mean: 0.830797, proj_loss: -0.421672, loss_mean_cls: 0.067947, grad_norm: 0.327180 +Steps: 1%| | 6419/1000000 [26:38<68:19:59, 4.04it/s, grad_norm=0.327, loss_final=0.477, loss_mean=0.831, loss_mean_cls=0.0679, proj_loss=-0.422][2026-03-26 12:57:42] Step: 6419, Training Logs: loss_final: 0.482371, loss_mean: 0.831550, proj_loss: -0.417459, loss_mean_cls: 0.068281, grad_norm: 0.445120 +Steps: 1%| | 6420/1000000 [26:39<68:20:06, 4.04it/s, grad_norm=0.445, loss_final=0.482, loss_mean=0.832, loss_mean_cls=0.0683, proj_loss=-0.417][2026-03-26 12:57:42] Step: 6420, Training Logs: loss_final: 0.463128, loss_mean: 0.813703, proj_loss: -0.418562, loss_mean_cls: 0.067987, grad_norm: 0.485563 +Steps: 1%| | 6421/1000000 [26:39<68:16:12, 4.04it/s, grad_norm=0.486, loss_final=0.463, loss_mean=0.814, loss_mean_cls=0.068, proj_loss=-0.419][2026-03-26 12:57:42] Step: 6421, Training Logs: loss_final: 0.484539, loss_mean: 0.833543, proj_loss: -0.417703, loss_mean_cls: 0.068699, grad_norm: 0.500233 +Steps: 1%| | 6422/1000000 [26:39<68:42:29, 4.02it/s, grad_norm=0.5, loss_final=0.485, loss_mean=0.834, loss_mean_cls=0.0687, proj_loss=-0.418][2026-03-26 12:57:42] Step: 6422, Training Logs: loss_final: 0.469660, loss_mean: 0.816617, proj_loss: -0.415327, loss_mean_cls: 0.068370, grad_norm: 0.392762 +Steps: 1%| | 6423/1000000 [26:39<69:12:45, 3.99it/s, grad_norm=0.393, loss_final=0.47, loss_mean=0.817, loss_mean_cls=0.0684, proj_loss=-0.415][2026-03-26 12:57:43] Step: 6423, Training Logs: loss_final: 0.464927, loss_mean: 0.808019, proj_loss: -0.412026, loss_mean_cls: 0.068934, grad_norm: 0.708581 +Steps: 1%| | 6424/1000000 [26:40<68:58:10, 4.00it/s, grad_norm=0.709, loss_final=0.465, loss_mean=0.808, loss_mean_cls=0.0689, proj_loss=-0.412][2026-03-26 12:57:43] Step: 6424, Training Logs: loss_final: 0.468427, loss_mean: 0.801457, proj_loss: -0.404373, loss_mean_cls: 0.071344, grad_norm: 0.402650 +Steps: 1%| | 6425/1000000 [26:40<68:44:01, 4.02it/s, grad_norm=0.403, loss_final=0.468, loss_mean=0.801, loss_mean_cls=0.0713, proj_loss=-0.404][2026-03-26 12:57:43] Step: 6425, Training Logs: loss_final: 0.470991, loss_mean: 0.816045, proj_loss: -0.414636, loss_mean_cls: 0.069583, grad_norm: 0.590638 +Steps: 1%| | 6426/1000000 [26:40<68:38:58, 4.02it/s, grad_norm=0.591, loss_final=0.471, loss_mean=0.816, loss_mean_cls=0.0696, proj_loss=-0.415][2026-03-26 12:57:43] Step: 6426, Training Logs: loss_final: 0.492186, loss_mean: 0.842532, proj_loss: -0.418165, loss_mean_cls: 0.067819, grad_norm: 0.570256 +Steps: 1%| | 6427/1000000 [26:40<68:31:26, 4.03it/s, grad_norm=0.57, loss_final=0.492, loss_mean=0.843, loss_mean_cls=0.0678, proj_loss=-0.418][2026-03-26 12:57:44] Step: 6427, Training Logs: loss_final: 0.471227, loss_mean: 0.817581, proj_loss: -0.415788, loss_mean_cls: 0.069434, grad_norm: 0.329082 +Steps: 1%| | 6428/1000000 [26:41<68:28:23, 4.03it/s, grad_norm=0.329, loss_final=0.471, loss_mean=0.818, loss_mean_cls=0.0694, proj_loss=-0.416][2026-03-26 12:57:44] Step: 6428, Training Logs: loss_final: 0.480256, loss_mean: 0.821136, proj_loss: -0.410788, loss_mean_cls: 0.069908, grad_norm: 0.419213 +Steps: 1%| | 6429/1000000 [26:41<68:23:18, 4.04it/s, grad_norm=0.419, loss_final=0.48, loss_mean=0.821, loss_mean_cls=0.0699, proj_loss=-0.411][2026-03-26 12:57:44] Step: 6429, Training Logs: loss_final: 0.483448, loss_mean: 0.824654, proj_loss: -0.411313, loss_mean_cls: 0.070106, grad_norm: 0.476817 +Steps: 1%| | 6430/1000000 [26:41<68:22:02, 4.04it/s, grad_norm=0.477, loss_final=0.483, loss_mean=0.825, loss_mean_cls=0.0701, proj_loss=-0.411][2026-03-26 12:57:44] Step: 6430, Training Logs: loss_final: 0.484132, loss_mean: 0.830203, proj_loss: -0.413765, loss_mean_cls: 0.067694, grad_norm: 0.715847 +Steps: 1%| | 6431/1000000 [26:41<68:18:27, 4.04it/s, grad_norm=0.716, loss_final=0.484, loss_mean=0.83, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 12:57:45] Step: 6431, Training Logs: loss_final: 0.466749, loss_mean: 0.829985, proj_loss: -0.427905, loss_mean_cls: 0.064669, grad_norm: 0.336338 +Steps: 1%| | 6432/1000000 [26:42<68:17:29, 4.04it/s, grad_norm=0.336, loss_final=0.467, loss_mean=0.83, loss_mean_cls=0.0647, proj_loss=-0.428][2026-03-26 12:57:45] Step: 6432, Training Logs: loss_final: 0.462233, loss_mean: 0.809391, proj_loss: -0.414863, loss_mean_cls: 0.067705, grad_norm: 0.583437 +Steps: 1%| | 6433/1000000 [26:42<68:21:41, 4.04it/s, grad_norm=0.583, loss_final=0.462, loss_mean=0.809, loss_mean_cls=0.0677, proj_loss=-0.415][2026-03-26 12:57:45] Step: 6433, Training Logs: loss_final: 0.457452, loss_mean: 0.810941, proj_loss: -0.420232, loss_mean_cls: 0.066743, grad_norm: 0.268663 +Steps: 1%| | 6434/1000000 [26:42<68:21:52, 4.04it/s, grad_norm=0.269, loss_final=0.457, loss_mean=0.811, loss_mean_cls=0.0667, proj_loss=-0.42][2026-03-26 12:57:45] Step: 6434, Training Logs: loss_final: 0.483826, loss_mean: 0.821960, proj_loss: -0.408617, loss_mean_cls: 0.070483, grad_norm: 0.475197 +Steps: 1%| | 6435/1000000 [26:42<68:19:28, 4.04it/s, grad_norm=0.475, loss_final=0.484, loss_mean=0.822, loss_mean_cls=0.0705, proj_loss=-0.409][2026-03-26 12:57:46] Step: 6435, Training Logs: loss_final: 0.471037, loss_mean: 0.819154, proj_loss: -0.417491, loss_mean_cls: 0.069373, grad_norm: 0.256905 +Steps: 1%| | 6436/1000000 [26:43<68:20:24, 4.04it/s, grad_norm=0.257, loss_final=0.471, loss_mean=0.819, loss_mean_cls=0.0694, proj_loss=-0.417][2026-03-26 12:57:46] Step: 6436, Training Logs: loss_final: 0.489051, loss_mean: 0.852892, proj_loss: -0.428773, loss_mean_cls: 0.064932, grad_norm: 0.439372 +Steps: 1%| | 6437/1000000 [26:43<68:17:34, 4.04it/s, grad_norm=0.439, loss_final=0.489, loss_mean=0.853, loss_mean_cls=0.0649, proj_loss=-0.429][2026-03-26 12:57:46] Step: 6437, Training Logs: loss_final: 0.478322, loss_mean: 0.825813, proj_loss: -0.416293, loss_mean_cls: 0.068802, grad_norm: 0.250269 +Steps: 1%| | 6438/1000000 [26:43<68:18:43, 4.04it/s, grad_norm=0.25, loss_final=0.478, loss_mean=0.826, loss_mean_cls=0.0688, proj_loss=-0.416][2026-03-26 12:57:46] Step: 6438, Training Logs: loss_final: 0.468548, loss_mean: 0.819664, proj_loss: -0.419374, loss_mean_cls: 0.068258, grad_norm: 0.430162 +Steps: 1%| | 6439/1000000 [26:43<68:17:16, 4.04it/s, grad_norm=0.43, loss_final=0.469, loss_mean=0.82, loss_mean_cls=0.0683, proj_loss=-0.419][2026-03-26 12:57:47] Step: 6439, Training Logs: loss_final: 0.466120, loss_mean: 0.823965, proj_loss: -0.423907, loss_mean_cls: 0.066062, grad_norm: 0.412236 +Steps: 1%| | 6440/1000000 [26:44<68:17:39, 4.04it/s, grad_norm=0.412, loss_final=0.466, loss_mean=0.824, loss_mean_cls=0.0661, proj_loss=-0.424][2026-03-26 12:57:47] Step: 6440, Training Logs: loss_final: 0.470551, loss_mean: 0.811070, proj_loss: -0.409795, loss_mean_cls: 0.069277, grad_norm: 0.386363 +Steps: 1%| | 6441/1000000 [26:44<68:20:16, 4.04it/s, grad_norm=0.386, loss_final=0.471, loss_mean=0.811, loss_mean_cls=0.0693, proj_loss=-0.41][2026-03-26 12:57:47] Step: 6441, Training Logs: loss_final: 0.488297, loss_mean: 0.837767, proj_loss: -0.418647, loss_mean_cls: 0.069178, grad_norm: 0.296683 +Steps: 1%| | 6442/1000000 [26:44<68:23:17, 4.04it/s, grad_norm=0.297, loss_final=0.488, loss_mean=0.838, loss_mean_cls=0.0692, proj_loss=-0.419][2026-03-26 12:57:47] Step: 6442, Training Logs: loss_final: 0.475266, loss_mean: 0.817034, proj_loss: -0.412672, loss_mean_cls: 0.070903, grad_norm: 0.534437 +Steps: 1%| | 6443/1000000 [26:44<68:22:48, 4.04it/s, grad_norm=0.534, loss_final=0.475, loss_mean=0.817, loss_mean_cls=0.0709, proj_loss=-0.413][2026-03-26 12:57:48] Step: 6443, Training Logs: loss_final: 0.485021, loss_mean: 0.824670, proj_loss: -0.410196, loss_mean_cls: 0.070548, grad_norm: 0.344356 +Steps: 1%| | 6444/1000000 [26:45<68:19:23, 4.04it/s, grad_norm=0.344, loss_final=0.485, loss_mean=0.825, loss_mean_cls=0.0705, proj_loss=-0.41][2026-03-26 12:57:48] Step: 6444, Training Logs: loss_final: 0.469155, loss_mean: 0.827393, proj_loss: -0.424259, loss_mean_cls: 0.066021, grad_norm: 0.323655 +Steps: 1%| | 6445/1000000 [26:45<68:18:55, 4.04it/s, grad_norm=0.324, loss_final=0.469, loss_mean=0.827, loss_mean_cls=0.066, proj_loss=-0.424][2026-03-26 12:57:48] Step: 6445, Training Logs: loss_final: 0.484913, loss_mean: 0.834882, proj_loss: -0.417758, loss_mean_cls: 0.067789, grad_norm: 0.297121 +Steps: 1%| | 6446/1000000 [26:45<68:19:56, 4.04it/s, grad_norm=0.297, loss_final=0.485, loss_mean=0.835, loss_mean_cls=0.0678, proj_loss=-0.418][2026-03-26 12:57:48] Step: 6446, Training Logs: loss_final: 0.480266, loss_mean: 0.818902, proj_loss: -0.410608, loss_mean_cls: 0.071972, grad_norm: 0.320233 +Steps: 1%| | 6447/1000000 [26:45<68:21:13, 4.04it/s, grad_norm=0.32, loss_final=0.48, loss_mean=0.819, loss_mean_cls=0.072, proj_loss=-0.411][2026-03-26 12:57:49] Step: 6447, Training Logs: loss_final: 0.477256, loss_mean: 0.821583, proj_loss: -0.414605, loss_mean_cls: 0.070278, grad_norm: 0.415847 +Steps: 1%| | 6448/1000000 [26:46<68:18:20, 4.04it/s, grad_norm=0.416, loss_final=0.477, loss_mean=0.822, loss_mean_cls=0.0703, proj_loss=-0.415][2026-03-26 12:57:49] Step: 6448, Training Logs: loss_final: 0.490882, loss_mean: 0.827094, proj_loss: -0.406817, loss_mean_cls: 0.070604, grad_norm: 0.413353 +Steps: 1%| | 6449/1000000 [26:46<68:18:00, 4.04it/s, grad_norm=0.413, loss_final=0.491, loss_mean=0.827, loss_mean_cls=0.0706, proj_loss=-0.407][2026-03-26 12:57:49] Step: 6449, Training Logs: loss_final: 0.474826, loss_mean: 0.819614, proj_loss: -0.413964, loss_mean_cls: 0.069176, grad_norm: 0.578783 +Steps: 1%| | 6450/1000000 [26:46<68:17:03, 4.04it/s, grad_norm=0.579, loss_final=0.475, loss_mean=0.82, loss_mean_cls=0.0692, proj_loss=-0.414][2026-03-26 12:57:49] Step: 6450, Training Logs: loss_final: 0.485223, loss_mean: 0.829301, proj_loss: -0.412339, loss_mean_cls: 0.068261, grad_norm: 0.441438 +Steps: 1%| | 6451/1000000 [26:46<68:18:50, 4.04it/s, grad_norm=0.441, loss_final=0.485, loss_mean=0.829, loss_mean_cls=0.0683, proj_loss=-0.412][2026-03-26 12:57:50] Step: 6451, Training Logs: loss_final: 0.474594, loss_mean: 0.822364, proj_loss: -0.417209, loss_mean_cls: 0.069440, grad_norm: 0.342969 +Steps: 1%| | 6452/1000000 [26:46<68:17:42, 4.04it/s, grad_norm=0.343, loss_final=0.475, loss_mean=0.822, loss_mean_cls=0.0694, proj_loss=-0.417][2026-03-26 12:57:50] Step: 6452, Training Logs: loss_final: 0.475863, loss_mean: 0.842109, proj_loss: -0.431643, loss_mean_cls: 0.065396, grad_norm: 0.360293 +Steps: 1%| | 6453/1000000 [26:47<68:17:13, 4.04it/s, grad_norm=0.36, loss_final=0.476, loss_mean=0.842, loss_mean_cls=0.0654, proj_loss=-0.432][2026-03-26 12:57:50] Step: 6453, Training Logs: loss_final: 0.485486, loss_mean: 0.849636, proj_loss: -0.429901, loss_mean_cls: 0.065751, grad_norm: 0.404070 +Steps: 1%| | 6454/1000000 [26:47<68:16:36, 4.04it/s, grad_norm=0.404, loss_final=0.485, loss_mean=0.85, loss_mean_cls=0.0658, proj_loss=-0.43][2026-03-26 12:57:50] Step: 6454, Training Logs: loss_final: 0.462788, loss_mean: 0.805888, proj_loss: -0.413049, loss_mean_cls: 0.069949, grad_norm: 0.628342 +Steps: 1%| | 6455/1000000 [26:47<68:18:13, 4.04it/s, grad_norm=0.628, loss_final=0.463, loss_mean=0.806, loss_mean_cls=0.0699, proj_loss=-0.413][2026-03-26 12:57:51] Step: 6455, Training Logs: loss_final: 0.482841, loss_mean: 0.826805, proj_loss: -0.413384, loss_mean_cls: 0.069420, grad_norm: 0.250186 +Steps: 1%| | 6456/1000000 [26:47<68:18:06, 4.04it/s, grad_norm=0.25, loss_final=0.483, loss_mean=0.827, loss_mean_cls=0.0694, proj_loss=-0.413][2026-03-26 12:57:51] Step: 6456, Training Logs: loss_final: 0.464647, loss_mean: 0.817132, proj_loss: -0.419763, loss_mean_cls: 0.067278, grad_norm: 0.460883 +Steps: 1%| | 6457/1000000 [26:48<68:18:55, 4.04it/s, grad_norm=0.461, loss_final=0.465, loss_mean=0.817, loss_mean_cls=0.0673, proj_loss=-0.42][2026-03-26 12:57:51] Step: 6457, Training Logs: loss_final: 0.469661, loss_mean: 0.809080, proj_loss: -0.409029, loss_mean_cls: 0.069611, grad_norm: 0.360156 +Steps: 1%| | 6458/1000000 [26:48<68:21:27, 4.04it/s, grad_norm=0.36, loss_final=0.47, loss_mean=0.809, loss_mean_cls=0.0696, proj_loss=-0.409][2026-03-26 12:57:51] Step: 6458, Training Logs: loss_final: 0.483227, loss_mean: 0.831727, proj_loss: -0.416557, loss_mean_cls: 0.068056, grad_norm: 0.269607 +Steps: 1%| | 6459/1000000 [26:48<68:47:48, 4.01it/s, grad_norm=0.27, loss_final=0.483, loss_mean=0.832, loss_mean_cls=0.0681, proj_loss=-0.417][2026-03-26 12:57:52] Step: 6459, Training Logs: loss_final: 0.463532, loss_mean: 0.806409, proj_loss: -0.413458, loss_mean_cls: 0.070582, grad_norm: 0.394515 +Steps: 1%| | 6460/1000000 [26:48<68:26:05, 4.03it/s, grad_norm=0.395, loss_final=0.464, loss_mean=0.806, loss_mean_cls=0.0706, proj_loss=-0.413][2026-03-26 12:57:52] Step: 6460, Training Logs: loss_final: 0.481386, loss_mean: 0.834751, proj_loss: -0.420585, loss_mean_cls: 0.067220, grad_norm: 0.289134 +Steps: 1%| | 6461/1000000 [26:49<68:23:17, 4.04it/s, grad_norm=0.289, loss_final=0.481, loss_mean=0.835, loss_mean_cls=0.0672, proj_loss=-0.421][2026-03-26 12:57:52] Step: 6461, Training Logs: loss_final: 0.468949, loss_mean: 0.826582, proj_loss: -0.424062, loss_mean_cls: 0.066429, grad_norm: 0.291499 +Steps: 1%| | 6462/1000000 [26:49<68:21:21, 4.04it/s, grad_norm=0.291, loss_final=0.469, loss_mean=0.827, loss_mean_cls=0.0664, proj_loss=-0.424][2026-03-26 12:57:52] Step: 6462, Training Logs: loss_final: 0.473658, loss_mean: 0.820134, proj_loss: -0.415286, loss_mean_cls: 0.068810, grad_norm: 0.292960 +Steps: 1%| | 6463/1000000 [26:49<68:19:39, 4.04it/s, grad_norm=0.293, loss_final=0.474, loss_mean=0.82, loss_mean_cls=0.0688, proj_loss=-0.415][2026-03-26 12:57:53] Step: 6463, Training Logs: loss_final: 0.479167, loss_mean: 0.825425, proj_loss: -0.415263, loss_mean_cls: 0.069005, grad_norm: 0.364984 +Steps: 1%| | 6464/1000000 [26:49<68:19:48, 4.04it/s, grad_norm=0.365, loss_final=0.479, loss_mean=0.825, loss_mean_cls=0.069, proj_loss=-0.415][2026-03-26 12:57:53] Step: 6464, Training Logs: loss_final: 0.470163, loss_mean: 0.814802, proj_loss: -0.412389, loss_mean_cls: 0.067750, grad_norm: 0.341313 +Steps: 1%| | 6465/1000000 [26:50<68:18:20, 4.04it/s, grad_norm=0.341, loss_final=0.47, loss_mean=0.815, loss_mean_cls=0.0677, proj_loss=-0.412][2026-03-26 12:57:53] Step: 6465, Training Logs: loss_final: 0.485534, loss_mean: 0.819967, proj_loss: -0.404004, loss_mean_cls: 0.069571, grad_norm: 0.451402 +Steps: 1%| | 6466/1000000 [26:50<68:17:53, 4.04it/s, grad_norm=0.451, loss_final=0.486, loss_mean=0.82, loss_mean_cls=0.0696, proj_loss=-0.404][2026-03-26 12:57:53] Step: 6466, Training Logs: loss_final: 0.486236, loss_mean: 0.834255, proj_loss: -0.417609, loss_mean_cls: 0.069590, grad_norm: 0.508843 +Steps: 1%| | 6467/1000000 [26:50<68:19:24, 4.04it/s, grad_norm=0.509, loss_final=0.486, loss_mean=0.834, loss_mean_cls=0.0696, proj_loss=-0.418][2026-03-26 12:57:54] Step: 6467, Training Logs: loss_final: 0.487415, loss_mean: 0.821133, proj_loss: -0.403514, loss_mean_cls: 0.069796, grad_norm: 0.246366 +Steps: 1%| | 6468/1000000 [26:50<68:19:26, 4.04it/s, grad_norm=0.246, loss_final=0.487, loss_mean=0.821, loss_mean_cls=0.0698, proj_loss=-0.404][2026-03-26 12:57:54] Step: 6468, Training Logs: loss_final: 0.486654, loss_mean: 0.828284, proj_loss: -0.411914, loss_mean_cls: 0.070285, grad_norm: 0.515950 +Steps: 1%| | 6469/1000000 [26:51<68:16:08, 4.04it/s, grad_norm=0.516, loss_final=0.487, loss_mean=0.828, loss_mean_cls=0.0703, proj_loss=-0.412][2026-03-26 12:57:54] Step: 6469, Training Logs: loss_final: 0.464242, loss_mean: 0.810544, proj_loss: -0.415816, loss_mean_cls: 0.069514, grad_norm: 0.302551 +Steps: 1%| | 6470/1000000 [26:51<68:15:43, 4.04it/s, grad_norm=0.303, loss_final=0.464, loss_mean=0.811, loss_mean_cls=0.0695, proj_loss=-0.416][2026-03-26 12:57:54] Step: 6470, Training Logs: loss_final: 0.472490, loss_mean: 0.823424, proj_loss: -0.419631, loss_mean_cls: 0.068697, grad_norm: 0.433904 +Steps: 1%| | 6471/1000000 [26:51<68:13:59, 4.04it/s, grad_norm=0.434, loss_final=0.472, loss_mean=0.823, loss_mean_cls=0.0687, proj_loss=-0.42][2026-03-26 12:57:55] Step: 6471, Training Logs: loss_final: 0.472663, loss_mean: 0.803210, proj_loss: -0.402323, loss_mean_cls: 0.071776, grad_norm: 0.294656 +Steps: 1%| | 6472/1000000 [26:51<68:17:39, 4.04it/s, grad_norm=0.295, loss_final=0.473, loss_mean=0.803, loss_mean_cls=0.0718, proj_loss=-0.402][2026-03-26 12:57:55] Step: 6472, Training Logs: loss_final: 0.475838, loss_mean: 0.832240, proj_loss: -0.422772, loss_mean_cls: 0.066370, grad_norm: 0.421063 +Steps: 1%| | 6473/1000000 [26:52<68:14:26, 4.04it/s, grad_norm=0.421, loss_final=0.476, loss_mean=0.832, loss_mean_cls=0.0664, proj_loss=-0.423][2026-03-26 12:57:55] Step: 6473, Training Logs: loss_final: 0.463172, loss_mean: 0.817435, proj_loss: -0.420761, loss_mean_cls: 0.066499, grad_norm: 0.395111 +Steps: 1%| | 6474/1000000 [26:52<68:15:50, 4.04it/s, grad_norm=0.395, loss_final=0.463, loss_mean=0.817, loss_mean_cls=0.0665, proj_loss=-0.421][2026-03-26 12:57:55] Step: 6474, Training Logs: loss_final: 0.477472, loss_mean: 0.822353, proj_loss: -0.413730, loss_mean_cls: 0.068849, grad_norm: 0.265986 +Steps: 1%| | 6475/1000000 [26:52<68:13:18, 4.05it/s, grad_norm=0.266, loss_final=0.477, loss_mean=0.822, loss_mean_cls=0.0688, proj_loss=-0.414][2026-03-26 12:57:56] Step: 6475, Training Logs: loss_final: 0.467025, loss_mean: 0.806131, proj_loss: -0.410113, loss_mean_cls: 0.071007, grad_norm: 0.312934 +Steps: 1%| | 6476/1000000 [26:52<68:15:27, 4.04it/s, grad_norm=0.313, loss_final=0.467, loss_mean=0.806, loss_mean_cls=0.071, proj_loss=-0.41][2026-03-26 12:57:56] Step: 6476, Training Logs: loss_final: 0.471074, loss_mean: 0.829061, proj_loss: -0.424109, loss_mean_cls: 0.066122, grad_norm: 0.201717 +Steps: 1%| | 6477/1000000 [26:53<68:14:20, 4.04it/s, grad_norm=0.202, loss_final=0.471, loss_mean=0.829, loss_mean_cls=0.0661, proj_loss=-0.424][2026-03-26 12:57:56] Step: 6477, Training Logs: loss_final: 0.469171, loss_mean: 0.819369, proj_loss: -0.418382, loss_mean_cls: 0.068184, grad_norm: 0.283523 +Steps: 1%| | 6478/1000000 [26:53<68:17:40, 4.04it/s, grad_norm=0.284, loss_final=0.469, loss_mean=0.819, loss_mean_cls=0.0682, proj_loss=-0.418][2026-03-26 12:57:56] Step: 6478, Training Logs: loss_final: 0.487700, loss_mean: 0.823694, proj_loss: -0.406854, loss_mean_cls: 0.070860, grad_norm: 0.292843 +Steps: 1%| | 6479/1000000 [26:53<68:15:58, 4.04it/s, grad_norm=0.293, loss_final=0.488, loss_mean=0.824, loss_mean_cls=0.0709, proj_loss=-0.407][2026-03-26 12:57:57] Step: 6479, Training Logs: loss_final: 0.482553, loss_mean: 0.831594, proj_loss: -0.416951, loss_mean_cls: 0.067911, grad_norm: 0.248173 +Steps: 1%| | 6480/1000000 [26:53<68:18:12, 4.04it/s, grad_norm=0.248, loss_final=0.483, loss_mean=0.832, loss_mean_cls=0.0679, proj_loss=-0.417][2026-03-26 12:57:57] Step: 6480, Training Logs: loss_final: 0.465552, loss_mean: 0.803644, proj_loss: -0.408049, loss_mean_cls: 0.069957, grad_norm: 0.303887 +Steps: 1%| | 6481/1000000 [26:54<68:18:42, 4.04it/s, grad_norm=0.304, loss_final=0.466, loss_mean=0.804, loss_mean_cls=0.07, proj_loss=-0.408][2026-03-26 12:57:57] Step: 6481, Training Logs: loss_final: 0.463414, loss_mean: 0.819907, proj_loss: -0.423680, loss_mean_cls: 0.067188, grad_norm: 0.342842 +Steps: 1%| | 6482/1000000 [26:54<68:20:08, 4.04it/s, grad_norm=0.343, loss_final=0.463, loss_mean=0.82, loss_mean_cls=0.0672, proj_loss=-0.424][2026-03-26 12:57:57] Step: 6482, Training Logs: loss_final: 0.474654, loss_mean: 0.822469, proj_loss: -0.416453, loss_mean_cls: 0.068637, grad_norm: 0.255955 +Steps: 1%| | 6483/1000000 [26:54<68:20:17, 4.04it/s, grad_norm=0.256, loss_final=0.475, loss_mean=0.822, loss_mean_cls=0.0686, proj_loss=-0.416][2026-03-26 12:57:58] Step: 6483, Training Logs: loss_final: 0.491140, loss_mean: 0.839315, proj_loss: -0.417187, loss_mean_cls: 0.069012, grad_norm: 0.381437 +Steps: 1%| | 6484/1000000 [26:54<68:20:07, 4.04it/s, grad_norm=0.381, loss_final=0.491, loss_mean=0.839, loss_mean_cls=0.069, proj_loss=-0.417][2026-03-26 12:57:58] Step: 6484, Training Logs: loss_final: 0.477748, loss_mean: 0.818353, proj_loss: -0.409797, loss_mean_cls: 0.069192, grad_norm: 0.264866 +Steps: 1%| | 6485/1000000 [26:55<68:18:53, 4.04it/s, grad_norm=0.265, loss_final=0.478, loss_mean=0.818, loss_mean_cls=0.0692, proj_loss=-0.41][2026-03-26 12:57:58] Step: 6485, Training Logs: loss_final: 0.466698, loss_mean: 0.819044, proj_loss: -0.419931, loss_mean_cls: 0.067585, grad_norm: 0.506929 +Steps: 1%| | 6486/1000000 [26:55<68:19:20, 4.04it/s, grad_norm=0.507, loss_final=0.467, loss_mean=0.819, loss_mean_cls=0.0676, proj_loss=-0.42][2026-03-26 12:57:58] Step: 6486, Training Logs: loss_final: 0.462579, loss_mean: 0.813548, proj_loss: -0.418646, loss_mean_cls: 0.067677, grad_norm: 0.376697 +Steps: 1%| | 6487/1000000 [26:55<69:33:08, 3.97it/s, grad_norm=0.377, loss_final=0.463, loss_mean=0.814, loss_mean_cls=0.0677, proj_loss=-0.419][2026-03-26 12:57:59] Step: 6487, Training Logs: loss_final: 0.471454, loss_mean: 0.818186, proj_loss: -0.414457, loss_mean_cls: 0.067725, grad_norm: 0.404770 +Steps: 1%| | 6488/1000000 [26:55<69:09:29, 3.99it/s, grad_norm=0.405, loss_final=0.471, loss_mean=0.818, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 12:57:59] Step: 6488, Training Logs: loss_final: 0.483855, loss_mean: 0.827369, proj_loss: -0.412085, loss_mean_cls: 0.068571, grad_norm: 0.743126 +Steps: 1%| | 6489/1000000 [26:56<68:51:46, 4.01it/s, grad_norm=0.743, loss_final=0.484, loss_mean=0.827, loss_mean_cls=0.0686, proj_loss=-0.412][2026-03-26 12:57:59] Step: 6489, Training Logs: loss_final: 0.467182, loss_mean: 0.818252, proj_loss: -0.419973, loss_mean_cls: 0.068904, grad_norm: 0.409095 +Steps: 1%| | 6490/1000000 [26:56<68:42:08, 4.02it/s, grad_norm=0.409, loss_final=0.467, loss_mean=0.818, loss_mean_cls=0.0689, proj_loss=-0.42][2026-03-26 12:57:59] Step: 6490, Training Logs: loss_final: 0.489310, loss_mean: 0.834608, proj_loss: -0.413742, loss_mean_cls: 0.068444, grad_norm: 0.601327 +Steps: 1%| | 6491/1000000 [26:56<68:35:14, 4.02it/s, grad_norm=0.601, loss_final=0.489, loss_mean=0.835, loss_mean_cls=0.0684, proj_loss=-0.414][2026-03-26 12:58:00] Step: 6491, Training Logs: loss_final: 0.465839, loss_mean: 0.810299, proj_loss: -0.413819, loss_mean_cls: 0.069358, grad_norm: 0.453988 +Steps: 1%| | 6492/1000000 [26:56<68:33:18, 4.03it/s, grad_norm=0.454, loss_final=0.466, loss_mean=0.81, loss_mean_cls=0.0694, proj_loss=-0.414][2026-03-26 12:58:00] Step: 6492, Training Logs: loss_final: 0.488437, loss_mean: 0.825423, proj_loss: -0.407718, loss_mean_cls: 0.070732, grad_norm: 0.560724 +Steps: 1%| | 6493/1000000 [26:57<68:28:22, 4.03it/s, grad_norm=0.561, loss_final=0.488, loss_mean=0.825, loss_mean_cls=0.0707, proj_loss=-0.408][2026-03-26 12:58:00] Step: 6493, Training Logs: loss_final: 0.480737, loss_mean: 0.822786, proj_loss: -0.411232, loss_mean_cls: 0.069183, grad_norm: 0.639503 +Steps: 1%| | 6494/1000000 [26:57<68:23:42, 4.04it/s, grad_norm=0.64, loss_final=0.481, loss_mean=0.823, loss_mean_cls=0.0692, proj_loss=-0.411][2026-03-26 12:58:00] Step: 6494, Training Logs: loss_final: 0.465458, loss_mean: 0.813255, proj_loss: -0.415865, loss_mean_cls: 0.068068, grad_norm: 0.303517 +Steps: 1%| | 6495/1000000 [26:57<68:21:13, 4.04it/s, grad_norm=0.304, loss_final=0.465, loss_mean=0.813, loss_mean_cls=0.0681, proj_loss=-0.416][2026-03-26 12:58:01] Step: 6495, Training Logs: loss_final: 0.477766, loss_mean: 0.823040, proj_loss: -0.414657, loss_mean_cls: 0.069383, grad_norm: 0.713999 +Steps: 1%| | 6496/1000000 [26:57<68:18:51, 4.04it/s, grad_norm=0.714, loss_final=0.478, loss_mean=0.823, loss_mean_cls=0.0694, proj_loss=-0.415][2026-03-26 12:58:01] Step: 6496, Training Logs: loss_final: 0.483755, loss_mean: 0.840290, proj_loss: -0.423485, loss_mean_cls: 0.066949, grad_norm: 0.396697 +Steps: 1%| | 6497/1000000 [26:58<68:20:55, 4.04it/s, grad_norm=0.397, loss_final=0.484, loss_mean=0.84, loss_mean_cls=0.0669, proj_loss=-0.423][2026-03-26 12:58:01] Step: 6497, Training Logs: loss_final: 0.474600, loss_mean: 0.827126, proj_loss: -0.420284, loss_mean_cls: 0.067757, grad_norm: 0.632299 +Steps: 1%| | 6498/1000000 [26:58<68:21:24, 4.04it/s, grad_norm=0.632, loss_final=0.475, loss_mean=0.827, loss_mean_cls=0.0678, proj_loss=-0.42][2026-03-26 12:58:01] Step: 6498, Training Logs: loss_final: 0.473754, loss_mean: 0.828151, proj_loss: -0.421382, loss_mean_cls: 0.066985, grad_norm: 0.299059 +Steps: 1%| | 6499/1000000 [26:58<68:16:25, 4.04it/s, grad_norm=0.299, loss_final=0.474, loss_mean=0.828, loss_mean_cls=0.067, proj_loss=-0.421][2026-03-26 12:58:02] Step: 6499, Training Logs: loss_final: 0.479742, loss_mean: 0.830042, proj_loss: -0.417498, loss_mean_cls: 0.067198, grad_norm: 0.802709 +Steps: 1%| | 6500/1000000 [26:58<68:16:51, 4.04it/s, grad_norm=0.803, loss_final=0.48, loss_mean=0.83, loss_mean_cls=0.0672, proj_loss=-0.417][2026-03-26 12:58:02] Step: 6500, Training Logs: loss_final: 0.465433, loss_mean: 0.805557, proj_loss: -0.411020, loss_mean_cls: 0.070897, grad_norm: 0.401220 +Steps: 1%| | 6501/1000000 [26:59<68:16:33, 4.04it/s, grad_norm=0.401, loss_final=0.465, loss_mean=0.806, loss_mean_cls=0.0709, proj_loss=-0.411][2026-03-26 12:58:02] Step: 6501, Training Logs: loss_final: 0.480544, loss_mean: 0.840318, proj_loss: -0.425701, loss_mean_cls: 0.065927, grad_norm: 0.687629 +Steps: 1%| | 6502/1000000 [26:59<68:14:36, 4.04it/s, grad_norm=0.688, loss_final=0.481, loss_mean=0.84, loss_mean_cls=0.0659, proj_loss=-0.426][2026-03-26 12:58:02] Step: 6502, Training Logs: loss_final: 0.469975, loss_mean: 0.812859, proj_loss: -0.412398, loss_mean_cls: 0.069514, grad_norm: 0.693430 +Steps: 1%| | 6503/1000000 [26:59<68:16:30, 4.04it/s, grad_norm=0.693, loss_final=0.47, loss_mean=0.813, loss_mean_cls=0.0695, proj_loss=-0.412][2026-03-26 12:58:03] Step: 6503, Training Logs: loss_final: 0.479728, loss_mean: 0.828459, proj_loss: -0.417228, loss_mean_cls: 0.068497, grad_norm: 0.456516 +Steps: 1%| | 6504/1000000 [26:59<68:16:24, 4.04it/s, grad_norm=0.457, loss_final=0.48, loss_mean=0.828, loss_mean_cls=0.0685, proj_loss=-0.417][2026-03-26 12:58:03] Step: 6504, Training Logs: loss_final: 0.482692, loss_mean: 0.835929, proj_loss: -0.420754, loss_mean_cls: 0.067517, grad_norm: 0.599624 +Steps: 1%| | 6505/1000000 [27:00<68:19:40, 4.04it/s, grad_norm=0.6, loss_final=0.483, loss_mean=0.836, loss_mean_cls=0.0675, proj_loss=-0.421][2026-03-26 12:58:03] Step: 6505, Training Logs: loss_final: 0.471447, loss_mean: 0.809219, proj_loss: -0.408248, loss_mean_cls: 0.070477, grad_norm: 0.380120 +Steps: 1%| | 6506/1000000 [27:00<68:19:49, 4.04it/s, grad_norm=0.38, loss_final=0.471, loss_mean=0.809, loss_mean_cls=0.0705, proj_loss=-0.408][2026-03-26 12:58:03] Step: 6506, Training Logs: loss_final: 0.488456, loss_mean: 0.844090, proj_loss: -0.422226, loss_mean_cls: 0.066592, grad_norm: 0.630950 +Steps: 1%| | 6507/1000000 [27:00<68:18:24, 4.04it/s, grad_norm=0.631, loss_final=0.488, loss_mean=0.844, loss_mean_cls=0.0666, proj_loss=-0.422][2026-03-26 12:58:04] Step: 6507, Training Logs: loss_final: 0.481827, loss_mean: 0.838318, proj_loss: -0.422633, loss_mean_cls: 0.066142, grad_norm: 0.387283 +Steps: 1%| | 6508/1000000 [27:00<68:21:12, 4.04it/s, grad_norm=0.387, loss_final=0.482, loss_mean=0.838, loss_mean_cls=0.0661, proj_loss=-0.423][2026-03-26 12:58:04] Step: 6508, Training Logs: loss_final: 0.491804, loss_mean: 0.842055, proj_loss: -0.418559, loss_mean_cls: 0.068309, grad_norm: 0.433401 +Steps: 1%| | 6509/1000000 [27:01<68:20:50, 4.04it/s, grad_norm=0.433, loss_final=0.492, loss_mean=0.842, loss_mean_cls=0.0683, proj_loss=-0.419][2026-03-26 12:58:04] Step: 6509, Training Logs: loss_final: 0.484342, loss_mean: 0.831539, proj_loss: -0.415629, loss_mean_cls: 0.068433, grad_norm: 0.271228 +Steps: 1%| | 6510/1000000 [27:01<68:20:03, 4.04it/s, grad_norm=0.271, loss_final=0.484, loss_mean=0.832, loss_mean_cls=0.0684, proj_loss=-0.416][2026-03-26 12:58:04] Step: 6510, Training Logs: loss_final: 0.457342, loss_mean: 0.803175, proj_loss: -0.416015, loss_mean_cls: 0.070181, grad_norm: 0.386721 +Steps: 1%| | 6511/1000000 [27:01<68:19:06, 4.04it/s, grad_norm=0.387, loss_final=0.457, loss_mean=0.803, loss_mean_cls=0.0702, proj_loss=-0.416][2026-03-26 12:58:05] Step: 6511, Training Logs: loss_final: 0.462626, loss_mean: 0.805989, proj_loss: -0.412619, loss_mean_cls: 0.069256, grad_norm: 0.282722 +Steps: 1%| | 6512/1000000 [27:01<68:19:06, 4.04it/s, grad_norm=0.283, loss_final=0.463, loss_mean=0.806, loss_mean_cls=0.0693, proj_loss=-0.413][2026-03-26 12:58:05] Step: 6512, Training Logs: loss_final: 0.461251, loss_mean: 0.810944, proj_loss: -0.417460, loss_mean_cls: 0.067767, grad_norm: 0.267705 +Steps: 1%| | 6513/1000000 [27:02<68:18:59, 4.04it/s, grad_norm=0.268, loss_final=0.461, loss_mean=0.811, loss_mean_cls=0.0678, proj_loss=-0.417][2026-03-26 12:58:05] Step: 6513, Training Logs: loss_final: 0.458432, loss_mean: 0.808488, proj_loss: -0.417833, loss_mean_cls: 0.067777, grad_norm: 0.239704 +Steps: 1%| | 6514/1000000 [27:02<68:18:38, 4.04it/s, grad_norm=0.24, loss_final=0.458, loss_mean=0.808, loss_mean_cls=0.0678, proj_loss=-0.418][2026-03-26 12:58:05] Step: 6514, Training Logs: loss_final: 0.458106, loss_mean: 0.805708, proj_loss: -0.416559, loss_mean_cls: 0.068957, grad_norm: 0.236167 +Steps: 1%| | 6515/1000000 [27:02<68:17:26, 4.04it/s, grad_norm=0.236, loss_final=0.458, loss_mean=0.806, loss_mean_cls=0.069, proj_loss=-0.417][2026-03-26 12:58:05] Step: 6515, Training Logs: loss_final: 0.483145, loss_mean: 0.823773, proj_loss: -0.410558, loss_mean_cls: 0.069929, grad_norm: 0.353983 +Steps: 1%| | 6516/1000000 [27:02<68:18:05, 4.04it/s, grad_norm=0.354, loss_final=0.483, loss_mean=0.824, loss_mean_cls=0.0699, proj_loss=-0.411][2026-03-26 12:58:06] Step: 6516, Training Logs: loss_final: 0.461717, loss_mean: 0.801499, proj_loss: -0.409736, loss_mean_cls: 0.069954, grad_norm: 0.319083 +Steps: 1%| | 6517/1000000 [27:03<68:15:43, 4.04it/s, grad_norm=0.319, loss_final=0.462, loss_mean=0.801, loss_mean_cls=0.07, proj_loss=-0.41][2026-03-26 12:58:06] Step: 6517, Training Logs: loss_final: 0.485006, loss_mean: 0.824679, proj_loss: -0.409573, loss_mean_cls: 0.069899, grad_norm: 0.343112 +Steps: 1%| | 6518/1000000 [27:03<68:16:41, 4.04it/s, grad_norm=0.343, loss_final=0.485, loss_mean=0.825, loss_mean_cls=0.0699, proj_loss=-0.41][2026-03-26 12:58:06] Step: 6518, Training Logs: loss_final: 0.456110, loss_mean: 0.798924, proj_loss: -0.412248, loss_mean_cls: 0.069434, grad_norm: 0.339946 +Steps: 1%| | 6519/1000000 [27:03<68:14:47, 4.04it/s, grad_norm=0.34, loss_final=0.456, loss_mean=0.799, loss_mean_cls=0.0694, proj_loss=-0.412][2026-03-26 12:58:06] Step: 6519, Training Logs: loss_final: 0.468375, loss_mean: 0.809630, proj_loss: -0.411356, loss_mean_cls: 0.070102, grad_norm: 0.264644 +Steps: 1%| | 6520/1000000 [27:03<68:18:43, 4.04it/s, grad_norm=0.265, loss_final=0.468, loss_mean=0.81, loss_mean_cls=0.0701, proj_loss=-0.411][2026-03-26 12:58:07] Step: 6520, Training Logs: loss_final: 0.470496, loss_mean: 0.806292, proj_loss: -0.406300, loss_mean_cls: 0.070504, grad_norm: 0.408539 +Steps: 1%| | 6521/1000000 [27:04<68:17:28, 4.04it/s, grad_norm=0.409, loss_final=0.47, loss_mean=0.806, loss_mean_cls=0.0705, proj_loss=-0.406][2026-03-26 12:58:07] Step: 6521, Training Logs: loss_final: 0.444350, loss_mean: 0.797948, proj_loss: -0.421294, loss_mean_cls: 0.067696, grad_norm: 0.289587 +Steps: 1%| | 6522/1000000 [27:04<68:18:58, 4.04it/s, grad_norm=0.29, loss_final=0.444, loss_mean=0.798, loss_mean_cls=0.0677, proj_loss=-0.421][2026-03-26 12:58:07] Step: 6522, Training Logs: loss_final: 0.488316, loss_mean: 0.831421, proj_loss: -0.411841, loss_mean_cls: 0.068737, grad_norm: 0.792026 +Steps: 1%| | 6523/1000000 [27:04<68:20:11, 4.04it/s, grad_norm=0.792, loss_final=0.488, loss_mean=0.831, loss_mean_cls=0.0687, proj_loss=-0.412][2026-03-26 12:58:07] Step: 6523, Training Logs: loss_final: 0.475638, loss_mean: 0.826285, proj_loss: -0.418606, loss_mean_cls: 0.067960, grad_norm: 0.542940 +Steps: 1%| | 6524/1000000 [27:04<68:18:02, 4.04it/s, grad_norm=0.543, loss_final=0.476, loss_mean=0.826, loss_mean_cls=0.068, proj_loss=-0.419][2026-03-26 12:58:08] Step: 6524, Training Logs: loss_final: 0.475116, loss_mean: 0.815969, proj_loss: -0.411255, loss_mean_cls: 0.070402, grad_norm: 0.293216 +Steps: 1%| | 6525/1000000 [27:05<68:16:35, 4.04it/s, grad_norm=0.293, loss_final=0.475, loss_mean=0.816, loss_mean_cls=0.0704, proj_loss=-0.411][2026-03-26 12:58:08] Step: 6525, Training Logs: loss_final: 0.483258, loss_mean: 0.829242, proj_loss: -0.414906, loss_mean_cls: 0.068921, grad_norm: 0.307028 +Steps: 1%| | 6526/1000000 [27:05<68:19:56, 4.04it/s, grad_norm=0.307, loss_final=0.483, loss_mean=0.829, loss_mean_cls=0.0689, proj_loss=-0.415][2026-03-26 12:58:08] Step: 6526, Training Logs: loss_final: 0.478277, loss_mean: 0.821342, proj_loss: -0.412688, loss_mean_cls: 0.069623, grad_norm: 0.635733 +Steps: 1%| | 6527/1000000 [27:05<68:19:36, 4.04it/s, grad_norm=0.636, loss_final=0.478, loss_mean=0.821, loss_mean_cls=0.0696, proj_loss=-0.413][2026-03-26 12:58:08] Step: 6527, Training Logs: loss_final: 0.477168, loss_mean: 0.828050, proj_loss: -0.419011, loss_mean_cls: 0.068129, grad_norm: 0.476137 +Steps: 1%| | 6528/1000000 [27:05<68:19:00, 4.04it/s, grad_norm=0.476, loss_final=0.477, loss_mean=0.828, loss_mean_cls=0.0681, proj_loss=-0.419][2026-03-26 12:58:09] Step: 6528, Training Logs: loss_final: 0.481117, loss_mean: 0.817381, proj_loss: -0.406514, loss_mean_cls: 0.070250, grad_norm: 0.363291 +Steps: 1%| | 6529/1000000 [27:06<68:18:52, 4.04it/s, grad_norm=0.363, loss_final=0.481, loss_mean=0.817, loss_mean_cls=0.0703, proj_loss=-0.407][2026-03-26 12:58:09] Step: 6529, Training Logs: loss_final: 0.465395, loss_mean: 0.810173, proj_loss: -0.413512, loss_mean_cls: 0.068734, grad_norm: 0.383227 +Steps: 1%| | 6530/1000000 [27:06<68:20:10, 4.04it/s, grad_norm=0.383, loss_final=0.465, loss_mean=0.81, loss_mean_cls=0.0687, proj_loss=-0.414][2026-03-26 12:58:09] Step: 6530, Training Logs: loss_final: 0.483596, loss_mean: 0.829295, proj_loss: -0.414496, loss_mean_cls: 0.068797, grad_norm: 0.334361 +Steps: 1%| | 6531/1000000 [27:06<68:19:55, 4.04it/s, grad_norm=0.334, loss_final=0.484, loss_mean=0.829, loss_mean_cls=0.0688, proj_loss=-0.414][2026-03-26 12:58:09] Step: 6531, Training Logs: loss_final: 0.487384, loss_mean: 0.831552, proj_loss: -0.412790, loss_mean_cls: 0.068622, grad_norm: 0.630527 +Steps: 1%| | 6532/1000000 [27:06<68:19:14, 4.04it/s, grad_norm=0.631, loss_final=0.487, loss_mean=0.832, loss_mean_cls=0.0686, proj_loss=-0.413][2026-03-26 12:58:10] Step: 6532, Training Logs: loss_final: 0.458544, loss_mean: 0.807881, proj_loss: -0.417512, loss_mean_cls: 0.068176, grad_norm: 0.251757 +Steps: 1%| | 6533/1000000 [27:07<68:17:43, 4.04it/s, grad_norm=0.252, loss_final=0.459, loss_mean=0.808, loss_mean_cls=0.0682, proj_loss=-0.418][2026-03-26 12:58:10] Step: 6533, Training Logs: loss_final: 0.478107, loss_mean: 0.826816, proj_loss: -0.416989, loss_mean_cls: 0.068280, grad_norm: 0.542613 +Steps: 1%| | 6534/1000000 [27:07<68:17:32, 4.04it/s, grad_norm=0.543, loss_final=0.478, loss_mean=0.827, loss_mean_cls=0.0683, proj_loss=-0.417][2026-03-26 12:58:10] Step: 6534, Training Logs: loss_final: 0.470597, loss_mean: 0.820318, proj_loss: -0.417324, loss_mean_cls: 0.067603, grad_norm: 0.278048 +Steps: 1%| | 6535/1000000 [27:07<68:18:47, 4.04it/s, grad_norm=0.278, loss_final=0.471, loss_mean=0.82, loss_mean_cls=0.0676, proj_loss=-0.417][2026-03-26 12:58:10] Step: 6535, Training Logs: loss_final: 0.456215, loss_mean: 0.812229, proj_loss: -0.423151, loss_mean_cls: 0.067137, grad_norm: 0.690670 +Steps: 1%| | 6536/1000000 [27:07<68:17:15, 4.04it/s, grad_norm=0.691, loss_final=0.456, loss_mean=0.812, loss_mean_cls=0.0671, proj_loss=-0.423][2026-03-26 12:58:11] Step: 6536, Training Logs: loss_final: 0.470475, loss_mean: 0.820869, proj_loss: -0.418810, loss_mean_cls: 0.068416, grad_norm: 0.404789 +Steps: 1%| | 6537/1000000 [27:08<68:31:52, 4.03it/s, grad_norm=0.405, loss_final=0.47, loss_mean=0.821, loss_mean_cls=0.0684, proj_loss=-0.419][2026-03-26 12:58:11] Step: 6537, Training Logs: loss_final: 0.490990, loss_mean: 0.845253, proj_loss: -0.421033, loss_mean_cls: 0.066770, grad_norm: 0.638145 +Steps: 1%| | 6538/1000000 [27:08<68:24:59, 4.03it/s, grad_norm=0.638, loss_final=0.491, loss_mean=0.845, loss_mean_cls=0.0668, proj_loss=-0.421][2026-03-26 12:58:11] Step: 6538, Training Logs: loss_final: 0.465916, loss_mean: 0.824504, proj_loss: -0.425557, loss_mean_cls: 0.066969, grad_norm: 0.467719 +Steps: 1%| | 6539/1000000 [27:08<68:23:30, 4.04it/s, grad_norm=0.468, loss_final=0.466, loss_mean=0.825, loss_mean_cls=0.067, proj_loss=-0.426][2026-03-26 12:58:11] Step: 6539, Training Logs: loss_final: 0.476382, loss_mean: 0.837847, proj_loss: -0.427603, loss_mean_cls: 0.066137, grad_norm: 0.373105 +Steps: 1%| | 6540/1000000 [27:08<68:21:38, 4.04it/s, grad_norm=0.373, loss_final=0.476, loss_mean=0.838, loss_mean_cls=0.0661, proj_loss=-0.428][2026-03-26 12:58:12] Step: 6540, Training Logs: loss_final: 0.481157, loss_mean: 0.838132, proj_loss: -0.423270, loss_mean_cls: 0.066296, grad_norm: 0.580107 +Steps: 1%| | 6541/1000000 [27:09<68:22:01, 4.04it/s, grad_norm=0.58, loss_final=0.481, loss_mean=0.838, loss_mean_cls=0.0663, proj_loss=-0.423][2026-03-26 12:58:12] Step: 6541, Training Logs: loss_final: 0.476323, loss_mean: 0.831864, proj_loss: -0.422073, loss_mean_cls: 0.066532, grad_norm: 0.348438 +Steps: 1%| | 6542/1000000 [27:09<68:20:35, 4.04it/s, grad_norm=0.348, loss_final=0.476, loss_mean=0.832, loss_mean_cls=0.0665, proj_loss=-0.422][2026-03-26 12:58:12] Step: 6542, Training Logs: loss_final: 0.477524, loss_mean: 0.829878, proj_loss: -0.419948, loss_mean_cls: 0.067593, grad_norm: 0.369138 +Steps: 1%| | 6543/1000000 [27:09<68:19:35, 4.04it/s, grad_norm=0.369, loss_final=0.478, loss_mean=0.83, loss_mean_cls=0.0676, proj_loss=-0.42][2026-03-26 12:58:12] Step: 6543, Training Logs: loss_final: 0.470035, loss_mean: 0.807128, proj_loss: -0.407740, loss_mean_cls: 0.070647, grad_norm: 0.259651 +Steps: 1%| | 6544/1000000 [27:09<68:19:50, 4.04it/s, grad_norm=0.26, loss_final=0.47, loss_mean=0.807, loss_mean_cls=0.0706, proj_loss=-0.408][2026-03-26 12:58:13] Step: 6544, Training Logs: loss_final: 0.470194, loss_mean: 0.817468, proj_loss: -0.416074, loss_mean_cls: 0.068800, grad_norm: 0.391931 +Steps: 1%| | 6545/1000000 [27:10<68:18:04, 4.04it/s, grad_norm=0.392, loss_final=0.47, loss_mean=0.817, loss_mean_cls=0.0688, proj_loss=-0.416][2026-03-26 12:58:13] Step: 6545, Training Logs: loss_final: 0.466941, loss_mean: 0.814969, proj_loss: -0.416674, loss_mean_cls: 0.068646, grad_norm: 0.211384 +Steps: 1%| | 6546/1000000 [27:10<68:17:29, 4.04it/s, grad_norm=0.211, loss_final=0.467, loss_mean=0.815, loss_mean_cls=0.0686, proj_loss=-0.417][2026-03-26 12:58:13] Step: 6546, Training Logs: loss_final: 0.470574, loss_mean: 0.822380, proj_loss: -0.418157, loss_mean_cls: 0.066351, grad_norm: 0.537555 +Steps: 1%| | 6547/1000000 [27:10<68:17:12, 4.04it/s, grad_norm=0.538, loss_final=0.471, loss_mean=0.822, loss_mean_cls=0.0664, proj_loss=-0.418][2026-03-26 12:58:13] Step: 6547, Training Logs: loss_final: 0.473963, loss_mean: 0.807345, proj_loss: -0.403894, loss_mean_cls: 0.070512, grad_norm: 0.485329 +Steps: 1%| | 6548/1000000 [27:10<68:16:14, 4.04it/s, grad_norm=0.485, loss_final=0.474, loss_mean=0.807, loss_mean_cls=0.0705, proj_loss=-0.404][2026-03-26 12:58:14] Step: 6548, Training Logs: loss_final: 0.483127, loss_mean: 0.827422, proj_loss: -0.413709, loss_mean_cls: 0.069415, grad_norm: 0.352191 +Steps: 1%| | 6549/1000000 [27:11<68:16:16, 4.04it/s, grad_norm=0.352, loss_final=0.483, loss_mean=0.827, loss_mean_cls=0.0694, proj_loss=-0.414][2026-03-26 12:58:14] Step: 6549, Training Logs: loss_final: 0.477696, loss_mean: 0.820660, proj_loss: -0.412633, loss_mean_cls: 0.069669, grad_norm: 0.312798 +Steps: 1%| | 6550/1000000 [27:11<68:17:24, 4.04it/s, grad_norm=0.313, loss_final=0.478, loss_mean=0.821, loss_mean_cls=0.0697, proj_loss=-0.413][2026-03-26 12:58:14] Step: 6550, Training Logs: loss_final: 0.476369, loss_mean: 0.813148, proj_loss: -0.407319, loss_mean_cls: 0.070541, grad_norm: 0.415145 +Steps: 1%| | 6551/1000000 [27:11<68:16:39, 4.04it/s, grad_norm=0.415, loss_final=0.476, loss_mean=0.813, loss_mean_cls=0.0705, proj_loss=-0.407][2026-03-26 12:58:14] Step: 6551, Training Logs: loss_final: 0.475185, loss_mean: 0.825760, proj_loss: -0.418025, loss_mean_cls: 0.067451, grad_norm: 0.398446 +Steps: 1%| | 6552/1000000 [27:11<68:17:21, 4.04it/s, grad_norm=0.398, loss_final=0.475, loss_mean=0.826, loss_mean_cls=0.0675, proj_loss=-0.418][2026-03-26 12:58:15] Step: 6552, Training Logs: loss_final: 0.471436, loss_mean: 0.821084, proj_loss: -0.417903, loss_mean_cls: 0.068255, grad_norm: 0.346386 +Steps: 1%| | 6553/1000000 [27:12<68:17:17, 4.04it/s, grad_norm=0.346, loss_final=0.471, loss_mean=0.821, loss_mean_cls=0.0683, proj_loss=-0.418][2026-03-26 12:58:15] Step: 6553, Training Logs: loss_final: 0.479861, loss_mean: 0.823583, proj_loss: -0.412877, loss_mean_cls: 0.069155, grad_norm: 0.554884 +Steps: 1%| | 6554/1000000 [27:12<68:18:58, 4.04it/s, grad_norm=0.555, loss_final=0.48, loss_mean=0.824, loss_mean_cls=0.0692, proj_loss=-0.413][2026-03-26 12:58:15] Step: 6554, Training Logs: loss_final: 0.491719, loss_mean: 0.840969, proj_loss: -0.417519, loss_mean_cls: 0.068270, grad_norm: 0.416668 +Steps: 1%| | 6555/1000000 [27:12<68:16:56, 4.04it/s, grad_norm=0.417, loss_final=0.492, loss_mean=0.841, loss_mean_cls=0.0683, proj_loss=-0.418][2026-03-26 12:58:15] Step: 6555, Training Logs: loss_final: 0.494888, loss_mean: 0.855682, proj_loss: -0.426337, loss_mean_cls: 0.065543, grad_norm: 0.353562 +Steps: 1%| | 6556/1000000 [27:12<68:20:21, 4.04it/s, grad_norm=0.354, loss_final=0.495, loss_mean=0.856, loss_mean_cls=0.0655, proj_loss=-0.426][2026-03-26 12:58:16] Step: 6556, Training Logs: loss_final: 0.470839, loss_mean: 0.812570, proj_loss: -0.411015, loss_mean_cls: 0.069283, grad_norm: 0.262355 +Steps: 1%| | 6557/1000000 [27:12<68:17:11, 4.04it/s, grad_norm=0.262, loss_final=0.471, loss_mean=0.813, loss_mean_cls=0.0693, proj_loss=-0.411][2026-03-26 12:58:16] Step: 6557, Training Logs: loss_final: 0.484912, loss_mean: 0.829729, proj_loss: -0.413176, loss_mean_cls: 0.068359, grad_norm: 0.331653 +Steps: 1%| | 6558/1000000 [27:13<68:18:28, 4.04it/s, grad_norm=0.332, loss_final=0.485, loss_mean=0.83, loss_mean_cls=0.0684, proj_loss=-0.413][2026-03-26 12:58:16] Step: 6558, Training Logs: loss_final: 0.470938, loss_mean: 0.836070, proj_loss: -0.429321, loss_mean_cls: 0.064188, grad_norm: 0.255342 +Steps: 1%| | 6559/1000000 [27:13<68:21:20, 4.04it/s, grad_norm=0.255, loss_final=0.471, loss_mean=0.836, loss_mean_cls=0.0642, proj_loss=-0.429][2026-03-26 12:58:16] Step: 6559, Training Logs: loss_final: 0.454710, loss_mean: 0.807489, proj_loss: -0.420525, loss_mean_cls: 0.067746, grad_norm: 0.213283 +Steps: 1%| | 6560/1000000 [27:13<68:18:51, 4.04it/s, grad_norm=0.213, loss_final=0.455, loss_mean=0.807, loss_mean_cls=0.0677, proj_loss=-0.421][2026-03-26 12:58:17] Step: 6560, Training Logs: loss_final: 0.480824, loss_mean: 0.833560, proj_loss: -0.420217, loss_mean_cls: 0.067481, grad_norm: 0.367347 +Steps: 1%| | 6561/1000000 [27:13<68:20:09, 4.04it/s, grad_norm=0.367, loss_final=0.481, loss_mean=0.834, loss_mean_cls=0.0675, proj_loss=-0.42][2026-03-26 12:58:17] Step: 6561, Training Logs: loss_final: 0.475839, loss_mean: 0.814432, proj_loss: -0.410056, loss_mean_cls: 0.071464, grad_norm: 0.246979 +Steps: 1%| | 6562/1000000 [27:14<68:21:40, 4.04it/s, grad_norm=0.247, loss_final=0.476, loss_mean=0.814, loss_mean_cls=0.0715, proj_loss=-0.41][2026-03-26 12:58:17] Step: 6562, Training Logs: loss_final: 0.473664, loss_mean: 0.826732, proj_loss: -0.420507, loss_mean_cls: 0.067439, grad_norm: 0.461358 +Steps: 1%| | 6563/1000000 [27:14<68:19:34, 4.04it/s, grad_norm=0.461, loss_final=0.474, loss_mean=0.827, loss_mean_cls=0.0674, proj_loss=-0.421][2026-03-26 12:58:17] Step: 6563, Training Logs: loss_final: 0.463141, loss_mean: 0.808083, proj_loss: -0.413674, loss_mean_cls: 0.068732, grad_norm: 0.400863 +Steps: 1%| | 6564/1000000 [27:14<68:19:29, 4.04it/s, grad_norm=0.401, loss_final=0.463, loss_mean=0.808, loss_mean_cls=0.0687, proj_loss=-0.414][2026-03-26 12:58:18] Step: 6564, Training Logs: loss_final: 0.472081, loss_mean: 0.817694, proj_loss: -0.413522, loss_mean_cls: 0.067909, grad_norm: 0.338286 +Steps: 1%| | 6565/1000000 [27:14<68:17:03, 4.04it/s, grad_norm=0.338, loss_final=0.472, loss_mean=0.818, loss_mean_cls=0.0679, proj_loss=-0.414][2026-03-26 12:58:18] Step: 6565, Training Logs: loss_final: 0.471352, loss_mean: 0.819291, proj_loss: -0.416418, loss_mean_cls: 0.068479, grad_norm: 0.367452 +Steps: 1%| | 6566/1000000 [27:15<68:15:56, 4.04it/s, grad_norm=0.367, loss_final=0.471, loss_mean=0.819, loss_mean_cls=0.0685, proj_loss=-0.416][2026-03-26 12:58:18] Step: 6566, Training Logs: loss_final: 0.461027, loss_mean: 0.791196, proj_loss: -0.402053, loss_mean_cls: 0.071883, grad_norm: 0.250809 +Steps: 1%| | 6567/1000000 [27:15<68:17:52, 4.04it/s, grad_norm=0.251, loss_final=0.461, loss_mean=0.791, loss_mean_cls=0.0719, proj_loss=-0.402][2026-03-26 12:58:18] Step: 6567, Training Logs: loss_final: 0.479717, loss_mean: 0.819195, proj_loss: -0.409277, loss_mean_cls: 0.069799, grad_norm: 0.445109 +Steps: 1%| | 6568/1000000 [27:15<68:18:18, 4.04it/s, grad_norm=0.445, loss_final=0.48, loss_mean=0.819, loss_mean_cls=0.0698, proj_loss=-0.409][2026-03-26 12:58:19] Step: 6568, Training Logs: loss_final: 0.491224, loss_mean: 0.854230, proj_loss: -0.429017, loss_mean_cls: 0.066011, grad_norm: 0.226056 +Steps: 1%| | 6569/1000000 [27:15<68:19:21, 4.04it/s, grad_norm=0.226, loss_final=0.491, loss_mean=0.854, loss_mean_cls=0.066, proj_loss=-0.429][2026-03-26 12:58:19] Step: 6569, Training Logs: loss_final: 0.470271, loss_mean: 0.815744, proj_loss: -0.413998, loss_mean_cls: 0.068524, grad_norm: 0.404328 +Steps: 1%| | 6570/1000000 [27:16<68:17:18, 4.04it/s, grad_norm=0.404, loss_final=0.47, loss_mean=0.816, loss_mean_cls=0.0685, proj_loss=-0.414][2026-03-26 12:58:19] Step: 6570, Training Logs: loss_final: 0.480851, loss_mean: 0.837120, proj_loss: -0.423685, loss_mean_cls: 0.067415, grad_norm: 0.228041 +Steps: 1%| | 6571/1000000 [27:16<68:18:03, 4.04it/s, grad_norm=0.228, loss_final=0.481, loss_mean=0.837, loss_mean_cls=0.0674, proj_loss=-0.424][2026-03-26 12:58:19] Step: 6571, Training Logs: loss_final: 0.464556, loss_mean: 0.818887, proj_loss: -0.421819, loss_mean_cls: 0.067489, grad_norm: 0.352568 +Steps: 1%| | 6572/1000000 [27:16<68:17:59, 4.04it/s, grad_norm=0.353, loss_final=0.465, loss_mean=0.819, loss_mean_cls=0.0675, proj_loss=-0.422][2026-03-26 12:58:20] Step: 6572, Training Logs: loss_final: 0.475452, loss_mean: 0.810871, proj_loss: -0.406515, loss_mean_cls: 0.071096, grad_norm: 0.280426 +Steps: 1%| | 6573/1000000 [27:16<68:22:45, 4.04it/s, grad_norm=0.28, loss_final=0.475, loss_mean=0.811, loss_mean_cls=0.0711, proj_loss=-0.407][2026-03-26 12:58:20] Step: 6573, Training Logs: loss_final: 0.465143, loss_mean: 0.814863, proj_loss: -0.418481, loss_mean_cls: 0.068761, grad_norm: 0.326749 +Steps: 1%| | 6574/1000000 [27:17<68:21:47, 4.04it/s, grad_norm=0.327, loss_final=0.465, loss_mean=0.815, loss_mean_cls=0.0688, proj_loss=-0.418][2026-03-26 12:58:20] Step: 6574, Training Logs: loss_final: 0.470678, loss_mean: 0.812819, proj_loss: -0.411904, loss_mean_cls: 0.069763, grad_norm: 0.229609 +Steps: 1%| | 6575/1000000 [27:17<68:19:56, 4.04it/s, grad_norm=0.23, loss_final=0.471, loss_mean=0.813, loss_mean_cls=0.0698, proj_loss=-0.412][2026-03-26 12:58:20] Step: 6575, Training Logs: loss_final: 0.472136, loss_mean: 0.818631, proj_loss: -0.414996, loss_mean_cls: 0.068502, grad_norm: 0.354908 +Steps: 1%| | 6576/1000000 [27:17<68:18:34, 4.04it/s, grad_norm=0.355, loss_final=0.472, loss_mean=0.819, loss_mean_cls=0.0685, proj_loss=-0.415][2026-03-26 12:58:21] Step: 6576, Training Logs: loss_final: 0.489501, loss_mean: 0.833835, proj_loss: -0.412708, loss_mean_cls: 0.068374, grad_norm: 0.267108 +Steps: 1%| | 6577/1000000 [27:17<68:22:26, 4.04it/s, grad_norm=0.267, loss_final=0.49, loss_mean=0.834, loss_mean_cls=0.0684, proj_loss=-0.413][2026-03-26 12:58:21] Step: 6577, Training Logs: loss_final: 0.469840, loss_mean: 0.815330, proj_loss: -0.414426, loss_mean_cls: 0.068936, grad_norm: 0.375736 +Steps: 1%| | 6578/1000000 [27:18<68:20:03, 4.04it/s, grad_norm=0.376, loss_final=0.47, loss_mean=0.815, loss_mean_cls=0.0689, proj_loss=-0.414][2026-03-26 12:58:21] Step: 6578, Training Logs: loss_final: 0.475770, loss_mean: 0.819560, proj_loss: -0.413200, loss_mean_cls: 0.069411, grad_norm: 0.441662 +Steps: 1%| | 6579/1000000 [27:18<68:20:35, 4.04it/s, grad_norm=0.442, loss_final=0.476, loss_mean=0.82, loss_mean_cls=0.0694, proj_loss=-0.413][2026-03-26 12:58:21] Step: 6579, Training Logs: loss_final: 0.469414, loss_mean: 0.822433, proj_loss: -0.420642, loss_mean_cls: 0.067623, grad_norm: 0.317407 +Steps: 1%| | 6580/1000000 [27:18<68:20:13, 4.04it/s, grad_norm=0.317, loss_final=0.469, loss_mean=0.822, loss_mean_cls=0.0676, proj_loss=-0.421][2026-03-26 12:58:22] Step: 6580, Training Logs: loss_final: 0.466709, loss_mean: 0.822761, proj_loss: -0.423349, loss_mean_cls: 0.067297, grad_norm: 0.430564 +Steps: 1%| | 6581/1000000 [27:18<68:20:35, 4.04it/s, grad_norm=0.431, loss_final=0.467, loss_mean=0.823, loss_mean_cls=0.0673, proj_loss=-0.423][2026-03-26 12:58:22] Step: 6581, Training Logs: loss_final: 0.471714, loss_mean: 0.824131, proj_loss: -0.419360, loss_mean_cls: 0.066943, grad_norm: 0.589823 +Steps: 1%| | 6582/1000000 [27:19<68:18:09, 4.04it/s, grad_norm=0.59, loss_final=0.472, loss_mean=0.824, loss_mean_cls=0.0669, proj_loss=-0.419][2026-03-26 12:58:22] Step: 6582, Training Logs: loss_final: 0.487225, loss_mean: 0.827998, proj_loss: -0.409920, loss_mean_cls: 0.069147, grad_norm: 0.298368 +Steps: 1%| | 6583/1000000 [27:19<68:18:59, 4.04it/s, grad_norm=0.298, loss_final=0.487, loss_mean=0.828, loss_mean_cls=0.0691, proj_loss=-0.41][2026-03-26 12:58:22] Step: 6583, Training Logs: loss_final: 0.482985, loss_mean: 0.827569, proj_loss: -0.413350, loss_mean_cls: 0.068766, grad_norm: 0.535189 +Steps: 1%| | 6584/1000000 [27:19<68:15:58, 4.04it/s, grad_norm=0.535, loss_final=0.483, loss_mean=0.828, loss_mean_cls=0.0688, proj_loss=-0.413][2026-03-26 12:58:23] Step: 6584, Training Logs: loss_final: 0.474079, loss_mean: 0.813957, proj_loss: -0.409994, loss_mean_cls: 0.070116, grad_norm: 0.515980 +Steps: 1%| | 6585/1000000 [27:19<68:16:58, 4.04it/s, grad_norm=0.516, loss_final=0.474, loss_mean=0.814, loss_mean_cls=0.0701, proj_loss=-0.41][2026-03-26 12:58:23] Step: 6585, Training Logs: loss_final: 0.485208, loss_mean: 0.829839, proj_loss: -0.412560, loss_mean_cls: 0.067929, grad_norm: 0.271226 +Steps: 1%| | 6586/1000000 [27:20<68:16:23, 4.04it/s, grad_norm=0.271, loss_final=0.485, loss_mean=0.83, loss_mean_cls=0.0679, proj_loss=-0.413][2026-03-26 12:58:23] Step: 6586, Training Logs: loss_final: 0.480223, loss_mean: 0.825092, proj_loss: -0.413925, loss_mean_cls: 0.069056, grad_norm: 0.571235 +Steps: 1%| | 6587/1000000 [27:20<68:15:39, 4.04it/s, grad_norm=0.571, loss_final=0.48, loss_mean=0.825, loss_mean_cls=0.0691, proj_loss=-0.414][2026-03-26 12:58:23] Step: 6587, Training Logs: loss_final: 0.476552, loss_mean: 0.816481, proj_loss: -0.410820, loss_mean_cls: 0.070891, grad_norm: 0.302272 +Steps: 1%| | 6588/1000000 [27:20<68:17:14, 4.04it/s, grad_norm=0.302, loss_final=0.477, loss_mean=0.816, loss_mean_cls=0.0709, proj_loss=-0.411][2026-03-26 12:58:24] Step: 6588, Training Logs: loss_final: 0.476704, loss_mean: 0.814314, proj_loss: -0.407902, loss_mean_cls: 0.070293, grad_norm: 0.328475 +Steps: 1%| | 6589/1000000 [27:20<68:16:42, 4.04it/s, grad_norm=0.328, loss_final=0.477, loss_mean=0.814, loss_mean_cls=0.0703, proj_loss=-0.408][2026-03-26 12:58:24] Step: 6589, Training Logs: loss_final: 0.480018, loss_mean: 0.826595, proj_loss: -0.414530, loss_mean_cls: 0.067953, grad_norm: 0.324325 +Steps: 1%| | 6590/1000000 [27:21<68:18:36, 4.04it/s, grad_norm=0.324, loss_final=0.48, loss_mean=0.827, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 12:58:24] Step: 6590, Training Logs: loss_final: 0.480960, loss_mean: 0.836185, proj_loss: -0.421560, loss_mean_cls: 0.066335, grad_norm: 0.231367 +Steps: 1%| | 6591/1000000 [27:21<68:18:15, 4.04it/s, grad_norm=0.231, loss_final=0.481, loss_mean=0.836, loss_mean_cls=0.0663, proj_loss=-0.422][2026-03-26 12:58:24] Step: 6591, Training Logs: loss_final: 0.470075, loss_mean: 0.821679, proj_loss: -0.419170, loss_mean_cls: 0.067566, grad_norm: 0.317302 +Steps: 1%| | 6592/1000000 [27:21<68:19:59, 4.04it/s, grad_norm=0.317, loss_final=0.47, loss_mean=0.822, loss_mean_cls=0.0676, proj_loss=-0.419][2026-03-26 12:58:25] Step: 6592, Training Logs: loss_final: 0.466309, loss_mean: 0.803174, proj_loss: -0.407283, loss_mean_cls: 0.070417, grad_norm: 0.329845 +Steps: 1%| | 6593/1000000 [27:21<68:18:15, 4.04it/s, grad_norm=0.33, loss_final=0.466, loss_mean=0.803, loss_mean_cls=0.0704, proj_loss=-0.407][2026-03-26 12:58:25] Step: 6593, Training Logs: loss_final: 0.474396, loss_mean: 0.818108, proj_loss: -0.412608, loss_mean_cls: 0.068895, grad_norm: 0.317764 +Steps: 1%| | 6594/1000000 [27:22<68:19:31, 4.04it/s, grad_norm=0.318, loss_final=0.474, loss_mean=0.818, loss_mean_cls=0.0689, proj_loss=-0.413][2026-03-26 12:58:25] Step: 6594, Training Logs: loss_final: 0.485888, loss_mean: 0.833586, proj_loss: -0.415324, loss_mean_cls: 0.067626, grad_norm: 0.278048 +Steps: 1%| | 6595/1000000 [27:22<68:16:54, 4.04it/s, grad_norm=0.278, loss_final=0.486, loss_mean=0.834, loss_mean_cls=0.0676, proj_loss=-0.415][2026-03-26 12:58:25] Step: 6595, Training Logs: loss_final: 0.473841, loss_mean: 0.819340, proj_loss: -0.413758, loss_mean_cls: 0.068259, grad_norm: 0.258911 +Steps: 1%| | 6596/1000000 [27:22<68:22:09, 4.04it/s, grad_norm=0.259, loss_final=0.474, loss_mean=0.819, loss_mean_cls=0.0683, proj_loss=-0.414][2026-03-26 12:58:26] Step: 6596, Training Logs: loss_final: 0.449764, loss_mean: 0.787445, proj_loss: -0.408363, loss_mean_cls: 0.070682, grad_norm: 0.328729 +Steps: 1%| | 6597/1000000 [27:22<68:19:19, 4.04it/s, grad_norm=0.329, loss_final=0.45, loss_mean=0.787, loss_mean_cls=0.0707, proj_loss=-0.408][2026-03-26 12:58:26] Step: 6597, Training Logs: loss_final: 0.478337, loss_mean: 0.830745, proj_loss: -0.419192, loss_mean_cls: 0.066784, grad_norm: 0.207413 +Steps: 1%| | 6598/1000000 [27:23<68:26:02, 4.03it/s, grad_norm=0.207, loss_final=0.478, loss_mean=0.831, loss_mean_cls=0.0668, proj_loss=-0.419][2026-03-26 12:58:26] Step: 6598, Training Logs: loss_final: 0.459137, loss_mean: 0.801986, proj_loss: -0.412713, loss_mean_cls: 0.069865, grad_norm: 0.389767 +Steps: 1%| | 6599/1000000 [27:23<68:22:43, 4.04it/s, grad_norm=0.39, loss_final=0.459, loss_mean=0.802, loss_mean_cls=0.0699, proj_loss=-0.413][2026-03-26 12:58:26] Step: 6599, Training Logs: loss_final: 0.459624, loss_mean: 0.800304, proj_loss: -0.410479, loss_mean_cls: 0.069799, grad_norm: 0.373934 +Steps: 1%| | 6600/1000000 [27:23<68:23:17, 4.03it/s, grad_norm=0.374, loss_final=0.46, loss_mean=0.8, loss_mean_cls=0.0698, proj_loss=-0.41][2026-03-26 12:58:27] Step: 6600, Training Logs: loss_final: 0.492193, loss_mean: 0.841768, proj_loss: -0.417115, loss_mean_cls: 0.067540, grad_norm: 0.274545 +Steps: 1%| | 6601/1000000 [27:23<68:20:03, 4.04it/s, grad_norm=0.275, loss_final=0.492, loss_mean=0.842, loss_mean_cls=0.0675, proj_loss=-0.417][2026-03-26 12:58:27] Step: 6601, Training Logs: loss_final: 0.471652, loss_mean: 0.825812, proj_loss: -0.421279, loss_mean_cls: 0.067118, grad_norm: 0.305547 +Steps: 1%| | 6602/1000000 [27:24<68:22:53, 4.04it/s, grad_norm=0.306, loss_final=0.472, loss_mean=0.826, loss_mean_cls=0.0671, proj_loss=-0.421][2026-03-26 12:58:27] Step: 6602, Training Logs: loss_final: 0.474234, loss_mean: 0.827607, proj_loss: -0.420845, loss_mean_cls: 0.067472, grad_norm: 0.391915 +Steps: 1%| | 6603/1000000 [27:24<68:21:28, 4.04it/s, grad_norm=0.392, loss_final=0.474, loss_mean=0.828, loss_mean_cls=0.0675, proj_loss=-0.421][2026-03-26 12:58:27] Step: 6603, Training Logs: loss_final: 0.465107, loss_mean: 0.804795, proj_loss: -0.410752, loss_mean_cls: 0.071065, grad_norm: 0.262237 +Steps: 1%| | 6604/1000000 [27:24<68:19:07, 4.04it/s, grad_norm=0.262, loss_final=0.465, loss_mean=0.805, loss_mean_cls=0.0711, proj_loss=-0.411][2026-03-26 12:58:28] Step: 6604, Training Logs: loss_final: 0.471147, loss_mean: 0.822464, proj_loss: -0.418807, loss_mean_cls: 0.067490, grad_norm: 0.468607 +Steps: 1%| | 6605/1000000 [27:24<68:14:27, 4.04it/s, grad_norm=0.469, loss_final=0.471, loss_mean=0.822, loss_mean_cls=0.0675, proj_loss=-0.419][2026-03-26 12:58:28] Step: 6605, Training Logs: loss_final: 0.466165, loss_mean: 0.818049, proj_loss: -0.419013, loss_mean_cls: 0.067129, grad_norm: 0.457499 +Steps: 1%| | 6606/1000000 [27:25<68:18:05, 4.04it/s, grad_norm=0.457, loss_final=0.466, loss_mean=0.818, loss_mean_cls=0.0671, proj_loss=-0.419][2026-03-26 12:58:28] Step: 6606, Training Logs: loss_final: 0.471329, loss_mean: 0.811983, proj_loss: -0.411280, loss_mean_cls: 0.070626, grad_norm: 0.307876 +Steps: 1%| | 6607/1000000 [27:25<68:17:00, 4.04it/s, grad_norm=0.308, loss_final=0.471, loss_mean=0.812, loss_mean_cls=0.0706, proj_loss=-0.411][2026-03-26 12:58:28] Step: 6607, Training Logs: loss_final: 0.467970, loss_mean: 0.817160, proj_loss: -0.417966, loss_mean_cls: 0.068777, grad_norm: 0.498014 +Steps: 1%| | 6608/1000000 [27:25<68:16:10, 4.04it/s, grad_norm=0.498, loss_final=0.468, loss_mean=0.817, loss_mean_cls=0.0688, proj_loss=-0.418][2026-03-26 12:58:29] Step: 6608, Training Logs: loss_final: 0.463717, loss_mean: 0.812264, proj_loss: -0.416070, loss_mean_cls: 0.067522, grad_norm: 0.336336 +Steps: 1%| | 6609/1000000 [27:25<68:19:12, 4.04it/s, grad_norm=0.336, loss_final=0.464, loss_mean=0.812, loss_mean_cls=0.0675, proj_loss=-0.416][2026-03-26 12:58:29] Step: 6609, Training Logs: loss_final: 0.468227, loss_mean: 0.807695, proj_loss: -0.409771, loss_mean_cls: 0.070303, grad_norm: 0.332008 +Steps: 1%| | 6610/1000000 [27:26<70:44:38, 3.90it/s, grad_norm=0.332, loss_final=0.468, loss_mean=0.808, loss_mean_cls=0.0703, proj_loss=-0.41][2026-03-26 12:58:29] Step: 6610, Training Logs: loss_final: 0.471722, loss_mean: 0.820095, proj_loss: -0.416275, loss_mean_cls: 0.067902, grad_norm: 0.294386 +Steps: 1%| | 6611/1000000 [27:26<69:59:22, 3.94it/s, grad_norm=0.294, loss_final=0.472, loss_mean=0.82, loss_mean_cls=0.0679, proj_loss=-0.416][2026-03-26 12:58:29] Step: 6611, Training Logs: loss_final: 0.459454, loss_mean: 0.810881, proj_loss: -0.418826, loss_mean_cls: 0.067399, grad_norm: 0.332757 +Steps: 1%| | 6612/1000000 [27:26<69:31:03, 3.97it/s, grad_norm=0.333, loss_final=0.459, loss_mean=0.811, loss_mean_cls=0.0674, proj_loss=-0.419][2026-03-26 12:58:30] Step: 6612, Training Logs: loss_final: 0.469283, loss_mean: 0.814415, proj_loss: -0.413333, loss_mean_cls: 0.068200, grad_norm: 0.254361 +Steps: 1%| | 6613/1000000 [27:26<69:07:48, 3.99it/s, grad_norm=0.254, loss_final=0.469, loss_mean=0.814, loss_mean_cls=0.0682, proj_loss=-0.413][2026-03-26 12:58:30] Step: 6613, Training Logs: loss_final: 0.467397, loss_mean: 0.817082, proj_loss: -0.417758, loss_mean_cls: 0.068073, grad_norm: 0.284690 +Steps: 1%| | 6614/1000000 [27:27<68:54:57, 4.00it/s, grad_norm=0.285, loss_final=0.467, loss_mean=0.817, loss_mean_cls=0.0681, proj_loss=-0.418][2026-03-26 12:58:30] Step: 6614, Training Logs: loss_final: 0.454199, loss_mean: 0.816651, proj_loss: -0.428602, loss_mean_cls: 0.066150, grad_norm: 0.310370 +Steps: 1%| | 6615/1000000 [27:27<68:41:09, 4.02it/s, grad_norm=0.31, loss_final=0.454, loss_mean=0.817, loss_mean_cls=0.0661, proj_loss=-0.429][2026-03-26 12:58:30] Step: 6615, Training Logs: loss_final: 0.456041, loss_mean: 0.797996, proj_loss: -0.411326, loss_mean_cls: 0.069371, grad_norm: 0.295141 +Steps: 1%| | 6616/1000000 [27:27<68:35:21, 4.02it/s, grad_norm=0.295, loss_final=0.456, loss_mean=0.798, loss_mean_cls=0.0694, proj_loss=-0.411][2026-03-26 12:58:31] Step: 6616, Training Logs: loss_final: 0.466989, loss_mean: 0.820441, proj_loss: -0.421048, loss_mean_cls: 0.067596, grad_norm: 0.233471 +Steps: 1%| | 6617/1000000 [27:27<68:28:45, 4.03it/s, grad_norm=0.233, loss_final=0.467, loss_mean=0.82, loss_mean_cls=0.0676, proj_loss=-0.421][2026-03-26 12:58:31] Step: 6617, Training Logs: loss_final: 0.490968, loss_mean: 0.838014, proj_loss: -0.416311, loss_mean_cls: 0.069265, grad_norm: 0.248385 +Steps: 1%| | 6618/1000000 [27:28<68:25:35, 4.03it/s, grad_norm=0.248, loss_final=0.491, loss_mean=0.838, loss_mean_cls=0.0693, proj_loss=-0.416][2026-03-26 12:58:31] Step: 6618, Training Logs: loss_final: 0.463634, loss_mean: 0.824127, proj_loss: -0.426022, loss_mean_cls: 0.065529, grad_norm: 0.507318 +Steps: 1%| | 6619/1000000 [27:28<68:24:55, 4.03it/s, grad_norm=0.507, loss_final=0.464, loss_mean=0.824, loss_mean_cls=0.0655, proj_loss=-0.426][2026-03-26 12:58:31] Step: 6619, Training Logs: loss_final: 0.473072, loss_mean: 0.822137, proj_loss: -0.418088, loss_mean_cls: 0.069023, grad_norm: 0.450939 +Steps: 1%| | 6620/1000000 [27:28<68:21:56, 4.04it/s, grad_norm=0.451, loss_final=0.473, loss_mean=0.822, loss_mean_cls=0.069, proj_loss=-0.418][2026-03-26 12:58:32] Step: 6620, Training Logs: loss_final: 0.477538, loss_mean: 0.816267, proj_loss: -0.407738, loss_mean_cls: 0.069009, grad_norm: 0.365035 +Steps: 1%| | 6621/1000000 [27:28<68:19:09, 4.04it/s, grad_norm=0.365, loss_final=0.478, loss_mean=0.816, loss_mean_cls=0.069, proj_loss=-0.408][2026-03-26 12:58:32] Step: 6621, Training Logs: loss_final: 0.455545, loss_mean: 0.807968, proj_loss: -0.419124, loss_mean_cls: 0.066701, grad_norm: 0.365341 +Steps: 1%| | 6622/1000000 [27:29<68:18:09, 4.04it/s, grad_norm=0.365, loss_final=0.456, loss_mean=0.808, loss_mean_cls=0.0667, proj_loss=-0.419][2026-03-26 12:58:32] Step: 6622, Training Logs: loss_final: 0.472031, loss_mean: 0.818329, proj_loss: -0.414451, loss_mean_cls: 0.068153, grad_norm: 0.333356 +Steps: 1%| | 6623/1000000 [27:29<68:19:25, 4.04it/s, grad_norm=0.333, loss_final=0.472, loss_mean=0.818, loss_mean_cls=0.0682, proj_loss=-0.414][2026-03-26 12:58:32] Step: 6623, Training Logs: loss_final: 0.475539, loss_mean: 0.833314, proj_loss: -0.424679, loss_mean_cls: 0.066904, grad_norm: 0.316114 +Steps: 1%| | 6624/1000000 [27:29<68:16:57, 4.04it/s, grad_norm=0.316, loss_final=0.476, loss_mean=0.833, loss_mean_cls=0.0669, proj_loss=-0.425][2026-03-26 12:58:33] Step: 6624, Training Logs: loss_final: 0.473453, loss_mean: 0.820848, proj_loss: -0.415337, loss_mean_cls: 0.067941, grad_norm: 0.440154 +Steps: 1%| | 6625/1000000 [27:29<68:15:31, 4.04it/s, grad_norm=0.44, loss_final=0.473, loss_mean=0.821, loss_mean_cls=0.0679, proj_loss=-0.415][2026-03-26 12:58:33] Step: 6625, Training Logs: loss_final: 0.487792, loss_mean: 0.838069, proj_loss: -0.418065, loss_mean_cls: 0.067789, grad_norm: 0.333931 +Steps: 1%| | 6626/1000000 [27:30<68:15:01, 4.04it/s, grad_norm=0.334, loss_final=0.488, loss_mean=0.838, loss_mean_cls=0.0678, proj_loss=-0.418][2026-03-26 12:58:33] Step: 6626, Training Logs: loss_final: 0.487106, loss_mean: 0.835914, proj_loss: -0.416532, loss_mean_cls: 0.067725, grad_norm: 0.507791 +Steps: 1%| | 6627/1000000 [27:30<68:16:51, 4.04it/s, grad_norm=0.508, loss_final=0.487, loss_mean=0.836, loss_mean_cls=0.0677, proj_loss=-0.417][2026-03-26 12:58:33] Step: 6627, Training Logs: loss_final: 0.463945, loss_mean: 0.803566, proj_loss: -0.410289, loss_mean_cls: 0.070669, grad_norm: 0.357263 +Steps: 1%| | 6628/1000000 [27:30<68:15:07, 4.04it/s, grad_norm=0.357, loss_final=0.464, loss_mean=0.804, loss_mean_cls=0.0707, proj_loss=-0.41][2026-03-26 12:58:33] Step: 6628, Training Logs: loss_final: 0.479042, loss_mean: 0.831664, proj_loss: -0.420248, loss_mean_cls: 0.067626, grad_norm: 0.238215 +Steps: 1%| | 6629/1000000 [27:30<68:16:51, 4.04it/s, grad_norm=0.238, loss_final=0.479, loss_mean=0.832, loss_mean_cls=0.0676, proj_loss=-0.42][2026-03-26 12:58:34] Step: 6629, Training Logs: loss_final: 0.467355, loss_mean: 0.832255, proj_loss: -0.429807, loss_mean_cls: 0.064907, grad_norm: 0.401000 +Steps: 1%| | 6630/1000000 [27:31<68:15:56, 4.04it/s, grad_norm=0.401, loss_final=0.467, loss_mean=0.832, loss_mean_cls=0.0649, proj_loss=-0.43][2026-03-26 12:58:34] Step: 6630, Training Logs: loss_final: 0.469594, loss_mean: 0.813678, proj_loss: -0.412857, loss_mean_cls: 0.068773, grad_norm: 0.262692 +Steps: 1%| | 6631/1000000 [27:31<68:19:32, 4.04it/s, grad_norm=0.263, loss_final=0.47, loss_mean=0.814, loss_mean_cls=0.0688, proj_loss=-0.413][2026-03-26 12:58:34] Step: 6631, Training Logs: loss_final: 0.453468, loss_mean: 0.808024, proj_loss: -0.421845, loss_mean_cls: 0.067289, grad_norm: 0.367528 +Steps: 1%| | 6632/1000000 [27:31<68:21:43, 4.04it/s, grad_norm=0.368, loss_final=0.453, loss_mean=0.808, loss_mean_cls=0.0673, proj_loss=-0.422][2026-03-26 12:58:34] Step: 6632, Training Logs: loss_final: 0.479141, loss_mean: 0.829593, proj_loss: -0.417625, loss_mean_cls: 0.067172, grad_norm: 0.362982 +Steps: 1%| | 6633/1000000 [27:31<68:27:08, 4.03it/s, grad_norm=0.363, loss_final=0.479, loss_mean=0.83, loss_mean_cls=0.0672, proj_loss=-0.418][2026-03-26 12:58:35] Step: 6633, Training Logs: loss_final: 0.469174, loss_mean: 0.815086, proj_loss: -0.414347, loss_mean_cls: 0.068435, grad_norm: 0.250904 +Steps: 1%| | 6634/1000000 [27:32<68:31:23, 4.03it/s, grad_norm=0.251, loss_final=0.469, loss_mean=0.815, loss_mean_cls=0.0684, proj_loss=-0.414][2026-03-26 12:58:35] Step: 6634, Training Logs: loss_final: 0.481529, loss_mean: 0.839881, proj_loss: -0.424543, loss_mean_cls: 0.066191, grad_norm: 0.469121 +Steps: 1%| | 6635/1000000 [27:32<68:25:06, 4.03it/s, grad_norm=0.469, loss_final=0.482, loss_mean=0.84, loss_mean_cls=0.0662, proj_loss=-0.425][2026-03-26 12:58:35] Step: 6635, Training Logs: loss_final: 0.472369, loss_mean: 0.820273, proj_loss: -0.415609, loss_mean_cls: 0.067704, grad_norm: 0.240694 +Steps: 1%| | 6636/1000000 [27:32<68:20:09, 4.04it/s, grad_norm=0.241, loss_final=0.472, loss_mean=0.82, loss_mean_cls=0.0677, proj_loss=-0.416][2026-03-26 12:58:35] Step: 6636, Training Logs: loss_final: 0.455272, loss_mean: 0.812489, proj_loss: -0.425235, loss_mean_cls: 0.068019, grad_norm: 0.438192 +Steps: 1%| | 6637/1000000 [27:32<68:18:48, 4.04it/s, grad_norm=0.438, loss_final=0.455, loss_mean=0.812, loss_mean_cls=0.068, proj_loss=-0.425][2026-03-26 12:58:36] Step: 6637, Training Logs: loss_final: 0.480739, loss_mean: 0.829836, proj_loss: -0.417479, loss_mean_cls: 0.068382, grad_norm: 0.252451 +Steps: 1%| | 6638/1000000 [27:33<68:17:22, 4.04it/s, grad_norm=0.252, loss_final=0.481, loss_mean=0.83, loss_mean_cls=0.0684, proj_loss=-0.417][2026-03-26 12:58:36] Step: 6638, Training Logs: loss_final: 0.500476, loss_mean: 0.855190, proj_loss: -0.421062, loss_mean_cls: 0.066348, grad_norm: 0.445742 +Steps: 1%| | 6639/1000000 [27:33<68:19:40, 4.04it/s, grad_norm=0.446, loss_final=0.5, loss_mean=0.855, loss_mean_cls=0.0663, proj_loss=-0.421][2026-03-26 12:58:36] Step: 6639, Training Logs: loss_final: 0.466088, loss_mean: 0.814100, proj_loss: -0.416921, loss_mean_cls: 0.068908, grad_norm: 0.391299 +Steps: 1%| | 6640/1000000 [27:33<68:21:01, 4.04it/s, grad_norm=0.391, loss_final=0.466, loss_mean=0.814, loss_mean_cls=0.0689, proj_loss=-0.417][2026-03-26 12:58:36] Step: 6640, Training Logs: loss_final: 0.463267, loss_mean: 0.820377, proj_loss: -0.423662, loss_mean_cls: 0.066551, grad_norm: 0.306164 +Steps: 1%| | 6641/1000000 [27:33<68:19:00, 4.04it/s, grad_norm=0.306, loss_final=0.463, loss_mean=0.82, loss_mean_cls=0.0666, proj_loss=-0.424][2026-03-26 12:58:37] Step: 6641, Training Logs: loss_final: 0.477188, loss_mean: 0.832250, proj_loss: -0.421697, loss_mean_cls: 0.066634, grad_norm: 0.489425 +Steps: 1%| | 6642/1000000 [27:34<68:20:06, 4.04it/s, grad_norm=0.489, loss_final=0.477, loss_mean=0.832, loss_mean_cls=0.0666, proj_loss=-0.422][2026-03-26 12:58:37] Step: 6642, Training Logs: loss_final: 0.472158, loss_mean: 0.809484, proj_loss: -0.407804, loss_mean_cls: 0.070477, grad_norm: 0.235298 +Steps: 1%| | 6643/1000000 [27:34<68:20:32, 4.04it/s, grad_norm=0.235, loss_final=0.472, loss_mean=0.809, loss_mean_cls=0.0705, proj_loss=-0.408][2026-03-26 12:58:37] Step: 6643, Training Logs: loss_final: 0.490428, loss_mean: 0.833129, proj_loss: -0.411949, loss_mean_cls: 0.069248, grad_norm: 0.357352 +Steps: 1%| | 6644/1000000 [27:34<68:20:36, 4.04it/s, grad_norm=0.357, loss_final=0.49, loss_mean=0.833, loss_mean_cls=0.0692, proj_loss=-0.412][2026-03-26 12:58:37] Step: 6644, Training Logs: loss_final: 0.479060, loss_mean: 0.831647, proj_loss: -0.420211, loss_mean_cls: 0.067625, grad_norm: 0.500168 +Steps: 1%| | 6645/1000000 [27:34<68:28:51, 4.03it/s, grad_norm=0.5, loss_final=0.479, loss_mean=0.832, loss_mean_cls=0.0676, proj_loss=-0.42][2026-03-26 12:58:38] Step: 6645, Training Logs: loss_final: 0.481149, loss_mean: 0.813108, proj_loss: -0.403758, loss_mean_cls: 0.071799, grad_norm: 0.527083 +Steps: 1%| | 6646/1000000 [27:35<68:29:04, 4.03it/s, grad_norm=0.527, loss_final=0.481, loss_mean=0.813, loss_mean_cls=0.0718, proj_loss=-0.404][2026-03-26 12:58:38] Step: 6646, Training Logs: loss_final: 0.461765, loss_mean: 0.804332, proj_loss: -0.411444, loss_mean_cls: 0.068877, grad_norm: 0.261947 +Steps: 1%| | 6647/1000000 [27:35<68:28:47, 4.03it/s, grad_norm=0.262, loss_final=0.462, loss_mean=0.804, loss_mean_cls=0.0689, proj_loss=-0.411][2026-03-26 12:58:38] Step: 6647, Training Logs: loss_final: 0.476369, loss_mean: 0.808388, proj_loss: -0.403308, loss_mean_cls: 0.071289, grad_norm: 0.558119 +Steps: 1%| | 6648/1000000 [27:35<68:28:32, 4.03it/s, grad_norm=0.558, loss_final=0.476, loss_mean=0.808, loss_mean_cls=0.0713, proj_loss=-0.403][2026-03-26 12:58:38] Step: 6648, Training Logs: loss_final: 0.465913, loss_mean: 0.818836, proj_loss: -0.420355, loss_mean_cls: 0.067433, grad_norm: 0.254749 +Steps: 1%| | 6649/1000000 [27:35<68:25:45, 4.03it/s, grad_norm=0.255, loss_final=0.466, loss_mean=0.819, loss_mean_cls=0.0674, proj_loss=-0.42][2026-03-26 12:58:39] Step: 6649, Training Logs: loss_final: 0.466577, loss_mean: 0.807524, proj_loss: -0.411071, loss_mean_cls: 0.070124, grad_norm: 0.531358 +Steps: 1%| | 6650/1000000 [27:36<68:21:27, 4.04it/s, grad_norm=0.531, loss_final=0.467, loss_mean=0.808, loss_mean_cls=0.0701, proj_loss=-0.411][2026-03-26 12:58:39] Step: 6650, Training Logs: loss_final: 0.469105, loss_mean: 0.825775, proj_loss: -0.422569, loss_mean_cls: 0.065899, grad_norm: 0.435443 +Steps: 1%| | 6651/1000000 [27:36<68:18:17, 4.04it/s, grad_norm=0.435, loss_final=0.469, loss_mean=0.826, loss_mean_cls=0.0659, proj_loss=-0.423][2026-03-26 12:58:39] Step: 6651, Training Logs: loss_final: 0.469380, loss_mean: 0.821977, proj_loss: -0.419573, loss_mean_cls: 0.066975, grad_norm: 0.300381 +Steps: 1%| | 6652/1000000 [27:36<68:19:35, 4.04it/s, grad_norm=0.3, loss_final=0.469, loss_mean=0.822, loss_mean_cls=0.067, proj_loss=-0.42][2026-03-26 12:58:39] Step: 6652, Training Logs: loss_final: 0.474044, loss_mean: 0.818139, proj_loss: -0.413073, loss_mean_cls: 0.068978, grad_norm: 0.369980 +Steps: 1%| | 6653/1000000 [27:36<68:18:11, 4.04it/s, grad_norm=0.37, loss_final=0.474, loss_mean=0.818, loss_mean_cls=0.069, proj_loss=-0.413][2026-03-26 12:58:40] Step: 6653, Training Logs: loss_final: 0.459708, loss_mean: 0.815635, proj_loss: -0.423158, loss_mean_cls: 0.067231, grad_norm: 0.319677 +Steps: 1%| | 6654/1000000 [27:37<68:18:32, 4.04it/s, grad_norm=0.32, loss_final=0.46, loss_mean=0.816, loss_mean_cls=0.0672, proj_loss=-0.423][2026-03-26 12:58:40] Step: 6654, Training Logs: loss_final: 0.474219, loss_mean: 0.811459, proj_loss: -0.407952, loss_mean_cls: 0.070713, grad_norm: 0.548366 +Steps: 1%| | 6655/1000000 [27:37<68:16:23, 4.04it/s, grad_norm=0.548, loss_final=0.474, loss_mean=0.811, loss_mean_cls=0.0707, proj_loss=-0.408][2026-03-26 12:58:40] Step: 6655, Training Logs: loss_final: 0.489028, loss_mean: 0.843029, proj_loss: -0.420443, loss_mean_cls: 0.066443, grad_norm: 0.291954 +Steps: 1%| | 6656/1000000 [27:37<68:14:59, 4.04it/s, grad_norm=0.292, loss_final=0.489, loss_mean=0.843, loss_mean_cls=0.0664, proj_loss=-0.42][2026-03-26 12:58:40] Step: 6656, Training Logs: loss_final: 0.485303, loss_mean: 0.843691, proj_loss: -0.424544, loss_mean_cls: 0.066156, grad_norm: 0.616558 +Steps: 1%| | 6657/1000000 [27:37<68:13:28, 4.04it/s, grad_norm=0.617, loss_final=0.485, loss_mean=0.844, loss_mean_cls=0.0662, proj_loss=-0.425][2026-03-26 12:58:41] Step: 6657, Training Logs: loss_final: 0.482719, loss_mean: 0.829978, proj_loss: -0.415189, loss_mean_cls: 0.067930, grad_norm: 0.291965 +Steps: 1%| | 6658/1000000 [27:38<68:14:25, 4.04it/s, grad_norm=0.292, loss_final=0.483, loss_mean=0.83, loss_mean_cls=0.0679, proj_loss=-0.415][2026-03-26 12:58:41] Step: 6658, Training Logs: loss_final: 0.500597, loss_mean: 0.854111, proj_loss: -0.420186, loss_mean_cls: 0.066672, grad_norm: 0.500659 +Steps: 1%| | 6659/1000000 [27:38<68:16:14, 4.04it/s, grad_norm=0.501, loss_final=0.501, loss_mean=0.854, loss_mean_cls=0.0667, proj_loss=-0.42][2026-03-26 12:58:41] Step: 6659, Training Logs: loss_final: 0.471338, loss_mean: 0.803510, proj_loss: -0.403330, loss_mean_cls: 0.071158, grad_norm: 0.815872 +Steps: 1%| | 6660/1000000 [27:38<68:22:43, 4.04it/s, grad_norm=0.816, loss_final=0.471, loss_mean=0.804, loss_mean_cls=0.0712, proj_loss=-0.403][2026-03-26 12:58:41] Step: 6660, Training Logs: loss_final: 0.488531, loss_mean: 0.832991, proj_loss: -0.412611, loss_mean_cls: 0.068151, grad_norm: 0.359666 +Steps: 1%| | 6661/1000000 [27:38<68:44:45, 4.01it/s, grad_norm=0.36, loss_final=0.489, loss_mean=0.833, loss_mean_cls=0.0682, proj_loss=-0.413][2026-03-26 12:58:42] Step: 6661, Training Logs: loss_final: 0.495066, loss_mean: 0.838926, proj_loss: -0.412557, loss_mean_cls: 0.068698, grad_norm: 0.736013 +Steps: 1%| | 6662/1000000 [27:39<68:23:05, 4.03it/s, grad_norm=0.736, loss_final=0.495, loss_mean=0.839, loss_mean_cls=0.0687, proj_loss=-0.413][2026-03-26 12:58:42] Step: 6662, Training Logs: loss_final: 0.480304, loss_mean: 0.829960, proj_loss: -0.417872, loss_mean_cls: 0.068216, grad_norm: 0.560502 +Steps: 1%| | 6663/1000000 [27:39<68:22:14, 4.04it/s, grad_norm=0.561, loss_final=0.48, loss_mean=0.83, loss_mean_cls=0.0682, proj_loss=-0.418][2026-03-26 12:58:42] Step: 6663, Training Logs: loss_final: 0.468932, loss_mean: 0.799407, proj_loss: -0.402694, loss_mean_cls: 0.072219, grad_norm: 0.613237 +Steps: 1%| | 6664/1000000 [27:39<68:20:43, 4.04it/s, grad_norm=0.613, loss_final=0.469, loss_mean=0.799, loss_mean_cls=0.0722, proj_loss=-0.403][2026-03-26 12:58:42] Step: 6664, Training Logs: loss_final: 0.486659, loss_mean: 0.823290, proj_loss: -0.407912, loss_mean_cls: 0.071281, grad_norm: 0.341692 +Steps: 1%| | 6665/1000000 [27:39<68:19:54, 4.04it/s, grad_norm=0.342, loss_final=0.487, loss_mean=0.823, loss_mean_cls=0.0713, proj_loss=-0.408][2026-03-26 12:58:43] Step: 6665, Training Logs: loss_final: 0.481500, loss_mean: 0.830072, proj_loss: -0.416312, loss_mean_cls: 0.067740, grad_norm: 0.861209 +Steps: 1%| | 6666/1000000 [27:40<68:22:17, 4.04it/s, grad_norm=0.861, loss_final=0.482, loss_mean=0.83, loss_mean_cls=0.0677, proj_loss=-0.416][2026-03-26 12:58:43] Step: 6666, Training Logs: loss_final: 0.464620, loss_mean: 0.813561, proj_loss: -0.416974, loss_mean_cls: 0.068032, grad_norm: 0.491409 +Steps: 1%| | 6667/1000000 [27:40<68:21:10, 4.04it/s, grad_norm=0.491, loss_final=0.465, loss_mean=0.814, loss_mean_cls=0.068, proj_loss=-0.417][2026-03-26 12:58:43] Step: 6667, Training Logs: loss_final: 0.465587, loss_mean: 0.820921, proj_loss: -0.422235, loss_mean_cls: 0.066901, grad_norm: 0.456619 +Steps: 1%| | 6668/1000000 [27:40<68:58:16, 4.00it/s, grad_norm=0.457, loss_final=0.466, loss_mean=0.821, loss_mean_cls=0.0669, proj_loss=-0.422][2026-03-26 12:58:43] Step: 6668, Training Logs: loss_final: 0.475552, loss_mean: 0.823695, proj_loss: -0.416324, loss_mean_cls: 0.068181, grad_norm: 0.376745 +Steps: 1%| | 6669/1000000 [27:40<68:44:14, 4.01it/s, grad_norm=0.377, loss_final=0.476, loss_mean=0.824, loss_mean_cls=0.0682, proj_loss=-0.416][2026-03-26 12:58:44] Step: 6669, Training Logs: loss_final: 0.487561, loss_mean: 0.842141, proj_loss: -0.420709, loss_mean_cls: 0.066128, grad_norm: 0.373655 +Steps: 1%| | 6670/1000000 [27:41<68:35:15, 4.02it/s, grad_norm=0.374, loss_final=0.488, loss_mean=0.842, loss_mean_cls=0.0661, proj_loss=-0.421][2026-03-26 12:58:44] Step: 6670, Training Logs: loss_final: 0.458928, loss_mean: 0.805033, proj_loss: -0.415533, loss_mean_cls: 0.069429, grad_norm: 0.457174 +Steps: 1%| | 6671/1000000 [27:41<68:29:42, 4.03it/s, grad_norm=0.457, loss_final=0.459, loss_mean=0.805, loss_mean_cls=0.0694, proj_loss=-0.416][2026-03-26 12:58:44] Step: 6671, Training Logs: loss_final: 0.469205, loss_mean: 0.825177, proj_loss: -0.422809, loss_mean_cls: 0.066837, grad_norm: 0.239962 +Steps: 1%| | 6672/1000000 [27:41<68:27:21, 4.03it/s, grad_norm=0.24, loss_final=0.469, loss_mean=0.825, loss_mean_cls=0.0668, proj_loss=-0.423][2026-03-26 12:58:44] Step: 6672, Training Logs: loss_final: 0.478689, loss_mean: 0.822625, proj_loss: -0.413483, loss_mean_cls: 0.069547, grad_norm: 0.568922 +Steps: 1%| | 6673/1000000 [27:41<68:26:43, 4.03it/s, grad_norm=0.569, loss_final=0.479, loss_mean=0.823, loss_mean_cls=0.0695, proj_loss=-0.413][2026-03-26 12:58:45] Step: 6673, Training Logs: loss_final: 0.469192, loss_mean: 0.811682, proj_loss: -0.411923, loss_mean_cls: 0.069433, grad_norm: 0.288869 +Steps: 1%| | 6674/1000000 [27:42<68:24:11, 4.03it/s, grad_norm=0.289, loss_final=0.469, loss_mean=0.812, loss_mean_cls=0.0694, proj_loss=-0.412][2026-03-26 12:58:45] Step: 6674, Training Logs: loss_final: 0.459539, loss_mean: 0.817029, proj_loss: -0.425008, loss_mean_cls: 0.067518, grad_norm: 0.634148 +Steps: 1%| | 6675/1000000 [27:42<68:25:02, 4.03it/s, grad_norm=0.634, loss_final=0.46, loss_mean=0.817, loss_mean_cls=0.0675, proj_loss=-0.425][2026-03-26 12:58:45] Step: 6675, Training Logs: loss_final: 0.487811, loss_mean: 0.831201, proj_loss: -0.411999, loss_mean_cls: 0.068608, grad_norm: 0.390807 +Steps: 1%| | 6676/1000000 [27:42<68:21:58, 4.04it/s, grad_norm=0.391, loss_final=0.488, loss_mean=0.831, loss_mean_cls=0.0686, proj_loss=-0.412][2026-03-26 12:58:45] Step: 6676, Training Logs: loss_final: 0.466641, loss_mean: 0.809353, proj_loss: -0.412745, loss_mean_cls: 0.070033, grad_norm: 0.434255 +Steps: 1%| | 6677/1000000 [27:42<68:23:39, 4.03it/s, grad_norm=0.434, loss_final=0.467, loss_mean=0.809, loss_mean_cls=0.07, proj_loss=-0.413][2026-03-26 12:58:46] Step: 6677, Training Logs: loss_final: 0.486234, loss_mean: 0.837950, proj_loss: -0.418417, loss_mean_cls: 0.066701, grad_norm: 0.484837 +Steps: 1%| | 6678/1000000 [27:42<68:20:00, 4.04it/s, grad_norm=0.485, loss_final=0.486, loss_mean=0.838, loss_mean_cls=0.0667, proj_loss=-0.418][2026-03-26 12:58:46] Step: 6678, Training Logs: loss_final: 0.479485, loss_mean: 0.830148, proj_loss: -0.417914, loss_mean_cls: 0.067250, grad_norm: 0.345522 +Steps: 1%| | 6679/1000000 [27:43<68:21:53, 4.04it/s, grad_norm=0.346, loss_final=0.479, loss_mean=0.83, loss_mean_cls=0.0672, proj_loss=-0.418][2026-03-26 12:58:46] Step: 6679, Training Logs: loss_final: 0.480645, loss_mean: 0.831133, proj_loss: -0.417463, loss_mean_cls: 0.066975, grad_norm: 0.553039 +Steps: 1%| | 6680/1000000 [27:43<68:27:54, 4.03it/s, grad_norm=0.553, loss_final=0.481, loss_mean=0.831, loss_mean_cls=0.067, proj_loss=-0.417][2026-03-26 12:58:46] Step: 6680, Training Logs: loss_final: 0.478004, loss_mean: 0.822789, proj_loss: -0.413720, loss_mean_cls: 0.068935, grad_norm: 0.436043 +Steps: 1%| | 6681/1000000 [27:43<68:25:03, 4.03it/s, grad_norm=0.436, loss_final=0.478, loss_mean=0.823, loss_mean_cls=0.0689, proj_loss=-0.414][2026-03-26 12:58:47] Step: 6681, Training Logs: loss_final: 0.472288, loss_mean: 0.828077, proj_loss: -0.421974, loss_mean_cls: 0.066184, grad_norm: 0.554161 +Steps: 1%| | 6682/1000000 [27:43<68:22:02, 4.04it/s, grad_norm=0.554, loss_final=0.472, loss_mean=0.828, loss_mean_cls=0.0662, proj_loss=-0.422][2026-03-26 12:58:47] Step: 6682, Training Logs: loss_final: 0.470317, loss_mean: 0.827188, proj_loss: -0.423475, loss_mean_cls: 0.066604, grad_norm: 0.420453 +Steps: 1%| | 6683/1000000 [27:44<68:20:33, 4.04it/s, grad_norm=0.42, loss_final=0.47, loss_mean=0.827, loss_mean_cls=0.0666, proj_loss=-0.423][2026-03-26 12:58:47] Step: 6683, Training Logs: loss_final: 0.458727, loss_mean: 0.798862, proj_loss: -0.410133, loss_mean_cls: 0.069998, grad_norm: 0.702389 +Steps: 1%| | 6684/1000000 [27:44<68:18:51, 4.04it/s, grad_norm=0.702, loss_final=0.459, loss_mean=0.799, loss_mean_cls=0.07, proj_loss=-0.41][2026-03-26 12:58:47] Step: 6684, Training Logs: loss_final: 0.460349, loss_mean: 0.806089, proj_loss: -0.414873, loss_mean_cls: 0.069133, grad_norm: 0.211022 +Steps: 1%| | 6685/1000000 [27:44<68:21:09, 4.04it/s, grad_norm=0.211, loss_final=0.46, loss_mean=0.806, loss_mean_cls=0.0691, proj_loss=-0.415][2026-03-26 12:58:48] Step: 6685, Training Logs: loss_final: 0.465042, loss_mean: 0.814440, proj_loss: -0.417598, loss_mean_cls: 0.068200, grad_norm: 0.541295 +Steps: 1%| | 6686/1000000 [27:44<68:20:56, 4.04it/s, grad_norm=0.541, loss_final=0.465, loss_mean=0.814, loss_mean_cls=0.0682, proj_loss=-0.418][2026-03-26 12:58:48] Step: 6686, Training Logs: loss_final: 0.485930, loss_mean: 0.829498, proj_loss: -0.412134, loss_mean_cls: 0.068566, grad_norm: 0.339475 +Steps: 1%| | 6687/1000000 [27:45<68:22:29, 4.04it/s, grad_norm=0.339, loss_final=0.486, loss_mean=0.829, loss_mean_cls=0.0686, proj_loss=-0.412][2026-03-26 12:58:48] Step: 6687, Training Logs: loss_final: 0.473058, loss_mean: 0.824932, proj_loss: -0.419299, loss_mean_cls: 0.067425, grad_norm: 0.329553 +Steps: 1%| | 6688/1000000 [27:45<68:24:06, 4.03it/s, grad_norm=0.33, loss_final=0.473, loss_mean=0.825, loss_mean_cls=0.0674, proj_loss=-0.419][2026-03-26 12:58:48] Step: 6688, Training Logs: loss_final: 0.484747, loss_mean: 0.826821, proj_loss: -0.411718, loss_mean_cls: 0.069644, grad_norm: 0.289846 +Steps: 1%| | 6689/1000000 [27:45<68:22:51, 4.04it/s, grad_norm=0.29, loss_final=0.485, loss_mean=0.827, loss_mean_cls=0.0696, proj_loss=-0.412][2026-03-26 12:58:49] Step: 6689, Training Logs: loss_final: 0.479772, loss_mean: 0.827032, proj_loss: -0.415375, loss_mean_cls: 0.068115, grad_norm: 0.231564 +Steps: 1%| | 6690/1000000 [27:45<68:25:23, 4.03it/s, grad_norm=0.232, loss_final=0.48, loss_mean=0.827, loss_mean_cls=0.0681, proj_loss=-0.415][2026-03-26 12:58:49] Step: 6690, Training Logs: loss_final: 0.476406, loss_mean: 0.828935, proj_loss: -0.419754, loss_mean_cls: 0.067225, grad_norm: 0.291289 +Steps: 1%| | 6691/1000000 [27:46<68:25:26, 4.03it/s, grad_norm=0.291, loss_final=0.476, loss_mean=0.829, loss_mean_cls=0.0672, proj_loss=-0.42][2026-03-26 12:58:49] Step: 6691, Training Logs: loss_final: 0.476062, loss_mean: 0.814254, proj_loss: -0.408009, loss_mean_cls: 0.069818, grad_norm: 0.277437 +Steps: 1%| | 6692/1000000 [27:46<68:28:12, 4.03it/s, grad_norm=0.277, loss_final=0.476, loss_mean=0.814, loss_mean_cls=0.0698, proj_loss=-0.408][2026-03-26 12:58:49] Step: 6692, Training Logs: loss_final: 0.477346, loss_mean: 0.819341, proj_loss: -0.411533, loss_mean_cls: 0.069538, grad_norm: 0.344095 +Steps: 1%| | 6693/1000000 [27:46<68:24:22, 4.03it/s, grad_norm=0.344, loss_final=0.477, loss_mean=0.819, loss_mean_cls=0.0695, proj_loss=-0.412][2026-03-26 12:58:50] Step: 6693, Training Logs: loss_final: 0.484922, loss_mean: 0.834143, proj_loss: -0.418083, loss_mean_cls: 0.068862, grad_norm: 0.275210 +Steps: 1%| | 6694/1000000 [27:46<68:23:22, 4.03it/s, grad_norm=0.275, loss_final=0.485, loss_mean=0.834, loss_mean_cls=0.0689, proj_loss=-0.418][2026-03-26 12:58:50] Step: 6694, Training Logs: loss_final: 0.475844, loss_mean: 0.828556, proj_loss: -0.419074, loss_mean_cls: 0.066362, grad_norm: 0.391806 +Steps: 1%| | 6695/1000000 [27:47<68:20:18, 4.04it/s, grad_norm=0.392, loss_final=0.476, loss_mean=0.829, loss_mean_cls=0.0664, proj_loss=-0.419][2026-03-26 12:58:50] Step: 6695, Training Logs: loss_final: 0.494983, loss_mean: 0.833375, proj_loss: -0.408656, loss_mean_cls: 0.070264, grad_norm: 0.391121 +Steps: 1%| | 6696/1000000 [27:47<68:21:30, 4.04it/s, grad_norm=0.391, loss_final=0.495, loss_mean=0.833, loss_mean_cls=0.0703, proj_loss=-0.409][2026-03-26 12:58:50] Step: 6696, Training Logs: loss_final: 0.493755, loss_mean: 0.844509, proj_loss: -0.418341, loss_mean_cls: 0.067588, grad_norm: 0.333437 +Steps: 1%| | 6697/1000000 [27:47<68:17:16, 4.04it/s, grad_norm=0.333, loss_final=0.494, loss_mean=0.845, loss_mean_cls=0.0676, proj_loss=-0.418][2026-03-26 12:58:51] Step: 6697, Training Logs: loss_final: 0.467248, loss_mean: 0.799270, proj_loss: -0.403314, loss_mean_cls: 0.071293, grad_norm: 0.450142 +Steps: 1%| | 6698/1000000 [27:47<68:21:26, 4.04it/s, grad_norm=0.45, loss_final=0.467, loss_mean=0.799, loss_mean_cls=0.0713, proj_loss=-0.403][2026-03-26 12:58:51] Step: 6698, Training Logs: loss_final: 0.480557, loss_mean: 0.815365, proj_loss: -0.405456, loss_mean_cls: 0.070647, grad_norm: 0.358518 +Steps: 1%| | 6699/1000000 [27:48<68:17:20, 4.04it/s, grad_norm=0.359, loss_final=0.481, loss_mean=0.815, loss_mean_cls=0.0706, proj_loss=-0.405][2026-03-26 12:58:51] Step: 6699, Training Logs: loss_final: 0.467523, loss_mean: 0.809389, proj_loss: -0.410609, loss_mean_cls: 0.068743, grad_norm: 0.390649 +Steps: 1%| | 6700/1000000 [27:48<68:22:32, 4.04it/s, grad_norm=0.391, loss_final=0.468, loss_mean=0.809, loss_mean_cls=0.0687, proj_loss=-0.411][2026-03-26 12:58:51] Step: 6700, Training Logs: loss_final: 0.482158, loss_mean: 0.823927, proj_loss: -0.410896, loss_mean_cls: 0.069127, grad_norm: 0.234074 +Steps: 1%| | 6701/1000000 [27:48<68:20:53, 4.04it/s, grad_norm=0.234, loss_final=0.482, loss_mean=0.824, loss_mean_cls=0.0691, proj_loss=-0.411][2026-03-26 12:58:52] Step: 6701, Training Logs: loss_final: 0.460189, loss_mean: 0.792138, proj_loss: -0.403068, loss_mean_cls: 0.071119, grad_norm: 0.539937 +Steps: 1%| | 6702/1000000 [27:48<68:18:59, 4.04it/s, grad_norm=0.54, loss_final=0.46, loss_mean=0.792, loss_mean_cls=0.0711, proj_loss=-0.403][2026-03-26 12:58:52] Step: 6702, Training Logs: loss_final: 0.477179, loss_mean: 0.824920, proj_loss: -0.415920, loss_mean_cls: 0.068179, grad_norm: 0.332805 +Steps: 1%| | 6703/1000000 [27:49<68:18:11, 4.04it/s, grad_norm=0.333, loss_final=0.477, loss_mean=0.825, loss_mean_cls=0.0682, proj_loss=-0.416][2026-03-26 12:58:52] Step: 6703, Training Logs: loss_final: 0.474945, loss_mean: 0.821021, proj_loss: -0.414140, loss_mean_cls: 0.068065, grad_norm: 0.331242 +Steps: 1%| | 6704/1000000 [27:49<68:18:33, 4.04it/s, grad_norm=0.331, loss_final=0.475, loss_mean=0.821, loss_mean_cls=0.0681, proj_loss=-0.414][2026-03-26 12:58:52] Step: 6704, Training Logs: loss_final: 0.473745, loss_mean: 0.820033, proj_loss: -0.415250, loss_mean_cls: 0.068963, grad_norm: 0.344611 +Steps: 1%| | 6705/1000000 [27:49<68:13:49, 4.04it/s, grad_norm=0.345, loss_final=0.474, loss_mean=0.82, loss_mean_cls=0.069, proj_loss=-0.415][2026-03-26 12:58:53] Step: 6705, Training Logs: loss_final: 0.469866, loss_mean: 0.812573, proj_loss: -0.411143, loss_mean_cls: 0.068435, grad_norm: 0.315546 +Steps: 1%| | 6706/1000000 [27:49<68:16:06, 4.04it/s, grad_norm=0.316, loss_final=0.47, loss_mean=0.813, loss_mean_cls=0.0684, proj_loss=-0.411][2026-03-26 12:58:53] Step: 6706, Training Logs: loss_final: 0.481273, loss_mean: 0.830436, proj_loss: -0.416421, loss_mean_cls: 0.067258, grad_norm: 0.408778 +Steps: 1%| | 6707/1000000 [27:50<68:14:32, 4.04it/s, grad_norm=0.409, loss_final=0.481, loss_mean=0.83, loss_mean_cls=0.0673, proj_loss=-0.416][2026-03-26 12:58:53] Step: 6707, Training Logs: loss_final: 0.474780, loss_mean: 0.829676, proj_loss: -0.421088, loss_mean_cls: 0.066193, grad_norm: 0.380930 +Steps: 1%| | 6708/1000000 [27:50<68:16:58, 4.04it/s, grad_norm=0.381, loss_final=0.475, loss_mean=0.83, loss_mean_cls=0.0662, proj_loss=-0.421][2026-03-26 12:58:53] Step: 6708, Training Logs: loss_final: 0.462571, loss_mean: 0.804118, proj_loss: -0.411398, loss_mean_cls: 0.069850, grad_norm: 0.533712 +Steps: 1%| | 6709/1000000 [27:50<68:17:38, 4.04it/s, grad_norm=0.534, loss_final=0.463, loss_mean=0.804, loss_mean_cls=0.0699, proj_loss=-0.411][2026-03-26 12:58:54] Step: 6709, Training Logs: loss_final: 0.474573, loss_mean: 0.821561, proj_loss: -0.415098, loss_mean_cls: 0.068110, grad_norm: 0.294221 +Steps: 1%| | 6710/1000000 [27:50<68:16:29, 4.04it/s, grad_norm=0.294, loss_final=0.475, loss_mean=0.822, loss_mean_cls=0.0681, proj_loss=-0.415][2026-03-26 12:58:54] Step: 6710, Training Logs: loss_final: 0.470427, loss_mean: 0.811767, proj_loss: -0.411033, loss_mean_cls: 0.069694, grad_norm: 0.327709 +Steps: 1%| | 6711/1000000 [27:51<68:18:16, 4.04it/s, grad_norm=0.328, loss_final=0.47, loss_mean=0.812, loss_mean_cls=0.0697, proj_loss=-0.411][2026-03-26 12:58:54] Step: 6711, Training Logs: loss_final: 0.480409, loss_mean: 0.827733, proj_loss: -0.414649, loss_mean_cls: 0.067324, grad_norm: 0.307892 +Steps: 1%| | 6712/1000000 [27:51<68:17:19, 4.04it/s, grad_norm=0.308, loss_final=0.48, loss_mean=0.828, loss_mean_cls=0.0673, proj_loss=-0.415][2026-03-26 12:58:54] Step: 6712, Training Logs: loss_final: 0.476141, loss_mean: 0.828123, proj_loss: -0.419886, loss_mean_cls: 0.067903, grad_norm: 0.403125 +Steps: 1%| | 6713/1000000 [27:51<68:17:37, 4.04it/s, grad_norm=0.403, loss_final=0.476, loss_mean=0.828, loss_mean_cls=0.0679, proj_loss=-0.42][2026-03-26 12:58:55] Step: 6713, Training Logs: loss_final: 0.476057, loss_mean: 0.829219, proj_loss: -0.420892, loss_mean_cls: 0.067731, grad_norm: 0.254195 +Steps: 1%| | 6714/1000000 [27:51<68:17:47, 4.04it/s, grad_norm=0.254, loss_final=0.476, loss_mean=0.829, loss_mean_cls=0.0677, proj_loss=-0.421][2026-03-26 12:58:55] Step: 6714, Training Logs: loss_final: 0.473923, loss_mean: 0.823254, proj_loss: -0.417578, loss_mean_cls: 0.068247, grad_norm: 0.473321 +Steps: 1%| | 6715/1000000 [27:52<68:18:11, 4.04it/s, grad_norm=0.473, loss_final=0.474, loss_mean=0.823, loss_mean_cls=0.0682, proj_loss=-0.418][2026-03-26 12:58:55] Step: 6715, Training Logs: loss_final: 0.474171, loss_mean: 0.819099, proj_loss: -0.414050, loss_mean_cls: 0.069122, grad_norm: 0.251621 +Steps: 1%| | 6716/1000000 [27:52<68:20:42, 4.04it/s, grad_norm=0.252, loss_final=0.474, loss_mean=0.819, loss_mean_cls=0.0691, proj_loss=-0.414][2026-03-26 12:58:55] Step: 6716, Training Logs: loss_final: 0.468730, loss_mean: 0.807542, proj_loss: -0.408387, loss_mean_cls: 0.069575, grad_norm: 0.542690 +Steps: 1%| | 6717/1000000 [27:52<68:22:19, 4.04it/s, grad_norm=0.543, loss_final=0.469, loss_mean=0.808, loss_mean_cls=0.0696, proj_loss=-0.408][2026-03-26 12:58:56] Step: 6717, Training Logs: loss_final: 0.458660, loss_mean: 0.801103, proj_loss: -0.411382, loss_mean_cls: 0.068940, grad_norm: 0.313804 +Steps: 1%| | 6718/1000000 [27:52<68:22:39, 4.04it/s, grad_norm=0.314, loss_final=0.459, loss_mean=0.801, loss_mean_cls=0.0689, proj_loss=-0.411][2026-03-26 12:58:56] Step: 6718, Training Logs: loss_final: 0.489218, loss_mean: 0.831972, proj_loss: -0.410427, loss_mean_cls: 0.067673, grad_norm: 0.449885 +Steps: 1%| | 6719/1000000 [27:53<68:21:22, 4.04it/s, grad_norm=0.45, loss_final=0.489, loss_mean=0.832, loss_mean_cls=0.0677, proj_loss=-0.41][2026-03-26 12:58:56] Step: 6719, Training Logs: loss_final: 0.455243, loss_mean: 0.800048, proj_loss: -0.414083, loss_mean_cls: 0.069278, grad_norm: 0.486137 +Steps: 1%| | 6720/1000000 [27:53<68:20:19, 4.04it/s, grad_norm=0.486, loss_final=0.455, loss_mean=0.8, loss_mean_cls=0.0693, proj_loss=-0.414][2026-03-26 12:58:56] Step: 6720, Training Logs: loss_final: 0.481593, loss_mean: 0.830599, proj_loss: -0.417362, loss_mean_cls: 0.068356, grad_norm: 0.397874 +Steps: 1%| | 6721/1000000 [27:53<68:18:26, 4.04it/s, grad_norm=0.398, loss_final=0.482, loss_mean=0.831, loss_mean_cls=0.0684, proj_loss=-0.417][2026-03-26 12:58:57] Step: 6721, Training Logs: loss_final: 0.459113, loss_mean: 0.803183, proj_loss: -0.413479, loss_mean_cls: 0.069409, grad_norm: 0.701212 +Steps: 1%| | 6722/1000000 [27:53<68:12:53, 4.04it/s, grad_norm=0.701, loss_final=0.459, loss_mean=0.803, loss_mean_cls=0.0694, proj_loss=-0.413][2026-03-26 12:58:57] Step: 6722, Training Logs: loss_final: 0.468996, loss_mean: 0.805475, proj_loss: -0.407064, loss_mean_cls: 0.070585, grad_norm: 0.339137 +Steps: 1%| | 6723/1000000 [27:54<68:12:53, 4.04it/s, grad_norm=0.339, loss_final=0.469, loss_mean=0.805, loss_mean_cls=0.0706, proj_loss=-0.407][2026-03-26 12:58:57] Step: 6723, Training Logs: loss_final: 0.481671, loss_mean: 0.837717, proj_loss: -0.422513, loss_mean_cls: 0.066467, grad_norm: 0.655788 +Steps: 1%| | 6724/1000000 [27:54<68:15:56, 4.04it/s, grad_norm=0.656, loss_final=0.482, loss_mean=0.838, loss_mean_cls=0.0665, proj_loss=-0.423][2026-03-26 12:58:57] Step: 6724, Training Logs: loss_final: 0.481915, loss_mean: 0.829639, proj_loss: -0.415960, loss_mean_cls: 0.068236, grad_norm: 0.633061 +Steps: 1%| | 6725/1000000 [27:54<68:15:16, 4.04it/s, grad_norm=0.633, loss_final=0.482, loss_mean=0.83, loss_mean_cls=0.0682, proj_loss=-0.416][2026-03-26 12:58:58] Step: 6725, Training Logs: loss_final: 0.478992, loss_mean: 0.826672, proj_loss: -0.414868, loss_mean_cls: 0.067188, grad_norm: 0.263036 +Steps: 1%| | 6726/1000000 [27:54<69:10:55, 3.99it/s, grad_norm=0.263, loss_final=0.479, loss_mean=0.827, loss_mean_cls=0.0672, proj_loss=-0.415][2026-03-26 12:58:58] Step: 6726, Training Logs: loss_final: 0.477974, loss_mean: 0.835198, proj_loss: -0.423472, loss_mean_cls: 0.066247, grad_norm: 0.522723 +Steps: 1%| | 6727/1000000 [27:55<68:56:03, 4.00it/s, grad_norm=0.523, loss_final=0.478, loss_mean=0.835, loss_mean_cls=0.0662, proj_loss=-0.423][2026-03-26 12:58:58] Step: 6727, Training Logs: loss_final: 0.455771, loss_mean: 0.800347, proj_loss: -0.412985, loss_mean_cls: 0.068409, grad_norm: 0.377116 +Steps: 1%| | 6728/1000000 [27:55<68:43:22, 4.01it/s, grad_norm=0.377, loss_final=0.456, loss_mean=0.8, loss_mean_cls=0.0684, proj_loss=-0.413][2026-03-26 12:58:58] Step: 6728, Training Logs: loss_final: 0.477184, loss_mean: 0.827332, proj_loss: -0.417343, loss_mean_cls: 0.067194, grad_norm: 0.350420 +Steps: 1%| | 6729/1000000 [27:55<68:34:16, 4.02it/s, grad_norm=0.35, loss_final=0.477, loss_mean=0.827, loss_mean_cls=0.0672, proj_loss=-0.417][2026-03-26 12:58:59] Step: 6729, Training Logs: loss_final: 0.475983, loss_mean: 0.810940, proj_loss: -0.405120, loss_mean_cls: 0.070162, grad_norm: 0.589034 +Steps: 1%| | 6730/1000000 [27:55<68:28:07, 4.03it/s, grad_norm=0.589, loss_final=0.476, loss_mean=0.811, loss_mean_cls=0.0702, proj_loss=-0.405][2026-03-26 12:58:59] Step: 6730, Training Logs: loss_final: 0.460523, loss_mean: 0.814175, proj_loss: -0.420613, loss_mean_cls: 0.066960, grad_norm: 0.211813 +Steps: 1%| | 6731/1000000 [27:56<68:26:20, 4.03it/s, grad_norm=0.212, loss_final=0.461, loss_mean=0.814, loss_mean_cls=0.067, proj_loss=-0.421][2026-03-26 12:58:59] Step: 6731, Training Logs: loss_final: 0.466504, loss_mean: 0.821325, proj_loss: -0.421319, loss_mean_cls: 0.066499, grad_norm: 0.501143 +Steps: 1%| | 6732/1000000 [27:56<68:23:39, 4.03it/s, grad_norm=0.501, loss_final=0.467, loss_mean=0.821, loss_mean_cls=0.0665, proj_loss=-0.421][2026-03-26 12:58:59] Step: 6732, Training Logs: loss_final: 0.451362, loss_mean: 0.808424, proj_loss: -0.424107, loss_mean_cls: 0.067045, grad_norm: 0.281175 +Steps: 1%| | 6733/1000000 [27:56<68:22:46, 4.03it/s, grad_norm=0.281, loss_final=0.451, loss_mean=0.808, loss_mean_cls=0.067, proj_loss=-0.424][2026-03-26 12:59:00] Step: 6733, Training Logs: loss_final: 0.494436, loss_mean: 0.838544, proj_loss: -0.412260, loss_mean_cls: 0.068152, grad_norm: 0.343798 +Steps: 1%| | 6734/1000000 [27:56<68:21:05, 4.04it/s, grad_norm=0.344, loss_final=0.494, loss_mean=0.839, loss_mean_cls=0.0682, proj_loss=-0.412][2026-03-26 12:59:00] Step: 6734, Training Logs: loss_final: 0.466948, loss_mean: 0.810464, proj_loss: -0.413598, loss_mean_cls: 0.070083, grad_norm: 0.301012 +Steps: 1%| | 6735/1000000 [27:57<68:20:30, 4.04it/s, grad_norm=0.301, loss_final=0.467, loss_mean=0.81, loss_mean_cls=0.0701, proj_loss=-0.414][2026-03-26 12:59:00] Step: 6735, Training Logs: loss_final: 0.487427, loss_mean: 0.832540, proj_loss: -0.414159, loss_mean_cls: 0.069046, grad_norm: 0.430176 +Steps: 1%| | 6736/1000000 [27:57<68:20:28, 4.04it/s, grad_norm=0.43, loss_final=0.487, loss_mean=0.833, loss_mean_cls=0.069, proj_loss=-0.414][2026-03-26 12:59:00] Step: 6736, Training Logs: loss_final: 0.489179, loss_mean: 0.827113, proj_loss: -0.408020, loss_mean_cls: 0.070087, grad_norm: 0.281856 +Steps: 1%| | 6737/1000000 [27:57<68:19:18, 4.04it/s, grad_norm=0.282, loss_final=0.489, loss_mean=0.827, loss_mean_cls=0.0701, proj_loss=-0.408][2026-03-26 12:59:01] Step: 6737, Training Logs: loss_final: 0.463705, loss_mean: 0.806072, proj_loss: -0.411559, loss_mean_cls: 0.069192, grad_norm: 0.451267 +Steps: 1%| | 6738/1000000 [27:57<68:22:12, 4.04it/s, grad_norm=0.451, loss_final=0.464, loss_mean=0.806, loss_mean_cls=0.0692, proj_loss=-0.412][2026-03-26 12:59:01] Step: 6738, Training Logs: loss_final: 0.478651, loss_mean: 0.816817, proj_loss: -0.408695, loss_mean_cls: 0.070528, grad_norm: 0.548268 +Steps: 1%| | 6739/1000000 [27:58<69:34:26, 3.97it/s, grad_norm=0.548, loss_final=0.479, loss_mean=0.817, loss_mean_cls=0.0705, proj_loss=-0.409][2026-03-26 12:59:01] Step: 6739, Training Logs: loss_final: 0.471590, loss_mean: 0.808183, proj_loss: -0.406385, loss_mean_cls: 0.069792, grad_norm: 0.342964 +Steps: 1%| | 6740/1000000 [27:58<69:14:41, 3.98it/s, grad_norm=0.343, loss_final=0.472, loss_mean=0.808, loss_mean_cls=0.0698, proj_loss=-0.406][2026-03-26 12:59:01] Step: 6740, Training Logs: loss_final: 0.473798, loss_mean: 0.815925, proj_loss: -0.411826, loss_mean_cls: 0.069699, grad_norm: 0.554139 +Steps: 1%| | 6741/1000000 [27:58<69:01:13, 4.00it/s, grad_norm=0.554, loss_final=0.474, loss_mean=0.816, loss_mean_cls=0.0697, proj_loss=-0.412][2026-03-26 12:59:02] Step: 6741, Training Logs: loss_final: 0.472386, loss_mean: 0.808168, proj_loss: -0.406061, loss_mean_cls: 0.070279, grad_norm: 0.219989 +Steps: 1%| | 6742/1000000 [27:58<68:49:15, 4.01it/s, grad_norm=0.22, loss_final=0.472, loss_mean=0.808, loss_mean_cls=0.0703, proj_loss=-0.406][2026-03-26 12:59:02] Step: 6742, Training Logs: loss_final: 0.485900, loss_mean: 0.838582, proj_loss: -0.419521, loss_mean_cls: 0.066840, grad_norm: 0.807692 +Steps: 1%| | 6743/1000000 [27:59<68:38:40, 4.02it/s, grad_norm=0.808, loss_final=0.486, loss_mean=0.839, loss_mean_cls=0.0668, proj_loss=-0.42][2026-03-26 12:59:02] Step: 6743, Training Logs: loss_final: 0.485231, loss_mean: 0.831501, proj_loss: -0.414532, loss_mean_cls: 0.068262, grad_norm: 0.642236 +Steps: 1%| | 6744/1000000 [27:59<68:33:39, 4.02it/s, grad_norm=0.642, loss_final=0.485, loss_mean=0.832, loss_mean_cls=0.0683, proj_loss=-0.415][2026-03-26 12:59:02] Step: 6744, Training Logs: loss_final: 0.482591, loss_mean: 0.823515, proj_loss: -0.409325, loss_mean_cls: 0.068401, grad_norm: 0.474761 +Steps: 1%| | 6745/1000000 [27:59<68:28:54, 4.03it/s, grad_norm=0.475, loss_final=0.483, loss_mean=0.824, loss_mean_cls=0.0684, proj_loss=-0.409][2026-03-26 12:59:03] Step: 6745, Training Logs: loss_final: 0.473001, loss_mean: 0.828439, proj_loss: -0.422757, loss_mean_cls: 0.067320, grad_norm: 0.672697 +Steps: 1%| | 6746/1000000 [27:59<68:24:11, 4.03it/s, grad_norm=0.673, loss_final=0.473, loss_mean=0.828, loss_mean_cls=0.0673, proj_loss=-0.423][2026-03-26 12:59:03] Step: 6746, Training Logs: loss_final: 0.477019, loss_mean: 0.846374, proj_loss: -0.433062, loss_mean_cls: 0.063708, grad_norm: 0.296455 +Steps: 1%| | 6747/1000000 [28:00<68:23:07, 4.03it/s, grad_norm=0.296, loss_final=0.477, loss_mean=0.846, loss_mean_cls=0.0637, proj_loss=-0.433][2026-03-26 12:59:03] Step: 6747, Training Logs: loss_final: 0.487484, loss_mean: 0.846458, proj_loss: -0.425050, loss_mean_cls: 0.066076, grad_norm: 0.626632 +Steps: 1%| | 6748/1000000 [28:00<68:21:53, 4.04it/s, grad_norm=0.627, loss_final=0.487, loss_mean=0.846, loss_mean_cls=0.0661, proj_loss=-0.425][2026-03-26 12:59:03] Step: 6748, Training Logs: loss_final: 0.480317, loss_mean: 0.818504, proj_loss: -0.407454, loss_mean_cls: 0.069267, grad_norm: 0.248378 +Steps: 1%| | 6749/1000000 [28:00<68:18:25, 4.04it/s, grad_norm=0.248, loss_final=0.48, loss_mean=0.819, loss_mean_cls=0.0693, proj_loss=-0.407][2026-03-26 12:59:03] Step: 6749, Training Logs: loss_final: 0.496251, loss_mean: 0.837290, proj_loss: -0.410280, loss_mean_cls: 0.069241, grad_norm: 0.619690 +Steps: 1%| | 6750/1000000 [28:00<68:19:15, 4.04it/s, grad_norm=0.62, loss_final=0.496, loss_mean=0.837, loss_mean_cls=0.0692, proj_loss=-0.41][2026-03-26 12:59:04] Step: 6750, Training Logs: loss_final: 0.478667, loss_mean: 0.824209, proj_loss: -0.413406, loss_mean_cls: 0.067864, grad_norm: 0.723597 +Steps: 1%| | 6751/1000000 [28:01<68:17:26, 4.04it/s, grad_norm=0.724, loss_final=0.479, loss_mean=0.824, loss_mean_cls=0.0679, proj_loss=-0.413][2026-03-26 12:59:04] Step: 6751, Training Logs: loss_final: 0.460787, loss_mean: 0.797884, proj_loss: -0.407073, loss_mean_cls: 0.069975, grad_norm: 0.434642 +Steps: 1%| | 6752/1000000 [28:01<68:20:43, 4.04it/s, grad_norm=0.435, loss_final=0.461, loss_mean=0.798, loss_mean_cls=0.07, proj_loss=-0.407][2026-03-26 12:59:04] Step: 6752, Training Logs: loss_final: 0.473609, loss_mean: 0.819657, proj_loss: -0.414662, loss_mean_cls: 0.068614, grad_norm: 0.618843 +Steps: 1%| | 6753/1000000 [28:01<68:16:48, 4.04it/s, grad_norm=0.619, loss_final=0.474, loss_mean=0.82, loss_mean_cls=0.0686, proj_loss=-0.415][2026-03-26 12:59:04] Step: 6753, Training Logs: loss_final: 0.493534, loss_mean: 0.837405, proj_loss: -0.411902, loss_mean_cls: 0.068031, grad_norm: 0.450929 +Steps: 1%| | 6754/1000000 [28:01<68:16:48, 4.04it/s, grad_norm=0.451, loss_final=0.494, loss_mean=0.837, loss_mean_cls=0.068, proj_loss=-0.412][2026-03-26 12:59:05] Step: 6754, Training Logs: loss_final: 0.465330, loss_mean: 0.813554, proj_loss: -0.416829, loss_mean_cls: 0.068605, grad_norm: 0.551316 +Steps: 1%| | 6755/1000000 [28:02<68:17:59, 4.04it/s, grad_norm=0.551, loss_final=0.465, loss_mean=0.814, loss_mean_cls=0.0686, proj_loss=-0.417][2026-03-26 12:59:05] Step: 6755, Training Logs: loss_final: 0.481538, loss_mean: 0.833250, proj_loss: -0.418575, loss_mean_cls: 0.066863, grad_norm: 0.412829 +Steps: 1%| | 6756/1000000 [28:02<68:16:43, 4.04it/s, grad_norm=0.413, loss_final=0.482, loss_mean=0.833, loss_mean_cls=0.0669, proj_loss=-0.419][2026-03-26 12:59:05] Step: 6756, Training Logs: loss_final: 0.488820, loss_mean: 0.826113, proj_loss: -0.407345, loss_mean_cls: 0.070052, grad_norm: 0.702640 +Steps: 1%| | 6757/1000000 [28:02<68:15:01, 4.04it/s, grad_norm=0.703, loss_final=0.489, loss_mean=0.826, loss_mean_cls=0.0701, proj_loss=-0.407][2026-03-26 12:59:05] Step: 6757, Training Logs: loss_final: 0.471599, loss_mean: 0.823521, proj_loss: -0.419187, loss_mean_cls: 0.067265, grad_norm: 0.317213 +Steps: 1%| | 6758/1000000 [28:02<68:16:58, 4.04it/s, grad_norm=0.317, loss_final=0.472, loss_mean=0.824, loss_mean_cls=0.0673, proj_loss=-0.419][2026-03-26 12:59:06] Step: 6758, Training Logs: loss_final: 0.473225, loss_mean: 0.805090, proj_loss: -0.403085, loss_mean_cls: 0.071219, grad_norm: 0.544310 +Steps: 1%| | 6759/1000000 [28:03<68:15:15, 4.04it/s, grad_norm=0.544, loss_final=0.473, loss_mean=0.805, loss_mean_cls=0.0712, proj_loss=-0.403][2026-03-26 12:59:06] Step: 6759, Training Logs: loss_final: 0.479453, loss_mean: 0.829262, proj_loss: -0.417649, loss_mean_cls: 0.067841, grad_norm: 0.387178 +Steps: 1%| | 6760/1000000 [28:03<68:16:59, 4.04it/s, grad_norm=0.387, loss_final=0.479, loss_mean=0.829, loss_mean_cls=0.0678, proj_loss=-0.418][2026-03-26 12:59:06] Step: 6760, Training Logs: loss_final: 0.467190, loss_mean: 0.814567, proj_loss: -0.414728, loss_mean_cls: 0.067351, grad_norm: 0.555078 +Steps: 1%| | 6761/1000000 [28:03<68:17:09, 4.04it/s, grad_norm=0.555, loss_final=0.467, loss_mean=0.815, loss_mean_cls=0.0674, proj_loss=-0.415][2026-03-26 12:59:06] Step: 6761, Training Logs: loss_final: 0.465074, loss_mean: 0.810825, proj_loss: -0.413328, loss_mean_cls: 0.067577, grad_norm: 0.287359 +Steps: 1%| | 6762/1000000 [28:03<68:21:36, 4.04it/s, grad_norm=0.287, loss_final=0.465, loss_mean=0.811, loss_mean_cls=0.0676, proj_loss=-0.413][2026-03-26 12:59:07] Step: 6762, Training Logs: loss_final: 0.474495, loss_mean: 0.818990, proj_loss: -0.413427, loss_mean_cls: 0.068932, grad_norm: 0.451003 +Steps: 1%| | 6763/1000000 [28:04<68:20:48, 4.04it/s, grad_norm=0.451, loss_final=0.474, loss_mean=0.819, loss_mean_cls=0.0689, proj_loss=-0.413][2026-03-26 12:59:07] Step: 6763, Training Logs: loss_final: 0.477153, loss_mean: 0.813541, proj_loss: -0.406292, loss_mean_cls: 0.069905, grad_norm: 0.414480 +Steps: 1%| | 6764/1000000 [28:04<68:18:06, 4.04it/s, grad_norm=0.414, loss_final=0.477, loss_mean=0.814, loss_mean_cls=0.0699, proj_loss=-0.406][2026-03-26 12:59:07] Step: 6764, Training Logs: loss_final: 0.467067, loss_mean: 0.803630, proj_loss: -0.407802, loss_mean_cls: 0.071239, grad_norm: 0.342222 +Steps: 1%| | 6765/1000000 [28:04<68:17:58, 4.04it/s, grad_norm=0.342, loss_final=0.467, loss_mean=0.804, loss_mean_cls=0.0712, proj_loss=-0.408][2026-03-26 12:59:07] Step: 6765, Training Logs: loss_final: 0.476134, loss_mean: 0.820549, proj_loss: -0.413565, loss_mean_cls: 0.069151, grad_norm: 0.378488 +Steps: 1%| | 6766/1000000 [28:04<68:18:03, 4.04it/s, grad_norm=0.378, loss_final=0.476, loss_mean=0.821, loss_mean_cls=0.0692, proj_loss=-0.414][2026-03-26 12:59:08] Step: 6766, Training Logs: loss_final: 0.479955, loss_mean: 0.831800, proj_loss: -0.418800, loss_mean_cls: 0.066956, grad_norm: 0.258678 +Steps: 1%| | 6767/1000000 [28:05<68:22:20, 4.04it/s, grad_norm=0.259, loss_final=0.48, loss_mean=0.832, loss_mean_cls=0.067, proj_loss=-0.419][2026-03-26 12:59:08] Step: 6767, Training Logs: loss_final: 0.468125, loss_mean: 0.819071, proj_loss: -0.418624, loss_mean_cls: 0.067678, grad_norm: 0.359494 +Steps: 1%| | 6768/1000000 [28:05<68:20:17, 4.04it/s, grad_norm=0.359, loss_final=0.468, loss_mean=0.819, loss_mean_cls=0.0677, proj_loss=-0.419][2026-03-26 12:59:08] Step: 6768, Training Logs: loss_final: 0.467572, loss_mean: 0.820724, proj_loss: -0.419849, loss_mean_cls: 0.066698, grad_norm: 0.228001 +Steps: 1%| | 6769/1000000 [28:05<68:21:11, 4.04it/s, grad_norm=0.228, loss_final=0.468, loss_mean=0.821, loss_mean_cls=0.0667, proj_loss=-0.42][2026-03-26 12:59:08] Step: 6769, Training Logs: loss_final: 0.469615, loss_mean: 0.817470, proj_loss: -0.416625, loss_mean_cls: 0.068770, grad_norm: 0.289180 +Steps: 1%| | 6770/1000000 [28:05<68:20:54, 4.04it/s, grad_norm=0.289, loss_final=0.47, loss_mean=0.817, loss_mean_cls=0.0688, proj_loss=-0.417][2026-03-26 12:59:09] Step: 6770, Training Logs: loss_final: 0.474182, loss_mean: 0.815043, proj_loss: -0.410062, loss_mean_cls: 0.069200, grad_norm: 0.307053 +Steps: 1%| | 6771/1000000 [28:06<68:23:31, 4.03it/s, grad_norm=0.307, loss_final=0.474, loss_mean=0.815, loss_mean_cls=0.0692, proj_loss=-0.41][2026-03-26 12:59:09] Step: 6771, Training Logs: loss_final: 0.472741, loss_mean: 0.822765, proj_loss: -0.417642, loss_mean_cls: 0.067619, grad_norm: 0.247932 +Steps: 1%| | 6772/1000000 [28:06<68:18:56, 4.04it/s, grad_norm=0.248, loss_final=0.473, loss_mean=0.823, loss_mean_cls=0.0676, proj_loss=-0.418][2026-03-26 12:59:09] Step: 6772, Training Logs: loss_final: 0.481139, loss_mean: 0.821414, proj_loss: -0.409478, loss_mean_cls: 0.069202, grad_norm: 0.329349 +Steps: 1%| | 6773/1000000 [28:06<68:21:22, 4.04it/s, grad_norm=0.329, loss_final=0.481, loss_mean=0.821, loss_mean_cls=0.0692, proj_loss=-0.409][2026-03-26 12:59:09] Step: 6773, Training Logs: loss_final: 0.467157, loss_mean: 0.815352, proj_loss: -0.415710, loss_mean_cls: 0.067515, grad_norm: 0.272034 +Steps: 1%| | 6774/1000000 [28:06<68:17:56, 4.04it/s, grad_norm=0.272, loss_final=0.467, loss_mean=0.815, loss_mean_cls=0.0675, proj_loss=-0.416][2026-03-26 12:59:10] Step: 6774, Training Logs: loss_final: 0.475217, loss_mean: 0.826378, proj_loss: -0.417792, loss_mean_cls: 0.066631, grad_norm: 0.335632 +Steps: 1%| | 6775/1000000 [28:07<68:18:30, 4.04it/s, grad_norm=0.336, loss_final=0.475, loss_mean=0.826, loss_mean_cls=0.0666, proj_loss=-0.418][2026-03-26 12:59:10] Step: 6775, Training Logs: loss_final: 0.481561, loss_mean: 0.821490, proj_loss: -0.409133, loss_mean_cls: 0.069205, grad_norm: 0.415456 +Steps: 1%| | 6776/1000000 [28:07<68:17:27, 4.04it/s, grad_norm=0.415, loss_final=0.482, loss_mean=0.821, loss_mean_cls=0.0692, proj_loss=-0.409][2026-03-26 12:59:10] Step: 6776, Training Logs: loss_final: 0.489343, loss_mean: 0.835328, proj_loss: -0.413842, loss_mean_cls: 0.067857, grad_norm: 0.265585 +Steps: 1%| | 6777/1000000 [28:07<68:16:50, 4.04it/s, grad_norm=0.266, loss_final=0.489, loss_mean=0.835, loss_mean_cls=0.0679, proj_loss=-0.414][2026-03-26 12:59:10] Step: 6777, Training Logs: loss_final: 0.470426, loss_mean: 0.824830, proj_loss: -0.420918, loss_mean_cls: 0.066513, grad_norm: 0.493297 +Steps: 1%| | 6778/1000000 [28:07<68:15:05, 4.04it/s, grad_norm=0.493, loss_final=0.47, loss_mean=0.825, loss_mean_cls=0.0665, proj_loss=-0.421][2026-03-26 12:59:11] Step: 6778, Training Logs: loss_final: 0.472516, loss_mean: 0.829327, proj_loss: -0.423645, loss_mean_cls: 0.066834, grad_norm: 0.400127 +Steps: 1%| | 6779/1000000 [28:08<68:15:39, 4.04it/s, grad_norm=0.4, loss_final=0.473, loss_mean=0.829, loss_mean_cls=0.0668, proj_loss=-0.424][2026-03-26 12:59:11] Step: 6779, Training Logs: loss_final: 0.467855, loss_mean: 0.818372, proj_loss: -0.419371, loss_mean_cls: 0.068855, grad_norm: 0.265684 +Steps: 1%| | 6780/1000000 [28:08<68:15:27, 4.04it/s, grad_norm=0.266, loss_final=0.468, loss_mean=0.818, loss_mean_cls=0.0689, proj_loss=-0.419][2026-03-26 12:59:11] Step: 6780, Training Logs: loss_final: 0.469684, loss_mean: 0.819241, proj_loss: -0.417290, loss_mean_cls: 0.067734, grad_norm: 0.382477 +Steps: 1%| | 6781/1000000 [28:08<68:14:26, 4.04it/s, grad_norm=0.382, loss_final=0.47, loss_mean=0.819, loss_mean_cls=0.0677, proj_loss=-0.417][2026-03-26 12:59:11] Step: 6781, Training Logs: loss_final: 0.476003, loss_mean: 0.825379, proj_loss: -0.417474, loss_mean_cls: 0.068098, grad_norm: 0.334428 +Steps: 1%| | 6782/1000000 [28:08<68:17:33, 4.04it/s, grad_norm=0.334, loss_final=0.476, loss_mean=0.825, loss_mean_cls=0.0681, proj_loss=-0.417][2026-03-26 12:59:12] Step: 6782, Training Logs: loss_final: 0.491047, loss_mean: 0.823653, proj_loss: -0.402635, loss_mean_cls: 0.070030, grad_norm: 0.378613 +Steps: 1%| | 6783/1000000 [28:09<68:18:00, 4.04it/s, grad_norm=0.379, loss_final=0.491, loss_mean=0.824, loss_mean_cls=0.07, proj_loss=-0.403][2026-03-26 12:59:12] Step: 6783, Training Logs: loss_final: 0.457484, loss_mean: 0.805636, proj_loss: -0.416218, loss_mean_cls: 0.068065, grad_norm: 0.350105 +Steps: 1%| | 6784/1000000 [28:09<68:17:09, 4.04it/s, grad_norm=0.35, loss_final=0.457, loss_mean=0.806, loss_mean_cls=0.0681, proj_loss=-0.416][2026-03-26 12:59:12] Step: 6784, Training Logs: loss_final: 0.462809, loss_mean: 0.810182, proj_loss: -0.414826, loss_mean_cls: 0.067453, grad_norm: 0.324590 +Steps: 1%| | 6785/1000000 [28:09<68:17:32, 4.04it/s, grad_norm=0.325, loss_final=0.463, loss_mean=0.81, loss_mean_cls=0.0675, proj_loss=-0.415][2026-03-26 12:59:12] Step: 6785, Training Logs: loss_final: 0.464166, loss_mean: 0.817995, proj_loss: -0.420642, loss_mean_cls: 0.066813, grad_norm: 0.484993 +Steps: 1%| | 6786/1000000 [28:09<68:16:30, 4.04it/s, grad_norm=0.485, loss_final=0.464, loss_mean=0.818, loss_mean_cls=0.0668, proj_loss=-0.421][2026-03-26 12:59:13] Step: 6786, Training Logs: loss_final: 0.474717, loss_mean: 0.827922, proj_loss: -0.419886, loss_mean_cls: 0.066681, grad_norm: 0.286337 +Steps: 1%| | 6787/1000000 [28:10<68:22:28, 4.04it/s, grad_norm=0.286, loss_final=0.475, loss_mean=0.828, loss_mean_cls=0.0667, proj_loss=-0.42][2026-03-26 12:59:13] Step: 6787, Training Logs: loss_final: 0.476501, loss_mean: 0.825469, proj_loss: -0.416914, loss_mean_cls: 0.067946, grad_norm: 0.397280 +Steps: 1%| | 6788/1000000 [28:10<68:15:20, 4.04it/s, grad_norm=0.397, loss_final=0.477, loss_mean=0.825, loss_mean_cls=0.0679, proj_loss=-0.417][2026-03-26 12:59:13] Step: 6788, Training Logs: loss_final: 0.469764, loss_mean: 0.807518, proj_loss: -0.408105, loss_mean_cls: 0.070351, grad_norm: 0.277752 +Steps: 1%| | 6789/1000000 [28:10<68:15:22, 4.04it/s, grad_norm=0.278, loss_final=0.47, loss_mean=0.808, loss_mean_cls=0.0704, proj_loss=-0.408][2026-03-26 12:59:13] Step: 6789, Training Logs: loss_final: 0.483519, loss_mean: 0.833315, proj_loss: -0.416680, loss_mean_cls: 0.066884, grad_norm: 0.295211 +Steps: 1%| | 6790/1000000 [28:10<68:17:06, 4.04it/s, grad_norm=0.295, loss_final=0.484, loss_mean=0.833, loss_mean_cls=0.0669, proj_loss=-0.417][2026-03-26 12:59:14] Step: 6790, Training Logs: loss_final: 0.480527, loss_mean: 0.814502, proj_loss: -0.404978, loss_mean_cls: 0.071003, grad_norm: 0.199884 +Steps: 1%| | 6791/1000000 [28:11<68:16:24, 4.04it/s, grad_norm=0.2, loss_final=0.481, loss_mean=0.815, loss_mean_cls=0.071, proj_loss=-0.405][2026-03-26 12:59:14] Step: 6791, Training Logs: loss_final: 0.466846, loss_mean: 0.818731, proj_loss: -0.419274, loss_mean_cls: 0.067389, grad_norm: 0.399033 +Steps: 1%| | 6792/1000000 [28:11<68:17:19, 4.04it/s, grad_norm=0.399, loss_final=0.467, loss_mean=0.819, loss_mean_cls=0.0674, proj_loss=-0.419][2026-03-26 12:59:14] Step: 6792, Training Logs: loss_final: 0.471145, loss_mean: 0.823595, proj_loss: -0.419730, loss_mean_cls: 0.067280, grad_norm: 0.379145 +Steps: 1%| | 6793/1000000 [28:11<68:15:17, 4.04it/s, grad_norm=0.379, loss_final=0.471, loss_mean=0.824, loss_mean_cls=0.0673, proj_loss=-0.42][2026-03-26 12:59:14] Step: 6793, Training Logs: loss_final: 0.477188, loss_mean: 0.812904, proj_loss: -0.406098, loss_mean_cls: 0.070382, grad_norm: 0.274582 +Steps: 1%| | 6794/1000000 [28:11<68:14:23, 4.04it/s, grad_norm=0.275, loss_final=0.477, loss_mean=0.813, loss_mean_cls=0.0704, proj_loss=-0.406][2026-03-26 12:59:15] Step: 6794, Training Logs: loss_final: 0.459445, loss_mean: 0.815945, proj_loss: -0.423144, loss_mean_cls: 0.066644, grad_norm: 0.322821 +Steps: 1%| | 6795/1000000 [28:11<68:13:07, 4.04it/s, grad_norm=0.323, loss_final=0.459, loss_mean=0.816, loss_mean_cls=0.0666, proj_loss=-0.423][2026-03-26 12:59:15] Step: 6795, Training Logs: loss_final: 0.479083, loss_mean: 0.815681, proj_loss: -0.407232, loss_mean_cls: 0.070634, grad_norm: 0.324146 +Steps: 1%| | 6796/1000000 [28:12<68:16:24, 4.04it/s, grad_norm=0.324, loss_final=0.479, loss_mean=0.816, loss_mean_cls=0.0706, proj_loss=-0.407][2026-03-26 12:59:15] Step: 6796, Training Logs: loss_final: 0.479365, loss_mean: 0.827855, proj_loss: -0.415941, loss_mean_cls: 0.067450, grad_norm: 0.346061 +Steps: 1%| | 6797/1000000 [28:12<68:17:07, 4.04it/s, grad_norm=0.346, loss_final=0.479, loss_mean=0.828, loss_mean_cls=0.0675, proj_loss=-0.416][2026-03-26 12:59:15] Step: 6797, Training Logs: loss_final: 0.477376, loss_mean: 0.829185, proj_loss: -0.419632, loss_mean_cls: 0.067823, grad_norm: 0.367983 +Steps: 1%| | 6798/1000000 [28:12<68:17:53, 4.04it/s, grad_norm=0.368, loss_final=0.477, loss_mean=0.829, loss_mean_cls=0.0678, proj_loss=-0.42][2026-03-26 12:59:16] Step: 6798, Training Logs: loss_final: 0.468482, loss_mean: 0.823503, proj_loss: -0.422071, loss_mean_cls: 0.067050, grad_norm: 0.243417 +Steps: 1%| | 6799/1000000 [28:12<68:18:49, 4.04it/s, grad_norm=0.243, loss_final=0.468, loss_mean=0.824, loss_mean_cls=0.067, proj_loss=-0.422][2026-03-26 12:59:16] Step: 6799, Training Logs: loss_final: 0.476137, loss_mean: 0.835895, proj_loss: -0.425291, loss_mean_cls: 0.065533, grad_norm: 0.391682 +Steps: 1%| | 6800/1000000 [28:13<68:21:08, 4.04it/s, grad_norm=0.392, loss_final=0.476, loss_mean=0.836, loss_mean_cls=0.0655, proj_loss=-0.425][2026-03-26 12:59:16] Step: 6800, Training Logs: loss_final: 0.471289, loss_mean: 0.824910, proj_loss: -0.419964, loss_mean_cls: 0.066343, grad_norm: 0.327126 +Steps: 1%| | 6801/1000000 [28:13<68:24:19, 4.03it/s, grad_norm=0.327, loss_final=0.471, loss_mean=0.825, loss_mean_cls=0.0663, proj_loss=-0.42][2026-03-26 12:59:16] Step: 6801, Training Logs: loss_final: 0.483418, loss_mean: 0.836432, proj_loss: -0.419388, loss_mean_cls: 0.066374, grad_norm: 0.337676 +Steps: 1%| | 6802/1000000 [28:13<68:25:46, 4.03it/s, grad_norm=0.338, loss_final=0.483, loss_mean=0.836, loss_mean_cls=0.0664, proj_loss=-0.419][2026-03-26 12:59:17] Step: 6802, Training Logs: loss_final: 0.466819, loss_mean: 0.815997, proj_loss: -0.417551, loss_mean_cls: 0.068373, grad_norm: 0.348851 +Steps: 1%| | 6803/1000000 [28:13<68:23:42, 4.03it/s, grad_norm=0.349, loss_final=0.467, loss_mean=0.816, loss_mean_cls=0.0684, proj_loss=-0.418][2026-03-26 12:59:17] Step: 6803, Training Logs: loss_final: 0.463389, loss_mean: 0.818691, proj_loss: -0.421728, loss_mean_cls: 0.066426, grad_norm: 0.451263 +Steps: 1%| | 6804/1000000 [28:14<68:24:23, 4.03it/s, grad_norm=0.451, loss_final=0.463, loss_mean=0.819, loss_mean_cls=0.0664, proj_loss=-0.422][2026-03-26 12:59:17] Step: 6804, Training Logs: loss_final: 0.466469, loss_mean: 0.812282, proj_loss: -0.414593, loss_mean_cls: 0.068780, grad_norm: 0.435441 +Steps: 1%| | 6805/1000000 [28:14<68:23:29, 4.03it/s, grad_norm=0.435, loss_final=0.466, loss_mean=0.812, loss_mean_cls=0.0688, proj_loss=-0.415][2026-03-26 12:59:17] Step: 6805, Training Logs: loss_final: 0.457937, loss_mean: 0.792342, proj_loss: -0.404802, loss_mean_cls: 0.070397, grad_norm: 0.344838 +Steps: 1%| | 6806/1000000 [28:14<68:21:29, 4.04it/s, grad_norm=0.345, loss_final=0.458, loss_mean=0.792, loss_mean_cls=0.0704, proj_loss=-0.405][2026-03-26 12:59:18] Step: 6806, Training Logs: loss_final: 0.476944, loss_mean: 0.825763, proj_loss: -0.416422, loss_mean_cls: 0.067604, grad_norm: 0.433822 +Steps: 1%| | 6807/1000000 [28:14<68:16:53, 4.04it/s, grad_norm=0.434, loss_final=0.477, loss_mean=0.826, loss_mean_cls=0.0676, proj_loss=-0.416][2026-03-26 12:59:18] Step: 6807, Training Logs: loss_final: 0.472197, loss_mean: 0.832454, proj_loss: -0.425372, loss_mean_cls: 0.065115, grad_norm: 0.193211 +Steps: 1%| | 6808/1000000 [28:15<68:17:38, 4.04it/s, grad_norm=0.193, loss_final=0.472, loss_mean=0.832, loss_mean_cls=0.0651, proj_loss=-0.425][2026-03-26 12:59:18] Step: 6808, Training Logs: loss_final: 0.472615, loss_mean: 0.828467, proj_loss: -0.422751, loss_mean_cls: 0.066900, grad_norm: 0.491606 +Steps: 1%| | 6809/1000000 [28:15<68:18:55, 4.04it/s, grad_norm=0.492, loss_final=0.473, loss_mean=0.828, loss_mean_cls=0.0669, proj_loss=-0.423][2026-03-26 12:59:18] Step: 6809, Training Logs: loss_final: 0.488932, loss_mean: 0.847425, proj_loss: -0.424316, loss_mean_cls: 0.065823, grad_norm: 0.406424 +Steps: 1%| | 6810/1000000 [28:15<68:19:35, 4.04it/s, grad_norm=0.406, loss_final=0.489, loss_mean=0.847, loss_mean_cls=0.0658, proj_loss=-0.424][2026-03-26 12:59:19] Step: 6810, Training Logs: loss_final: 0.475198, loss_mean: 0.827872, proj_loss: -0.419727, loss_mean_cls: 0.067052, grad_norm: 0.307867 +Steps: 1%| | 6811/1000000 [28:15<68:20:32, 4.04it/s, grad_norm=0.308, loss_final=0.475, loss_mean=0.828, loss_mean_cls=0.0671, proj_loss=-0.42][2026-03-26 12:59:19] Step: 6811, Training Logs: loss_final: 0.469352, loss_mean: 0.826631, proj_loss: -0.424370, loss_mean_cls: 0.067092, grad_norm: 0.342682 +Steps: 1%| | 6812/1000000 [28:16<68:16:37, 4.04it/s, grad_norm=0.343, loss_final=0.469, loss_mean=0.827, loss_mean_cls=0.0671, proj_loss=-0.424][2026-03-26 12:59:19] Step: 6812, Training Logs: loss_final: 0.466571, loss_mean: 0.812283, proj_loss: -0.414056, loss_mean_cls: 0.068344, grad_norm: 0.293569 +Steps: 1%| | 6813/1000000 [28:16<68:17:00, 4.04it/s, grad_norm=0.294, loss_final=0.467, loss_mean=0.812, loss_mean_cls=0.0683, proj_loss=-0.414][2026-03-26 12:59:19] Step: 6813, Training Logs: loss_final: 0.482448, loss_mean: 0.829428, proj_loss: -0.414809, loss_mean_cls: 0.067829, grad_norm: 0.319712 +Steps: 1%| | 6814/1000000 [28:16<68:17:32, 4.04it/s, grad_norm=0.32, loss_final=0.482, loss_mean=0.829, loss_mean_cls=0.0678, proj_loss=-0.415][2026-03-26 12:59:20] Step: 6814, Training Logs: loss_final: 0.472055, loss_mean: 0.822136, proj_loss: -0.417055, loss_mean_cls: 0.066974, grad_norm: 0.254299 +Steps: 1%| | 6815/1000000 [28:16<68:17:38, 4.04it/s, grad_norm=0.254, loss_final=0.472, loss_mean=0.822, loss_mean_cls=0.067, proj_loss=-0.417][2026-03-26 12:59:20] Step: 6815, Training Logs: loss_final: 0.461215, loss_mean: 0.818660, proj_loss: -0.423607, loss_mean_cls: 0.066162, grad_norm: 0.300222 +Steps: 1%| | 6816/1000000 [28:17<68:17:48, 4.04it/s, grad_norm=0.3, loss_final=0.461, loss_mean=0.819, loss_mean_cls=0.0662, proj_loss=-0.424][2026-03-26 12:59:20] Step: 6816, Training Logs: loss_final: 0.448133, loss_mean: 0.795931, proj_loss: -0.416364, loss_mean_cls: 0.068566, grad_norm: 0.259625 +Steps: 1%| | 6817/1000000 [28:17<68:17:44, 4.04it/s, grad_norm=0.26, loss_final=0.448, loss_mean=0.796, loss_mean_cls=0.0686, proj_loss=-0.416][2026-03-26 12:59:20] Step: 6817, Training Logs: loss_final: 0.469450, loss_mean: 0.814979, proj_loss: -0.413461, loss_mean_cls: 0.067933, grad_norm: 0.453899 +Steps: 1%| | 6818/1000000 [28:17<68:15:58, 4.04it/s, grad_norm=0.454, loss_final=0.469, loss_mean=0.815, loss_mean_cls=0.0679, proj_loss=-0.413][2026-03-26 12:59:21] Step: 6818, Training Logs: loss_final: 0.473432, loss_mean: 0.820214, proj_loss: -0.414673, loss_mean_cls: 0.067891, grad_norm: 0.277288 +Steps: 1%| | 6819/1000000 [28:17<68:15:32, 4.04it/s, grad_norm=0.277, loss_final=0.473, loss_mean=0.82, loss_mean_cls=0.0679, proj_loss=-0.415][2026-03-26 12:59:21] Step: 6819, Training Logs: loss_final: 0.461295, loss_mean: 0.809496, proj_loss: -0.416449, loss_mean_cls: 0.068248, grad_norm: 0.251174 +Steps: 1%| | 6820/1000000 [28:18<68:15:45, 4.04it/s, grad_norm=0.251, loss_final=0.461, loss_mean=0.809, loss_mean_cls=0.0682, proj_loss=-0.416][2026-03-26 12:59:21] Step: 6820, Training Logs: loss_final: 0.453279, loss_mean: 0.814946, proj_loss: -0.427162, loss_mean_cls: 0.065495, grad_norm: 0.401696 +Steps: 1%| | 6821/1000000 [28:18<68:17:17, 4.04it/s, grad_norm=0.402, loss_final=0.453, loss_mean=0.815, loss_mean_cls=0.0655, proj_loss=-0.427][2026-03-26 12:59:21] Step: 6821, Training Logs: loss_final: 0.468556, loss_mean: 0.815640, proj_loss: -0.415931, loss_mean_cls: 0.068846, grad_norm: 0.224923 +Steps: 1%| | 6822/1000000 [28:18<68:17:37, 4.04it/s, grad_norm=0.225, loss_final=0.469, loss_mean=0.816, loss_mean_cls=0.0688, proj_loss=-0.416][2026-03-26 12:59:22] Step: 6822, Training Logs: loss_final: 0.477319, loss_mean: 0.826237, proj_loss: -0.415681, loss_mean_cls: 0.066763, grad_norm: 0.526796 +Steps: 1%| | 6823/1000000 [28:18<68:26:46, 4.03it/s, grad_norm=0.527, loss_final=0.477, loss_mean=0.826, loss_mean_cls=0.0668, proj_loss=-0.416][2026-03-26 12:59:22] Step: 6823, Training Logs: loss_final: 0.470024, loss_mean: 0.818092, proj_loss: -0.416124, loss_mean_cls: 0.068056, grad_norm: 0.279266 +Steps: 1%| | 6824/1000000 [28:19<68:23:57, 4.03it/s, grad_norm=0.279, loss_final=0.47, loss_mean=0.818, loss_mean_cls=0.0681, proj_loss=-0.416][2026-03-26 12:59:22] Step: 6824, Training Logs: loss_final: 0.466185, loss_mean: 0.811285, proj_loss: -0.413089, loss_mean_cls: 0.067988, grad_norm: 0.441919 +Steps: 1%| | 6825/1000000 [28:19<68:20:35, 4.04it/s, grad_norm=0.442, loss_final=0.466, loss_mean=0.811, loss_mean_cls=0.068, proj_loss=-0.413][2026-03-26 12:59:22] Step: 6825, Training Logs: loss_final: 0.471016, loss_mean: 0.810851, proj_loss: -0.409495, loss_mean_cls: 0.069660, grad_norm: 0.444449 +Steps: 1%| | 6826/1000000 [28:19<68:22:03, 4.04it/s, grad_norm=0.444, loss_final=0.471, loss_mean=0.811, loss_mean_cls=0.0697, proj_loss=-0.409][2026-03-26 12:59:23] Step: 6826, Training Logs: loss_final: 0.475151, loss_mean: 0.819604, proj_loss: -0.413100, loss_mean_cls: 0.068647, grad_norm: 0.304962 +Steps: 1%| | 6827/1000000 [28:19<68:22:37, 4.03it/s, grad_norm=0.305, loss_final=0.475, loss_mean=0.82, loss_mean_cls=0.0686, proj_loss=-0.413][2026-03-26 12:59:23] Step: 6827, Training Logs: loss_final: 0.470884, loss_mean: 0.824184, proj_loss: -0.419372, loss_mean_cls: 0.066072, grad_norm: 0.328786 +Steps: 1%| | 6828/1000000 [28:20<68:20:59, 4.04it/s, grad_norm=0.329, loss_final=0.471, loss_mean=0.824, loss_mean_cls=0.0661, proj_loss=-0.419][2026-03-26 12:59:23] Step: 6828, Training Logs: loss_final: 0.463680, loss_mean: 0.804494, proj_loss: -0.410172, loss_mean_cls: 0.069358, grad_norm: 0.389842 +Steps: 1%| | 6829/1000000 [28:20<68:18:30, 4.04it/s, grad_norm=0.39, loss_final=0.464, loss_mean=0.804, loss_mean_cls=0.0694, proj_loss=-0.41][2026-03-26 12:59:23] Step: 6829, Training Logs: loss_final: 0.474772, loss_mean: 0.823504, proj_loss: -0.415673, loss_mean_cls: 0.066941, grad_norm: 0.254870 +Steps: 1%| | 6830/1000000 [28:20<68:18:25, 4.04it/s, grad_norm=0.255, loss_final=0.475, loss_mean=0.824, loss_mean_cls=0.0669, proj_loss=-0.416][2026-03-26 12:59:24] Step: 6830, Training Logs: loss_final: 0.464735, loss_mean: 0.807338, proj_loss: -0.412477, loss_mean_cls: 0.069874, grad_norm: 0.299669 +Steps: 1%| | 6831/1000000 [28:20<68:15:24, 4.04it/s, grad_norm=0.3, loss_final=0.465, loss_mean=0.807, loss_mean_cls=0.0699, proj_loss=-0.412][2026-03-26 12:59:24] Step: 6831, Training Logs: loss_final: 0.467634, loss_mean: 0.814783, proj_loss: -0.414290, loss_mean_cls: 0.067140, grad_norm: 0.315158 +Steps: 1%| | 6832/1000000 [28:21<68:18:35, 4.04it/s, grad_norm=0.315, loss_final=0.468, loss_mean=0.815, loss_mean_cls=0.0671, proj_loss=-0.414][2026-03-26 12:59:24] Step: 6832, Training Logs: loss_final: 0.473478, loss_mean: 0.804995, proj_loss: -0.401848, loss_mean_cls: 0.070331, grad_norm: 0.250813 +Steps: 1%| | 6833/1000000 [28:21<68:14:09, 4.04it/s, grad_norm=0.251, loss_final=0.473, loss_mean=0.805, loss_mean_cls=0.0703, proj_loss=-0.402][2026-03-26 12:59:24] Step: 6833, Training Logs: loss_final: 0.471774, loss_mean: 0.817545, proj_loss: -0.413149, loss_mean_cls: 0.067378, grad_norm: 0.228437 +Steps: 1%| | 6834/1000000 [28:21<68:16:43, 4.04it/s, grad_norm=0.228, loss_final=0.472, loss_mean=0.818, loss_mean_cls=0.0674, proj_loss=-0.413][2026-03-26 12:59:25] Step: 6834, Training Logs: loss_final: 0.463223, loss_mean: 0.812650, proj_loss: -0.418335, loss_mean_cls: 0.068908, grad_norm: 0.304305 +Steps: 1%| | 6835/1000000 [28:21<68:13:08, 4.04it/s, grad_norm=0.304, loss_final=0.463, loss_mean=0.813, loss_mean_cls=0.0689, proj_loss=-0.418][2026-03-26 12:59:25] Step: 6835, Training Logs: loss_final: 0.469066, loss_mean: 0.805985, proj_loss: -0.407388, loss_mean_cls: 0.070469, grad_norm: 0.206580 +Steps: 1%| | 6836/1000000 [28:22<68:15:08, 4.04it/s, grad_norm=0.207, loss_final=0.469, loss_mean=0.806, loss_mean_cls=0.0705, proj_loss=-0.407][2026-03-26 12:59:25] Step: 6836, Training Logs: loss_final: 0.468381, loss_mean: 0.812242, proj_loss: -0.412968, loss_mean_cls: 0.069108, grad_norm: 0.254492 +Steps: 1%| | 6837/1000000 [28:22<68:10:36, 4.05it/s, grad_norm=0.254, loss_final=0.468, loss_mean=0.812, loss_mean_cls=0.0691, proj_loss=-0.413][2026-03-26 12:59:25] Step: 6837, Training Logs: loss_final: 0.488489, loss_mean: 0.836416, proj_loss: -0.417397, loss_mean_cls: 0.069470, grad_norm: 0.292322 +Steps: 1%| | 6838/1000000 [28:22<68:11:47, 4.05it/s, grad_norm=0.292, loss_final=0.488, loss_mean=0.836, loss_mean_cls=0.0695, proj_loss=-0.417][2026-03-26 12:59:26] Step: 6838, Training Logs: loss_final: 0.471832, loss_mean: 0.822417, proj_loss: -0.418224, loss_mean_cls: 0.067638, grad_norm: 0.345763 +Steps: 1%| | 6839/1000000 [28:22<68:09:18, 4.05it/s, grad_norm=0.346, loss_final=0.472, loss_mean=0.822, loss_mean_cls=0.0676, proj_loss=-0.418][2026-03-26 12:59:26] Step: 6839, Training Logs: loss_final: 0.456879, loss_mean: 0.804786, proj_loss: -0.415978, loss_mean_cls: 0.068071, grad_norm: 0.324432 +Steps: 1%| | 6840/1000000 [28:23<68:12:39, 4.04it/s, grad_norm=0.324, loss_final=0.457, loss_mean=0.805, loss_mean_cls=0.0681, proj_loss=-0.416][2026-03-26 12:59:26] Step: 6840, Training Logs: loss_final: 0.492993, loss_mean: 0.843478, proj_loss: -0.417978, loss_mean_cls: 0.067494, grad_norm: 0.360911 +Steps: 1%| | 6841/1000000 [28:23<68:12:02, 4.05it/s, grad_norm=0.361, loss_final=0.493, loss_mean=0.843, loss_mean_cls=0.0675, proj_loss=-0.418][2026-03-26 12:59:26] Step: 6841, Training Logs: loss_final: 0.467823, loss_mean: 0.810269, proj_loss: -0.410490, loss_mean_cls: 0.068044, grad_norm: 0.298973 +Steps: 1%| | 6842/1000000 [28:23<68:13:37, 4.04it/s, grad_norm=0.299, loss_final=0.468, loss_mean=0.81, loss_mean_cls=0.068, proj_loss=-0.41][2026-03-26 12:59:27] Step: 6842, Training Logs: loss_final: 0.470715, loss_mean: 0.816579, proj_loss: -0.414514, loss_mean_cls: 0.068651, grad_norm: 0.236613 +Steps: 1%| | 6843/1000000 [28:23<68:13:31, 4.04it/s, grad_norm=0.237, loss_final=0.471, loss_mean=0.817, loss_mean_cls=0.0687, proj_loss=-0.415][2026-03-26 12:59:27] Step: 6843, Training Logs: loss_final: 0.457660, loss_mean: 0.798118, proj_loss: -0.409553, loss_mean_cls: 0.069094, grad_norm: 0.361908 +Steps: 1%| | 6844/1000000 [28:24<68:17:41, 4.04it/s, grad_norm=0.362, loss_final=0.458, loss_mean=0.798, loss_mean_cls=0.0691, proj_loss=-0.41][2026-03-26 12:59:27] Step: 6844, Training Logs: loss_final: 0.455792, loss_mean: 0.809004, proj_loss: -0.420994, loss_mean_cls: 0.067782, grad_norm: 0.323514 +Steps: 1%| | 6845/1000000 [28:24<68:20:03, 4.04it/s, grad_norm=0.324, loss_final=0.456, loss_mean=0.809, loss_mean_cls=0.0678, proj_loss=-0.421][2026-03-26 12:59:27] Step: 6845, Training Logs: loss_final: 0.468599, loss_mean: 0.818927, proj_loss: -0.418077, loss_mean_cls: 0.067749, grad_norm: 0.227995 +Steps: 1%| | 6846/1000000 [28:24<68:18:31, 4.04it/s, grad_norm=0.228, loss_final=0.469, loss_mean=0.819, loss_mean_cls=0.0677, proj_loss=-0.418][2026-03-26 12:59:28] Step: 6846, Training Logs: loss_final: 0.477881, loss_mean: 0.822864, proj_loss: -0.413357, loss_mean_cls: 0.068374, grad_norm: 0.301419 +Steps: 1%| | 6847/1000000 [28:24<68:17:43, 4.04it/s, grad_norm=0.301, loss_final=0.478, loss_mean=0.823, loss_mean_cls=0.0684, proj_loss=-0.413][2026-03-26 12:59:28] Step: 6847, Training Logs: loss_final: 0.453198, loss_mean: 0.802324, proj_loss: -0.416106, loss_mean_cls: 0.066979, grad_norm: 0.345787 +Steps: 1%| | 6848/1000000 [28:25<68:20:41, 4.04it/s, grad_norm=0.346, loss_final=0.453, loss_mean=0.802, loss_mean_cls=0.067, proj_loss=-0.416][2026-03-26 12:59:28] Step: 6848, Training Logs: loss_final: 0.471370, loss_mean: 0.811668, proj_loss: -0.408882, loss_mean_cls: 0.068584, grad_norm: 0.217287 +Steps: 1%| | 6849/1000000 [28:25<68:19:18, 4.04it/s, grad_norm=0.217, loss_final=0.471, loss_mean=0.812, loss_mean_cls=0.0686, proj_loss=-0.409][2026-03-26 12:59:28] Step: 6849, Training Logs: loss_final: 0.455018, loss_mean: 0.802438, proj_loss: -0.413953, loss_mean_cls: 0.066534, grad_norm: 0.264388 +Steps: 1%| | 6850/1000000 [28:25<68:20:06, 4.04it/s, grad_norm=0.264, loss_final=0.455, loss_mean=0.802, loss_mean_cls=0.0665, proj_loss=-0.414][2026-03-26 12:59:29] Step: 6850, Training Logs: loss_final: 0.464689, loss_mean: 0.812112, proj_loss: -0.415710, loss_mean_cls: 0.068287, grad_norm: 0.258327 +Steps: 1%| | 6851/1000000 [28:25<68:17:35, 4.04it/s, grad_norm=0.258, loss_final=0.465, loss_mean=0.812, loss_mean_cls=0.0683, proj_loss=-0.416][2026-03-26 12:59:29] Step: 6851, Training Logs: loss_final: 0.473682, loss_mean: 0.818750, proj_loss: -0.414108, loss_mean_cls: 0.069040, grad_norm: 0.397075 +Steps: 1%| | 6852/1000000 [28:26<68:12:54, 4.04it/s, grad_norm=0.397, loss_final=0.474, loss_mean=0.819, loss_mean_cls=0.069, proj_loss=-0.414][2026-03-26 12:59:29] Step: 6852, Training Logs: loss_final: 0.469947, loss_mean: 0.816787, proj_loss: -0.414889, loss_mean_cls: 0.068048, grad_norm: 0.252912 +Steps: 1%| | 6853/1000000 [28:26<68:18:24, 4.04it/s, grad_norm=0.253, loss_final=0.47, loss_mean=0.817, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 12:59:29] Step: 6853, Training Logs: loss_final: 0.455297, loss_mean: 0.802855, proj_loss: -0.415626, loss_mean_cls: 0.068067, grad_norm: 0.437400 +Steps: 1%| | 6854/1000000 [28:26<68:15:47, 4.04it/s, grad_norm=0.437, loss_final=0.455, loss_mean=0.803, loss_mean_cls=0.0681, proj_loss=-0.416][2026-03-26 12:59:29] Step: 6854, Training Logs: loss_final: 0.479674, loss_mean: 0.828699, proj_loss: -0.416568, loss_mean_cls: 0.067543, grad_norm: 0.415741 +Steps: 1%| | 6855/1000000 [28:26<68:13:55, 4.04it/s, grad_norm=0.416, loss_final=0.48, loss_mean=0.829, loss_mean_cls=0.0675, proj_loss=-0.417][2026-03-26 12:59:30] Step: 6855, Training Logs: loss_final: 0.486996, loss_mean: 0.839921, proj_loss: -0.419345, loss_mean_cls: 0.066420, grad_norm: 0.250811 +Steps: 1%| | 6856/1000000 [28:27<68:14:38, 4.04it/s, grad_norm=0.251, loss_final=0.487, loss_mean=0.84, loss_mean_cls=0.0664, proj_loss=-0.419][2026-03-26 12:59:30] Step: 6856, Training Logs: loss_final: 0.481690, loss_mean: 0.830853, proj_loss: -0.416325, loss_mean_cls: 0.067163, grad_norm: 0.348952 +Steps: 1%| | 6857/1000000 [28:27<68:16:52, 4.04it/s, grad_norm=0.349, loss_final=0.482, loss_mean=0.831, loss_mean_cls=0.0672, proj_loss=-0.416][2026-03-26 12:59:30] Step: 6857, Training Logs: loss_final: 0.468978, loss_mean: 0.822405, proj_loss: -0.420680, loss_mean_cls: 0.067253, grad_norm: 0.346977 +Steps: 1%| | 6858/1000000 [28:27<68:19:29, 4.04it/s, grad_norm=0.347, loss_final=0.469, loss_mean=0.822, loss_mean_cls=0.0673, proj_loss=-0.421][2026-03-26 12:59:30] Step: 6858, Training Logs: loss_final: 0.458387, loss_mean: 0.799149, proj_loss: -0.410625, loss_mean_cls: 0.069863, grad_norm: 0.417532 +Steps: 1%| | 6859/1000000 [28:27<68:21:01, 4.04it/s, grad_norm=0.418, loss_final=0.458, loss_mean=0.799, loss_mean_cls=0.0699, proj_loss=-0.411][2026-03-26 12:59:31] Step: 6859, Training Logs: loss_final: 0.460904, loss_mean: 0.795218, proj_loss: -0.405630, loss_mean_cls: 0.071315, grad_norm: 0.264261 +Steps: 1%| | 6860/1000000 [28:28<68:20:49, 4.04it/s, grad_norm=0.264, loss_final=0.461, loss_mean=0.795, loss_mean_cls=0.0713, proj_loss=-0.406][2026-03-26 12:59:31] Step: 6860, Training Logs: loss_final: 0.481471, loss_mean: 0.832182, proj_loss: -0.417730, loss_mean_cls: 0.067019, grad_norm: 0.263807 +Steps: 1%| | 6861/1000000 [28:28<68:20:29, 4.04it/s, grad_norm=0.264, loss_final=0.481, loss_mean=0.832, loss_mean_cls=0.067, proj_loss=-0.418][2026-03-26 12:59:31] Step: 6861, Training Logs: loss_final: 0.475106, loss_mean: 0.824945, proj_loss: -0.416072, loss_mean_cls: 0.066234, grad_norm: 0.239717 +Steps: 1%| | 6862/1000000 [28:28<68:24:34, 4.03it/s, grad_norm=0.24, loss_final=0.475, loss_mean=0.825, loss_mean_cls=0.0662, proj_loss=-0.416][2026-03-26 12:59:31] Step: 6862, Training Logs: loss_final: 0.473904, loss_mean: 0.831732, proj_loss: -0.423107, loss_mean_cls: 0.065280, grad_norm: 0.228373 +Steps: 1%| | 6863/1000000 [28:28<68:30:05, 4.03it/s, grad_norm=0.228, loss_final=0.474, loss_mean=0.832, loss_mean_cls=0.0653, proj_loss=-0.423][2026-03-26 12:59:32] Step: 6863, Training Logs: loss_final: 0.485245, loss_mean: 0.828813, proj_loss: -0.412790, loss_mean_cls: 0.069222, grad_norm: 0.231765 +Steps: 1%| | 6864/1000000 [28:29<69:11:38, 3.99it/s, grad_norm=0.232, loss_final=0.485, loss_mean=0.829, loss_mean_cls=0.0692, proj_loss=-0.413][2026-03-26 12:59:32] Step: 6864, Training Logs: loss_final: 0.484661, loss_mean: 0.828305, proj_loss: -0.411098, loss_mean_cls: 0.067454, grad_norm: 0.297979 +Steps: 1%| | 6865/1000000 [28:29<69:13:49, 3.98it/s, grad_norm=0.298, loss_final=0.485, loss_mean=0.828, loss_mean_cls=0.0675, proj_loss=-0.411][2026-03-26 12:59:32] Step: 6865, Training Logs: loss_final: 0.471885, loss_mean: 0.816014, proj_loss: -0.412591, loss_mean_cls: 0.068461, grad_norm: 0.256072 +Steps: 1%| | 6866/1000000 [28:29<68:59:04, 4.00it/s, grad_norm=0.256, loss_final=0.472, loss_mean=0.816, loss_mean_cls=0.0685, proj_loss=-0.413][2026-03-26 12:59:32] Step: 6866, Training Logs: loss_final: 0.475574, loss_mean: 0.828590, proj_loss: -0.420609, loss_mean_cls: 0.067594, grad_norm: 0.269064 +Steps: 1%| | 6867/1000000 [28:29<68:45:16, 4.01it/s, grad_norm=0.269, loss_final=0.476, loss_mean=0.829, loss_mean_cls=0.0676, proj_loss=-0.421][2026-03-26 12:59:33] Step: 6867, Training Logs: loss_final: 0.479557, loss_mean: 0.846184, proj_loss: -0.431096, loss_mean_cls: 0.064469, grad_norm: 0.286454 +Steps: 1%| | 6868/1000000 [28:30<68:39:17, 4.02it/s, grad_norm=0.286, loss_final=0.48, loss_mean=0.846, loss_mean_cls=0.0645, proj_loss=-0.431][2026-03-26 12:59:33] Step: 6868, Training Logs: loss_final: 0.476838, loss_mean: 0.817673, proj_loss: -0.410455, loss_mean_cls: 0.069620, grad_norm: 0.335291 +Steps: 1%| | 6869/1000000 [28:30<68:34:15, 4.02it/s, grad_norm=0.335, loss_final=0.477, loss_mean=0.818, loss_mean_cls=0.0696, proj_loss=-0.41][2026-03-26 12:59:33] Step: 6869, Training Logs: loss_final: 0.455965, loss_mean: 0.810887, proj_loss: -0.420736, loss_mean_cls: 0.065814, grad_norm: 0.324658 +Steps: 1%| | 6870/1000000 [28:30<68:26:28, 4.03it/s, grad_norm=0.325, loss_final=0.456, loss_mean=0.811, loss_mean_cls=0.0658, proj_loss=-0.421][2026-03-26 12:59:33] Step: 6870, Training Logs: loss_final: 0.464834, loss_mean: 0.804216, proj_loss: -0.409088, loss_mean_cls: 0.069706, grad_norm: 0.458905 +Steps: 1%| | 6871/1000000 [28:30<68:27:25, 4.03it/s, grad_norm=0.459, loss_final=0.465, loss_mean=0.804, loss_mean_cls=0.0697, proj_loss=-0.409][2026-03-26 12:59:34] Step: 6871, Training Logs: loss_final: 0.456909, loss_mean: 0.819068, proj_loss: -0.427710, loss_mean_cls: 0.065551, grad_norm: 0.272475 +Steps: 1%| | 6872/1000000 [28:31<68:21:49, 4.04it/s, grad_norm=0.272, loss_final=0.457, loss_mean=0.819, loss_mean_cls=0.0656, proj_loss=-0.428][2026-03-26 12:59:34] Step: 6872, Training Logs: loss_final: 0.470160, loss_mean: 0.814377, proj_loss: -0.412742, loss_mean_cls: 0.068525, grad_norm: 0.356596 +Steps: 1%| | 6873/1000000 [28:31<68:20:47, 4.04it/s, grad_norm=0.357, loss_final=0.47, loss_mean=0.814, loss_mean_cls=0.0685, proj_loss=-0.413][2026-03-26 12:59:34] Step: 6873, Training Logs: loss_final: 0.468295, loss_mean: 0.820365, proj_loss: -0.419264, loss_mean_cls: 0.067194, grad_norm: 0.344875 +Steps: 1%| | 6874/1000000 [28:31<68:19:08, 4.04it/s, grad_norm=0.345, loss_final=0.468, loss_mean=0.82, loss_mean_cls=0.0672, proj_loss=-0.419][2026-03-26 12:59:34] Step: 6874, Training Logs: loss_final: 0.469728, loss_mean: 0.816857, proj_loss: -0.415941, loss_mean_cls: 0.068812, grad_norm: 0.365709 +Steps: 1%| | 6875/1000000 [28:31<68:18:04, 4.04it/s, grad_norm=0.366, loss_final=0.47, loss_mean=0.817, loss_mean_cls=0.0688, proj_loss=-0.416][2026-03-26 12:59:35] Step: 6875, Training Logs: loss_final: 0.475428, loss_mean: 0.828602, proj_loss: -0.420142, loss_mean_cls: 0.066968, grad_norm: 0.388533 +Steps: 1%| | 6876/1000000 [28:32<68:18:25, 4.04it/s, grad_norm=0.389, loss_final=0.475, loss_mean=0.829, loss_mean_cls=0.067, proj_loss=-0.42][2026-03-26 12:59:35] Step: 6876, Training Logs: loss_final: 0.489146, loss_mean: 0.835460, proj_loss: -0.415299, loss_mean_cls: 0.068985, grad_norm: 0.262360 +Steps: 1%| | 6877/1000000 [28:32<68:19:47, 4.04it/s, grad_norm=0.262, loss_final=0.489, loss_mean=0.835, loss_mean_cls=0.069, proj_loss=-0.415][2026-03-26 12:59:35] Step: 6877, Training Logs: loss_final: 0.473555, loss_mean: 0.815807, proj_loss: -0.411037, loss_mean_cls: 0.068786, grad_norm: 0.289188 +Steps: 1%| | 6878/1000000 [28:32<68:21:02, 4.04it/s, grad_norm=0.289, loss_final=0.474, loss_mean=0.816, loss_mean_cls=0.0688, proj_loss=-0.411][2026-03-26 12:59:35] Step: 6878, Training Logs: loss_final: 0.468348, loss_mean: 0.824277, proj_loss: -0.422518, loss_mean_cls: 0.066589, grad_norm: 0.221442 +Steps: 1%| | 6879/1000000 [28:32<68:22:27, 4.03it/s, grad_norm=0.221, loss_final=0.468, loss_mean=0.824, loss_mean_cls=0.0666, proj_loss=-0.423][2026-03-26 12:59:36] Step: 6879, Training Logs: loss_final: 0.474921, loss_mean: 0.822591, proj_loss: -0.415988, loss_mean_cls: 0.068318, grad_norm: 0.310062 +Steps: 1%| | 6880/1000000 [28:33<68:22:49, 4.03it/s, grad_norm=0.31, loss_final=0.475, loss_mean=0.823, loss_mean_cls=0.0683, proj_loss=-0.416][2026-03-26 12:59:36] Step: 6880, Training Logs: loss_final: 0.466719, loss_mean: 0.812128, proj_loss: -0.414081, loss_mean_cls: 0.068672, grad_norm: 0.255525 +Steps: 1%| | 6881/1000000 [28:33<68:25:30, 4.03it/s, grad_norm=0.256, loss_final=0.467, loss_mean=0.812, loss_mean_cls=0.0687, proj_loss=-0.414][2026-03-26 12:59:36] Step: 6881, Training Logs: loss_final: 0.494140, loss_mean: 0.848088, proj_loss: -0.420524, loss_mean_cls: 0.066576, grad_norm: 0.480424 +Steps: 1%| | 6882/1000000 [28:33<68:24:27, 4.03it/s, grad_norm=0.48, loss_final=0.494, loss_mean=0.848, loss_mean_cls=0.0666, proj_loss=-0.421][2026-03-26 12:59:36] Step: 6882, Training Logs: loss_final: 0.480781, loss_mean: 0.817111, proj_loss: -0.406633, loss_mean_cls: 0.070303, grad_norm: 0.394517 +Steps: 1%| | 6883/1000000 [28:33<68:25:59, 4.03it/s, grad_norm=0.395, loss_final=0.481, loss_mean=0.817, loss_mean_cls=0.0703, proj_loss=-0.407][2026-03-26 12:59:37] Step: 6883, Training Logs: loss_final: 0.457675, loss_mean: 0.817853, proj_loss: -0.426343, loss_mean_cls: 0.066165, grad_norm: 0.400868 +Steps: 1%| | 6884/1000000 [28:34<68:18:14, 4.04it/s, grad_norm=0.401, loss_final=0.458, loss_mean=0.818, loss_mean_cls=0.0662, proj_loss=-0.426][2026-03-26 12:59:37] Step: 6884, Training Logs: loss_final: 0.486544, loss_mean: 0.830267, proj_loss: -0.412974, loss_mean_cls: 0.069252, grad_norm: 0.561232 +Steps: 1%| | 6885/1000000 [28:34<68:13:33, 4.04it/s, grad_norm=0.561, loss_final=0.487, loss_mean=0.83, loss_mean_cls=0.0693, proj_loss=-0.413][2026-03-26 12:59:37] Step: 6885, Training Logs: loss_final: 0.469587, loss_mean: 0.811067, proj_loss: -0.410685, loss_mean_cls: 0.069205, grad_norm: 0.270874 +Steps: 1%| | 6886/1000000 [28:34<68:14:37, 4.04it/s, grad_norm=0.271, loss_final=0.47, loss_mean=0.811, loss_mean_cls=0.0692, proj_loss=-0.411][2026-03-26 12:59:37] Step: 6886, Training Logs: loss_final: 0.487846, loss_mean: 0.824691, proj_loss: -0.408192, loss_mean_cls: 0.071347, grad_norm: 0.483358 +Steps: 1%| | 6887/1000000 [28:34<68:12:31, 4.04it/s, grad_norm=0.483, loss_final=0.488, loss_mean=0.825, loss_mean_cls=0.0713, proj_loss=-0.408][2026-03-26 12:59:38] Step: 6887, Training Logs: loss_final: 0.478042, loss_mean: 0.822485, proj_loss: -0.411879, loss_mean_cls: 0.067435, grad_norm: 0.341477 +Steps: 1%| | 6888/1000000 [28:35<68:15:38, 4.04it/s, grad_norm=0.341, loss_final=0.478, loss_mean=0.822, loss_mean_cls=0.0674, proj_loss=-0.412][2026-03-26 12:59:38] Step: 6888, Training Logs: loss_final: 0.480029, loss_mean: 0.841312, proj_loss: -0.426962, loss_mean_cls: 0.065679, grad_norm: 0.533764 +Steps: 1%| | 6889/1000000 [28:35<68:12:42, 4.04it/s, grad_norm=0.534, loss_final=0.48, loss_mean=0.841, loss_mean_cls=0.0657, proj_loss=-0.427][2026-03-26 12:59:38] Step: 6889, Training Logs: loss_final: 0.464464, loss_mean: 0.808193, proj_loss: -0.412202, loss_mean_cls: 0.068473, grad_norm: 0.350739 +Steps: 1%| | 6890/1000000 [28:35<68:18:46, 4.04it/s, grad_norm=0.351, loss_final=0.464, loss_mean=0.808, loss_mean_cls=0.0685, proj_loss=-0.412][2026-03-26 12:59:38] Step: 6890, Training Logs: loss_final: 0.471492, loss_mean: 0.803542, proj_loss: -0.402710, loss_mean_cls: 0.070661, grad_norm: 0.518580 +Steps: 1%| | 6891/1000000 [28:35<68:18:22, 4.04it/s, grad_norm=0.519, loss_final=0.471, loss_mean=0.804, loss_mean_cls=0.0707, proj_loss=-0.403][2026-03-26 12:59:39] Step: 6891, Training Logs: loss_final: 0.463633, loss_mean: 0.815539, proj_loss: -0.419864, loss_mean_cls: 0.067958, grad_norm: 0.328896 +Steps: 1%| | 6892/1000000 [28:36<68:18:31, 4.04it/s, grad_norm=0.329, loss_final=0.464, loss_mean=0.816, loss_mean_cls=0.068, proj_loss=-0.42][2026-03-26 12:59:39] Step: 6892, Training Logs: loss_final: 0.451601, loss_mean: 0.796649, proj_loss: -0.414018, loss_mean_cls: 0.068970, grad_norm: 0.463376 +Steps: 1%| | 6893/1000000 [28:36<68:15:59, 4.04it/s, grad_norm=0.463, loss_final=0.452, loss_mean=0.797, loss_mean_cls=0.069, proj_loss=-0.414][2026-03-26 12:59:39] Step: 6893, Training Logs: loss_final: 0.474045, loss_mean: 0.817525, proj_loss: -0.412465, loss_mean_cls: 0.068985, grad_norm: 0.392546 +Steps: 1%| | 6894/1000000 [28:36<68:17:09, 4.04it/s, grad_norm=0.393, loss_final=0.474, loss_mean=0.818, loss_mean_cls=0.069, proj_loss=-0.412][2026-03-26 12:59:39] Step: 6894, Training Logs: loss_final: 0.462776, loss_mean: 0.810524, proj_loss: -0.416475, loss_mean_cls: 0.068727, grad_norm: 0.394531 +Steps: 1%| | 6895/1000000 [28:36<68:17:43, 4.04it/s, grad_norm=0.395, loss_final=0.463, loss_mean=0.811, loss_mean_cls=0.0687, proj_loss=-0.416][2026-03-26 12:59:40] Step: 6895, Training Logs: loss_final: 0.458686, loss_mean: 0.806159, proj_loss: -0.415314, loss_mean_cls: 0.067840, grad_norm: 0.407232 +Steps: 1%| | 6896/1000000 [28:37<68:17:41, 4.04it/s, grad_norm=0.407, loss_final=0.459, loss_mean=0.806, loss_mean_cls=0.0678, proj_loss=-0.415][2026-03-26 12:59:40] Step: 6896, Training Logs: loss_final: 0.467558, loss_mean: 0.811026, proj_loss: -0.413321, loss_mean_cls: 0.069854, grad_norm: 0.314312 +Steps: 1%| | 6897/1000000 [28:37<68:20:02, 4.04it/s, grad_norm=0.314, loss_final=0.468, loss_mean=0.811, loss_mean_cls=0.0699, proj_loss=-0.413][2026-03-26 12:59:40] Step: 6897, Training Logs: loss_final: 0.477115, loss_mean: 0.828119, proj_loss: -0.418383, loss_mean_cls: 0.067378, grad_norm: 0.516557 +Steps: 1%| | 6898/1000000 [28:37<68:20:46, 4.04it/s, grad_norm=0.517, loss_final=0.477, loss_mean=0.828, loss_mean_cls=0.0674, proj_loss=-0.418][2026-03-26 12:59:40] Step: 6898, Training Logs: loss_final: 0.467519, loss_mean: 0.811841, proj_loss: -0.412626, loss_mean_cls: 0.068305, grad_norm: 0.258190 +Steps: 1%| | 6899/1000000 [28:37<68:16:53, 4.04it/s, grad_norm=0.258, loss_final=0.468, loss_mean=0.812, loss_mean_cls=0.0683, proj_loss=-0.413][2026-03-26 12:59:41] Step: 6899, Training Logs: loss_final: 0.496210, loss_mean: 0.833804, proj_loss: -0.407501, loss_mean_cls: 0.069906, grad_norm: 0.390260 +Steps: 1%| | 6900/1000000 [28:38<68:15:36, 4.04it/s, grad_norm=0.39, loss_final=0.496, loss_mean=0.834, loss_mean_cls=0.0699, proj_loss=-0.408][2026-03-26 12:59:41] Step: 6900, Training Logs: loss_final: 0.479213, loss_mean: 0.826032, proj_loss: -0.415002, loss_mean_cls: 0.068183, grad_norm: 0.231476 +Steps: 1%| | 6901/1000000 [28:38<68:20:32, 4.04it/s, grad_norm=0.231, loss_final=0.479, loss_mean=0.826, loss_mean_cls=0.0682, proj_loss=-0.415][2026-03-26 12:59:41] Step: 6901, Training Logs: loss_final: 0.490148, loss_mean: 0.844248, proj_loss: -0.421236, loss_mean_cls: 0.067136, grad_norm: 0.431469 +Steps: 1%| | 6902/1000000 [28:38<68:19:01, 4.04it/s, grad_norm=0.431, loss_final=0.49, loss_mean=0.844, loss_mean_cls=0.0671, proj_loss=-0.421][2026-03-26 12:59:41] Step: 6902, Training Logs: loss_final: 0.487128, loss_mean: 0.822804, proj_loss: -0.405789, loss_mean_cls: 0.070113, grad_norm: 0.312411 +Steps: 1%| | 6903/1000000 [28:38<68:20:59, 4.04it/s, grad_norm=0.312, loss_final=0.487, loss_mean=0.823, loss_mean_cls=0.0701, proj_loss=-0.406][2026-03-26 12:59:42] Step: 6903, Training Logs: loss_final: 0.454691, loss_mean: 0.796966, proj_loss: -0.411160, loss_mean_cls: 0.068884, grad_norm: 0.261224 +Steps: 1%| | 6904/1000000 [28:38<68:18:28, 4.04it/s, grad_norm=0.261, loss_final=0.455, loss_mean=0.797, loss_mean_cls=0.0689, proj_loss=-0.411][2026-03-26 12:59:42] Step: 6904, Training Logs: loss_final: 0.465744, loss_mean: 0.814224, proj_loss: -0.416354, loss_mean_cls: 0.067873, grad_norm: 0.422335 +Steps: 1%| | 6905/1000000 [28:39<68:18:04, 4.04it/s, grad_norm=0.422, loss_final=0.466, loss_mean=0.814, loss_mean_cls=0.0679, proj_loss=-0.416][2026-03-26 12:59:42] Step: 6905, Training Logs: loss_final: 0.482729, loss_mean: 0.827284, proj_loss: -0.412311, loss_mean_cls: 0.067756, grad_norm: 0.313627 +Steps: 1%| | 6906/1000000 [28:39<68:14:49, 4.04it/s, grad_norm=0.314, loss_final=0.483, loss_mean=0.827, loss_mean_cls=0.0678, proj_loss=-0.412][2026-03-26 12:59:42] Step: 6906, Training Logs: loss_final: 0.476670, loss_mean: 0.831265, proj_loss: -0.420420, loss_mean_cls: 0.065825, grad_norm: 0.351110 +Steps: 1%| | 6907/1000000 [28:39<68:15:41, 4.04it/s, grad_norm=0.351, loss_final=0.477, loss_mean=0.831, loss_mean_cls=0.0658, proj_loss=-0.42][2026-03-26 12:59:43] Step: 6907, Training Logs: loss_final: 0.458370, loss_mean: 0.806316, proj_loss: -0.416695, loss_mean_cls: 0.068749, grad_norm: 0.553273 +Steps: 1%| | 6908/1000000 [28:39<68:15:03, 4.04it/s, grad_norm=0.553, loss_final=0.458, loss_mean=0.806, loss_mean_cls=0.0687, proj_loss=-0.417][2026-03-26 12:59:43] Step: 6908, Training Logs: loss_final: 0.480260, loss_mean: 0.829195, proj_loss: -0.416885, loss_mean_cls: 0.067950, grad_norm: 0.297250 +Steps: 1%| | 6909/1000000 [28:40<68:14:24, 4.04it/s, grad_norm=0.297, loss_final=0.48, loss_mean=0.829, loss_mean_cls=0.0679, proj_loss=-0.417][2026-03-26 12:59:43] Step: 6909, Training Logs: loss_final: 0.465147, loss_mean: 0.801316, proj_loss: -0.405483, loss_mean_cls: 0.069314, grad_norm: 0.643299 +Steps: 1%| | 6910/1000000 [28:40<68:19:01, 4.04it/s, grad_norm=0.643, loss_final=0.465, loss_mean=0.801, loss_mean_cls=0.0693, proj_loss=-0.405][2026-03-26 12:59:43] Step: 6910, Training Logs: loss_final: 0.457225, loss_mean: 0.811934, proj_loss: -0.421655, loss_mean_cls: 0.066946, grad_norm: 0.725775 +Steps: 1%| | 6911/1000000 [28:40<68:17:37, 4.04it/s, grad_norm=0.726, loss_final=0.457, loss_mean=0.812, loss_mean_cls=0.0669, proj_loss=-0.422][2026-03-26 12:59:44] Step: 6911, Training Logs: loss_final: 0.486722, loss_mean: 0.829754, proj_loss: -0.411775, loss_mean_cls: 0.068744, grad_norm: 0.354305 +Steps: 1%| | 6912/1000000 [28:40<68:18:24, 4.04it/s, grad_norm=0.354, loss_final=0.487, loss_mean=0.83, loss_mean_cls=0.0687, proj_loss=-0.412][2026-03-26 12:59:44] Step: 6912, Training Logs: loss_final: 0.465572, loss_mean: 0.801830, proj_loss: -0.406952, loss_mean_cls: 0.070694, grad_norm: 0.613046 +Steps: 1%| | 6913/1000000 [28:41<68:22:01, 4.03it/s, grad_norm=0.613, loss_final=0.466, loss_mean=0.802, loss_mean_cls=0.0707, proj_loss=-0.407][2026-03-26 12:59:44] Step: 6913, Training Logs: loss_final: 0.467644, loss_mean: 0.812748, proj_loss: -0.413825, loss_mean_cls: 0.068721, grad_norm: 0.398618 +Steps: 1%| | 6914/1000000 [28:41<68:59:37, 4.00it/s, grad_norm=0.399, loss_final=0.468, loss_mean=0.813, loss_mean_cls=0.0687, proj_loss=-0.414][2026-03-26 12:59:44] Step: 6914, Training Logs: loss_final: 0.461731, loss_mean: 0.812261, proj_loss: -0.417186, loss_mean_cls: 0.066656, grad_norm: 0.631979 +Steps: 1%| | 6915/1000000 [28:41<68:50:02, 4.01it/s, grad_norm=0.632, loss_final=0.462, loss_mean=0.812, loss_mean_cls=0.0667, proj_loss=-0.417][2026-03-26 12:59:45] Step: 6915, Training Logs: loss_final: 0.467182, loss_mean: 0.800937, proj_loss: -0.404254, loss_mean_cls: 0.070499, grad_norm: 0.525912 +Steps: 1%| | 6916/1000000 [28:41<68:42:47, 4.01it/s, grad_norm=0.526, loss_final=0.467, loss_mean=0.801, loss_mean_cls=0.0705, proj_loss=-0.404][2026-03-26 12:59:45] Step: 6916, Training Logs: loss_final: 0.458586, loss_mean: 0.804108, proj_loss: -0.413535, loss_mean_cls: 0.068013, grad_norm: 0.401793 +Steps: 1%| | 6917/1000000 [28:42<68:36:03, 4.02it/s, grad_norm=0.402, loss_final=0.459, loss_mean=0.804, loss_mean_cls=0.068, proj_loss=-0.414][2026-03-26 12:59:45] Step: 6917, Training Logs: loss_final: 0.464924, loss_mean: 0.804259, proj_loss: -0.409412, loss_mean_cls: 0.070077, grad_norm: 0.396194 +Steps: 1%| | 6918/1000000 [28:42<68:29:09, 4.03it/s, grad_norm=0.396, loss_final=0.465, loss_mean=0.804, loss_mean_cls=0.0701, proj_loss=-0.409][2026-03-26 12:59:45] Step: 6918, Training Logs: loss_final: 0.463235, loss_mean: 0.810254, proj_loss: -0.415902, loss_mean_cls: 0.068883, grad_norm: 0.500755 +Steps: 1%| | 6919/1000000 [28:42<68:36:07, 4.02it/s, grad_norm=0.501, loss_final=0.463, loss_mean=0.81, loss_mean_cls=0.0689, proj_loss=-0.416][2026-03-26 12:59:46] Step: 6919, Training Logs: loss_final: 0.456173, loss_mean: 0.799623, proj_loss: -0.411980, loss_mean_cls: 0.068530, grad_norm: 0.478866 +Steps: 1%| | 6920/1000000 [28:42<69:08:38, 3.99it/s, grad_norm=0.479, loss_final=0.456, loss_mean=0.8, loss_mean_cls=0.0685, proj_loss=-0.412][2026-03-26 12:59:46] Step: 6920, Training Logs: loss_final: 0.464639, loss_mean: 0.826363, proj_loss: -0.427214, loss_mean_cls: 0.065491, grad_norm: 0.369121 +Steps: 1%| | 6921/1000000 [28:43<68:53:42, 4.00it/s, grad_norm=0.369, loss_final=0.465, loss_mean=0.826, loss_mean_cls=0.0655, proj_loss=-0.427][2026-03-26 12:59:46] Step: 6921, Training Logs: loss_final: 0.476632, loss_mean: 0.820261, proj_loss: -0.412026, loss_mean_cls: 0.068396, grad_norm: 0.498718 +Steps: 1%| | 6922/1000000 [28:43<68:41:48, 4.02it/s, grad_norm=0.499, loss_final=0.477, loss_mean=0.82, loss_mean_cls=0.0684, proj_loss=-0.412][2026-03-26 12:59:46] Step: 6922, Training Logs: loss_final: 0.479245, loss_mean: 0.831372, proj_loss: -0.419182, loss_mean_cls: 0.067055, grad_norm: 0.298024 +Steps: 1%| | 6923/1000000 [28:43<68:31:00, 4.03it/s, grad_norm=0.298, loss_final=0.479, loss_mean=0.831, loss_mean_cls=0.0671, proj_loss=-0.419][2026-03-26 12:59:47] Step: 6923, Training Logs: loss_final: 0.477291, loss_mean: 0.817649, proj_loss: -0.409236, loss_mean_cls: 0.068878, grad_norm: 0.386160 +Steps: 1%| | 6924/1000000 [28:43<68:29:44, 4.03it/s, grad_norm=0.386, loss_final=0.477, loss_mean=0.818, loss_mean_cls=0.0689, proj_loss=-0.409][2026-03-26 12:59:47] Step: 6924, Training Logs: loss_final: 0.476885, loss_mean: 0.829267, proj_loss: -0.419120, loss_mean_cls: 0.066738, grad_norm: 0.497391 +Steps: 1%| | 6925/1000000 [28:44<68:26:13, 4.03it/s, grad_norm=0.497, loss_final=0.477, loss_mean=0.829, loss_mean_cls=0.0667, proj_loss=-0.419][2026-03-26 12:59:47] Step: 6925, Training Logs: loss_final: 0.471945, loss_mean: 0.812955, proj_loss: -0.410562, loss_mean_cls: 0.069552, grad_norm: 0.713372 +Steps: 1%| | 6926/1000000 [28:44<68:24:58, 4.03it/s, grad_norm=0.713, loss_final=0.472, loss_mean=0.813, loss_mean_cls=0.0696, proj_loss=-0.411][2026-03-26 12:59:47] Step: 6926, Training Logs: loss_final: 0.479393, loss_mean: 0.820567, proj_loss: -0.410130, loss_mean_cls: 0.068956, grad_norm: 0.266561 +Steps: 1%| | 6927/1000000 [28:44<68:22:23, 4.03it/s, grad_norm=0.267, loss_final=0.479, loss_mean=0.821, loss_mean_cls=0.069, proj_loss=-0.41][2026-03-26 12:59:48] Step: 6927, Training Logs: loss_final: 0.477307, loss_mean: 0.805212, proj_loss: -0.399379, loss_mean_cls: 0.071474, grad_norm: 0.471964 +Steps: 1%| | 6928/1000000 [28:44<68:25:02, 4.03it/s, grad_norm=0.472, loss_final=0.477, loss_mean=0.805, loss_mean_cls=0.0715, proj_loss=-0.399][2026-03-26 12:59:48] Step: 6928, Training Logs: loss_final: 0.488422, loss_mean: 0.816569, proj_loss: -0.400199, loss_mean_cls: 0.072051, grad_norm: 0.395693 +Steps: 1%| | 6929/1000000 [28:45<68:21:32, 4.04it/s, grad_norm=0.396, loss_final=0.488, loss_mean=0.817, loss_mean_cls=0.0721, proj_loss=-0.4][2026-03-26 12:59:48] Step: 6929, Training Logs: loss_final: 0.488160, loss_mean: 0.837281, proj_loss: -0.416911, loss_mean_cls: 0.067790, grad_norm: 0.584661 +Steps: 1%| | 6930/1000000 [28:45<68:19:25, 4.04it/s, grad_norm=0.585, loss_final=0.488, loss_mean=0.837, loss_mean_cls=0.0678, proj_loss=-0.417][2026-03-26 12:59:48] Step: 6930, Training Logs: loss_final: 0.482384, loss_mean: 0.831483, proj_loss: -0.416026, loss_mean_cls: 0.066927, grad_norm: 0.278595 +Steps: 1%| | 6931/1000000 [28:45<68:16:55, 4.04it/s, grad_norm=0.279, loss_final=0.482, loss_mean=0.831, loss_mean_cls=0.0669, proj_loss=-0.416][2026-03-26 12:59:49] Step: 6931, Training Logs: loss_final: 0.471531, loss_mean: 0.819243, proj_loss: -0.415324, loss_mean_cls: 0.067613, grad_norm: 0.464930 +Steps: 1%| | 6932/1000000 [28:45<68:15:18, 4.04it/s, grad_norm=0.465, loss_final=0.472, loss_mean=0.819, loss_mean_cls=0.0676, proj_loss=-0.415][2026-03-26 12:59:49] Step: 6932, Training Logs: loss_final: 0.476003, loss_mean: 0.810487, proj_loss: -0.404997, loss_mean_cls: 0.070513, grad_norm: 0.393233 +Steps: 1%| | 6933/1000000 [28:46<68:10:54, 4.05it/s, grad_norm=0.393, loss_final=0.476, loss_mean=0.81, loss_mean_cls=0.0705, proj_loss=-0.405][2026-03-26 12:59:49] Step: 6933, Training Logs: loss_final: 0.454526, loss_mean: 0.800933, proj_loss: -0.414609, loss_mean_cls: 0.068201, grad_norm: 0.371389 +Steps: 1%| | 6934/1000000 [28:46<68:18:02, 4.04it/s, grad_norm=0.371, loss_final=0.455, loss_mean=0.801, loss_mean_cls=0.0682, proj_loss=-0.415][2026-03-26 12:59:49] Step: 6934, Training Logs: loss_final: 0.474718, loss_mean: 0.819473, proj_loss: -0.412651, loss_mean_cls: 0.067895, grad_norm: 0.372851 +Steps: 1%| | 6935/1000000 [28:46<68:16:00, 4.04it/s, grad_norm=0.373, loss_final=0.475, loss_mean=0.819, loss_mean_cls=0.0679, proj_loss=-0.413][2026-03-26 12:59:50] Step: 6935, Training Logs: loss_final: 0.478805, loss_mean: 0.826997, proj_loss: -0.415566, loss_mean_cls: 0.067374, grad_norm: 0.285665 +Steps: 1%| | 6936/1000000 [28:46<68:22:59, 4.03it/s, grad_norm=0.286, loss_final=0.479, loss_mean=0.827, loss_mean_cls=0.0674, proj_loss=-0.416][2026-03-26 12:59:50] Step: 6936, Training Logs: loss_final: 0.455858, loss_mean: 0.799208, proj_loss: -0.411687, loss_mean_cls: 0.068337, grad_norm: 0.265927 +Steps: 1%| | 6937/1000000 [28:47<68:16:36, 4.04it/s, grad_norm=0.266, loss_final=0.456, loss_mean=0.799, loss_mean_cls=0.0683, proj_loss=-0.412][2026-03-26 12:59:50] Step: 6937, Training Logs: loss_final: 0.483922, loss_mean: 0.841294, proj_loss: -0.422278, loss_mean_cls: 0.064906, grad_norm: 0.348131 +Steps: 1%| | 6938/1000000 [28:47<68:17:08, 4.04it/s, grad_norm=0.348, loss_final=0.484, loss_mean=0.841, loss_mean_cls=0.0649, proj_loss=-0.422][2026-03-26 12:59:50] Step: 6938, Training Logs: loss_final: 0.460363, loss_mean: 0.813469, proj_loss: -0.420356, loss_mean_cls: 0.067251, grad_norm: 0.214407 +Steps: 1%| | 6939/1000000 [28:47<68:17:13, 4.04it/s, grad_norm=0.214, loss_final=0.46, loss_mean=0.813, loss_mean_cls=0.0673, proj_loss=-0.42][2026-03-26 12:59:51] Step: 6939, Training Logs: loss_final: 0.463398, loss_mean: 0.804755, proj_loss: -0.409583, loss_mean_cls: 0.068225, grad_norm: 0.334044 +Steps: 1%| | 6940/1000000 [28:47<68:17:44, 4.04it/s, grad_norm=0.334, loss_final=0.463, loss_mean=0.805, loss_mean_cls=0.0682, proj_loss=-0.41][2026-03-26 12:59:51] Step: 6940, Training Logs: loss_final: 0.470934, loss_mean: 0.818583, proj_loss: -0.415589, loss_mean_cls: 0.067940, grad_norm: 0.282064 +Steps: 1%| | 6941/1000000 [28:48<68:16:58, 4.04it/s, grad_norm=0.282, loss_final=0.471, loss_mean=0.819, loss_mean_cls=0.0679, proj_loss=-0.416][2026-03-26 12:59:51] Step: 6941, Training Logs: loss_final: 0.466321, loss_mean: 0.824102, proj_loss: -0.423346, loss_mean_cls: 0.065564, grad_norm: 0.531686 +Steps: 1%| | 6942/1000000 [28:48<68:19:53, 4.04it/s, grad_norm=0.532, loss_final=0.466, loss_mean=0.824, loss_mean_cls=0.0656, proj_loss=-0.423][2026-03-26 12:59:51] Step: 6942, Training Logs: loss_final: 0.460417, loss_mean: 0.817443, proj_loss: -0.422601, loss_mean_cls: 0.065575, grad_norm: 0.551133 +Steps: 1%| | 6943/1000000 [28:48<68:15:56, 4.04it/s, grad_norm=0.551, loss_final=0.46, loss_mean=0.817, loss_mean_cls=0.0656, proj_loss=-0.423][2026-03-26 12:59:52] Step: 6943, Training Logs: loss_final: 0.467465, loss_mean: 0.807848, proj_loss: -0.409653, loss_mean_cls: 0.069270, grad_norm: 0.423060 +Steps: 1%| | 6944/1000000 [28:48<68:16:13, 4.04it/s, grad_norm=0.423, loss_final=0.467, loss_mean=0.808, loss_mean_cls=0.0693, proj_loss=-0.41][2026-03-26 12:59:52] Step: 6944, Training Logs: loss_final: 0.479359, loss_mean: 0.833806, proj_loss: -0.422182, loss_mean_cls: 0.067734, grad_norm: 0.785565 +Steps: 1%| | 6945/1000000 [28:49<68:16:50, 4.04it/s, grad_norm=0.786, loss_final=0.479, loss_mean=0.834, loss_mean_cls=0.0677, proj_loss=-0.422][2026-03-26 12:59:52] Step: 6945, Training Logs: loss_final: 0.457244, loss_mean: 0.804678, proj_loss: -0.414820, loss_mean_cls: 0.067387, grad_norm: 0.407691 +Steps: 1%| | 6946/1000000 [28:49<68:16:31, 4.04it/s, grad_norm=0.408, loss_final=0.457, loss_mean=0.805, loss_mean_cls=0.0674, proj_loss=-0.415][2026-03-26 12:59:52] Step: 6946, Training Logs: loss_final: 0.447492, loss_mean: 0.792422, proj_loss: -0.413963, loss_mean_cls: 0.069034, grad_norm: 0.528217 +Steps: 1%| | 6947/1000000 [28:49<68:20:03, 4.04it/s, grad_norm=0.528, loss_final=0.447, loss_mean=0.792, loss_mean_cls=0.069, proj_loss=-0.414][2026-03-26 12:59:53] Step: 6947, Training Logs: loss_final: 0.481166, loss_mean: 0.823224, proj_loss: -0.411104, loss_mean_cls: 0.069046, grad_norm: 0.294141 +Steps: 1%| | 6948/1000000 [28:49<68:19:05, 4.04it/s, grad_norm=0.294, loss_final=0.481, loss_mean=0.823, loss_mean_cls=0.069, proj_loss=-0.411][2026-03-26 12:59:53] Step: 6948, Training Logs: loss_final: 0.461975, loss_mean: 0.807934, proj_loss: -0.415393, loss_mean_cls: 0.069434, grad_norm: 0.683064 +Steps: 1%| | 6949/1000000 [28:50<68:21:53, 4.03it/s, grad_norm=0.683, loss_final=0.462, loss_mean=0.808, loss_mean_cls=0.0694, proj_loss=-0.415][2026-03-26 12:59:53] Step: 6949, Training Logs: loss_final: 0.468124, loss_mean: 0.814523, proj_loss: -0.414976, loss_mean_cls: 0.068577, grad_norm: 0.613394 +Steps: 1%| | 6950/1000000 [28:50<68:18:17, 4.04it/s, grad_norm=0.613, loss_final=0.468, loss_mean=0.815, loss_mean_cls=0.0686, proj_loss=-0.415][2026-03-26 12:59:53] Step: 6950, Training Logs: loss_final: 0.454255, loss_mean: 0.803127, proj_loss: -0.416798, loss_mean_cls: 0.067927, grad_norm: 0.477791 +Steps: 1%| | 6951/1000000 [28:50<68:20:28, 4.04it/s, grad_norm=0.478, loss_final=0.454, loss_mean=0.803, loss_mean_cls=0.0679, proj_loss=-0.417][2026-03-26 12:59:54] Step: 6951, Training Logs: loss_final: 0.499894, loss_mean: 0.835465, proj_loss: -0.405515, loss_mean_cls: 0.069944, grad_norm: 0.391880 +Steps: 1%| | 6952/1000000 [28:50<68:19:25, 4.04it/s, grad_norm=0.392, loss_final=0.5, loss_mean=0.835, loss_mean_cls=0.0699, proj_loss=-0.406][2026-03-26 12:59:54] Step: 6952, Training Logs: loss_final: 0.470916, loss_mean: 0.819357, proj_loss: -0.415867, loss_mean_cls: 0.067427, grad_norm: 0.570132 +Steps: 1%| | 6953/1000000 [28:51<68:20:24, 4.04it/s, grad_norm=0.57, loss_final=0.471, loss_mean=0.819, loss_mean_cls=0.0674, proj_loss=-0.416][2026-03-26 12:59:54] Step: 6953, Training Logs: loss_final: 0.489223, loss_mean: 0.840326, proj_loss: -0.418203, loss_mean_cls: 0.067100, grad_norm: 0.495800 +Steps: 1%| | 6954/1000000 [28:51<68:16:04, 4.04it/s, grad_norm=0.496, loss_final=0.489, loss_mean=0.84, loss_mean_cls=0.0671, proj_loss=-0.418][2026-03-26 12:59:54] Step: 6954, Training Logs: loss_final: 0.473911, loss_mean: 0.820469, proj_loss: -0.415050, loss_mean_cls: 0.068491, grad_norm: 0.406231 +Steps: 1%| | 6955/1000000 [28:51<68:19:28, 4.04it/s, grad_norm=0.406, loss_final=0.474, loss_mean=0.82, loss_mean_cls=0.0685, proj_loss=-0.415][2026-03-26 12:59:55] Step: 6955, Training Logs: loss_final: 0.478443, loss_mean: 0.818817, proj_loss: -0.409065, loss_mean_cls: 0.068692, grad_norm: 0.599687 +Steps: 1%| | 6956/1000000 [28:51<68:18:26, 4.04it/s, grad_norm=0.6, loss_final=0.478, loss_mean=0.819, loss_mean_cls=0.0687, proj_loss=-0.409][2026-03-26 12:59:55] Step: 6956, Training Logs: loss_final: 0.471553, loss_mean: 0.820437, proj_loss: -0.417096, loss_mean_cls: 0.068213, grad_norm: 0.295263 +Steps: 1%| | 6957/1000000 [28:52<68:16:11, 4.04it/s, grad_norm=0.295, loss_final=0.472, loss_mean=0.82, loss_mean_cls=0.0682, proj_loss=-0.417][2026-03-26 12:59:55] Step: 6957, Training Logs: loss_final: 0.480622, loss_mean: 0.822386, proj_loss: -0.411619, loss_mean_cls: 0.069855, grad_norm: 0.499108 +Steps: 1%| | 6958/1000000 [28:52<68:15:38, 4.04it/s, grad_norm=0.499, loss_final=0.481, loss_mean=0.822, loss_mean_cls=0.0699, proj_loss=-0.412][2026-03-26 12:59:55] Step: 6958, Training Logs: loss_final: 0.470924, loss_mean: 0.809452, proj_loss: -0.407926, loss_mean_cls: 0.069398, grad_norm: 0.331887 +Steps: 1%| | 6959/1000000 [28:52<68:17:17, 4.04it/s, grad_norm=0.332, loss_final=0.471, loss_mean=0.809, loss_mean_cls=0.0694, proj_loss=-0.408][2026-03-26 12:59:56] Step: 6959, Training Logs: loss_final: 0.458028, loss_mean: 0.794651, proj_loss: -0.407902, loss_mean_cls: 0.071279, grad_norm: 0.651521 +Steps: 1%| | 6960/1000000 [28:52<68:16:22, 4.04it/s, grad_norm=0.652, loss_final=0.458, loss_mean=0.795, loss_mean_cls=0.0713, proj_loss=-0.408][2026-03-26 12:59:56] Step: 6960, Training Logs: loss_final: 0.461848, loss_mean: 0.806346, proj_loss: -0.413475, loss_mean_cls: 0.068976, grad_norm: 0.271249 +Steps: 1%| | 6961/1000000 [28:53<68:17:21, 4.04it/s, grad_norm=0.271, loss_final=0.462, loss_mean=0.806, loss_mean_cls=0.069, proj_loss=-0.413][2026-03-26 12:59:56] Step: 6961, Training Logs: loss_final: 0.468437, loss_mean: 0.819265, proj_loss: -0.418636, loss_mean_cls: 0.067808, grad_norm: 0.382209 +Steps: 1%| | 6962/1000000 [28:53<68:17:49, 4.04it/s, grad_norm=0.382, loss_final=0.468, loss_mean=0.819, loss_mean_cls=0.0678, proj_loss=-0.419][2026-03-26 12:59:56] Step: 6962, Training Logs: loss_final: 0.485523, loss_mean: 0.845085, proj_loss: -0.424396, loss_mean_cls: 0.064834, grad_norm: 0.387797 +Steps: 1%| | 6963/1000000 [28:53<68:17:32, 4.04it/s, grad_norm=0.388, loss_final=0.486, loss_mean=0.845, loss_mean_cls=0.0648, proj_loss=-0.424][2026-03-26 12:59:57] Step: 6963, Training Logs: loss_final: 0.482575, loss_mean: 0.832184, proj_loss: -0.417109, loss_mean_cls: 0.067500, grad_norm: 0.537007 +Steps: 1%| | 6964/1000000 [28:53<68:21:24, 4.04it/s, grad_norm=0.537, loss_final=0.483, loss_mean=0.832, loss_mean_cls=0.0675, proj_loss=-0.417][2026-03-26 12:59:57] Step: 6964, Training Logs: loss_final: 0.463209, loss_mean: 0.811073, proj_loss: -0.416267, loss_mean_cls: 0.068404, grad_norm: 0.306840 +Steps: 1%| | 6965/1000000 [28:54<70:34:12, 3.91it/s, grad_norm=0.307, loss_final=0.463, loss_mean=0.811, loss_mean_cls=0.0684, proj_loss=-0.416][2026-03-26 12:59:57] Step: 6965, Training Logs: loss_final: 0.480780, loss_mean: 0.817164, proj_loss: -0.406525, loss_mean_cls: 0.070141, grad_norm: 0.452490 +Steps: 1%| | 6966/1000000 [28:54<70:51:38, 3.89it/s, grad_norm=0.452, loss_final=0.481, loss_mean=0.817, loss_mean_cls=0.0701, proj_loss=-0.407][2026-03-26 12:59:57] Step: 6966, Training Logs: loss_final: 0.472311, loss_mean: 0.822303, proj_loss: -0.417833, loss_mean_cls: 0.067841, grad_norm: 0.224681 +Steps: 1%| | 6967/1000000 [28:54<70:06:42, 3.93it/s, grad_norm=0.225, loss_final=0.472, loss_mean=0.822, loss_mean_cls=0.0678, proj_loss=-0.418][2026-03-26 12:59:58] Step: 6967, Training Logs: loss_final: 0.471604, loss_mean: 0.825358, proj_loss: -0.420262, loss_mean_cls: 0.066507, grad_norm: 0.561358 +Steps: 1%| | 6968/1000000 [28:54<69:35:45, 3.96it/s, grad_norm=0.561, loss_final=0.472, loss_mean=0.825, loss_mean_cls=0.0665, proj_loss=-0.42][2026-03-26 12:59:58] Step: 6968, Training Logs: loss_final: 0.482193, loss_mean: 0.825697, proj_loss: -0.411415, loss_mean_cls: 0.067911, grad_norm: 0.311897 +Steps: 1%| | 6969/1000000 [28:55<69:19:21, 3.98it/s, grad_norm=0.312, loss_final=0.482, loss_mean=0.826, loss_mean_cls=0.0679, proj_loss=-0.411][2026-03-26 12:59:58] Step: 6969, Training Logs: loss_final: 0.467477, loss_mean: 0.800711, proj_loss: -0.403283, loss_mean_cls: 0.070049, grad_norm: 0.390312 +Steps: 1%| | 6970/1000000 [28:55<68:59:11, 4.00it/s, grad_norm=0.39, loss_final=0.467, loss_mean=0.801, loss_mean_cls=0.07, proj_loss=-0.403][2026-03-26 12:59:58] Step: 6970, Training Logs: loss_final: 0.471493, loss_mean: 0.800144, proj_loss: -0.399357, loss_mean_cls: 0.070705, grad_norm: 0.368006 +Steps: 1%| | 6971/1000000 [28:55<68:46:00, 4.01it/s, grad_norm=0.368, loss_final=0.471, loss_mean=0.8, loss_mean_cls=0.0707, proj_loss=-0.399][2026-03-26 12:59:59] Step: 6971, Training Logs: loss_final: 0.480667, loss_mean: 0.815180, proj_loss: -0.404186, loss_mean_cls: 0.069673, grad_norm: 0.343384 +Steps: 1%| | 6972/1000000 [28:55<68:35:57, 4.02it/s, grad_norm=0.343, loss_final=0.481, loss_mean=0.815, loss_mean_cls=0.0697, proj_loss=-0.404][2026-03-26 12:59:59] Step: 6972, Training Logs: loss_final: 0.482141, loss_mean: 0.834343, proj_loss: -0.418492, loss_mean_cls: 0.066290, grad_norm: 0.431171 +Steps: 1%| | 6973/1000000 [28:56<68:28:10, 4.03it/s, grad_norm=0.431, loss_final=0.482, loss_mean=0.834, loss_mean_cls=0.0663, proj_loss=-0.418][2026-03-26 12:59:59] Step: 6973, Training Logs: loss_final: 0.481548, loss_mean: 0.832753, proj_loss: -0.417690, loss_mean_cls: 0.066485, grad_norm: 0.291557 +Steps: 1%| | 6974/1000000 [28:56<68:24:35, 4.03it/s, grad_norm=0.292, loss_final=0.482, loss_mean=0.833, loss_mean_cls=0.0665, proj_loss=-0.418][2026-03-26 12:59:59] Step: 6974, Training Logs: loss_final: 0.471972, loss_mean: 0.822851, proj_loss: -0.417279, loss_mean_cls: 0.066401, grad_norm: 0.321955 +Steps: 1%| | 6975/1000000 [28:56<68:21:45, 4.03it/s, grad_norm=0.322, loss_final=0.472, loss_mean=0.823, loss_mean_cls=0.0664, proj_loss=-0.417][2026-03-26 13:00:00] Step: 6975, Training Logs: loss_final: 0.484393, loss_mean: 0.829887, proj_loss: -0.412947, loss_mean_cls: 0.067453, grad_norm: 0.268118 +Steps: 1%| | 6976/1000000 [28:56<68:20:27, 4.04it/s, grad_norm=0.268, loss_final=0.484, loss_mean=0.83, loss_mean_cls=0.0675, proj_loss=-0.413][2026-03-26 13:00:00] Step: 6976, Training Logs: loss_final: 0.470270, loss_mean: 0.812376, proj_loss: -0.411096, loss_mean_cls: 0.068990, grad_norm: 0.252837 +Steps: 1%| | 6977/1000000 [28:57<68:18:30, 4.04it/s, grad_norm=0.253, loss_final=0.47, loss_mean=0.812, loss_mean_cls=0.069, proj_loss=-0.411][2026-03-26 13:00:00] Step: 6977, Training Logs: loss_final: 0.482463, loss_mean: 0.828119, proj_loss: -0.413989, loss_mean_cls: 0.068333, grad_norm: 0.220001 +Steps: 1%| | 6978/1000000 [28:57<68:20:48, 4.04it/s, grad_norm=0.22, loss_final=0.482, loss_mean=0.828, loss_mean_cls=0.0683, proj_loss=-0.414][2026-03-26 13:00:00] Step: 6978, Training Logs: loss_final: 0.467819, loss_mean: 0.812060, proj_loss: -0.412001, loss_mean_cls: 0.067760, grad_norm: 0.514245 +Steps: 1%| | 6979/1000000 [28:57<68:18:37, 4.04it/s, grad_norm=0.514, loss_final=0.468, loss_mean=0.812, loss_mean_cls=0.0678, proj_loss=-0.412][2026-03-26 13:00:01] Step: 6979, Training Logs: loss_final: 0.453634, loss_mean: 0.808480, proj_loss: -0.421491, loss_mean_cls: 0.066645, grad_norm: 0.353775 +Steps: 1%| | 6980/1000000 [28:57<68:20:17, 4.04it/s, grad_norm=0.354, loss_final=0.454, loss_mean=0.808, loss_mean_cls=0.0666, proj_loss=-0.421][2026-03-26 13:00:01] Step: 6980, Training Logs: loss_final: 0.477174, loss_mean: 0.825094, proj_loss: -0.416483, loss_mean_cls: 0.068563, grad_norm: 0.317219 +Steps: 1%| | 6981/1000000 [28:58<68:19:04, 4.04it/s, grad_norm=0.317, loss_final=0.477, loss_mean=0.825, loss_mean_cls=0.0686, proj_loss=-0.416][2026-03-26 13:00:01] Step: 6981, Training Logs: loss_final: 0.457361, loss_mean: 0.796789, proj_loss: -0.408724, loss_mean_cls: 0.069297, grad_norm: 0.328256 +Steps: 1%| | 6982/1000000 [28:58<68:21:02, 4.04it/s, grad_norm=0.328, loss_final=0.457, loss_mean=0.797, loss_mean_cls=0.0693, proj_loss=-0.409][2026-03-26 13:00:01] Step: 6982, Training Logs: loss_final: 0.474412, loss_mean: 0.822805, proj_loss: -0.416591, loss_mean_cls: 0.068199, grad_norm: 0.270782 +Steps: 1%| | 6983/1000000 [28:58<68:17:24, 4.04it/s, grad_norm=0.271, loss_final=0.474, loss_mean=0.823, loss_mean_cls=0.0682, proj_loss=-0.417][2026-03-26 13:00:02] Step: 6983, Training Logs: loss_final: 0.465428, loss_mean: 0.807374, proj_loss: -0.411430, loss_mean_cls: 0.069485, grad_norm: 0.385441 +Steps: 1%| | 6984/1000000 [28:58<68:22:29, 4.03it/s, grad_norm=0.385, loss_final=0.465, loss_mean=0.807, loss_mean_cls=0.0695, proj_loss=-0.411][2026-03-26 13:00:02] Step: 6984, Training Logs: loss_final: 0.465433, loss_mean: 0.803264, proj_loss: -0.408458, loss_mean_cls: 0.070628, grad_norm: 0.274876 +Steps: 1%| | 6985/1000000 [28:59<68:22:25, 4.03it/s, grad_norm=0.275, loss_final=0.465, loss_mean=0.803, loss_mean_cls=0.0706, proj_loss=-0.408][2026-03-26 13:00:02] Step: 6985, Training Logs: loss_final: 0.468265, loss_mean: 0.811956, proj_loss: -0.413203, loss_mean_cls: 0.069512, grad_norm: 0.415043 +Steps: 1%| | 6986/1000000 [28:59<69:38:49, 3.96it/s, grad_norm=0.415, loss_final=0.468, loss_mean=0.812, loss_mean_cls=0.0695, proj_loss=-0.413][2026-03-26 13:00:02] Step: 6986, Training Logs: loss_final: 0.485458, loss_mean: 0.830251, proj_loss: -0.412446, loss_mean_cls: 0.067653, grad_norm: 0.243268 +Steps: 1%| | 6987/1000000 [28:59<69:12:34, 3.99it/s, grad_norm=0.243, loss_final=0.485, loss_mean=0.83, loss_mean_cls=0.0677, proj_loss=-0.412][2026-03-26 13:00:03] Step: 6987, Training Logs: loss_final: 0.466554, loss_mean: 0.822244, proj_loss: -0.421684, loss_mean_cls: 0.065994, grad_norm: 0.331067 +Steps: 1%| | 6988/1000000 [28:59<68:57:03, 4.00it/s, grad_norm=0.331, loss_final=0.467, loss_mean=0.822, loss_mean_cls=0.066, proj_loss=-0.422][2026-03-26 13:00:03] Step: 6988, Training Logs: loss_final: 0.469800, loss_mean: 0.816418, proj_loss: -0.414650, loss_mean_cls: 0.068032, grad_norm: 0.328805 +Steps: 1%| | 6989/1000000 [29:00<68:42:27, 4.01it/s, grad_norm=0.329, loss_final=0.47, loss_mean=0.816, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 13:00:03] Step: 6989, Training Logs: loss_final: 0.484819, loss_mean: 0.832982, proj_loss: -0.416022, loss_mean_cls: 0.067859, grad_norm: 0.254890 +Steps: 1%| | 6990/1000000 [29:00<68:34:14, 4.02it/s, grad_norm=0.255, loss_final=0.485, loss_mean=0.833, loss_mean_cls=0.0679, proj_loss=-0.416][2026-03-26 13:00:03] Step: 6990, Training Logs: loss_final: 0.481435, loss_mean: 0.828023, proj_loss: -0.414992, loss_mean_cls: 0.068404, grad_norm: 0.279228 +Steps: 1%| | 6991/1000000 [29:00<68:26:34, 4.03it/s, grad_norm=0.279, loss_final=0.481, loss_mean=0.828, loss_mean_cls=0.0684, proj_loss=-0.415][2026-03-26 13:00:04] Step: 6991, Training Logs: loss_final: 0.470882, loss_mean: 0.803582, proj_loss: -0.403260, loss_mean_cls: 0.070559, grad_norm: 0.360731 +Steps: 1%| | 6992/1000000 [29:00<68:27:00, 4.03it/s, grad_norm=0.361, loss_final=0.471, loss_mean=0.804, loss_mean_cls=0.0706, proj_loss=-0.403][2026-03-26 13:00:04] Step: 6992, Training Logs: loss_final: 0.483176, loss_mean: 0.832023, proj_loss: -0.416062, loss_mean_cls: 0.067214, grad_norm: 0.568363 +Steps: 1%| | 6993/1000000 [29:01<68:22:22, 4.03it/s, grad_norm=0.568, loss_final=0.483, loss_mean=0.832, loss_mean_cls=0.0672, proj_loss=-0.416][2026-03-26 13:00:04] Step: 6993, Training Logs: loss_final: 0.462354, loss_mean: 0.810186, proj_loss: -0.415109, loss_mean_cls: 0.067277, grad_norm: 0.237574 +Steps: 1%| | 6994/1000000 [29:01<68:21:23, 4.04it/s, grad_norm=0.238, loss_final=0.462, loss_mean=0.81, loss_mean_cls=0.0673, proj_loss=-0.415][2026-03-26 13:00:04] Step: 6994, Training Logs: loss_final: 0.482046, loss_mean: 0.817793, proj_loss: -0.406008, loss_mean_cls: 0.070260, grad_norm: 0.549449 +Steps: 1%| | 6995/1000000 [29:01<68:20:10, 4.04it/s, grad_norm=0.549, loss_final=0.482, loss_mean=0.818, loss_mean_cls=0.0703, proj_loss=-0.406][2026-03-26 13:00:04] Step: 6995, Training Logs: loss_final: 0.475223, loss_mean: 0.821898, proj_loss: -0.414349, loss_mean_cls: 0.067674, grad_norm: 0.279365 +Steps: 1%| | 6996/1000000 [29:01<68:21:36, 4.04it/s, grad_norm=0.279, loss_final=0.475, loss_mean=0.822, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 13:00:05] Step: 6996, Training Logs: loss_final: 0.470508, loss_mean: 0.833393, proj_loss: -0.427777, loss_mean_cls: 0.064892, grad_norm: 0.335541 +Steps: 1%| | 6997/1000000 [29:02<68:19:57, 4.04it/s, grad_norm=0.336, loss_final=0.471, loss_mean=0.833, loss_mean_cls=0.0649, proj_loss=-0.428][2026-03-26 13:00:05] Step: 6997, Training Logs: loss_final: 0.459416, loss_mean: 0.825906, proj_loss: -0.429953, loss_mean_cls: 0.063463, grad_norm: 0.426814 +Steps: 1%| | 6998/1000000 [29:02<68:18:27, 4.04it/s, grad_norm=0.427, loss_final=0.459, loss_mean=0.826, loss_mean_cls=0.0635, proj_loss=-0.43][2026-03-26 13:00:05] Step: 6998, Training Logs: loss_final: 0.468301, loss_mean: 0.813323, proj_loss: -0.413090, loss_mean_cls: 0.068068, grad_norm: 0.335374 +Steps: 1%| | 6999/1000000 [29:02<68:16:14, 4.04it/s, grad_norm=0.335, loss_final=0.468, loss_mean=0.813, loss_mean_cls=0.0681, proj_loss=-0.413][2026-03-26 13:00:05] Step: 6999, Training Logs: loss_final: 0.467103, loss_mean: 0.825144, proj_loss: -0.424116, loss_mean_cls: 0.066075, grad_norm: 0.670093 +Steps: 1%| | 7000/1000000 [29:02<68:16:34, 4.04it/s, grad_norm=0.67, loss_final=0.467, loss_mean=0.825, loss_mean_cls=0.0661, proj_loss=-0.424][2026-03-26 13:00:06] Step: 7000, Training Logs: loss_final: 0.479626, loss_mean: 0.826254, proj_loss: -0.414636, loss_mean_cls: 0.068008, grad_norm: 0.404986 +Steps: 1%| | 7001/1000000 [29:03<68:14:55, 4.04it/s, grad_norm=0.405, loss_final=0.48, loss_mean=0.826, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 13:00:06] Step: 7001, Training Logs: loss_final: 0.476446, loss_mean: 0.819005, proj_loss: -0.410276, loss_mean_cls: 0.067717, grad_norm: 0.325738 +Steps: 1%| | 7002/1000000 [29:03<68:15:47, 4.04it/s, grad_norm=0.326, loss_final=0.476, loss_mean=0.819, loss_mean_cls=0.0677, proj_loss=-0.41][2026-03-26 13:00:06] Step: 7002, Training Logs: loss_final: 0.469345, loss_mean: 0.812426, proj_loss: -0.411855, loss_mean_cls: 0.068774, grad_norm: 0.308608 +Steps: 1%| | 7003/1000000 [29:03<68:16:21, 4.04it/s, grad_norm=0.309, loss_final=0.469, loss_mean=0.812, loss_mean_cls=0.0688, proj_loss=-0.412][2026-03-26 13:00:06] Step: 7003, Training Logs: loss_final: 0.478233, loss_mean: 0.824893, proj_loss: -0.414909, loss_mean_cls: 0.068249, grad_norm: 0.320124 +Steps: 1%| | 7004/1000000 [29:03<68:16:07, 4.04it/s, grad_norm=0.32, loss_final=0.478, loss_mean=0.825, loss_mean_cls=0.0682, proj_loss=-0.415][2026-03-26 13:00:07] Step: 7004, Training Logs: loss_final: 0.471011, loss_mean: 0.823963, proj_loss: -0.419495, loss_mean_cls: 0.066544, grad_norm: 0.197661 +Steps: 1%| | 7005/1000000 [29:04<68:15:06, 4.04it/s, grad_norm=0.198, loss_final=0.471, loss_mean=0.824, loss_mean_cls=0.0665, proj_loss=-0.419][2026-03-26 13:00:07] Step: 7005, Training Logs: loss_final: 0.466582, loss_mean: 0.819790, proj_loss: -0.420261, loss_mean_cls: 0.067052, grad_norm: 0.260154 +Steps: 1%| | 7006/1000000 [29:04<68:14:28, 4.04it/s, grad_norm=0.26, loss_final=0.467, loss_mean=0.82, loss_mean_cls=0.0671, proj_loss=-0.42][2026-03-26 13:00:07] Step: 7006, Training Logs: loss_final: 0.465475, loss_mean: 0.824877, proj_loss: -0.424764, loss_mean_cls: 0.065363, grad_norm: 0.251386 +Steps: 1%| | 7007/1000000 [29:04<68:14:34, 4.04it/s, grad_norm=0.251, loss_final=0.465, loss_mean=0.825, loss_mean_cls=0.0654, proj_loss=-0.425][2026-03-26 13:00:07] Step: 7007, Training Logs: loss_final: 0.460414, loss_mean: 0.807021, proj_loss: -0.414625, loss_mean_cls: 0.068018, grad_norm: 0.442030 +Steps: 1%| | 7008/1000000 [29:04<68:15:28, 4.04it/s, grad_norm=0.442, loss_final=0.46, loss_mean=0.807, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 13:00:08] Step: 7008, Training Logs: loss_final: 0.467054, loss_mean: 0.800104, proj_loss: -0.403955, loss_mean_cls: 0.070905, grad_norm: 0.322039 +Steps: 1%| | 7009/1000000 [29:05<68:15:11, 4.04it/s, grad_norm=0.322, loss_final=0.467, loss_mean=0.8, loss_mean_cls=0.0709, proj_loss=-0.404][2026-03-26 13:00:08] Step: 7009, Training Logs: loss_final: 0.454489, loss_mean: 0.810140, proj_loss: -0.423112, loss_mean_cls: 0.067462, grad_norm: 0.316598 +Steps: 1%| | 7010/1000000 [29:05<68:14:27, 4.04it/s, grad_norm=0.317, loss_final=0.454, loss_mean=0.81, loss_mean_cls=0.0675, proj_loss=-0.423][2026-03-26 13:00:08] Step: 7010, Training Logs: loss_final: 0.477349, loss_mean: 0.826168, proj_loss: -0.415791, loss_mean_cls: 0.066972, grad_norm: 0.316571 +Steps: 1%| | 7011/1000000 [29:05<68:18:54, 4.04it/s, grad_norm=0.317, loss_final=0.477, loss_mean=0.826, loss_mean_cls=0.067, proj_loss=-0.416][2026-03-26 13:00:08] Step: 7011, Training Logs: loss_final: 0.487727, loss_mean: 0.830868, proj_loss: -0.411275, loss_mean_cls: 0.068133, grad_norm: 0.212562 +Steps: 1%| | 7012/1000000 [29:05<68:19:06, 4.04it/s, grad_norm=0.213, loss_final=0.488, loss_mean=0.831, loss_mean_cls=0.0681, proj_loss=-0.411][2026-03-26 13:00:09] Step: 7012, Training Logs: loss_final: 0.478754, loss_mean: 0.832736, proj_loss: -0.420478, loss_mean_cls: 0.066495, grad_norm: 0.261179 +Steps: 1%| | 7013/1000000 [29:06<68:18:39, 4.04it/s, grad_norm=0.261, loss_final=0.479, loss_mean=0.833, loss_mean_cls=0.0665, proj_loss=-0.42][2026-03-26 13:00:09] Step: 7013, Training Logs: loss_final: 0.463610, loss_mean: 0.815821, proj_loss: -0.419172, loss_mean_cls: 0.066961, grad_norm: 0.378389 +Steps: 1%| | 7014/1000000 [29:06<68:18:07, 4.04it/s, grad_norm=0.378, loss_final=0.464, loss_mean=0.816, loss_mean_cls=0.067, proj_loss=-0.419][2026-03-26 13:00:09] Step: 7014, Training Logs: loss_final: 0.483952, loss_mean: 0.833833, proj_loss: -0.415851, loss_mean_cls: 0.065970, grad_norm: 0.363225 +Steps: 1%| | 7015/1000000 [29:06<68:16:47, 4.04it/s, grad_norm=0.363, loss_final=0.484, loss_mean=0.834, loss_mean_cls=0.066, proj_loss=-0.416][2026-03-26 13:00:09] Step: 7015, Training Logs: loss_final: 0.457764, loss_mean: 0.804072, proj_loss: -0.414241, loss_mean_cls: 0.067933, grad_norm: 0.304758 +Steps: 1%| | 7016/1000000 [29:06<68:18:06, 4.04it/s, grad_norm=0.305, loss_final=0.458, loss_mean=0.804, loss_mean_cls=0.0679, proj_loss=-0.414][2026-03-26 13:00:10] Step: 7016, Training Logs: loss_final: 0.466500, loss_mean: 0.819913, proj_loss: -0.420383, loss_mean_cls: 0.066970, grad_norm: 0.409699 +Steps: 1%| | 7017/1000000 [29:07<68:14:01, 4.04it/s, grad_norm=0.41, loss_final=0.466, loss_mean=0.82, loss_mean_cls=0.067, proj_loss=-0.42][2026-03-26 13:00:10] Step: 7017, Training Logs: loss_final: 0.457673, loss_mean: 0.810738, proj_loss: -0.420542, loss_mean_cls: 0.067477, grad_norm: 0.325355 +Steps: 1%| | 7018/1000000 [29:07<68:13:08, 4.04it/s, grad_norm=0.325, loss_final=0.458, loss_mean=0.811, loss_mean_cls=0.0675, proj_loss=-0.421][2026-03-26 13:00:10] Step: 7018, Training Logs: loss_final: 0.478013, loss_mean: 0.838772, proj_loss: -0.426362, loss_mean_cls: 0.065603, grad_norm: 0.383601 +Steps: 1%| | 7019/1000000 [29:07<68:12:18, 4.04it/s, grad_norm=0.384, loss_final=0.478, loss_mean=0.839, loss_mean_cls=0.0656, proj_loss=-0.426][2026-03-26 13:00:10] Step: 7019, Training Logs: loss_final: 0.486109, loss_mean: 0.839769, proj_loss: -0.420898, loss_mean_cls: 0.067239, grad_norm: 0.501065 +Steps: 1%| | 7020/1000000 [29:07<68:13:48, 4.04it/s, grad_norm=0.501, loss_final=0.486, loss_mean=0.84, loss_mean_cls=0.0672, proj_loss=-0.421][2026-03-26 13:00:11] Step: 7020, Training Logs: loss_final: 0.461491, loss_mean: 0.814195, proj_loss: -0.419386, loss_mean_cls: 0.066682, grad_norm: 0.290880 +Steps: 1%| | 7021/1000000 [29:08<68:14:09, 4.04it/s, grad_norm=0.291, loss_final=0.461, loss_mean=0.814, loss_mean_cls=0.0667, proj_loss=-0.419][2026-03-26 13:00:11] Step: 7021, Training Logs: loss_final: 0.468725, loss_mean: 0.813039, proj_loss: -0.412445, loss_mean_cls: 0.068132, grad_norm: 0.381002 +Steps: 1%| | 7022/1000000 [29:08<68:13:14, 4.04it/s, grad_norm=0.381, loss_final=0.469, loss_mean=0.813, loss_mean_cls=0.0681, proj_loss=-0.412][2026-03-26 13:00:11] Step: 7022, Training Logs: loss_final: 0.461551, loss_mean: 0.807982, proj_loss: -0.414761, loss_mean_cls: 0.068331, grad_norm: 0.338239 +Steps: 1%| | 7023/1000000 [29:08<68:19:00, 4.04it/s, grad_norm=0.338, loss_final=0.462, loss_mean=0.808, loss_mean_cls=0.0683, proj_loss=-0.415][2026-03-26 13:00:11] Step: 7023, Training Logs: loss_final: 0.485763, loss_mean: 0.817749, proj_loss: -0.402074, loss_mean_cls: 0.070088, grad_norm: 0.587380 +Steps: 1%| | 7024/1000000 [29:08<68:18:51, 4.04it/s, grad_norm=0.587, loss_final=0.486, loss_mean=0.818, loss_mean_cls=0.0701, proj_loss=-0.402][2026-03-26 13:00:12] Step: 7024, Training Logs: loss_final: 0.470317, loss_mean: 0.821579, proj_loss: -0.417929, loss_mean_cls: 0.066666, grad_norm: 0.269784 +Steps: 1%| | 7025/1000000 [29:09<68:16:18, 4.04it/s, grad_norm=0.27, loss_final=0.47, loss_mean=0.822, loss_mean_cls=0.0667, proj_loss=-0.418][2026-03-26 13:00:12] Step: 7025, Training Logs: loss_final: 0.479868, loss_mean: 0.837010, proj_loss: -0.423526, loss_mean_cls: 0.066384, grad_norm: 0.372306 +Steps: 1%| | 7026/1000000 [29:09<68:18:18, 4.04it/s, grad_norm=0.372, loss_final=0.48, loss_mean=0.837, loss_mean_cls=0.0664, proj_loss=-0.424][2026-03-26 13:00:12] Step: 7026, Training Logs: loss_final: 0.466623, loss_mean: 0.811214, proj_loss: -0.413245, loss_mean_cls: 0.068654, grad_norm: 0.356545 +Steps: 1%| | 7027/1000000 [29:09<68:17:34, 4.04it/s, grad_norm=0.357, loss_final=0.467, loss_mean=0.811, loss_mean_cls=0.0687, proj_loss=-0.413][2026-03-26 13:00:12] Step: 7027, Training Logs: loss_final: 0.452183, loss_mean: 0.790764, proj_loss: -0.408674, loss_mean_cls: 0.070093, grad_norm: 0.447736 +Steps: 1%| | 7028/1000000 [29:09<68:17:51, 4.04it/s, grad_norm=0.448, loss_final=0.452, loss_mean=0.791, loss_mean_cls=0.0701, proj_loss=-0.409][2026-03-26 13:00:13] Step: 7028, Training Logs: loss_final: 0.445598, loss_mean: 0.790069, proj_loss: -0.413288, loss_mean_cls: 0.068817, grad_norm: 0.389952 +Steps: 1%| | 7029/1000000 [29:10<68:18:30, 4.04it/s, grad_norm=0.39, loss_final=0.446, loss_mean=0.79, loss_mean_cls=0.0688, proj_loss=-0.413][2026-03-26 13:00:13] Step: 7029, Training Logs: loss_final: 0.470963, loss_mean: 0.811852, proj_loss: -0.410902, loss_mean_cls: 0.070013, grad_norm: 0.463525 +Steps: 1%| | 7030/1000000 [29:10<68:16:15, 4.04it/s, grad_norm=0.464, loss_final=0.471, loss_mean=0.812, loss_mean_cls=0.07, proj_loss=-0.411][2026-03-26 13:00:13] Step: 7030, Training Logs: loss_final: 0.469574, loss_mean: 0.819800, proj_loss: -0.418547, loss_mean_cls: 0.068320, grad_norm: 0.420480 +Steps: 1%| | 7031/1000000 [29:10<68:18:32, 4.04it/s, grad_norm=0.42, loss_final=0.47, loss_mean=0.82, loss_mean_cls=0.0683, proj_loss=-0.419][2026-03-26 13:00:13] Step: 7031, Training Logs: loss_final: 0.473002, loss_mean: 0.815843, proj_loss: -0.411646, loss_mean_cls: 0.068805, grad_norm: 0.497769 +Steps: 1%| | 7032/1000000 [29:10<68:16:36, 4.04it/s, grad_norm=0.498, loss_final=0.473, loss_mean=0.816, loss_mean_cls=0.0688, proj_loss=-0.412][2026-03-26 13:00:14] Step: 7032, Training Logs: loss_final: 0.453317, loss_mean: 0.810194, proj_loss: -0.422758, loss_mean_cls: 0.065881, grad_norm: 0.444499 +Steps: 1%| | 7033/1000000 [29:11<68:14:24, 4.04it/s, grad_norm=0.444, loss_final=0.453, loss_mean=0.81, loss_mean_cls=0.0659, proj_loss=-0.423][2026-03-26 13:00:14] Step: 7033, Training Logs: loss_final: 0.468348, loss_mean: 0.813798, proj_loss: -0.414887, loss_mean_cls: 0.069437, grad_norm: 0.430182 +Steps: 1%| | 7034/1000000 [29:11<68:12:30, 4.04it/s, grad_norm=0.43, loss_final=0.468, loss_mean=0.814, loss_mean_cls=0.0694, proj_loss=-0.415][2026-03-26 13:00:14] Step: 7034, Training Logs: loss_final: 0.477817, loss_mean: 0.824030, proj_loss: -0.414324, loss_mean_cls: 0.068111, grad_norm: 0.389449 +Steps: 1%| | 7035/1000000 [29:11<68:12:10, 4.04it/s, grad_norm=0.389, loss_final=0.478, loss_mean=0.824, loss_mean_cls=0.0681, proj_loss=-0.414][2026-03-26 13:00:14] Step: 7035, Training Logs: loss_final: 0.466714, loss_mean: 0.804201, proj_loss: -0.407138, loss_mean_cls: 0.069650, grad_norm: 0.228047 +Steps: 1%| | 7036/1000000 [29:11<68:24:33, 4.03it/s, grad_norm=0.228, loss_final=0.467, loss_mean=0.804, loss_mean_cls=0.0697, proj_loss=-0.407][2026-03-26 13:00:15] Step: 7036, Training Logs: loss_final: 0.462589, loss_mean: 0.821887, proj_loss: -0.425021, loss_mean_cls: 0.065724, grad_norm: 0.357198 +Steps: 1%| | 7037/1000000 [29:11<68:22:09, 4.03it/s, grad_norm=0.357, loss_final=0.463, loss_mean=0.822, loss_mean_cls=0.0657, proj_loss=-0.425][2026-03-26 13:00:15] Step: 7037, Training Logs: loss_final: 0.479775, loss_mean: 0.824426, proj_loss: -0.412762, loss_mean_cls: 0.068112, grad_norm: 0.409742 +Steps: 1%| | 7038/1000000 [29:12<68:20:37, 4.04it/s, grad_norm=0.41, loss_final=0.48, loss_mean=0.824, loss_mean_cls=0.0681, proj_loss=-0.413][2026-03-26 13:00:15] Step: 7038, Training Logs: loss_final: 0.463203, loss_mean: 0.818212, proj_loss: -0.421267, loss_mean_cls: 0.066258, grad_norm: 0.453656 +Steps: 1%| | 7039/1000000 [29:12<68:20:05, 4.04it/s, grad_norm=0.454, loss_final=0.463, loss_mean=0.818, loss_mean_cls=0.0663, proj_loss=-0.421][2026-03-26 13:00:15] Step: 7039, Training Logs: loss_final: 0.466007, loss_mean: 0.817683, proj_loss: -0.418870, loss_mean_cls: 0.067194, grad_norm: 0.241362 +Steps: 1%| | 7040/1000000 [29:12<68:19:57, 4.04it/s, grad_norm=0.241, loss_final=0.466, loss_mean=0.818, loss_mean_cls=0.0672, proj_loss=-0.419][2026-03-26 13:00:16] Step: 7040, Training Logs: loss_final: 0.485495, loss_mean: 0.831456, proj_loss: -0.414595, loss_mean_cls: 0.068634, grad_norm: 0.608223 +Steps: 1%| | 7041/1000000 [29:12<68:17:13, 4.04it/s, grad_norm=0.608, loss_final=0.485, loss_mean=0.831, loss_mean_cls=0.0686, proj_loss=-0.415][2026-03-26 13:00:16] Step: 7041, Training Logs: loss_final: 0.470420, loss_mean: 0.829660, proj_loss: -0.425128, loss_mean_cls: 0.065889, grad_norm: 0.372781 +Steps: 1%| | 7042/1000000 [29:13<68:16:09, 4.04it/s, grad_norm=0.373, loss_final=0.47, loss_mean=0.83, loss_mean_cls=0.0659, proj_loss=-0.425][2026-03-26 13:00:16] Step: 7042, Training Logs: loss_final: 0.464643, loss_mean: 0.807458, proj_loss: -0.411691, loss_mean_cls: 0.068877, grad_norm: 0.600137 +Steps: 1%| | 7043/1000000 [29:13<68:17:05, 4.04it/s, grad_norm=0.6, loss_final=0.465, loss_mean=0.807, loss_mean_cls=0.0689, proj_loss=-0.412][2026-03-26 13:00:16] Step: 7043, Training Logs: loss_final: 0.467092, loss_mean: 0.816730, proj_loss: -0.417521, loss_mean_cls: 0.067883, grad_norm: 0.750251 +Steps: 1%| | 7044/1000000 [29:13<68:18:27, 4.04it/s, grad_norm=0.75, loss_final=0.467, loss_mean=0.817, loss_mean_cls=0.0679, proj_loss=-0.418][2026-03-26 13:00:17] Step: 7044, Training Logs: loss_final: 0.465109, loss_mean: 0.807121, proj_loss: -0.410627, loss_mean_cls: 0.068615, grad_norm: 0.190335 +Steps: 1%| | 7045/1000000 [29:13<68:15:46, 4.04it/s, grad_norm=0.19, loss_final=0.465, loss_mean=0.807, loss_mean_cls=0.0686, proj_loss=-0.411][2026-03-26 13:00:17] Step: 7045, Training Logs: loss_final: 0.471380, loss_mean: 0.814227, proj_loss: -0.411451, loss_mean_cls: 0.068604, grad_norm: 0.465361 +Steps: 1%| | 7046/1000000 [29:14<68:14:53, 4.04it/s, grad_norm=0.465, loss_final=0.471, loss_mean=0.814, loss_mean_cls=0.0686, proj_loss=-0.411][2026-03-26 13:00:17] Step: 7046, Training Logs: loss_final: 0.475350, loss_mean: 0.830846, proj_loss: -0.422512, loss_mean_cls: 0.067015, grad_norm: 0.302583 +Steps: 1%| | 7047/1000000 [29:14<68:10:01, 4.05it/s, grad_norm=0.303, loss_final=0.475, loss_mean=0.831, loss_mean_cls=0.067, proj_loss=-0.423][2026-03-26 13:00:17] Step: 7047, Training Logs: loss_final: 0.475134, loss_mean: 0.821426, proj_loss: -0.415493, loss_mean_cls: 0.069201, grad_norm: 0.492282 +Steps: 1%| | 7048/1000000 [29:14<68:11:06, 4.05it/s, grad_norm=0.492, loss_final=0.475, loss_mean=0.821, loss_mean_cls=0.0692, proj_loss=-0.415][2026-03-26 13:00:18] Step: 7048, Training Logs: loss_final: 0.467236, loss_mean: 0.802206, proj_loss: -0.405523, loss_mean_cls: 0.070553, grad_norm: 0.280512 +Steps: 1%| | 7049/1000000 [29:14<68:11:05, 4.05it/s, grad_norm=0.281, loss_final=0.467, loss_mean=0.802, loss_mean_cls=0.0706, proj_loss=-0.406][2026-03-26 13:00:18] Step: 7049, Training Logs: loss_final: 0.474169, loss_mean: 0.828977, proj_loss: -0.420734, loss_mean_cls: 0.065927, grad_norm: 0.273303 +Steps: 1%| | 7050/1000000 [29:15<68:11:55, 4.04it/s, grad_norm=0.273, loss_final=0.474, loss_mean=0.829, loss_mean_cls=0.0659, proj_loss=-0.421][2026-03-26 13:00:18] Step: 7050, Training Logs: loss_final: 0.473855, loss_mean: 0.820627, proj_loss: -0.414118, loss_mean_cls: 0.067346, grad_norm: 0.305774 +Steps: 1%| | 7051/1000000 [29:15<68:14:14, 4.04it/s, grad_norm=0.306, loss_final=0.474, loss_mean=0.821, loss_mean_cls=0.0673, proj_loss=-0.414][2026-03-26 13:00:18] Step: 7051, Training Logs: loss_final: 0.452348, loss_mean: 0.807522, proj_loss: -0.421256, loss_mean_cls: 0.066083, grad_norm: 0.262990 +Steps: 1%| | 7052/1000000 [29:15<68:14:26, 4.04it/s, grad_norm=0.263, loss_final=0.452, loss_mean=0.808, loss_mean_cls=0.0661, proj_loss=-0.421][2026-03-26 13:00:19] Step: 7052, Training Logs: loss_final: 0.478987, loss_mean: 0.837353, proj_loss: -0.424049, loss_mean_cls: 0.065683, grad_norm: 0.261057 +Steps: 1%| | 7053/1000000 [29:15<68:14:22, 4.04it/s, grad_norm=0.261, loss_final=0.479, loss_mean=0.837, loss_mean_cls=0.0657, proj_loss=-0.424][2026-03-26 13:00:19] Step: 7053, Training Logs: loss_final: 0.483788, loss_mean: 0.826220, proj_loss: -0.411978, loss_mean_cls: 0.069545, grad_norm: 0.453315 +Steps: 1%| | 7054/1000000 [29:16<68:13:14, 4.04it/s, grad_norm=0.453, loss_final=0.484, loss_mean=0.826, loss_mean_cls=0.0695, proj_loss=-0.412][2026-03-26 13:00:19] Step: 7054, Training Logs: loss_final: 0.464132, loss_mean: 0.825596, proj_loss: -0.426551, loss_mean_cls: 0.065087, grad_norm: 0.424843 +Steps: 1%| | 7055/1000000 [29:16<68:15:42, 4.04it/s, grad_norm=0.425, loss_final=0.464, loss_mean=0.826, loss_mean_cls=0.0651, proj_loss=-0.427][2026-03-26 13:00:19] Step: 7055, Training Logs: loss_final: 0.467258, loss_mean: 0.830248, proj_loss: -0.428493, loss_mean_cls: 0.065503, grad_norm: 0.234259 +Steps: 1%| | 7056/1000000 [29:16<68:14:17, 4.04it/s, grad_norm=0.234, loss_final=0.467, loss_mean=0.83, loss_mean_cls=0.0655, proj_loss=-0.428][2026-03-26 13:00:20] Step: 7056, Training Logs: loss_final: 0.473968, loss_mean: 0.806416, proj_loss: -0.403423, loss_mean_cls: 0.070975, grad_norm: 0.374916 +Steps: 1%| | 7057/1000000 [29:16<68:14:21, 4.04it/s, grad_norm=0.375, loss_final=0.474, loss_mean=0.806, loss_mean_cls=0.071, proj_loss=-0.403][2026-03-26 13:00:20] Step: 7057, Training Logs: loss_final: 0.462154, loss_mean: 0.819674, proj_loss: -0.423349, loss_mean_cls: 0.065829, grad_norm: 0.283397 +Steps: 1%| | 7058/1000000 [29:17<68:16:46, 4.04it/s, grad_norm=0.283, loss_final=0.462, loss_mean=0.82, loss_mean_cls=0.0658, proj_loss=-0.423][2026-03-26 13:00:20] Step: 7058, Training Logs: loss_final: 0.462206, loss_mean: 0.799482, proj_loss: -0.407668, loss_mean_cls: 0.070392, grad_norm: 0.347849 +Steps: 1%| | 7059/1000000 [29:17<68:19:03, 4.04it/s, grad_norm=0.348, loss_final=0.462, loss_mean=0.799, loss_mean_cls=0.0704, proj_loss=-0.408][2026-03-26 13:00:20] Step: 7059, Training Logs: loss_final: 0.461860, loss_mean: 0.821466, proj_loss: -0.425129, loss_mean_cls: 0.065523, grad_norm: 0.248355 +Steps: 1%| | 7060/1000000 [29:17<68:18:33, 4.04it/s, grad_norm=0.248, loss_final=0.462, loss_mean=0.821, loss_mean_cls=0.0655, proj_loss=-0.425][2026-03-26 13:00:21] Step: 7060, Training Logs: loss_final: 0.456114, loss_mean: 0.809720, proj_loss: -0.419423, loss_mean_cls: 0.065817, grad_norm: 0.331077 +Steps: 1%| | 7061/1000000 [29:17<68:45:37, 4.01it/s, grad_norm=0.331, loss_final=0.456, loss_mean=0.81, loss_mean_cls=0.0658, proj_loss=-0.419][2026-03-26 13:00:21] Step: 7061, Training Logs: loss_final: 0.446916, loss_mean: 0.788434, proj_loss: -0.411210, loss_mean_cls: 0.069693, grad_norm: 0.369450 +Steps: 1%| | 7062/1000000 [29:18<68:09:40, 4.05it/s, grad_norm=0.369, loss_final=0.447, loss_mean=0.788, loss_mean_cls=0.0697, proj_loss=-0.411][2026-03-26 13:00:21] Step: 7062, Training Logs: loss_final: 0.476928, loss_mean: 0.818006, proj_loss: -0.410229, loss_mean_cls: 0.069151, grad_norm: 0.242771 +Steps: 1%| | 7063/1000000 [29:18<68:11:27, 4.04it/s, grad_norm=0.243, loss_final=0.477, loss_mean=0.818, loss_mean_cls=0.0692, proj_loss=-0.41][2026-03-26 13:00:21] Step: 7063, Training Logs: loss_final: 0.458764, loss_mean: 0.789476, proj_loss: -0.401850, loss_mean_cls: 0.071138, grad_norm: 0.373779 +Steps: 1%| | 7064/1000000 [29:18<68:12:43, 4.04it/s, grad_norm=0.374, loss_final=0.459, loss_mean=0.789, loss_mean_cls=0.0711, proj_loss=-0.402][2026-03-26 13:00:22] Step: 7064, Training Logs: loss_final: 0.466357, loss_mean: 0.811144, proj_loss: -0.412751, loss_mean_cls: 0.067964, grad_norm: 0.207586 +Steps: 1%| | 7065/1000000 [29:18<68:24:26, 4.03it/s, grad_norm=0.208, loss_final=0.466, loss_mean=0.811, loss_mean_cls=0.068, proj_loss=-0.413][2026-03-26 13:00:22] Step: 7065, Training Logs: loss_final: 0.469996, loss_mean: 0.820208, proj_loss: -0.417556, loss_mean_cls: 0.067345, grad_norm: 0.219948 +Steps: 1%| | 7066/1000000 [29:19<68:21:46, 4.03it/s, grad_norm=0.22, loss_final=0.47, loss_mean=0.82, loss_mean_cls=0.0673, proj_loss=-0.418][2026-03-26 13:00:22] Step: 7066, Training Logs: loss_final: 0.470299, loss_mean: 0.831006, proj_loss: -0.425039, loss_mean_cls: 0.064331, grad_norm: 0.257110 +Steps: 1%| | 7067/1000000 [29:19<68:31:11, 4.03it/s, grad_norm=0.257, loss_final=0.47, loss_mean=0.831, loss_mean_cls=0.0643, proj_loss=-0.425][2026-03-26 13:00:22] Step: 7067, Training Logs: loss_final: 0.469414, loss_mean: 0.832839, proj_loss: -0.428860, loss_mean_cls: 0.065435, grad_norm: 0.211934 +Steps: 1%| | 7068/1000000 [29:19<68:26:17, 4.03it/s, grad_norm=0.212, loss_final=0.469, loss_mean=0.833, loss_mean_cls=0.0654, proj_loss=-0.429][2026-03-26 13:00:23] Step: 7068, Training Logs: loss_final: 0.473318, loss_mean: 0.824591, proj_loss: -0.418662, loss_mean_cls: 0.067389, grad_norm: 0.228002 +Steps: 1%| | 7069/1000000 [29:19<68:21:39, 4.03it/s, grad_norm=0.228, loss_final=0.473, loss_mean=0.825, loss_mean_cls=0.0674, proj_loss=-0.419][2026-03-26 13:00:23] Step: 7069, Training Logs: loss_final: 0.468016, loss_mean: 0.807323, proj_loss: -0.408754, loss_mean_cls: 0.069447, grad_norm: 0.272801 +Steps: 1%| | 7070/1000000 [29:20<68:20:51, 4.04it/s, grad_norm=0.273, loss_final=0.468, loss_mean=0.807, loss_mean_cls=0.0694, proj_loss=-0.409][2026-03-26 13:00:23] Step: 7070, Training Logs: loss_final: 0.472560, loss_mean: 0.826311, proj_loss: -0.420440, loss_mean_cls: 0.066689, grad_norm: 0.211182 +Steps: 1%| | 7071/1000000 [29:20<68:19:05, 4.04it/s, grad_norm=0.211, loss_final=0.473, loss_mean=0.826, loss_mean_cls=0.0667, proj_loss=-0.42][2026-03-26 13:00:23] Step: 7071, Training Logs: loss_final: 0.467904, loss_mean: 0.816577, proj_loss: -0.416546, loss_mean_cls: 0.067874, grad_norm: 0.323064 +Steps: 1%| | 7072/1000000 [29:20<68:18:32, 4.04it/s, grad_norm=0.323, loss_final=0.468, loss_mean=0.817, loss_mean_cls=0.0679, proj_loss=-0.417][2026-03-26 13:00:24] Step: 7072, Training Logs: loss_final: 0.489058, loss_mean: 0.824189, proj_loss: -0.406411, loss_mean_cls: 0.071280, grad_norm: 0.339265 +Steps: 1%| | 7073/1000000 [29:20<68:16:34, 4.04it/s, grad_norm=0.339, loss_final=0.489, loss_mean=0.824, loss_mean_cls=0.0713, proj_loss=-0.406][2026-03-26 13:00:24] Step: 7073, Training Logs: loss_final: 0.462755, loss_mean: 0.818429, proj_loss: -0.420744, loss_mean_cls: 0.065071, grad_norm: 0.294594 +Steps: 1%| | 7074/1000000 [29:21<68:19:13, 4.04it/s, grad_norm=0.295, loss_final=0.463, loss_mean=0.818, loss_mean_cls=0.0651, proj_loss=-0.421][2026-03-26 13:00:24] Step: 7074, Training Logs: loss_final: 0.459476, loss_mean: 0.805845, proj_loss: -0.415337, loss_mean_cls: 0.068968, grad_norm: 0.542049 +Steps: 1%| | 7075/1000000 [29:21<68:16:04, 4.04it/s, grad_norm=0.542, loss_final=0.459, loss_mean=0.806, loss_mean_cls=0.069, proj_loss=-0.415][2026-03-26 13:00:24] Step: 7075, Training Logs: loss_final: 0.473472, loss_mean: 0.817281, proj_loss: -0.413495, loss_mean_cls: 0.069686, grad_norm: 0.303896 +Steps: 1%| | 7076/1000000 [29:21<68:17:46, 4.04it/s, grad_norm=0.304, loss_final=0.473, loss_mean=0.817, loss_mean_cls=0.0697, proj_loss=-0.413][2026-03-26 13:00:25] Step: 7076, Training Logs: loss_final: 0.459141, loss_mean: 0.811014, proj_loss: -0.418765, loss_mean_cls: 0.066892, grad_norm: 0.479042 +Steps: 1%| | 7077/1000000 [29:21<68:15:32, 4.04it/s, grad_norm=0.479, loss_final=0.459, loss_mean=0.811, loss_mean_cls=0.0669, proj_loss=-0.419][2026-03-26 13:00:25] Step: 7077, Training Logs: loss_final: 0.451601, loss_mean: 0.794624, proj_loss: -0.411928, loss_mean_cls: 0.068905, grad_norm: 0.253938 +Steps: 1%| | 7078/1000000 [29:22<68:16:20, 4.04it/s, grad_norm=0.254, loss_final=0.452, loss_mean=0.795, loss_mean_cls=0.0689, proj_loss=-0.412][2026-03-26 13:00:25] Step: 7078, Training Logs: loss_final: 0.455634, loss_mean: 0.808279, proj_loss: -0.419535, loss_mean_cls: 0.066890, grad_norm: 0.377327 +Steps: 1%| | 7079/1000000 [29:22<68:14:05, 4.04it/s, grad_norm=0.377, loss_final=0.456, loss_mean=0.808, loss_mean_cls=0.0669, proj_loss=-0.42][2026-03-26 13:00:25] Step: 7079, Training Logs: loss_final: 0.488438, loss_mean: 0.832147, proj_loss: -0.412087, loss_mean_cls: 0.068378, grad_norm: 0.394992 +Steps: 1%| | 7080/1000000 [29:22<68:14:41, 4.04it/s, grad_norm=0.395, loss_final=0.488, loss_mean=0.832, loss_mean_cls=0.0684, proj_loss=-0.412][2026-03-26 13:00:26] Step: 7080, Training Logs: loss_final: 0.471489, loss_mean: 0.826187, proj_loss: -0.421233, loss_mean_cls: 0.066535, grad_norm: 0.408765 +Steps: 1%| | 7081/1000000 [29:22<68:13:01, 4.04it/s, grad_norm=0.409, loss_final=0.471, loss_mean=0.826, loss_mean_cls=0.0665, proj_loss=-0.421][2026-03-26 13:00:26] Step: 7081, Training Logs: loss_final: 0.473025, loss_mean: 0.814945, proj_loss: -0.411363, loss_mean_cls: 0.069443, grad_norm: 0.464684 +Steps: 1%| | 7082/1000000 [29:23<68:15:16, 4.04it/s, grad_norm=0.465, loss_final=0.473, loss_mean=0.815, loss_mean_cls=0.0694, proj_loss=-0.411][2026-03-26 13:00:26] Step: 7082, Training Logs: loss_final: 0.472705, loss_mean: 0.833405, proj_loss: -0.425219, loss_mean_cls: 0.064520, grad_norm: 0.368868 +Steps: 1%| | 7083/1000000 [29:23<68:13:50, 4.04it/s, grad_norm=0.369, loss_final=0.473, loss_mean=0.833, loss_mean_cls=0.0645, proj_loss=-0.425][2026-03-26 13:00:26] Step: 7083, Training Logs: loss_final: 0.452795, loss_mean: 0.805508, proj_loss: -0.419301, loss_mean_cls: 0.066588, grad_norm: 0.449788 +Steps: 1%| | 7084/1000000 [29:23<68:13:45, 4.04it/s, grad_norm=0.45, loss_final=0.453, loss_mean=0.806, loss_mean_cls=0.0666, proj_loss=-0.419][2026-03-26 13:00:27] Step: 7084, Training Logs: loss_final: 0.471347, loss_mean: 0.822602, proj_loss: -0.418504, loss_mean_cls: 0.067250, grad_norm: 0.288996 +Steps: 1%| | 7085/1000000 [29:23<68:15:11, 4.04it/s, grad_norm=0.289, loss_final=0.471, loss_mean=0.823, loss_mean_cls=0.0672, proj_loss=-0.419][2026-03-26 13:00:27] Step: 7085, Training Logs: loss_final: 0.463756, loss_mean: 0.814430, proj_loss: -0.417384, loss_mean_cls: 0.066710, grad_norm: 0.299984 +Steps: 1%| | 7086/1000000 [29:24<68:17:16, 4.04it/s, grad_norm=0.3, loss_final=0.464, loss_mean=0.814, loss_mean_cls=0.0667, proj_loss=-0.417][2026-03-26 13:00:27] Step: 7086, Training Logs: loss_final: 0.474016, loss_mean: 0.820410, proj_loss: -0.414213, loss_mean_cls: 0.067818, grad_norm: 0.352727 +Steps: 1%| | 7087/1000000 [29:24<68:17:58, 4.04it/s, grad_norm=0.353, loss_final=0.474, loss_mean=0.82, loss_mean_cls=0.0678, proj_loss=-0.414][2026-03-26 13:00:27] Step: 7087, Training Logs: loss_final: 0.463086, loss_mean: 0.818412, proj_loss: -0.421354, loss_mean_cls: 0.066028, grad_norm: 0.417927 +Steps: 1%| | 7088/1000000 [29:24<68:17:11, 4.04it/s, grad_norm=0.418, loss_final=0.463, loss_mean=0.818, loss_mean_cls=0.066, proj_loss=-0.421][2026-03-26 13:00:28] Step: 7088, Training Logs: loss_final: 0.469350, loss_mean: 0.807894, proj_loss: -0.408126, loss_mean_cls: 0.069583, grad_norm: 0.354764 +Steps: 1%| | 7089/1000000 [29:24<68:16:47, 4.04it/s, grad_norm=0.355, loss_final=0.469, loss_mean=0.808, loss_mean_cls=0.0696, proj_loss=-0.408][2026-03-26 13:00:28] Step: 7089, Training Logs: loss_final: 0.474322, loss_mean: 0.826946, proj_loss: -0.418941, loss_mean_cls: 0.066316, grad_norm: 0.386665 +Steps: 1%| | 7090/1000000 [29:25<68:30:55, 4.03it/s, grad_norm=0.387, loss_final=0.474, loss_mean=0.827, loss_mean_cls=0.0663, proj_loss=-0.419][2026-03-26 13:00:28] Step: 7090, Training Logs: loss_final: 0.449201, loss_mean: 0.801833, proj_loss: -0.420060, loss_mean_cls: 0.067428, grad_norm: 0.506243 +Steps: 1%| | 7091/1000000 [29:25<68:42:20, 4.01it/s, grad_norm=0.506, loss_final=0.449, loss_mean=0.802, loss_mean_cls=0.0674, proj_loss=-0.42][2026-03-26 13:00:28] Step: 7091, Training Logs: loss_final: 0.475301, loss_mean: 0.806631, proj_loss: -0.402526, loss_mean_cls: 0.071196, grad_norm: 0.330028 +Steps: 1%| | 7092/1000000 [29:25<68:35:00, 4.02it/s, grad_norm=0.33, loss_final=0.475, loss_mean=0.807, loss_mean_cls=0.0712, proj_loss=-0.403][2026-03-26 13:00:29] Step: 7092, Training Logs: loss_final: 0.475311, loss_mean: 0.828619, proj_loss: -0.419637, loss_mean_cls: 0.066329, grad_norm: 0.531027 +Steps: 1%| | 7093/1000000 [29:25<68:26:26, 4.03it/s, grad_norm=0.531, loss_final=0.475, loss_mean=0.829, loss_mean_cls=0.0663, proj_loss=-0.42][2026-03-26 13:00:29] Step: 7093, Training Logs: loss_final: 0.467098, loss_mean: 0.807844, proj_loss: -0.409647, loss_mean_cls: 0.068901, grad_norm: 0.361542 +Steps: 1%| | 7094/1000000 [29:26<68:24:06, 4.03it/s, grad_norm=0.362, loss_final=0.467, loss_mean=0.808, loss_mean_cls=0.0689, proj_loss=-0.41][2026-03-26 13:00:29] Step: 7094, Training Logs: loss_final: 0.476026, loss_mean: 0.837396, proj_loss: -0.426196, loss_mean_cls: 0.064826, grad_norm: 0.286864 +Steps: 1%| | 7095/1000000 [29:26<68:20:57, 4.04it/s, grad_norm=0.287, loss_final=0.476, loss_mean=0.837, loss_mean_cls=0.0648, proj_loss=-0.426][2026-03-26 13:00:29] Step: 7095, Training Logs: loss_final: 0.469014, loss_mean: 0.821070, proj_loss: -0.419814, loss_mean_cls: 0.067758, grad_norm: 0.244997 +Steps: 1%| | 7096/1000000 [29:26<68:20:18, 4.04it/s, grad_norm=0.245, loss_final=0.469, loss_mean=0.821, loss_mean_cls=0.0678, proj_loss=-0.42][2026-03-26 13:00:30] Step: 7096, Training Logs: loss_final: 0.461438, loss_mean: 0.815525, proj_loss: -0.420775, loss_mean_cls: 0.066688, grad_norm: 0.285436 +Steps: 1%| | 7097/1000000 [29:26<68:16:35, 4.04it/s, grad_norm=0.285, loss_final=0.461, loss_mean=0.816, loss_mean_cls=0.0667, proj_loss=-0.421][2026-03-26 13:00:30] Step: 7097, Training Logs: loss_final: 0.467076, loss_mean: 0.821700, proj_loss: -0.420815, loss_mean_cls: 0.066192, grad_norm: 0.344192 +Steps: 1%| | 7098/1000000 [29:27<68:15:21, 4.04it/s, grad_norm=0.344, loss_final=0.467, loss_mean=0.822, loss_mean_cls=0.0662, proj_loss=-0.421][2026-03-26 13:00:30] Step: 7098, Training Logs: loss_final: 0.473648, loss_mean: 0.827268, proj_loss: -0.420179, loss_mean_cls: 0.066559, grad_norm: 0.251254 +Steps: 1%| | 7099/1000000 [29:27<68:14:24, 4.04it/s, grad_norm=0.251, loss_final=0.474, loss_mean=0.827, loss_mean_cls=0.0666, proj_loss=-0.42][2026-03-26 13:00:30] Step: 7099, Training Logs: loss_final: 0.474478, loss_mean: 0.824493, proj_loss: -0.417518, loss_mean_cls: 0.067504, grad_norm: 0.378930 +Steps: 1%| | 7100/1000000 [29:27<68:13:27, 4.04it/s, grad_norm=0.379, loss_final=0.474, loss_mean=0.824, loss_mean_cls=0.0675, proj_loss=-0.418][2026-03-26 13:00:30] Step: 7100, Training Logs: loss_final: 0.442223, loss_mean: 0.792414, proj_loss: -0.417962, loss_mean_cls: 0.067771, grad_norm: 0.552657 +Steps: 1%| | 7101/1000000 [29:27<68:14:58, 4.04it/s, grad_norm=0.553, loss_final=0.442, loss_mean=0.792, loss_mean_cls=0.0678, proj_loss=-0.418][2026-03-26 13:00:31] Step: 7101, Training Logs: loss_final: 0.461688, loss_mean: 0.822942, proj_loss: -0.426412, loss_mean_cls: 0.065158, grad_norm: 0.252259 +Steps: 1%| | 7102/1000000 [29:28<68:14:25, 4.04it/s, grad_norm=0.252, loss_final=0.462, loss_mean=0.823, loss_mean_cls=0.0652, proj_loss=-0.426][2026-03-26 13:00:31] Step: 7102, Training Logs: loss_final: 0.458076, loss_mean: 0.810040, proj_loss: -0.419525, loss_mean_cls: 0.067560, grad_norm: 0.196864 +Steps: 1%| | 7103/1000000 [29:28<68:17:23, 4.04it/s, grad_norm=0.197, loss_final=0.458, loss_mean=0.81, loss_mean_cls=0.0676, proj_loss=-0.42][2026-03-26 13:00:31] Step: 7103, Training Logs: loss_final: 0.477058, loss_mean: 0.833157, proj_loss: -0.421381, loss_mean_cls: 0.065283, grad_norm: 0.188014 +Steps: 1%| | 7104/1000000 [29:28<68:19:31, 4.04it/s, grad_norm=0.188, loss_final=0.477, loss_mean=0.833, loss_mean_cls=0.0653, proj_loss=-0.421][2026-03-26 13:00:31] Step: 7104, Training Logs: loss_final: 0.481070, loss_mean: 0.831592, proj_loss: -0.417930, loss_mean_cls: 0.067408, grad_norm: 0.278600 +Steps: 1%| | 7105/1000000 [29:28<68:19:26, 4.04it/s, grad_norm=0.279, loss_final=0.481, loss_mean=0.832, loss_mean_cls=0.0674, proj_loss=-0.418][2026-03-26 13:00:32] Step: 7105, Training Logs: loss_final: 0.470219, loss_mean: 0.809964, proj_loss: -0.409206, loss_mean_cls: 0.069462, grad_norm: 0.290889 +Steps: 1%| | 7106/1000000 [29:29<68:15:43, 4.04it/s, grad_norm=0.291, loss_final=0.47, loss_mean=0.81, loss_mean_cls=0.0695, proj_loss=-0.409][2026-03-26 13:00:32] Step: 7106, Training Logs: loss_final: 0.484382, loss_mean: 0.834929, proj_loss: -0.416464, loss_mean_cls: 0.065917, grad_norm: 0.353593 +Steps: 1%| | 7107/1000000 [29:29<68:17:27, 4.04it/s, grad_norm=0.354, loss_final=0.484, loss_mean=0.835, loss_mean_cls=0.0659, proj_loss=-0.416][2026-03-26 13:00:32] Step: 7107, Training Logs: loss_final: 0.459855, loss_mean: 0.799606, proj_loss: -0.409522, loss_mean_cls: 0.069771, grad_norm: 0.315392 +Steps: 1%| | 7108/1000000 [29:29<68:17:31, 4.04it/s, grad_norm=0.315, loss_final=0.46, loss_mean=0.8, loss_mean_cls=0.0698, proj_loss=-0.41][2026-03-26 13:00:32] Step: 7108, Training Logs: loss_final: 0.475206, loss_mean: 0.819746, proj_loss: -0.413499, loss_mean_cls: 0.068959, grad_norm: 0.389556 +Steps: 1%| | 7109/1000000 [29:29<68:19:19, 4.04it/s, grad_norm=0.39, loss_final=0.475, loss_mean=0.82, loss_mean_cls=0.069, proj_loss=-0.413][2026-03-26 13:00:33] Step: 7109, Training Logs: loss_final: 0.469535, loss_mean: 0.814074, proj_loss: -0.412501, loss_mean_cls: 0.067962, grad_norm: 0.425939 +Steps: 1%| | 7110/1000000 [29:30<68:19:11, 4.04it/s, grad_norm=0.426, loss_final=0.47, loss_mean=0.814, loss_mean_cls=0.068, proj_loss=-0.413][2026-03-26 13:00:33] Step: 7110, Training Logs: loss_final: 0.491373, loss_mean: 0.832627, proj_loss: -0.410084, loss_mean_cls: 0.068831, grad_norm: 0.267513 +Steps: 1%| | 7111/1000000 [29:30<68:18:41, 4.04it/s, grad_norm=0.268, loss_final=0.491, loss_mean=0.833, loss_mean_cls=0.0688, proj_loss=-0.41][2026-03-26 13:00:33] Step: 7111, Training Logs: loss_final: 0.472676, loss_mean: 0.820829, proj_loss: -0.415670, loss_mean_cls: 0.067517, grad_norm: 0.302786 +Steps: 1%| | 7112/1000000 [29:30<68:19:13, 4.04it/s, grad_norm=0.303, loss_final=0.473, loss_mean=0.821, loss_mean_cls=0.0675, proj_loss=-0.416][2026-03-26 13:00:33] Step: 7112, Training Logs: loss_final: 0.484027, loss_mean: 0.831595, proj_loss: -0.415436, loss_mean_cls: 0.067868, grad_norm: 0.265555 +Steps: 1%| | 7113/1000000 [29:30<68:18:05, 4.04it/s, grad_norm=0.266, loss_final=0.484, loss_mean=0.832, loss_mean_cls=0.0679, proj_loss=-0.415][2026-03-26 13:00:34] Step: 7113, Training Logs: loss_final: 0.467181, loss_mean: 0.824041, proj_loss: -0.423089, loss_mean_cls: 0.066229, grad_norm: 0.513988 +Steps: 1%| | 7114/1000000 [29:31<68:20:37, 4.04it/s, grad_norm=0.514, loss_final=0.467, loss_mean=0.824, loss_mean_cls=0.0662, proj_loss=-0.423][2026-03-26 13:00:34] Step: 7114, Training Logs: loss_final: 0.492812, loss_mean: 0.846148, proj_loss: -0.419558, loss_mean_cls: 0.066222, grad_norm: 0.239396 +Steps: 1%| | 7115/1000000 [29:31<68:19:25, 4.04it/s, grad_norm=0.239, loss_final=0.493, loss_mean=0.846, loss_mean_cls=0.0662, proj_loss=-0.42][2026-03-26 13:00:34] Step: 7115, Training Logs: loss_final: 0.470234, loss_mean: 0.826339, proj_loss: -0.422249, loss_mean_cls: 0.066144, grad_norm: 0.418535 +Steps: 1%| | 7116/1000000 [29:31<68:20:04, 4.04it/s, grad_norm=0.419, loss_final=0.47, loss_mean=0.826, loss_mean_cls=0.0661, proj_loss=-0.422][2026-03-26 13:00:34] Step: 7116, Training Logs: loss_final: 0.464848, loss_mean: 0.825260, proj_loss: -0.425743, loss_mean_cls: 0.065330, grad_norm: 0.452615 +Steps: 1%| | 7117/1000000 [29:31<68:19:38, 4.04it/s, grad_norm=0.453, loss_final=0.465, loss_mean=0.825, loss_mean_cls=0.0653, proj_loss=-0.426][2026-03-26 13:00:35] Step: 7117, Training Logs: loss_final: 0.463626, loss_mean: 0.813181, proj_loss: -0.416997, loss_mean_cls: 0.067443, grad_norm: 0.285331 +Steps: 1%| | 7118/1000000 [29:32<68:17:47, 4.04it/s, grad_norm=0.285, loss_final=0.464, loss_mean=0.813, loss_mean_cls=0.0674, proj_loss=-0.417][2026-03-26 13:00:35] Step: 7118, Training Logs: loss_final: 0.483648, loss_mean: 0.839126, proj_loss: -0.421815, loss_mean_cls: 0.066337, grad_norm: 0.362575 +Steps: 1%| | 7119/1000000 [29:32<68:15:07, 4.04it/s, grad_norm=0.363, loss_final=0.484, loss_mean=0.839, loss_mean_cls=0.0663, proj_loss=-0.422][2026-03-26 13:00:35] Step: 7119, Training Logs: loss_final: 0.484744, loss_mean: 0.839938, proj_loss: -0.420911, loss_mean_cls: 0.065717, grad_norm: 0.351134 +Steps: 1%| | 7120/1000000 [29:32<68:17:33, 4.04it/s, grad_norm=0.351, loss_final=0.485, loss_mean=0.84, loss_mean_cls=0.0657, proj_loss=-0.421][2026-03-26 13:00:35] Step: 7120, Training Logs: loss_final: 0.466776, loss_mean: 0.813164, proj_loss: -0.414938, loss_mean_cls: 0.068550, grad_norm: 0.326056 +Steps: 1%| | 7121/1000000 [29:32<68:14:53, 4.04it/s, grad_norm=0.326, loss_final=0.467, loss_mean=0.813, loss_mean_cls=0.0686, proj_loss=-0.415][2026-03-26 13:00:36] Step: 7121, Training Logs: loss_final: 0.465198, loss_mean: 0.817659, proj_loss: -0.419377, loss_mean_cls: 0.066917, grad_norm: 0.389256 +Steps: 1%| | 7122/1000000 [29:33<68:17:10, 4.04it/s, grad_norm=0.389, loss_final=0.465, loss_mean=0.818, loss_mean_cls=0.0669, proj_loss=-0.419][2026-03-26 13:00:36] Step: 7122, Training Logs: loss_final: 0.472612, loss_mean: 0.817234, proj_loss: -0.413355, loss_mean_cls: 0.068733, grad_norm: 0.272111 +Steps: 1%| | 7123/1000000 [29:33<68:17:06, 4.04it/s, grad_norm=0.272, loss_final=0.473, loss_mean=0.817, loss_mean_cls=0.0687, proj_loss=-0.413][2026-03-26 13:00:36] Step: 7123, Training Logs: loss_final: 0.463660, loss_mean: 0.821530, proj_loss: -0.423199, loss_mean_cls: 0.065329, grad_norm: 0.540629 +Steps: 1%| | 7124/1000000 [29:33<68:18:20, 4.04it/s, grad_norm=0.541, loss_final=0.464, loss_mean=0.822, loss_mean_cls=0.0653, proj_loss=-0.423][2026-03-26 13:00:36] Step: 7124, Training Logs: loss_final: 0.464160, loss_mean: 0.808313, proj_loss: -0.412351, loss_mean_cls: 0.068198, grad_norm: 0.362443 +Steps: 1%| | 7125/1000000 [29:33<68:17:47, 4.04it/s, grad_norm=0.362, loss_final=0.464, loss_mean=0.808, loss_mean_cls=0.0682, proj_loss=-0.412][2026-03-26 13:00:37] Step: 7125, Training Logs: loss_final: 0.484183, loss_mean: 0.832540, proj_loss: -0.416854, loss_mean_cls: 0.068497, grad_norm: 0.380232 +Steps: 1%| | 7126/1000000 [29:34<68:16:35, 4.04it/s, grad_norm=0.38, loss_final=0.484, loss_mean=0.833, loss_mean_cls=0.0685, proj_loss=-0.417][2026-03-26 13:00:37] Step: 7126, Training Logs: loss_final: 0.467894, loss_mean: 0.818563, proj_loss: -0.418005, loss_mean_cls: 0.067336, grad_norm: 0.594104 +Steps: 1%| | 7127/1000000 [29:34<68:14:20, 4.04it/s, grad_norm=0.594, loss_final=0.468, loss_mean=0.819, loss_mean_cls=0.0673, proj_loss=-0.418][2026-03-26 13:00:37] Step: 7127, Training Logs: loss_final: 0.474955, loss_mean: 0.827027, proj_loss: -0.418974, loss_mean_cls: 0.066902, grad_norm: 0.294001 +Steps: 1%| | 7128/1000000 [29:34<68:20:07, 4.04it/s, grad_norm=0.294, loss_final=0.475, loss_mean=0.827, loss_mean_cls=0.0669, proj_loss=-0.419][2026-03-26 13:00:37] Step: 7128, Training Logs: loss_final: 0.463661, loss_mean: 0.807236, proj_loss: -0.411850, loss_mean_cls: 0.068275, grad_norm: 0.560741 +Steps: 1%| | 7129/1000000 [29:34<68:18:26, 4.04it/s, grad_norm=0.561, loss_final=0.464, loss_mean=0.807, loss_mean_cls=0.0683, proj_loss=-0.412][2026-03-26 13:00:38] Step: 7129, Training Logs: loss_final: 0.465536, loss_mean: 0.814181, proj_loss: -0.416990, loss_mean_cls: 0.068345, grad_norm: 0.306560 +Steps: 1%| | 7130/1000000 [29:35<68:20:53, 4.04it/s, grad_norm=0.307, loss_final=0.466, loss_mean=0.814, loss_mean_cls=0.0683, proj_loss=-0.417][2026-03-26 13:00:38] Step: 7130, Training Logs: loss_final: 0.486314, loss_mean: 0.825856, proj_loss: -0.408676, loss_mean_cls: 0.069134, grad_norm: 0.448811 +Steps: 1%| | 7131/1000000 [29:35<68:18:10, 4.04it/s, grad_norm=0.449, loss_final=0.486, loss_mean=0.826, loss_mean_cls=0.0691, proj_loss=-0.409][2026-03-26 13:00:38] Step: 7131, Training Logs: loss_final: 0.490053, loss_mean: 0.833802, proj_loss: -0.412603, loss_mean_cls: 0.068855, grad_norm: 0.372348 +Steps: 1%| | 7132/1000000 [29:35<68:16:27, 4.04it/s, grad_norm=0.372, loss_final=0.49, loss_mean=0.834, loss_mean_cls=0.0689, proj_loss=-0.413][2026-03-26 13:00:38] Step: 7132, Training Logs: loss_final: 0.452419, loss_mean: 0.801609, proj_loss: -0.416602, loss_mean_cls: 0.067412, grad_norm: 0.456471 +Steps: 1%| | 7133/1000000 [29:35<68:16:04, 4.04it/s, grad_norm=0.456, loss_final=0.452, loss_mean=0.802, loss_mean_cls=0.0674, proj_loss=-0.417][2026-03-26 13:00:39] Step: 7133, Training Logs: loss_final: 0.477865, loss_mean: 0.835251, proj_loss: -0.423227, loss_mean_cls: 0.065841, grad_norm: 0.782207 +Steps: 1%| | 7134/1000000 [29:36<68:18:30, 4.04it/s, grad_norm=0.782, loss_final=0.478, loss_mean=0.835, loss_mean_cls=0.0658, proj_loss=-0.423][2026-03-26 13:00:39] Step: 7134, Training Logs: loss_final: 0.459135, loss_mean: 0.797469, proj_loss: -0.407641, loss_mean_cls: 0.069306, grad_norm: 0.309711 +Steps: 1%| | 7135/1000000 [29:36<68:18:46, 4.04it/s, grad_norm=0.31, loss_final=0.459, loss_mean=0.797, loss_mean_cls=0.0693, proj_loss=-0.408][2026-03-26 13:00:39] Step: 7135, Training Logs: loss_final: 0.483156, loss_mean: 0.832430, proj_loss: -0.416604, loss_mean_cls: 0.067330, grad_norm: 0.662414 +Steps: 1%| | 7136/1000000 [29:36<68:16:05, 4.04it/s, grad_norm=0.662, loss_final=0.483, loss_mean=0.832, loss_mean_cls=0.0673, proj_loss=-0.417][2026-03-26 13:00:39] Step: 7136, Training Logs: loss_final: 0.485497, loss_mean: 0.819503, proj_loss: -0.403945, loss_mean_cls: 0.069939, grad_norm: 0.212899 +Steps: 1%| | 7137/1000000 [29:36<68:15:38, 4.04it/s, grad_norm=0.213, loss_final=0.485, loss_mean=0.82, loss_mean_cls=0.0699, proj_loss=-0.404][2026-03-26 13:00:40] Step: 7137, Training Logs: loss_final: 0.478271, loss_mean: 0.830728, proj_loss: -0.419947, loss_mean_cls: 0.067490, grad_norm: 0.808425 +Steps: 1%| | 7138/1000000 [29:37<68:18:50, 4.04it/s, grad_norm=0.808, loss_final=0.478, loss_mean=0.831, loss_mean_cls=0.0675, proj_loss=-0.42][2026-03-26 13:00:40] Step: 7138, Training Logs: loss_final: 0.474070, loss_mean: 0.814850, proj_loss: -0.409948, loss_mean_cls: 0.069169, grad_norm: 0.453588 +Steps: 1%| | 7139/1000000 [29:37<68:19:08, 4.04it/s, grad_norm=0.454, loss_final=0.474, loss_mean=0.815, loss_mean_cls=0.0692, proj_loss=-0.41][2026-03-26 13:00:40] Step: 7139, Training Logs: loss_final: 0.477743, loss_mean: 0.823574, proj_loss: -0.413853, loss_mean_cls: 0.068022, grad_norm: 0.697504 +Steps: 1%| | 7140/1000000 [29:37<68:16:38, 4.04it/s, grad_norm=0.698, loss_final=0.478, loss_mean=0.824, loss_mean_cls=0.068, proj_loss=-0.414][2026-03-26 13:00:40] Step: 7140, Training Logs: loss_final: 0.469112, loss_mean: 0.814777, proj_loss: -0.413174, loss_mean_cls: 0.067509, grad_norm: 0.590068 +Steps: 1%| | 7141/1000000 [29:37<68:18:31, 4.04it/s, grad_norm=0.59, loss_final=0.469, loss_mean=0.815, loss_mean_cls=0.0675, proj_loss=-0.413][2026-03-26 13:00:41] Step: 7141, Training Logs: loss_final: 0.477121, loss_mean: 0.820904, proj_loss: -0.412546, loss_mean_cls: 0.068763, grad_norm: 0.398680 +Steps: 1%| | 7142/1000000 [29:37<68:16:58, 4.04it/s, grad_norm=0.399, loss_final=0.477, loss_mean=0.821, loss_mean_cls=0.0688, proj_loss=-0.413][2026-03-26 13:00:41] Step: 7142, Training Logs: loss_final: 0.464881, loss_mean: 0.811232, proj_loss: -0.414057, loss_mean_cls: 0.067707, grad_norm: 0.510326 +Steps: 1%| | 7143/1000000 [29:38<68:17:26, 4.04it/s, grad_norm=0.51, loss_final=0.465, loss_mean=0.811, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 13:00:41] Step: 7143, Training Logs: loss_final: 0.456967, loss_mean: 0.812056, proj_loss: -0.421010, loss_mean_cls: 0.065920, grad_norm: 0.336750 +Steps: 1%| | 7144/1000000 [29:38<68:14:34, 4.04it/s, grad_norm=0.337, loss_final=0.457, loss_mean=0.812, loss_mean_cls=0.0659, proj_loss=-0.421][2026-03-26 13:00:41] Step: 7144, Training Logs: loss_final: 0.464986, loss_mean: 0.806600, proj_loss: -0.411181, loss_mean_cls: 0.069567, grad_norm: 0.451140 +Steps: 1%| | 7145/1000000 [29:38<68:16:09, 4.04it/s, grad_norm=0.451, loss_final=0.465, loss_mean=0.807, loss_mean_cls=0.0696, proj_loss=-0.411][2026-03-26 13:00:42] Step: 7145, Training Logs: loss_final: 0.482108, loss_mean: 0.838658, proj_loss: -0.422791, loss_mean_cls: 0.066241, grad_norm: 0.343299 +Steps: 1%| | 7146/1000000 [29:38<68:16:25, 4.04it/s, grad_norm=0.343, loss_final=0.482, loss_mean=0.839, loss_mean_cls=0.0662, proj_loss=-0.423][2026-03-26 13:00:42] Step: 7146, Training Logs: loss_final: 0.489200, loss_mean: 0.833411, proj_loss: -0.412140, loss_mean_cls: 0.067929, grad_norm: 0.597078 +Steps: 1%| | 7147/1000000 [29:39<68:14:44, 4.04it/s, grad_norm=0.597, loss_final=0.489, loss_mean=0.833, loss_mean_cls=0.0679, proj_loss=-0.412][2026-03-26 13:00:42] Step: 7147, Training Logs: loss_final: 0.465555, loss_mean: 0.826476, proj_loss: -0.425992, loss_mean_cls: 0.065072, grad_norm: 0.263947 +Steps: 1%| | 7148/1000000 [29:39<68:15:14, 4.04it/s, grad_norm=0.264, loss_final=0.466, loss_mean=0.826, loss_mean_cls=0.0651, proj_loss=-0.426][2026-03-26 13:00:42] Step: 7148, Training Logs: loss_final: 0.465814, loss_mean: 0.812451, proj_loss: -0.413913, loss_mean_cls: 0.067276, grad_norm: 0.436859 +Steps: 1%| | 7149/1000000 [29:39<68:13:55, 4.04it/s, grad_norm=0.437, loss_final=0.466, loss_mean=0.812, loss_mean_cls=0.0673, proj_loss=-0.414][2026-03-26 13:00:43] Step: 7149, Training Logs: loss_final: 0.453202, loss_mean: 0.804045, proj_loss: -0.417888, loss_mean_cls: 0.067045, grad_norm: 0.308351 +Steps: 1%| | 7150/1000000 [29:39<68:13:36, 4.04it/s, grad_norm=0.308, loss_final=0.453, loss_mean=0.804, loss_mean_cls=0.067, proj_loss=-0.418][2026-03-26 13:00:43] Step: 7150, Training Logs: loss_final: 0.467112, loss_mean: 0.807505, proj_loss: -0.409949, loss_mean_cls: 0.069556, grad_norm: 0.437080 +Steps: 1%| | 7151/1000000 [29:40<68:15:35, 4.04it/s, grad_norm=0.437, loss_final=0.467, loss_mean=0.808, loss_mean_cls=0.0696, proj_loss=-0.41][2026-03-26 13:00:43] Step: 7151, Training Logs: loss_final: 0.457124, loss_mean: 0.810060, proj_loss: -0.420773, loss_mean_cls: 0.067837, grad_norm: 0.390492 +Steps: 1%| | 7152/1000000 [29:40<68:13:49, 4.04it/s, grad_norm=0.39, loss_final=0.457, loss_mean=0.81, loss_mean_cls=0.0678, proj_loss=-0.421][2026-03-26 13:00:43] Step: 7152, Training Logs: loss_final: 0.475744, loss_mean: 0.818753, proj_loss: -0.410513, loss_mean_cls: 0.067504, grad_norm: 0.325836 +Steps: 1%| | 7153/1000000 [29:40<68:13:25, 4.04it/s, grad_norm=0.326, loss_final=0.476, loss_mean=0.819, loss_mean_cls=0.0675, proj_loss=-0.411][2026-03-26 13:00:44] Step: 7153, Training Logs: loss_final: 0.482142, loss_mean: 0.835203, proj_loss: -0.419451, loss_mean_cls: 0.066390, grad_norm: 0.347856 +Steps: 1%| | 7154/1000000 [29:40<68:14:47, 4.04it/s, grad_norm=0.348, loss_final=0.482, loss_mean=0.835, loss_mean_cls=0.0664, proj_loss=-0.419][2026-03-26 13:00:44] Step: 7154, Training Logs: loss_final: 0.482715, loss_mean: 0.823734, proj_loss: -0.409260, loss_mean_cls: 0.068241, grad_norm: 0.375426 +Steps: 1%| | 7155/1000000 [29:41<68:14:50, 4.04it/s, grad_norm=0.375, loss_final=0.483, loss_mean=0.824, loss_mean_cls=0.0682, proj_loss=-0.409][2026-03-26 13:00:44] Step: 7155, Training Logs: loss_final: 0.470212, loss_mean: 0.818205, proj_loss: -0.415348, loss_mean_cls: 0.067355, grad_norm: 0.357104 +Steps: 1%| | 7156/1000000 [29:41<68:12:36, 4.04it/s, grad_norm=0.357, loss_final=0.47, loss_mean=0.818, loss_mean_cls=0.0674, proj_loss=-0.415][2026-03-26 13:00:44] Step: 7156, Training Logs: loss_final: 0.467093, loss_mean: 0.816869, proj_loss: -0.417061, loss_mean_cls: 0.067285, grad_norm: 0.445699 +Steps: 1%| | 7157/1000000 [29:41<68:14:02, 4.04it/s, grad_norm=0.446, loss_final=0.467, loss_mean=0.817, loss_mean_cls=0.0673, proj_loss=-0.417][2026-03-26 13:00:45] Step: 7157, Training Logs: loss_final: 0.469405, loss_mean: 0.817618, proj_loss: -0.415171, loss_mean_cls: 0.066958, grad_norm: 0.298332 +Steps: 1%| | 7158/1000000 [29:41<68:17:10, 4.04it/s, grad_norm=0.298, loss_final=0.469, loss_mean=0.818, loss_mean_cls=0.067, proj_loss=-0.415][2026-03-26 13:00:45] Step: 7158, Training Logs: loss_final: 0.476834, loss_mean: 0.836040, proj_loss: -0.424990, loss_mean_cls: 0.065784, grad_norm: 0.520511 +Steps: 1%| | 7159/1000000 [29:42<68:23:08, 4.03it/s, grad_norm=0.521, loss_final=0.477, loss_mean=0.836, loss_mean_cls=0.0658, proj_loss=-0.425][2026-03-26 13:00:45] Step: 7159, Training Logs: loss_final: 0.468849, loss_mean: 0.831992, proj_loss: -0.427684, loss_mean_cls: 0.064542, grad_norm: 0.641939 +Steps: 1%| | 7160/1000000 [29:42<69:01:08, 4.00it/s, grad_norm=0.642, loss_final=0.469, loss_mean=0.832, loss_mean_cls=0.0645, proj_loss=-0.428][2026-03-26 13:00:45] Step: 7160, Training Logs: loss_final: 0.472516, loss_mean: 0.825349, proj_loss: -0.419178, loss_mean_cls: 0.066345, grad_norm: 0.196319 +Steps: 1%| | 7161/1000000 [29:42<68:46:31, 4.01it/s, grad_norm=0.196, loss_final=0.473, loss_mean=0.825, loss_mean_cls=0.0663, proj_loss=-0.419][2026-03-26 13:00:46] Step: 7161, Training Logs: loss_final: 0.469957, loss_mean: 0.824300, proj_loss: -0.420895, loss_mean_cls: 0.066552, grad_norm: 0.557554 +Steps: 1%| | 7162/1000000 [29:42<68:39:08, 4.02it/s, grad_norm=0.558, loss_final=0.47, loss_mean=0.824, loss_mean_cls=0.0666, proj_loss=-0.421][2026-03-26 13:00:46] Step: 7162, Training Logs: loss_final: 0.455335, loss_mean: 0.806631, proj_loss: -0.419110, loss_mean_cls: 0.067814, grad_norm: 0.314914 +Steps: 1%| | 7163/1000000 [29:43<68:31:13, 4.02it/s, grad_norm=0.315, loss_final=0.455, loss_mean=0.807, loss_mean_cls=0.0678, proj_loss=-0.419][2026-03-26 13:00:46] Step: 7163, Training Logs: loss_final: 0.487733, loss_mean: 0.843491, proj_loss: -0.421983, loss_mean_cls: 0.066225, grad_norm: 0.632879 +Steps: 1%| | 7164/1000000 [29:43<68:26:52, 4.03it/s, grad_norm=0.633, loss_final=0.488, loss_mean=0.843, loss_mean_cls=0.0662, proj_loss=-0.422][2026-03-26 13:00:46] Step: 7164, Training Logs: loss_final: 0.448936, loss_mean: 0.800771, proj_loss: -0.417974, loss_mean_cls: 0.066139, grad_norm: 0.310338 +Steps: 1%| | 7165/1000000 [29:43<68:20:00, 4.04it/s, grad_norm=0.31, loss_final=0.449, loss_mean=0.801, loss_mean_cls=0.0661, proj_loss=-0.418][2026-03-26 13:00:47] Step: 7165, Training Logs: loss_final: 0.462634, loss_mean: 0.804889, proj_loss: -0.411299, loss_mean_cls: 0.069044, grad_norm: 0.595688 +Steps: 1%| | 7166/1000000 [29:43<68:20:23, 4.04it/s, grad_norm=0.596, loss_final=0.463, loss_mean=0.805, loss_mean_cls=0.069, proj_loss=-0.411][2026-03-26 13:00:47] Step: 7166, Training Logs: loss_final: 0.473305, loss_mean: 0.816216, proj_loss: -0.411629, loss_mean_cls: 0.068718, grad_norm: 0.554895 +Steps: 1%| | 7167/1000000 [29:44<70:30:56, 3.91it/s, grad_norm=0.555, loss_final=0.473, loss_mean=0.816, loss_mean_cls=0.0687, proj_loss=-0.412][2026-03-26 13:00:47] Step: 7167, Training Logs: loss_final: 0.472421, loss_mean: 0.821288, proj_loss: -0.416602, loss_mean_cls: 0.067734, grad_norm: 0.337306 +Steps: 1%| | 7168/1000000 [29:44<70:10:16, 3.93it/s, grad_norm=0.337, loss_final=0.472, loss_mean=0.821, loss_mean_cls=0.0677, proj_loss=-0.417][2026-03-26 13:00:47] Step: 7168, Training Logs: loss_final: 0.449288, loss_mean: 0.800421, proj_loss: -0.418520, loss_mean_cls: 0.067387, grad_norm: 0.464075 +Steps: 1%| | 7169/1000000 [29:44<69:35:51, 3.96it/s, grad_norm=0.464, loss_final=0.449, loss_mean=0.8, loss_mean_cls=0.0674, proj_loss=-0.419][2026-03-26 13:00:48] Step: 7169, Training Logs: loss_final: 0.459926, loss_mean: 0.812638, proj_loss: -0.420881, loss_mean_cls: 0.068169, grad_norm: 0.272881 +Steps: 1%| | 7170/1000000 [29:44<69:14:21, 3.98it/s, grad_norm=0.273, loss_final=0.46, loss_mean=0.813, loss_mean_cls=0.0682, proj_loss=-0.421][2026-03-26 13:00:48] Step: 7170, Training Logs: loss_final: 0.510252, loss_mean: 0.840751, proj_loss: -0.401307, loss_mean_cls: 0.070808, grad_norm: 0.410042 +Steps: 1%| | 7171/1000000 [29:45<68:55:53, 4.00it/s, grad_norm=0.41, loss_final=0.51, loss_mean=0.841, loss_mean_cls=0.0708, proj_loss=-0.401][2026-03-26 13:00:48] Step: 7171, Training Logs: loss_final: 0.475535, loss_mean: 0.830882, proj_loss: -0.421032, loss_mean_cls: 0.065685, grad_norm: 0.213657 +Steps: 1%| | 7172/1000000 [29:45<68:44:38, 4.01it/s, grad_norm=0.214, loss_final=0.476, loss_mean=0.831, loss_mean_cls=0.0657, proj_loss=-0.421][2026-03-26 13:00:48] Step: 7172, Training Logs: loss_final: 0.471354, loss_mean: 0.822196, proj_loss: -0.418117, loss_mean_cls: 0.067275, grad_norm: 0.311413 +Steps: 1%| | 7173/1000000 [29:45<68:33:59, 4.02it/s, grad_norm=0.311, loss_final=0.471, loss_mean=0.822, loss_mean_cls=0.0673, proj_loss=-0.418][2026-03-26 13:00:49] Step: 7173, Training Logs: loss_final: 0.466507, loss_mean: 0.814313, proj_loss: -0.414969, loss_mean_cls: 0.067163, grad_norm: 0.202832 +Steps: 1%| | 7174/1000000 [29:45<68:30:09, 4.03it/s, grad_norm=0.203, loss_final=0.467, loss_mean=0.814, loss_mean_cls=0.0672, proj_loss=-0.415][2026-03-26 13:00:49] Step: 7174, Training Logs: loss_final: 0.466090, loss_mean: 0.819095, proj_loss: -0.419352, loss_mean_cls: 0.066347, grad_norm: 0.488755 +Steps: 1%| | 7175/1000000 [29:46<68:24:09, 4.03it/s, grad_norm=0.489, loss_final=0.466, loss_mean=0.819, loss_mean_cls=0.0663, proj_loss=-0.419][2026-03-26 13:00:49] Step: 7175, Training Logs: loss_final: 0.466897, loss_mean: 0.820764, proj_loss: -0.419987, loss_mean_cls: 0.066120, grad_norm: 0.459059 +Steps: 1%| | 7176/1000000 [29:46<68:22:06, 4.03it/s, grad_norm=0.459, loss_final=0.467, loss_mean=0.821, loss_mean_cls=0.0661, proj_loss=-0.42][2026-03-26 13:00:49] Step: 7176, Training Logs: loss_final: 0.466545, loss_mean: 0.816329, proj_loss: -0.417592, loss_mean_cls: 0.067809, grad_norm: 0.304496 +Steps: 1%| | 7177/1000000 [29:46<68:18:36, 4.04it/s, grad_norm=0.304, loss_final=0.467, loss_mean=0.816, loss_mean_cls=0.0678, proj_loss=-0.418][2026-03-26 13:00:50] Step: 7177, Training Logs: loss_final: 0.473713, loss_mean: 0.821530, proj_loss: -0.414945, loss_mean_cls: 0.067128, grad_norm: 0.424133 +Steps: 1%| | 7178/1000000 [29:46<68:21:47, 4.03it/s, grad_norm=0.424, loss_final=0.474, loss_mean=0.822, loss_mean_cls=0.0671, proj_loss=-0.415][2026-03-26 13:00:50] Step: 7178, Training Logs: loss_final: 0.461698, loss_mean: 0.798758, proj_loss: -0.407315, loss_mean_cls: 0.070256, grad_norm: 0.241314 +Steps: 1%| | 7179/1000000 [29:47<68:19:05, 4.04it/s, grad_norm=0.241, loss_final=0.462, loss_mean=0.799, loss_mean_cls=0.0703, proj_loss=-0.407][2026-03-26 13:00:50] Step: 7179, Training Logs: loss_final: 0.480088, loss_mean: 0.827706, proj_loss: -0.414465, loss_mean_cls: 0.066847, grad_norm: 0.382024 +Steps: 1%| | 7180/1000000 [29:47<68:17:52, 4.04it/s, grad_norm=0.382, loss_final=0.48, loss_mean=0.828, loss_mean_cls=0.0668, proj_loss=-0.414][2026-03-26 13:00:50] Step: 7180, Training Logs: loss_final: 0.493349, loss_mean: 0.832898, proj_loss: -0.408584, loss_mean_cls: 0.069034, grad_norm: 0.312238 +Steps: 1%| | 7181/1000000 [29:47<68:16:45, 4.04it/s, grad_norm=0.312, loss_final=0.493, loss_mean=0.833, loss_mean_cls=0.069, proj_loss=-0.409][2026-03-26 13:00:51] Step: 7181, Training Logs: loss_final: 0.459727, loss_mean: 0.804246, proj_loss: -0.412652, loss_mean_cls: 0.068133, grad_norm: 0.336918 +Steps: 1%| | 7182/1000000 [29:47<68:15:20, 4.04it/s, grad_norm=0.337, loss_final=0.46, loss_mean=0.804, loss_mean_cls=0.0681, proj_loss=-0.413][2026-03-26 13:00:51] Step: 7182, Training Logs: loss_final: 0.463063, loss_mean: 0.813107, proj_loss: -0.417277, loss_mean_cls: 0.067234, grad_norm: 0.302318 +Steps: 1%| | 7183/1000000 [29:48<68:15:38, 4.04it/s, grad_norm=0.302, loss_final=0.463, loss_mean=0.813, loss_mean_cls=0.0672, proj_loss=-0.417][2026-03-26 13:00:51] Step: 7183, Training Logs: loss_final: 0.480961, loss_mean: 0.829520, proj_loss: -0.416529, loss_mean_cls: 0.067970, grad_norm: 0.250661 +Steps: 1%| | 7184/1000000 [29:48<68:16:28, 4.04it/s, grad_norm=0.251, loss_final=0.481, loss_mean=0.83, loss_mean_cls=0.068, proj_loss=-0.417][2026-03-26 13:00:51] Step: 7184, Training Logs: loss_final: 0.461173, loss_mean: 0.805234, proj_loss: -0.413267, loss_mean_cls: 0.069205, grad_norm: 0.537390 +Steps: 1%| | 7185/1000000 [29:48<68:15:46, 4.04it/s, grad_norm=0.537, loss_final=0.461, loss_mean=0.805, loss_mean_cls=0.0692, proj_loss=-0.413][2026-03-26 13:00:52] Step: 7185, Training Logs: loss_final: 0.480613, loss_mean: 0.832005, proj_loss: -0.419000, loss_mean_cls: 0.067608, grad_norm: 0.266742 +Steps: 1%| | 7186/1000000 [29:48<68:17:35, 4.04it/s, grad_norm=0.267, loss_final=0.481, loss_mean=0.832, loss_mean_cls=0.0676, proj_loss=-0.419][2026-03-26 13:00:52] Step: 7186, Training Logs: loss_final: 0.480160, loss_mean: 0.826320, proj_loss: -0.413886, loss_mean_cls: 0.067727, grad_norm: 0.354128 +Steps: 1%| | 7187/1000000 [29:49<68:16:49, 4.04it/s, grad_norm=0.354, loss_final=0.48, loss_mean=0.826, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 13:00:52] Step: 7187, Training Logs: loss_final: 0.450344, loss_mean: 0.802073, proj_loss: -0.418798, loss_mean_cls: 0.067069, grad_norm: 0.333728 +Steps: 1%| | 7188/1000000 [29:49<68:15:14, 4.04it/s, grad_norm=0.334, loss_final=0.45, loss_mean=0.802, loss_mean_cls=0.0671, proj_loss=-0.419][2026-03-26 13:00:52] Step: 7188, Training Logs: loss_final: 0.450432, loss_mean: 0.801638, proj_loss: -0.418017, loss_mean_cls: 0.066811, grad_norm: 0.346827 +Steps: 1%| | 7189/1000000 [29:49<68:14:58, 4.04it/s, grad_norm=0.347, loss_final=0.45, loss_mean=0.802, loss_mean_cls=0.0668, proj_loss=-0.418][2026-03-26 13:00:53] Step: 7189, Training Logs: loss_final: 0.447852, loss_mean: 0.804921, proj_loss: -0.422929, loss_mean_cls: 0.065859, grad_norm: 0.369085 +Steps: 1%| | 7190/1000000 [29:49<68:15:09, 4.04it/s, grad_norm=0.369, loss_final=0.448, loss_mean=0.805, loss_mean_cls=0.0659, proj_loss=-0.423][2026-03-26 13:00:53] Step: 7190, Training Logs: loss_final: 0.468152, loss_mean: 0.816206, proj_loss: -0.415359, loss_mean_cls: 0.067305, grad_norm: 0.355395 +Steps: 1%| | 7191/1000000 [29:50<68:18:01, 4.04it/s, grad_norm=0.355, loss_final=0.468, loss_mean=0.816, loss_mean_cls=0.0673, proj_loss=-0.415][2026-03-26 13:00:53] Step: 7191, Training Logs: loss_final: 0.459878, loss_mean: 0.822494, proj_loss: -0.427964, loss_mean_cls: 0.065348, grad_norm: 0.325182 +Steps: 1%| | 7192/1000000 [29:50<68:18:07, 4.04it/s, grad_norm=0.325, loss_final=0.46, loss_mean=0.822, loss_mean_cls=0.0653, proj_loss=-0.428][2026-03-26 13:00:53] Step: 7192, Training Logs: loss_final: 0.454542, loss_mean: 0.809539, proj_loss: -0.421870, loss_mean_cls: 0.066873, grad_norm: 0.605552 +Steps: 1%| | 7193/1000000 [29:50<68:17:11, 4.04it/s, grad_norm=0.606, loss_final=0.455, loss_mean=0.81, loss_mean_cls=0.0669, proj_loss=-0.422][2026-03-26 13:00:54] Step: 7193, Training Logs: loss_final: 0.471681, loss_mean: 0.829728, proj_loss: -0.424145, loss_mean_cls: 0.066098, grad_norm: 0.417333 +Steps: 1%| | 7194/1000000 [29:50<68:17:27, 4.04it/s, grad_norm=0.417, loss_final=0.472, loss_mean=0.83, loss_mean_cls=0.0661, proj_loss=-0.424][2026-03-26 13:00:54] Step: 7194, Training Logs: loss_final: 0.474057, loss_mean: 0.821608, proj_loss: -0.415804, loss_mean_cls: 0.068254, grad_norm: 0.540384 +Steps: 1%| | 7195/1000000 [29:51<68:17:19, 4.04it/s, grad_norm=0.54, loss_final=0.474, loss_mean=0.822, loss_mean_cls=0.0683, proj_loss=-0.416][2026-03-26 13:00:54] Step: 7195, Training Logs: loss_final: 0.454999, loss_mean: 0.798116, proj_loss: -0.412488, loss_mean_cls: 0.069372, grad_norm: 0.238627 +Steps: 1%| | 7196/1000000 [29:51<68:17:55, 4.04it/s, grad_norm=0.239, loss_final=0.455, loss_mean=0.798, loss_mean_cls=0.0694, proj_loss=-0.412][2026-03-26 13:00:54] Step: 7196, Training Logs: loss_final: 0.488724, loss_mean: 0.819560, proj_loss: -0.401681, loss_mean_cls: 0.070845, grad_norm: 0.435104 +Steps: 1%| | 7197/1000000 [29:51<68:16:34, 4.04it/s, grad_norm=0.435, loss_final=0.489, loss_mean=0.82, loss_mean_cls=0.0708, proj_loss=-0.402][2026-03-26 13:00:55] Step: 7197, Training Logs: loss_final: 0.464401, loss_mean: 0.812937, proj_loss: -0.416950, loss_mean_cls: 0.068414, grad_norm: 0.318319 +Steps: 1%| | 7198/1000000 [29:51<68:14:36, 4.04it/s, grad_norm=0.318, loss_final=0.464, loss_mean=0.813, loss_mean_cls=0.0684, proj_loss=-0.417][2026-03-26 13:00:55] Step: 7198, Training Logs: loss_final: 0.469938, loss_mean: 0.814297, proj_loss: -0.411522, loss_mean_cls: 0.067164, grad_norm: 0.412915 +Steps: 1%| | 7199/1000000 [29:52<68:16:03, 4.04it/s, grad_norm=0.413, loss_final=0.47, loss_mean=0.814, loss_mean_cls=0.0672, proj_loss=-0.412][2026-03-26 13:00:55] Step: 7199, Training Logs: loss_final: 0.453176, loss_mean: 0.803237, proj_loss: -0.417382, loss_mean_cls: 0.067320, grad_norm: 0.265262 +Steps: 1%| | 7200/1000000 [29:52<68:16:40, 4.04it/s, grad_norm=0.265, loss_final=0.453, loss_mean=0.803, loss_mean_cls=0.0673, proj_loss=-0.417][2026-03-26 13:00:55] Step: 7200, Training Logs: loss_final: 0.472266, loss_mean: 0.815461, proj_loss: -0.411637, loss_mean_cls: 0.068442, grad_norm: 0.521744 +Steps: 1%| | 7201/1000000 [29:52<68:14:45, 4.04it/s, grad_norm=0.522, loss_final=0.472, loss_mean=0.815, loss_mean_cls=0.0684, proj_loss=-0.412][2026-03-26 13:00:56] Step: 7201, Training Logs: loss_final: 0.468959, loss_mean: 0.815779, proj_loss: -0.415073, loss_mean_cls: 0.068254, grad_norm: 0.458405 +Steps: 1%| | 7202/1000000 [29:52<68:15:30, 4.04it/s, grad_norm=0.458, loss_final=0.469, loss_mean=0.816, loss_mean_cls=0.0683, proj_loss=-0.415][2026-03-26 13:00:56] Step: 7202, Training Logs: loss_final: 0.481888, loss_mean: 0.826861, proj_loss: -0.412930, loss_mean_cls: 0.067957, grad_norm: 0.311072 +Steps: 1%| | 7203/1000000 [29:53<68:16:19, 4.04it/s, grad_norm=0.311, loss_final=0.482, loss_mean=0.827, loss_mean_cls=0.068, proj_loss=-0.413][2026-03-26 13:00:56] Step: 7203, Training Logs: loss_final: 0.472329, loss_mean: 0.817136, proj_loss: -0.413580, loss_mean_cls: 0.068773, grad_norm: 0.426397 +Steps: 1%| | 7204/1000000 [29:53<68:15:01, 4.04it/s, grad_norm=0.426, loss_final=0.472, loss_mean=0.817, loss_mean_cls=0.0688, proj_loss=-0.414][2026-03-26 13:00:56] Step: 7204, Training Logs: loss_final: 0.473665, loss_mean: 0.818336, proj_loss: -0.413175, loss_mean_cls: 0.068503, grad_norm: 0.245613 +Steps: 1%| | 7205/1000000 [29:53<68:52:47, 4.00it/s, grad_norm=0.246, loss_final=0.474, loss_mean=0.818, loss_mean_cls=0.0685, proj_loss=-0.413][2026-03-26 13:00:57] Step: 7205, Training Logs: loss_final: 0.463053, loss_mean: 0.816033, proj_loss: -0.419436, loss_mean_cls: 0.066456, grad_norm: 0.342719 +Steps: 1%| | 7206/1000000 [29:53<69:23:27, 3.97it/s, grad_norm=0.343, loss_final=0.463, loss_mean=0.816, loss_mean_cls=0.0665, proj_loss=-0.419][2026-03-26 13:00:57] Step: 7206, Training Logs: loss_final: 0.467492, loss_mean: 0.818127, proj_loss: -0.417435, loss_mean_cls: 0.066800, grad_norm: 0.199416 +Steps: 1%| | 7207/1000000 [29:54<69:03:19, 3.99it/s, grad_norm=0.199, loss_final=0.467, loss_mean=0.818, loss_mean_cls=0.0668, proj_loss=-0.417][2026-03-26 13:00:57] Step: 7207, Training Logs: loss_final: 0.468682, loss_mean: 0.821424, proj_loss: -0.420546, loss_mean_cls: 0.067803, grad_norm: 0.256262 +Steps: 1%| | 7208/1000000 [29:54<68:49:15, 4.01it/s, grad_norm=0.256, loss_final=0.469, loss_mean=0.821, loss_mean_cls=0.0678, proj_loss=-0.421][2026-03-26 13:00:57] Step: 7208, Training Logs: loss_final: 0.479286, loss_mean: 0.836814, proj_loss: -0.424135, loss_mean_cls: 0.066607, grad_norm: 0.254769 +Steps: 1%| | 7209/1000000 [29:54<68:41:14, 4.01it/s, grad_norm=0.255, loss_final=0.479, loss_mean=0.837, loss_mean_cls=0.0666, proj_loss=-0.424][2026-03-26 13:00:58] Step: 7209, Training Logs: loss_final: 0.486116, loss_mean: 0.810503, proj_loss: -0.396568, loss_mean_cls: 0.072181, grad_norm: 0.283659 +Steps: 1%| | 7210/1000000 [29:54<68:31:42, 4.02it/s, grad_norm=0.284, loss_final=0.486, loss_mean=0.811, loss_mean_cls=0.0722, proj_loss=-0.397][2026-03-26 13:00:58] Step: 7210, Training Logs: loss_final: 0.475755, loss_mean: 0.815214, proj_loss: -0.409632, loss_mean_cls: 0.070174, grad_norm: 0.318132 +Steps: 1%| | 7211/1000000 [29:55<68:26:11, 4.03it/s, grad_norm=0.318, loss_final=0.476, loss_mean=0.815, loss_mean_cls=0.0702, proj_loss=-0.41][2026-03-26 13:00:58] Step: 7211, Training Logs: loss_final: 0.470843, loss_mean: 0.816557, proj_loss: -0.414018, loss_mean_cls: 0.068304, grad_norm: 0.299410 +Steps: 1%| | 7212/1000000 [29:55<68:23:46, 4.03it/s, grad_norm=0.299, loss_final=0.471, loss_mean=0.817, loss_mean_cls=0.0683, proj_loss=-0.414][2026-03-26 13:00:58] Step: 7212, Training Logs: loss_final: 0.462515, loss_mean: 0.813901, proj_loss: -0.418530, loss_mean_cls: 0.067144, grad_norm: 0.340702 +Steps: 1%| | 7213/1000000 [29:55<68:21:52, 4.03it/s, grad_norm=0.341, loss_final=0.463, loss_mean=0.814, loss_mean_cls=0.0671, proj_loss=-0.419][2026-03-26 13:00:59] Step: 7213, Training Logs: loss_final: 0.468871, loss_mean: 0.818892, proj_loss: -0.417778, loss_mean_cls: 0.067757, grad_norm: 0.235525 +Steps: 1%| | 7214/1000000 [29:55<68:18:54, 4.04it/s, grad_norm=0.236, loss_final=0.469, loss_mean=0.819, loss_mean_cls=0.0678, proj_loss=-0.418][2026-03-26 13:00:59] Step: 7214, Training Logs: loss_final: 0.473920, loss_mean: 0.829914, proj_loss: -0.422068, loss_mean_cls: 0.066074, grad_norm: 0.234602 +Steps: 1%| | 7215/1000000 [29:56<68:24:20, 4.03it/s, grad_norm=0.235, loss_final=0.474, loss_mean=0.83, loss_mean_cls=0.0661, proj_loss=-0.422][2026-03-26 13:00:59] Step: 7215, Training Logs: loss_final: 0.475265, loss_mean: 0.834473, proj_loss: -0.424452, loss_mean_cls: 0.065244, grad_norm: 0.198317 +Steps: 1%| | 7216/1000000 [29:56<68:15:55, 4.04it/s, grad_norm=0.198, loss_final=0.475, loss_mean=0.834, loss_mean_cls=0.0652, proj_loss=-0.424][2026-03-26 13:00:59] Step: 7216, Training Logs: loss_final: 0.478738, loss_mean: 0.836661, proj_loss: -0.423390, loss_mean_cls: 0.065467, grad_norm: 0.286372 +Steps: 1%| | 7217/1000000 [29:56<68:17:49, 4.04it/s, grad_norm=0.286, loss_final=0.479, loss_mean=0.837, loss_mean_cls=0.0655, proj_loss=-0.423][2026-03-26 13:01:00] Step: 7217, Training Logs: loss_final: 0.458269, loss_mean: 0.813167, proj_loss: -0.422153, loss_mean_cls: 0.067255, grad_norm: 0.272056 +Steps: 1%| | 7218/1000000 [29:56<68:18:31, 4.04it/s, grad_norm=0.272, loss_final=0.458, loss_mean=0.813, loss_mean_cls=0.0673, proj_loss=-0.422][2026-03-26 13:01:00] Step: 7218, Training Logs: loss_final: 0.462600, loss_mean: 0.807246, proj_loss: -0.413975, loss_mean_cls: 0.069329, grad_norm: 0.424967 +Steps: 1%| | 7219/1000000 [29:57<96:28:37, 2.86it/s, grad_norm=0.425, loss_final=0.463, loss_mean=0.807, loss_mean_cls=0.0693, proj_loss=-0.414][2026-03-26 13:01:00] Step: 7219, Training Logs: loss_final: 0.463300, loss_mean: 0.811805, proj_loss: -0.416628, loss_mean_cls: 0.068122, grad_norm: 0.252817 +Steps: 1%| | 7220/1000000 [29:57<87:59:34, 3.13it/s, grad_norm=0.253, loss_final=0.463, loss_mean=0.812, loss_mean_cls=0.0681, proj_loss=-0.417][2026-03-26 13:01:01] Step: 7220, Training Logs: loss_final: 0.487420, loss_mean: 0.823544, proj_loss: -0.405918, loss_mean_cls: 0.069793, grad_norm: 0.598920 +Steps: 1%| | 7221/1000000 [29:57<82:02:22, 3.36it/s, grad_norm=0.599, loss_final=0.487, loss_mean=0.824, loss_mean_cls=0.0698, proj_loss=-0.406][2026-03-26 13:01:01] Step: 7221, Training Logs: loss_final: 0.469895, loss_mean: 0.812163, proj_loss: -0.412034, loss_mean_cls: 0.069765, grad_norm: 0.688081 +Steps: 1%| | 7222/1000000 [29:58<77:54:05, 3.54it/s, grad_norm=0.688, loss_final=0.47, loss_mean=0.812, loss_mean_cls=0.0698, proj_loss=-0.412][2026-03-26 13:01:01] Step: 7222, Training Logs: loss_final: 0.465554, loss_mean: 0.820812, proj_loss: -0.421370, loss_mean_cls: 0.066112, grad_norm: 0.270851 +Steps: 1%| | 7223/1000000 [29:58<75:01:02, 3.68it/s, grad_norm=0.271, loss_final=0.466, loss_mean=0.821, loss_mean_cls=0.0661, proj_loss=-0.421][2026-03-26 13:01:01] Step: 7223, Training Logs: loss_final: 0.480705, loss_mean: 0.817363, proj_loss: -0.406644, loss_mean_cls: 0.069987, grad_norm: 0.721036 +Steps: 1%| | 7224/1000000 [29:58<73:02:01, 3.78it/s, grad_norm=0.721, loss_final=0.481, loss_mean=0.817, loss_mean_cls=0.07, proj_loss=-0.407][2026-03-26 13:01:02] Step: 7224, Training Logs: loss_final: 0.476252, loss_mean: 0.821295, proj_loss: -0.412918, loss_mean_cls: 0.067874, grad_norm: 0.335703 +Steps: 1%| | 7225/1000000 [29:58<71:32:29, 3.85it/s, grad_norm=0.336, loss_final=0.476, loss_mean=0.821, loss_mean_cls=0.0679, proj_loss=-0.413][2026-03-26 13:01:02] Step: 7225, Training Logs: loss_final: 0.454040, loss_mean: 0.810003, proj_loss: -0.422764, loss_mean_cls: 0.066801, grad_norm: 0.515071 +Steps: 1%| | 7226/1000000 [29:59<70:32:29, 3.91it/s, grad_norm=0.515, loss_final=0.454, loss_mean=0.81, loss_mean_cls=0.0668, proj_loss=-0.423][2026-03-26 13:01:02] Step: 7226, Training Logs: loss_final: 0.488133, loss_mean: 0.846647, proj_loss: -0.424836, loss_mean_cls: 0.066322, grad_norm: 0.394684 +Steps: 1%| | 7227/1000000 [29:59<69:50:15, 3.95it/s, grad_norm=0.395, loss_final=0.488, loss_mean=0.847, loss_mean_cls=0.0663, proj_loss=-0.425][2026-03-26 13:01:02] Step: 7227, Training Logs: loss_final: 0.451175, loss_mean: 0.804782, proj_loss: -0.419865, loss_mean_cls: 0.066259, grad_norm: 0.274752 +Steps: 1%| | 7228/1000000 [29:59<69:25:08, 3.97it/s, grad_norm=0.275, loss_final=0.451, loss_mean=0.805, loss_mean_cls=0.0663, proj_loss=-0.42][2026-03-26 13:01:03] Step: 7228, Training Logs: loss_final: 0.460423, loss_mean: 0.819805, proj_loss: -0.424859, loss_mean_cls: 0.065477, grad_norm: 0.483699 +Steps: 1%| | 7229/1000000 [29:59<69:03:53, 3.99it/s, grad_norm=0.484, loss_final=0.46, loss_mean=0.82, loss_mean_cls=0.0655, proj_loss=-0.425][2026-03-26 13:01:03] Step: 7229, Training Logs: loss_final: 0.476504, loss_mean: 0.810939, proj_loss: -0.404392, loss_mean_cls: 0.069956, grad_norm: 0.294501 +Steps: 1%| | 7230/1000000 [30:00<68:50:22, 4.01it/s, grad_norm=0.295, loss_final=0.477, loss_mean=0.811, loss_mean_cls=0.07, proj_loss=-0.404][2026-03-26 13:01:03] Step: 7230, Training Logs: loss_final: 0.467695, loss_mean: 0.827461, proj_loss: -0.425094, loss_mean_cls: 0.065328, grad_norm: 0.341725 +Steps: 1%| | 7231/1000000 [30:00<68:38:10, 4.02it/s, grad_norm=0.342, loss_final=0.468, loss_mean=0.827, loss_mean_cls=0.0653, proj_loss=-0.425][2026-03-26 13:01:03] Step: 7231, Training Logs: loss_final: 0.452012, loss_mean: 0.791737, proj_loss: -0.408798, loss_mean_cls: 0.069074, grad_norm: 0.318229 +Steps: 1%| | 7232/1000000 [30:00<68:29:39, 4.03it/s, grad_norm=0.318, loss_final=0.452, loss_mean=0.792, loss_mean_cls=0.0691, proj_loss=-0.409][2026-03-26 13:01:04] Step: 7232, Training Logs: loss_final: 0.463177, loss_mean: 0.815954, proj_loss: -0.419771, loss_mean_cls: 0.066994, grad_norm: 0.341223 +Steps: 1%| | 7233/1000000 [30:00<68:24:50, 4.03it/s, grad_norm=0.341, loss_final=0.463, loss_mean=0.816, loss_mean_cls=0.067, proj_loss=-0.42][2026-03-26 13:01:04] Step: 7233, Training Logs: loss_final: 0.465613, loss_mean: 0.822020, proj_loss: -0.421982, loss_mean_cls: 0.065575, grad_norm: 0.425037 +Steps: 1%| | 7234/1000000 [30:01<68:22:29, 4.03it/s, grad_norm=0.425, loss_final=0.466, loss_mean=0.822, loss_mean_cls=0.0656, proj_loss=-0.422][2026-03-26 13:01:04] Step: 7234, Training Logs: loss_final: 0.476557, loss_mean: 0.830339, proj_loss: -0.419820, loss_mean_cls: 0.066037, grad_norm: 0.347176 +Steps: 1%| | 7235/1000000 [30:01<68:20:59, 4.03it/s, grad_norm=0.347, loss_final=0.477, loss_mean=0.83, loss_mean_cls=0.066, proj_loss=-0.42][2026-03-26 13:01:04] Step: 7235, Training Logs: loss_final: 0.459136, loss_mean: 0.807939, proj_loss: -0.416160, loss_mean_cls: 0.067358, grad_norm: 0.479251 +Steps: 1%| | 7236/1000000 [30:01<68:17:55, 4.04it/s, grad_norm=0.479, loss_final=0.459, loss_mean=0.808, loss_mean_cls=0.0674, proj_loss=-0.416][2026-03-26 13:01:05] Step: 7236, Training Logs: loss_final: 0.482171, loss_mean: 0.828007, proj_loss: -0.413514, loss_mean_cls: 0.067679, grad_norm: 0.444437 +Steps: 1%| | 7237/1000000 [30:01<68:15:58, 4.04it/s, grad_norm=0.444, loss_final=0.482, loss_mean=0.828, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 13:01:05] Step: 7237, Training Logs: loss_final: 0.472539, loss_mean: 0.816340, proj_loss: -0.413269, loss_mean_cls: 0.069468, grad_norm: 0.491196 +Steps: 1%| | 7238/1000000 [30:02<68:16:44, 4.04it/s, grad_norm=0.491, loss_final=0.473, loss_mean=0.816, loss_mean_cls=0.0695, proj_loss=-0.413][2026-03-26 13:01:05] Step: 7238, Training Logs: loss_final: 0.486187, loss_mean: 0.825324, proj_loss: -0.409417, loss_mean_cls: 0.070280, grad_norm: 0.221196 +Steps: 1%| | 7239/1000000 [30:02<69:25:43, 3.97it/s, grad_norm=0.221, loss_final=0.486, loss_mean=0.825, loss_mean_cls=0.0703, proj_loss=-0.409][2026-03-26 13:01:05] Step: 7239, Training Logs: loss_final: 0.462614, loss_mean: 0.816320, proj_loss: -0.420395, loss_mean_cls: 0.066690, grad_norm: 0.555241 +Steps: 1%| | 7240/1000000 [30:02<69:02:46, 3.99it/s, grad_norm=0.555, loss_final=0.463, loss_mean=0.816, loss_mean_cls=0.0667, proj_loss=-0.42][2026-03-26 13:01:06] Step: 7240, Training Logs: loss_final: 0.448515, loss_mean: 0.798086, proj_loss: -0.417309, loss_mean_cls: 0.067738, grad_norm: 0.386417 +Steps: 1%| | 7241/1000000 [30:02<68:44:05, 4.01it/s, grad_norm=0.386, loss_final=0.449, loss_mean=0.798, loss_mean_cls=0.0677, proj_loss=-0.417][2026-03-26 13:01:06] Step: 7241, Training Logs: loss_final: 0.466244, loss_mean: 0.805756, proj_loss: -0.409307, loss_mean_cls: 0.069796, grad_norm: 0.517026 +Steps: 1%| | 7242/1000000 [30:03<68:34:54, 4.02it/s, grad_norm=0.517, loss_final=0.466, loss_mean=0.806, loss_mean_cls=0.0698, proj_loss=-0.409][2026-03-26 13:01:06] Step: 7242, Training Logs: loss_final: 0.456993, loss_mean: 0.802911, proj_loss: -0.414566, loss_mean_cls: 0.068648, grad_norm: 0.456053 +Steps: 1%| | 7243/1000000 [30:03<68:27:31, 4.03it/s, grad_norm=0.456, loss_final=0.457, loss_mean=0.803, loss_mean_cls=0.0686, proj_loss=-0.415][2026-03-26 13:01:06] Step: 7243, Training Logs: loss_final: 0.472309, loss_mean: 0.825600, proj_loss: -0.419294, loss_mean_cls: 0.066003, grad_norm: 0.483592 +Steps: 1%| | 7244/1000000 [30:03<68:27:09, 4.03it/s, grad_norm=0.484, loss_final=0.472, loss_mean=0.826, loss_mean_cls=0.066, proj_loss=-0.419][2026-03-26 13:01:07] Step: 7244, Training Logs: loss_final: 0.467950, loss_mean: 0.808176, proj_loss: -0.409472, loss_mean_cls: 0.069246, grad_norm: 0.703308 +Steps: 1%| | 7245/1000000 [30:03<68:22:01, 4.03it/s, grad_norm=0.703, loss_final=0.468, loss_mean=0.808, loss_mean_cls=0.0692, proj_loss=-0.409][2026-03-26 13:01:07] Step: 7245, Training Logs: loss_final: 0.461349, loss_mean: 0.797069, proj_loss: -0.405444, loss_mean_cls: 0.069724, grad_norm: 0.374310 +Steps: 1%| | 7246/1000000 [30:04<68:21:43, 4.03it/s, grad_norm=0.374, loss_final=0.461, loss_mean=0.797, loss_mean_cls=0.0697, proj_loss=-0.405][2026-03-26 13:01:07] Step: 7246, Training Logs: loss_final: 0.471620, loss_mean: 0.820531, proj_loss: -0.416163, loss_mean_cls: 0.067252, grad_norm: 0.511064 +Steps: 1%| | 7247/1000000 [30:04<68:18:02, 4.04it/s, grad_norm=0.511, loss_final=0.472, loss_mean=0.821, loss_mean_cls=0.0673, proj_loss=-0.416][2026-03-26 13:01:07] Step: 7247, Training Logs: loss_final: 0.486313, loss_mean: 0.828328, proj_loss: -0.411112, loss_mean_cls: 0.069098, grad_norm: 0.348154 +Steps: 1%| | 7248/1000000 [30:04<68:21:23, 4.03it/s, grad_norm=0.348, loss_final=0.486, loss_mean=0.828, loss_mean_cls=0.0691, proj_loss=-0.411][2026-03-26 13:01:08] Step: 7248, Training Logs: loss_final: 0.478185, loss_mean: 0.830334, proj_loss: -0.419307, loss_mean_cls: 0.067158, grad_norm: 0.623762 +Steps: 1%| | 7249/1000000 [30:04<68:18:10, 4.04it/s, grad_norm=0.624, loss_final=0.478, loss_mean=0.83, loss_mean_cls=0.0672, proj_loss=-0.419][2026-03-26 13:01:08] Step: 7249, Training Logs: loss_final: 0.463336, loss_mean: 0.815685, proj_loss: -0.418932, loss_mean_cls: 0.066583, grad_norm: 0.236583 +Steps: 1%| | 7250/1000000 [30:05<68:19:20, 4.04it/s, grad_norm=0.237, loss_final=0.463, loss_mean=0.816, loss_mean_cls=0.0666, proj_loss=-0.419][2026-03-26 13:01:08] Step: 7250, Training Logs: loss_final: 0.461105, loss_mean: 0.810248, proj_loss: -0.415300, loss_mean_cls: 0.066157, grad_norm: 0.662655 +Steps: 1%| | 7251/1000000 [30:05<68:16:49, 4.04it/s, grad_norm=0.663, loss_final=0.461, loss_mean=0.81, loss_mean_cls=0.0662, proj_loss=-0.415][2026-03-26 13:01:08] Step: 7251, Training Logs: loss_final: 0.468534, loss_mean: 0.830404, proj_loss: -0.426035, loss_mean_cls: 0.064165, grad_norm: 0.378631 +Steps: 1%| | 7252/1000000 [30:05<69:19:42, 3.98it/s, grad_norm=0.379, loss_final=0.469, loss_mean=0.83, loss_mean_cls=0.0642, proj_loss=-0.426][2026-03-26 13:01:09] Step: 7252, Training Logs: loss_final: 0.490074, loss_mean: 0.830371, proj_loss: -0.409624, loss_mean_cls: 0.069327, grad_norm: 0.468448 +Steps: 1%| | 7253/1000000 [30:05<69:02:11, 3.99it/s, grad_norm=0.468, loss_final=0.49, loss_mean=0.83, loss_mean_cls=0.0693, proj_loss=-0.41][2026-03-26 13:01:09] Step: 7253, Training Logs: loss_final: 0.476945, loss_mean: 0.830821, proj_loss: -0.420475, loss_mean_cls: 0.066598, grad_norm: 0.319274 +Steps: 1%| | 7254/1000000 [30:06<68:47:58, 4.01it/s, grad_norm=0.319, loss_final=0.477, loss_mean=0.831, loss_mean_cls=0.0666, proj_loss=-0.42][2026-03-26 13:01:09] Step: 7254, Training Logs: loss_final: 0.470423, loss_mean: 0.817795, proj_loss: -0.414785, loss_mean_cls: 0.067413, grad_norm: 0.332024 +Steps: 1%| | 7255/1000000 [30:06<68:36:16, 4.02it/s, grad_norm=0.332, loss_final=0.47, loss_mean=0.818, loss_mean_cls=0.0674, proj_loss=-0.415][2026-03-26 13:01:09] Step: 7255, Training Logs: loss_final: 0.460302, loss_mean: 0.795933, proj_loss: -0.405881, loss_mean_cls: 0.070250, grad_norm: 0.274653 +Steps: 1%| | 7256/1000000 [30:06<68:31:23, 4.02it/s, grad_norm=0.275, loss_final=0.46, loss_mean=0.796, loss_mean_cls=0.0702, proj_loss=-0.406][2026-03-26 13:01:10] Step: 7256, Training Logs: loss_final: 0.478704, loss_mean: 0.812387, proj_loss: -0.404515, loss_mean_cls: 0.070831, grad_norm: 0.326620 +Steps: 1%| | 7257/1000000 [30:06<68:24:34, 4.03it/s, grad_norm=0.327, loss_final=0.479, loss_mean=0.812, loss_mean_cls=0.0708, proj_loss=-0.405][2026-03-26 13:01:10] Step: 7257, Training Logs: loss_final: 0.463809, loss_mean: 0.817664, proj_loss: -0.420154, loss_mean_cls: 0.066299, grad_norm: 0.411951 +Steps: 1%| | 7258/1000000 [30:07<68:22:39, 4.03it/s, grad_norm=0.412, loss_final=0.464, loss_mean=0.818, loss_mean_cls=0.0663, proj_loss=-0.42][2026-03-26 13:01:10] Step: 7258, Training Logs: loss_final: 0.475112, loss_mean: 0.815889, proj_loss: -0.409699, loss_mean_cls: 0.068922, grad_norm: 0.295738 +Steps: 1%| | 7259/1000000 [30:07<68:21:34, 4.03it/s, grad_norm=0.296, loss_final=0.475, loss_mean=0.816, loss_mean_cls=0.0689, proj_loss=-0.41][2026-03-26 13:01:10] Step: 7259, Training Logs: loss_final: 0.468442, loss_mean: 0.820872, proj_loss: -0.419241, loss_mean_cls: 0.066812, grad_norm: 0.194325 +Steps: 1%| | 7260/1000000 [30:07<68:20:10, 4.04it/s, grad_norm=0.194, loss_final=0.468, loss_mean=0.821, loss_mean_cls=0.0668, proj_loss=-0.419][2026-03-26 13:01:11] Step: 7260, Training Logs: loss_final: 0.457308, loss_mean: 0.804011, proj_loss: -0.415583, loss_mean_cls: 0.068879, grad_norm: 0.259738 +Steps: 1%| | 7261/1000000 [30:07<68:15:21, 4.04it/s, grad_norm=0.26, loss_final=0.457, loss_mean=0.804, loss_mean_cls=0.0689, proj_loss=-0.416][2026-03-26 13:01:11] Step: 7261, Training Logs: loss_final: 0.474565, loss_mean: 0.818232, proj_loss: -0.412249, loss_mean_cls: 0.068581, grad_norm: 0.287423 +Steps: 1%| | 7262/1000000 [30:08<68:17:02, 4.04it/s, grad_norm=0.287, loss_final=0.475, loss_mean=0.818, loss_mean_cls=0.0686, proj_loss=-0.412][2026-03-26 13:01:11] Step: 7262, Training Logs: loss_final: 0.471934, loss_mean: 0.830835, proj_loss: -0.425181, loss_mean_cls: 0.066280, grad_norm: 0.384356 +Steps: 1%| | 7263/1000000 [30:08<68:18:24, 4.04it/s, grad_norm=0.384, loss_final=0.472, loss_mean=0.831, loss_mean_cls=0.0663, proj_loss=-0.425][2026-03-26 13:01:11] Step: 7263, Training Logs: loss_final: 0.456331, loss_mean: 0.806243, proj_loss: -0.416875, loss_mean_cls: 0.066964, grad_norm: 0.361727 +Steps: 1%| | 7264/1000000 [30:08<68:20:03, 4.04it/s, grad_norm=0.362, loss_final=0.456, loss_mean=0.806, loss_mean_cls=0.067, proj_loss=-0.417][2026-03-26 13:01:12] Step: 7264, Training Logs: loss_final: 0.465687, loss_mean: 0.821372, proj_loss: -0.421200, loss_mean_cls: 0.065514, grad_norm: 0.249443 +Steps: 1%| | 7265/1000000 [30:08<68:17:38, 4.04it/s, grad_norm=0.249, loss_final=0.466, loss_mean=0.821, loss_mean_cls=0.0655, proj_loss=-0.421][2026-03-26 13:01:12] Step: 7265, Training Logs: loss_final: 0.483753, loss_mean: 0.826484, proj_loss: -0.411205, loss_mean_cls: 0.068475, grad_norm: 0.370206 +Steps: 1%| | 7266/1000000 [30:09<68:18:39, 4.04it/s, grad_norm=0.37, loss_final=0.484, loss_mean=0.826, loss_mean_cls=0.0685, proj_loss=-0.411][2026-03-26 13:01:12] Step: 7266, Training Logs: loss_final: 0.461715, loss_mean: 0.822079, proj_loss: -0.424557, loss_mean_cls: 0.064193, grad_norm: 0.225831 +Steps: 1%| | 7267/1000000 [30:09<68:20:32, 4.03it/s, grad_norm=0.226, loss_final=0.462, loss_mean=0.822, loss_mean_cls=0.0642, proj_loss=-0.425][2026-03-26 13:01:12] Step: 7267, Training Logs: loss_final: 0.467605, loss_mean: 0.804842, proj_loss: -0.406718, loss_mean_cls: 0.069481, grad_norm: 0.422451 +Steps: 1%| | 7268/1000000 [30:09<68:20:34, 4.03it/s, grad_norm=0.422, loss_final=0.468, loss_mean=0.805, loss_mean_cls=0.0695, proj_loss=-0.407][2026-03-26 13:01:13] Step: 7268, Training Logs: loss_final: 0.481128, loss_mean: 0.835882, proj_loss: -0.421263, loss_mean_cls: 0.066510, grad_norm: 0.363776 +Steps: 1%| | 7269/1000000 [30:09<68:21:24, 4.03it/s, grad_norm=0.364, loss_final=0.481, loss_mean=0.836, loss_mean_cls=0.0665, proj_loss=-0.421][2026-03-26 13:01:13] Step: 7269, Training Logs: loss_final: 0.470252, loss_mean: 0.829600, proj_loss: -0.424399, loss_mean_cls: 0.065050, grad_norm: 0.265941 +Steps: 1%| | 7270/1000000 [30:10<68:23:01, 4.03it/s, grad_norm=0.266, loss_final=0.47, loss_mean=0.83, loss_mean_cls=0.0651, proj_loss=-0.424][2026-03-26 13:01:13] Step: 7270, Training Logs: loss_final: 0.476687, loss_mean: 0.817581, proj_loss: -0.409412, loss_mean_cls: 0.068518, grad_norm: 0.412969 +Steps: 1%| | 7271/1000000 [30:10<68:25:14, 4.03it/s, grad_norm=0.413, loss_final=0.477, loss_mean=0.818, loss_mean_cls=0.0685, proj_loss=-0.409][2026-03-26 13:01:13] Step: 7271, Training Logs: loss_final: 0.487111, loss_mean: 0.836316, proj_loss: -0.416504, loss_mean_cls: 0.067299, grad_norm: 0.302810 +Steps: 1%| | 7272/1000000 [30:10<68:20:16, 4.04it/s, grad_norm=0.303, loss_final=0.487, loss_mean=0.836, loss_mean_cls=0.0673, proj_loss=-0.417][2026-03-26 13:01:13] Step: 7272, Training Logs: loss_final: 0.481341, loss_mean: 0.817037, proj_loss: -0.406198, loss_mean_cls: 0.070502, grad_norm: 0.249262 +Steps: 1%| | 7273/1000000 [30:10<68:21:27, 4.03it/s, grad_norm=0.249, loss_final=0.481, loss_mean=0.817, loss_mean_cls=0.0705, proj_loss=-0.406][2026-03-26 13:01:14] Step: 7273, Training Logs: loss_final: 0.466492, loss_mean: 0.815558, proj_loss: -0.417660, loss_mean_cls: 0.068595, grad_norm: 0.245766 +Steps: 1%| | 7274/1000000 [30:11<68:18:01, 4.04it/s, grad_norm=0.246, loss_final=0.466, loss_mean=0.816, loss_mean_cls=0.0686, proj_loss=-0.418][2026-03-26 13:01:14] Step: 7274, Training Logs: loss_final: 0.483450, loss_mean: 0.834224, proj_loss: -0.417161, loss_mean_cls: 0.066387, grad_norm: 0.323280 +Steps: 1%| | 7275/1000000 [30:11<68:17:09, 4.04it/s, grad_norm=0.323, loss_final=0.483, loss_mean=0.834, loss_mean_cls=0.0664, proj_loss=-0.417][2026-03-26 13:01:14] Step: 7275, Training Logs: loss_final: 0.476003, loss_mean: 0.822970, proj_loss: -0.414517, loss_mean_cls: 0.067550, grad_norm: 0.306159 +Steps: 1%| | 7276/1000000 [30:11<68:16:32, 4.04it/s, grad_norm=0.306, loss_final=0.476, loss_mean=0.823, loss_mean_cls=0.0675, proj_loss=-0.415][2026-03-26 13:01:14] Step: 7276, Training Logs: loss_final: 0.469549, loss_mean: 0.817358, proj_loss: -0.416410, loss_mean_cls: 0.068602, grad_norm: 0.215982 +Steps: 1%| | 7277/1000000 [30:11<68:14:01, 4.04it/s, grad_norm=0.216, loss_final=0.47, loss_mean=0.817, loss_mean_cls=0.0686, proj_loss=-0.416][2026-03-26 13:01:15] Step: 7277, Training Logs: loss_final: 0.463150, loss_mean: 0.823373, proj_loss: -0.425747, loss_mean_cls: 0.065524, grad_norm: 0.268037 +Steps: 1%| | 7278/1000000 [30:12<68:11:19, 4.04it/s, grad_norm=0.268, loss_final=0.463, loss_mean=0.823, loss_mean_cls=0.0655, proj_loss=-0.426][2026-03-26 13:01:15] Step: 7278, Training Logs: loss_final: 0.480022, loss_mean: 0.837689, proj_loss: -0.423608, loss_mean_cls: 0.065941, grad_norm: 0.275391 +Steps: 1%| | 7279/1000000 [30:12<68:09:33, 4.05it/s, grad_norm=0.275, loss_final=0.48, loss_mean=0.838, loss_mean_cls=0.0659, proj_loss=-0.424][2026-03-26 13:01:15] Step: 7279, Training Logs: loss_final: 0.466899, loss_mean: 0.831121, proj_loss: -0.428058, loss_mean_cls: 0.063835, grad_norm: 0.304755 +Steps: 1%| | 7280/1000000 [30:12<68:11:56, 4.04it/s, grad_norm=0.305, loss_final=0.467, loss_mean=0.831, loss_mean_cls=0.0638, proj_loss=-0.428][2026-03-26 13:01:15] Step: 7280, Training Logs: loss_final: 0.465424, loss_mean: 0.825213, proj_loss: -0.424700, loss_mean_cls: 0.064911, grad_norm: 0.289430 +Steps: 1%| | 7281/1000000 [30:12<68:11:23, 4.04it/s, grad_norm=0.289, loss_final=0.465, loss_mean=0.825, loss_mean_cls=0.0649, proj_loss=-0.425][2026-03-26 13:01:16] Step: 7281, Training Logs: loss_final: 0.460072, loss_mean: 0.797934, proj_loss: -0.407318, loss_mean_cls: 0.069457, grad_norm: 0.271927 +Steps: 1%| | 7282/1000000 [30:13<68:12:35, 4.04it/s, grad_norm=0.272, loss_final=0.46, loss_mean=0.798, loss_mean_cls=0.0695, proj_loss=-0.407][2026-03-26 13:01:16] Step: 7282, Training Logs: loss_final: 0.464743, loss_mean: 0.820367, proj_loss: -0.421869, loss_mean_cls: 0.066245, grad_norm: 0.406135 +Steps: 1%| | 7283/1000000 [30:13<68:13:57, 4.04it/s, grad_norm=0.406, loss_final=0.465, loss_mean=0.82, loss_mean_cls=0.0662, proj_loss=-0.422][2026-03-26 13:01:16] Step: 7283, Training Logs: loss_final: 0.465132, loss_mean: 0.808557, proj_loss: -0.412701, loss_mean_cls: 0.069276, grad_norm: 0.218448 +Steps: 1%| | 7284/1000000 [30:13<68:15:22, 4.04it/s, grad_norm=0.218, loss_final=0.465, loss_mean=0.809, loss_mean_cls=0.0693, proj_loss=-0.413][2026-03-26 13:01:16] Step: 7284, Training Logs: loss_final: 0.455475, loss_mean: 0.799222, proj_loss: -0.411881, loss_mean_cls: 0.068133, grad_norm: 0.358379 +Steps: 1%| | 7285/1000000 [30:13<68:13:07, 4.04it/s, grad_norm=0.358, loss_final=0.455, loss_mean=0.799, loss_mean_cls=0.0681, proj_loss=-0.412][2026-03-26 13:01:17] Step: 7285, Training Logs: loss_final: 0.461579, loss_mean: 0.810162, proj_loss: -0.416224, loss_mean_cls: 0.067641, grad_norm: 0.220377 +Steps: 1%| | 7286/1000000 [30:14<68:14:39, 4.04it/s, grad_norm=0.22, loss_final=0.462, loss_mean=0.81, loss_mean_cls=0.0676, proj_loss=-0.416][2026-03-26 13:01:17] Step: 7286, Training Logs: loss_final: 0.455115, loss_mean: 0.790105, proj_loss: -0.404955, loss_mean_cls: 0.069966, grad_norm: 0.300645 +Steps: 1%| | 7287/1000000 [30:14<68:14:30, 4.04it/s, grad_norm=0.301, loss_final=0.455, loss_mean=0.79, loss_mean_cls=0.07, proj_loss=-0.405][2026-03-26 13:01:17] Step: 7287, Training Logs: loss_final: 0.465936, loss_mean: 0.817263, proj_loss: -0.418176, loss_mean_cls: 0.066849, grad_norm: 0.240606 +Steps: 1%| | 7288/1000000 [30:14<68:15:48, 4.04it/s, grad_norm=0.241, loss_final=0.466, loss_mean=0.817, loss_mean_cls=0.0668, proj_loss=-0.418][2026-03-26 13:01:17] Step: 7288, Training Logs: loss_final: 0.459962, loss_mean: 0.813181, proj_loss: -0.420456, loss_mean_cls: 0.067237, grad_norm: 0.246086 +Steps: 1%| | 7289/1000000 [30:14<68:16:00, 4.04it/s, grad_norm=0.246, loss_final=0.46, loss_mean=0.813, loss_mean_cls=0.0672, proj_loss=-0.42][2026-03-26 13:01:18] Step: 7289, Training Logs: loss_final: 0.479541, loss_mean: 0.844916, proj_loss: -0.429486, loss_mean_cls: 0.064110, grad_norm: 0.286571 +Steps: 1%| | 7290/1000000 [30:15<68:15:20, 4.04it/s, grad_norm=0.287, loss_final=0.48, loss_mean=0.845, loss_mean_cls=0.0641, proj_loss=-0.429][2026-03-26 13:01:18] Step: 7290, Training Logs: loss_final: 0.464243, loss_mean: 0.817357, proj_loss: -0.419683, loss_mean_cls: 0.066569, grad_norm: 0.219512 +Steps: 1%| | 7291/1000000 [30:15<68:10:47, 4.04it/s, grad_norm=0.22, loss_final=0.464, loss_mean=0.817, loss_mean_cls=0.0666, proj_loss=-0.42][2026-03-26 13:01:18] Step: 7291, Training Logs: loss_final: 0.469042, loss_mean: 0.815637, proj_loss: -0.415297, loss_mean_cls: 0.068703, grad_norm: 0.275995 +Steps: 1%| | 7292/1000000 [30:15<68:12:32, 4.04it/s, grad_norm=0.276, loss_final=0.469, loss_mean=0.816, loss_mean_cls=0.0687, proj_loss=-0.415][2026-03-26 13:01:18] Step: 7292, Training Logs: loss_final: 0.461684, loss_mean: 0.814054, proj_loss: -0.419076, loss_mean_cls: 0.066706, grad_norm: 0.306445 +Steps: 1%| | 7293/1000000 [30:15<68:11:32, 4.04it/s, grad_norm=0.306, loss_final=0.462, loss_mean=0.814, loss_mean_cls=0.0667, proj_loss=-0.419][2026-03-26 13:01:19] Step: 7293, Training Logs: loss_final: 0.486729, loss_mean: 0.841135, proj_loss: -0.420523, loss_mean_cls: 0.066117, grad_norm: 0.207077 +Steps: 1%| | 7294/1000000 [30:16<68:12:25, 4.04it/s, grad_norm=0.207, loss_final=0.487, loss_mean=0.841, loss_mean_cls=0.0661, proj_loss=-0.421][2026-03-26 13:01:19] Step: 7294, Training Logs: loss_final: 0.471926, loss_mean: 0.827972, proj_loss: -0.422595, loss_mean_cls: 0.066548, grad_norm: 0.279882 +Steps: 1%| | 7295/1000000 [30:16<68:09:37, 4.05it/s, grad_norm=0.28, loss_final=0.472, loss_mean=0.828, loss_mean_cls=0.0665, proj_loss=-0.423][2026-03-26 13:01:19] Step: 7295, Training Logs: loss_final: 0.472528, loss_mean: 0.826664, proj_loss: -0.420790, loss_mean_cls: 0.066654, grad_norm: 0.229320 +Steps: 1%| | 7296/1000000 [30:16<68:13:13, 4.04it/s, grad_norm=0.229, loss_final=0.473, loss_mean=0.827, loss_mean_cls=0.0667, proj_loss=-0.421][2026-03-26 13:01:19] Step: 7296, Training Logs: loss_final: 0.473748, loss_mean: 0.822380, proj_loss: -0.416508, loss_mean_cls: 0.067875, grad_norm: 0.293434 +Steps: 1%| | 7297/1000000 [30:16<68:12:40, 4.04it/s, grad_norm=0.293, loss_final=0.474, loss_mean=0.822, loss_mean_cls=0.0679, proj_loss=-0.417][2026-03-26 13:01:20] Step: 7297, Training Logs: loss_final: 0.475571, loss_mean: 0.825973, proj_loss: -0.417654, loss_mean_cls: 0.067252, grad_norm: 0.348311 +Steps: 1%| | 7298/1000000 [30:17<68:12:48, 4.04it/s, grad_norm=0.348, loss_final=0.476, loss_mean=0.826, loss_mean_cls=0.0673, proj_loss=-0.418][2026-03-26 13:01:20] Step: 7298, Training Logs: loss_final: 0.468476, loss_mean: 0.818246, proj_loss: -0.417641, loss_mean_cls: 0.067870, grad_norm: 0.317680 +Steps: 1%| | 7299/1000000 [30:17<68:14:21, 4.04it/s, grad_norm=0.318, loss_final=0.468, loss_mean=0.818, loss_mean_cls=0.0679, proj_loss=-0.418][2026-03-26 13:01:20] Step: 7299, Training Logs: loss_final: 0.464608, loss_mean: 0.822372, proj_loss: -0.423603, loss_mean_cls: 0.065838, grad_norm: 0.293485 +Steps: 1%| | 7300/1000000 [30:17<68:15:22, 4.04it/s, grad_norm=0.293, loss_final=0.465, loss_mean=0.822, loss_mean_cls=0.0658, proj_loss=-0.424][2026-03-26 13:01:20] Step: 7300, Training Logs: loss_final: 0.475693, loss_mean: 0.814814, proj_loss: -0.408297, loss_mean_cls: 0.069175, grad_norm: 0.342960 +Steps: 1%| | 7301/1000000 [30:17<68:16:32, 4.04it/s, grad_norm=0.343, loss_final=0.476, loss_mean=0.815, loss_mean_cls=0.0692, proj_loss=-0.408][2026-03-26 13:01:21] Step: 7301, Training Logs: loss_final: 0.475410, loss_mean: 0.836842, proj_loss: -0.426779, loss_mean_cls: 0.065346, grad_norm: 0.410461 +Steps: 1%| | 7302/1000000 [30:18<68:14:47, 4.04it/s, grad_norm=0.41, loss_final=0.475, loss_mean=0.837, loss_mean_cls=0.0653, proj_loss=-0.427][2026-03-26 13:01:21] Step: 7302, Training Logs: loss_final: 0.457346, loss_mean: 0.808492, proj_loss: -0.418689, loss_mean_cls: 0.067544, grad_norm: 0.330708 +Steps: 1%| | 7303/1000000 [30:18<68:13:59, 4.04it/s, grad_norm=0.331, loss_final=0.457, loss_mean=0.808, loss_mean_cls=0.0675, proj_loss=-0.419][2026-03-26 13:01:21] Step: 7303, Training Logs: loss_final: 0.465361, loss_mean: 0.822198, proj_loss: -0.422747, loss_mean_cls: 0.065911, grad_norm: 0.210278 +Steps: 1%| | 7304/1000000 [30:18<68:15:31, 4.04it/s, grad_norm=0.21, loss_final=0.465, loss_mean=0.822, loss_mean_cls=0.0659, proj_loss=-0.423][2026-03-26 13:01:21] Step: 7304, Training Logs: loss_final: 0.476772, loss_mean: 0.828897, proj_loss: -0.418694, loss_mean_cls: 0.066569, grad_norm: 0.400480 +Steps: 1%| | 7305/1000000 [30:18<68:14:42, 4.04it/s, grad_norm=0.4, loss_final=0.477, loss_mean=0.829, loss_mean_cls=0.0666, proj_loss=-0.419][2026-03-26 13:01:22] Step: 7305, Training Logs: loss_final: 0.468716, loss_mean: 0.822414, proj_loss: -0.419946, loss_mean_cls: 0.066248, grad_norm: 0.382516 +Steps: 1%| | 7306/1000000 [30:19<68:13:49, 4.04it/s, grad_norm=0.383, loss_final=0.469, loss_mean=0.822, loss_mean_cls=0.0662, proj_loss=-0.42][2026-03-26 13:01:22] Step: 7306, Training Logs: loss_final: 0.471705, loss_mean: 0.819228, proj_loss: -0.415574, loss_mean_cls: 0.068052, grad_norm: 0.363048 +Steps: 1%| | 7307/1000000 [30:19<68:14:37, 4.04it/s, grad_norm=0.363, loss_final=0.472, loss_mean=0.819, loss_mean_cls=0.0681, proj_loss=-0.416][2026-03-26 13:01:22] Step: 7307, Training Logs: loss_final: 0.482903, loss_mean: 0.818935, proj_loss: -0.404587, loss_mean_cls: 0.068555, grad_norm: 0.562454 +Steps: 1%| | 7308/1000000 [30:19<68:15:30, 4.04it/s, grad_norm=0.562, loss_final=0.483, loss_mean=0.819, loss_mean_cls=0.0686, proj_loss=-0.405][2026-03-26 13:01:22] Step: 7308, Training Logs: loss_final: 0.479407, loss_mean: 0.836384, proj_loss: -0.422461, loss_mean_cls: 0.065484, grad_norm: 0.290644 +Steps: 1%| | 7309/1000000 [30:19<68:17:34, 4.04it/s, grad_norm=0.291, loss_final=0.479, loss_mean=0.836, loss_mean_cls=0.0655, proj_loss=-0.422][2026-03-26 13:01:23] Step: 7309, Training Logs: loss_final: 0.491659, loss_mean: 0.849006, proj_loss: -0.423337, loss_mean_cls: 0.065991, grad_norm: 0.545440 +Steps: 1%| | 7310/1000000 [30:20<68:15:32, 4.04it/s, grad_norm=0.545, loss_final=0.492, loss_mean=0.849, loss_mean_cls=0.066, proj_loss=-0.423][2026-03-26 13:01:23] Step: 7310, Training Logs: loss_final: 0.456362, loss_mean: 0.793057, proj_loss: -0.406813, loss_mean_cls: 0.070117, grad_norm: 0.617351 +Steps: 1%| | 7311/1000000 [30:20<68:17:25, 4.04it/s, grad_norm=0.617, loss_final=0.456, loss_mean=0.793, loss_mean_cls=0.0701, proj_loss=-0.407][2026-03-26 13:01:23] Step: 7311, Training Logs: loss_final: 0.469631, loss_mean: 0.832985, proj_loss: -0.427855, loss_mean_cls: 0.064502, grad_norm: 0.413947 +Steps: 1%| | 7312/1000000 [30:20<68:15:43, 4.04it/s, grad_norm=0.414, loss_final=0.47, loss_mean=0.833, loss_mean_cls=0.0645, proj_loss=-0.428][2026-03-26 13:01:23] Step: 7312, Training Logs: loss_final: 0.480415, loss_mean: 0.832226, proj_loss: -0.418418, loss_mean_cls: 0.066607, grad_norm: 0.734758 +Steps: 1%| | 7313/1000000 [30:20<68:14:31, 4.04it/s, grad_norm=0.735, loss_final=0.48, loss_mean=0.832, loss_mean_cls=0.0666, proj_loss=-0.418][2026-03-26 13:01:24] Step: 7313, Training Logs: loss_final: 0.480820, loss_mean: 0.825643, proj_loss: -0.412571, loss_mean_cls: 0.067748, grad_norm: 0.276687 +Steps: 1%| | 7314/1000000 [30:20<68:12:58, 4.04it/s, grad_norm=0.277, loss_final=0.481, loss_mean=0.826, loss_mean_cls=0.0677, proj_loss=-0.413][2026-03-26 13:01:24] Step: 7314, Training Logs: loss_final: 0.439506, loss_mean: 0.791534, proj_loss: -0.418949, loss_mean_cls: 0.066921, grad_norm: 0.533787 +Steps: 1%| | 7315/1000000 [30:21<68:14:04, 4.04it/s, grad_norm=0.534, loss_final=0.44, loss_mean=0.792, loss_mean_cls=0.0669, proj_loss=-0.419][2026-03-26 13:01:24] Step: 7315, Training Logs: loss_final: 0.481252, loss_mean: 0.833266, proj_loss: -0.418778, loss_mean_cls: 0.066764, grad_norm: 0.224391 +Steps: 1%| | 7316/1000000 [30:21<68:13:56, 4.04it/s, grad_norm=0.224, loss_final=0.481, loss_mean=0.833, loss_mean_cls=0.0668, proj_loss=-0.419][2026-03-26 13:01:24] Step: 7316, Training Logs: loss_final: 0.468448, loss_mean: 0.828014, proj_loss: -0.424937, loss_mean_cls: 0.065371, grad_norm: 0.612940 +Steps: 1%| | 7317/1000000 [30:21<68:12:03, 4.04it/s, grad_norm=0.613, loss_final=0.468, loss_mean=0.828, loss_mean_cls=0.0654, proj_loss=-0.425][2026-03-26 13:01:25] Step: 7317, Training Logs: loss_final: 0.473734, loss_mean: 0.814634, proj_loss: -0.409605, loss_mean_cls: 0.068705, grad_norm: 0.578903 +Steps: 1%| | 7318/1000000 [30:21<68:10:45, 4.04it/s, grad_norm=0.579, loss_final=0.474, loss_mean=0.815, loss_mean_cls=0.0687, proj_loss=-0.41][2026-03-26 13:01:25] Step: 7318, Training Logs: loss_final: 0.459431, loss_mean: 0.822544, proj_loss: -0.428765, loss_mean_cls: 0.065652, grad_norm: 0.397039 +Steps: 1%| | 7319/1000000 [30:22<68:14:48, 4.04it/s, grad_norm=0.397, loss_final=0.459, loss_mean=0.823, loss_mean_cls=0.0657, proj_loss=-0.429][2026-03-26 13:01:25] Step: 7319, Training Logs: loss_final: 0.464096, loss_mean: 0.814220, proj_loss: -0.418167, loss_mean_cls: 0.068043, grad_norm: 0.621187 +Steps: 1%| | 7320/1000000 [30:22<68:18:44, 4.04it/s, grad_norm=0.621, loss_final=0.464, loss_mean=0.814, loss_mean_cls=0.068, proj_loss=-0.418][2026-03-26 13:01:25] Step: 7320, Training Logs: loss_final: 0.467873, loss_mean: 0.814715, proj_loss: -0.415291, loss_mean_cls: 0.068449, grad_norm: 0.226055 +Steps: 1%| | 7321/1000000 [30:22<68:18:05, 4.04it/s, grad_norm=0.226, loss_final=0.468, loss_mean=0.815, loss_mean_cls=0.0684, proj_loss=-0.415][2026-03-26 13:01:26] Step: 7321, Training Logs: loss_final: 0.466046, loss_mean: 0.814246, proj_loss: -0.416107, loss_mean_cls: 0.067906, grad_norm: 0.425151 +Steps: 1%| | 7322/1000000 [30:22<68:16:10, 4.04it/s, grad_norm=0.425, loss_final=0.466, loss_mean=0.814, loss_mean_cls=0.0679, proj_loss=-0.416][2026-03-26 13:01:26] Step: 7322, Training Logs: loss_final: 0.475143, loss_mean: 0.823639, proj_loss: -0.416004, loss_mean_cls: 0.067508, grad_norm: 0.239156 +Steps: 1%| | 7323/1000000 [30:23<68:14:54, 4.04it/s, grad_norm=0.239, loss_final=0.475, loss_mean=0.824, loss_mean_cls=0.0675, proj_loss=-0.416][2026-03-26 13:01:26] Step: 7323, Training Logs: loss_final: 0.477803, loss_mean: 0.825561, proj_loss: -0.415338, loss_mean_cls: 0.067580, grad_norm: 0.591119 +Steps: 1%| | 7324/1000000 [30:23<68:15:14, 4.04it/s, grad_norm=0.591, loss_final=0.478, loss_mean=0.826, loss_mean_cls=0.0676, proj_loss=-0.415][2026-03-26 13:01:26] Step: 7324, Training Logs: loss_final: 0.458397, loss_mean: 0.804404, proj_loss: -0.413866, loss_mean_cls: 0.067859, grad_norm: 0.226607 +Steps: 1%| | 7325/1000000 [30:23<68:15:50, 4.04it/s, grad_norm=0.227, loss_final=0.458, loss_mean=0.804, loss_mean_cls=0.0679, proj_loss=-0.414][2026-03-26 13:01:27] Step: 7325, Training Logs: loss_final: 0.459325, loss_mean: 0.815014, proj_loss: -0.421982, loss_mean_cls: 0.066293, grad_norm: 0.673362 +Steps: 1%| | 7326/1000000 [30:23<68:14:26, 4.04it/s, grad_norm=0.673, loss_final=0.459, loss_mean=0.815, loss_mean_cls=0.0663, proj_loss=-0.422][2026-03-26 13:01:27] Step: 7326, Training Logs: loss_final: 0.472807, loss_mean: 0.823249, proj_loss: -0.417383, loss_mean_cls: 0.066941, grad_norm: 0.320180 +Steps: 1%| | 7327/1000000 [30:24<68:12:47, 4.04it/s, grad_norm=0.32, loss_final=0.473, loss_mean=0.823, loss_mean_cls=0.0669, proj_loss=-0.417][2026-03-26 13:01:27] Step: 7327, Training Logs: loss_final: 0.493043, loss_mean: 0.857679, proj_loss: -0.429221, loss_mean_cls: 0.064584, grad_norm: 0.828555 +Steps: 1%| | 7328/1000000 [30:24<68:16:40, 4.04it/s, grad_norm=0.829, loss_final=0.493, loss_mean=0.858, loss_mean_cls=0.0646, proj_loss=-0.429][2026-03-26 13:01:27] Step: 7328, Training Logs: loss_final: 0.472274, loss_mean: 0.835628, proj_loss: -0.427371, loss_mean_cls: 0.064016, grad_norm: 0.549992 +Steps: 1%| | 7329/1000000 [30:24<68:15:19, 4.04it/s, grad_norm=0.55, loss_final=0.472, loss_mean=0.836, loss_mean_cls=0.064, proj_loss=-0.427][2026-03-26 13:01:28] Step: 7329, Training Logs: loss_final: 0.485299, loss_mean: 0.838403, proj_loss: -0.419678, loss_mean_cls: 0.066573, grad_norm: 0.614026 +Steps: 1%| | 7330/1000000 [30:24<68:14:55, 4.04it/s, grad_norm=0.614, loss_final=0.485, loss_mean=0.838, loss_mean_cls=0.0666, proj_loss=-0.42][2026-03-26 13:01:28] Step: 7330, Training Logs: loss_final: 0.465458, loss_mean: 0.799201, proj_loss: -0.404756, loss_mean_cls: 0.071013, grad_norm: 0.673530 +Steps: 1%| | 7331/1000000 [30:25<68:16:00, 4.04it/s, grad_norm=0.674, loss_final=0.465, loss_mean=0.799, loss_mean_cls=0.071, proj_loss=-0.405][2026-03-26 13:01:28] Step: 7331, Training Logs: loss_final: 0.480570, loss_mean: 0.844686, proj_loss: -0.428745, loss_mean_cls: 0.064629, grad_norm: 0.283047 +Steps: 1%| | 7332/1000000 [30:25<68:20:32, 4.03it/s, grad_norm=0.283, loss_final=0.481, loss_mean=0.845, loss_mean_cls=0.0646, proj_loss=-0.429][2026-03-26 13:01:28] Step: 7332, Training Logs: loss_final: 0.456474, loss_mean: 0.804169, proj_loss: -0.415463, loss_mean_cls: 0.067768, grad_norm: 0.483111 +Steps: 1%| | 7333/1000000 [30:25<68:17:04, 4.04it/s, grad_norm=0.483, loss_final=0.456, loss_mean=0.804, loss_mean_cls=0.0678, proj_loss=-0.415][2026-03-26 13:01:29] Step: 7333, Training Logs: loss_final: 0.460469, loss_mean: 0.806827, proj_loss: -0.414572, loss_mean_cls: 0.068214, grad_norm: 0.468481 +Steps: 1%| | 7334/1000000 [30:25<68:18:41, 4.04it/s, grad_norm=0.468, loss_final=0.46, loss_mean=0.807, loss_mean_cls=0.0682, proj_loss=-0.415][2026-03-26 13:01:29] Step: 7334, Training Logs: loss_final: 0.465947, loss_mean: 0.817180, proj_loss: -0.418861, loss_mean_cls: 0.067628, grad_norm: 0.576566 +Steps: 1%| | 7335/1000000 [30:26<68:15:12, 4.04it/s, grad_norm=0.577, loss_final=0.466, loss_mean=0.817, loss_mean_cls=0.0676, proj_loss=-0.419][2026-03-26 13:01:29] Step: 7335, Training Logs: loss_final: 0.478666, loss_mean: 0.832738, proj_loss: -0.420343, loss_mean_cls: 0.066271, grad_norm: 0.287495 +Steps: 1%| | 7336/1000000 [30:26<68:15:46, 4.04it/s, grad_norm=0.287, loss_final=0.479, loss_mean=0.833, loss_mean_cls=0.0663, proj_loss=-0.42][2026-03-26 13:01:29] Step: 7336, Training Logs: loss_final: 0.466480, loss_mean: 0.810616, proj_loss: -0.412401, loss_mean_cls: 0.068265, grad_norm: 0.475800 +Steps: 1%| | 7337/1000000 [30:26<68:15:08, 4.04it/s, grad_norm=0.476, loss_final=0.466, loss_mean=0.811, loss_mean_cls=0.0683, proj_loss=-0.412][2026-03-26 13:01:30] Step: 7337, Training Logs: loss_final: 0.453003, loss_mean: 0.806872, proj_loss: -0.419972, loss_mean_cls: 0.066103, grad_norm: 0.274396 +Steps: 1%| | 7338/1000000 [30:26<68:13:42, 4.04it/s, grad_norm=0.274, loss_final=0.453, loss_mean=0.807, loss_mean_cls=0.0661, proj_loss=-0.42][2026-03-26 13:01:30] Step: 7338, Training Logs: loss_final: 0.469589, loss_mean: 0.814469, proj_loss: -0.413772, loss_mean_cls: 0.068892, grad_norm: 0.453681 +Steps: 1%| | 7339/1000000 [30:27<68:12:02, 4.04it/s, grad_norm=0.454, loss_final=0.47, loss_mean=0.814, loss_mean_cls=0.0689, proj_loss=-0.414][2026-03-26 13:01:30] Step: 7339, Training Logs: loss_final: 0.478614, loss_mean: 0.828833, proj_loss: -0.417869, loss_mean_cls: 0.067651, grad_norm: 0.372111 +Steps: 1%| | 7340/1000000 [30:27<68:11:28, 4.04it/s, grad_norm=0.372, loss_final=0.479, loss_mean=0.829, loss_mean_cls=0.0677, proj_loss=-0.418][2026-03-26 13:01:30] Step: 7340, Training Logs: loss_final: 0.463124, loss_mean: 0.792986, proj_loss: -0.402028, loss_mean_cls: 0.072166, grad_norm: 0.489852 +Steps: 1%| | 7341/1000000 [30:27<68:11:57, 4.04it/s, grad_norm=0.49, loss_final=0.463, loss_mean=0.793, loss_mean_cls=0.0722, proj_loss=-0.402][2026-03-26 13:01:31] Step: 7341, Training Logs: loss_final: 0.461414, loss_mean: 0.811201, proj_loss: -0.417487, loss_mean_cls: 0.067700, grad_norm: 0.210941 +Steps: 1%| | 7342/1000000 [30:27<68:12:17, 4.04it/s, grad_norm=0.211, loss_final=0.461, loss_mean=0.811, loss_mean_cls=0.0677, proj_loss=-0.417][2026-03-26 13:01:31] Step: 7342, Training Logs: loss_final: 0.457740, loss_mean: 0.805351, proj_loss: -0.415722, loss_mean_cls: 0.068111, grad_norm: 0.377797 +Steps: 1%| | 7343/1000000 [30:28<68:11:25, 4.04it/s, grad_norm=0.378, loss_final=0.458, loss_mean=0.805, loss_mean_cls=0.0681, proj_loss=-0.416][2026-03-26 13:01:31] Step: 7343, Training Logs: loss_final: 0.456343, loss_mean: 0.805899, proj_loss: -0.417578, loss_mean_cls: 0.068021, grad_norm: 0.297093 +Steps: 1%| | 7344/1000000 [30:28<68:13:05, 4.04it/s, grad_norm=0.297, loss_final=0.456, loss_mean=0.806, loss_mean_cls=0.068, proj_loss=-0.418][2026-03-26 13:01:31] Step: 7344, Training Logs: loss_final: 0.468093, loss_mean: 0.814326, proj_loss: -0.413680, loss_mean_cls: 0.067447, grad_norm: 0.272424 +Steps: 1%| | 7345/1000000 [30:28<68:12:42, 4.04it/s, grad_norm=0.272, loss_final=0.468, loss_mean=0.814, loss_mean_cls=0.0674, proj_loss=-0.414][2026-03-26 13:01:32] Step: 7345, Training Logs: loss_final: 0.455657, loss_mean: 0.798200, proj_loss: -0.411126, loss_mean_cls: 0.068582, grad_norm: 0.241077 +Steps: 1%| | 7346/1000000 [30:28<68:16:58, 4.04it/s, grad_norm=0.241, loss_final=0.456, loss_mean=0.798, loss_mean_cls=0.0686, proj_loss=-0.411][2026-03-26 13:01:32] Step: 7346, Training Logs: loss_final: 0.477905, loss_mean: 0.833461, proj_loss: -0.422128, loss_mean_cls: 0.066571, grad_norm: 0.305922 +Steps: 1%| | 7347/1000000 [30:29<68:15:17, 4.04it/s, grad_norm=0.306, loss_final=0.478, loss_mean=0.833, loss_mean_cls=0.0666, proj_loss=-0.422][2026-03-26 13:01:32] Step: 7347, Training Logs: loss_final: 0.469666, loss_mean: 0.816341, proj_loss: -0.415021, loss_mean_cls: 0.068346, grad_norm: 0.325684 +Steps: 1%| | 7348/1000000 [30:29<68:20:07, 4.04it/s, grad_norm=0.326, loss_final=0.47, loss_mean=0.816, loss_mean_cls=0.0683, proj_loss=-0.415][2026-03-26 13:01:32] Step: 7348, Training Logs: loss_final: 0.465283, loss_mean: 0.802570, proj_loss: -0.407248, loss_mean_cls: 0.069961, grad_norm: 0.237945 +Steps: 1%| | 7349/1000000 [30:29<68:19:49, 4.04it/s, grad_norm=0.238, loss_final=0.465, loss_mean=0.803, loss_mean_cls=0.07, proj_loss=-0.407][2026-03-26 13:01:33] Step: 7349, Training Logs: loss_final: 0.487199, loss_mean: 0.839720, proj_loss: -0.418722, loss_mean_cls: 0.066201, grad_norm: 0.186483 +Steps: 1%| | 7350/1000000 [30:29<68:16:47, 4.04it/s, grad_norm=0.186, loss_final=0.487, loss_mean=0.84, loss_mean_cls=0.0662, proj_loss=-0.419][2026-03-26 13:01:33] Step: 7350, Training Logs: loss_final: 0.478619, loss_mean: 0.833569, proj_loss: -0.420560, loss_mean_cls: 0.065611, grad_norm: 0.206954 +Steps: 1%| | 7351/1000000 [30:30<68:20:09, 4.04it/s, grad_norm=0.207, loss_final=0.479, loss_mean=0.834, loss_mean_cls=0.0656, proj_loss=-0.421][2026-03-26 13:01:33] Step: 7351, Training Logs: loss_final: 0.452151, loss_mean: 0.792850, proj_loss: -0.410541, loss_mean_cls: 0.069842, grad_norm: 0.318255 +Steps: 1%| | 7352/1000000 [30:30<68:12:52, 4.04it/s, grad_norm=0.318, loss_final=0.452, loss_mean=0.793, loss_mean_cls=0.0698, proj_loss=-0.411][2026-03-26 13:01:33] Step: 7352, Training Logs: loss_final: 0.470591, loss_mean: 0.823759, proj_loss: -0.419643, loss_mean_cls: 0.066475, grad_norm: 0.192208 +Steps: 1%| | 7353/1000000 [30:30<68:14:34, 4.04it/s, grad_norm=0.192, loss_final=0.471, loss_mean=0.824, loss_mean_cls=0.0665, proj_loss=-0.42][2026-03-26 13:01:34] Step: 7353, Training Logs: loss_final: 0.476641, loss_mean: 0.825052, proj_loss: -0.415544, loss_mean_cls: 0.067133, grad_norm: 0.338697 +Steps: 1%| | 7354/1000000 [30:30<68:11:52, 4.04it/s, grad_norm=0.339, loss_final=0.477, loss_mean=0.825, loss_mean_cls=0.0671, proj_loss=-0.416][2026-03-26 13:01:34] Step: 7354, Training Logs: loss_final: 0.485685, loss_mean: 0.839405, proj_loss: -0.419025, loss_mean_cls: 0.065305, grad_norm: 0.340120 +Steps: 1%| | 7355/1000000 [30:31<68:13:43, 4.04it/s, grad_norm=0.34, loss_final=0.486, loss_mean=0.839, loss_mean_cls=0.0653, proj_loss=-0.419][2026-03-26 13:01:34] Step: 7355, Training Logs: loss_final: 0.473285, loss_mean: 0.832669, proj_loss: -0.424926, loss_mean_cls: 0.065542, grad_norm: 0.219073 +Steps: 1%| | 7356/1000000 [30:31<68:11:25, 4.04it/s, grad_norm=0.219, loss_final=0.473, loss_mean=0.833, loss_mean_cls=0.0655, proj_loss=-0.425][2026-03-26 13:01:34] Step: 7356, Training Logs: loss_final: 0.461128, loss_mean: 0.817279, proj_loss: -0.422792, loss_mean_cls: 0.066641, grad_norm: 0.360638 +Steps: 1%| | 7357/1000000 [30:31<68:15:17, 4.04it/s, grad_norm=0.361, loss_final=0.461, loss_mean=0.817, loss_mean_cls=0.0666, proj_loss=-0.423][2026-03-26 13:01:35] Step: 7357, Training Logs: loss_final: 0.452434, loss_mean: 0.802288, proj_loss: -0.418404, loss_mean_cls: 0.068550, grad_norm: 0.253594 +Steps: 1%| | 7358/1000000 [30:31<68:15:25, 4.04it/s, grad_norm=0.254, loss_final=0.452, loss_mean=0.802, loss_mean_cls=0.0685, proj_loss=-0.418][2026-03-26 13:01:35] Step: 7358, Training Logs: loss_final: 0.456027, loss_mean: 0.798130, proj_loss: -0.411511, loss_mean_cls: 0.069408, grad_norm: 0.615304 +Steps: 1%| | 7359/1000000 [30:32<68:50:18, 4.01it/s, grad_norm=0.615, loss_final=0.456, loss_mean=0.798, loss_mean_cls=0.0694, proj_loss=-0.412][2026-03-26 13:01:35] Step: 7359, Training Logs: loss_final: 0.488458, loss_mean: 0.834841, proj_loss: -0.414382, loss_mean_cls: 0.067999, grad_norm: 0.433204 +Steps: 1%| | 7360/1000000 [30:32<68:42:03, 4.01it/s, grad_norm=0.433, loss_final=0.488, loss_mean=0.835, loss_mean_cls=0.068, proj_loss=-0.414][2026-03-26 13:01:35] Step: 7360, Training Logs: loss_final: 0.462270, loss_mean: 0.812860, proj_loss: -0.417256, loss_mean_cls: 0.066665, grad_norm: 0.345702 +Steps: 1%| | 7361/1000000 [30:32<68:34:17, 4.02it/s, grad_norm=0.346, loss_final=0.462, loss_mean=0.813, loss_mean_cls=0.0667, proj_loss=-0.417][2026-03-26 13:01:36] Step: 7361, Training Logs: loss_final: 0.464464, loss_mean: 0.810969, proj_loss: -0.415033, loss_mean_cls: 0.068529, grad_norm: 0.315210 +Steps: 1%| | 7362/1000000 [30:32<68:24:41, 4.03it/s, grad_norm=0.315, loss_final=0.464, loss_mean=0.811, loss_mean_cls=0.0685, proj_loss=-0.415][2026-03-26 13:01:36] Step: 7362, Training Logs: loss_final: 0.470966, loss_mean: 0.824165, proj_loss: -0.418870, loss_mean_cls: 0.065671, grad_norm: 0.356435 +Steps: 1%| | 7363/1000000 [30:33<68:22:40, 4.03it/s, grad_norm=0.356, loss_final=0.471, loss_mean=0.824, loss_mean_cls=0.0657, proj_loss=-0.419][2026-03-26 13:01:36] Step: 7363, Training Logs: loss_final: 0.476814, loss_mean: 0.828155, proj_loss: -0.418160, loss_mean_cls: 0.066819, grad_norm: 0.458614 +Steps: 1%| | 7364/1000000 [30:33<68:20:36, 4.03it/s, grad_norm=0.459, loss_final=0.477, loss_mean=0.828, loss_mean_cls=0.0668, proj_loss=-0.418][2026-03-26 13:01:36] Step: 7364, Training Logs: loss_final: 0.462493, loss_mean: 0.814030, proj_loss: -0.417824, loss_mean_cls: 0.066287, grad_norm: 0.254417 +Steps: 1%| | 7365/1000000 [30:33<68:19:17, 4.04it/s, grad_norm=0.254, loss_final=0.462, loss_mean=0.814, loss_mean_cls=0.0663, proj_loss=-0.418][2026-03-26 13:01:37] Step: 7365, Training Logs: loss_final: 0.464727, loss_mean: 0.824121, proj_loss: -0.423996, loss_mean_cls: 0.064603, grad_norm: 0.396812 +Steps: 1%| | 7366/1000000 [30:33<68:18:04, 4.04it/s, grad_norm=0.397, loss_final=0.465, loss_mean=0.824, loss_mean_cls=0.0646, proj_loss=-0.424][2026-03-26 13:01:37] Step: 7366, Training Logs: loss_final: 0.469212, loss_mean: 0.808951, proj_loss: -0.408512, loss_mean_cls: 0.068773, grad_norm: 0.201358 +Steps: 1%| | 7367/1000000 [30:34<68:15:20, 4.04it/s, grad_norm=0.201, loss_final=0.469, loss_mean=0.809, loss_mean_cls=0.0688, proj_loss=-0.409][2026-03-26 13:01:37] Step: 7367, Training Logs: loss_final: 0.474918, loss_mean: 0.823781, proj_loss: -0.416414, loss_mean_cls: 0.067551, grad_norm: 0.355275 +Steps: 1%| | 7368/1000000 [30:34<68:16:36, 4.04it/s, grad_norm=0.355, loss_final=0.475, loss_mean=0.824, loss_mean_cls=0.0676, proj_loss=-0.416][2026-03-26 13:01:37] Step: 7368, Training Logs: loss_final: 0.485951, loss_mean: 0.837880, proj_loss: -0.417919, loss_mean_cls: 0.065990, grad_norm: 0.233173 +Steps: 1%| | 7369/1000000 [30:34<68:17:23, 4.04it/s, grad_norm=0.233, loss_final=0.486, loss_mean=0.838, loss_mean_cls=0.066, proj_loss=-0.418][2026-03-26 13:01:38] Step: 7369, Training Logs: loss_final: 0.464234, loss_mean: 0.820554, proj_loss: -0.422747, loss_mean_cls: 0.066426, grad_norm: 0.308073 +Steps: 1%| | 7370/1000000 [30:34<71:20:24, 3.87it/s, grad_norm=0.308, loss_final=0.464, loss_mean=0.821, loss_mean_cls=0.0664, proj_loss=-0.423][2026-03-26 13:01:38] Step: 7370, Training Logs: loss_final: 0.488505, loss_mean: 0.835826, proj_loss: -0.414552, loss_mean_cls: 0.067231, grad_norm: 0.318124 +Steps: 1%| | 7371/1000000 [30:35<70:42:30, 3.90it/s, grad_norm=0.318, loss_final=0.489, loss_mean=0.836, loss_mean_cls=0.0672, proj_loss=-0.415][2026-03-26 13:01:38] Step: 7371, Training Logs: loss_final: 0.477727, loss_mean: 0.836714, proj_loss: -0.424259, loss_mean_cls: 0.065272, grad_norm: 0.243889 +Steps: 1%| | 7372/1000000 [30:35<69:57:15, 3.94it/s, grad_norm=0.244, loss_final=0.478, loss_mean=0.837, loss_mean_cls=0.0653, proj_loss=-0.424][2026-03-26 13:01:38] Step: 7372, Training Logs: loss_final: 0.474708, loss_mean: 0.818149, proj_loss: -0.411903, loss_mean_cls: 0.068463, grad_norm: 0.297713 +Steps: 1%| | 7373/1000000 [30:35<69:28:39, 3.97it/s, grad_norm=0.298, loss_final=0.475, loss_mean=0.818, loss_mean_cls=0.0685, proj_loss=-0.412][2026-03-26 13:01:39] Step: 7373, Training Logs: loss_final: 0.466203, loss_mean: 0.813511, proj_loss: -0.415463, loss_mean_cls: 0.068155, grad_norm: 0.250269 +Steps: 1%| | 7374/1000000 [30:35<69:07:36, 3.99it/s, grad_norm=0.25, loss_final=0.466, loss_mean=0.814, loss_mean_cls=0.0682, proj_loss=-0.415][2026-03-26 13:01:39] Step: 7374, Training Logs: loss_final: 0.459154, loss_mean: 0.800326, proj_loss: -0.410164, loss_mean_cls: 0.068992, grad_norm: 0.311947 +Steps: 1%| | 7375/1000000 [30:36<68:56:45, 4.00it/s, grad_norm=0.312, loss_final=0.459, loss_mean=0.8, loss_mean_cls=0.069, proj_loss=-0.41][2026-03-26 13:01:39] Step: 7375, Training Logs: loss_final: 0.476603, loss_mean: 0.828962, proj_loss: -0.418895, loss_mean_cls: 0.066537, grad_norm: 0.301046 +Steps: 1%| | 7376/1000000 [30:36<68:43:36, 4.01it/s, grad_norm=0.301, loss_final=0.477, loss_mean=0.829, loss_mean_cls=0.0665, proj_loss=-0.419][2026-03-26 13:01:39] Step: 7376, Training Logs: loss_final: 0.473645, loss_mean: 0.820153, proj_loss: -0.414218, loss_mean_cls: 0.067709, grad_norm: 0.213847 +Steps: 1%| | 7377/1000000 [30:36<68:35:03, 4.02it/s, grad_norm=0.214, loss_final=0.474, loss_mean=0.82, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 13:01:40] Step: 7377, Training Logs: loss_final: 0.459334, loss_mean: 0.817681, proj_loss: -0.423563, loss_mean_cls: 0.065216, grad_norm: 0.259503 +Steps: 1%| | 7378/1000000 [30:36<68:29:17, 4.03it/s, grad_norm=0.26, loss_final=0.459, loss_mean=0.818, loss_mean_cls=0.0652, proj_loss=-0.424][2026-03-26 13:01:40] Step: 7378, Training Logs: loss_final: 0.490871, loss_mean: 0.843973, proj_loss: -0.418851, loss_mean_cls: 0.065749, grad_norm: 0.287225 +Steps: 1%| | 7379/1000000 [30:37<68:22:51, 4.03it/s, grad_norm=0.287, loss_final=0.491, loss_mean=0.844, loss_mean_cls=0.0657, proj_loss=-0.419][2026-03-26 13:01:40] Step: 7379, Training Logs: loss_final: 0.467059, loss_mean: 0.812152, proj_loss: -0.413904, loss_mean_cls: 0.068811, grad_norm: 0.559397 +Steps: 1%| | 7380/1000000 [30:37<68:18:53, 4.04it/s, grad_norm=0.559, loss_final=0.467, loss_mean=0.812, loss_mean_cls=0.0688, proj_loss=-0.414][2026-03-26 13:01:40] Step: 7380, Training Logs: loss_final: 0.471078, loss_mean: 0.822190, proj_loss: -0.417660, loss_mean_cls: 0.066549, grad_norm: 0.278754 +Steps: 1%| | 7381/1000000 [30:37<68:18:24, 4.04it/s, grad_norm=0.279, loss_final=0.471, loss_mean=0.822, loss_mean_cls=0.0665, proj_loss=-0.418][2026-03-26 13:01:41] Step: 7381, Training Logs: loss_final: 0.471216, loss_mean: 0.818347, proj_loss: -0.414017, loss_mean_cls: 0.066886, grad_norm: 0.370009 +Steps: 1%| | 7382/1000000 [30:37<68:16:53, 4.04it/s, grad_norm=0.37, loss_final=0.471, loss_mean=0.818, loss_mean_cls=0.0669, proj_loss=-0.414][2026-03-26 13:01:41] Step: 7382, Training Logs: loss_final: 0.466257, loss_mean: 0.806033, proj_loss: -0.409044, loss_mean_cls: 0.069267, grad_norm: 0.343752 +Steps: 1%| | 7383/1000000 [30:38<68:14:47, 4.04it/s, grad_norm=0.344, loss_final=0.466, loss_mean=0.806, loss_mean_cls=0.0693, proj_loss=-0.409][2026-03-26 13:01:41] Step: 7383, Training Logs: loss_final: 0.462711, loss_mean: 0.814079, proj_loss: -0.418406, loss_mean_cls: 0.067038, grad_norm: 0.334845 +Steps: 1%| | 7384/1000000 [30:38<68:14:48, 4.04it/s, grad_norm=0.335, loss_final=0.463, loss_mean=0.814, loss_mean_cls=0.067, proj_loss=-0.418][2026-03-26 13:01:41] Step: 7384, Training Logs: loss_final: 0.464885, loss_mean: 0.806121, proj_loss: -0.410422, loss_mean_cls: 0.069185, grad_norm: 0.447104 +Steps: 1%| | 7385/1000000 [30:38<68:13:45, 4.04it/s, grad_norm=0.447, loss_final=0.465, loss_mean=0.806, loss_mean_cls=0.0692, proj_loss=-0.41][2026-03-26 13:01:42] Step: 7385, Training Logs: loss_final: 0.467822, loss_mean: 0.821895, proj_loss: -0.420425, loss_mean_cls: 0.066352, grad_norm: 0.179218 +Steps: 1%| | 7386/1000000 [30:38<68:13:43, 4.04it/s, grad_norm=0.179, loss_final=0.468, loss_mean=0.822, loss_mean_cls=0.0664, proj_loss=-0.42][2026-03-26 13:01:42] Step: 7386, Training Logs: loss_final: 0.463006, loss_mean: 0.815826, proj_loss: -0.419609, loss_mean_cls: 0.066789, grad_norm: 0.268754 +Steps: 1%| | 7387/1000000 [30:39<68:13:46, 4.04it/s, grad_norm=0.269, loss_final=0.463, loss_mean=0.816, loss_mean_cls=0.0668, proj_loss=-0.42][2026-03-26 13:01:42] Step: 7387, Training Logs: loss_final: 0.487028, loss_mean: 0.834057, proj_loss: -0.414710, loss_mean_cls: 0.067680, grad_norm: 0.296864 +Steps: 1%| | 7388/1000000 [30:39<68:18:01, 4.04it/s, grad_norm=0.297, loss_final=0.487, loss_mean=0.834, loss_mean_cls=0.0677, proj_loss=-0.415][2026-03-26 13:01:42] Step: 7388, Training Logs: loss_final: 0.466291, loss_mean: 0.821028, proj_loss: -0.420611, loss_mean_cls: 0.065874, grad_norm: 0.280405 +Steps: 1%| | 7389/1000000 [30:39<68:15:09, 4.04it/s, grad_norm=0.28, loss_final=0.466, loss_mean=0.821, loss_mean_cls=0.0659, proj_loss=-0.421][2026-03-26 13:01:42] Step: 7389, Training Logs: loss_final: 0.476286, loss_mean: 0.822122, proj_loss: -0.413547, loss_mean_cls: 0.067711, grad_norm: 0.259936 +Steps: 1%| | 7390/1000000 [30:39<68:17:47, 4.04it/s, grad_norm=0.26, loss_final=0.476, loss_mean=0.822, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 13:01:43] Step: 7390, Training Logs: loss_final: 0.472930, loss_mean: 0.816367, proj_loss: -0.411383, loss_mean_cls: 0.067947, grad_norm: 0.216991 +Steps: 1%| | 7391/1000000 [30:40<68:13:24, 4.04it/s, grad_norm=0.217, loss_final=0.473, loss_mean=0.816, loss_mean_cls=0.0679, proj_loss=-0.411][2026-03-26 13:01:43] Step: 7391, Training Logs: loss_final: 0.465573, loss_mean: 0.821142, proj_loss: -0.421492, loss_mean_cls: 0.065924, grad_norm: 0.341306 +Steps: 1%| | 7392/1000000 [30:40<68:15:50, 4.04it/s, grad_norm=0.341, loss_final=0.466, loss_mean=0.821, loss_mean_cls=0.0659, proj_loss=-0.421][2026-03-26 13:01:43] Step: 7392, Training Logs: loss_final: 0.468691, loss_mean: 0.823265, proj_loss: -0.420780, loss_mean_cls: 0.066206, grad_norm: 0.206687 +Steps: 1%| | 7393/1000000 [30:40<68:12:32, 4.04it/s, grad_norm=0.207, loss_final=0.469, loss_mean=0.823, loss_mean_cls=0.0662, proj_loss=-0.421][2026-03-26 13:01:43] Step: 7393, Training Logs: loss_final: 0.468372, loss_mean: 0.820251, proj_loss: -0.418936, loss_mean_cls: 0.067057, grad_norm: 0.200053 +Steps: 1%| | 7394/1000000 [30:40<68:11:42, 4.04it/s, grad_norm=0.2, loss_final=0.468, loss_mean=0.82, loss_mean_cls=0.0671, proj_loss=-0.419][2026-03-26 13:01:44] Step: 7394, Training Logs: loss_final: 0.473619, loss_mean: 0.817228, proj_loss: -0.412503, loss_mean_cls: 0.068894, grad_norm: 0.235180 +Steps: 1%| | 7395/1000000 [30:41<68:11:31, 4.04it/s, grad_norm=0.235, loss_final=0.474, loss_mean=0.817, loss_mean_cls=0.0689, proj_loss=-0.413][2026-03-26 13:01:44] Step: 7395, Training Logs: loss_final: 0.470096, loss_mean: 0.818134, proj_loss: -0.415215, loss_mean_cls: 0.067177, grad_norm: 0.307001 +Steps: 1%| | 7396/1000000 [30:41<68:12:58, 4.04it/s, grad_norm=0.307, loss_final=0.47, loss_mean=0.818, loss_mean_cls=0.0672, proj_loss=-0.415][2026-03-26 13:01:44] Step: 7396, Training Logs: loss_final: 0.473735, loss_mean: 0.828113, proj_loss: -0.421477, loss_mean_cls: 0.067100, grad_norm: 0.396029 +Steps: 1%| | 7397/1000000 [30:41<68:13:53, 4.04it/s, grad_norm=0.396, loss_final=0.474, loss_mean=0.828, loss_mean_cls=0.0671, proj_loss=-0.421][2026-03-26 13:01:44] Step: 7397, Training Logs: loss_final: 0.462373, loss_mean: 0.798204, proj_loss: -0.406226, loss_mean_cls: 0.070395, grad_norm: 0.278239 +Steps: 1%| | 7398/1000000 [30:41<68:15:55, 4.04it/s, grad_norm=0.278, loss_final=0.462, loss_mean=0.798, loss_mean_cls=0.0704, proj_loss=-0.406][2026-03-26 13:01:45] Step: 7398, Training Logs: loss_final: 0.491074, loss_mean: 0.830918, proj_loss: -0.409096, loss_mean_cls: 0.069251, grad_norm: 0.459480 +Steps: 1%| | 7399/1000000 [30:42<68:13:15, 4.04it/s, grad_norm=0.459, loss_final=0.491, loss_mean=0.831, loss_mean_cls=0.0693, proj_loss=-0.409][2026-03-26 13:01:45] Step: 7399, Training Logs: loss_final: 0.461648, loss_mean: 0.812791, proj_loss: -0.418140, loss_mean_cls: 0.066997, grad_norm: 0.187048 +Steps: 1%| | 7400/1000000 [30:42<68:17:35, 4.04it/s, grad_norm=0.187, loss_final=0.462, loss_mean=0.813, loss_mean_cls=0.067, proj_loss=-0.418][2026-03-26 13:01:45] Step: 7400, Training Logs: loss_final: 0.463895, loss_mean: 0.808196, proj_loss: -0.412469, loss_mean_cls: 0.068168, grad_norm: 0.528465 +Steps: 1%| | 7401/1000000 [30:42<68:17:10, 4.04it/s, grad_norm=0.528, loss_final=0.464, loss_mean=0.808, loss_mean_cls=0.0682, proj_loss=-0.412][2026-03-26 13:01:45] Step: 7401, Training Logs: loss_final: 0.480044, loss_mean: 0.836506, proj_loss: -0.422711, loss_mean_cls: 0.066249, grad_norm: 0.238774 +Steps: 1%| | 7402/1000000 [30:42<68:19:17, 4.04it/s, grad_norm=0.239, loss_final=0.48, loss_mean=0.837, loss_mean_cls=0.0662, proj_loss=-0.423][2026-03-26 13:01:46] Step: 7402, Training Logs: loss_final: 0.487953, loss_mean: 0.831671, proj_loss: -0.411312, loss_mean_cls: 0.067593, grad_norm: 0.580302 +Steps: 1%| | 7403/1000000 [30:43<68:18:23, 4.04it/s, grad_norm=0.58, loss_final=0.488, loss_mean=0.832, loss_mean_cls=0.0676, proj_loss=-0.411][2026-03-26 13:01:46] Step: 7403, Training Logs: loss_final: 0.462886, loss_mean: 0.821266, proj_loss: -0.423911, loss_mean_cls: 0.065532, grad_norm: 0.527188 +Steps: 1%| | 7404/1000000 [30:43<68:17:20, 4.04it/s, grad_norm=0.527, loss_final=0.463, loss_mean=0.821, loss_mean_cls=0.0655, proj_loss=-0.424][2026-03-26 13:01:46] Step: 7404, Training Logs: loss_final: 0.464187, loss_mean: 0.811403, proj_loss: -0.414890, loss_mean_cls: 0.067674, grad_norm: 0.353335 +Steps: 1%| | 7405/1000000 [30:43<83:57:37, 3.28it/s, grad_norm=0.353, loss_final=0.464, loss_mean=0.811, loss_mean_cls=0.0677, proj_loss=-0.415][2026-03-26 13:01:47] Step: 7405, Training Logs: loss_final: 0.467199, loss_mean: 0.809445, proj_loss: -0.411689, loss_mean_cls: 0.069443, grad_norm: 0.487654 +Steps: 1%| | 7406/1000000 [30:44<79:12:43, 3.48it/s, grad_norm=0.488, loss_final=0.467, loss_mean=0.809, loss_mean_cls=0.0694, proj_loss=-0.412][2026-03-26 13:01:47] Step: 7406, Training Logs: loss_final: 0.461247, loss_mean: 0.810906, proj_loss: -0.417683, loss_mean_cls: 0.068024, grad_norm: 0.191559 +Steps: 1%| | 7407/1000000 [30:44<75:54:50, 3.63it/s, grad_norm=0.192, loss_final=0.461, loss_mean=0.811, loss_mean_cls=0.068, proj_loss=-0.418][2026-03-26 13:01:47] Step: 7407, Training Logs: loss_final: 0.466294, loss_mean: 0.821500, proj_loss: -0.421673, loss_mean_cls: 0.066468, grad_norm: 0.534097 +Steps: 1%| | 7408/1000000 [30:44<74:17:47, 3.71it/s, grad_norm=0.534, loss_final=0.466, loss_mean=0.821, loss_mean_cls=0.0665, proj_loss=-0.422][2026-03-26 13:01:47] Step: 7408, Training Logs: loss_final: 0.459552, loss_mean: 0.806587, proj_loss: -0.414575, loss_mean_cls: 0.067540, grad_norm: 0.333249 +Steps: 1%| | 7409/1000000 [30:44<72:30:37, 3.80it/s, grad_norm=0.333, loss_final=0.46, loss_mean=0.807, loss_mean_cls=0.0675, proj_loss=-0.415][2026-03-26 13:01:48] Step: 7409, Training Logs: loss_final: 0.466361, loss_mean: 0.814728, proj_loss: -0.416246, loss_mean_cls: 0.067879, grad_norm: 0.327982 +Steps: 1%| | 7410/1000000 [30:45<71:19:21, 3.87it/s, grad_norm=0.328, loss_final=0.466, loss_mean=0.815, loss_mean_cls=0.0679, proj_loss=-0.416][2026-03-26 13:01:48] Step: 7410, Training Logs: loss_final: 0.472910, loss_mean: 0.823511, proj_loss: -0.417766, loss_mean_cls: 0.067165, grad_norm: 0.264356 +Steps: 1%| | 7411/1000000 [30:45<70:19:35, 3.92it/s, grad_norm=0.264, loss_final=0.473, loss_mean=0.824, loss_mean_cls=0.0672, proj_loss=-0.418][2026-03-26 13:01:48] Step: 7411, Training Logs: loss_final: 0.454609, loss_mean: 0.804855, proj_loss: -0.418104, loss_mean_cls: 0.067857, grad_norm: 0.214126 +Steps: 1%| | 7412/1000000 [30:45<69:39:29, 3.96it/s, grad_norm=0.214, loss_final=0.455, loss_mean=0.805, loss_mean_cls=0.0679, proj_loss=-0.418][2026-03-26 13:01:48] Step: 7412, Training Logs: loss_final: 0.460393, loss_mean: 0.819032, proj_loss: -0.424023, loss_mean_cls: 0.065384, grad_norm: 0.255751 +Steps: 1%| | 7413/1000000 [30:45<69:14:19, 3.98it/s, grad_norm=0.256, loss_final=0.46, loss_mean=0.819, loss_mean_cls=0.0654, proj_loss=-0.424][2026-03-26 13:01:49] Step: 7413, Training Logs: loss_final: 0.473919, loss_mean: 0.822950, proj_loss: -0.416755, loss_mean_cls: 0.067724, grad_norm: 0.297737 +Steps: 1%| | 7414/1000000 [30:45<68:56:35, 4.00it/s, grad_norm=0.298, loss_final=0.474, loss_mean=0.823, loss_mean_cls=0.0677, proj_loss=-0.417][2026-03-26 13:01:49] Step: 7414, Training Logs: loss_final: 0.453676, loss_mean: 0.812242, proj_loss: -0.423947, loss_mean_cls: 0.065380, grad_norm: 0.427053 +Steps: 1%| | 7415/1000000 [30:46<68:43:41, 4.01it/s, grad_norm=0.427, loss_final=0.454, loss_mean=0.812, loss_mean_cls=0.0654, proj_loss=-0.424][2026-03-26 13:01:49] Step: 7415, Training Logs: loss_final: 0.453694, loss_mean: 0.788479, proj_loss: -0.405380, loss_mean_cls: 0.070595, grad_norm: 0.376047 +Steps: 1%| | 7416/1000000 [30:46<68:36:03, 4.02it/s, grad_norm=0.376, loss_final=0.454, loss_mean=0.788, loss_mean_cls=0.0706, proj_loss=-0.405][2026-03-26 13:01:49] Step: 7416, Training Logs: loss_final: 0.465833, loss_mean: 0.815922, proj_loss: -0.417888, loss_mean_cls: 0.067798, grad_norm: 0.364170 +Steps: 1%| | 7417/1000000 [30:46<68:29:53, 4.03it/s, grad_norm=0.364, loss_final=0.466, loss_mean=0.816, loss_mean_cls=0.0678, proj_loss=-0.418][2026-03-26 13:01:50] Step: 7417, Training Logs: loss_final: 0.474149, loss_mean: 0.825301, proj_loss: -0.417932, loss_mean_cls: 0.066781, grad_norm: 0.450824 +Steps: 1%| | 7418/1000000 [30:47<84:52:11, 3.25it/s, grad_norm=0.451, loss_final=0.474, loss_mean=0.825, loss_mean_cls=0.0668, proj_loss=-0.418][2026-03-26 13:01:50] Step: 7418, Training Logs: loss_final: 0.469055, loss_mean: 0.811200, proj_loss: -0.411176, loss_mean_cls: 0.069031, grad_norm: 0.176409 +Steps: 1%| | 7419/1000000 [30:47<79:45:16, 3.46it/s, grad_norm=0.176, loss_final=0.469, loss_mean=0.811, loss_mean_cls=0.069, proj_loss=-0.411][2026-03-26 13:01:50] Step: 7419, Training Logs: loss_final: 0.458293, loss_mean: 0.816364, proj_loss: -0.423775, loss_mean_cls: 0.065704, grad_norm: 0.331672 +Steps: 1%| | 7420/1000000 [30:47<76:17:02, 3.61it/s, grad_norm=0.332, loss_final=0.458, loss_mean=0.816, loss_mean_cls=0.0657, proj_loss=-0.424][2026-03-26 13:01:51] Step: 7420, Training Logs: loss_final: 0.475469, loss_mean: 0.820019, proj_loss: -0.412295, loss_mean_cls: 0.067745, grad_norm: 0.285755 +Steps: 1%| | 7421/1000000 [30:47<73:52:00, 3.73it/s, grad_norm=0.286, loss_final=0.475, loss_mean=0.82, loss_mean_cls=0.0677, proj_loss=-0.412][2026-03-26 13:01:51] Step: 7421, Training Logs: loss_final: 0.474994, loss_mean: 0.812116, proj_loss: -0.406678, loss_mean_cls: 0.069556, grad_norm: 0.251375 +Steps: 1%| | 7422/1000000 [30:48<72:11:42, 3.82it/s, grad_norm=0.251, loss_final=0.475, loss_mean=0.812, loss_mean_cls=0.0696, proj_loss=-0.407][2026-03-26 13:01:51] Step: 7422, Training Logs: loss_final: 0.479009, loss_mean: 0.817568, proj_loss: -0.408369, loss_mean_cls: 0.069810, grad_norm: 0.233698 +Steps: 1%| | 7423/1000000 [30:48<71:12:47, 3.87it/s, grad_norm=0.234, loss_final=0.479, loss_mean=0.818, loss_mean_cls=0.0698, proj_loss=-0.408][2026-03-26 13:01:51] Step: 7423, Training Logs: loss_final: 0.475479, loss_mean: 0.831640, proj_loss: -0.423037, loss_mean_cls: 0.066876, grad_norm: 0.298602 +Steps: 1%| | 7424/1000000 [30:48<70:43:12, 3.90it/s, grad_norm=0.299, loss_final=0.475, loss_mean=0.832, loss_mean_cls=0.0669, proj_loss=-0.423][2026-03-26 13:01:52] Step: 7424, Training Logs: loss_final: 0.462799, loss_mean: 0.809568, proj_loss: -0.414276, loss_mean_cls: 0.067508, grad_norm: 0.362826 +Steps: 1%| | 7425/1000000 [30:48<69:58:34, 3.94it/s, grad_norm=0.363, loss_final=0.463, loss_mean=0.81, loss_mean_cls=0.0675, proj_loss=-0.414][2026-03-26 13:01:52] Step: 7425, Training Logs: loss_final: 0.458067, loss_mean: 0.819358, proj_loss: -0.426352, loss_mean_cls: 0.065062, grad_norm: 0.358925 +Steps: 1%| | 7426/1000000 [30:49<69:30:01, 3.97it/s, grad_norm=0.359, loss_final=0.458, loss_mean=0.819, loss_mean_cls=0.0651, proj_loss=-0.426][2026-03-26 13:01:52] Step: 7426, Training Logs: loss_final: 0.479551, loss_mean: 0.815589, proj_loss: -0.405955, loss_mean_cls: 0.069917, grad_norm: 0.355589 +Steps: 1%| | 7427/1000000 [30:49<69:08:11, 3.99it/s, grad_norm=0.356, loss_final=0.48, loss_mean=0.816, loss_mean_cls=0.0699, proj_loss=-0.406][2026-03-26 13:01:52] Step: 7427, Training Logs: loss_final: 0.470192, loss_mean: 0.812957, proj_loss: -0.412239, loss_mean_cls: 0.069473, grad_norm: 0.262448 +Steps: 1%| | 7428/1000000 [30:49<68:54:06, 4.00it/s, grad_norm=0.262, loss_final=0.47, loss_mean=0.813, loss_mean_cls=0.0695, proj_loss=-0.412][2026-03-26 13:01:53] Step: 7428, Training Logs: loss_final: 0.480537, loss_mean: 0.830611, proj_loss: -0.417285, loss_mean_cls: 0.067211, grad_norm: 0.283316 +Steps: 1%| | 7429/1000000 [30:49<68:40:07, 4.02it/s, grad_norm=0.283, loss_final=0.481, loss_mean=0.831, loss_mean_cls=0.0672, proj_loss=-0.417][2026-03-26 13:01:53] Step: 7429, Training Logs: loss_final: 0.474044, loss_mean: 0.804246, proj_loss: -0.401318, loss_mean_cls: 0.071116, grad_norm: 0.413166 +Steps: 1%| | 7430/1000000 [30:50<84:15:56, 3.27it/s, grad_norm=0.413, loss_final=0.474, loss_mean=0.804, loss_mean_cls=0.0711, proj_loss=-0.401][2026-03-26 13:01:53] Step: 7430, Training Logs: loss_final: 0.483414, loss_mean: 0.847064, proj_loss: -0.427569, loss_mean_cls: 0.063918, grad_norm: 0.275217 +Steps: 1%| | 7431/1000000 [30:50<79:25:28, 3.47it/s, grad_norm=0.275, loss_final=0.483, loss_mean=0.847, loss_mean_cls=0.0639, proj_loss=-0.428][2026-03-26 13:01:53] Step: 7431, Training Logs: loss_final: 0.462766, loss_mean: 0.814295, proj_loss: -0.417246, loss_mean_cls: 0.065717, grad_norm: 0.227364 +Steps: 1%| | 7432/1000000 [30:50<76:05:52, 3.62it/s, grad_norm=0.227, loss_final=0.463, loss_mean=0.814, loss_mean_cls=0.0657, proj_loss=-0.417][2026-03-26 13:01:54] Step: 7432, Training Logs: loss_final: 0.479922, loss_mean: 0.821182, proj_loss: -0.410322, loss_mean_cls: 0.069063, grad_norm: 0.233300 +Steps: 1%| | 7433/1000000 [30:51<73:41:45, 3.74it/s, grad_norm=0.233, loss_final=0.48, loss_mean=0.821, loss_mean_cls=0.0691, proj_loss=-0.41][2026-03-26 13:01:54] Step: 7433, Training Logs: loss_final: 0.483717, loss_mean: 0.835955, proj_loss: -0.419487, loss_mean_cls: 0.067249, grad_norm: 0.384544 +Steps: 1%| | 7434/1000000 [30:51<72:04:59, 3.82it/s, grad_norm=0.385, loss_final=0.484, loss_mean=0.836, loss_mean_cls=0.0672, proj_loss=-0.419][2026-03-26 13:01:54] Step: 7434, Training Logs: loss_final: 0.487528, loss_mean: 0.835801, proj_loss: -0.415216, loss_mean_cls: 0.066943, grad_norm: 0.286654 +Steps: 1%| | 7435/1000000 [30:51<70:53:46, 3.89it/s, grad_norm=0.287, loss_final=0.488, loss_mean=0.836, loss_mean_cls=0.0669, proj_loss=-0.415][2026-03-26 13:01:54] Step: 7435, Training Logs: loss_final: 0.472255, loss_mean: 0.820581, proj_loss: -0.415809, loss_mean_cls: 0.067483, grad_norm: 0.306070 +Steps: 1%| | 7436/1000000 [30:51<70:05:01, 3.93it/s, grad_norm=0.306, loss_final=0.472, loss_mean=0.821, loss_mean_cls=0.0675, proj_loss=-0.416][2026-03-26 13:01:55] Step: 7436, Training Logs: loss_final: 0.473813, loss_mean: 0.836249, proj_loss: -0.426309, loss_mean_cls: 0.063872, grad_norm: 0.523398 +Steps: 1%| | 7437/1000000 [30:52<69:28:31, 3.97it/s, grad_norm=0.523, loss_final=0.474, loss_mean=0.836, loss_mean_cls=0.0639, proj_loss=-0.426][2026-03-26 13:01:55] Step: 7437, Training Logs: loss_final: 0.460019, loss_mean: 0.803608, proj_loss: -0.412922, loss_mean_cls: 0.069332, grad_norm: 0.291057 +Steps: 1%| | 7438/1000000 [30:52<69:09:53, 3.99it/s, grad_norm=0.291, loss_final=0.46, loss_mean=0.804, loss_mean_cls=0.0693, proj_loss=-0.413][2026-03-26 13:01:55] Step: 7438, Training Logs: loss_final: 0.470768, loss_mean: 0.825390, proj_loss: -0.421586, loss_mean_cls: 0.066964, grad_norm: 0.307157 +Steps: 1%| | 7439/1000000 [30:52<68:51:39, 4.00it/s, grad_norm=0.307, loss_final=0.471, loss_mean=0.825, loss_mean_cls=0.067, proj_loss=-0.422][2026-03-26 13:01:55] Step: 7439, Training Logs: loss_final: 0.468782, loss_mean: 0.801559, proj_loss: -0.403368, loss_mean_cls: 0.070591, grad_norm: 0.275327 +Steps: 1%| | 7440/1000000 [30:52<68:39:04, 4.02it/s, grad_norm=0.275, loss_final=0.469, loss_mean=0.802, loss_mean_cls=0.0706, proj_loss=-0.403][2026-03-26 13:01:56] Step: 7440, Training Logs: loss_final: 0.474082, loss_mean: 0.822261, proj_loss: -0.416376, loss_mean_cls: 0.068197, grad_norm: 0.400717 +Steps: 1%| | 7441/1000000 [30:53<68:31:16, 4.02it/s, grad_norm=0.401, loss_final=0.474, loss_mean=0.822, loss_mean_cls=0.0682, proj_loss=-0.416][2026-03-26 13:01:56] Step: 7441, Training Logs: loss_final: 0.468848, loss_mean: 0.809532, proj_loss: -0.410596, loss_mean_cls: 0.069912, grad_norm: 0.402688 +Steps: 1%| | 7442/1000000 [30:53<68:28:08, 4.03it/s, grad_norm=0.403, loss_final=0.469, loss_mean=0.81, loss_mean_cls=0.0699, proj_loss=-0.411][2026-03-26 13:01:56] Step: 7442, Training Logs: loss_final: 0.473989, loss_mean: 0.828253, proj_loss: -0.420207, loss_mean_cls: 0.065943, grad_norm: 0.229514 +Steps: 1%| | 7443/1000000 [30:53<68:23:46, 4.03it/s, grad_norm=0.23, loss_final=0.474, loss_mean=0.828, loss_mean_cls=0.0659, proj_loss=-0.42][2026-03-26 13:01:56] Step: 7443, Training Logs: loss_final: 0.469976, loss_mean: 0.822139, proj_loss: -0.418969, loss_mean_cls: 0.066806, grad_norm: 0.350313 +Steps: 1%| | 7444/1000000 [30:53<68:18:46, 4.04it/s, grad_norm=0.35, loss_final=0.47, loss_mean=0.822, loss_mean_cls=0.0668, proj_loss=-0.419][2026-03-26 13:01:57] Step: 7444, Training Logs: loss_final: 0.477304, loss_mean: 0.824703, proj_loss: -0.415368, loss_mean_cls: 0.067968, grad_norm: 0.272110 +Steps: 1%| | 7445/1000000 [30:54<68:19:23, 4.04it/s, grad_norm=0.272, loss_final=0.477, loss_mean=0.825, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 13:01:57] Step: 7445, Training Logs: loss_final: 0.458208, loss_mean: 0.809115, proj_loss: -0.418503, loss_mean_cls: 0.067596, grad_norm: 0.477446 +Steps: 1%| | 7446/1000000 [30:54<69:08:35, 3.99it/s, grad_norm=0.477, loss_final=0.458, loss_mean=0.809, loss_mean_cls=0.0676, proj_loss=-0.419][2026-03-26 13:01:57] Step: 7446, Training Logs: loss_final: 0.462184, loss_mean: 0.811405, proj_loss: -0.416536, loss_mean_cls: 0.067315, grad_norm: 0.210383 +Steps: 1%| | 7447/1000000 [30:54<68:54:49, 4.00it/s, grad_norm=0.21, loss_final=0.462, loss_mean=0.811, loss_mean_cls=0.0673, proj_loss=-0.417][2026-03-26 13:01:57] Step: 7447, Training Logs: loss_final: 0.462918, loss_mean: 0.812041, proj_loss: -0.416426, loss_mean_cls: 0.067303, grad_norm: 0.481343 +Steps: 1%| | 7448/1000000 [30:54<68:42:09, 4.01it/s, grad_norm=0.481, loss_final=0.463, loss_mean=0.812, loss_mean_cls=0.0673, proj_loss=-0.416][2026-03-26 13:01:58] Step: 7448, Training Logs: loss_final: 0.479167, loss_mean: 0.833380, proj_loss: -0.420658, loss_mean_cls: 0.066445, grad_norm: 0.409129 +Steps: 1%| | 7449/1000000 [30:55<68:32:36, 4.02it/s, grad_norm=0.409, loss_final=0.479, loss_mean=0.833, loss_mean_cls=0.0664, proj_loss=-0.421][2026-03-26 13:01:58] Step: 7449, Training Logs: loss_final: 0.461690, loss_mean: 0.798779, proj_loss: -0.406776, loss_mean_cls: 0.069687, grad_norm: 0.255245 +Steps: 1%| | 7450/1000000 [30:55<68:28:45, 4.03it/s, grad_norm=0.255, loss_final=0.462, loss_mean=0.799, loss_mean_cls=0.0697, proj_loss=-0.407][2026-03-26 13:01:58] Step: 7450, Training Logs: loss_final: 0.455159, loss_mean: 0.799641, proj_loss: -0.413283, loss_mean_cls: 0.068801, grad_norm: 0.239542 +Steps: 1%| | 7451/1000000 [30:55<68:24:14, 4.03it/s, grad_norm=0.24, loss_final=0.455, loss_mean=0.8, loss_mean_cls=0.0688, proj_loss=-0.413][2026-03-26 13:01:58] Step: 7451, Training Logs: loss_final: 0.488465, loss_mean: 0.843048, proj_loss: -0.421052, loss_mean_cls: 0.066469, grad_norm: 0.304874 +Steps: 1%| | 7452/1000000 [30:55<68:19:17, 4.04it/s, grad_norm=0.305, loss_final=0.488, loss_mean=0.843, loss_mean_cls=0.0665, proj_loss=-0.421][2026-03-26 13:01:59] Step: 7452, Training Logs: loss_final: 0.464695, loss_mean: 0.809308, proj_loss: -0.413311, loss_mean_cls: 0.068698, grad_norm: 0.523925 +Steps: 1%| | 7453/1000000 [30:56<68:16:13, 4.04it/s, grad_norm=0.524, loss_final=0.465, loss_mean=0.809, loss_mean_cls=0.0687, proj_loss=-0.413][2026-03-26 13:01:59] Step: 7453, Training Logs: loss_final: 0.457926, loss_mean: 0.803489, proj_loss: -0.413825, loss_mean_cls: 0.068261, grad_norm: 0.268693 +Steps: 1%| | 7454/1000000 [30:56<68:15:09, 4.04it/s, grad_norm=0.269, loss_final=0.458, loss_mean=0.803, loss_mean_cls=0.0683, proj_loss=-0.414][2026-03-26 13:01:59] Step: 7454, Training Logs: loss_final: 0.456384, loss_mean: 0.802959, proj_loss: -0.414811, loss_mean_cls: 0.068236, grad_norm: 0.280162 +Steps: 1%| | 7455/1000000 [30:56<68:14:07, 4.04it/s, grad_norm=0.28, loss_final=0.456, loss_mean=0.803, loss_mean_cls=0.0682, proj_loss=-0.415][2026-03-26 13:01:59] Step: 7455, Training Logs: loss_final: 0.473435, loss_mean: 0.827643, proj_loss: -0.420083, loss_mean_cls: 0.065875, grad_norm: 0.250643 +Steps: 1%| | 7456/1000000 [30:56<68:12:07, 4.04it/s, grad_norm=0.251, loss_final=0.473, loss_mean=0.828, loss_mean_cls=0.0659, proj_loss=-0.42][2026-03-26 13:02:00] Step: 7456, Training Logs: loss_final: 0.472217, loss_mean: 0.822135, proj_loss: -0.417361, loss_mean_cls: 0.067443, grad_norm: 0.330768 +Steps: 1%| | 7457/1000000 [30:57<68:11:51, 4.04it/s, grad_norm=0.331, loss_final=0.472, loss_mean=0.822, loss_mean_cls=0.0674, proj_loss=-0.417][2026-03-26 13:02:00] Step: 7457, Training Logs: loss_final: 0.460846, loss_mean: 0.810929, proj_loss: -0.417812, loss_mean_cls: 0.067730, grad_norm: 0.228856 +Steps: 1%| | 7458/1000000 [30:57<68:12:17, 4.04it/s, grad_norm=0.229, loss_final=0.461, loss_mean=0.811, loss_mean_cls=0.0677, proj_loss=-0.418][2026-03-26 13:02:00] Step: 7458, Training Logs: loss_final: 0.468375, loss_mean: 0.811663, proj_loss: -0.412002, loss_mean_cls: 0.068715, grad_norm: 0.393740 +Steps: 1%| | 7459/1000000 [30:57<68:11:37, 4.04it/s, grad_norm=0.394, loss_final=0.468, loss_mean=0.812, loss_mean_cls=0.0687, proj_loss=-0.412][2026-03-26 13:02:00] Step: 7459, Training Logs: loss_final: 0.475718, loss_mean: 0.824536, proj_loss: -0.416402, loss_mean_cls: 0.067583, grad_norm: 0.263209 +Steps: 1%| | 7460/1000000 [30:57<68:12:24, 4.04it/s, grad_norm=0.263, loss_final=0.476, loss_mean=0.825, loss_mean_cls=0.0676, proj_loss=-0.416][2026-03-26 13:02:01] Step: 7460, Training Logs: loss_final: 0.457117, loss_mean: 0.806333, proj_loss: -0.416667, loss_mean_cls: 0.067451, grad_norm: 0.291454 +Steps: 1%| | 7461/1000000 [30:58<68:10:18, 4.04it/s, grad_norm=0.291, loss_final=0.457, loss_mean=0.806, loss_mean_cls=0.0675, proj_loss=-0.417][2026-03-26 13:02:01] Step: 7461, Training Logs: loss_final: 0.461295, loss_mean: 0.813211, proj_loss: -0.418545, loss_mean_cls: 0.066629, grad_norm: 0.390636 +Steps: 1%| | 7462/1000000 [30:58<68:13:27, 4.04it/s, grad_norm=0.391, loss_final=0.461, loss_mean=0.813, loss_mean_cls=0.0666, proj_loss=-0.419][2026-03-26 13:02:01] Step: 7462, Training Logs: loss_final: 0.459862, loss_mean: 0.825979, proj_loss: -0.430513, loss_mean_cls: 0.064396, grad_norm: 0.238438 +Steps: 1%| | 7463/1000000 [30:58<68:13:39, 4.04it/s, grad_norm=0.238, loss_final=0.46, loss_mean=0.826, loss_mean_cls=0.0644, proj_loss=-0.431][2026-03-26 13:02:01] Step: 7463, Training Logs: loss_final: 0.457334, loss_mean: 0.795618, proj_loss: -0.407922, loss_mean_cls: 0.069638, grad_norm: 0.508430 +Steps: 1%| | 7464/1000000 [30:58<68:14:37, 4.04it/s, grad_norm=0.508, loss_final=0.457, loss_mean=0.796, loss_mean_cls=0.0696, proj_loss=-0.408][2026-03-26 13:02:02] Step: 7464, Training Logs: loss_final: 0.475752, loss_mean: 0.821011, proj_loss: -0.413861, loss_mean_cls: 0.068602, grad_norm: 0.327594 +Steps: 1%| | 7465/1000000 [30:59<68:17:16, 4.04it/s, grad_norm=0.328, loss_final=0.476, loss_mean=0.821, loss_mean_cls=0.0686, proj_loss=-0.414][2026-03-26 13:02:02] Step: 7465, Training Logs: loss_final: 0.463092, loss_mean: 0.819900, proj_loss: -0.422447, loss_mean_cls: 0.065639, grad_norm: 0.459490 +Steps: 1%| | 7466/1000000 [30:59<68:20:47, 4.03it/s, grad_norm=0.459, loss_final=0.463, loss_mean=0.82, loss_mean_cls=0.0656, proj_loss=-0.422][2026-03-26 13:02:02] Step: 7466, Training Logs: loss_final: 0.468849, loss_mean: 0.826704, proj_loss: -0.423316, loss_mean_cls: 0.065461, grad_norm: 0.596298 +Steps: 1%| | 7467/1000000 [30:59<68:20:37, 4.03it/s, grad_norm=0.596, loss_final=0.469, loss_mean=0.827, loss_mean_cls=0.0655, proj_loss=-0.423][2026-03-26 13:02:02] Step: 7467, Training Logs: loss_final: 0.480459, loss_mean: 0.828318, proj_loss: -0.414982, loss_mean_cls: 0.067124, grad_norm: 0.429906 +Steps: 1%| | 7468/1000000 [30:59<68:17:47, 4.04it/s, grad_norm=0.43, loss_final=0.48, loss_mean=0.828, loss_mean_cls=0.0671, proj_loss=-0.415][2026-03-26 13:02:03] Step: 7468, Training Logs: loss_final: 0.454818, loss_mean: 0.807045, proj_loss: -0.419253, loss_mean_cls: 0.067026, grad_norm: 0.373754 +Steps: 1%| | 7469/1000000 [31:00<68:22:22, 4.03it/s, grad_norm=0.374, loss_final=0.455, loss_mean=0.807, loss_mean_cls=0.067, proj_loss=-0.419][2026-03-26 13:02:03] Step: 7469, Training Logs: loss_final: 0.460110, loss_mean: 0.814296, proj_loss: -0.420810, loss_mean_cls: 0.066624, grad_norm: 0.526199 +Steps: 1%| | 7470/1000000 [31:00<68:23:57, 4.03it/s, grad_norm=0.526, loss_final=0.46, loss_mean=0.814, loss_mean_cls=0.0666, proj_loss=-0.421][2026-03-26 13:02:03] Step: 7470, Training Logs: loss_final: 0.451488, loss_mean: 0.788424, proj_loss: -0.406483, loss_mean_cls: 0.069546, grad_norm: 0.274031 +Steps: 1%| | 7471/1000000 [31:00<68:21:50, 4.03it/s, grad_norm=0.274, loss_final=0.451, loss_mean=0.788, loss_mean_cls=0.0695, proj_loss=-0.406][2026-03-26 13:02:03] Step: 7471, Training Logs: loss_final: 0.481517, loss_mean: 0.820980, proj_loss: -0.408705, loss_mean_cls: 0.069241, grad_norm: 0.353570 +Steps: 1%| | 7472/1000000 [31:00<68:19:33, 4.04it/s, grad_norm=0.354, loss_final=0.482, loss_mean=0.821, loss_mean_cls=0.0692, proj_loss=-0.409][2026-03-26 13:02:04] Step: 7472, Training Logs: loss_final: 0.449752, loss_mean: 0.798975, proj_loss: -0.417393, loss_mean_cls: 0.068169, grad_norm: 0.226282 +Steps: 1%| | 7473/1000000 [31:01<68:16:16, 4.04it/s, grad_norm=0.226, loss_final=0.45, loss_mean=0.799, loss_mean_cls=0.0682, proj_loss=-0.417][2026-03-26 13:02:04] Step: 7473, Training Logs: loss_final: 0.456323, loss_mean: 0.813400, proj_loss: -0.422756, loss_mean_cls: 0.065679, grad_norm: 0.246641 +Steps: 1%| | 7474/1000000 [31:01<68:18:30, 4.04it/s, grad_norm=0.247, loss_final=0.456, loss_mean=0.813, loss_mean_cls=0.0657, proj_loss=-0.423][2026-03-26 13:02:04] Step: 7474, Training Logs: loss_final: 0.481617, loss_mean: 0.827218, proj_loss: -0.414333, loss_mean_cls: 0.068732, grad_norm: 0.428267 +Steps: 1%| | 7475/1000000 [31:01<68:14:21, 4.04it/s, grad_norm=0.428, loss_final=0.482, loss_mean=0.827, loss_mean_cls=0.0687, proj_loss=-0.414][2026-03-26 13:02:04] Step: 7475, Training Logs: loss_final: 0.466480, loss_mean: 0.811900, proj_loss: -0.413711, loss_mean_cls: 0.068291, grad_norm: 0.449408 +Steps: 1%| | 7476/1000000 [31:01<68:16:29, 4.04it/s, grad_norm=0.449, loss_final=0.466, loss_mean=0.812, loss_mean_cls=0.0683, proj_loss=-0.414][2026-03-26 13:02:05] Step: 7476, Training Logs: loss_final: 0.470801, loss_mean: 0.829430, proj_loss: -0.424386, loss_mean_cls: 0.065757, grad_norm: 0.320583 +Steps: 1%| | 7477/1000000 [31:01<68:14:51, 4.04it/s, grad_norm=0.321, loss_final=0.471, loss_mean=0.829, loss_mean_cls=0.0658, proj_loss=-0.424][2026-03-26 13:02:05] Step: 7477, Training Logs: loss_final: 0.449826, loss_mean: 0.805704, proj_loss: -0.421727, loss_mean_cls: 0.065849, grad_norm: 0.254513 +Steps: 1%| | 7478/1000000 [31:02<68:15:04, 4.04it/s, grad_norm=0.255, loss_final=0.45, loss_mean=0.806, loss_mean_cls=0.0658, proj_loss=-0.422][2026-03-26 13:02:05] Step: 7478, Training Logs: loss_final: 0.470065, loss_mean: 0.830402, proj_loss: -0.424637, loss_mean_cls: 0.064301, grad_norm: 0.204972 +Steps: 1%| | 7479/1000000 [31:02<68:13:55, 4.04it/s, grad_norm=0.205, loss_final=0.47, loss_mean=0.83, loss_mean_cls=0.0643, proj_loss=-0.425][2026-03-26 13:02:05] Step: 7479, Training Logs: loss_final: 0.468832, loss_mean: 0.818111, proj_loss: -0.416618, loss_mean_cls: 0.067340, grad_norm: 0.460754 +Steps: 1%| | 7480/1000000 [31:02<68:13:10, 4.04it/s, grad_norm=0.461, loss_final=0.469, loss_mean=0.818, loss_mean_cls=0.0673, proj_loss=-0.417][2026-03-26 13:02:06] Step: 7480, Training Logs: loss_final: 0.458636, loss_mean: 0.809136, proj_loss: -0.418505, loss_mean_cls: 0.068006, grad_norm: 0.371996 +Steps: 1%| | 7481/1000000 [31:02<68:14:06, 4.04it/s, grad_norm=0.372, loss_final=0.459, loss_mean=0.809, loss_mean_cls=0.068, proj_loss=-0.419][2026-03-26 13:02:06] Step: 7481, Training Logs: loss_final: 0.458901, loss_mean: 0.814822, proj_loss: -0.421497, loss_mean_cls: 0.065575, grad_norm: 0.211599 +Steps: 1%| | 7482/1000000 [31:03<68:15:02, 4.04it/s, grad_norm=0.212, loss_final=0.459, loss_mean=0.815, loss_mean_cls=0.0656, proj_loss=-0.421][2026-03-26 13:02:06] Step: 7482, Training Logs: loss_final: 0.468274, loss_mean: 0.824951, proj_loss: -0.423448, loss_mean_cls: 0.066770, grad_norm: 0.258898 +Steps: 1%| | 7483/1000000 [31:03<68:16:32, 4.04it/s, grad_norm=0.259, loss_final=0.468, loss_mean=0.825, loss_mean_cls=0.0668, proj_loss=-0.423][2026-03-26 13:02:06] Step: 7483, Training Logs: loss_final: 0.472115, loss_mean: 0.819572, proj_loss: -0.414844, loss_mean_cls: 0.067388, grad_norm: 0.226734 +Steps: 1%| | 7484/1000000 [31:03<68:14:32, 4.04it/s, grad_norm=0.227, loss_final=0.472, loss_mean=0.82, loss_mean_cls=0.0674, proj_loss=-0.415][2026-03-26 13:02:07] Step: 7484, Training Logs: loss_final: 0.456480, loss_mean: 0.814243, proj_loss: -0.423227, loss_mean_cls: 0.065464, grad_norm: 0.272145 +Steps: 1%| | 7485/1000000 [31:03<69:30:24, 3.97it/s, grad_norm=0.272, loss_final=0.456, loss_mean=0.814, loss_mean_cls=0.0655, proj_loss=-0.423][2026-03-26 13:02:07] Step: 7485, Training Logs: loss_final: 0.479986, loss_mean: 0.831825, proj_loss: -0.417887, loss_mean_cls: 0.066048, grad_norm: 0.342808 +Steps: 1%| | 7486/1000000 [31:04<69:09:28, 3.99it/s, grad_norm=0.343, loss_final=0.48, loss_mean=0.832, loss_mean_cls=0.066, proj_loss=-0.418][2026-03-26 13:02:07] Step: 7486, Training Logs: loss_final: 0.478607, loss_mean: 0.830631, proj_loss: -0.418683, loss_mean_cls: 0.066660, grad_norm: 0.208772 +Steps: 1%| | 7487/1000000 [31:04<68:52:33, 4.00it/s, grad_norm=0.209, loss_final=0.479, loss_mean=0.831, loss_mean_cls=0.0667, proj_loss=-0.419][2026-03-26 13:02:07] Step: 7487, Training Logs: loss_final: 0.463654, loss_mean: 0.815041, proj_loss: -0.417274, loss_mean_cls: 0.065887, grad_norm: 0.276726 +Steps: 1%| | 7488/1000000 [31:04<68:41:19, 4.01it/s, grad_norm=0.277, loss_final=0.464, loss_mean=0.815, loss_mean_cls=0.0659, proj_loss=-0.417][2026-03-26 13:02:08] Step: 7488, Training Logs: loss_final: 0.452374, loss_mean: 0.807948, proj_loss: -0.421600, loss_mean_cls: 0.066026, grad_norm: 0.261586 +Steps: 1%| | 7489/1000000 [31:04<68:32:01, 4.02it/s, grad_norm=0.262, loss_final=0.452, loss_mean=0.808, loss_mean_cls=0.066, proj_loss=-0.422][2026-03-26 13:02:08] Step: 7489, Training Logs: loss_final: 0.472275, loss_mean: 0.812675, proj_loss: -0.409157, loss_mean_cls: 0.068758, grad_norm: 0.212558 +Steps: 1%| | 7490/1000000 [31:05<68:30:05, 4.02it/s, grad_norm=0.213, loss_final=0.472, loss_mean=0.813, loss_mean_cls=0.0688, proj_loss=-0.409][2026-03-26 13:02:08] Step: 7490, Training Logs: loss_final: 0.481676, loss_mean: 0.836398, proj_loss: -0.420815, loss_mean_cls: 0.066093, grad_norm: 0.266896 +Steps: 1%| | 7491/1000000 [31:05<68:26:02, 4.03it/s, grad_norm=0.267, loss_final=0.482, loss_mean=0.836, loss_mean_cls=0.0661, proj_loss=-0.421][2026-03-26 13:02:08] Step: 7491, Training Logs: loss_final: 0.459227, loss_mean: 0.808537, proj_loss: -0.416508, loss_mean_cls: 0.067198, grad_norm: 0.260761 +Steps: 1%| | 7492/1000000 [31:05<68:20:16, 4.03it/s, grad_norm=0.261, loss_final=0.459, loss_mean=0.809, loss_mean_cls=0.0672, proj_loss=-0.417][2026-03-26 13:02:09] Step: 7492, Training Logs: loss_final: 0.467484, loss_mean: 0.810162, proj_loss: -0.410925, loss_mean_cls: 0.068247, grad_norm: 0.217102 +Steps: 1%| | 7493/1000000 [31:05<68:21:15, 4.03it/s, grad_norm=0.217, loss_final=0.467, loss_mean=0.81, loss_mean_cls=0.0682, proj_loss=-0.411][2026-03-26 13:02:09] Step: 7493, Training Logs: loss_final: 0.449636, loss_mean: 0.804171, proj_loss: -0.420634, loss_mean_cls: 0.066098, grad_norm: 0.277941 +Steps: 1%| | 7494/1000000 [31:06<68:24:20, 4.03it/s, grad_norm=0.278, loss_final=0.45, loss_mean=0.804, loss_mean_cls=0.0661, proj_loss=-0.421][2026-03-26 13:02:09] Step: 7494, Training Logs: loss_final: 0.460379, loss_mean: 0.813273, proj_loss: -0.420343, loss_mean_cls: 0.067449, grad_norm: 0.267566 +Steps: 1%| | 7495/1000000 [31:06<68:23:18, 4.03it/s, grad_norm=0.268, loss_final=0.46, loss_mean=0.813, loss_mean_cls=0.0674, proj_loss=-0.42][2026-03-26 13:02:09] Step: 7495, Training Logs: loss_final: 0.485830, loss_mean: 0.831215, proj_loss: -0.413859, loss_mean_cls: 0.068474, grad_norm: 0.354050 +Steps: 1%| | 7496/1000000 [31:06<68:19:53, 4.03it/s, grad_norm=0.354, loss_final=0.486, loss_mean=0.831, loss_mean_cls=0.0685, proj_loss=-0.414][2026-03-26 13:02:10] Step: 7496, Training Logs: loss_final: 0.483213, loss_mean: 0.824151, proj_loss: -0.410976, loss_mean_cls: 0.070037, grad_norm: 0.343775 +Steps: 1%| | 7497/1000000 [31:06<68:18:23, 4.04it/s, grad_norm=0.344, loss_final=0.483, loss_mean=0.824, loss_mean_cls=0.07, proj_loss=-0.411][2026-03-26 13:02:10] Step: 7497, Training Logs: loss_final: 0.460975, loss_mean: 0.809909, proj_loss: -0.416818, loss_mean_cls: 0.067884, grad_norm: 0.206170 +Steps: 1%| | 7498/1000000 [31:07<68:18:05, 4.04it/s, grad_norm=0.206, loss_final=0.461, loss_mean=0.81, loss_mean_cls=0.0679, proj_loss=-0.417][2026-03-26 13:02:10] Step: 7498, Training Logs: loss_final: 0.473881, loss_mean: 0.817043, proj_loss: -0.411311, loss_mean_cls: 0.068150, grad_norm: 0.327354 +Steps: 1%| | 7499/1000000 [31:07<68:16:06, 4.04it/s, grad_norm=0.327, loss_final=0.474, loss_mean=0.817, loss_mean_cls=0.0681, proj_loss=-0.411][2026-03-26 13:02:10] Step: 7499, Training Logs: loss_final: 0.475572, loss_mean: 0.826955, proj_loss: -0.418738, loss_mean_cls: 0.067354, grad_norm: 0.396224 +Steps: 1%| | 7500/1000000 [31:07<68:14:48, 4.04it/s, grad_norm=0.396, loss_final=0.476, loss_mean=0.827, loss_mean_cls=0.0674, proj_loss=-0.419][2026-03-26 13:02:11] Step: 7500, Training Logs: loss_final: 0.480407, loss_mean: 0.826573, proj_loss: -0.414405, loss_mean_cls: 0.068239, grad_norm: 0.279319 +Steps: 1%| | 7501/1000000 [31:07<68:14:01, 4.04it/s, grad_norm=0.279, loss_final=0.48, loss_mean=0.827, loss_mean_cls=0.0682, proj_loss=-0.414][2026-03-26 13:02:11] Step: 7501, Training Logs: loss_final: 0.461578, loss_mean: 0.812470, proj_loss: -0.418181, loss_mean_cls: 0.067288, grad_norm: 0.217969 +Steps: 1%| | 7502/1000000 [31:08<68:13:24, 4.04it/s, grad_norm=0.218, loss_final=0.462, loss_mean=0.812, loss_mean_cls=0.0673, proj_loss=-0.418][2026-03-26 13:02:11] Step: 7502, Training Logs: loss_final: 0.459939, loss_mean: 0.815055, proj_loss: -0.421633, loss_mean_cls: 0.066517, grad_norm: 0.360686 +Steps: 1%| | 7503/1000000 [31:08<68:15:01, 4.04it/s, grad_norm=0.361, loss_final=0.46, loss_mean=0.815, loss_mean_cls=0.0665, proj_loss=-0.422][2026-03-26 13:02:11] Step: 7503, Training Logs: loss_final: 0.471668, loss_mean: 0.831100, proj_loss: -0.424676, loss_mean_cls: 0.065244, grad_norm: 0.265340 +Steps: 1%| | 7504/1000000 [31:08<68:14:29, 4.04it/s, grad_norm=0.265, loss_final=0.472, loss_mean=0.831, loss_mean_cls=0.0652, proj_loss=-0.425][2026-03-26 13:02:12] Step: 7504, Training Logs: loss_final: 0.467485, loss_mean: 0.813635, proj_loss: -0.413980, loss_mean_cls: 0.067829, grad_norm: 0.258873 +Steps: 1%| | 7505/1000000 [31:08<68:12:31, 4.04it/s, grad_norm=0.259, loss_final=0.467, loss_mean=0.814, loss_mean_cls=0.0678, proj_loss=-0.414][2026-03-26 13:02:12] Step: 7505, Training Logs: loss_final: 0.470783, loss_mean: 0.813290, proj_loss: -0.411540, loss_mean_cls: 0.069034, grad_norm: 0.384283 +Steps: 1%| | 7506/1000000 [31:09<68:13:52, 4.04it/s, grad_norm=0.384, loss_final=0.471, loss_mean=0.813, loss_mean_cls=0.069, proj_loss=-0.412][2026-03-26 13:02:12] Step: 7506, Training Logs: loss_final: 0.461191, loss_mean: 0.807509, proj_loss: -0.413701, loss_mean_cls: 0.067383, grad_norm: 0.308696 +Steps: 1%| | 7507/1000000 [31:09<68:15:01, 4.04it/s, grad_norm=0.309, loss_final=0.461, loss_mean=0.808, loss_mean_cls=0.0674, proj_loss=-0.414][2026-03-26 13:02:12] Step: 7507, Training Logs: loss_final: 0.468658, loss_mean: 0.809538, proj_loss: -0.409892, loss_mean_cls: 0.069012, grad_norm: 0.473282 +Steps: 1%| | 7508/1000000 [31:09<68:18:09, 4.04it/s, grad_norm=0.473, loss_final=0.469, loss_mean=0.81, loss_mean_cls=0.069, proj_loss=-0.41][2026-03-26 13:02:13] Step: 7508, Training Logs: loss_final: 0.483802, loss_mean: 0.823513, proj_loss: -0.408444, loss_mean_cls: 0.068733, grad_norm: 0.308091 +Steps: 1%| | 7509/1000000 [31:09<68:13:10, 4.04it/s, grad_norm=0.308, loss_final=0.484, loss_mean=0.824, loss_mean_cls=0.0687, proj_loss=-0.408][2026-03-26 13:02:13] Step: 7509, Training Logs: loss_final: 0.462723, loss_mean: 0.812513, proj_loss: -0.417264, loss_mean_cls: 0.067474, grad_norm: 0.380509 +Steps: 1%| | 7510/1000000 [31:10<68:15:10, 4.04it/s, grad_norm=0.381, loss_final=0.463, loss_mean=0.813, loss_mean_cls=0.0675, proj_loss=-0.417][2026-03-26 13:02:13] Step: 7510, Training Logs: loss_final: 0.455731, loss_mean: 0.810774, proj_loss: -0.421898, loss_mean_cls: 0.066855, grad_norm: 0.386277 +Steps: 1%| | 7511/1000000 [31:10<68:11:36, 4.04it/s, grad_norm=0.386, loss_final=0.456, loss_mean=0.811, loss_mean_cls=0.0669, proj_loss=-0.422][2026-03-26 13:02:13] Step: 7511, Training Logs: loss_final: 0.466602, loss_mean: 0.807970, proj_loss: -0.410049, loss_mean_cls: 0.068680, grad_norm: 0.334108 +Steps: 1%| | 7512/1000000 [31:10<68:11:53, 4.04it/s, grad_norm=0.334, loss_final=0.467, loss_mean=0.808, loss_mean_cls=0.0687, proj_loss=-0.41][2026-03-26 13:02:14] Step: 7512, Training Logs: loss_final: 0.464359, loss_mean: 0.817730, proj_loss: -0.419919, loss_mean_cls: 0.066548, grad_norm: 0.395645 +Steps: 1%| | 7513/1000000 [31:10<68:12:29, 4.04it/s, grad_norm=0.396, loss_final=0.464, loss_mean=0.818, loss_mean_cls=0.0665, proj_loss=-0.42][2026-03-26 13:02:14] Step: 7513, Training Logs: loss_final: 0.460135, loss_mean: 0.803290, proj_loss: -0.411819, loss_mean_cls: 0.068664, grad_norm: 0.228792 +Steps: 1%| | 7514/1000000 [31:11<68:13:37, 4.04it/s, grad_norm=0.229, loss_final=0.46, loss_mean=0.803, loss_mean_cls=0.0687, proj_loss=-0.412][2026-03-26 13:02:14] Step: 7514, Training Logs: loss_final: 0.458680, loss_mean: 0.822340, proj_loss: -0.428308, loss_mean_cls: 0.064648, grad_norm: 0.311956 +Steps: 1%| | 7515/1000000 [31:11<68:12:22, 4.04it/s, grad_norm=0.312, loss_final=0.459, loss_mean=0.822, loss_mean_cls=0.0646, proj_loss=-0.428][2026-03-26 13:02:14] Step: 7515, Training Logs: loss_final: 0.464019, loss_mean: 0.822376, proj_loss: -0.423650, loss_mean_cls: 0.065294, grad_norm: 0.267872 +Steps: 1%| | 7516/1000000 [31:11<68:12:13, 4.04it/s, grad_norm=0.268, loss_final=0.464, loss_mean=0.822, loss_mean_cls=0.0653, proj_loss=-0.424][2026-03-26 13:02:15] Step: 7516, Training Logs: loss_final: 0.457327, loss_mean: 0.807341, proj_loss: -0.418065, loss_mean_cls: 0.068051, grad_norm: 0.359962 +Steps: 1%| | 7517/1000000 [31:11<68:12:55, 4.04it/s, grad_norm=0.36, loss_final=0.457, loss_mean=0.807, loss_mean_cls=0.0681, proj_loss=-0.418][2026-03-26 13:02:15] Step: 7517, Training Logs: loss_final: 0.456530, loss_mean: 0.796410, proj_loss: -0.408733, loss_mean_cls: 0.068853, grad_norm: 0.308142 +Steps: 1%| | 7518/1000000 [31:12<68:14:24, 4.04it/s, grad_norm=0.308, loss_final=0.457, loss_mean=0.796, loss_mean_cls=0.0689, proj_loss=-0.409][2026-03-26 13:02:15] Step: 7518, Training Logs: loss_final: 0.465177, loss_mean: 0.825409, proj_loss: -0.425652, loss_mean_cls: 0.065419, grad_norm: 0.233755 +Steps: 1%| | 7519/1000000 [31:12<68:12:07, 4.04it/s, grad_norm=0.234, loss_final=0.465, loss_mean=0.825, loss_mean_cls=0.0654, proj_loss=-0.426][2026-03-26 13:02:15] Step: 7519, Training Logs: loss_final: 0.469385, loss_mean: 0.802570, proj_loss: -0.403735, loss_mean_cls: 0.070550, grad_norm: 0.250797 +Steps: 1%| | 7520/1000000 [31:12<68:13:14, 4.04it/s, grad_norm=0.251, loss_final=0.469, loss_mean=0.803, loss_mean_cls=0.0706, proj_loss=-0.404][2026-03-26 13:02:16] Step: 7520, Training Logs: loss_final: 0.478088, loss_mean: 0.824440, proj_loss: -0.413547, loss_mean_cls: 0.067195, grad_norm: 0.238653 +Steps: 1%| | 7521/1000000 [31:12<68:12:19, 4.04it/s, grad_norm=0.239, loss_final=0.478, loss_mean=0.824, loss_mean_cls=0.0672, proj_loss=-0.414][2026-03-26 13:02:16] Step: 7521, Training Logs: loss_final: 0.472756, loss_mean: 0.817409, proj_loss: -0.413288, loss_mean_cls: 0.068636, grad_norm: 0.459335 +Steps: 1%| | 7522/1000000 [31:13<68:13:37, 4.04it/s, grad_norm=0.459, loss_final=0.473, loss_mean=0.817, loss_mean_cls=0.0686, proj_loss=-0.413][2026-03-26 13:02:16] Step: 7522, Training Logs: loss_final: 0.456074, loss_mean: 0.813747, proj_loss: -0.423305, loss_mean_cls: 0.065632, grad_norm: 0.215168 +Steps: 1%| | 7523/1000000 [31:13<68:21:24, 4.03it/s, grad_norm=0.215, loss_final=0.456, loss_mean=0.814, loss_mean_cls=0.0656, proj_loss=-0.423][2026-03-26 13:02:16] Step: 7523, Training Logs: loss_final: 0.495885, loss_mean: 0.844838, proj_loss: -0.416365, loss_mean_cls: 0.067412, grad_norm: 0.212478 +Steps: 1%| | 7524/1000000 [31:13<68:18:27, 4.04it/s, grad_norm=0.212, loss_final=0.496, loss_mean=0.845, loss_mean_cls=0.0674, proj_loss=-0.416][2026-03-26 13:02:17] Step: 7524, Training Logs: loss_final: 0.453471, loss_mean: 0.796434, proj_loss: -0.411531, loss_mean_cls: 0.068568, grad_norm: 0.264814 +Steps: 1%| | 7525/1000000 [31:13<68:17:10, 4.04it/s, grad_norm=0.265, loss_final=0.453, loss_mean=0.796, loss_mean_cls=0.0686, proj_loss=-0.412][2026-03-26 13:02:17] Step: 7525, Training Logs: loss_final: 0.457793, loss_mean: 0.807350, proj_loss: -0.416414, loss_mean_cls: 0.066857, grad_norm: 0.304558 +Steps: 1%| | 7526/1000000 [31:14<68:12:56, 4.04it/s, grad_norm=0.305, loss_final=0.458, loss_mean=0.807, loss_mean_cls=0.0669, proj_loss=-0.416][2026-03-26 13:02:17] Step: 7526, Training Logs: loss_final: 0.472073, loss_mean: 0.822588, proj_loss: -0.417869, loss_mean_cls: 0.067354, grad_norm: 0.250602 +Steps: 1%| | 7527/1000000 [31:14<68:13:01, 4.04it/s, grad_norm=0.251, loss_final=0.472, loss_mean=0.823, loss_mean_cls=0.0674, proj_loss=-0.418][2026-03-26 13:02:17] Step: 7527, Training Logs: loss_final: 0.472808, loss_mean: 0.827340, proj_loss: -0.420847, loss_mean_cls: 0.066314, grad_norm: 0.225722 +Steps: 1%| | 7528/1000000 [31:14<68:14:26, 4.04it/s, grad_norm=0.226, loss_final=0.473, loss_mean=0.827, loss_mean_cls=0.0663, proj_loss=-0.421][2026-03-26 13:02:18] Step: 7528, Training Logs: loss_final: 0.483346, loss_mean: 0.832078, proj_loss: -0.416654, loss_mean_cls: 0.067921, grad_norm: 0.214273 +Steps: 1%| | 7529/1000000 [31:14<68:12:32, 4.04it/s, grad_norm=0.214, loss_final=0.483, loss_mean=0.832, loss_mean_cls=0.0679, proj_loss=-0.417][2026-03-26 13:02:18] Step: 7529, Training Logs: loss_final: 0.473841, loss_mean: 0.816195, proj_loss: -0.411643, loss_mean_cls: 0.069289, grad_norm: 0.232935 +Steps: 1%| | 7530/1000000 [31:15<68:15:24, 4.04it/s, grad_norm=0.233, loss_final=0.474, loss_mean=0.816, loss_mean_cls=0.0693, proj_loss=-0.412][2026-03-26 13:02:18] Step: 7530, Training Logs: loss_final: 0.481096, loss_mean: 0.831789, proj_loss: -0.418254, loss_mean_cls: 0.067561, grad_norm: 0.269564 +Steps: 1%| | 7531/1000000 [31:15<68:15:00, 4.04it/s, grad_norm=0.27, loss_final=0.481, loss_mean=0.832, loss_mean_cls=0.0676, proj_loss=-0.418][2026-03-26 13:02:18] Step: 7531, Training Logs: loss_final: 0.466809, loss_mean: 0.825520, proj_loss: -0.425336, loss_mean_cls: 0.066625, grad_norm: 0.249588 +Steps: 1%| | 7532/1000000 [31:15<68:17:02, 4.04it/s, grad_norm=0.25, loss_final=0.467, loss_mean=0.826, loss_mean_cls=0.0666, proj_loss=-0.425][2026-03-26 13:02:19] Step: 7532, Training Logs: loss_final: 0.470175, loss_mean: 0.812639, proj_loss: -0.410478, loss_mean_cls: 0.068014, grad_norm: 0.362726 +Steps: 1%| | 7533/1000000 [31:15<68:14:22, 4.04it/s, grad_norm=0.363, loss_final=0.47, loss_mean=0.813, loss_mean_cls=0.068, proj_loss=-0.41][2026-03-26 13:02:19] Step: 7533, Training Logs: loss_final: 0.479557, loss_mean: 0.842801, proj_loss: -0.428069, loss_mean_cls: 0.064825, grad_norm: 0.231768 +Steps: 1%| | 7534/1000000 [31:16<68:13:10, 4.04it/s, grad_norm=0.232, loss_final=0.48, loss_mean=0.843, loss_mean_cls=0.0648, proj_loss=-0.428][2026-03-26 13:02:19] Step: 7534, Training Logs: loss_final: 0.460211, loss_mean: 0.816307, proj_loss: -0.421554, loss_mean_cls: 0.065458, grad_norm: 0.319691 +Steps: 1%| | 7535/1000000 [31:16<68:12:55, 4.04it/s, grad_norm=0.32, loss_final=0.46, loss_mean=0.816, loss_mean_cls=0.0655, proj_loss=-0.422][2026-03-26 13:02:19] Step: 7535, Training Logs: loss_final: 0.481317, loss_mean: 0.831653, proj_loss: -0.417609, loss_mean_cls: 0.067273, grad_norm: 0.437219 +Steps: 1%| | 7536/1000000 [31:16<68:13:07, 4.04it/s, grad_norm=0.437, loss_final=0.481, loss_mean=0.832, loss_mean_cls=0.0673, proj_loss=-0.418][2026-03-26 13:02:20] Step: 7536, Training Logs: loss_final: 0.481161, loss_mean: 0.825189, proj_loss: -0.413074, loss_mean_cls: 0.069046, grad_norm: 0.369616 +Steps: 1%| | 7537/1000000 [31:16<68:13:08, 4.04it/s, grad_norm=0.37, loss_final=0.481, loss_mean=0.825, loss_mean_cls=0.069, proj_loss=-0.413][2026-03-26 13:02:20] Step: 7537, Training Logs: loss_final: 0.460528, loss_mean: 0.814913, proj_loss: -0.421664, loss_mean_cls: 0.067280, grad_norm: 0.241181 +Steps: 1%| | 7538/1000000 [31:17<68:15:41, 4.04it/s, grad_norm=0.241, loss_final=0.461, loss_mean=0.815, loss_mean_cls=0.0673, proj_loss=-0.422][2026-03-26 13:02:20] Step: 7538, Training Logs: loss_final: 0.464371, loss_mean: 0.811345, proj_loss: -0.414832, loss_mean_cls: 0.067859, grad_norm: 0.236164 +Steps: 1%| | 7539/1000000 [31:17<68:15:29, 4.04it/s, grad_norm=0.236, loss_final=0.464, loss_mean=0.811, loss_mean_cls=0.0679, proj_loss=-0.415][2026-03-26 13:02:20] Step: 7539, Training Logs: loss_final: 0.456017, loss_mean: 0.803287, proj_loss: -0.415891, loss_mean_cls: 0.068621, grad_norm: 0.263283 +Steps: 1%| | 7540/1000000 [31:17<68:17:44, 4.04it/s, grad_norm=0.263, loss_final=0.456, loss_mean=0.803, loss_mean_cls=0.0686, proj_loss=-0.416][2026-03-26 13:02:20] Step: 7540, Training Logs: loss_final: 0.476204, loss_mean: 0.830347, proj_loss: -0.420964, loss_mean_cls: 0.066821, grad_norm: 0.228484 +Steps: 1%| | 7541/1000000 [31:17<68:15:43, 4.04it/s, grad_norm=0.228, loss_final=0.476, loss_mean=0.83, loss_mean_cls=0.0668, proj_loss=-0.421][2026-03-26 13:02:21] Step: 7541, Training Logs: loss_final: 0.476287, loss_mean: 0.816995, proj_loss: -0.409986, loss_mean_cls: 0.069278, grad_norm: 0.248793 +Steps: 1%| | 7542/1000000 [31:18<68:16:30, 4.04it/s, grad_norm=0.249, loss_final=0.476, loss_mean=0.817, loss_mean_cls=0.0693, proj_loss=-0.41][2026-03-26 13:02:21] Step: 7542, Training Logs: loss_final: 0.485000, loss_mean: 0.824504, proj_loss: -0.408111, loss_mean_cls: 0.068607, grad_norm: 0.224783 +Steps: 1%| | 7543/1000000 [31:18<68:17:56, 4.04it/s, grad_norm=0.225, loss_final=0.485, loss_mean=0.825, loss_mean_cls=0.0686, proj_loss=-0.408][2026-03-26 13:02:21] Step: 7543, Training Logs: loss_final: 0.461838, loss_mean: 0.817414, proj_loss: -0.421188, loss_mean_cls: 0.065612, grad_norm: 0.317782 +Steps: 1%| | 7544/1000000 [31:18<68:13:34, 4.04it/s, grad_norm=0.318, loss_final=0.462, loss_mean=0.817, loss_mean_cls=0.0656, proj_loss=-0.421][2026-03-26 13:02:21] Step: 7544, Training Logs: loss_final: 0.448072, loss_mean: 0.801616, proj_loss: -0.419870, loss_mean_cls: 0.066325, grad_norm: 0.225919 +Steps: 1%| | 7545/1000000 [31:18<68:17:18, 4.04it/s, grad_norm=0.226, loss_final=0.448, loss_mean=0.802, loss_mean_cls=0.0663, proj_loss=-0.42][2026-03-26 13:02:22] Step: 7545, Training Logs: loss_final: 0.475907, loss_mean: 0.823277, proj_loss: -0.415402, loss_mean_cls: 0.068032, grad_norm: 0.346322 +Steps: 1%| | 7546/1000000 [31:19<68:15:31, 4.04it/s, grad_norm=0.346, loss_final=0.476, loss_mean=0.823, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 13:02:22] Step: 7546, Training Logs: loss_final: 0.467377, loss_mean: 0.820334, proj_loss: -0.419168, loss_mean_cls: 0.066210, grad_norm: 0.492132 +Steps: 1%| | 7547/1000000 [31:19<68:16:24, 4.04it/s, grad_norm=0.492, loss_final=0.467, loss_mean=0.82, loss_mean_cls=0.0662, proj_loss=-0.419][2026-03-26 13:02:22] Step: 7547, Training Logs: loss_final: 0.469550, loss_mean: 0.823629, proj_loss: -0.421100, loss_mean_cls: 0.067021, grad_norm: 0.218121 +Steps: 1%| | 7548/1000000 [31:19<68:13:51, 4.04it/s, grad_norm=0.218, loss_final=0.47, loss_mean=0.824, loss_mean_cls=0.067, proj_loss=-0.421][2026-03-26 13:02:22] Step: 7548, Training Logs: loss_final: 0.462271, loss_mean: 0.805694, proj_loss: -0.411307, loss_mean_cls: 0.067884, grad_norm: 0.353488 +Steps: 1%| | 7549/1000000 [31:19<68:13:06, 4.04it/s, grad_norm=0.353, loss_final=0.462, loss_mean=0.806, loss_mean_cls=0.0679, proj_loss=-0.411][2026-03-26 13:02:23] Step: 7549, Training Logs: loss_final: 0.471674, loss_mean: 0.821475, proj_loss: -0.417032, loss_mean_cls: 0.067230, grad_norm: 0.214524 +Steps: 1%| | 7550/1000000 [31:20<68:11:29, 4.04it/s, grad_norm=0.215, loss_final=0.472, loss_mean=0.821, loss_mean_cls=0.0672, proj_loss=-0.417][2026-03-26 13:02:23] Step: 7550, Training Logs: loss_final: 0.465267, loss_mean: 0.815820, proj_loss: -0.417554, loss_mean_cls: 0.067000, grad_norm: 0.740529 +Steps: 1%| | 7551/1000000 [31:20<68:13:30, 4.04it/s, grad_norm=0.741, loss_final=0.465, loss_mean=0.816, loss_mean_cls=0.067, proj_loss=-0.418][2026-03-26 13:02:23] Step: 7551, Training Logs: loss_final: 0.463050, loss_mean: 0.809923, proj_loss: -0.414758, loss_mean_cls: 0.067886, grad_norm: 0.598285 +Steps: 1%| | 7552/1000000 [31:20<68:12:17, 4.04it/s, grad_norm=0.598, loss_final=0.463, loss_mean=0.81, loss_mean_cls=0.0679, proj_loss=-0.415][2026-03-26 13:02:23] Step: 7552, Training Logs: loss_final: 0.471261, loss_mean: 0.818498, proj_loss: -0.414546, loss_mean_cls: 0.067308, grad_norm: 0.200762 +Steps: 1%| | 7553/1000000 [31:20<68:11:56, 4.04it/s, grad_norm=0.201, loss_final=0.471, loss_mean=0.818, loss_mean_cls=0.0673, proj_loss=-0.415][2026-03-26 13:02:24] Step: 7553, Training Logs: loss_final: 0.478209, loss_mean: 0.843584, proj_loss: -0.429636, loss_mean_cls: 0.064260, grad_norm: 0.441430 +Steps: 1%| | 7554/1000000 [31:21<68:11:41, 4.04it/s, grad_norm=0.441, loss_final=0.478, loss_mean=0.844, loss_mean_cls=0.0643, proj_loss=-0.43][2026-03-26 13:02:24] Step: 7554, Training Logs: loss_final: 0.464312, loss_mean: 0.810040, proj_loss: -0.414398, loss_mean_cls: 0.068670, grad_norm: 0.305969 +Steps: 1%| | 7555/1000000 [31:21<68:13:43, 4.04it/s, grad_norm=0.306, loss_final=0.464, loss_mean=0.81, loss_mean_cls=0.0687, proj_loss=-0.414][2026-03-26 13:02:24] Step: 7555, Training Logs: loss_final: 0.478371, loss_mean: 0.814325, proj_loss: -0.406292, loss_mean_cls: 0.070338, grad_norm: 0.415662 +Steps: 1%| | 7556/1000000 [31:21<68:10:27, 4.04it/s, grad_norm=0.416, loss_final=0.478, loss_mean=0.814, loss_mean_cls=0.0703, proj_loss=-0.406][2026-03-26 13:02:24] Step: 7556, Training Logs: loss_final: 0.475881, loss_mean: 0.817652, proj_loss: -0.411255, loss_mean_cls: 0.069484, grad_norm: 0.452836 +Steps: 1%| | 7557/1000000 [31:21<68:09:55, 4.04it/s, grad_norm=0.453, loss_final=0.476, loss_mean=0.818, loss_mean_cls=0.0695, proj_loss=-0.411][2026-03-26 13:02:25] Step: 7557, Training Logs: loss_final: 0.470213, loss_mean: 0.814735, proj_loss: -0.412887, loss_mean_cls: 0.068365, grad_norm: 0.415638 +Steps: 1%| | 7558/1000000 [31:22<68:14:03, 4.04it/s, grad_norm=0.416, loss_final=0.47, loss_mean=0.815, loss_mean_cls=0.0684, proj_loss=-0.413][2026-03-26 13:02:25] Step: 7558, Training Logs: loss_final: 0.468553, loss_mean: 0.819283, proj_loss: -0.418236, loss_mean_cls: 0.067506, grad_norm: 0.490751 +Steps: 1%| | 7559/1000000 [31:22<68:13:38, 4.04it/s, grad_norm=0.491, loss_final=0.469, loss_mean=0.819, loss_mean_cls=0.0675, proj_loss=-0.418][2026-03-26 13:02:25] Step: 7559, Training Logs: loss_final: 0.460956, loss_mean: 0.812251, proj_loss: -0.418116, loss_mean_cls: 0.066820, grad_norm: 0.235297 +Steps: 1%| | 7560/1000000 [31:22<68:15:31, 4.04it/s, grad_norm=0.235, loss_final=0.461, loss_mean=0.812, loss_mean_cls=0.0668, proj_loss=-0.418][2026-03-26 13:02:25] Step: 7560, Training Logs: loss_final: 0.469322, loss_mean: 0.815959, proj_loss: -0.414873, loss_mean_cls: 0.068237, grad_norm: 0.497835 +Steps: 1%| | 7561/1000000 [31:22<68:11:15, 4.04it/s, grad_norm=0.498, loss_final=0.469, loss_mean=0.816, loss_mean_cls=0.0682, proj_loss=-0.415][2026-03-26 13:02:26] Step: 7561, Training Logs: loss_final: 0.477574, loss_mean: 0.820413, proj_loss: -0.411892, loss_mean_cls: 0.069053, grad_norm: 0.424203 +Steps: 1%| | 7562/1000000 [31:23<68:14:53, 4.04it/s, grad_norm=0.424, loss_final=0.478, loss_mean=0.82, loss_mean_cls=0.0691, proj_loss=-0.412][2026-03-26 13:02:26] Step: 7562, Training Logs: loss_final: 0.497111, loss_mean: 0.850088, proj_loss: -0.419016, loss_mean_cls: 0.066039, grad_norm: 0.228297 +Steps: 1%| | 7563/1000000 [31:23<68:13:59, 4.04it/s, grad_norm=0.228, loss_final=0.497, loss_mean=0.85, loss_mean_cls=0.066, proj_loss=-0.419][2026-03-26 13:02:26] Step: 7563, Training Logs: loss_final: 0.462690, loss_mean: 0.806904, proj_loss: -0.411546, loss_mean_cls: 0.067331, grad_norm: 0.378746 +Steps: 1%| | 7564/1000000 [31:23<68:11:51, 4.04it/s, grad_norm=0.379, loss_final=0.463, loss_mean=0.807, loss_mean_cls=0.0673, proj_loss=-0.412][2026-03-26 13:02:26] Step: 7564, Training Logs: loss_final: 0.476879, loss_mean: 0.830102, proj_loss: -0.419745, loss_mean_cls: 0.066522, grad_norm: 0.274981 +Steps: 1%| | 7565/1000000 [31:23<68:10:44, 4.04it/s, grad_norm=0.275, loss_final=0.477, loss_mean=0.83, loss_mean_cls=0.0665, proj_loss=-0.42][2026-03-26 13:02:27] Step: 7565, Training Logs: loss_final: 0.464330, loss_mean: 0.811229, proj_loss: -0.413836, loss_mean_cls: 0.066936, grad_norm: 0.224561 +Steps: 1%| | 7566/1000000 [31:24<68:12:38, 4.04it/s, grad_norm=0.225, loss_final=0.464, loss_mean=0.811, loss_mean_cls=0.0669, proj_loss=-0.414][2026-03-26 13:02:27] Step: 7566, Training Logs: loss_final: 0.468575, loss_mean: 0.819338, proj_loss: -0.417647, loss_mean_cls: 0.066884, grad_norm: 0.269188 +Steps: 1%| | 7567/1000000 [31:24<68:12:57, 4.04it/s, grad_norm=0.269, loss_final=0.469, loss_mean=0.819, loss_mean_cls=0.0669, proj_loss=-0.418][2026-03-26 13:02:27] Step: 7567, Training Logs: loss_final: 0.478397, loss_mean: 0.825839, proj_loss: -0.414726, loss_mean_cls: 0.067284, grad_norm: 0.229320 +Steps: 1%| | 7568/1000000 [31:24<68:23:31, 4.03it/s, grad_norm=0.229, loss_final=0.478, loss_mean=0.826, loss_mean_cls=0.0673, proj_loss=-0.415][2026-03-26 13:02:27] Step: 7568, Training Logs: loss_final: 0.467715, loss_mean: 0.819738, proj_loss: -0.418858, loss_mean_cls: 0.066835, grad_norm: 0.303099 +Steps: 1%| | 7569/1000000 [31:24<68:19:34, 4.03it/s, grad_norm=0.303, loss_final=0.468, loss_mean=0.82, loss_mean_cls=0.0668, proj_loss=-0.419][2026-03-26 13:02:28] Step: 7569, Training Logs: loss_final: 0.465267, loss_mean: 0.823572, proj_loss: -0.423889, loss_mean_cls: 0.065585, grad_norm: 0.315604 +Steps: 1%| | 7570/1000000 [31:25<68:24:37, 4.03it/s, grad_norm=0.316, loss_final=0.465, loss_mean=0.824, loss_mean_cls=0.0656, proj_loss=-0.424][2026-03-26 13:02:28] Step: 7570, Training Logs: loss_final: 0.456846, loss_mean: 0.804484, proj_loss: -0.415083, loss_mean_cls: 0.067444, grad_norm: 0.321197 +Steps: 1%| | 7571/1000000 [31:25<68:19:41, 4.03it/s, grad_norm=0.321, loss_final=0.457, loss_mean=0.804, loss_mean_cls=0.0674, proj_loss=-0.415][2026-03-26 13:02:28] Step: 7571, Training Logs: loss_final: 0.463026, loss_mean: 0.812845, proj_loss: -0.416572, loss_mean_cls: 0.066753, grad_norm: 0.317032 +Steps: 1%| | 7572/1000000 [31:25<68:20:35, 4.03it/s, grad_norm=0.317, loss_final=0.463, loss_mean=0.813, loss_mean_cls=0.0668, proj_loss=-0.417][2026-03-26 13:02:28] Step: 7572, Training Logs: loss_final: 0.473281, loss_mean: 0.824052, proj_loss: -0.418208, loss_mean_cls: 0.067437, grad_norm: 0.360658 +Steps: 1%| | 7573/1000000 [31:25<68:14:10, 4.04it/s, grad_norm=0.361, loss_final=0.473, loss_mean=0.824, loss_mean_cls=0.0674, proj_loss=-0.418][2026-03-26 13:02:29] Step: 7573, Training Logs: loss_final: 0.457195, loss_mean: 0.804592, proj_loss: -0.415276, loss_mean_cls: 0.067880, grad_norm: 0.213765 +Steps: 1%| | 7574/1000000 [31:26<68:16:36, 4.04it/s, grad_norm=0.214, loss_final=0.457, loss_mean=0.805, loss_mean_cls=0.0679, proj_loss=-0.415][2026-03-26 13:02:29] Step: 7574, Training Logs: loss_final: 0.478956, loss_mean: 0.833483, proj_loss: -0.420539, loss_mean_cls: 0.066012, grad_norm: 0.194291 +Steps: 1%| | 7575/1000000 [31:26<68:17:02, 4.04it/s, grad_norm=0.194, loss_final=0.479, loss_mean=0.833, loss_mean_cls=0.066, proj_loss=-0.421][2026-03-26 13:02:29] Step: 7575, Training Logs: loss_final: 0.464159, loss_mean: 0.820788, proj_loss: -0.422701, loss_mean_cls: 0.066072, grad_norm: 0.319618 +Steps: 1%| | 7576/1000000 [31:26<68:13:54, 4.04it/s, grad_norm=0.32, loss_final=0.464, loss_mean=0.821, loss_mean_cls=0.0661, proj_loss=-0.423][2026-03-26 13:02:29] Step: 7576, Training Logs: loss_final: 0.461977, loss_mean: 0.800773, proj_loss: -0.408857, loss_mean_cls: 0.070061, grad_norm: 0.415067 +Steps: 1%| | 7577/1000000 [31:26<68:18:17, 4.04it/s, grad_norm=0.415, loss_final=0.462, loss_mean=0.801, loss_mean_cls=0.0701, proj_loss=-0.409][2026-03-26 13:02:30] Step: 7577, Training Logs: loss_final: 0.473697, loss_mean: 0.806643, proj_loss: -0.403203, loss_mean_cls: 0.070257, grad_norm: 0.239469 +Steps: 1%| | 7578/1000000 [31:27<68:16:24, 4.04it/s, grad_norm=0.239, loss_final=0.474, loss_mean=0.807, loss_mean_cls=0.0703, proj_loss=-0.403][2026-03-26 13:02:30] Step: 7578, Training Logs: loss_final: 0.476535, loss_mean: 0.819911, proj_loss: -0.411773, loss_mean_cls: 0.068398, grad_norm: 0.455891 +Steps: 1%| | 7579/1000000 [31:27<68:15:34, 4.04it/s, grad_norm=0.456, loss_final=0.477, loss_mean=0.82, loss_mean_cls=0.0684, proj_loss=-0.412][2026-03-26 13:02:30] Step: 7579, Training Logs: loss_final: 0.468867, loss_mean: 0.829351, proj_loss: -0.424877, loss_mean_cls: 0.064393, grad_norm: 0.444292 +Steps: 1%| | 7580/1000000 [31:27<68:13:02, 4.04it/s, grad_norm=0.444, loss_final=0.469, loss_mean=0.829, loss_mean_cls=0.0644, proj_loss=-0.425][2026-03-26 13:02:30] Step: 7580, Training Logs: loss_final: 0.460491, loss_mean: 0.811023, proj_loss: -0.418078, loss_mean_cls: 0.067546, grad_norm: 0.287925 +Steps: 1%| | 7581/1000000 [31:27<68:16:38, 4.04it/s, grad_norm=0.288, loss_final=0.46, loss_mean=0.811, loss_mean_cls=0.0675, proj_loss=-0.418][2026-03-26 13:02:31] Step: 7581, Training Logs: loss_final: 0.470840, loss_mean: 0.823830, proj_loss: -0.419510, loss_mean_cls: 0.066520, grad_norm: 0.321653 +Steps: 1%| | 7582/1000000 [31:28<68:15:22, 4.04it/s, grad_norm=0.322, loss_final=0.471, loss_mean=0.824, loss_mean_cls=0.0665, proj_loss=-0.42][2026-03-26 13:02:31] Step: 7582, Training Logs: loss_final: 0.466626, loss_mean: 0.817925, proj_loss: -0.418606, loss_mean_cls: 0.067306, grad_norm: 0.297164 +Steps: 1%| | 7583/1000000 [31:28<68:18:06, 4.04it/s, grad_norm=0.297, loss_final=0.467, loss_mean=0.818, loss_mean_cls=0.0673, proj_loss=-0.419][2026-03-26 13:02:31] Step: 7583, Training Logs: loss_final: 0.458474, loss_mean: 0.806968, proj_loss: -0.416585, loss_mean_cls: 0.068091, grad_norm: 0.282462 +Steps: 1%| | 7584/1000000 [31:28<68:15:58, 4.04it/s, grad_norm=0.282, loss_final=0.458, loss_mean=0.807, loss_mean_cls=0.0681, proj_loss=-0.417][2026-03-26 13:02:31] Step: 7584, Training Logs: loss_final: 0.469362, loss_mean: 0.803646, proj_loss: -0.404611, loss_mean_cls: 0.070327, grad_norm: 0.376630 +Steps: 1%| | 7585/1000000 [31:28<68:15:00, 4.04it/s, grad_norm=0.377, loss_final=0.469, loss_mean=0.804, loss_mean_cls=0.0703, proj_loss=-0.405][2026-03-26 13:02:32] Step: 7585, Training Logs: loss_final: 0.451298, loss_mean: 0.798013, proj_loss: -0.413956, loss_mean_cls: 0.067241, grad_norm: 0.246748 +Steps: 1%| | 7586/1000000 [31:28<68:27:39, 4.03it/s, grad_norm=0.247, loss_final=0.451, loss_mean=0.798, loss_mean_cls=0.0672, proj_loss=-0.414][2026-03-26 13:02:32] Step: 7586, Training Logs: loss_final: 0.465984, loss_mean: 0.816597, proj_loss: -0.417451, loss_mean_cls: 0.066839, grad_norm: 0.271046 +Steps: 1%| | 7587/1000000 [31:29<68:22:28, 4.03it/s, grad_norm=0.271, loss_final=0.466, loss_mean=0.817, loss_mean_cls=0.0668, proj_loss=-0.417][2026-03-26 13:02:32] Step: 7587, Training Logs: loss_final: 0.476306, loss_mean: 0.835547, proj_loss: -0.423918, loss_mean_cls: 0.064677, grad_norm: 0.345675 +Steps: 1%| | 7588/1000000 [31:29<68:17:48, 4.04it/s, grad_norm=0.346, loss_final=0.476, loss_mean=0.836, loss_mean_cls=0.0647, proj_loss=-0.424][2026-03-26 13:02:32] Step: 7588, Training Logs: loss_final: 0.453942, loss_mean: 0.806941, proj_loss: -0.420128, loss_mean_cls: 0.067129, grad_norm: 0.381230 +Steps: 1%| | 7589/1000000 [31:29<68:15:27, 4.04it/s, grad_norm=0.381, loss_final=0.454, loss_mean=0.807, loss_mean_cls=0.0671, proj_loss=-0.42][2026-03-26 13:02:33] Step: 7589, Training Logs: loss_final: 0.469194, loss_mean: 0.819709, proj_loss: -0.417973, loss_mean_cls: 0.067458, grad_norm: 0.264503 +Steps: 1%| | 7590/1000000 [31:29<68:12:51, 4.04it/s, grad_norm=0.265, loss_final=0.469, loss_mean=0.82, loss_mean_cls=0.0675, proj_loss=-0.418][2026-03-26 13:02:33] Step: 7590, Training Logs: loss_final: 0.473468, loss_mean: 0.818328, proj_loss: -0.413776, loss_mean_cls: 0.068916, grad_norm: 0.185271 +Steps: 1%| | 7591/1000000 [31:30<68:14:52, 4.04it/s, grad_norm=0.185, loss_final=0.473, loss_mean=0.818, loss_mean_cls=0.0689, proj_loss=-0.414][2026-03-26 13:02:33] Step: 7591, Training Logs: loss_final: 0.456237, loss_mean: 0.794323, proj_loss: -0.408136, loss_mean_cls: 0.070051, grad_norm: 0.250251 +Steps: 1%| | 7592/1000000 [31:30<68:14:47, 4.04it/s, grad_norm=0.25, loss_final=0.456, loss_mean=0.794, loss_mean_cls=0.0701, proj_loss=-0.408][2026-03-26 13:02:33] Step: 7592, Training Logs: loss_final: 0.483289, loss_mean: 0.829402, proj_loss: -0.414279, loss_mean_cls: 0.068166, grad_norm: 0.403262 +Steps: 1%| | 7593/1000000 [31:30<68:11:42, 4.04it/s, grad_norm=0.403, loss_final=0.483, loss_mean=0.829, loss_mean_cls=0.0682, proj_loss=-0.414][2026-03-26 13:02:34] Step: 7593, Training Logs: loss_final: 0.462990, loss_mean: 0.814435, proj_loss: -0.419059, loss_mean_cls: 0.067613, grad_norm: 0.295140 +Steps: 1%| | 7594/1000000 [31:30<68:11:05, 4.04it/s, grad_norm=0.295, loss_final=0.463, loss_mean=0.814, loss_mean_cls=0.0676, proj_loss=-0.419][2026-03-26 13:02:34] Step: 7594, Training Logs: loss_final: 0.477073, loss_mean: 0.823186, proj_loss: -0.414262, loss_mean_cls: 0.068149, grad_norm: 0.391313 +Steps: 1%| | 7595/1000000 [31:31<68:10:46, 4.04it/s, grad_norm=0.391, loss_final=0.477, loss_mean=0.823, loss_mean_cls=0.0681, proj_loss=-0.414][2026-03-26 13:02:34] Step: 7595, Training Logs: loss_final: 0.466557, loss_mean: 0.811709, proj_loss: -0.412894, loss_mean_cls: 0.067742, grad_norm: 0.417628 +Steps: 1%| | 7596/1000000 [31:31<68:11:19, 4.04it/s, grad_norm=0.418, loss_final=0.467, loss_mean=0.812, loss_mean_cls=0.0677, proj_loss=-0.413][2026-03-26 13:02:34] Step: 7596, Training Logs: loss_final: 0.469900, loss_mean: 0.812670, proj_loss: -0.410653, loss_mean_cls: 0.067883, grad_norm: 0.392108 +Steps: 1%| | 7597/1000000 [31:31<68:11:42, 4.04it/s, grad_norm=0.392, loss_final=0.47, loss_mean=0.813, loss_mean_cls=0.0679, proj_loss=-0.411][2026-03-26 13:02:35] Step: 7597, Training Logs: loss_final: 0.462560, loss_mean: 0.811741, proj_loss: -0.416431, loss_mean_cls: 0.067250, grad_norm: 0.199208 +Steps: 1%| | 7598/1000000 [31:31<68:13:08, 4.04it/s, grad_norm=0.199, loss_final=0.463, loss_mean=0.812, loss_mean_cls=0.0673, proj_loss=-0.416][2026-03-26 13:02:35] Step: 7598, Training Logs: loss_final: 0.447031, loss_mean: 0.797631, proj_loss: -0.417722, loss_mean_cls: 0.067122, grad_norm: 0.316192 +Steps: 1%| | 7599/1000000 [31:32<68:12:13, 4.04it/s, grad_norm=0.316, loss_final=0.447, loss_mean=0.798, loss_mean_cls=0.0671, proj_loss=-0.418][2026-03-26 13:02:35] Step: 7599, Training Logs: loss_final: 0.476433, loss_mean: 0.839118, proj_loss: -0.426873, loss_mean_cls: 0.064188, grad_norm: 0.381569 +Steps: 1%| | 7600/1000000 [31:32<68:11:39, 4.04it/s, grad_norm=0.382, loss_final=0.476, loss_mean=0.839, loss_mean_cls=0.0642, proj_loss=-0.427][2026-03-26 13:02:35] Step: 7600, Training Logs: loss_final: 0.476294, loss_mean: 0.826672, proj_loss: -0.417672, loss_mean_cls: 0.067294, grad_norm: 0.472412 +Steps: 1%| | 7601/1000000 [31:32<68:08:41, 4.05it/s, grad_norm=0.472, loss_final=0.476, loss_mean=0.827, loss_mean_cls=0.0673, proj_loss=-0.418][2026-03-26 13:02:36] Step: 7601, Training Logs: loss_final: 0.489426, loss_mean: 0.833659, proj_loss: -0.411647, loss_mean_cls: 0.067414, grad_norm: 0.427022 +Steps: 1%| | 7602/1000000 [31:32<68:09:50, 4.04it/s, grad_norm=0.427, loss_final=0.489, loss_mean=0.834, loss_mean_cls=0.0674, proj_loss=-0.412][2026-03-26 13:02:36] Step: 7602, Training Logs: loss_final: 0.475085, loss_mean: 0.817117, proj_loss: -0.410350, loss_mean_cls: 0.068318, grad_norm: 0.246520 +Steps: 1%| | 7603/1000000 [31:33<68:10:32, 4.04it/s, grad_norm=0.247, loss_final=0.475, loss_mean=0.817, loss_mean_cls=0.0683, proj_loss=-0.41][2026-03-26 13:02:36] Step: 7603, Training Logs: loss_final: 0.458041, loss_mean: 0.816591, proj_loss: -0.423653, loss_mean_cls: 0.065103, grad_norm: 0.554582 +Steps: 1%| | 7604/1000000 [31:33<68:11:55, 4.04it/s, grad_norm=0.555, loss_final=0.458, loss_mean=0.817, loss_mean_cls=0.0651, proj_loss=-0.424][2026-03-26 13:02:36] Step: 7604, Training Logs: loss_final: 0.475138, loss_mean: 0.815337, proj_loss: -0.408875, loss_mean_cls: 0.068676, grad_norm: 0.352778 +Steps: 1%| | 7605/1000000 [31:33<68:06:09, 4.05it/s, grad_norm=0.353, loss_final=0.475, loss_mean=0.815, loss_mean_cls=0.0687, proj_loss=-0.409][2026-03-26 13:02:37] Step: 7605, Training Logs: loss_final: 0.471041, loss_mean: 0.811595, proj_loss: -0.409360, loss_mean_cls: 0.068806, grad_norm: 0.464143 +Steps: 1%| | 7606/1000000 [31:33<68:06:03, 4.05it/s, grad_norm=0.464, loss_final=0.471, loss_mean=0.812, loss_mean_cls=0.0688, proj_loss=-0.409][2026-03-26 13:02:37] Step: 7606, Training Logs: loss_final: 0.467383, loss_mean: 0.806153, proj_loss: -0.407855, loss_mean_cls: 0.069084, grad_norm: 0.541616 +Steps: 1%| | 7607/1000000 [31:34<68:09:40, 4.04it/s, grad_norm=0.542, loss_final=0.467, loss_mean=0.806, loss_mean_cls=0.0691, proj_loss=-0.408][2026-03-26 13:02:37] Step: 7607, Training Logs: loss_final: 0.490386, loss_mean: 0.829563, proj_loss: -0.408457, loss_mean_cls: 0.069281, grad_norm: 0.302361 +Steps: 1%| | 7608/1000000 [31:34<68:11:52, 4.04it/s, grad_norm=0.302, loss_final=0.49, loss_mean=0.83, loss_mean_cls=0.0693, proj_loss=-0.408][2026-03-26 13:02:37] Step: 7608, Training Logs: loss_final: 0.457480, loss_mean: 0.810503, proj_loss: -0.419823, loss_mean_cls: 0.066800, grad_norm: 0.670583 +Steps: 1%| | 7609/1000000 [31:34<68:08:36, 4.05it/s, grad_norm=0.671, loss_final=0.457, loss_mean=0.811, loss_mean_cls=0.0668, proj_loss=-0.42][2026-03-26 13:02:38] Step: 7609, Training Logs: loss_final: 0.446855, loss_mean: 0.794620, proj_loss: -0.416462, loss_mean_cls: 0.068696, grad_norm: 0.316124 +Steps: 1%| | 7610/1000000 [31:34<68:09:39, 4.04it/s, grad_norm=0.316, loss_final=0.447, loss_mean=0.795, loss_mean_cls=0.0687, proj_loss=-0.416][2026-03-26 13:02:38] Step: 7610, Training Logs: loss_final: 0.453800, loss_mean: 0.806564, proj_loss: -0.420299, loss_mean_cls: 0.067535, grad_norm: 0.506397 +Steps: 1%| | 7611/1000000 [31:35<68:12:54, 4.04it/s, grad_norm=0.506, loss_final=0.454, loss_mean=0.807, loss_mean_cls=0.0675, proj_loss=-0.42][2026-03-26 13:02:38] Step: 7611, Training Logs: loss_final: 0.479787, loss_mean: 0.823641, proj_loss: -0.411258, loss_mean_cls: 0.067404, grad_norm: 0.530869 +Steps: 1%| | 7612/1000000 [31:35<68:14:15, 4.04it/s, grad_norm=0.531, loss_final=0.48, loss_mean=0.824, loss_mean_cls=0.0674, proj_loss=-0.411][2026-03-26 13:02:38] Step: 7612, Training Logs: loss_final: 0.477907, loss_mean: 0.827287, proj_loss: -0.416476, loss_mean_cls: 0.067096, grad_norm: 0.229376 +Steps: 1%| | 7613/1000000 [31:35<68:12:30, 4.04it/s, grad_norm=0.229, loss_final=0.478, loss_mean=0.827, loss_mean_cls=0.0671, proj_loss=-0.416][2026-03-26 13:02:39] Step: 7613, Training Logs: loss_final: 0.465051, loss_mean: 0.815765, proj_loss: -0.418021, loss_mean_cls: 0.067307, grad_norm: 0.496978 +Steps: 1%| | 7614/1000000 [31:35<68:12:42, 4.04it/s, grad_norm=0.497, loss_final=0.465, loss_mean=0.816, loss_mean_cls=0.0673, proj_loss=-0.418][2026-03-26 13:02:39] Step: 7614, Training Logs: loss_final: 0.475012, loss_mean: 0.821433, proj_loss: -0.414375, loss_mean_cls: 0.067954, grad_norm: 0.417880 +Steps: 1%| | 7615/1000000 [31:36<68:14:40, 4.04it/s, grad_norm=0.418, loss_final=0.475, loss_mean=0.821, loss_mean_cls=0.068, proj_loss=-0.414][2026-03-26 13:02:39] Step: 7615, Training Logs: loss_final: 0.484301, loss_mean: 0.840275, proj_loss: -0.422188, loss_mean_cls: 0.066214, grad_norm: 0.263526 +Steps: 1%| | 7616/1000000 [31:36<68:16:12, 4.04it/s, grad_norm=0.264, loss_final=0.484, loss_mean=0.84, loss_mean_cls=0.0662, proj_loss=-0.422][2026-03-26 13:02:39] Step: 7616, Training Logs: loss_final: 0.480300, loss_mean: 0.823388, proj_loss: -0.411582, loss_mean_cls: 0.068494, grad_norm: 0.341210 +Steps: 1%| | 7617/1000000 [31:36<68:14:14, 4.04it/s, grad_norm=0.341, loss_final=0.48, loss_mean=0.823, loss_mean_cls=0.0685, proj_loss=-0.412][2026-03-26 13:02:40] Step: 7617, Training Logs: loss_final: 0.461809, loss_mean: 0.813342, proj_loss: -0.417549, loss_mean_cls: 0.066017, grad_norm: 0.348283 +Steps: 1%| | 7618/1000000 [31:36<68:15:33, 4.04it/s, grad_norm=0.348, loss_final=0.462, loss_mean=0.813, loss_mean_cls=0.066, proj_loss=-0.418][2026-03-26 13:02:40] Step: 7618, Training Logs: loss_final: 0.466647, loss_mean: 0.817379, proj_loss: -0.417231, loss_mean_cls: 0.066500, grad_norm: 0.322448 +Steps: 1%| | 7619/1000000 [31:37<68:16:00, 4.04it/s, grad_norm=0.322, loss_final=0.467, loss_mean=0.817, loss_mean_cls=0.0665, proj_loss=-0.417][2026-03-26 13:02:40] Step: 7619, Training Logs: loss_final: 0.470355, loss_mean: 0.831509, proj_loss: -0.426605, loss_mean_cls: 0.065451, grad_norm: 0.406173 +Steps: 1%| | 7620/1000000 [31:37<68:15:47, 4.04it/s, grad_norm=0.406, loss_final=0.47, loss_mean=0.832, loss_mean_cls=0.0655, proj_loss=-0.427][2026-03-26 13:02:40] Step: 7620, Training Logs: loss_final: 0.464350, loss_mean: 0.813328, proj_loss: -0.415717, loss_mean_cls: 0.066738, grad_norm: 0.188461 +Steps: 1%| | 7621/1000000 [31:37<68:16:12, 4.04it/s, grad_norm=0.188, loss_final=0.464, loss_mean=0.813, loss_mean_cls=0.0667, proj_loss=-0.416][2026-03-26 13:02:41] Step: 7621, Training Logs: loss_final: 0.486439, loss_mean: 0.833593, proj_loss: -0.414943, loss_mean_cls: 0.067789, grad_norm: 0.347655 +Steps: 1%| | 7622/1000000 [31:37<68:13:55, 4.04it/s, grad_norm=0.348, loss_final=0.486, loss_mean=0.834, loss_mean_cls=0.0678, proj_loss=-0.415][2026-03-26 13:02:41] Step: 7622, Training Logs: loss_final: 0.458671, loss_mean: 0.794521, proj_loss: -0.406796, loss_mean_cls: 0.070946, grad_norm: 0.244856 +Steps: 1%| | 7623/1000000 [31:38<68:15:06, 4.04it/s, grad_norm=0.245, loss_final=0.459, loss_mean=0.795, loss_mean_cls=0.0709, proj_loss=-0.407][2026-03-26 13:02:41] Step: 7623, Training Logs: loss_final: 0.465506, loss_mean: 0.820594, proj_loss: -0.422518, loss_mean_cls: 0.067429, grad_norm: 0.352182 +Steps: 1%| | 7624/1000000 [31:38<72:35:13, 3.80it/s, grad_norm=0.352, loss_final=0.466, loss_mean=0.821, loss_mean_cls=0.0674, proj_loss=-0.423][2026-03-26 13:02:41] Step: 7624, Training Logs: loss_final: 0.476251, loss_mean: 0.822169, proj_loss: -0.413641, loss_mean_cls: 0.067723, grad_norm: 0.450803 +Steps: 1%| | 7625/1000000 [31:38<74:33:39, 3.70it/s, grad_norm=0.451, loss_final=0.476, loss_mean=0.822, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 13:02:42] Step: 7625, Training Logs: loss_final: 0.461743, loss_mean: 0.806090, proj_loss: -0.413384, loss_mean_cls: 0.069037, grad_norm: 0.386273 +Steps: 1%| | 7626/1000000 [31:38<72:40:16, 3.79it/s, grad_norm=0.386, loss_final=0.462, loss_mean=0.806, loss_mean_cls=0.069, proj_loss=-0.413][2026-03-26 13:02:42] Step: 7626, Training Logs: loss_final: 0.474699, loss_mean: 0.814199, proj_loss: -0.408812, loss_mean_cls: 0.069312, grad_norm: 0.358335 +Steps: 1%| | 7627/1000000 [31:39<71:22:16, 3.86it/s, grad_norm=0.358, loss_final=0.475, loss_mean=0.814, loss_mean_cls=0.0693, proj_loss=-0.409][2026-03-26 13:02:42] Step: 7627, Training Logs: loss_final: 0.459953, loss_mean: 0.817253, proj_loss: -0.423723, loss_mean_cls: 0.066424, grad_norm: 0.255602 +Steps: 1%| | 7628/1000000 [31:39<70:25:56, 3.91it/s, grad_norm=0.256, loss_final=0.46, loss_mean=0.817, loss_mean_cls=0.0664, proj_loss=-0.424][2026-03-26 13:02:42] Step: 7628, Training Logs: loss_final: 0.477628, loss_mean: 0.830174, proj_loss: -0.418567, loss_mean_cls: 0.066021, grad_norm: 0.529944 +Steps: 1%| | 7629/1000000 [31:39<69:47:34, 3.95it/s, grad_norm=0.53, loss_final=0.478, loss_mean=0.83, loss_mean_cls=0.066, proj_loss=-0.419][2026-03-26 13:02:43] Step: 7629, Training Logs: loss_final: 0.469125, loss_mean: 0.829310, proj_loss: -0.424910, loss_mean_cls: 0.064726, grad_norm: 0.348361 +Steps: 1%| | 7630/1000000 [31:39<69:19:19, 3.98it/s, grad_norm=0.348, loss_final=0.469, loss_mean=0.829, loss_mean_cls=0.0647, proj_loss=-0.425][2026-03-26 13:02:43] Step: 7630, Training Logs: loss_final: 0.479567, loss_mean: 0.829331, proj_loss: -0.417245, loss_mean_cls: 0.067482, grad_norm: 0.400065 +Steps: 1%| | 7631/1000000 [31:40<69:00:10, 3.99it/s, grad_norm=0.4, loss_final=0.48, loss_mean=0.829, loss_mean_cls=0.0675, proj_loss=-0.417][2026-03-26 13:02:43] Step: 7631, Training Logs: loss_final: 0.455094, loss_mean: 0.796387, proj_loss: -0.410974, loss_mean_cls: 0.069681, grad_norm: 0.390166 +Steps: 1%| | 7632/1000000 [31:40<68:45:32, 4.01it/s, grad_norm=0.39, loss_final=0.455, loss_mean=0.796, loss_mean_cls=0.0697, proj_loss=-0.411][2026-03-26 13:02:43] Step: 7632, Training Logs: loss_final: 0.463414, loss_mean: 0.818962, proj_loss: -0.421813, loss_mean_cls: 0.066264, grad_norm: 0.203174 +Steps: 1%| | 7633/1000000 [31:40<68:36:01, 4.02it/s, grad_norm=0.203, loss_final=0.463, loss_mean=0.819, loss_mean_cls=0.0663, proj_loss=-0.422][2026-03-26 13:02:44] Step: 7633, Training Logs: loss_final: 0.461927, loss_mean: 0.808296, proj_loss: -0.414681, loss_mean_cls: 0.068312, grad_norm: 0.374613 +Steps: 1%| | 7634/1000000 [31:40<68:28:48, 4.03it/s, grad_norm=0.375, loss_final=0.462, loss_mean=0.808, loss_mean_cls=0.0683, proj_loss=-0.415][2026-03-26 13:02:44] Step: 7634, Training Logs: loss_final: 0.472986, loss_mean: 0.831833, proj_loss: -0.424559, loss_mean_cls: 0.065712, grad_norm: 0.250500 +Steps: 1%| | 7635/1000000 [31:41<68:23:46, 4.03it/s, grad_norm=0.251, loss_final=0.473, loss_mean=0.832, loss_mean_cls=0.0657, proj_loss=-0.425][2026-03-26 13:02:44] Step: 7635, Training Logs: loss_final: 0.466171, loss_mean: 0.806920, proj_loss: -0.409608, loss_mean_cls: 0.068858, grad_norm: 0.407038 +Steps: 1%| | 7636/1000000 [31:41<68:21:24, 4.03it/s, grad_norm=0.407, loss_final=0.466, loss_mean=0.807, loss_mean_cls=0.0689, proj_loss=-0.41][2026-03-26 13:02:44] Step: 7636, Training Logs: loss_final: 0.460183, loss_mean: 0.809217, proj_loss: -0.417044, loss_mean_cls: 0.068010, grad_norm: 0.391910 +Steps: 1%| | 7637/1000000 [31:41<68:18:10, 4.04it/s, grad_norm=0.392, loss_final=0.46, loss_mean=0.809, loss_mean_cls=0.068, proj_loss=-0.417][2026-03-26 13:02:45] Step: 7637, Training Logs: loss_final: 0.453429, loss_mean: 0.814477, proj_loss: -0.425787, loss_mean_cls: 0.064740, grad_norm: 0.457163 +Steps: 1%| | 7638/1000000 [31:41<68:18:20, 4.04it/s, grad_norm=0.457, loss_final=0.453, loss_mean=0.814, loss_mean_cls=0.0647, proj_loss=-0.426][2026-03-26 13:02:45] Step: 7638, Training Logs: loss_final: 0.458065, loss_mean: 0.817795, proj_loss: -0.425545, loss_mean_cls: 0.065814, grad_norm: 0.432863 +Steps: 1%| | 7639/1000000 [31:42<68:14:43, 4.04it/s, grad_norm=0.433, loss_final=0.458, loss_mean=0.818, loss_mean_cls=0.0658, proj_loss=-0.426][2026-03-26 13:02:45] Step: 7639, Training Logs: loss_final: 0.453074, loss_mean: 0.797999, proj_loss: -0.413828, loss_mean_cls: 0.068903, grad_norm: 0.274358 +Steps: 1%| | 7640/1000000 [31:42<68:14:20, 4.04it/s, grad_norm=0.274, loss_final=0.453, loss_mean=0.798, loss_mean_cls=0.0689, proj_loss=-0.414][2026-03-26 13:02:45] Step: 7640, Training Logs: loss_final: 0.474439, loss_mean: 0.833844, proj_loss: -0.424467, loss_mean_cls: 0.065063, grad_norm: 0.397550 +Steps: 1%| | 7641/1000000 [31:42<68:12:05, 4.04it/s, grad_norm=0.398, loss_final=0.474, loss_mean=0.834, loss_mean_cls=0.0651, proj_loss=-0.424][2026-03-26 13:02:46] Step: 7641, Training Logs: loss_final: 0.478871, loss_mean: 0.834996, proj_loss: -0.421416, loss_mean_cls: 0.065291, grad_norm: 0.349191 +Steps: 1%| | 7642/1000000 [31:42<68:14:03, 4.04it/s, grad_norm=0.349, loss_final=0.479, loss_mean=0.835, loss_mean_cls=0.0653, proj_loss=-0.421][2026-03-26 13:02:46] Step: 7642, Training Logs: loss_final: 0.476226, loss_mean: 0.823258, proj_loss: -0.415156, loss_mean_cls: 0.068124, grad_norm: 0.344226 +Steps: 1%| | 7643/1000000 [31:43<68:11:32, 4.04it/s, grad_norm=0.344, loss_final=0.476, loss_mean=0.823, loss_mean_cls=0.0681, proj_loss=-0.415][2026-03-26 13:02:46] Step: 7643, Training Logs: loss_final: 0.451616, loss_mean: 0.815693, proj_loss: -0.428554, loss_mean_cls: 0.064477, grad_norm: 0.337339 +Steps: 1%| | 7644/1000000 [31:43<68:11:23, 4.04it/s, grad_norm=0.337, loss_final=0.452, loss_mean=0.816, loss_mean_cls=0.0645, proj_loss=-0.429][2026-03-26 13:02:46] Step: 7644, Training Logs: loss_final: 0.461722, loss_mean: 0.820029, proj_loss: -0.424037, loss_mean_cls: 0.065730, grad_norm: 0.232600 +Steps: 1%| | 7645/1000000 [31:43<68:08:25, 4.05it/s, grad_norm=0.233, loss_final=0.462, loss_mean=0.82, loss_mean_cls=0.0657, proj_loss=-0.424][2026-03-26 13:02:47] Step: 7645, Training Logs: loss_final: 0.466097, loss_mean: 0.820873, proj_loss: -0.420701, loss_mean_cls: 0.065925, grad_norm: 0.298585 +Steps: 1%| | 7646/1000000 [31:43<68:09:55, 4.04it/s, grad_norm=0.299, loss_final=0.466, loss_mean=0.821, loss_mean_cls=0.0659, proj_loss=-0.421][2026-03-26 13:02:47] Step: 7646, Training Logs: loss_final: 0.484256, loss_mean: 0.817266, proj_loss: -0.403608, loss_mean_cls: 0.070599, grad_norm: 0.574450 +Steps: 1%| | 7647/1000000 [31:44<68:11:09, 4.04it/s, grad_norm=0.574, loss_final=0.484, loss_mean=0.817, loss_mean_cls=0.0706, proj_loss=-0.404][2026-03-26 13:02:47] Step: 7647, Training Logs: loss_final: 0.478120, loss_mean: 0.822127, proj_loss: -0.412019, loss_mean_cls: 0.068012, grad_norm: 0.327361 +Steps: 1%| | 7648/1000000 [31:44<68:15:00, 4.04it/s, grad_norm=0.327, loss_final=0.478, loss_mean=0.822, loss_mean_cls=0.068, proj_loss=-0.412][2026-03-26 13:02:47] Step: 7648, Training Logs: loss_final: 0.468834, loss_mean: 0.824122, proj_loss: -0.420654, loss_mean_cls: 0.065366, grad_norm: 0.445718 +Steps: 1%| | 7649/1000000 [31:44<68:12:59, 4.04it/s, grad_norm=0.446, loss_final=0.469, loss_mean=0.824, loss_mean_cls=0.0654, proj_loss=-0.421][2026-03-26 13:02:48] Step: 7649, Training Logs: loss_final: 0.459640, loss_mean: 0.803530, proj_loss: -0.411968, loss_mean_cls: 0.068078, grad_norm: 0.404437 +Steps: 1%| | 7650/1000000 [31:44<68:12:45, 4.04it/s, grad_norm=0.404, loss_final=0.46, loss_mean=0.804, loss_mean_cls=0.0681, proj_loss=-0.412][2026-03-26 13:02:48] Step: 7650, Training Logs: loss_final: 0.472797, loss_mean: 0.824355, proj_loss: -0.418302, loss_mean_cls: 0.066745, grad_norm: 0.370073 +Steps: 1%| | 7651/1000000 [31:45<68:12:18, 4.04it/s, grad_norm=0.37, loss_final=0.473, loss_mean=0.824, loss_mean_cls=0.0667, proj_loss=-0.418][2026-03-26 13:02:48] Step: 7651, Training Logs: loss_final: 0.466783, loss_mean: 0.807676, proj_loss: -0.409900, loss_mean_cls: 0.069007, grad_norm: 0.476280 +Steps: 1%| | 7652/1000000 [31:45<68:12:23, 4.04it/s, grad_norm=0.476, loss_final=0.467, loss_mean=0.808, loss_mean_cls=0.069, proj_loss=-0.41][2026-03-26 13:02:48] Step: 7652, Training Logs: loss_final: 0.479108, loss_mean: 0.827116, proj_loss: -0.415427, loss_mean_cls: 0.067420, grad_norm: 0.280695 +Steps: 1%| | 7653/1000000 [31:45<68:50:15, 4.00it/s, grad_norm=0.281, loss_final=0.479, loss_mean=0.827, loss_mean_cls=0.0674, proj_loss=-0.415][2026-03-26 13:02:49] Step: 7653, Training Logs: loss_final: 0.466421, loss_mean: 0.818391, proj_loss: -0.417171, loss_mean_cls: 0.065201, grad_norm: 0.580493 +Steps: 1%| | 7654/1000000 [31:45<68:40:01, 4.01it/s, grad_norm=0.58, loss_final=0.466, loss_mean=0.818, loss_mean_cls=0.0652, proj_loss=-0.417][2026-03-26 13:02:49] Step: 7654, Training Logs: loss_final: 0.469931, loss_mean: 0.822447, proj_loss: -0.420396, loss_mean_cls: 0.067880, grad_norm: 0.243640 +Steps: 1%| | 7655/1000000 [31:46<68:31:40, 4.02it/s, grad_norm=0.244, loss_final=0.47, loss_mean=0.822, loss_mean_cls=0.0679, proj_loss=-0.42][2026-03-26 13:02:49] Step: 7655, Training Logs: loss_final: 0.472705, loss_mean: 0.824221, proj_loss: -0.417650, loss_mean_cls: 0.066134, grad_norm: 0.719793 +Steps: 1%| | 7656/1000000 [31:46<68:24:42, 4.03it/s, grad_norm=0.72, loss_final=0.473, loss_mean=0.824, loss_mean_cls=0.0661, proj_loss=-0.418][2026-03-26 13:02:49] Step: 7656, Training Logs: loss_final: 0.486095, loss_mean: 0.836594, proj_loss: -0.417695, loss_mean_cls: 0.067196, grad_norm: 0.423024 +Steps: 1%| | 7657/1000000 [31:46<68:22:07, 4.03it/s, grad_norm=0.423, loss_final=0.486, loss_mean=0.837, loss_mean_cls=0.0672, proj_loss=-0.418][2026-03-26 13:02:50] Step: 7657, Training Logs: loss_final: 0.464215, loss_mean: 0.815673, proj_loss: -0.418684, loss_mean_cls: 0.067226, grad_norm: 0.513707 +Steps: 1%| | 7658/1000000 [31:46<68:19:00, 4.03it/s, grad_norm=0.514, loss_final=0.464, loss_mean=0.816, loss_mean_cls=0.0672, proj_loss=-0.419][2026-03-26 13:02:50] Step: 7658, Training Logs: loss_final: 0.457088, loss_mean: 0.800714, proj_loss: -0.412276, loss_mean_cls: 0.068650, grad_norm: 0.613671 +Steps: 1%| | 7659/1000000 [31:47<68:18:52, 4.04it/s, grad_norm=0.614, loss_final=0.457, loss_mean=0.801, loss_mean_cls=0.0687, proj_loss=-0.412][2026-03-26 13:02:50] Step: 7659, Training Logs: loss_final: 0.474946, loss_mean: 0.822527, proj_loss: -0.414715, loss_mean_cls: 0.067134, grad_norm: 0.344078 +Steps: 1%| | 7660/1000000 [31:47<68:16:24, 4.04it/s, grad_norm=0.344, loss_final=0.475, loss_mean=0.823, loss_mean_cls=0.0671, proj_loss=-0.415][2026-03-26 13:02:50] Step: 7660, Training Logs: loss_final: 0.456904, loss_mean: 0.803414, proj_loss: -0.414715, loss_mean_cls: 0.068204, grad_norm: 0.230038 +Steps: 1%| | 7661/1000000 [31:47<68:15:02, 4.04it/s, grad_norm=0.23, loss_final=0.457, loss_mean=0.803, loss_mean_cls=0.0682, proj_loss=-0.415][2026-03-26 13:02:51] Step: 7661, Training Logs: loss_final: 0.451125, loss_mean: 0.811611, proj_loss: -0.425533, loss_mean_cls: 0.065047, grad_norm: 0.208688 +Steps: 1%| | 7662/1000000 [31:47<68:13:42, 4.04it/s, grad_norm=0.209, loss_final=0.451, loss_mean=0.812, loss_mean_cls=0.065, proj_loss=-0.426][2026-03-26 13:02:51] Step: 7662, Training Logs: loss_final: 0.464902, loss_mean: 0.807281, proj_loss: -0.411006, loss_mean_cls: 0.068627, grad_norm: 0.319297 +Steps: 1%| | 7663/1000000 [31:48<68:14:31, 4.04it/s, grad_norm=0.319, loss_final=0.465, loss_mean=0.807, loss_mean_cls=0.0686, proj_loss=-0.411][2026-03-26 13:02:51] Step: 7663, Training Logs: loss_final: 0.462082, loss_mean: 0.807154, proj_loss: -0.412928, loss_mean_cls: 0.067856, grad_norm: 0.233702 +Steps: 1%| | 7664/1000000 [31:48<68:12:42, 4.04it/s, grad_norm=0.234, loss_final=0.462, loss_mean=0.807, loss_mean_cls=0.0679, proj_loss=-0.413][2026-03-26 13:02:51] Step: 7664, Training Logs: loss_final: 0.467840, loss_mean: 0.815668, proj_loss: -0.415316, loss_mean_cls: 0.067488, grad_norm: 0.372723 +Steps: 1%| | 7665/1000000 [31:48<68:12:04, 4.04it/s, grad_norm=0.373, loss_final=0.468, loss_mean=0.816, loss_mean_cls=0.0675, proj_loss=-0.415][2026-03-26 13:02:52] Step: 7665, Training Logs: loss_final: 0.470865, loss_mean: 0.820340, proj_loss: -0.417311, loss_mean_cls: 0.067837, grad_norm: 0.193944 +Steps: 1%| | 7666/1000000 [31:48<68:11:39, 4.04it/s, grad_norm=0.194, loss_final=0.471, loss_mean=0.82, loss_mean_cls=0.0678, proj_loss=-0.417][2026-03-26 13:02:52] Step: 7666, Training Logs: loss_final: 0.470663, loss_mean: 0.825715, proj_loss: -0.421209, loss_mean_cls: 0.066157, grad_norm: 0.270808 +Steps: 1%| | 7667/1000000 [31:49<68:13:48, 4.04it/s, grad_norm=0.271, loss_final=0.471, loss_mean=0.826, loss_mean_cls=0.0662, proj_loss=-0.421][2026-03-26 13:02:52] Step: 7667, Training Logs: loss_final: 0.471232, loss_mean: 0.809094, proj_loss: -0.406733, loss_mean_cls: 0.068871, grad_norm: 0.321238 +Steps: 1%| | 7668/1000000 [31:49<68:14:39, 4.04it/s, grad_norm=0.321, loss_final=0.471, loss_mean=0.809, loss_mean_cls=0.0689, proj_loss=-0.407][2026-03-26 13:02:52] Step: 7668, Training Logs: loss_final: 0.480461, loss_mean: 0.833008, proj_loss: -0.419215, loss_mean_cls: 0.066669, grad_norm: 0.274425 +Steps: 1%| | 7669/1000000 [31:49<68:13:01, 4.04it/s, grad_norm=0.274, loss_final=0.48, loss_mean=0.833, loss_mean_cls=0.0667, proj_loss=-0.419][2026-03-26 13:02:53] Step: 7669, Training Logs: loss_final: 0.465096, loss_mean: 0.822460, proj_loss: -0.423148, loss_mean_cls: 0.065785, grad_norm: 0.372480 +Steps: 1%| | 7670/1000000 [31:49<68:13:16, 4.04it/s, grad_norm=0.372, loss_final=0.465, loss_mean=0.822, loss_mean_cls=0.0658, proj_loss=-0.423][2026-03-26 13:02:53] Step: 7670, Training Logs: loss_final: 0.465715, loss_mean: 0.819247, proj_loss: -0.420729, loss_mean_cls: 0.067197, grad_norm: 0.293517 +Steps: 1%| | 7671/1000000 [31:50<68:13:09, 4.04it/s, grad_norm=0.294, loss_final=0.466, loss_mean=0.819, loss_mean_cls=0.0672, proj_loss=-0.421][2026-03-26 13:02:53] Step: 7671, Training Logs: loss_final: 0.479156, loss_mean: 0.825311, proj_loss: -0.413357, loss_mean_cls: 0.067202, grad_norm: 0.521628 +Steps: 1%| | 7672/1000000 [31:50<68:11:39, 4.04it/s, grad_norm=0.522, loss_final=0.479, loss_mean=0.825, loss_mean_cls=0.0672, proj_loss=-0.413][2026-03-26 13:02:53] Step: 7672, Training Logs: loss_final: 0.468838, loss_mean: 0.819696, proj_loss: -0.418056, loss_mean_cls: 0.067197, grad_norm: 0.284532 +Steps: 1%| | 7673/1000000 [31:50<68:11:31, 4.04it/s, grad_norm=0.285, loss_final=0.469, loss_mean=0.82, loss_mean_cls=0.0672, proj_loss=-0.418][2026-03-26 13:02:54] Step: 7673, Training Logs: loss_final: 0.461339, loss_mean: 0.800497, proj_loss: -0.409321, loss_mean_cls: 0.070162, grad_norm: 0.291077 +Steps: 1%| | 7674/1000000 [31:50<68:14:13, 4.04it/s, grad_norm=0.291, loss_final=0.461, loss_mean=0.8, loss_mean_cls=0.0702, proj_loss=-0.409][2026-03-26 13:02:54] Step: 7674, Training Logs: loss_final: 0.463275, loss_mean: 0.815755, proj_loss: -0.419649, loss_mean_cls: 0.067169, grad_norm: 0.362316 +Steps: 1%| | 7675/1000000 [31:51<68:14:39, 4.04it/s, grad_norm=0.362, loss_final=0.463, loss_mean=0.816, loss_mean_cls=0.0672, proj_loss=-0.42][2026-03-26 13:02:54] Step: 7675, Training Logs: loss_final: 0.450461, loss_mean: 0.807343, proj_loss: -0.422820, loss_mean_cls: 0.065938, grad_norm: 0.264249 +Steps: 1%| | 7676/1000000 [31:51<68:16:11, 4.04it/s, grad_norm=0.264, loss_final=0.45, loss_mean=0.807, loss_mean_cls=0.0659, proj_loss=-0.423][2026-03-26 13:02:54] Step: 7676, Training Logs: loss_final: 0.471277, loss_mean: 0.814633, proj_loss: -0.411043, loss_mean_cls: 0.067687, grad_norm: 0.317801 +Steps: 1%| | 7677/1000000 [31:51<68:14:08, 4.04it/s, grad_norm=0.318, loss_final=0.471, loss_mean=0.815, loss_mean_cls=0.0677, proj_loss=-0.411][2026-03-26 13:02:55] Step: 7677, Training Logs: loss_final: 0.466746, loss_mean: 0.818543, proj_loss: -0.418925, loss_mean_cls: 0.067129, grad_norm: 0.446065 +Steps: 1%| | 7678/1000000 [31:51<68:13:02, 4.04it/s, grad_norm=0.446, loss_final=0.467, loss_mean=0.819, loss_mean_cls=0.0671, proj_loss=-0.419][2026-03-26 13:02:55] Step: 7678, Training Logs: loss_final: 0.468185, loss_mean: 0.821794, proj_loss: -0.419660, loss_mean_cls: 0.066051, grad_norm: 0.290872 +Steps: 1%| | 7679/1000000 [31:52<68:13:37, 4.04it/s, grad_norm=0.291, loss_final=0.468, loss_mean=0.822, loss_mean_cls=0.0661, proj_loss=-0.42][2026-03-26 13:02:55] Step: 7679, Training Logs: loss_final: 0.476909, loss_mean: 0.817105, proj_loss: -0.409196, loss_mean_cls: 0.069000, grad_norm: 0.464521 +Steps: 1%| | 7680/1000000 [31:52<68:15:03, 4.04it/s, grad_norm=0.465, loss_final=0.477, loss_mean=0.817, loss_mean_cls=0.069, proj_loss=-0.409][2026-03-26 13:02:55] Step: 7680, Training Logs: loss_final: 0.461043, loss_mean: 0.799167, proj_loss: -0.407658, loss_mean_cls: 0.069534, grad_norm: 0.292505 +Steps: 1%| | 7681/1000000 [31:52<68:11:51, 4.04it/s, grad_norm=0.293, loss_final=0.461, loss_mean=0.799, loss_mean_cls=0.0695, proj_loss=-0.408][2026-03-26 13:02:55] Step: 7681, Training Logs: loss_final: 0.457137, loss_mean: 0.813781, proj_loss: -0.421753, loss_mean_cls: 0.065108, grad_norm: 0.407759 +Steps: 1%| | 7682/1000000 [31:52<68:12:34, 4.04it/s, grad_norm=0.408, loss_final=0.457, loss_mean=0.814, loss_mean_cls=0.0651, proj_loss=-0.422][2026-03-26 13:02:56] Step: 7682, Training Logs: loss_final: 0.462277, loss_mean: 0.809310, proj_loss: -0.414652, loss_mean_cls: 0.067618, grad_norm: 0.318669 +Steps: 1%| | 7683/1000000 [31:53<68:12:28, 4.04it/s, grad_norm=0.319, loss_final=0.462, loss_mean=0.809, loss_mean_cls=0.0676, proj_loss=-0.415][2026-03-26 13:02:56] Step: 7683, Training Logs: loss_final: 0.481897, loss_mean: 0.835128, proj_loss: -0.419327, loss_mean_cls: 0.066095, grad_norm: 0.369268 +Steps: 1%| | 7684/1000000 [31:53<69:52:55, 3.94it/s, grad_norm=0.369, loss_final=0.482, loss_mean=0.835, loss_mean_cls=0.0661, proj_loss=-0.419][2026-03-26 13:02:56] Step: 7684, Training Logs: loss_final: 0.478155, loss_mean: 0.831688, proj_loss: -0.421066, loss_mean_cls: 0.067533, grad_norm: 0.351556 +Steps: 1%| | 7685/1000000 [31:53<69:29:43, 3.97it/s, grad_norm=0.352, loss_final=0.478, loss_mean=0.832, loss_mean_cls=0.0675, proj_loss=-0.421][2026-03-26 13:02:57] Step: 7685, Training Logs: loss_final: 0.458178, loss_mean: 0.821716, proj_loss: -0.428069, loss_mean_cls: 0.064531, grad_norm: 0.338424 +Steps: 1%| | 7686/1000000 [31:53<69:09:34, 3.99it/s, grad_norm=0.338, loss_final=0.458, loss_mean=0.822, loss_mean_cls=0.0645, proj_loss=-0.428][2026-03-26 13:02:57] Step: 7686, Training Logs: loss_final: 0.464828, loss_mean: 0.815684, proj_loss: -0.417807, loss_mean_cls: 0.066952, grad_norm: 0.404151 +Steps: 1%| | 7687/1000000 [31:54<68:54:04, 4.00it/s, grad_norm=0.404, loss_final=0.465, loss_mean=0.816, loss_mean_cls=0.067, proj_loss=-0.418][2026-03-26 13:02:57] Step: 7687, Training Logs: loss_final: 0.461288, loss_mean: 0.813715, proj_loss: -0.418801, loss_mean_cls: 0.066375, grad_norm: 0.215612 +Steps: 1%| | 7688/1000000 [31:54<68:42:37, 4.01it/s, grad_norm=0.216, loss_final=0.461, loss_mean=0.814, loss_mean_cls=0.0664, proj_loss=-0.419][2026-03-26 13:02:57] Step: 7688, Training Logs: loss_final: 0.466258, loss_mean: 0.810858, proj_loss: -0.412136, loss_mean_cls: 0.067536, grad_norm: 0.235466 +Steps: 1%| | 7689/1000000 [31:54<68:31:44, 4.02it/s, grad_norm=0.235, loss_final=0.466, loss_mean=0.811, loss_mean_cls=0.0675, proj_loss=-0.412][2026-03-26 13:02:57] Step: 7689, Training Logs: loss_final: 0.470421, loss_mean: 0.813652, proj_loss: -0.412269, loss_mean_cls: 0.069038, grad_norm: 0.249239 +Steps: 1%| | 7690/1000000 [31:54<68:27:35, 4.03it/s, grad_norm=0.249, loss_final=0.47, loss_mean=0.814, loss_mean_cls=0.069, proj_loss=-0.412][2026-03-26 13:02:58] Step: 7690, Training Logs: loss_final: 0.454441, loss_mean: 0.809012, proj_loss: -0.420710, loss_mean_cls: 0.066139, grad_norm: 0.252736 +Steps: 1%| | 7691/1000000 [31:55<68:20:21, 4.03it/s, grad_norm=0.253, loss_final=0.454, loss_mean=0.809, loss_mean_cls=0.0661, proj_loss=-0.421][2026-03-26 13:02:58] Step: 7691, Training Logs: loss_final: 0.471880, loss_mean: 0.821610, proj_loss: -0.416447, loss_mean_cls: 0.066717, grad_norm: 0.188542 +Steps: 1%| | 7692/1000000 [31:55<68:19:30, 4.03it/s, grad_norm=0.189, loss_final=0.472, loss_mean=0.822, loss_mean_cls=0.0667, proj_loss=-0.416][2026-03-26 13:02:58] Step: 7692, Training Logs: loss_final: 0.459917, loss_mean: 0.810859, proj_loss: -0.417510, loss_mean_cls: 0.066568, grad_norm: 0.309112 +Steps: 1%| | 7693/1000000 [31:55<68:14:17, 4.04it/s, grad_norm=0.309, loss_final=0.46, loss_mean=0.811, loss_mean_cls=0.0666, proj_loss=-0.418][2026-03-26 13:02:58] Step: 7693, Training Logs: loss_final: 0.467469, loss_mean: 0.809986, proj_loss: -0.411182, loss_mean_cls: 0.068665, grad_norm: 0.235311 +Steps: 1%| | 7694/1000000 [31:55<68:14:32, 4.04it/s, grad_norm=0.235, loss_final=0.467, loss_mean=0.81, loss_mean_cls=0.0687, proj_loss=-0.411][2026-03-26 13:02:59] Step: 7694, Training Logs: loss_final: 0.463100, loss_mean: 0.818623, proj_loss: -0.421938, loss_mean_cls: 0.066416, grad_norm: 0.324703 +Steps: 1%| | 7695/1000000 [31:56<68:12:21, 4.04it/s, grad_norm=0.325, loss_final=0.463, loss_mean=0.819, loss_mean_cls=0.0664, proj_loss=-0.422][2026-03-26 13:02:59] Step: 7695, Training Logs: loss_final: 0.455414, loss_mean: 0.813779, proj_loss: -0.423744, loss_mean_cls: 0.065378, grad_norm: 0.230065 +Steps: 1%| | 7696/1000000 [31:56<68:10:20, 4.04it/s, grad_norm=0.23, loss_final=0.455, loss_mean=0.814, loss_mean_cls=0.0654, proj_loss=-0.424][2026-03-26 13:02:59] Step: 7696, Training Logs: loss_final: 0.469648, loss_mean: 0.809488, proj_loss: -0.409141, loss_mean_cls: 0.069300, grad_norm: 0.214816 +Steps: 1%| | 7697/1000000 [31:56<68:06:44, 4.05it/s, grad_norm=0.215, loss_final=0.47, loss_mean=0.809, loss_mean_cls=0.0693, proj_loss=-0.409][2026-03-26 13:02:59] Step: 7697, Training Logs: loss_final: 0.475597, loss_mean: 0.817351, proj_loss: -0.409553, loss_mean_cls: 0.067798, grad_norm: 0.388686 +Steps: 1%| | 7698/1000000 [31:56<68:09:21, 4.04it/s, grad_norm=0.389, loss_final=0.476, loss_mean=0.817, loss_mean_cls=0.0678, proj_loss=-0.41][2026-03-26 13:03:00] Step: 7698, Training Logs: loss_final: 0.448697, loss_mean: 0.797672, proj_loss: -0.416550, loss_mean_cls: 0.067576, grad_norm: 0.219880 +Steps: 1%| | 7699/1000000 [31:57<68:10:05, 4.04it/s, grad_norm=0.22, loss_final=0.449, loss_mean=0.798, loss_mean_cls=0.0676, proj_loss=-0.417][2026-03-26 13:03:00] Step: 7699, Training Logs: loss_final: 0.484352, loss_mean: 0.820073, proj_loss: -0.404902, loss_mean_cls: 0.069181, grad_norm: 0.293783 +Steps: 1%| | 7700/1000000 [31:57<68:10:07, 4.04it/s, grad_norm=0.294, loss_final=0.484, loss_mean=0.82, loss_mean_cls=0.0692, proj_loss=-0.405][2026-03-26 13:03:00] Step: 7700, Training Logs: loss_final: 0.467275, loss_mean: 0.802123, proj_loss: -0.404772, loss_mean_cls: 0.069924, grad_norm: 0.273632 +Steps: 1%| | 7701/1000000 [31:57<68:10:33, 4.04it/s, grad_norm=0.274, loss_final=0.467, loss_mean=0.802, loss_mean_cls=0.0699, proj_loss=-0.405][2026-03-26 13:03:00] Step: 7701, Training Logs: loss_final: 0.480572, loss_mean: 0.827602, proj_loss: -0.415036, loss_mean_cls: 0.068006, grad_norm: 0.272656 +Steps: 1%| | 7702/1000000 [31:57<68:10:41, 4.04it/s, grad_norm=0.273, loss_final=0.481, loss_mean=0.828, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 13:03:01] Step: 7702, Training Logs: loss_final: 0.462629, loss_mean: 0.807498, proj_loss: -0.412909, loss_mean_cls: 0.068040, grad_norm: 0.300885 +Steps: 1%| | 7703/1000000 [31:58<68:11:27, 4.04it/s, grad_norm=0.301, loss_final=0.463, loss_mean=0.807, loss_mean_cls=0.068, proj_loss=-0.413][2026-03-26 13:03:01] Step: 7703, Training Logs: loss_final: 0.480484, loss_mean: 0.821818, proj_loss: -0.411334, loss_mean_cls: 0.070000, grad_norm: 0.344852 +Steps: 1%| | 7704/1000000 [31:58<68:13:10, 4.04it/s, grad_norm=0.345, loss_final=0.48, loss_mean=0.822, loss_mean_cls=0.07, proj_loss=-0.411][2026-03-26 13:03:01] Step: 7704, Training Logs: loss_final: 0.471225, loss_mean: 0.826916, proj_loss: -0.421403, loss_mean_cls: 0.065713, grad_norm: 0.268502 +Steps: 1%| | 7705/1000000 [31:58<68:13:03, 4.04it/s, grad_norm=0.269, loss_final=0.471, loss_mean=0.827, loss_mean_cls=0.0657, proj_loss=-0.421][2026-03-26 13:03:01] Step: 7705, Training Logs: loss_final: 0.457879, loss_mean: 0.801571, proj_loss: -0.411592, loss_mean_cls: 0.067901, grad_norm: 0.378400 +Steps: 1%| | 7706/1000000 [31:58<68:14:45, 4.04it/s, grad_norm=0.378, loss_final=0.458, loss_mean=0.802, loss_mean_cls=0.0679, proj_loss=-0.412][2026-03-26 13:03:02] Step: 7706, Training Logs: loss_final: 0.477271, loss_mean: 0.836088, proj_loss: -0.424372, loss_mean_cls: 0.065554, grad_norm: 0.279900 +Steps: 1%| | 7707/1000000 [31:59<68:16:17, 4.04it/s, grad_norm=0.28, loss_final=0.477, loss_mean=0.836, loss_mean_cls=0.0656, proj_loss=-0.424][2026-03-26 13:03:02] Step: 7707, Training Logs: loss_final: 0.472511, loss_mean: 0.817200, proj_loss: -0.412022, loss_mean_cls: 0.067333, grad_norm: 0.294120 +Steps: 1%| | 7708/1000000 [31:59<68:13:56, 4.04it/s, grad_norm=0.294, loss_final=0.473, loss_mean=0.817, loss_mean_cls=0.0673, proj_loss=-0.412][2026-03-26 13:03:02] Step: 7708, Training Logs: loss_final: 0.467995, loss_mean: 0.822001, proj_loss: -0.420930, loss_mean_cls: 0.066924, grad_norm: 0.216882 +Steps: 1%| | 7709/1000000 [31:59<68:13:03, 4.04it/s, grad_norm=0.217, loss_final=0.468, loss_mean=0.822, loss_mean_cls=0.0669, proj_loss=-0.421][2026-03-26 13:03:02] Step: 7709, Training Logs: loss_final: 0.463526, loss_mean: 0.818365, proj_loss: -0.421036, loss_mean_cls: 0.066198, grad_norm: 0.571909 +Steps: 1%| | 7710/1000000 [31:59<68:10:27, 4.04it/s, grad_norm=0.572, loss_final=0.464, loss_mean=0.818, loss_mean_cls=0.0662, proj_loss=-0.421][2026-03-26 13:03:03] Step: 7710, Training Logs: loss_final: 0.469930, loss_mean: 0.816109, proj_loss: -0.413849, loss_mean_cls: 0.067670, grad_norm: 0.399921 +Steps: 1%| | 7711/1000000 [32:00<68:13:39, 4.04it/s, grad_norm=0.4, loss_final=0.47, loss_mean=0.816, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 13:03:03] Step: 7711, Training Logs: loss_final: 0.457025, loss_mean: 0.803449, proj_loss: -0.414586, loss_mean_cls: 0.068162, grad_norm: 0.329663 +Steps: 1%| | 7712/1000000 [32:00<68:12:20, 4.04it/s, grad_norm=0.33, loss_final=0.457, loss_mean=0.803, loss_mean_cls=0.0682, proj_loss=-0.415][2026-03-26 13:03:03] Step: 7712, Training Logs: loss_final: 0.473319, loss_mean: 0.823409, proj_loss: -0.416846, loss_mean_cls: 0.066756, grad_norm: 0.555964 +Steps: 1%| | 7713/1000000 [32:00<68:14:42, 4.04it/s, grad_norm=0.556, loss_final=0.473, loss_mean=0.823, loss_mean_cls=0.0668, proj_loss=-0.417][2026-03-26 13:03:03] Step: 7713, Training Logs: loss_final: 0.440278, loss_mean: 0.781667, proj_loss: -0.409960, loss_mean_cls: 0.068571, grad_norm: 0.316724 +Steps: 1%| | 7714/1000000 [32:00<68:17:54, 4.04it/s, grad_norm=0.317, loss_final=0.44, loss_mean=0.782, loss_mean_cls=0.0686, proj_loss=-0.41][2026-03-26 13:03:04] Step: 7714, Training Logs: loss_final: 0.464168, loss_mean: 0.806260, proj_loss: -0.409722, loss_mean_cls: 0.067630, grad_norm: 0.444925 +Steps: 1%| | 7715/1000000 [32:01<68:15:11, 4.04it/s, grad_norm=0.445, loss_final=0.464, loss_mean=0.806, loss_mean_cls=0.0676, proj_loss=-0.41][2026-03-26 13:03:04] Step: 7715, Training Logs: loss_final: 0.462711, loss_mean: 0.810170, proj_loss: -0.415229, loss_mean_cls: 0.067770, grad_norm: 0.350774 +Steps: 1%| | 7716/1000000 [32:01<68:13:17, 4.04it/s, grad_norm=0.351, loss_final=0.463, loss_mean=0.81, loss_mean_cls=0.0678, proj_loss=-0.415][2026-03-26 13:03:04] Step: 7716, Training Logs: loss_final: 0.474031, loss_mean: 0.829395, proj_loss: -0.421694, loss_mean_cls: 0.066331, grad_norm: 0.404441 +Steps: 1%| | 7717/1000000 [32:01<68:13:30, 4.04it/s, grad_norm=0.404, loss_final=0.474, loss_mean=0.829, loss_mean_cls=0.0663, proj_loss=-0.422][2026-03-26 13:03:04] Step: 7717, Training Logs: loss_final: 0.478139, loss_mean: 0.802564, proj_loss: -0.395816, loss_mean_cls: 0.071391, grad_norm: 0.440725 +Steps: 1%| | 7718/1000000 [32:01<68:10:48, 4.04it/s, grad_norm=0.441, loss_final=0.478, loss_mean=0.803, loss_mean_cls=0.0714, proj_loss=-0.396][2026-03-26 13:03:05] Step: 7718, Training Logs: loss_final: 0.483950, loss_mean: 0.814610, proj_loss: -0.401358, loss_mean_cls: 0.070698, grad_norm: 0.437070 +Steps: 1%| | 7719/1000000 [32:02<68:12:26, 4.04it/s, grad_norm=0.437, loss_final=0.484, loss_mean=0.815, loss_mean_cls=0.0707, proj_loss=-0.401][2026-03-26 13:03:05] Step: 7719, Training Logs: loss_final: 0.462510, loss_mean: 0.822528, proj_loss: -0.424810, loss_mean_cls: 0.064793, grad_norm: 0.543715 +Steps: 1%| | 7720/1000000 [32:02<68:18:41, 4.03it/s, grad_norm=0.544, loss_final=0.463, loss_mean=0.823, loss_mean_cls=0.0648, proj_loss=-0.425][2026-03-26 13:03:05] Step: 7720, Training Logs: loss_final: 0.463578, loss_mean: 0.818232, proj_loss: -0.420182, loss_mean_cls: 0.065527, grad_norm: 0.286143 +Steps: 1%| | 7721/1000000 [32:02<68:20:14, 4.03it/s, grad_norm=0.286, loss_final=0.464, loss_mean=0.818, loss_mean_cls=0.0655, proj_loss=-0.42][2026-03-26 13:03:05] Step: 7721, Training Logs: loss_final: 0.457060, loss_mean: 0.798402, proj_loss: -0.410918, loss_mean_cls: 0.069577, grad_norm: 0.469431 +Steps: 1%| | 7722/1000000 [32:02<68:23:55, 4.03it/s, grad_norm=0.469, loss_final=0.457, loss_mean=0.798, loss_mean_cls=0.0696, proj_loss=-0.411][2026-03-26 13:03:06] Step: 7722, Training Logs: loss_final: 0.463419, loss_mean: 0.817730, proj_loss: -0.421002, loss_mean_cls: 0.066690, grad_norm: 0.212557 +Steps: 1%| | 7723/1000000 [32:03<68:29:01, 4.02it/s, grad_norm=0.213, loss_final=0.463, loss_mean=0.818, loss_mean_cls=0.0667, proj_loss=-0.421][2026-03-26 13:03:06] Step: 7723, Training Logs: loss_final: 0.480596, loss_mean: 0.827350, proj_loss: -0.414198, loss_mean_cls: 0.067443, grad_norm: 0.464792 +Steps: 1%| | 7724/1000000 [32:03<68:27:24, 4.03it/s, grad_norm=0.465, loss_final=0.481, loss_mean=0.827, loss_mean_cls=0.0674, proj_loss=-0.414][2026-03-26 13:03:06] Step: 7724, Training Logs: loss_final: 0.475346, loss_mean: 0.818009, proj_loss: -0.412279, loss_mean_cls: 0.069615, grad_norm: 0.268322 +Steps: 1%| | 7725/1000000 [32:03<68:25:48, 4.03it/s, grad_norm=0.268, loss_final=0.475, loss_mean=0.818, loss_mean_cls=0.0696, proj_loss=-0.412][2026-03-26 13:03:06] Step: 7725, Training Logs: loss_final: 0.478276, loss_mean: 0.839395, proj_loss: -0.425693, loss_mean_cls: 0.064574, grad_norm: 0.455405 +Steps: 1%| | 7726/1000000 [32:03<68:23:59, 4.03it/s, grad_norm=0.455, loss_final=0.478, loss_mean=0.839, loss_mean_cls=0.0646, proj_loss=-0.426][2026-03-26 13:03:07] Step: 7726, Training Logs: loss_final: 0.481730, loss_mean: 0.840348, proj_loss: -0.423947, loss_mean_cls: 0.065329, grad_norm: 0.528915 +Steps: 1%| | 7727/1000000 [32:04<68:25:05, 4.03it/s, grad_norm=0.529, loss_final=0.482, loss_mean=0.84, loss_mean_cls=0.0653, proj_loss=-0.424][2026-03-26 13:03:07] Step: 7727, Training Logs: loss_final: 0.471069, loss_mean: 0.811795, proj_loss: -0.410079, loss_mean_cls: 0.069353, grad_norm: 0.340476 +Steps: 1%| | 7728/1000000 [32:04<68:28:42, 4.03it/s, grad_norm=0.34, loss_final=0.471, loss_mean=0.812, loss_mean_cls=0.0694, proj_loss=-0.41][2026-03-26 13:03:07] Step: 7728, Training Logs: loss_final: 0.467816, loss_mean: 0.828705, proj_loss: -0.425912, loss_mean_cls: 0.065023, grad_norm: 0.263669 +Steps: 1%| | 7729/1000000 [32:04<68:28:37, 4.03it/s, grad_norm=0.264, loss_final=0.468, loss_mean=0.829, loss_mean_cls=0.065, proj_loss=-0.426][2026-03-26 13:03:07] Step: 7729, Training Logs: loss_final: 0.464870, loss_mean: 0.809742, proj_loss: -0.413339, loss_mean_cls: 0.068468, grad_norm: 0.345706 +Steps: 1%| | 7730/1000000 [32:04<68:30:37, 4.02it/s, grad_norm=0.346, loss_final=0.465, loss_mean=0.81, loss_mean_cls=0.0685, proj_loss=-0.413][2026-03-26 13:03:08] Step: 7730, Training Logs: loss_final: 0.464328, loss_mean: 0.815480, proj_loss: -0.417996, loss_mean_cls: 0.066844, grad_norm: 0.338926 +Steps: 1%| | 7731/1000000 [32:05<68:28:59, 4.02it/s, grad_norm=0.339, loss_final=0.464, loss_mean=0.815, loss_mean_cls=0.0668, proj_loss=-0.418][2026-03-26 13:03:08] Step: 7731, Training Logs: loss_final: 0.484448, loss_mean: 0.810197, proj_loss: -0.397354, loss_mean_cls: 0.071606, grad_norm: 0.298739 +Steps: 1%| | 7732/1000000 [32:05<68:26:55, 4.03it/s, grad_norm=0.299, loss_final=0.484, loss_mean=0.81, loss_mean_cls=0.0716, proj_loss=-0.397][2026-03-26 13:03:08] Step: 7732, Training Logs: loss_final: 0.473222, loss_mean: 0.823983, proj_loss: -0.417745, loss_mean_cls: 0.066984, grad_norm: 0.266629 +Steps: 1%| | 7733/1000000 [32:05<68:21:11, 4.03it/s, grad_norm=0.267, loss_final=0.473, loss_mean=0.824, loss_mean_cls=0.067, proj_loss=-0.418][2026-03-26 13:03:08] Step: 7733, Training Logs: loss_final: 0.456829, loss_mean: 0.809940, proj_loss: -0.419721, loss_mean_cls: 0.066610, grad_norm: 0.275803 +Steps: 1%| | 7734/1000000 [32:05<68:25:46, 4.03it/s, grad_norm=0.276, loss_final=0.457, loss_mean=0.81, loss_mean_cls=0.0666, proj_loss=-0.42][2026-03-26 13:03:09] Step: 7734, Training Logs: loss_final: 0.463948, loss_mean: 0.821125, proj_loss: -0.422295, loss_mean_cls: 0.065117, grad_norm: 0.246415 +Steps: 1%| | 7735/1000000 [32:06<68:24:32, 4.03it/s, grad_norm=0.246, loss_final=0.464, loss_mean=0.821, loss_mean_cls=0.0651, proj_loss=-0.422][2026-03-26 13:03:09] Step: 7735, Training Logs: loss_final: 0.456994, loss_mean: 0.812163, proj_loss: -0.422256, loss_mean_cls: 0.067087, grad_norm: 0.184782 +Steps: 1%| | 7736/1000000 [32:06<68:30:15, 4.02it/s, grad_norm=0.185, loss_final=0.457, loss_mean=0.812, loss_mean_cls=0.0671, proj_loss=-0.422][2026-03-26 13:03:09] Step: 7736, Training Logs: loss_final: 0.458694, loss_mean: 0.811257, proj_loss: -0.419196, loss_mean_cls: 0.066632, grad_norm: 0.373420 +Steps: 1%| | 7737/1000000 [32:06<68:30:16, 4.02it/s, grad_norm=0.373, loss_final=0.459, loss_mean=0.811, loss_mean_cls=0.0666, proj_loss=-0.419][2026-03-26 13:03:09] Step: 7737, Training Logs: loss_final: 0.461646, loss_mean: 0.808015, proj_loss: -0.416124, loss_mean_cls: 0.069756, grad_norm: 0.335720 +Steps: 1%| | 7738/1000000 [32:06<69:47:33, 3.95it/s, grad_norm=0.336, loss_final=0.462, loss_mean=0.808, loss_mean_cls=0.0698, proj_loss=-0.416][2026-03-26 13:03:10] Step: 7738, Training Logs: loss_final: 0.470375, loss_mean: 0.822354, proj_loss: -0.417538, loss_mean_cls: 0.065558, grad_norm: 0.261629 +Steps: 1%| | 7739/1000000 [32:07<69:21:07, 3.97it/s, grad_norm=0.262, loss_final=0.47, loss_mean=0.822, loss_mean_cls=0.0656, proj_loss=-0.418][2026-03-26 13:03:10] Step: 7739, Training Logs: loss_final: 0.495362, loss_mean: 0.849826, proj_loss: -0.420252, loss_mean_cls: 0.065788, grad_norm: 0.400043 +Steps: 1%| | 7740/1000000 [32:07<69:04:22, 3.99it/s, grad_norm=0.4, loss_final=0.495, loss_mean=0.85, loss_mean_cls=0.0658, proj_loss=-0.42][2026-03-26 13:03:10] Step: 7740, Training Logs: loss_final: 0.464934, loss_mean: 0.809279, proj_loss: -0.413008, loss_mean_cls: 0.068663, grad_norm: 0.333893 +Steps: 1%| | 7741/1000000 [32:07<68:52:23, 4.00it/s, grad_norm=0.334, loss_final=0.465, loss_mean=0.809, loss_mean_cls=0.0687, proj_loss=-0.413][2026-03-26 13:03:10] Step: 7741, Training Logs: loss_final: 0.480122, loss_mean: 0.828267, proj_loss: -0.415131, loss_mean_cls: 0.066986, grad_norm: 0.502528 +Steps: 1%| | 7742/1000000 [32:07<68:44:15, 4.01it/s, grad_norm=0.503, loss_final=0.48, loss_mean=0.828, loss_mean_cls=0.067, proj_loss=-0.415][2026-03-26 13:03:11] Step: 7742, Training Logs: loss_final: 0.474086, loss_mean: 0.814791, proj_loss: -0.409565, loss_mean_cls: 0.068859, grad_norm: 0.500514 +Steps: 1%| | 7743/1000000 [32:08<68:40:24, 4.01it/s, grad_norm=0.501, loss_final=0.474, loss_mean=0.815, loss_mean_cls=0.0689, proj_loss=-0.41][2026-03-26 13:03:11] Step: 7743, Training Logs: loss_final: 0.459557, loss_mean: 0.802232, proj_loss: -0.410793, loss_mean_cls: 0.068118, grad_norm: 0.301264 +Steps: 1%| | 7744/1000000 [32:08<68:39:29, 4.01it/s, grad_norm=0.301, loss_final=0.46, loss_mean=0.802, loss_mean_cls=0.0681, proj_loss=-0.411][2026-03-26 13:03:11] Step: 7744, Training Logs: loss_final: 0.468148, loss_mean: 0.819197, proj_loss: -0.417879, loss_mean_cls: 0.066829, grad_norm: 0.574177 +Steps: 1%| | 7745/1000000 [32:08<68:36:04, 4.02it/s, grad_norm=0.574, loss_final=0.468, loss_mean=0.819, loss_mean_cls=0.0668, proj_loss=-0.418][2026-03-26 13:03:11] Step: 7745, Training Logs: loss_final: 0.461930, loss_mean: 0.812379, proj_loss: -0.418183, loss_mean_cls: 0.067734, grad_norm: 0.328745 +Steps: 1%| | 7746/1000000 [32:08<68:37:38, 4.02it/s, grad_norm=0.329, loss_final=0.462, loss_mean=0.812, loss_mean_cls=0.0677, proj_loss=-0.418][2026-03-26 13:03:12] Step: 7746, Training Logs: loss_final: 0.475384, loss_mean: 0.823747, proj_loss: -0.415619, loss_mean_cls: 0.067256, grad_norm: 0.303836 +Steps: 1%| | 7747/1000000 [32:09<68:33:45, 4.02it/s, grad_norm=0.304, loss_final=0.475, loss_mean=0.824, loss_mean_cls=0.0673, proj_loss=-0.416][2026-03-26 13:03:12] Step: 7747, Training Logs: loss_final: 0.465679, loss_mean: 0.817434, proj_loss: -0.418138, loss_mean_cls: 0.066383, grad_norm: 0.422231 +Steps: 1%| | 7748/1000000 [32:09<68:33:45, 4.02it/s, grad_norm=0.422, loss_final=0.466, loss_mean=0.817, loss_mean_cls=0.0664, proj_loss=-0.418][2026-03-26 13:03:12] Step: 7748, Training Logs: loss_final: 0.474730, loss_mean: 0.817615, proj_loss: -0.411211, loss_mean_cls: 0.068326, grad_norm: 0.490978 +Steps: 1%| | 7749/1000000 [32:09<68:27:59, 4.03it/s, grad_norm=0.491, loss_final=0.475, loss_mean=0.818, loss_mean_cls=0.0683, proj_loss=-0.411][2026-03-26 13:03:12] Step: 7749, Training Logs: loss_final: 0.456804, loss_mean: 0.813801, proj_loss: -0.423502, loss_mean_cls: 0.066506, grad_norm: 0.244355 +Steps: 1%| | 7750/1000000 [32:09<68:29:56, 4.02it/s, grad_norm=0.244, loss_final=0.457, loss_mean=0.814, loss_mean_cls=0.0665, proj_loss=-0.424][2026-03-26 13:03:13] Step: 7750, Training Logs: loss_final: 0.468375, loss_mean: 0.822502, proj_loss: -0.418609, loss_mean_cls: 0.064481, grad_norm: 0.676849 +Steps: 1%| | 7751/1000000 [32:09<68:28:01, 4.03it/s, grad_norm=0.677, loss_final=0.468, loss_mean=0.823, loss_mean_cls=0.0645, proj_loss=-0.419][2026-03-26 13:03:13] Step: 7751, Training Logs: loss_final: 0.460579, loss_mean: 0.816313, proj_loss: -0.421573, loss_mean_cls: 0.065838, grad_norm: 0.584475 +Steps: 1%| | 7752/1000000 [32:10<68:28:47, 4.02it/s, grad_norm=0.584, loss_final=0.461, loss_mean=0.816, loss_mean_cls=0.0658, proj_loss=-0.422][2026-03-26 13:03:13] Step: 7752, Training Logs: loss_final: 0.453419, loss_mean: 0.806234, proj_loss: -0.420152, loss_mean_cls: 0.067337, grad_norm: 0.214174 +Steps: 1%| | 7753/1000000 [32:10<68:25:47, 4.03it/s, grad_norm=0.214, loss_final=0.453, loss_mean=0.806, loss_mean_cls=0.0673, proj_loss=-0.42][2026-03-26 13:03:13] Step: 7753, Training Logs: loss_final: 0.473997, loss_mean: 0.826520, proj_loss: -0.419948, loss_mean_cls: 0.067425, grad_norm: 0.359231 +Steps: 1%| | 7754/1000000 [32:10<68:26:18, 4.03it/s, grad_norm=0.359, loss_final=0.474, loss_mean=0.827, loss_mean_cls=0.0674, proj_loss=-0.42][2026-03-26 13:03:14] Step: 7754, Training Logs: loss_final: 0.465125, loss_mean: 0.822402, proj_loss: -0.422991, loss_mean_cls: 0.065714, grad_norm: 0.230293 +Steps: 1%| | 7755/1000000 [32:10<68:23:51, 4.03it/s, grad_norm=0.23, loss_final=0.465, loss_mean=0.822, loss_mean_cls=0.0657, proj_loss=-0.423][2026-03-26 13:03:14] Step: 7755, Training Logs: loss_final: 0.453444, loss_mean: 0.800034, proj_loss: -0.414601, loss_mean_cls: 0.068011, grad_norm: 0.305806 +Steps: 1%| | 7756/1000000 [32:11<68:25:38, 4.03it/s, grad_norm=0.306, loss_final=0.453, loss_mean=0.8, loss_mean_cls=0.068, proj_loss=-0.415][2026-03-26 13:03:14] Step: 7756, Training Logs: loss_final: 0.460382, loss_mean: 0.812400, proj_loss: -0.418729, loss_mean_cls: 0.066711, grad_norm: 0.308458 +Steps: 1%| | 7757/1000000 [32:11<68:30:09, 4.02it/s, grad_norm=0.308, loss_final=0.46, loss_mean=0.812, loss_mean_cls=0.0667, proj_loss=-0.419][2026-03-26 13:03:14] Step: 7757, Training Logs: loss_final: 0.482840, loss_mean: 0.838681, proj_loss: -0.422105, loss_mean_cls: 0.066265, grad_norm: 0.257134 +Steps: 1%| | 7758/1000000 [32:11<68:32:14, 4.02it/s, grad_norm=0.257, loss_final=0.483, loss_mean=0.839, loss_mean_cls=0.0663, proj_loss=-0.422][2026-03-26 13:03:15] Step: 7758, Training Logs: loss_final: 0.460000, loss_mean: 0.812343, proj_loss: -0.418762, loss_mean_cls: 0.066419, grad_norm: 0.254917 +Steps: 1%| | 7759/1000000 [32:11<68:34:33, 4.02it/s, grad_norm=0.255, loss_final=0.46, loss_mean=0.812, loss_mean_cls=0.0664, proj_loss=-0.419][2026-03-26 13:03:15] Step: 7759, Training Logs: loss_final: 0.476940, loss_mean: 0.829498, proj_loss: -0.418469, loss_mean_cls: 0.065912, grad_norm: 0.335100 +Steps: 1%| | 7760/1000000 [32:12<68:38:55, 4.01it/s, grad_norm=0.335, loss_final=0.477, loss_mean=0.829, loss_mean_cls=0.0659, proj_loss=-0.418][2026-03-26 13:03:15] Step: 7760, Training Logs: loss_final: 0.474776, loss_mean: 0.820155, proj_loss: -0.412935, loss_mean_cls: 0.067557, grad_norm: 0.249097 +Steps: 1%| | 7761/1000000 [32:12<68:32:35, 4.02it/s, grad_norm=0.249, loss_final=0.475, loss_mean=0.82, loss_mean_cls=0.0676, proj_loss=-0.413][2026-03-26 13:03:15] Step: 7761, Training Logs: loss_final: 0.459141, loss_mean: 0.815479, proj_loss: -0.422824, loss_mean_cls: 0.066487, grad_norm: 0.342259 +Steps: 1%| | 7762/1000000 [32:12<68:31:48, 4.02it/s, grad_norm=0.342, loss_final=0.459, loss_mean=0.815, loss_mean_cls=0.0665, proj_loss=-0.423][2026-03-26 13:03:16] Step: 7762, Training Logs: loss_final: 0.455575, loss_mean: 0.802040, proj_loss: -0.414417, loss_mean_cls: 0.067951, grad_norm: 0.326046 +Steps: 1%| | 7763/1000000 [32:12<68:33:31, 4.02it/s, grad_norm=0.326, loss_final=0.456, loss_mean=0.802, loss_mean_cls=0.068, proj_loss=-0.414][2026-03-26 13:03:16] Step: 7763, Training Logs: loss_final: 0.452722, loss_mean: 0.809386, proj_loss: -0.422224, loss_mean_cls: 0.065560, grad_norm: 0.245584 +Steps: 1%| | 7764/1000000 [32:13<68:32:27, 4.02it/s, grad_norm=0.246, loss_final=0.453, loss_mean=0.809, loss_mean_cls=0.0656, proj_loss=-0.422][2026-03-26 13:03:16] Step: 7764, Training Logs: loss_final: 0.474877, loss_mean: 0.824943, proj_loss: -0.417347, loss_mean_cls: 0.067281, grad_norm: 0.338570 +Steps: 1%| | 7765/1000000 [32:13<68:34:13, 4.02it/s, grad_norm=0.339, loss_final=0.475, loss_mean=0.825, loss_mean_cls=0.0673, proj_loss=-0.417][2026-03-26 13:03:16] Step: 7765, Training Logs: loss_final: 0.468772, loss_mean: 0.818864, proj_loss: -0.416502, loss_mean_cls: 0.066410, grad_norm: 0.225851 +Steps: 1%| | 7766/1000000 [32:13<68:32:17, 4.02it/s, grad_norm=0.226, loss_final=0.469, loss_mean=0.819, loss_mean_cls=0.0664, proj_loss=-0.417][2026-03-26 13:03:17] Step: 7766, Training Logs: loss_final: 0.482907, loss_mean: 0.843026, proj_loss: -0.424421, loss_mean_cls: 0.064302, grad_norm: 0.362712 +Steps: 1%| | 7767/1000000 [32:13<68:31:33, 4.02it/s, grad_norm=0.363, loss_final=0.483, loss_mean=0.843, loss_mean_cls=0.0643, proj_loss=-0.424][2026-03-26 13:03:17] Step: 7767, Training Logs: loss_final: 0.482243, loss_mean: 0.842503, proj_loss: -0.424939, loss_mean_cls: 0.064678, grad_norm: 0.237826 +Steps: 1%| | 7768/1000000 [32:14<68:28:23, 4.03it/s, grad_norm=0.238, loss_final=0.482, loss_mean=0.843, loss_mean_cls=0.0647, proj_loss=-0.425][2026-03-26 13:03:17] Step: 7768, Training Logs: loss_final: 0.455041, loss_mean: 0.792019, proj_loss: -0.407060, loss_mean_cls: 0.070082, grad_norm: 0.246109 +Steps: 1%| | 7769/1000000 [32:14<68:29:03, 4.02it/s, grad_norm=0.246, loss_final=0.455, loss_mean=0.792, loss_mean_cls=0.0701, proj_loss=-0.407][2026-03-26 13:03:17] Step: 7769, Training Logs: loss_final: 0.470850, loss_mean: 0.816363, proj_loss: -0.412557, loss_mean_cls: 0.067044, grad_norm: 0.395738 +Steps: 1%| | 7770/1000000 [32:14<68:28:45, 4.02it/s, grad_norm=0.396, loss_final=0.471, loss_mean=0.816, loss_mean_cls=0.067, proj_loss=-0.413][2026-03-26 13:03:18] Step: 7770, Training Logs: loss_final: 0.461694, loss_mean: 0.805552, proj_loss: -0.412677, loss_mean_cls: 0.068820, grad_norm: 0.274594 +Steps: 1%| | 7771/1000000 [32:14<68:29:16, 4.02it/s, grad_norm=0.275, loss_final=0.462, loss_mean=0.806, loss_mean_cls=0.0688, proj_loss=-0.413][2026-03-26 13:03:18] Step: 7771, Training Logs: loss_final: 0.471883, loss_mean: 0.819356, proj_loss: -0.414846, loss_mean_cls: 0.067374, grad_norm: 0.317791 +Steps: 1%| | 7772/1000000 [32:15<68:31:24, 4.02it/s, grad_norm=0.318, loss_final=0.472, loss_mean=0.819, loss_mean_cls=0.0674, proj_loss=-0.415][2026-03-26 13:03:18] Step: 7772, Training Logs: loss_final: 0.455303, loss_mean: 0.808466, proj_loss: -0.419476, loss_mean_cls: 0.066312, grad_norm: 0.391955 +Steps: 1%| | 7773/1000000 [32:15<68:28:59, 4.02it/s, grad_norm=0.392, loss_final=0.455, loss_mean=0.808, loss_mean_cls=0.0663, proj_loss=-0.419][2026-03-26 13:03:18] Step: 7773, Training Logs: loss_final: 0.468828, loss_mean: 0.821876, proj_loss: -0.419868, loss_mean_cls: 0.066819, grad_norm: 0.221423 +Steps: 1%| | 7774/1000000 [32:15<68:28:18, 4.03it/s, grad_norm=0.221, loss_final=0.469, loss_mean=0.822, loss_mean_cls=0.0668, proj_loss=-0.42][2026-03-26 13:03:19] Step: 7774, Training Logs: loss_final: 0.451714, loss_mean: 0.803047, proj_loss: -0.417613, loss_mean_cls: 0.066280, grad_norm: 0.469490 +Steps: 1%| | 7775/1000000 [32:15<68:28:58, 4.02it/s, grad_norm=0.469, loss_final=0.452, loss_mean=0.803, loss_mean_cls=0.0663, proj_loss=-0.418][2026-03-26 13:03:19] Step: 7775, Training Logs: loss_final: 0.463723, loss_mean: 0.810523, proj_loss: -0.414451, loss_mean_cls: 0.067651, grad_norm: 0.453423 +Steps: 1%| | 7776/1000000 [32:16<68:26:07, 4.03it/s, grad_norm=0.453, loss_final=0.464, loss_mean=0.811, loss_mean_cls=0.0677, proj_loss=-0.414][2026-03-26 13:03:19] Step: 7776, Training Logs: loss_final: 0.460010, loss_mean: 0.810791, proj_loss: -0.416900, loss_mean_cls: 0.066119, grad_norm: 0.217196 +Steps: 1%| | 7777/1000000 [32:16<68:23:34, 4.03it/s, grad_norm=0.217, loss_final=0.46, loss_mean=0.811, loss_mean_cls=0.0661, proj_loss=-0.417][2026-03-26 13:03:19] Step: 7777, Training Logs: loss_final: 0.475792, loss_mean: 0.825122, proj_loss: -0.415916, loss_mean_cls: 0.066585, grad_norm: 0.508239 +Steps: 1%| | 7778/1000000 [32:16<68:24:56, 4.03it/s, grad_norm=0.508, loss_final=0.476, loss_mean=0.825, loss_mean_cls=0.0666, proj_loss=-0.416][2026-03-26 13:03:20] Step: 7778, Training Logs: loss_final: 0.459849, loss_mean: 0.806191, proj_loss: -0.414385, loss_mean_cls: 0.068043, grad_norm: 0.320176 +Steps: 1%| | 7779/1000000 [32:16<68:29:23, 4.02it/s, grad_norm=0.32, loss_final=0.46, loss_mean=0.806, loss_mean_cls=0.068, proj_loss=-0.414][2026-03-26 13:03:20] Step: 7779, Training Logs: loss_final: 0.453300, loss_mean: 0.794327, proj_loss: -0.410482, loss_mean_cls: 0.069455, grad_norm: 0.380067 +Steps: 1%| | 7780/1000000 [32:17<68:27:56, 4.03it/s, grad_norm=0.38, loss_final=0.453, loss_mean=0.794, loss_mean_cls=0.0695, proj_loss=-0.41][2026-03-26 13:03:20] Step: 7780, Training Logs: loss_final: 0.457948, loss_mean: 0.807671, proj_loss: -0.416355, loss_mean_cls: 0.066632, grad_norm: 0.296983 +Steps: 1%| | 7781/1000000 [32:17<68:32:56, 4.02it/s, grad_norm=0.297, loss_final=0.458, loss_mean=0.808, loss_mean_cls=0.0666, proj_loss=-0.416][2026-03-26 13:03:20] Step: 7781, Training Logs: loss_final: 0.470362, loss_mean: 0.824624, proj_loss: -0.420181, loss_mean_cls: 0.065919, grad_norm: 0.376052 +Steps: 1%| | 7782/1000000 [32:17<68:30:44, 4.02it/s, grad_norm=0.376, loss_final=0.47, loss_mean=0.825, loss_mean_cls=0.0659, proj_loss=-0.42][2026-03-26 13:03:21] Step: 7782, Training Logs: loss_final: 0.455522, loss_mean: 0.812013, proj_loss: -0.422408, loss_mean_cls: 0.065916, grad_norm: 0.357397 +Steps: 1%| | 7783/1000000 [32:17<68:29:40, 4.02it/s, grad_norm=0.357, loss_final=0.456, loss_mean=0.812, loss_mean_cls=0.0659, proj_loss=-0.422][2026-03-26 13:03:21] Step: 7783, Training Logs: loss_final: 0.484892, loss_mean: 0.836103, proj_loss: -0.417795, loss_mean_cls: 0.066584, grad_norm: 0.373109 +Steps: 1%| | 7784/1000000 [32:18<68:26:39, 4.03it/s, grad_norm=0.373, loss_final=0.485, loss_mean=0.836, loss_mean_cls=0.0666, proj_loss=-0.418][2026-03-26 13:03:21] Step: 7784, Training Logs: loss_final: 0.457338, loss_mean: 0.810992, proj_loss: -0.420503, loss_mean_cls: 0.066849, grad_norm: 0.392394 +Steps: 1%| | 7785/1000000 [32:18<68:27:07, 4.03it/s, grad_norm=0.392, loss_final=0.457, loss_mean=0.811, loss_mean_cls=0.0668, proj_loss=-0.421][2026-03-26 13:03:21] Step: 7785, Training Logs: loss_final: 0.475585, loss_mean: 0.811216, proj_loss: -0.405252, loss_mean_cls: 0.069621, grad_norm: 0.339426 +Steps: 1%| | 7786/1000000 [32:18<68:29:58, 4.02it/s, grad_norm=0.339, loss_final=0.476, loss_mean=0.811, loss_mean_cls=0.0696, proj_loss=-0.405][2026-03-26 13:03:22] Step: 7786, Training Logs: loss_final: 0.455602, loss_mean: 0.802977, proj_loss: -0.414956, loss_mean_cls: 0.067581, grad_norm: 0.481424 +Steps: 1%| | 7787/1000000 [32:18<68:28:57, 4.02it/s, grad_norm=0.481, loss_final=0.456, loss_mean=0.803, loss_mean_cls=0.0676, proj_loss=-0.415][2026-03-26 13:03:22] Step: 7787, Training Logs: loss_final: 0.464133, loss_mean: 0.826667, proj_loss: -0.426709, loss_mean_cls: 0.064175, grad_norm: 0.244952 +Steps: 1%| | 7788/1000000 [32:19<68:29:30, 4.02it/s, grad_norm=0.245, loss_final=0.464, loss_mean=0.827, loss_mean_cls=0.0642, proj_loss=-0.427][2026-03-26 13:03:22] Step: 7788, Training Logs: loss_final: 0.460658, loss_mean: 0.818088, proj_loss: -0.423096, loss_mean_cls: 0.065666, grad_norm: 0.719423 +Steps: 1%| | 7789/1000000 [32:19<68:29:50, 4.02it/s, grad_norm=0.719, loss_final=0.461, loss_mean=0.818, loss_mean_cls=0.0657, proj_loss=-0.423][2026-03-26 13:03:22] Step: 7789, Training Logs: loss_final: 0.459149, loss_mean: 0.811074, proj_loss: -0.418671, loss_mean_cls: 0.066747, grad_norm: 0.462956 +Steps: 1%| | 7790/1000000 [32:19<68:27:44, 4.03it/s, grad_norm=0.463, loss_final=0.459, loss_mean=0.811, loss_mean_cls=0.0667, proj_loss=-0.419][2026-03-26 13:03:23] Step: 7790, Training Logs: loss_final: 0.470221, loss_mean: 0.821541, proj_loss: -0.417601, loss_mean_cls: 0.066280, grad_norm: 0.454475 +Steps: 1%| | 7791/1000000 [32:19<68:25:49, 4.03it/s, grad_norm=0.454, loss_final=0.47, loss_mean=0.822, loss_mean_cls=0.0663, proj_loss=-0.418][2026-03-26 13:03:23] Step: 7791, Training Logs: loss_final: 0.458644, loss_mean: 0.804284, proj_loss: -0.413153, loss_mean_cls: 0.067513, grad_norm: 0.580988 +Steps: 1%| | 7792/1000000 [32:20<68:26:08, 4.03it/s, grad_norm=0.581, loss_final=0.459, loss_mean=0.804, loss_mean_cls=0.0675, proj_loss=-0.413][2026-03-26 13:03:23] Step: 7792, Training Logs: loss_final: 0.441917, loss_mean: 0.800499, proj_loss: -0.423657, loss_mean_cls: 0.065075, grad_norm: 0.218239 +Steps: 1%| | 7793/1000000 [32:20<68:29:46, 4.02it/s, grad_norm=0.218, loss_final=0.442, loss_mean=0.8, loss_mean_cls=0.0651, proj_loss=-0.424][2026-03-26 13:03:23] Step: 7793, Training Logs: loss_final: 0.468117, loss_mean: 0.821692, proj_loss: -0.420089, loss_mean_cls: 0.066514, grad_norm: 0.618020 +Steps: 1%| | 7794/1000000 [32:20<68:26:43, 4.03it/s, grad_norm=0.618, loss_final=0.468, loss_mean=0.822, loss_mean_cls=0.0665, proj_loss=-0.42][2026-03-26 13:03:24] Step: 7794, Training Logs: loss_final: 0.454350, loss_mean: 0.805191, proj_loss: -0.418704, loss_mean_cls: 0.067862, grad_norm: 0.370690 +Steps: 1%| | 7795/1000000 [32:20<68:26:35, 4.03it/s, grad_norm=0.371, loss_final=0.454, loss_mean=0.805, loss_mean_cls=0.0679, proj_loss=-0.419][2026-03-26 13:03:24] Step: 7795, Training Logs: loss_final: 0.467014, loss_mean: 0.807548, proj_loss: -0.408776, loss_mean_cls: 0.068242, grad_norm: 0.221653 +Steps: 1%| | 7796/1000000 [32:21<68:26:29, 4.03it/s, grad_norm=0.222, loss_final=0.467, loss_mean=0.808, loss_mean_cls=0.0682, proj_loss=-0.409][2026-03-26 13:03:24] Step: 7796, Training Logs: loss_final: 0.465092, loss_mean: 0.820069, proj_loss: -0.421386, loss_mean_cls: 0.066408, grad_norm: 0.329558 +Steps: 1%| | 7797/1000000 [32:21<68:26:05, 4.03it/s, grad_norm=0.33, loss_final=0.465, loss_mean=0.82, loss_mean_cls=0.0664, proj_loss=-0.421][2026-03-26 13:03:24] Step: 7797, Training Logs: loss_final: 0.458709, loss_mean: 0.805574, proj_loss: -0.414591, loss_mean_cls: 0.067725, grad_norm: 0.317594 +Steps: 1%| | 7798/1000000 [32:21<68:24:59, 4.03it/s, grad_norm=0.318, loss_final=0.459, loss_mean=0.806, loss_mean_cls=0.0677, proj_loss=-0.415][2026-03-26 13:03:25] Step: 7798, Training Logs: loss_final: 0.460000, loss_mean: 0.813665, proj_loss: -0.418943, loss_mean_cls: 0.065278, grad_norm: 0.202811 +Steps: 1%| | 7799/1000000 [32:21<68:25:35, 4.03it/s, grad_norm=0.203, loss_final=0.46, loss_mean=0.814, loss_mean_cls=0.0653, proj_loss=-0.419][2026-03-26 13:03:25] Step: 7799, Training Logs: loss_final: 0.466856, loss_mean: 0.824968, proj_loss: -0.424136, loss_mean_cls: 0.066024, grad_norm: 0.208057 +Steps: 1%| | 7800/1000000 [32:22<68:25:51, 4.03it/s, grad_norm=0.208, loss_final=0.467, loss_mean=0.825, loss_mean_cls=0.066, proj_loss=-0.424][2026-03-26 13:03:25] Step: 7800, Training Logs: loss_final: 0.461360, loss_mean: 0.805004, proj_loss: -0.411890, loss_mean_cls: 0.068246, grad_norm: 0.196837 +Steps: 1%| | 7801/1000000 [32:22<68:26:23, 4.03it/s, grad_norm=0.197, loss_final=0.461, loss_mean=0.805, loss_mean_cls=0.0682, proj_loss=-0.412][2026-03-26 13:03:25] Step: 7801, Training Logs: loss_final: 0.469879, loss_mean: 0.810440, proj_loss: -0.409094, loss_mean_cls: 0.068533, grad_norm: 0.234363 +Steps: 1%| | 7802/1000000 [32:22<68:28:20, 4.03it/s, grad_norm=0.234, loss_final=0.47, loss_mean=0.81, loss_mean_cls=0.0685, proj_loss=-0.409][2026-03-26 13:03:26] Step: 7802, Training Logs: loss_final: 0.480183, loss_mean: 0.829903, proj_loss: -0.417271, loss_mean_cls: 0.067551, grad_norm: 0.254053 +Steps: 1%| | 7803/1000000 [32:22<68:27:59, 4.03it/s, grad_norm=0.254, loss_final=0.48, loss_mean=0.83, loss_mean_cls=0.0676, proj_loss=-0.417][2026-03-26 13:03:26] Step: 7803, Training Logs: loss_final: 0.458055, loss_mean: 0.814027, proj_loss: -0.421353, loss_mean_cls: 0.065381, grad_norm: 0.231886 +Steps: 1%| | 7804/1000000 [32:23<68:27:15, 4.03it/s, grad_norm=0.232, loss_final=0.458, loss_mean=0.814, loss_mean_cls=0.0654, proj_loss=-0.421][2026-03-26 13:03:26] Step: 7804, Training Logs: loss_final: 0.462030, loss_mean: 0.821263, proj_loss: -0.424013, loss_mean_cls: 0.064780, grad_norm: 0.351875 +Steps: 1%| | 7805/1000000 [32:23<68:29:38, 4.02it/s, grad_norm=0.352, loss_final=0.462, loss_mean=0.821, loss_mean_cls=0.0648, proj_loss=-0.424][2026-03-26 13:03:26] Step: 7805, Training Logs: loss_final: 0.481320, loss_mean: 0.829975, proj_loss: -0.416264, loss_mean_cls: 0.067609, grad_norm: 0.205394 +Steps: 1%| | 7806/1000000 [32:23<68:31:13, 4.02it/s, grad_norm=0.205, loss_final=0.481, loss_mean=0.83, loss_mean_cls=0.0676, proj_loss=-0.416][2026-03-26 13:03:27] Step: 7806, Training Logs: loss_final: 0.481830, loss_mean: 0.844496, proj_loss: -0.427747, loss_mean_cls: 0.065082, grad_norm: 0.601300 +Steps: 1%| | 7807/1000000 [32:23<68:27:55, 4.03it/s, grad_norm=0.601, loss_final=0.482, loss_mean=0.844, loss_mean_cls=0.0651, proj_loss=-0.428][2026-03-26 13:03:27] Step: 7807, Training Logs: loss_final: 0.477077, loss_mean: 0.819623, proj_loss: -0.410530, loss_mean_cls: 0.067985, grad_norm: 0.541683 +Steps: 1%| | 7808/1000000 [32:24<68:28:54, 4.02it/s, grad_norm=0.542, loss_final=0.477, loss_mean=0.82, loss_mean_cls=0.068, proj_loss=-0.411][2026-03-26 13:03:27] Step: 7808, Training Logs: loss_final: 0.447406, loss_mean: 0.803850, proj_loss: -0.422113, loss_mean_cls: 0.065668, grad_norm: 0.250003 +Steps: 1%| | 7809/1000000 [32:24<68:27:43, 4.03it/s, grad_norm=0.25, loss_final=0.447, loss_mean=0.804, loss_mean_cls=0.0657, proj_loss=-0.422][2026-03-26 13:03:27] Step: 7809, Training Logs: loss_final: 0.468637, loss_mean: 0.833401, proj_loss: -0.429456, loss_mean_cls: 0.064692, grad_norm: 0.757023 +Steps: 1%| | 7810/1000000 [32:24<68:29:23, 4.02it/s, grad_norm=0.757, loss_final=0.469, loss_mean=0.833, loss_mean_cls=0.0647, proj_loss=-0.429][2026-03-26 13:03:28] Step: 7810, Training Logs: loss_final: 0.458362, loss_mean: 0.814318, proj_loss: -0.421985, loss_mean_cls: 0.066029, grad_norm: 0.344641 +Steps: 1%| | 7811/1000000 [32:24<68:29:01, 4.02it/s, grad_norm=0.345, loss_final=0.458, loss_mean=0.814, loss_mean_cls=0.066, proj_loss=-0.422][2026-03-26 13:03:28] Step: 7811, Training Logs: loss_final: 0.473074, loss_mean: 0.834328, proj_loss: -0.426457, loss_mean_cls: 0.065202, grad_norm: 0.715331 +Steps: 1%| | 7812/1000000 [32:25<68:31:05, 4.02it/s, grad_norm=0.715, loss_final=0.473, loss_mean=0.834, loss_mean_cls=0.0652, proj_loss=-0.426][2026-03-26 13:03:28] Step: 7812, Training Logs: loss_final: 0.468822, loss_mean: 0.816118, proj_loss: -0.415598, loss_mean_cls: 0.068303, grad_norm: 0.724394 +Steps: 1%| | 7813/1000000 [32:25<68:28:12, 4.03it/s, grad_norm=0.724, loss_final=0.469, loss_mean=0.816, loss_mean_cls=0.0683, proj_loss=-0.416][2026-03-26 13:03:28] Step: 7813, Training Logs: loss_final: 0.481905, loss_mean: 0.833601, proj_loss: -0.417643, loss_mean_cls: 0.065948, grad_norm: 0.346782 +Steps: 1%| | 7814/1000000 [32:25<68:28:12, 4.03it/s, grad_norm=0.347, loss_final=0.482, loss_mean=0.834, loss_mean_cls=0.0659, proj_loss=-0.418][2026-03-26 13:03:29] Step: 7814, Training Logs: loss_final: 0.448604, loss_mean: 0.801021, proj_loss: -0.419364, loss_mean_cls: 0.066947, grad_norm: 0.859948 +Steps: 1%| | 7815/1000000 [32:25<68:24:23, 4.03it/s, grad_norm=0.86, loss_final=0.449, loss_mean=0.801, loss_mean_cls=0.0669, proj_loss=-0.419][2026-03-26 13:03:29] Step: 7815, Training Logs: loss_final: 0.477331, loss_mean: 0.833300, proj_loss: -0.421604, loss_mean_cls: 0.065634, grad_norm: 0.334596 +Steps: 1%| | 7816/1000000 [32:26<68:26:36, 4.03it/s, grad_norm=0.335, loss_final=0.477, loss_mean=0.833, loss_mean_cls=0.0656, proj_loss=-0.422][2026-03-26 13:03:29] Step: 7816, Training Logs: loss_final: 0.460199, loss_mean: 0.814702, proj_loss: -0.420860, loss_mean_cls: 0.066358, grad_norm: 0.516917 +Steps: 1%| | 7817/1000000 [32:26<68:29:08, 4.02it/s, grad_norm=0.517, loss_final=0.46, loss_mean=0.815, loss_mean_cls=0.0664, proj_loss=-0.421][2026-03-26 13:03:29] Step: 7817, Training Logs: loss_final: 0.464290, loss_mean: 0.812345, proj_loss: -0.415294, loss_mean_cls: 0.067238, grad_norm: 0.295369 +Steps: 1%| | 7818/1000000 [32:26<68:31:00, 4.02it/s, grad_norm=0.295, loss_final=0.464, loss_mean=0.812, loss_mean_cls=0.0672, proj_loss=-0.415][2026-03-26 13:03:30] Step: 7818, Training Logs: loss_final: 0.464332, loss_mean: 0.811126, proj_loss: -0.414619, loss_mean_cls: 0.067824, grad_norm: 0.633540 +Steps: 1%| | 7819/1000000 [32:26<68:29:20, 4.02it/s, grad_norm=0.634, loss_final=0.464, loss_mean=0.811, loss_mean_cls=0.0678, proj_loss=-0.415][2026-03-26 13:03:30] Step: 7819, Training Logs: loss_final: 0.466494, loss_mean: 0.817026, proj_loss: -0.418275, loss_mean_cls: 0.067743, grad_norm: 0.660033 +Steps: 1%| | 7820/1000000 [32:27<68:29:40, 4.02it/s, grad_norm=0.66, loss_final=0.466, loss_mean=0.817, loss_mean_cls=0.0677, proj_loss=-0.418][2026-03-26 13:03:30] Step: 7820, Training Logs: loss_final: 0.460119, loss_mean: 0.814683, proj_loss: -0.421250, loss_mean_cls: 0.066686, grad_norm: 0.237878 +Steps: 1%| | 7821/1000000 [32:27<68:35:03, 4.02it/s, grad_norm=0.238, loss_final=0.46, loss_mean=0.815, loss_mean_cls=0.0667, proj_loss=-0.421][2026-03-26 13:03:30] Step: 7821, Training Logs: loss_final: 0.468213, loss_mean: 0.813083, proj_loss: -0.411994, loss_mean_cls: 0.067123, grad_norm: 0.564224 +Steps: 1%| | 7822/1000000 [32:27<68:32:31, 4.02it/s, grad_norm=0.564, loss_final=0.468, loss_mean=0.813, loss_mean_cls=0.0671, proj_loss=-0.412][2026-03-26 13:03:31] Step: 7822, Training Logs: loss_final: 0.447990, loss_mean: 0.794366, proj_loss: -0.414435, loss_mean_cls: 0.068058, grad_norm: 0.232772 +Steps: 1%| | 7823/1000000 [32:27<68:31:51, 4.02it/s, grad_norm=0.233, loss_final=0.448, loss_mean=0.794, loss_mean_cls=0.0681, proj_loss=-0.414][2026-03-26 13:03:31] Step: 7823, Training Logs: loss_final: 0.459512, loss_mean: 0.820265, proj_loss: -0.425586, loss_mean_cls: 0.064832, grad_norm: 0.539908 +Steps: 1%| | 7824/1000000 [32:28<68:30:58, 4.02it/s, grad_norm=0.54, loss_final=0.46, loss_mean=0.82, loss_mean_cls=0.0648, proj_loss=-0.426][2026-03-26 13:03:31] Step: 7824, Training Logs: loss_final: 0.457026, loss_mean: 0.806979, proj_loss: -0.417330, loss_mean_cls: 0.067378, grad_norm: 0.298712 +Steps: 1%| | 7825/1000000 [32:28<68:32:33, 4.02it/s, grad_norm=0.299, loss_final=0.457, loss_mean=0.807, loss_mean_cls=0.0674, proj_loss=-0.417][2026-03-26 13:03:31] Step: 7825, Training Logs: loss_final: 0.458247, loss_mean: 0.810247, proj_loss: -0.418192, loss_mean_cls: 0.066193, grad_norm: 0.590294 +Steps: 1%| | 7826/1000000 [32:28<68:32:36, 4.02it/s, grad_norm=0.59, loss_final=0.458, loss_mean=0.81, loss_mean_cls=0.0662, proj_loss=-0.418][2026-03-26 13:03:32] Step: 7826, Training Logs: loss_final: 0.451419, loss_mean: 0.804677, proj_loss: -0.419534, loss_mean_cls: 0.066276, grad_norm: 0.446811 +Steps: 1%| | 7827/1000000 [32:28<68:30:55, 4.02it/s, grad_norm=0.447, loss_final=0.451, loss_mean=0.805, loss_mean_cls=0.0663, proj_loss=-0.42][2026-03-26 13:03:32] Step: 7827, Training Logs: loss_final: 0.464500, loss_mean: 0.825354, proj_loss: -0.425864, loss_mean_cls: 0.065009, grad_norm: 0.288783 +Steps: 1%| | 7828/1000000 [32:29<68:30:38, 4.02it/s, grad_norm=0.289, loss_final=0.464, loss_mean=0.825, loss_mean_cls=0.065, proj_loss=-0.426][2026-03-26 13:03:32] Step: 7828, Training Logs: loss_final: 0.460319, loss_mean: 0.802034, proj_loss: -0.411031, loss_mean_cls: 0.069316, grad_norm: 0.286545 +Steps: 1%| | 7829/1000000 [32:29<68:31:47, 4.02it/s, grad_norm=0.287, loss_final=0.46, loss_mean=0.802, loss_mean_cls=0.0693, proj_loss=-0.411][2026-03-26 13:03:32] Step: 7829, Training Logs: loss_final: 0.464953, loss_mean: 0.815808, proj_loss: -0.418174, loss_mean_cls: 0.067318, grad_norm: 0.223292 +Steps: 1%| | 7830/1000000 [32:29<68:31:41, 4.02it/s, grad_norm=0.223, loss_final=0.465, loss_mean=0.816, loss_mean_cls=0.0673, proj_loss=-0.418][2026-03-26 13:03:33] Step: 7830, Training Logs: loss_final: 0.497398, loss_mean: 0.840346, proj_loss: -0.411407, loss_mean_cls: 0.068459, grad_norm: 0.305230 +Steps: 1%| | 7831/1000000 [32:29<68:26:40, 4.03it/s, grad_norm=0.305, loss_final=0.497, loss_mean=0.84, loss_mean_cls=0.0685, proj_loss=-0.411][2026-03-26 13:03:33] Step: 7831, Training Logs: loss_final: 0.465491, loss_mean: 0.806480, proj_loss: -0.409098, loss_mean_cls: 0.068108, grad_norm: 0.275828 +Steps: 1%| | 7832/1000000 [32:30<68:21:57, 4.03it/s, grad_norm=0.276, loss_final=0.465, loss_mean=0.806, loss_mean_cls=0.0681, proj_loss=-0.409][2026-03-26 13:03:33] Step: 7832, Training Logs: loss_final: 0.477644, loss_mean: 0.829929, proj_loss: -0.419134, loss_mean_cls: 0.066849, grad_norm: 0.252900 +Steps: 1%| | 7833/1000000 [32:30<68:21:09, 4.03it/s, grad_norm=0.253, loss_final=0.478, loss_mean=0.83, loss_mean_cls=0.0668, proj_loss=-0.419][2026-03-26 13:03:33] Step: 7833, Training Logs: loss_final: 0.458776, loss_mean: 0.812112, proj_loss: -0.420178, loss_mean_cls: 0.066842, grad_norm: 0.275326 +Steps: 1%| | 7834/1000000 [32:30<68:18:45, 4.03it/s, grad_norm=0.275, loss_final=0.459, loss_mean=0.812, loss_mean_cls=0.0668, proj_loss=-0.42][2026-03-26 13:03:34] Step: 7834, Training Logs: loss_final: 0.483046, loss_mean: 0.829564, proj_loss: -0.413983, loss_mean_cls: 0.067465, grad_norm: 0.240069 +Steps: 1%| | 7835/1000000 [32:30<68:17:30, 4.04it/s, grad_norm=0.24, loss_final=0.483, loss_mean=0.83, loss_mean_cls=0.0675, proj_loss=-0.414][2026-03-26 13:03:34] Step: 7835, Training Logs: loss_final: 0.465537, loss_mean: 0.822455, proj_loss: -0.422914, loss_mean_cls: 0.065997, grad_norm: 0.202601 +Steps: 1%| | 7836/1000000 [32:31<68:18:24, 4.03it/s, grad_norm=0.203, loss_final=0.466, loss_mean=0.822, loss_mean_cls=0.066, proj_loss=-0.423][2026-03-26 13:03:34] Step: 7836, Training Logs: loss_final: 0.467402, loss_mean: 0.799514, proj_loss: -0.402058, loss_mean_cls: 0.069945, grad_norm: 0.238342 +Steps: 1%| | 7837/1000000 [32:31<68:16:45, 4.04it/s, grad_norm=0.238, loss_final=0.467, loss_mean=0.8, loss_mean_cls=0.0699, proj_loss=-0.402][2026-03-26 13:03:34] Step: 7837, Training Logs: loss_final: 0.465366, loss_mean: 0.810790, proj_loss: -0.413150, loss_mean_cls: 0.067726, grad_norm: 0.204325 +Steps: 1%| | 7838/1000000 [32:31<68:12:59, 4.04it/s, grad_norm=0.204, loss_final=0.465, loss_mean=0.811, loss_mean_cls=0.0677, proj_loss=-0.413][2026-03-26 13:03:34] Step: 7838, Training Logs: loss_final: 0.462792, loss_mean: 0.808393, proj_loss: -0.413194, loss_mean_cls: 0.067593, grad_norm: 0.246272 +Steps: 1%| | 7839/1000000 [32:31<68:15:07, 4.04it/s, grad_norm=0.246, loss_final=0.463, loss_mean=0.808, loss_mean_cls=0.0676, proj_loss=-0.413][2026-03-26 13:03:35] Step: 7839, Training Logs: loss_final: 0.461725, loss_mean: 0.809726, proj_loss: -0.414865, loss_mean_cls: 0.066864, grad_norm: 0.233138 +Steps: 1%| | 7840/1000000 [32:32<68:14:54, 4.04it/s, grad_norm=0.233, loss_final=0.462, loss_mean=0.81, loss_mean_cls=0.0669, proj_loss=-0.415][2026-03-26 13:03:35] Step: 7840, Training Logs: loss_final: 0.455370, loss_mean: 0.798873, proj_loss: -0.411756, loss_mean_cls: 0.068253, grad_norm: 0.202041 +Steps: 1%| | 7841/1000000 [32:32<68:16:30, 4.04it/s, grad_norm=0.202, loss_final=0.455, loss_mean=0.799, loss_mean_cls=0.0683, proj_loss=-0.412][2026-03-26 13:03:35] Step: 7841, Training Logs: loss_final: 0.453880, loss_mean: 0.812381, proj_loss: -0.423135, loss_mean_cls: 0.064634, grad_norm: 0.224009 +Steps: 1%| | 7842/1000000 [32:32<68:15:18, 4.04it/s, grad_norm=0.224, loss_final=0.454, loss_mean=0.812, loss_mean_cls=0.0646, proj_loss=-0.423][2026-03-26 13:03:35] Step: 7842, Training Logs: loss_final: 0.457091, loss_mean: 0.814354, proj_loss: -0.423357, loss_mean_cls: 0.066095, grad_norm: 0.368250 +Steps: 1%| | 7843/1000000 [32:32<68:14:53, 4.04it/s, grad_norm=0.368, loss_final=0.457, loss_mean=0.814, loss_mean_cls=0.0661, proj_loss=-0.423][2026-03-26 13:03:36] Step: 7843, Training Logs: loss_final: 0.478246, loss_mean: 0.805956, proj_loss: -0.400259, loss_mean_cls: 0.072549, grad_norm: 0.229751 +Steps: 1%| | 7844/1000000 [32:33<68:14:20, 4.04it/s, grad_norm=0.23, loss_final=0.478, loss_mean=0.806, loss_mean_cls=0.0725, proj_loss=-0.4][2026-03-26 13:03:36] Step: 7844, Training Logs: loss_final: 0.472674, loss_mean: 0.812865, proj_loss: -0.409154, loss_mean_cls: 0.068963, grad_norm: 0.362994 +Steps: 1%| | 7845/1000000 [32:33<68:12:06, 4.04it/s, grad_norm=0.363, loss_final=0.473, loss_mean=0.813, loss_mean_cls=0.069, proj_loss=-0.409][2026-03-26 13:03:36] Step: 7845, Training Logs: loss_final: 0.486858, loss_mean: 0.839478, proj_loss: -0.418542, loss_mean_cls: 0.065922, grad_norm: 0.332159 +Steps: 1%| | 7846/1000000 [32:33<68:13:10, 4.04it/s, grad_norm=0.332, loss_final=0.487, loss_mean=0.839, loss_mean_cls=0.0659, proj_loss=-0.419][2026-03-26 13:03:36] Step: 7846, Training Logs: loss_final: 0.462936, loss_mean: 0.818617, proj_loss: -0.421261, loss_mean_cls: 0.065580, grad_norm: 0.456255 +Steps: 1%| | 7847/1000000 [32:33<68:11:55, 4.04it/s, grad_norm=0.456, loss_final=0.463, loss_mean=0.819, loss_mean_cls=0.0656, proj_loss=-0.421][2026-03-26 13:03:37] Step: 7847, Training Logs: loss_final: 0.450150, loss_mean: 0.797611, proj_loss: -0.414861, loss_mean_cls: 0.067400, grad_norm: 0.594554 +Steps: 1%| | 7848/1000000 [32:34<68:11:57, 4.04it/s, grad_norm=0.595, loss_final=0.45, loss_mean=0.798, loss_mean_cls=0.0674, proj_loss=-0.415][2026-03-26 13:03:37] Step: 7848, Training Logs: loss_final: 0.456516, loss_mean: 0.805708, proj_loss: -0.416961, loss_mean_cls: 0.067768, grad_norm: 0.328969 +Steps: 1%| | 7849/1000000 [32:34<68:13:50, 4.04it/s, grad_norm=0.329, loss_final=0.457, loss_mean=0.806, loss_mean_cls=0.0678, proj_loss=-0.417][2026-03-26 13:03:37] Step: 7849, Training Logs: loss_final: 0.455857, loss_mean: 0.794130, proj_loss: -0.408276, loss_mean_cls: 0.070004, grad_norm: 0.348082 +Steps: 1%| | 7850/1000000 [32:34<68:15:36, 4.04it/s, grad_norm=0.348, loss_final=0.456, loss_mean=0.794, loss_mean_cls=0.07, proj_loss=-0.408][2026-03-26 13:03:37] Step: 7850, Training Logs: loss_final: 0.483584, loss_mean: 0.823413, proj_loss: -0.408821, loss_mean_cls: 0.068992, grad_norm: 0.360269 +Steps: 1%| | 7851/1000000 [32:34<68:15:27, 4.04it/s, grad_norm=0.36, loss_final=0.484, loss_mean=0.823, loss_mean_cls=0.069, proj_loss=-0.409][2026-03-26 13:03:38] Step: 7851, Training Logs: loss_final: 0.466822, loss_mean: 0.820831, proj_loss: -0.419888, loss_mean_cls: 0.065880, grad_norm: 0.284645 +Steps: 1%| | 7851/1000000 [32:34<68:15:27, 4.04it/s, grad_norm=0.285, loss_final=0.467, loss_mean=0.821, loss_mean_cls=0.0659, proj_loss=-0.42]