diff --git "a/runs/big_run/exp_c_8b/run_20260409_205717/train.log" "b/runs/big_run/exp_c_8b/run_20260409_205717/train.log" new file mode 100644--- /dev/null +++ "b/runs/big_run/exp_c_8b/run_20260409_205717/train.log" @@ -0,0 +1,2999 @@ +2026-04-09 20:57:18 - INFO - parrotllm - Logging initialised -> runs/big_run/exp_c_8b/run_20260409_205717/train.log +2026-04-09 20:57:18 - INFO - parrotllm.training - device=cuda:0 | rank=0 | world_size=8 | distributed=yes +2026-04-09 20:57:18 - INFO - parrotllm.training - checkpoint directory=/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints | keep_last=10 | keep_best=10 +2026-04-09 20:57:18 - INFO - parrotllm.training - train windows=random-overlapping | tokens=8083182257 | valid_starts=8083181233 | sampled_windows_per_epoch=8083181233 +2026-04-09 20:57:18 - INFO - parrotllm.training - val windows=sequential | tokens=81648305 | stride=512 | windows=159468 +2026-04-09 20:57:19 - INFO - parrotllm.training - compiling model with torch.compile... +2026-04-09 20:57:20 - INFO - parrotllm.training - +============================================================ +MODEL ARCHITECTURE SUMMARY +============================================================ + +Configuration: + Vocab size: 50258 + Block size (context): 1024 + Layers: 14 + Heads: 6 + Embedding dim: 384 + FFN hidden dim: 768 + Dropout: 0.01287921140461293 + Bias: False + +Parameters (unique, weight-tied layers counted once): + Total: 39,966,592 (39.97M) + Trainable: 39,966,592 (39.97M) + Non-trainable: 0 + Non-embedding: 39,966,592 (39.97M) + Position embeddings: 0 (RoPE: 0 learned params) + Size (MB): 159.87 + +Layer-wise breakdown (torchinfo): + Note: torchinfo double-counts weight-tied layers (tok_emb/lm_head). +============================================================================================================================================ +Layer (type:depth-idx) Input Shape Output Shape Param # Trainable +============================================================================================================================================ +ParrotLLM [32, 1024] [32, 1024, 50258] -- True +├─Embedding: 1-1 [32, 1024] [32, 1024, 384] 19,299,072 True +├─Dropout: 1-2 [32, 1024, 384] [32, 1024, 384] -- -- +├─ModuleList: 1-3 -- -- -- True +│ └─TransformerBlock: 2-1 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-1 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-2 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-3 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-4 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-5 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-6 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-2 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-7 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-8 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-9 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-10 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-11 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-12 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-3 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-13 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-14 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-15 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-16 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-17 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-18 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-4 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-19 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-20 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-21 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-22 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-23 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-24 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-5 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-25 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-26 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-27 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-28 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-29 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-30 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-6 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-31 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-32 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-33 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-34 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-35 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-36 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-7 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-37 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-38 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-39 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-40 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-41 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-42 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-8 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-43 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-44 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-45 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-46 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-47 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-48 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-9 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-49 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-50 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-51 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-52 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-53 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-54 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-10 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-55 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-56 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-57 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-58 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-59 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-60 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-11 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-61 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-62 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-63 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-64 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-65 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-66 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-12 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-67 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-68 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-69 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-70 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-71 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-72 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-13 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-73 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-74 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-75 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-76 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-77 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-78 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-14 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-79 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-80 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-81 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-82 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-83 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-84 [32, 1024, 384] [32, 1024, 384] 384 True +├─RMSNorm: 1-4 [32, 1024, 384] [32, 1024, 384] 384 True +├─Linear: 1-5 [32, 1024, 384] [32, 1024, 50258] 19,299,072 True +============================================================================================================================================ +Total params: 59,265,664 +Trainable params: 59,265,664 +Non-trainable params: 0 +Total mult-adds (Units.GIGABYTES): 1.90 +============================================================================================================================================ +Input size (MB): 0.26 +Forward/backward pass size (MB): 34515.45 +Params size (MB): 237.06 +Estimated Total Size (MB): 34752.78 +============================================================================================================================================ +============================================================ +2026-04-09 20:57:20 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 20:57:23 - INFO - parrotllm.training - Initial val: loss=10.8942, ppl=53862.54 +2026-04-09 20:57:23 - INFO - parrotllm.training - +============================================================ +Starting training... + Steps per epoch (approx): 31574926 + Max steps: 122070 +============================================================ +2026-04-09 20:57:23 - INFO - parrotllm.training - LR schedule=wsd | decay_ratio=0.25826046407623926 | z_loss_coeff=0.00013887413811122497 +2026-04-09 20:57:23 - INFO - parrotllm.training - Early stopping enabled | patience=5 evals | min_delta=0.001000 +2026-04-09 20:57:41 - INFO - parrotllm.training - step 1 | epoch 0 | loss 10.9076 | lr 1.75e-06 | grad 7.0844 +2026-04-09 20:57:44 - INFO - parrotllm.training - step 10 | epoch 0 | loss 10.4171 | lr 1.75e-05 | grad 3.3182 +2026-04-09 20:57:47 - INFO - parrotllm.training - step 20 | epoch 0 | loss 10.1428 | lr 3.51e-05 | grad 1.6992 +2026-04-09 20:57:50 - INFO - parrotllm.training - step 30 | epoch 0 | loss 9.8734 | lr 5.26e-05 | grad 1.6849 +2026-04-09 20:57:53 - INFO - parrotllm.training - step 40 | epoch 0 | loss 9.5334 | lr 7.01e-05 | grad 1.6301 +2026-04-09 20:57:56 - INFO - parrotllm.training - step 50 | epoch 0 | loss 9.1426 | lr 8.76e-05 | grad 1.5705 +2026-04-09 20:57:59 - INFO - parrotllm.training - step 60 | epoch 0 | loss 8.8005 | lr 1.05e-04 | grad 1.5589 +2026-04-09 20:58:02 - INFO - parrotllm.training - step 70 | epoch 0 | loss 8.2721 | lr 1.23e-04 | grad 1.5379 +2026-04-09 20:58:05 - INFO - parrotllm.training - step 80 | epoch 0 | loss 8.0198 | lr 1.40e-04 | grad 1.3195 +2026-04-09 20:58:08 - INFO - parrotllm.training - step 90 | epoch 0 | loss 7.5926 | lr 1.58e-04 | grad 1.1542 +2026-04-09 20:58:11 - INFO - parrotllm.training - step 100 | epoch 0 | loss 7.3171 | lr 1.75e-04 | grad 1.0949 +2026-04-09 20:58:14 - INFO - parrotllm.training - step 110 | epoch 0 | loss 7.1642 | lr 1.93e-04 | grad 0.9382 +2026-04-09 20:58:18 - INFO - parrotllm.training - step 120 | epoch 0 | loss 7.0070 | lr 2.10e-04 | grad 0.5934 +2026-04-09 20:58:21 - INFO - parrotllm.training - step 130 | epoch 0 | loss 6.8397 | lr 2.28e-04 | grad 1.3176 +2026-04-09 20:58:24 - INFO - parrotllm.training - step 140 | epoch 0 | loss 6.6803 | lr 2.45e-04 | grad 0.6868 +2026-04-09 20:58:27 - INFO - parrotllm.training - step 150 | epoch 0 | loss 6.5429 | lr 2.63e-04 | grad 0.7972 +2026-04-09 20:58:30 - INFO - parrotllm.training - step 160 | epoch 0 | loss 6.3811 | lr 2.80e-04 | grad 0.9517 +2026-04-09 20:58:33 - INFO - parrotllm.training - step 170 | epoch 0 | loss 6.2879 | lr 2.98e-04 | grad 1.1569 +2026-04-09 20:58:36 - INFO - parrotllm.training - step 180 | epoch 0 | loss 6.2406 | lr 3.16e-04 | grad 1.3275 +2026-04-09 20:58:39 - INFO - parrotllm.training - step 190 | epoch 0 | loss 6.2866 | lr 3.33e-04 | grad 1.1334 +2026-04-09 20:58:42 - INFO - parrotllm.training - step 200 | epoch 0 | loss 6.1365 | lr 3.51e-04 | grad 1.3085 +2026-04-09 20:58:45 - INFO - parrotllm.training - step 210 | epoch 0 | loss 5.9613 | lr 3.68e-04 | grad 0.9725 +2026-04-09 20:58:48 - INFO - parrotllm.training - step 220 | epoch 0 | loss 5.9314 | lr 3.86e-04 | grad 0.9849 +2026-04-09 20:58:51 - INFO - parrotllm.training - step 230 | epoch 0 | loss 5.9664 | lr 4.03e-04 | grad 1.2253 +2026-04-09 20:58:54 - INFO - parrotllm.training - step 240 | epoch 0 | loss 5.8495 | lr 4.21e-04 | grad 1.1720 +2026-04-09 20:58:57 - INFO - parrotllm.training - step 250 | epoch 0 | loss 5.8320 | lr 4.38e-04 | grad 1.0126 +2026-04-09 20:59:00 - INFO - parrotllm.training - step 260 | epoch 0 | loss 5.6836 | lr 4.56e-04 | grad 1.3296 +2026-04-09 20:59:03 - INFO - parrotllm.training - step 270 | epoch 0 | loss 5.6749 | lr 4.73e-04 | grad 1.1185 +2026-04-09 20:59:07 - INFO - parrotllm.training - step 280 | epoch 0 | loss 5.7278 | lr 4.91e-04 | grad 1.0457 +2026-04-09 20:59:10 - INFO - parrotllm.training - step 290 | epoch 0 | loss 5.5052 | lr 5.08e-04 | grad 1.2953 +2026-04-09 20:59:13 - INFO - parrotllm.training - step 300 | epoch 0 | loss 5.5044 | lr 5.26e-04 | grad 0.8182 +2026-04-09 20:59:16 - INFO - parrotllm.training - step 310 | epoch 0 | loss 5.5753 | lr 5.43e-04 | grad 1.3712 +2026-04-09 20:59:19 - INFO - parrotllm.training - step 320 | epoch 0 | loss 5.5068 | lr 5.61e-04 | grad 1.2363 +2026-04-09 20:59:22 - INFO - parrotllm.training - step 330 | epoch 0 | loss 5.4790 | lr 5.78e-04 | grad 0.8789 +2026-04-09 20:59:25 - INFO - parrotllm.training - step 340 | epoch 0 | loss 5.3873 | lr 5.96e-04 | grad 0.7224 +2026-04-09 20:59:28 - INFO - parrotllm.training - step 350 | epoch 0 | loss 5.3656 | lr 6.13e-04 | grad 0.8745 +2026-04-09 20:59:31 - INFO - parrotllm.training - step 360 | epoch 0 | loss 5.2891 | lr 6.31e-04 | grad 1.1492 +2026-04-09 20:59:34 - INFO - parrotllm.training - step 370 | epoch 0 | loss 5.2196 | lr 6.49e-04 | grad 1.1031 +2026-04-09 20:59:37 - INFO - parrotllm.training - step 380 | epoch 0 | loss 5.1232 | lr 6.66e-04 | grad 0.7398 +2026-04-09 20:59:40 - INFO - parrotllm.training - step 390 | epoch 0 | loss 5.1211 | lr 6.84e-04 | grad 0.6801 +2026-04-09 20:59:43 - INFO - parrotllm.training - step 400 | epoch 0 | loss 5.0634 | lr 7.01e-04 | grad 0.8986 +2026-04-09 20:59:46 - INFO - parrotllm.training - step 410 | epoch 0 | loss 5.1907 | lr 7.19e-04 | grad 0.8567 +2026-04-09 20:59:49 - INFO - parrotllm.training - step 420 | epoch 0 | loss 5.0500 | lr 7.36e-04 | grad 0.6550 +2026-04-09 20:59:52 - INFO - parrotllm.training - step 430 | epoch 0 | loss 5.1327 | lr 7.54e-04 | grad 0.6631 +2026-04-09 20:59:55 - INFO - parrotllm.training - step 440 | epoch 0 | loss 5.0934 | lr 7.71e-04 | grad 0.6713 +2026-04-09 20:59:59 - INFO - parrotllm.training - step 450 | epoch 0 | loss 4.9388 | lr 7.89e-04 | grad 0.8150 +2026-04-09 21:00:02 - INFO - parrotllm.training - step 460 | epoch 0 | loss 4.9740 | lr 8.06e-04 | grad 0.7573 +2026-04-09 21:00:05 - INFO - parrotllm.training - step 470 | epoch 0 | loss 4.9498 | lr 8.24e-04 | grad 0.6280 +2026-04-09 21:00:08 - INFO - parrotllm.training - step 480 | epoch 0 | loss 4.9669 | lr 8.41e-04 | grad 0.6773 +2026-04-09 21:00:11 - INFO - parrotllm.training - step 490 | epoch 0 | loss 4.9454 | lr 8.59e-04 | grad 0.5298 +2026-04-09 21:00:14 - INFO - parrotllm.training - step 500 | epoch 0 | loss 4.8500 | lr 8.76e-04 | grad 0.5679 +2026-04-09 21:00:14 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:00:14 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:00:17 - INFO - parrotllm.training - Train: loss=4.8500, ppl=127.74 +2026-04-09 21:00:17 - INFO - parrotllm.training - Val: loss=4.7747, ppl=118.48 +2026-04-09 21:00:17 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:00:18 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_4p7747_epoch_0000_step_0000500.pt +2026-04-09 21:00:19 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:00:22 - INFO - parrotllm.training - step 510 | epoch 0 | loss 4.7024 | lr 8.94e-04 | grad 0.4380 +2026-04-09 21:00:25 - INFO - parrotllm.training - step 520 | epoch 0 | loss 4.8236 | lr 9.11e-04 | grad 0.9173 +2026-04-09 21:00:28 - INFO - parrotllm.training - step 530 | epoch 0 | loss 4.8002 | lr 9.29e-04 | grad 0.6427 +2026-04-09 21:00:32 - INFO - parrotllm.training - step 540 | epoch 0 | loss 4.8347 | lr 9.47e-04 | grad 0.5033 +2026-04-09 21:00:35 - INFO - parrotllm.training - step 550 | epoch 0 | loss 4.7250 | lr 9.64e-04 | grad 0.6872 +2026-04-09 21:00:38 - INFO - parrotllm.training - step 560 | epoch 0 | loss 4.5905 | lr 9.82e-04 | grad 0.5336 +2026-04-09 21:00:41 - INFO - parrotllm.training - step 570 | epoch 0 | loss 4.5785 | lr 9.99e-04 | grad 0.5821 +2026-04-09 21:00:44 - INFO - parrotllm.training - step 580 | epoch 0 | loss 4.6148 | lr 1.02e-03 | grad 0.5346 +2026-04-09 21:00:47 - INFO - parrotllm.training - step 590 | epoch 0 | loss 4.5642 | lr 1.03e-03 | grad 0.5751 +2026-04-09 21:00:50 - INFO - parrotllm.training - step 600 | epoch 0 | loss 4.5837 | lr 1.05e-03 | grad 0.5851 +2026-04-09 21:00:53 - INFO - parrotllm.training - step 610 | epoch 0 | loss 4.6206 | lr 1.07e-03 | grad 0.5647 +2026-04-09 21:00:56 - INFO - parrotllm.training - step 620 | epoch 0 | loss 4.4768 | lr 1.09e-03 | grad 0.6404 +2026-04-09 21:01:00 - INFO - parrotllm.training - step 630 | epoch 0 | loss 4.5334 | lr 1.10e-03 | grad 0.4749 +2026-04-09 21:01:03 - INFO - parrotllm.training - step 640 | epoch 0 | loss 4.4119 | lr 1.12e-03 | grad 0.4506 +2026-04-09 21:01:06 - INFO - parrotllm.training - step 650 | epoch 0 | loss 4.4367 | lr 1.14e-03 | grad 0.5355 +2026-04-09 21:01:09 - INFO - parrotllm.training - step 660 | epoch 0 | loss 4.4357 | lr 1.16e-03 | grad 0.4830 +2026-04-09 21:01:12 - INFO - parrotllm.training - step 670 | epoch 0 | loss 4.4959 | lr 1.17e-03 | grad 0.5099 +2026-04-09 21:01:15 - INFO - parrotllm.training - step 680 | epoch 0 | loss 4.4822 | lr 1.19e-03 | grad 0.5851 +2026-04-09 21:01:18 - INFO - parrotllm.training - step 690 | epoch 0 | loss 4.3583 | lr 1.21e-03 | grad 0.3927 +2026-04-09 21:01:21 - INFO - parrotllm.training - step 700 | epoch 0 | loss 4.3063 | lr 1.23e-03 | grad 0.4789 +2026-04-09 21:01:24 - INFO - parrotllm.training - step 710 | epoch 0 | loss 4.3974 | lr 1.24e-03 | grad 0.6268 +2026-04-09 21:01:27 - INFO - parrotllm.training - step 720 | epoch 0 | loss 4.3421 | lr 1.26e-03 | grad 0.3796 +2026-04-09 21:01:30 - INFO - parrotllm.training - step 730 | epoch 0 | loss 4.3590 | lr 1.28e-03 | grad 0.3717 +2026-04-09 21:01:33 - INFO - parrotllm.training - step 740 | epoch 0 | loss 4.3662 | lr 1.30e-03 | grad 0.6237 +2026-04-09 21:01:36 - INFO - parrotllm.training - step 750 | epoch 0 | loss 4.2838 | lr 1.31e-03 | grad 0.4276 +2026-04-09 21:01:39 - INFO - parrotllm.training - step 760 | epoch 0 | loss 4.3296 | lr 1.33e-03 | grad 0.3971 +2026-04-09 21:01:42 - INFO - parrotllm.training - step 770 | epoch 0 | loss 4.2521 | lr 1.35e-03 | grad 0.3305 +2026-04-09 21:01:45 - INFO - parrotllm.training - step 780 | epoch 0 | loss 4.3756 | lr 1.37e-03 | grad 0.3810 +2026-04-09 21:01:48 - INFO - parrotllm.training - step 790 | epoch 0 | loss 4.3893 | lr 1.38e-03 | grad 0.3796 +2026-04-09 21:01:51 - INFO - parrotllm.training - step 800 | epoch 0 | loss 4.3371 | lr 1.40e-03 | grad 0.4289 +2026-04-09 21:01:55 - INFO - parrotllm.training - step 810 | epoch 0 | loss 4.1388 | lr 1.42e-03 | grad 0.3570 +2026-04-09 21:01:58 - INFO - parrotllm.training - step 820 | epoch 0 | loss 4.1782 | lr 1.44e-03 | grad 0.3842 +2026-04-09 21:02:01 - INFO - parrotllm.training - step 830 | epoch 0 | loss 4.2766 | lr 1.45e-03 | grad 0.3143 +2026-04-09 21:02:04 - INFO - parrotllm.training - step 840 | epoch 0 | loss 4.2508 | lr 1.47e-03 | grad 0.4546 +2026-04-09 21:02:07 - INFO - parrotllm.training - step 850 | epoch 0 | loss 4.2617 | lr 1.49e-03 | grad 0.3464 +2026-04-09 21:02:10 - INFO - parrotllm.training - step 860 | epoch 0 | loss 4.3340 | lr 1.51e-03 | grad 0.3785 +2026-04-09 21:02:13 - INFO - parrotllm.training - step 870 | epoch 0 | loss 4.2858 | lr 1.52e-03 | grad 0.3566 +2026-04-09 21:02:16 - INFO - parrotllm.training - step 880 | epoch 0 | loss 4.1352 | lr 1.54e-03 | grad 0.3574 +2026-04-09 21:02:19 - INFO - parrotllm.training - step 890 | epoch 0 | loss 4.2341 | lr 1.56e-03 | grad 0.3571 +2026-04-09 21:02:22 - INFO - parrotllm.training - step 900 | epoch 0 | loss 4.2752 | lr 1.58e-03 | grad 0.2973 +2026-04-09 21:02:25 - INFO - parrotllm.training - step 910 | epoch 0 | loss 4.2057 | lr 1.60e-03 | grad 0.4907 +2026-04-09 21:02:28 - INFO - parrotllm.training - step 920 | epoch 0 | loss 4.1994 | lr 1.61e-03 | grad 0.3442 +2026-04-09 21:02:31 - INFO - parrotllm.training - step 930 | epoch 0 | loss 4.1589 | lr 1.63e-03 | grad 0.2912 +2026-04-09 21:02:34 - INFO - parrotllm.training - step 940 | epoch 0 | loss 4.0849 | lr 1.65e-03 | grad 0.3245 +2026-04-09 21:02:37 - INFO - parrotllm.training - step 950 | epoch 0 | loss 4.0451 | lr 1.67e-03 | grad 0.3263 +2026-04-09 21:02:40 - INFO - parrotllm.training - step 960 | epoch 0 | loss 4.1362 | lr 1.68e-03 | grad 0.3894 +2026-04-09 21:02:43 - INFO - parrotllm.training - step 970 | epoch 0 | loss 4.2121 | lr 1.70e-03 | grad 0.3424 +2026-04-09 21:02:46 - INFO - parrotllm.training - step 980 | epoch 0 | loss 4.0902 | lr 1.72e-03 | grad 0.3041 +2026-04-09 21:02:49 - INFO - parrotllm.training - step 990 | epoch 0 | loss 4.1260 | lr 1.74e-03 | grad 0.3811 +2026-04-09 21:02:53 - INFO - parrotllm.training - step 1000 | epoch 0 | loss 4.0827 | lr 1.75e-03 | grad 0.2298 +2026-04-09 21:02:53 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:02:53 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:02:56 - INFO - parrotllm.training - Train: loss=4.0827, ppl=59.30 +2026-04-09 21:02:56 - INFO - parrotllm.training - Val: loss=4.0267, ppl=56.08 +2026-04-09 21:02:56 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:02:57 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_4p0267_epoch_0000_step_0001000.pt +2026-04-09 21:02:57 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:03:00 - INFO - parrotllm.training - step 1010 | epoch 0 | loss 4.0440 | lr 1.77e-03 | grad 0.3713 +2026-04-09 21:03:03 - INFO - parrotllm.training - step 1020 | epoch 0 | loss 4.1306 | lr 1.79e-03 | grad 0.3436 +2026-04-09 21:03:07 - INFO - parrotllm.training - step 1030 | epoch 0 | loss 4.1543 | lr 1.81e-03 | grad 0.3365 +2026-04-09 21:03:10 - INFO - parrotllm.training - step 1040 | epoch 0 | loss 3.9688 | lr 1.82e-03 | grad 0.2555 +2026-04-09 21:03:13 - INFO - parrotllm.training - step 1050 | epoch 0 | loss 4.0905 | lr 1.84e-03 | grad 0.4406 +2026-04-09 21:03:16 - INFO - parrotllm.training - step 1060 | epoch 0 | loss 4.0935 | lr 1.86e-03 | grad 0.3052 +2026-04-09 21:03:19 - INFO - parrotllm.training - step 1070 | epoch 0 | loss 4.0624 | lr 1.88e-03 | grad 0.2737 +2026-04-09 21:03:22 - INFO - parrotllm.training - step 1080 | epoch 0 | loss 3.9767 | lr 1.89e-03 | grad 0.2416 +2026-04-09 21:03:25 - INFO - parrotllm.training - step 1090 | epoch 0 | loss 4.0537 | lr 1.91e-03 | grad 0.2741 +2026-04-09 21:03:28 - INFO - parrotllm.training - step 1100 | epoch 0 | loss 4.0598 | lr 1.93e-03 | grad 0.3706 +2026-04-09 21:03:31 - INFO - parrotllm.training - step 1110 | epoch 0 | loss 3.9287 | lr 1.95e-03 | grad 0.2955 +2026-04-09 21:03:34 - INFO - parrotllm.training - step 1120 | epoch 0 | loss 3.9829 | lr 1.96e-03 | grad 0.2468 +2026-04-09 21:03:37 - INFO - parrotllm.training - step 1130 | epoch 0 | loss 4.0578 | lr 1.98e-03 | grad 0.2901 +2026-04-09 21:03:40 - INFO - parrotllm.training - step 1140 | epoch 0 | loss 4.0898 | lr 2.00e-03 | grad 0.2697 +2026-04-09 21:03:43 - INFO - parrotllm.training - step 1150 | epoch 0 | loss 4.0487 | lr 2.02e-03 | grad 0.3316 +2026-04-09 21:03:46 - INFO - parrotllm.training - step 1160 | epoch 0 | loss 4.0543 | lr 2.03e-03 | grad 0.2852 +2026-04-09 21:03:49 - INFO - parrotllm.training - step 1170 | epoch 0 | loss 3.9806 | lr 2.05e-03 | grad 0.2765 +2026-04-09 21:03:52 - INFO - parrotllm.training - step 1180 | epoch 0 | loss 4.0435 | lr 2.07e-03 | grad 0.2825 +2026-04-09 21:03:55 - INFO - parrotllm.training - step 1190 | epoch 0 | loss 3.9186 | lr 2.09e-03 | grad 0.3245 +2026-04-09 21:03:59 - INFO - parrotllm.training - step 1200 | epoch 0 | loss 4.0618 | lr 2.10e-03 | grad 0.2617 +2026-04-09 21:04:02 - INFO - parrotllm.training - step 1210 | epoch 0 | loss 4.0750 | lr 2.12e-03 | grad 0.2416 +2026-04-09 21:04:05 - INFO - parrotllm.training - step 1220 | epoch 0 | loss 3.9123 | lr 2.14e-03 | grad 0.2644 +2026-04-09 21:04:08 - INFO - parrotllm.training - step 1230 | epoch 0 | loss 4.0545 | lr 2.16e-03 | grad 0.3101 +2026-04-09 21:04:11 - INFO - parrotllm.training - step 1240 | epoch 0 | loss 3.9627 | lr 2.17e-03 | grad 0.2576 +2026-04-09 21:04:14 - INFO - parrotllm.training - step 1250 | epoch 0 | loss 4.0210 | lr 2.19e-03 | grad 0.2627 +2026-04-09 21:04:17 - INFO - parrotllm.training - step 1260 | epoch 0 | loss 3.9069 | lr 2.21e-03 | grad 0.2481 +2026-04-09 21:04:20 - INFO - parrotllm.training - step 1270 | epoch 0 | loss 3.9566 | lr 2.23e-03 | grad 0.2878 +2026-04-09 21:04:23 - INFO - parrotllm.training - step 1280 | epoch 0 | loss 4.0710 | lr 2.24e-03 | grad 0.2335 +2026-04-09 21:04:26 - INFO - parrotllm.training - step 1290 | epoch 0 | loss 4.0440 | lr 2.26e-03 | grad 0.2902 +2026-04-09 21:04:29 - INFO - parrotllm.training - step 1300 | epoch 0 | loss 3.9647 | lr 2.28e-03 | grad 0.2636 +2026-04-09 21:04:32 - INFO - parrotllm.training - step 1310 | epoch 0 | loss 3.9686 | lr 2.30e-03 | grad 0.2017 +2026-04-09 21:04:35 - INFO - parrotllm.training - step 1320 | epoch 0 | loss 4.0663 | lr 2.31e-03 | grad 0.2450 +2026-04-09 21:04:38 - INFO - parrotllm.training - step 1330 | epoch 0 | loss 3.9733 | lr 2.33e-03 | grad 0.2469 +2026-04-09 21:04:41 - INFO - parrotllm.training - step 1340 | epoch 0 | loss 3.9458 | lr 2.35e-03 | grad 0.2700 +2026-04-09 21:04:44 - INFO - parrotllm.training - step 1350 | epoch 0 | loss 3.8984 | lr 2.37e-03 | grad 0.2401 +2026-04-09 21:04:47 - INFO - parrotllm.training - step 1360 | epoch 0 | loss 3.8945 | lr 2.38e-03 | grad 0.2281 +2026-04-09 21:04:50 - INFO - parrotllm.training - step 1370 | epoch 0 | loss 3.9120 | lr 2.40e-03 | grad 0.2102 +2026-04-09 21:04:54 - INFO - parrotllm.training - step 1380 | epoch 0 | loss 3.9827 | lr 2.42e-03 | grad 0.2192 +2026-04-09 21:04:57 - INFO - parrotllm.training - step 1390 | epoch 0 | loss 4.0654 | lr 2.44e-03 | grad 0.2443 +2026-04-09 21:05:00 - INFO - parrotllm.training - step 1400 | epoch 0 | loss 3.9025 | lr 2.45e-03 | grad 0.2164 +2026-04-09 21:05:03 - INFO - parrotllm.training - step 1410 | epoch 0 | loss 3.9400 | lr 2.47e-03 | grad 0.2484 +2026-04-09 21:05:06 - INFO - parrotllm.training - step 1420 | epoch 0 | loss 4.0685 | lr 2.49e-03 | grad 0.2513 +2026-04-09 21:05:09 - INFO - parrotllm.training - step 1430 | epoch 0 | loss 3.9851 | lr 2.51e-03 | grad 0.1919 +2026-04-09 21:05:12 - INFO - parrotllm.training - step 1440 | epoch 0 | loss 3.8662 | lr 2.52e-03 | grad 0.2900 +2026-04-09 21:05:15 - INFO - parrotllm.training - step 1450 | epoch 0 | loss 3.9613 | lr 2.54e-03 | grad 0.2331 +2026-04-09 21:05:18 - INFO - parrotllm.training - step 1460 | epoch 0 | loss 3.9794 | lr 2.56e-03 | grad 0.2146 +2026-04-09 21:05:21 - INFO - parrotllm.training - step 1470 | epoch 0 | loss 3.9520 | lr 2.58e-03 | grad 0.2151 +2026-04-09 21:05:24 - INFO - parrotllm.training - step 1480 | epoch 0 | loss 3.9784 | lr 2.59e-03 | grad 0.2345 +2026-04-09 21:05:27 - INFO - parrotllm.training - step 1490 | epoch 0 | loss 3.9420 | lr 2.61e-03 | grad 0.2134 +2026-04-09 21:05:30 - INFO - parrotllm.training - step 1500 | epoch 0 | loss 3.8511 | lr 2.63e-03 | grad 0.2191 +2026-04-09 21:05:30 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:05:30 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:05:33 - INFO - parrotllm.training - Train: loss=3.8511, ppl=47.04 +2026-04-09 21:05:33 - INFO - parrotllm.training - Val: loss=3.8691, ppl=47.90 +2026-04-09 21:05:33 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:05:34 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p8691_epoch_0000_step_0001500.pt +2026-04-09 21:05:35 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:05:38 - INFO - parrotllm.training - step 1510 | epoch 0 | loss 4.0508 | lr 2.63e-03 | grad 0.1971 +2026-04-09 21:05:41 - INFO - parrotllm.training - step 1520 | epoch 0 | loss 3.9700 | lr 2.63e-03 | grad 0.2107 +2026-04-09 21:05:44 - INFO - parrotllm.training - step 1530 | epoch 0 | loss 3.8881 | lr 2.63e-03 | grad 0.1977 +2026-04-09 21:05:47 - INFO - parrotllm.training - step 1540 | epoch 0 | loss 3.9729 | lr 2.63e-03 | grad 0.2347 +2026-04-09 21:05:50 - INFO - parrotllm.training - step 1550 | epoch 0 | loss 3.9222 | lr 2.63e-03 | grad 0.1931 +2026-04-09 21:05:53 - INFO - parrotllm.training - step 1560 | epoch 0 | loss 3.9449 | lr 2.63e-03 | grad 0.2760 +2026-04-09 21:05:56 - INFO - parrotllm.training - step 1570 | epoch 0 | loss 3.9634 | lr 2.63e-03 | grad 0.1878 +2026-04-09 21:05:59 - INFO - parrotllm.training - step 1580 | epoch 0 | loss 3.9254 | lr 2.63e-03 | grad 0.1951 +2026-04-09 21:06:03 - INFO - parrotllm.training - step 1590 | epoch 0 | loss 3.9091 | lr 2.63e-03 | grad 0.2139 +2026-04-09 21:06:06 - INFO - parrotllm.training - step 1600 | epoch 0 | loss 3.8384 | lr 2.63e-03 | grad 0.1735 +2026-04-09 21:06:09 - INFO - parrotllm.training - step 1610 | epoch 0 | loss 3.9703 | lr 2.63e-03 | grad 0.1962 +2026-04-09 21:06:12 - INFO - parrotllm.training - step 1620 | epoch 0 | loss 3.8264 | lr 2.63e-03 | grad 0.2164 +2026-04-09 21:06:15 - INFO - parrotllm.training - step 1630 | epoch 0 | loss 3.8477 | lr 2.63e-03 | grad 0.1962 +2026-04-09 21:06:18 - INFO - parrotllm.training - step 1640 | epoch 0 | loss 3.8227 | lr 2.63e-03 | grad 0.1857 +2026-04-09 21:06:21 - INFO - parrotllm.training - step 1650 | epoch 0 | loss 3.9288 | lr 2.63e-03 | grad 0.1878 +2026-04-09 21:06:24 - INFO - parrotllm.training - step 1660 | epoch 0 | loss 3.8314 | lr 2.63e-03 | grad 0.1779 +2026-04-09 21:06:27 - INFO - parrotllm.training - step 1670 | epoch 0 | loss 3.9554 | lr 2.63e-03 | grad 0.1744 +2026-04-09 21:06:30 - INFO - parrotllm.training - step 1680 | epoch 0 | loss 3.9282 | lr 2.63e-03 | grad 0.2858 +2026-04-09 21:06:33 - INFO - parrotllm.training - step 1690 | epoch 0 | loss 3.7921 | lr 2.63e-03 | grad 0.1885 +2026-04-09 21:06:36 - INFO - parrotllm.training - step 1700 | epoch 0 | loss 3.9810 | lr 2.63e-03 | grad 0.1901 +2026-04-09 21:06:39 - INFO - parrotllm.training - step 1710 | epoch 0 | loss 3.7861 | lr 2.63e-03 | grad 0.2378 +2026-04-09 21:06:42 - INFO - parrotllm.training - step 1720 | epoch 0 | loss 3.8601 | lr 2.63e-03 | grad 0.1764 +2026-04-09 21:06:45 - INFO - parrotllm.training - step 1730 | epoch 0 | loss 3.7498 | lr 2.63e-03 | grad 0.1634 +2026-04-09 21:06:48 - INFO - parrotllm.training - step 1740 | epoch 0 | loss 3.8580 | lr 2.63e-03 | grad 0.2268 +2026-04-09 21:06:52 - INFO - parrotllm.training - step 1750 | epoch 0 | loss 3.8002 | lr 2.63e-03 | grad 0.1802 +2026-04-09 21:06:55 - INFO - parrotllm.training - step 1760 | epoch 0 | loss 3.7370 | lr 2.63e-03 | grad 0.1620 +2026-04-09 21:06:58 - INFO - parrotllm.training - step 1770 | epoch 0 | loss 3.7643 | lr 2.63e-03 | grad 0.2176 +2026-04-09 21:07:01 - INFO - parrotllm.training - step 1780 | epoch 0 | loss 3.7768 | lr 2.63e-03 | grad 0.1738 +2026-04-09 21:07:04 - INFO - parrotllm.training - step 1790 | epoch 0 | loss 3.8956 | lr 2.63e-03 | grad 0.2378 +2026-04-09 21:07:07 - INFO - parrotllm.training - step 1800 | epoch 0 | loss 3.7632 | lr 2.63e-03 | grad 0.1709 +2026-04-09 21:07:10 - INFO - parrotllm.training - step 1810 | epoch 0 | loss 3.8711 | lr 2.63e-03 | grad 0.1566 +2026-04-09 21:07:13 - INFO - parrotllm.training - step 1820 | epoch 0 | loss 3.8566 | lr 2.63e-03 | grad 0.2047 +2026-04-09 21:07:16 - INFO - parrotllm.training - step 1830 | epoch 0 | loss 3.7982 | lr 2.63e-03 | grad 0.1676 +2026-04-09 21:07:19 - INFO - parrotllm.training - step 1840 | epoch 0 | loss 3.8854 | lr 2.63e-03 | grad 0.1650 +2026-04-09 21:07:22 - INFO - parrotllm.training - step 1850 | epoch 0 | loss 3.9154 | lr 2.63e-03 | grad 0.1694 +2026-04-09 21:07:25 - INFO - parrotllm.training - step 1860 | epoch 0 | loss 3.9097 | lr 2.63e-03 | grad 0.1977 +2026-04-09 21:07:28 - INFO - parrotllm.training - step 1870 | epoch 0 | loss 3.9063 | lr 2.63e-03 | grad 0.1699 +2026-04-09 21:07:31 - INFO - parrotllm.training - step 1880 | epoch 0 | loss 3.8375 | lr 2.63e-03 | grad 0.1673 +2026-04-09 21:07:34 - INFO - parrotllm.training - step 1890 | epoch 0 | loss 3.7477 | lr 2.63e-03 | grad 0.1681 +2026-04-09 21:07:37 - INFO - parrotllm.training - step 1900 | epoch 0 | loss 3.8881 | lr 2.63e-03 | grad 0.2216 +2026-04-09 21:07:40 - INFO - parrotllm.training - step 1910 | epoch 0 | loss 3.9595 | lr 2.63e-03 | grad 0.2011 +2026-04-09 21:07:43 - INFO - parrotllm.training - step 1920 | epoch 0 | loss 3.7530 | lr 2.63e-03 | grad 0.1654 +2026-04-09 21:07:47 - INFO - parrotllm.training - step 1930 | epoch 0 | loss 3.8422 | lr 2.63e-03 | grad 0.1840 +2026-04-09 21:07:50 - INFO - parrotllm.training - step 1940 | epoch 0 | loss 3.6920 | lr 2.63e-03 | grad 0.1712 +2026-04-09 21:07:53 - INFO - parrotllm.training - step 1950 | epoch 0 | loss 3.6856 | lr 2.63e-03 | grad 0.1652 +2026-04-09 21:07:56 - INFO - parrotllm.training - step 1960 | epoch 0 | loss 3.8612 | lr 2.63e-03 | grad 0.1843 +2026-04-09 21:07:59 - INFO - parrotllm.training - step 1970 | epoch 0 | loss 3.8167 | lr 2.63e-03 | grad 0.1863 +2026-04-09 21:08:02 - INFO - parrotllm.training - step 1980 | epoch 0 | loss 3.8358 | lr 2.63e-03 | grad 0.1773 +2026-04-09 21:08:05 - INFO - parrotllm.training - step 1990 | epoch 0 | loss 3.8780 | lr 2.63e-03 | grad 0.2071 +2026-04-09 21:08:08 - INFO - parrotllm.training - step 2000 | epoch 0 | loss 3.8238 | lr 2.63e-03 | grad 0.1751 +2026-04-09 21:08:08 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:08:08 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:08:11 - INFO - parrotllm.training - Train: loss=3.8238, ppl=45.78 +2026-04-09 21:08:11 - INFO - parrotllm.training - Val: loss=3.7550, ppl=42.73 +2026-04-09 21:08:11 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:08:12 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p7550_epoch_0000_step_0002000.pt +2026-04-09 21:08:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:08:16 - INFO - parrotllm.training - step 2010 | epoch 0 | loss 3.7377 | lr 2.63e-03 | grad 0.1594 +2026-04-09 21:08:19 - INFO - parrotllm.training - step 2020 | epoch 0 | loss 3.8157 | lr 2.63e-03 | grad 0.1955 +2026-04-09 21:08:22 - INFO - parrotllm.training - step 2030 | epoch 0 | loss 3.8375 | lr 2.63e-03 | grad 0.1637 +2026-04-09 21:08:25 - INFO - parrotllm.training - step 2040 | epoch 0 | loss 3.8820 | lr 2.63e-03 | grad 0.1625 +2026-04-09 21:08:28 - INFO - parrotllm.training - step 2050 | epoch 0 | loss 3.8608 | lr 2.63e-03 | grad 0.1660 +2026-04-09 21:08:31 - INFO - parrotllm.training - step 2060 | epoch 0 | loss 3.7606 | lr 2.63e-03 | grad 0.1678 +2026-04-09 21:08:34 - INFO - parrotllm.training - step 2070 | epoch 0 | loss 3.8014 | lr 2.63e-03 | grad 0.1882 +2026-04-09 21:08:37 - INFO - parrotllm.training - step 2080 | epoch 0 | loss 3.7938 | lr 2.63e-03 | grad 0.1738 +2026-04-09 21:08:40 - INFO - parrotllm.training - step 2090 | epoch 0 | loss 3.8426 | lr 2.63e-03 | grad 0.1571 +2026-04-09 21:08:43 - INFO - parrotllm.training - step 2100 | epoch 0 | loss 3.9352 | lr 2.63e-03 | grad 0.1619 +2026-04-09 21:08:46 - INFO - parrotllm.training - step 2110 | epoch 0 | loss 3.7642 | lr 2.63e-03 | grad 0.1687 +2026-04-09 21:08:49 - INFO - parrotllm.training - step 2120 | epoch 0 | loss 3.7757 | lr 2.63e-03 | grad 0.1733 +2026-04-09 21:08:52 - INFO - parrotllm.training - step 2130 | epoch 0 | loss 3.7347 | lr 2.63e-03 | grad 0.1509 +2026-04-09 21:08:55 - INFO - parrotllm.training - step 2140 | epoch 0 | loss 3.9008 | lr 2.63e-03 | grad 0.1777 +2026-04-09 21:08:58 - INFO - parrotllm.training - step 2150 | epoch 0 | loss 3.7257 | lr 2.63e-03 | grad 0.1719 +2026-04-09 21:09:01 - INFO - parrotllm.training - step 2160 | epoch 0 | loss 3.7437 | lr 2.63e-03 | grad 0.1841 +2026-04-09 21:09:04 - INFO - parrotllm.training - step 2170 | epoch 0 | loss 3.6669 | lr 2.63e-03 | grad 0.1638 +2026-04-09 21:09:07 - INFO - parrotllm.training - step 2180 | epoch 0 | loss 3.7600 | lr 2.63e-03 | grad 0.1571 +2026-04-09 21:09:11 - INFO - parrotllm.training - step 2190 | epoch 0 | loss 3.8084 | lr 2.63e-03 | grad 0.1942 +2026-04-09 21:09:14 - INFO - parrotllm.training - step 2200 | epoch 0 | loss 3.7357 | lr 2.63e-03 | grad 0.1577 +2026-04-09 21:09:17 - INFO - parrotllm.training - step 2210 | epoch 0 | loss 3.7422 | lr 2.63e-03 | grad 0.1613 +2026-04-09 21:09:20 - INFO - parrotllm.training - step 2220 | epoch 0 | loss 3.7315 | lr 2.63e-03 | grad 0.1724 +2026-04-09 21:09:23 - INFO - parrotllm.training - step 2230 | epoch 0 | loss 3.8986 | lr 2.63e-03 | grad 0.1992 +2026-04-09 21:09:26 - INFO - parrotllm.training - step 2240 | epoch 0 | loss 3.6898 | lr 2.63e-03 | grad 0.1631 +2026-04-09 21:09:29 - INFO - parrotllm.training - step 2250 | epoch 0 | loss 3.7999 | lr 2.63e-03 | grad 0.1554 +2026-04-09 21:09:32 - INFO - parrotllm.training - step 2260 | epoch 0 | loss 3.5940 | lr 2.63e-03 | grad 0.1667 +2026-04-09 21:09:35 - INFO - parrotllm.training - step 2270 | epoch 0 | loss 3.7632 | lr 2.63e-03 | grad 0.1773 +2026-04-09 21:09:38 - INFO - parrotllm.training - step 2280 | epoch 0 | loss 3.7288 | lr 2.63e-03 | grad 0.1541 +2026-04-09 21:09:41 - INFO - parrotllm.training - step 2290 | epoch 0 | loss 3.8617 | lr 2.63e-03 | grad 0.2051 +2026-04-09 21:09:44 - INFO - parrotllm.training - step 2300 | epoch 0 | loss 3.7310 | lr 2.63e-03 | grad 0.1514 +2026-04-09 21:09:47 - INFO - parrotllm.training - step 2310 | epoch 0 | loss 3.6514 | lr 2.63e-03 | grad 0.1893 +2026-04-09 21:09:50 - INFO - parrotllm.training - step 2320 | epoch 0 | loss 3.7655 | lr 2.63e-03 | grad 0.1429 +2026-04-09 21:09:53 - INFO - parrotllm.training - step 2330 | epoch 0 | loss 3.6546 | lr 2.63e-03 | grad 0.1663 +2026-04-09 21:09:56 - INFO - parrotllm.training - step 2340 | epoch 0 | loss 3.7557 | lr 2.63e-03 | grad 0.1588 +2026-04-09 21:09:59 - INFO - parrotllm.training - step 2350 | epoch 0 | loss 3.8121 | lr 2.63e-03 | grad 0.1598 +2026-04-09 21:10:02 - INFO - parrotllm.training - step 2360 | epoch 0 | loss 3.7529 | lr 2.63e-03 | grad 0.1427 +2026-04-09 21:10:05 - INFO - parrotllm.training - step 2370 | epoch 0 | loss 3.8128 | lr 2.63e-03 | grad 0.1525 +2026-04-09 21:10:09 - INFO - parrotllm.training - step 2380 | epoch 0 | loss 3.8129 | lr 2.63e-03 | grad 0.1493 +2026-04-09 21:10:12 - INFO - parrotllm.training - step 2390 | epoch 0 | loss 3.7246 | lr 2.63e-03 | grad 0.2129 +2026-04-09 21:10:15 - INFO - parrotllm.training - step 2400 | epoch 0 | loss 3.7047 | lr 2.63e-03 | grad 0.1465 +2026-04-09 21:10:18 - INFO - parrotllm.training - step 2410 | epoch 0 | loss 3.7443 | lr 2.63e-03 | grad 0.1714 +2026-04-09 21:10:21 - INFO - parrotllm.training - step 2420 | epoch 0 | loss 3.7664 | lr 2.63e-03 | grad 0.1402 +2026-04-09 21:10:24 - INFO - parrotllm.training - step 2430 | epoch 0 | loss 3.7496 | lr 2.63e-03 | grad 0.1528 +2026-04-09 21:10:27 - INFO - parrotllm.training - step 2440 | epoch 0 | loss 3.7463 | lr 2.63e-03 | grad 0.1623 +2026-04-09 21:10:30 - INFO - parrotllm.training - step 2450 | epoch 0 | loss 3.8032 | lr 2.63e-03 | grad 0.1739 +2026-04-09 21:10:33 - INFO - parrotllm.training - step 2460 | epoch 0 | loss 3.6769 | lr 2.63e-03 | grad 0.1698 +2026-04-09 21:10:36 - INFO - parrotllm.training - step 2470 | epoch 0 | loss 3.8078 | lr 2.63e-03 | grad 0.1642 +2026-04-09 21:10:39 - INFO - parrotllm.training - step 2480 | epoch 0 | loss 3.7097 | lr 2.63e-03 | grad 0.1631 +2026-04-09 21:10:42 - INFO - parrotllm.training - step 2490 | epoch 0 | loss 3.7254 | lr 2.63e-03 | grad 0.1492 +2026-04-09 21:10:45 - INFO - parrotllm.training - step 2500 | epoch 0 | loss 3.6688 | lr 2.63e-03 | grad 0.1726 +2026-04-09 21:10:45 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:10:45 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:10:48 - INFO - parrotllm.training - Train: loss=3.6688, ppl=39.21 +2026-04-09 21:10:48 - INFO - parrotllm.training - Val: loss=3.6886, ppl=39.99 +2026-04-09 21:10:48 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:10:49 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p6886_epoch_0000_step_0002500.pt +2026-04-09 21:10:50 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:10:51 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0002500.pt +2026-04-09 21:10:55 - INFO - parrotllm.training - step 2510 | epoch 0 | loss 3.7657 | lr 2.63e-03 | grad 0.1574 +2026-04-09 21:10:58 - INFO - parrotllm.training - step 2520 | epoch 0 | loss 3.6438 | lr 2.63e-03 | grad 0.1721 +2026-04-09 21:11:01 - INFO - parrotllm.training - step 2530 | epoch 0 | loss 3.7347 | lr 2.63e-03 | grad 0.1727 +2026-04-09 21:11:04 - INFO - parrotllm.training - step 2540 | epoch 0 | loss 3.6656 | lr 2.63e-03 | grad 0.1512 +2026-04-09 21:11:07 - INFO - parrotllm.training - step 2550 | epoch 0 | loss 3.6988 | lr 2.63e-03 | grad 0.1707 +2026-04-09 21:11:10 - INFO - parrotllm.training - step 2560 | epoch 0 | loss 3.6521 | lr 2.63e-03 | grad 0.1605 +2026-04-09 21:11:13 - INFO - parrotllm.training - step 2570 | epoch 0 | loss 3.7339 | lr 2.63e-03 | grad 0.1659 +2026-04-09 21:11:16 - INFO - parrotllm.training - step 2580 | epoch 0 | loss 3.7464 | lr 2.63e-03 | grad 0.1822 +2026-04-09 21:11:19 - INFO - parrotllm.training - step 2590 | epoch 0 | loss 3.8766 | lr 2.63e-03 | grad 0.1485 +2026-04-09 21:11:23 - INFO - parrotllm.training - step 2600 | epoch 0 | loss 3.7301 | lr 2.63e-03 | grad 0.1548 +2026-04-09 21:11:26 - INFO - parrotllm.training - step 2610 | epoch 0 | loss 3.6877 | lr 2.63e-03 | grad 0.1602 +2026-04-09 21:11:29 - INFO - parrotllm.training - step 2620 | epoch 0 | loss 3.5787 | lr 2.63e-03 | grad 0.1651 +2026-04-09 21:11:32 - INFO - parrotllm.training - step 2630 | epoch 0 | loss 3.7781 | lr 2.63e-03 | grad 0.1629 +2026-04-09 21:11:35 - INFO - parrotllm.training - step 2640 | epoch 0 | loss 3.6658 | lr 2.63e-03 | grad 0.1448 +2026-04-09 21:11:38 - INFO - parrotllm.training - step 2650 | epoch 0 | loss 3.8100 | lr 2.63e-03 | grad 0.1653 +2026-04-09 21:11:41 - INFO - parrotllm.training - step 2660 | epoch 0 | loss 3.6855 | lr 2.63e-03 | grad 0.1741 +2026-04-09 21:11:44 - INFO - parrotllm.training - step 2670 | epoch 0 | loss 3.7711 | lr 2.63e-03 | grad 0.1485 +2026-04-09 21:11:47 - INFO - parrotllm.training - step 2680 | epoch 0 | loss 3.6807 | lr 2.63e-03 | grad 0.1588 +2026-04-09 21:11:50 - INFO - parrotllm.training - step 2690 | epoch 0 | loss 3.7364 | lr 2.63e-03 | grad 0.1905 +2026-04-09 21:11:53 - INFO - parrotllm.training - step 2700 | epoch 0 | loss 3.7071 | lr 2.63e-03 | grad 0.1695 +2026-04-09 21:11:56 - INFO - parrotllm.training - step 2710 | epoch 0 | loss 3.6922 | lr 2.63e-03 | grad 0.1468 +2026-04-09 21:11:59 - INFO - parrotllm.training - step 2720 | epoch 0 | loss 3.7410 | lr 2.63e-03 | grad 0.1738 +2026-04-09 21:12:02 - INFO - parrotllm.training - step 2730 | epoch 0 | loss 3.7294 | lr 2.63e-03 | grad 0.1492 +2026-04-09 21:12:05 - INFO - parrotllm.training - step 2740 | epoch 0 | loss 3.6398 | lr 2.63e-03 | grad 0.1603 +2026-04-09 21:12:08 - INFO - parrotllm.training - step 2750 | epoch 0 | loss 3.6812 | lr 2.63e-03 | grad 0.1695 +2026-04-09 21:12:11 - INFO - parrotllm.training - step 2760 | epoch 0 | loss 3.7837 | lr 2.63e-03 | grad 0.1549 +2026-04-09 21:12:14 - INFO - parrotllm.training - step 2770 | epoch 0 | loss 3.7276 | lr 2.63e-03 | grad 0.1445 +2026-04-09 21:12:17 - INFO - parrotllm.training - step 2780 | epoch 0 | loss 3.7563 | lr 2.63e-03 | grad 0.2181 +2026-04-09 21:12:21 - INFO - parrotllm.training - step 2790 | epoch 0 | loss 3.7809 | lr 2.63e-03 | grad 0.1582 +2026-04-09 21:12:24 - INFO - parrotllm.training - step 2800 | epoch 0 | loss 3.6880 | lr 2.63e-03 | grad 0.1377 +2026-04-09 21:12:27 - INFO - parrotllm.training - step 2810 | epoch 0 | loss 3.6957 | lr 2.63e-03 | grad 0.1463 +2026-04-09 21:12:30 - INFO - parrotllm.training - step 2820 | epoch 0 | loss 3.6497 | lr 2.63e-03 | grad 0.1675 +2026-04-09 21:12:33 - INFO - parrotllm.training - step 2830 | epoch 0 | loss 3.7150 | lr 2.63e-03 | grad 0.1906 +2026-04-09 21:12:36 - INFO - parrotllm.training - step 2840 | epoch 0 | loss 3.7512 | lr 2.63e-03 | grad 0.1650 +2026-04-09 21:12:39 - INFO - parrotllm.training - step 2850 | epoch 0 | loss 3.6350 | lr 2.63e-03 | grad 0.1526 +2026-04-09 21:12:42 - INFO - parrotllm.training - step 2860 | epoch 0 | loss 3.6283 | lr 2.63e-03 | grad 0.1244 +2026-04-09 21:12:45 - INFO - parrotllm.training - step 2870 | epoch 0 | loss 3.7159 | lr 2.63e-03 | grad 0.1891 +2026-04-09 21:12:48 - INFO - parrotllm.training - step 2880 | epoch 0 | loss 3.8224 | lr 2.63e-03 | grad 0.1554 +2026-04-09 21:12:51 - INFO - parrotllm.training - step 2890 | epoch 0 | loss 3.7465 | lr 2.63e-03 | grad 0.1721 +2026-04-09 21:12:54 - INFO - parrotllm.training - step 2900 | epoch 0 | loss 3.6996 | lr 2.63e-03 | grad 0.1591 +2026-04-09 21:12:57 - INFO - parrotllm.training - step 2910 | epoch 0 | loss 3.6344 | lr 2.63e-03 | grad 0.1521 +2026-04-09 21:13:00 - INFO - parrotllm.training - step 2920 | epoch 0 | loss 3.6075 | lr 2.63e-03 | grad 0.1431 +2026-04-09 21:13:03 - INFO - parrotllm.training - step 2930 | epoch 0 | loss 3.7568 | lr 2.63e-03 | grad 0.1520 +2026-04-09 21:13:06 - INFO - parrotllm.training - step 2940 | epoch 0 | loss 3.6712 | lr 2.63e-03 | grad 0.1870 +2026-04-09 21:13:09 - INFO - parrotllm.training - step 2950 | epoch 0 | loss 3.6501 | lr 2.63e-03 | grad 0.1619 +2026-04-09 21:13:13 - INFO - parrotllm.training - step 2960 | epoch 0 | loss 3.7417 | lr 2.63e-03 | grad 0.1686 +2026-04-09 21:13:16 - INFO - parrotllm.training - step 2970 | epoch 0 | loss 3.6663 | lr 2.63e-03 | grad 0.1627 +2026-04-09 21:13:19 - INFO - parrotllm.training - step 2980 | epoch 0 | loss 3.7288 | lr 2.63e-03 | grad 0.1530 +2026-04-09 21:13:22 - INFO - parrotllm.training - step 2990 | epoch 0 | loss 3.6970 | lr 2.63e-03 | grad 0.1644 +2026-04-09 21:13:25 - INFO - parrotllm.training - step 3000 | epoch 0 | loss 3.7678 | lr 2.63e-03 | grad 0.1739 +2026-04-09 21:13:25 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:13:25 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:13:28 - INFO - parrotllm.training - Train: loss=3.7678, ppl=43.28 +2026-04-09 21:13:28 - INFO - parrotllm.training - Val: loss=3.6427, ppl=38.19 +2026-04-09 21:13:28 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:13:29 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p6427_epoch_0000_step_0003000.pt +2026-04-09 21:13:29 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:13:32 - INFO - parrotllm.training - step 3010 | epoch 0 | loss 3.7653 | lr 2.63e-03 | grad 0.1709 +2026-04-09 21:13:36 - INFO - parrotllm.training - step 3020 | epoch 0 | loss 3.7349 | lr 2.63e-03 | grad 0.1634 +2026-04-09 21:13:39 - INFO - parrotllm.training - step 3030 | epoch 0 | loss 3.6699 | lr 2.63e-03 | grad 0.1580 +2026-04-09 21:13:42 - INFO - parrotllm.training - step 3040 | epoch 0 | loss 3.7583 | lr 2.63e-03 | grad 0.1363 +2026-04-09 21:13:45 - INFO - parrotllm.training - step 3050 | epoch 0 | loss 3.6753 | lr 2.63e-03 | grad 0.1462 +2026-04-09 21:13:48 - INFO - parrotllm.training - step 3060 | epoch 0 | loss 3.6552 | lr 2.63e-03 | grad 0.1368 +2026-04-09 21:13:51 - INFO - parrotllm.training - step 3070 | epoch 0 | loss 3.6180 | lr 2.63e-03 | grad 0.1733 +2026-04-09 21:13:54 - INFO - parrotllm.training - step 3080 | epoch 0 | loss 3.6665 | lr 2.63e-03 | grad 0.1851 +2026-04-09 21:13:57 - INFO - parrotllm.training - step 3090 | epoch 0 | loss 3.6158 | lr 2.63e-03 | grad 0.1618 +2026-04-09 21:14:00 - INFO - parrotllm.training - step 3100 | epoch 0 | loss 3.5985 | lr 2.63e-03 | grad 0.1592 +2026-04-09 21:14:03 - INFO - parrotllm.training - step 3110 | epoch 0 | loss 3.7877 | lr 2.63e-03 | grad 0.1484 +2026-04-09 21:14:06 - INFO - parrotllm.training - step 3120 | epoch 0 | loss 3.6751 | lr 2.63e-03 | grad 0.1548 +2026-04-09 21:14:09 - INFO - parrotllm.training - step 3130 | epoch 0 | loss 3.5962 | lr 2.63e-03 | grad 0.1533 +2026-04-09 21:14:12 - INFO - parrotllm.training - step 3140 | epoch 0 | loss 3.7436 | lr 2.63e-03 | grad 0.1615 +2026-04-09 21:14:15 - INFO - parrotllm.training - step 3150 | epoch 0 | loss 3.6221 | lr 2.63e-03 | grad 0.1604 +2026-04-09 21:14:18 - INFO - parrotllm.training - step 3160 | epoch 0 | loss 3.6741 | lr 2.63e-03 | grad 0.1469 +2026-04-09 21:14:21 - INFO - parrotllm.training - step 3170 | epoch 0 | loss 3.6341 | lr 2.63e-03 | grad 0.1708 +2026-04-09 21:14:24 - INFO - parrotllm.training - step 3180 | epoch 0 | loss 3.6837 | lr 2.63e-03 | grad 0.1460 +2026-04-09 21:14:28 - INFO - parrotllm.training - step 3190 | epoch 0 | loss 3.6714 | lr 2.63e-03 | grad 0.1706 +2026-04-09 21:14:31 - INFO - parrotllm.training - step 3200 | epoch 0 | loss 3.6481 | lr 2.63e-03 | grad 0.1739 +2026-04-09 21:14:34 - INFO - parrotllm.training - step 3210 | epoch 0 | loss 3.6443 | lr 2.63e-03 | grad 0.1536 +2026-04-09 21:14:37 - INFO - parrotllm.training - step 3220 | epoch 0 | loss 3.7641 | lr 2.63e-03 | grad 0.1666 +2026-04-09 21:14:40 - INFO - parrotllm.training - step 3230 | epoch 0 | loss 3.6281 | lr 2.63e-03 | grad 0.1361 +2026-04-09 21:14:43 - INFO - parrotllm.training - step 3240 | epoch 0 | loss 3.6542 | lr 2.63e-03 | grad 0.1793 +2026-04-09 21:14:46 - INFO - parrotllm.training - step 3250 | epoch 0 | loss 3.8321 | lr 2.63e-03 | grad 0.1557 +2026-04-09 21:14:49 - INFO - parrotllm.training - step 3260 | epoch 0 | loss 3.7504 | lr 2.63e-03 | grad 0.1580 +2026-04-09 21:14:52 - INFO - parrotllm.training - step 3270 | epoch 0 | loss 3.6294 | lr 2.63e-03 | grad 0.1455 +2026-04-09 21:14:55 - INFO - parrotllm.training - step 3280 | epoch 0 | loss 3.6668 | lr 2.63e-03 | grad 0.1575 +2026-04-09 21:14:58 - INFO - parrotllm.training - step 3290 | epoch 0 | loss 3.7776 | lr 2.63e-03 | grad 0.1826 +2026-04-09 21:15:01 - INFO - parrotllm.training - step 3300 | epoch 0 | loss 3.7297 | lr 2.63e-03 | grad 0.1640 +2026-04-09 21:15:04 - INFO - parrotllm.training - step 3310 | epoch 0 | loss 3.7239 | lr 2.63e-03 | grad 0.1838 +2026-04-09 21:15:07 - INFO - parrotllm.training - step 3320 | epoch 0 | loss 3.7124 | lr 2.63e-03 | grad 0.1731 +2026-04-09 21:15:10 - INFO - parrotllm.training - step 3330 | epoch 0 | loss 3.6504 | lr 2.63e-03 | grad 0.1746 +2026-04-09 21:15:13 - INFO - parrotllm.training - step 3340 | epoch 0 | loss 3.6762 | lr 2.63e-03 | grad 0.1550 +2026-04-09 21:15:16 - INFO - parrotllm.training - step 3350 | epoch 0 | loss 3.6259 | lr 2.63e-03 | grad 0.1569 +2026-04-09 21:15:19 - INFO - parrotllm.training - step 3360 | epoch 0 | loss 3.6563 | lr 2.63e-03 | grad 0.1530 +2026-04-09 21:15:23 - INFO - parrotllm.training - step 3370 | epoch 0 | loss 3.7894 | lr 2.63e-03 | grad 0.1379 +2026-04-09 21:15:26 - INFO - parrotllm.training - step 3380 | epoch 0 | loss 3.6267 | lr 2.63e-03 | grad 0.1565 +2026-04-09 21:15:29 - INFO - parrotllm.training - step 3390 | epoch 0 | loss 3.7947 | lr 2.63e-03 | grad 0.1741 +2026-04-09 21:15:32 - INFO - parrotllm.training - step 3400 | epoch 0 | loss 3.6693 | lr 2.63e-03 | grad 0.1420 +2026-04-09 21:15:35 - INFO - parrotllm.training - step 3410 | epoch 0 | loss 3.6616 | lr 2.63e-03 | grad 0.1513 +2026-04-09 21:15:38 - INFO - parrotllm.training - step 3420 | epoch 0 | loss 3.6723 | lr 2.63e-03 | grad 0.1529 +2026-04-09 21:15:41 - INFO - parrotllm.training - step 3430 | epoch 0 | loss 3.6776 | lr 2.63e-03 | grad 0.1652 +2026-04-09 21:15:44 - INFO - parrotllm.training - step 3440 | epoch 0 | loss 3.7129 | lr 2.63e-03 | grad 0.1576 +2026-04-09 21:15:47 - INFO - parrotllm.training - step 3450 | epoch 0 | loss 3.6925 | lr 2.63e-03 | grad 0.1735 +2026-04-09 21:15:50 - INFO - parrotllm.training - step 3460 | epoch 0 | loss 3.7323 | lr 2.63e-03 | grad 0.1770 +2026-04-09 21:15:53 - INFO - parrotllm.training - step 3470 | epoch 0 | loss 3.5392 | lr 2.63e-03 | grad 0.1542 +2026-04-09 21:15:56 - INFO - parrotllm.training - step 3480 | epoch 0 | loss 3.6568 | lr 2.63e-03 | grad 0.1563 +2026-04-09 21:15:59 - INFO - parrotllm.training - step 3490 | epoch 0 | loss 3.6222 | lr 2.63e-03 | grad 0.1864 +2026-04-09 21:16:02 - INFO - parrotllm.training - step 3500 | epoch 0 | loss 3.7263 | lr 2.63e-03 | grad 0.1822 +2026-04-09 21:16:02 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:16:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:16:05 - INFO - parrotllm.training - Train: loss=3.7263, ppl=41.53 +2026-04-09 21:16:05 - INFO - parrotllm.training - Val: loss=3.6163, ppl=37.20 +2026-04-09 21:16:05 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:16:06 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p6163_epoch_0000_step_0003500.pt +2026-04-09 21:16:07 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:16:10 - INFO - parrotllm.training - step 3510 | epoch 0 | loss 3.7635 | lr 2.63e-03 | grad 0.1678 +2026-04-09 21:16:13 - INFO - parrotllm.training - step 3520 | epoch 0 | loss 3.5930 | lr 2.63e-03 | grad 0.1414 +2026-04-09 21:16:16 - INFO - parrotllm.training - step 3530 | epoch 0 | loss 3.6398 | lr 2.63e-03 | grad 0.1527 +2026-04-09 21:16:19 - INFO - parrotllm.training - step 3540 | epoch 0 | loss 3.7074 | lr 2.63e-03 | grad 0.1754 +2026-04-09 21:16:22 - INFO - parrotllm.training - step 3550 | epoch 0 | loss 3.5297 | lr 2.63e-03 | grad 0.1602 +2026-04-09 21:16:25 - INFO - parrotllm.training - step 3560 | epoch 0 | loss 3.7537 | lr 2.63e-03 | grad 0.1777 +2026-04-09 21:16:28 - INFO - parrotllm.training - step 3570 | epoch 0 | loss 3.6817 | lr 2.63e-03 | grad 0.1469 +2026-04-09 21:16:31 - INFO - parrotllm.training - step 3580 | epoch 0 | loss 3.6355 | lr 2.63e-03 | grad 0.1608 +2026-04-09 21:16:34 - INFO - parrotllm.training - step 3590 | epoch 0 | loss 3.6963 | lr 2.63e-03 | grad 0.1750 +2026-04-09 21:16:37 - INFO - parrotllm.training - step 3600 | epoch 0 | loss 3.6234 | lr 2.63e-03 | grad 0.1732 +2026-04-09 21:16:41 - INFO - parrotllm.training - step 3610 | epoch 0 | loss 3.7208 | lr 2.63e-03 | grad 0.1672 +2026-04-09 21:16:44 - INFO - parrotllm.training - step 3620 | epoch 0 | loss 3.6120 | lr 2.63e-03 | grad 0.1670 +2026-04-09 21:16:47 - INFO - parrotllm.training - step 3630 | epoch 0 | loss 3.6236 | lr 2.63e-03 | grad 0.1454 +2026-04-09 21:16:50 - INFO - parrotllm.training - step 3640 | epoch 0 | loss 3.6800 | lr 2.63e-03 | grad 0.1626 +2026-04-09 21:16:53 - INFO - parrotllm.training - step 3650 | epoch 0 | loss 3.6869 | lr 2.63e-03 | grad 0.1759 +2026-04-09 21:16:56 - INFO - parrotllm.training - step 3660 | epoch 0 | loss 3.7615 | lr 2.63e-03 | grad 0.1522 +2026-04-09 21:16:59 - INFO - parrotllm.training - step 3670 | epoch 0 | loss 3.6561 | lr 2.63e-03 | grad 0.1512 +2026-04-09 21:17:02 - INFO - parrotllm.training - step 3680 | epoch 0 | loss 3.6198 | lr 2.63e-03 | grad 0.1396 +2026-04-09 21:17:05 - INFO - parrotllm.training - step 3690 | epoch 0 | loss 3.5825 | lr 2.63e-03 | grad 0.1641 +2026-04-09 21:17:08 - INFO - parrotllm.training - step 3700 | epoch 0 | loss 3.5990 | lr 2.63e-03 | grad 0.1614 +2026-04-09 21:17:11 - INFO - parrotllm.training - step 3710 | epoch 0 | loss 3.7060 | lr 2.63e-03 | grad 0.1532 +2026-04-09 21:17:14 - INFO - parrotllm.training - step 3720 | epoch 0 | loss 3.6651 | lr 2.63e-03 | grad 0.1700 +2026-04-09 21:17:17 - INFO - parrotllm.training - step 3730 | epoch 0 | loss 3.6959 | lr 2.63e-03 | grad 0.1902 +2026-04-09 21:17:20 - INFO - parrotllm.training - step 3740 | epoch 0 | loss 3.6156 | lr 2.63e-03 | grad 0.1771 +2026-04-09 21:17:23 - INFO - parrotllm.training - step 3750 | epoch 0 | loss 3.7669 | lr 2.63e-03 | grad 0.1362 +2026-04-09 21:17:26 - INFO - parrotllm.training - step 3760 | epoch 0 | loss 3.6896 | lr 2.63e-03 | grad 0.1502 +2026-04-09 21:17:29 - INFO - parrotllm.training - step 3770 | epoch 0 | loss 3.6107 | lr 2.63e-03 | grad 0.1504 +2026-04-09 21:17:32 - INFO - parrotllm.training - step 3780 | epoch 0 | loss 3.6583 | lr 2.63e-03 | grad 0.1750 +2026-04-09 21:17:35 - INFO - parrotllm.training - step 3790 | epoch 0 | loss 3.7185 | lr 2.63e-03 | grad 0.1466 +2026-04-09 21:17:38 - INFO - parrotllm.training - step 3800 | epoch 0 | loss 3.7052 | lr 2.63e-03 | grad 0.1468 +2026-04-09 21:17:41 - INFO - parrotllm.training - step 3810 | epoch 0 | loss 3.5358 | lr 2.63e-03 | grad 0.1593 +2026-04-09 21:17:44 - INFO - parrotllm.training - step 3820 | epoch 0 | loss 3.6279 | lr 2.63e-03 | grad 0.1581 +2026-04-09 21:17:48 - INFO - parrotllm.training - step 3830 | epoch 0 | loss 3.6500 | lr 2.63e-03 | grad 0.1355 +2026-04-09 21:17:51 - INFO - parrotllm.training - step 3840 | epoch 0 | loss 3.6376 | lr 2.63e-03 | grad 0.1574 +2026-04-09 21:17:54 - INFO - parrotllm.training - step 3850 | epoch 0 | loss 3.5680 | lr 2.63e-03 | grad 0.1616 +2026-04-09 21:17:57 - INFO - parrotllm.training - step 3860 | epoch 0 | loss 3.7019 | lr 2.63e-03 | grad 0.1534 +2026-04-09 21:18:00 - INFO - parrotllm.training - step 3870 | epoch 0 | loss 3.5924 | lr 2.63e-03 | grad 0.1548 +2026-04-09 21:18:03 - INFO - parrotllm.training - step 3880 | epoch 0 | loss 3.5229 | lr 2.63e-03 | grad 0.1881 +2026-04-09 21:18:06 - INFO - parrotllm.training - step 3890 | epoch 0 | loss 3.5939 | lr 2.63e-03 | grad 0.1479 +2026-04-09 21:18:09 - INFO - parrotllm.training - step 3900 | epoch 0 | loss 3.6430 | lr 2.63e-03 | grad 0.1391 +2026-04-09 21:18:12 - INFO - parrotllm.training - step 3910 | epoch 0 | loss 3.6441 | lr 2.63e-03 | grad 0.1726 +2026-04-09 21:18:15 - INFO - parrotllm.training - step 3920 | epoch 0 | loss 3.6517 | lr 2.63e-03 | grad 0.1636 +2026-04-09 21:18:18 - INFO - parrotllm.training - step 3930 | epoch 0 | loss 3.7147 | lr 2.63e-03 | grad 0.1537 +2026-04-09 21:18:21 - INFO - parrotllm.training - step 3940 | epoch 0 | loss 3.7640 | lr 2.63e-03 | grad 0.1554 +2026-04-09 21:18:24 - INFO - parrotllm.training - step 3950 | epoch 0 | loss 3.5427 | lr 2.63e-03 | grad 0.1447 +2026-04-09 21:18:27 - INFO - parrotllm.training - step 3960 | epoch 0 | loss 3.6981 | lr 2.63e-03 | grad 0.1570 +2026-04-09 21:18:30 - INFO - parrotllm.training - step 3970 | epoch 0 | loss 3.7180 | lr 2.63e-03 | grad 0.1918 +2026-04-09 21:18:33 - INFO - parrotllm.training - step 3980 | epoch 0 | loss 3.6657 | lr 2.63e-03 | grad 0.1506 +2026-04-09 21:18:36 - INFO - parrotllm.training - step 3990 | epoch 0 | loss 3.5889 | lr 2.63e-03 | grad 0.1671 +2026-04-09 21:18:39 - INFO - parrotllm.training - step 4000 | epoch 0 | loss 3.6779 | lr 2.63e-03 | grad 0.1637 +2026-04-09 21:18:39 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:18:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:18:42 - INFO - parrotllm.training - Train: loss=3.6779, ppl=39.56 +2026-04-09 21:18:42 - INFO - parrotllm.training - Val: loss=3.5910, ppl=36.27 +2026-04-09 21:18:42 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:18:43 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5910_epoch_0000_step_0004000.pt +2026-04-09 21:18:44 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:18:47 - INFO - parrotllm.training - step 4010 | epoch 0 | loss 3.6711 | lr 2.63e-03 | grad 0.1956 +2026-04-09 21:18:50 - INFO - parrotllm.training - step 4020 | epoch 0 | loss 3.5827 | lr 2.63e-03 | grad 0.1539 +2026-04-09 21:18:53 - INFO - parrotllm.training - step 4030 | epoch 0 | loss 3.6870 | lr 2.63e-03 | grad 0.1632 +2026-04-09 21:18:56 - INFO - parrotllm.training - step 4040 | epoch 0 | loss 3.6525 | lr 2.63e-03 | grad 0.1650 +2026-04-09 21:18:59 - INFO - parrotllm.training - step 4050 | epoch 0 | loss 3.5033 | lr 2.63e-03 | grad 0.1794 +2026-04-09 21:19:02 - INFO - parrotllm.training - step 4060 | epoch 0 | loss 3.6702 | lr 2.63e-03 | grad 0.1483 +2026-04-09 21:19:05 - INFO - parrotllm.training - step 4070 | epoch 0 | loss 3.6469 | lr 2.63e-03 | grad 0.1524 +2026-04-09 21:19:09 - INFO - parrotllm.training - step 4080 | epoch 0 | loss 3.6640 | lr 2.63e-03 | grad 0.1566 +2026-04-09 21:19:12 - INFO - parrotllm.training - step 4090 | epoch 0 | loss 3.7168 | lr 2.63e-03 | grad 0.1742 +2026-04-09 21:19:15 - INFO - parrotllm.training - step 4100 | epoch 0 | loss 3.6025 | lr 2.63e-03 | grad 0.1513 +2026-04-09 21:19:18 - INFO - parrotllm.training - step 4110 | epoch 0 | loss 3.5856 | lr 2.63e-03 | grad 0.1599 +2026-04-09 21:19:21 - INFO - parrotllm.training - step 4120 | epoch 0 | loss 3.6014 | lr 2.63e-03 | grad 0.1698 +2026-04-09 21:19:24 - INFO - parrotllm.training - step 4130 | epoch 0 | loss 3.6453 | lr 2.63e-03 | grad 0.1648 +2026-04-09 21:19:27 - INFO - parrotllm.training - step 4140 | epoch 0 | loss 3.6432 | lr 2.63e-03 | grad 0.1536 +2026-04-09 21:19:30 - INFO - parrotllm.training - step 4150 | epoch 0 | loss 3.6212 | lr 2.63e-03 | grad 0.1734 +2026-04-09 21:19:33 - INFO - parrotllm.training - step 4160 | epoch 0 | loss 3.7035 | lr 2.63e-03 | grad 0.1543 +2026-04-09 21:19:36 - INFO - parrotllm.training - step 4170 | epoch 0 | loss 3.6564 | lr 2.63e-03 | grad 0.1685 +2026-04-09 21:19:39 - INFO - parrotllm.training - step 4180 | epoch 0 | loss 3.6322 | lr 2.63e-03 | grad 0.1462 +2026-04-09 21:19:42 - INFO - parrotllm.training - step 4190 | epoch 0 | loss 3.6932 | lr 2.63e-03 | grad 0.1525 +2026-04-09 21:19:45 - INFO - parrotllm.training - step 4200 | epoch 0 | loss 3.6107 | lr 2.63e-03 | grad 0.1473 +2026-04-09 21:19:48 - INFO - parrotllm.training - step 4210 | epoch 0 | loss 3.6769 | lr 2.63e-03 | grad 0.1483 +2026-04-09 21:19:51 - INFO - parrotllm.training - step 4220 | epoch 0 | loss 3.5667 | lr 2.63e-03 | grad 0.1471 +2026-04-09 21:19:54 - INFO - parrotllm.training - step 4230 | epoch 0 | loss 3.7050 | lr 2.63e-03 | grad 0.1596 +2026-04-09 21:19:57 - INFO - parrotllm.training - step 4240 | epoch 0 | loss 3.6700 | lr 2.63e-03 | grad 0.1547 +2026-04-09 21:20:00 - INFO - parrotllm.training - step 4250 | epoch 0 | loss 3.5578 | lr 2.63e-03 | grad 0.1665 +2026-04-09 21:20:03 - INFO - parrotllm.training - step 4260 | epoch 0 | loss 3.5834 | lr 2.63e-03 | grad 0.1490 +2026-04-09 21:20:06 - INFO - parrotllm.training - step 4270 | epoch 0 | loss 3.5972 | lr 2.63e-03 | grad 0.1431 +2026-04-09 21:20:10 - INFO - parrotllm.training - step 4280 | epoch 0 | loss 3.6305 | lr 2.63e-03 | grad 0.1418 +2026-04-09 21:20:13 - INFO - parrotllm.training - step 4290 | epoch 0 | loss 3.6281 | lr 2.63e-03 | grad 0.1651 +2026-04-09 21:20:16 - INFO - parrotllm.training - step 4300 | epoch 0 | loss 3.6454 | lr 2.63e-03 | grad 0.1446 +2026-04-09 21:20:19 - INFO - parrotllm.training - step 4310 | epoch 0 | loss 3.6810 | lr 2.63e-03 | grad 0.1753 +2026-04-09 21:20:22 - INFO - parrotllm.training - step 4320 | epoch 0 | loss 3.6205 | lr 2.63e-03 | grad 0.1447 +2026-04-09 21:20:25 - INFO - parrotllm.training - step 4330 | epoch 0 | loss 3.6202 | lr 2.63e-03 | grad 0.1567 +2026-04-09 21:20:28 - INFO - parrotllm.training - step 4340 | epoch 0 | loss 3.5048 | lr 2.63e-03 | grad 0.1568 +2026-04-09 21:20:31 - INFO - parrotllm.training - step 4350 | epoch 0 | loss 3.5593 | lr 2.63e-03 | grad 0.1777 +2026-04-09 21:20:34 - INFO - parrotllm.training - step 4360 | epoch 0 | loss 3.6710 | lr 2.63e-03 | grad 0.1694 +2026-04-09 21:20:37 - INFO - parrotllm.training - step 4370 | epoch 0 | loss 3.5992 | lr 2.63e-03 | grad 0.1552 +2026-04-09 21:20:40 - INFO - parrotllm.training - step 4380 | epoch 0 | loss 3.6329 | lr 2.63e-03 | grad 0.1670 +2026-04-09 21:20:43 - INFO - parrotllm.training - step 4390 | epoch 0 | loss 3.6516 | lr 2.63e-03 | grad 0.1763 +2026-04-09 21:20:46 - INFO - parrotllm.training - step 4400 | epoch 0 | loss 3.6604 | lr 2.63e-03 | grad 0.1531 +2026-04-09 21:20:49 - INFO - parrotllm.training - step 4410 | epoch 0 | loss 3.6375 | lr 2.63e-03 | grad 0.1963 +2026-04-09 21:20:52 - INFO - parrotllm.training - step 4420 | epoch 0 | loss 3.7126 | lr 2.63e-03 | grad 0.1607 +2026-04-09 21:20:55 - INFO - parrotllm.training - step 4430 | epoch 0 | loss 3.6892 | lr 2.63e-03 | grad 0.1617 +2026-04-09 21:20:58 - INFO - parrotllm.training - step 4440 | epoch 0 | loss 3.5860 | lr 2.63e-03 | grad 0.1419 +2026-04-09 21:21:01 - INFO - parrotllm.training - step 4450 | epoch 0 | loss 3.5598 | lr 2.63e-03 | grad 0.1623 +2026-04-09 21:21:04 - INFO - parrotllm.training - step 4460 | epoch 0 | loss 3.5912 | lr 2.63e-03 | grad 0.1787 +2026-04-09 21:21:08 - INFO - parrotllm.training - step 4470 | epoch 0 | loss 3.6709 | lr 2.63e-03 | grad 0.1471 +2026-04-09 21:21:11 - INFO - parrotllm.training - step 4480 | epoch 0 | loss 3.5421 | lr 2.63e-03 | grad 0.1548 +2026-04-09 21:21:14 - INFO - parrotllm.training - step 4490 | epoch 0 | loss 3.6676 | lr 2.63e-03 | grad 0.1632 +2026-04-09 21:21:17 - INFO - parrotllm.training - step 4500 | epoch 0 | loss 3.7244 | lr 2.63e-03 | grad 0.1568 +2026-04-09 21:21:17 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:21:17 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:21:20 - INFO - parrotllm.training - Train: loss=3.7244, ppl=41.45 +2026-04-09 21:21:20 - INFO - parrotllm.training - Val: loss=3.5757, ppl=35.72 +2026-04-09 21:21:20 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:21:21 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5757_epoch_0000_step_0004500.pt +2026-04-09 21:21:21 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:21:24 - INFO - parrotllm.training - step 4510 | epoch 0 | loss 3.6613 | lr 2.63e-03 | grad 0.1500 +2026-04-09 21:21:28 - INFO - parrotllm.training - step 4520 | epoch 0 | loss 3.7299 | lr 2.63e-03 | grad 0.1733 +2026-04-09 21:21:31 - INFO - parrotllm.training - step 4530 | epoch 0 | loss 3.5978 | lr 2.63e-03 | grad 0.1484 +2026-04-09 21:21:34 - INFO - parrotllm.training - step 4540 | epoch 0 | loss 3.5526 | lr 2.63e-03 | grad 0.1524 +2026-04-09 21:21:37 - INFO - parrotllm.training - step 4550 | epoch 0 | loss 3.6691 | lr 2.63e-03 | grad 0.1470 +2026-04-09 21:21:40 - INFO - parrotllm.training - step 4560 | epoch 0 | loss 3.6975 | lr 2.63e-03 | grad 0.1402 +2026-04-09 21:21:43 - INFO - parrotllm.training - step 4570 | epoch 0 | loss 3.6284 | lr 2.63e-03 | grad 0.1670 +2026-04-09 21:21:46 - INFO - parrotllm.training - step 4580 | epoch 0 | loss 3.5834 | lr 2.63e-03 | grad 0.1637 +2026-04-09 21:21:49 - INFO - parrotllm.training - step 4590 | epoch 0 | loss 3.7202 | lr 2.63e-03 | grad 0.1769 +2026-04-09 21:21:52 - INFO - parrotllm.training - step 4600 | epoch 0 | loss 3.5401 | lr 2.63e-03 | grad 0.1692 +2026-04-09 21:21:55 - INFO - parrotllm.training - step 4610 | epoch 0 | loss 3.5445 | lr 2.63e-03 | grad 0.1466 +2026-04-09 21:21:58 - INFO - parrotllm.training - step 4620 | epoch 0 | loss 3.6889 | lr 2.63e-03 | grad 0.1832 +2026-04-09 21:22:01 - INFO - parrotllm.training - step 4630 | epoch 0 | loss 3.6040 | lr 2.63e-03 | grad 0.1549 +2026-04-09 21:22:04 - INFO - parrotllm.training - step 4640 | epoch 0 | loss 3.6137 | lr 2.63e-03 | grad 0.1489 +2026-04-09 21:22:07 - INFO - parrotllm.training - step 4650 | epoch 0 | loss 3.5991 | lr 2.63e-03 | grad 0.1534 +2026-04-09 21:22:10 - INFO - parrotllm.training - step 4660 | epoch 0 | loss 3.5851 | lr 2.63e-03 | grad 0.1597 +2026-04-09 21:22:13 - INFO - parrotllm.training - step 4670 | epoch 0 | loss 3.6590 | lr 2.63e-03 | grad 0.1567 +2026-04-09 21:22:16 - INFO - parrotllm.training - step 4680 | epoch 0 | loss 3.6623 | lr 2.63e-03 | grad 0.1603 +2026-04-09 21:22:19 - INFO - parrotllm.training - step 4690 | epoch 0 | loss 3.6727 | lr 2.63e-03 | grad 0.1627 +2026-04-09 21:22:22 - INFO - parrotllm.training - step 4700 | epoch 0 | loss 3.6118 | lr 2.63e-03 | grad 0.1595 +2026-04-09 21:22:26 - INFO - parrotllm.training - step 4710 | epoch 0 | loss 3.6010 | lr 2.63e-03 | grad 0.1820 +2026-04-09 21:22:29 - INFO - parrotllm.training - step 4720 | epoch 0 | loss 3.7403 | lr 2.63e-03 | grad 0.1491 +2026-04-09 21:22:32 - INFO - parrotllm.training - step 4730 | epoch 0 | loss 3.5137 | lr 2.63e-03 | grad 0.1581 +2026-04-09 21:22:35 - INFO - parrotllm.training - step 4740 | epoch 0 | loss 3.6146 | lr 2.63e-03 | grad 0.1819 +2026-04-09 21:22:38 - INFO - parrotllm.training - step 4750 | epoch 0 | loss 3.7322 | lr 2.63e-03 | grad 0.1692 +2026-04-09 21:22:41 - INFO - parrotllm.training - step 4760 | epoch 0 | loss 3.6364 | lr 2.63e-03 | grad 0.1695 +2026-04-09 21:22:44 - INFO - parrotllm.training - step 4770 | epoch 0 | loss 3.5453 | lr 2.63e-03 | grad 0.1494 +2026-04-09 21:22:47 - INFO - parrotllm.training - step 4780 | epoch 0 | loss 3.5750 | lr 2.63e-03 | grad 0.1481 +2026-04-09 21:22:50 - INFO - parrotllm.training - step 4790 | epoch 0 | loss 3.5795 | lr 2.63e-03 | grad 0.1639 +2026-04-09 21:22:53 - INFO - parrotllm.training - step 4800 | epoch 0 | loss 3.6128 | lr 2.63e-03 | grad 0.1487 +2026-04-09 21:22:56 - INFO - parrotllm.training - step 4810 | epoch 0 | loss 3.5191 | lr 2.63e-03 | grad 0.1455 +2026-04-09 21:22:59 - INFO - parrotllm.training - step 4820 | epoch 0 | loss 3.5810 | lr 2.63e-03 | grad 0.1577 +2026-04-09 21:23:02 - INFO - parrotllm.training - step 4830 | epoch 0 | loss 3.6733 | lr 2.63e-03 | grad 0.1789 +2026-04-09 21:23:05 - INFO - parrotllm.training - step 4840 | epoch 0 | loss 3.5635 | lr 2.63e-03 | grad 0.1787 +2026-04-09 21:23:08 - INFO - parrotllm.training - step 4850 | epoch 0 | loss 3.5455 | lr 2.63e-03 | grad 0.1504 +2026-04-09 21:23:11 - INFO - parrotllm.training - step 4860 | epoch 0 | loss 3.5725 | lr 2.63e-03 | grad 0.1598 +2026-04-09 21:23:14 - INFO - parrotllm.training - step 4870 | epoch 0 | loss 3.6290 | lr 2.63e-03 | grad 0.1585 +2026-04-09 21:23:17 - INFO - parrotllm.training - step 4880 | epoch 0 | loss 3.5362 | lr 2.63e-03 | grad 0.1969 +2026-04-09 21:23:20 - INFO - parrotllm.training - step 4890 | epoch 0 | loss 3.6174 | lr 2.63e-03 | grad 0.1560 +2026-04-09 21:23:23 - INFO - parrotllm.training - step 4900 | epoch 0 | loss 3.5722 | lr 2.63e-03 | grad 0.1405 +2026-04-09 21:23:27 - INFO - parrotllm.training - step 4910 | epoch 0 | loss 3.5916 | lr 2.63e-03 | grad 0.1398 +2026-04-09 21:23:30 - INFO - parrotllm.training - step 4920 | epoch 0 | loss 3.7467 | lr 2.63e-03 | grad 0.1528 +2026-04-09 21:23:33 - INFO - parrotllm.training - step 4930 | epoch 0 | loss 3.5629 | lr 2.63e-03 | grad 0.1486 +2026-04-09 21:23:36 - INFO - parrotllm.training - step 4940 | epoch 0 | loss 3.6422 | lr 2.63e-03 | grad 0.1925 +2026-04-09 21:23:39 - INFO - parrotllm.training - step 4950 | epoch 0 | loss 3.7433 | lr 2.63e-03 | grad 0.1546 +2026-04-09 21:23:42 - INFO - parrotllm.training - step 4960 | epoch 0 | loss 3.5598 | lr 2.63e-03 | grad 0.1419 +2026-04-09 21:23:45 - INFO - parrotllm.training - step 4970 | epoch 0 | loss 3.6137 | lr 2.63e-03 | grad 0.1660 +2026-04-09 21:23:48 - INFO - parrotllm.training - step 4980 | epoch 0 | loss 3.6145 | lr 2.63e-03 | grad 0.1446 +2026-04-09 21:23:51 - INFO - parrotllm.training - step 4990 | epoch 0 | loss 3.5468 | lr 2.63e-03 | grad 0.1523 +2026-04-09 21:23:54 - INFO - parrotllm.training - step 5000 | epoch 0 | loss 3.6727 | lr 2.63e-03 | grad 0.1614 +2026-04-09 21:23:54 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:23:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:23:57 - INFO - parrotllm.training - Train: loss=3.6727, ppl=39.36 +2026-04-09 21:23:57 - INFO - parrotllm.training - Val: loss=3.5621, ppl=35.24 +2026-04-09 21:23:57 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:23:58 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5621_epoch_0000_step_0005000.pt +2026-04-09 21:23:59 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:24:00 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0005000.pt +2026-04-09 21:24:04 - INFO - parrotllm.training - step 5010 | epoch 0 | loss 3.6919 | lr 2.63e-03 | grad 0.1655 +2026-04-09 21:24:07 - INFO - parrotllm.training - step 5020 | epoch 0 | loss 3.6671 | lr 2.63e-03 | grad 0.1558 +2026-04-09 21:24:10 - INFO - parrotllm.training - step 5030 | epoch 0 | loss 3.6165 | lr 2.63e-03 | grad 0.1636 +2026-04-09 21:24:13 - INFO - parrotllm.training - step 5040 | epoch 0 | loss 3.4790 | lr 2.63e-03 | grad 0.1608 +2026-04-09 21:24:16 - INFO - parrotllm.training - step 5050 | epoch 0 | loss 3.5920 | lr 2.63e-03 | grad 0.1791 +2026-04-09 21:24:19 - INFO - parrotllm.training - step 5060 | epoch 0 | loss 3.6272 | lr 2.63e-03 | grad 0.1764 +2026-04-09 21:24:22 - INFO - parrotllm.training - step 5070 | epoch 0 | loss 3.5425 | lr 2.63e-03 | grad 0.1433 +2026-04-09 21:24:25 - INFO - parrotllm.training - step 5080 | epoch 0 | loss 3.5648 | lr 2.63e-03 | grad 0.1865 +2026-04-09 21:24:28 - INFO - parrotllm.training - step 5090 | epoch 0 | loss 3.5400 | lr 2.63e-03 | grad 0.1647 +2026-04-09 21:24:31 - INFO - parrotllm.training - step 5100 | epoch 0 | loss 3.7185 | lr 2.63e-03 | grad 0.1810 +2026-04-09 21:24:34 - INFO - parrotllm.training - step 5110 | epoch 0 | loss 3.5995 | lr 2.63e-03 | grad 0.1420 +2026-04-09 21:24:37 - INFO - parrotllm.training - step 5120 | epoch 0 | loss 3.5950 | lr 2.63e-03 | grad 0.1507 +2026-04-09 21:24:40 - INFO - parrotllm.training - step 5130 | epoch 0 | loss 3.6464 | lr 2.63e-03 | grad 0.1429 +2026-04-09 21:24:43 - INFO - parrotllm.training - step 5140 | epoch 0 | loss 3.6567 | lr 2.63e-03 | grad 0.1867 +2026-04-09 21:24:46 - INFO - parrotllm.training - step 5150 | epoch 0 | loss 3.5104 | lr 2.63e-03 | grad 0.1443 +2026-04-09 21:24:50 - INFO - parrotllm.training - step 5160 | epoch 0 | loss 3.5223 | lr 2.63e-03 | grad 0.1785 +2026-04-09 21:24:53 - INFO - parrotllm.training - step 5170 | epoch 0 | loss 3.5085 | lr 2.63e-03 | grad 0.1500 +2026-04-09 21:24:56 - INFO - parrotllm.training - step 5180 | epoch 0 | loss 3.5702 | lr 2.63e-03 | grad 0.1664 +2026-04-09 21:24:59 - INFO - parrotllm.training - step 5190 | epoch 0 | loss 3.5629 | lr 2.63e-03 | grad 0.1524 +2026-04-09 21:25:02 - INFO - parrotllm.training - step 5200 | epoch 0 | loss 3.6022 | lr 2.63e-03 | grad 0.1616 +2026-04-09 21:25:05 - INFO - parrotllm.training - step 5210 | epoch 0 | loss 3.5668 | lr 2.63e-03 | grad 0.1533 +2026-04-09 21:25:08 - INFO - parrotllm.training - step 5220 | epoch 0 | loss 3.5766 | lr 2.63e-03 | grad 0.1617 +2026-04-09 21:25:11 - INFO - parrotllm.training - step 5230 | epoch 0 | loss 3.6293 | lr 2.63e-03 | grad 0.1501 +2026-04-09 21:25:14 - INFO - parrotllm.training - step 5240 | epoch 0 | loss 3.5899 | lr 2.63e-03 | grad 0.1655 +2026-04-09 21:25:17 - INFO - parrotllm.training - step 5250 | epoch 0 | loss 3.5443 | lr 2.63e-03 | grad 0.1841 +2026-04-09 21:25:20 - INFO - parrotllm.training - step 5260 | epoch 0 | loss 3.5885 | lr 2.63e-03 | grad 0.1739 +2026-04-09 21:25:23 - INFO - parrotllm.training - step 5270 | epoch 0 | loss 3.6339 | lr 2.63e-03 | grad 0.1437 +2026-04-09 21:25:26 - INFO - parrotllm.training - step 5280 | epoch 0 | loss 3.5592 | lr 2.63e-03 | grad 0.1464 +2026-04-09 21:25:29 - INFO - parrotllm.training - step 5290 | epoch 0 | loss 3.4978 | lr 2.63e-03 | grad 0.1669 +2026-04-09 21:25:32 - INFO - parrotllm.training - step 5300 | epoch 0 | loss 3.5799 | lr 2.63e-03 | grad 0.1517 +2026-04-09 21:25:35 - INFO - parrotllm.training - step 5310 | epoch 0 | loss 3.5836 | lr 2.63e-03 | grad 0.1566 +2026-04-09 21:25:38 - INFO - parrotllm.training - step 5320 | epoch 0 | loss 3.4986 | lr 2.63e-03 | grad 0.1851 +2026-04-09 21:25:42 - INFO - parrotllm.training - step 5330 | epoch 0 | loss 3.6115 | lr 2.63e-03 | grad 0.1568 +2026-04-09 21:25:45 - INFO - parrotllm.training - step 5340 | epoch 0 | loss 3.5980 | lr 2.63e-03 | grad 0.1534 +2026-04-09 21:25:48 - INFO - parrotllm.training - step 5350 | epoch 0 | loss 3.5933 | lr 2.63e-03 | grad 0.1485 +2026-04-09 21:25:51 - INFO - parrotllm.training - step 5360 | epoch 0 | loss 3.5921 | lr 2.63e-03 | grad 0.1499 +2026-04-09 21:25:54 - INFO - parrotllm.training - step 5370 | epoch 0 | loss 3.6248 | lr 2.63e-03 | grad 0.1485 +2026-04-09 21:25:57 - INFO - parrotllm.training - step 5380 | epoch 0 | loss 3.6388 | lr 2.63e-03 | grad 0.1676 +2026-04-09 21:26:00 - INFO - parrotllm.training - step 5390 | epoch 0 | loss 3.5291 | lr 2.63e-03 | grad 0.1551 +2026-04-09 21:26:03 - INFO - parrotllm.training - step 5400 | epoch 0 | loss 3.6425 | lr 2.63e-03 | grad 0.1361 +2026-04-09 21:26:06 - INFO - parrotllm.training - step 5410 | epoch 0 | loss 3.5890 | lr 2.63e-03 | grad 0.1362 +2026-04-09 21:26:09 - INFO - parrotllm.training - step 5420 | epoch 0 | loss 3.5226 | lr 2.63e-03 | grad 0.2366 +2026-04-09 21:26:12 - INFO - parrotllm.training - step 5430 | epoch 0 | loss 3.6211 | lr 2.63e-03 | grad 0.1416 +2026-04-09 21:26:15 - INFO - parrotllm.training - step 5440 | epoch 0 | loss 3.6572 | lr 2.63e-03 | grad 0.1771 +2026-04-09 21:26:18 - INFO - parrotllm.training - step 5450 | epoch 0 | loss 3.6260 | lr 2.63e-03 | grad 0.1582 +2026-04-09 21:26:21 - INFO - parrotllm.training - step 5460 | epoch 0 | loss 3.6040 | lr 2.63e-03 | grad 0.1668 +2026-04-09 21:26:24 - INFO - parrotllm.training - step 5470 | epoch 0 | loss 3.6638 | lr 2.63e-03 | grad 0.1460 +2026-04-09 21:26:27 - INFO - parrotllm.training - step 5480 | epoch 0 | loss 3.5147 | lr 2.63e-03 | grad 0.1821 +2026-04-09 21:26:30 - INFO - parrotllm.training - step 5490 | epoch 0 | loss 3.6176 | lr 2.63e-03 | grad 0.1496 +2026-04-09 21:26:34 - INFO - parrotllm.training - step 5500 | epoch 0 | loss 3.6282 | lr 2.63e-03 | grad 0.1402 +2026-04-09 21:26:34 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:26:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:26:37 - INFO - parrotllm.training - Train: loss=3.6282, ppl=37.65 +2026-04-09 21:26:37 - INFO - parrotllm.training - Val: loss=3.5501, ppl=34.82 +2026-04-09 21:26:37 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:26:38 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5501_epoch_0000_step_0005500.pt +2026-04-09 21:26:38 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:26:41 - INFO - parrotllm.training - step 5510 | epoch 0 | loss 3.5237 | lr 2.63e-03 | grad 0.1644 +2026-04-09 21:26:44 - INFO - parrotllm.training - step 5520 | epoch 0 | loss 3.5647 | lr 2.63e-03 | grad 0.1481 +2026-04-09 21:26:47 - INFO - parrotllm.training - step 5530 | epoch 0 | loss 3.6132 | lr 2.63e-03 | grad 0.1608 +2026-04-09 21:26:51 - INFO - parrotllm.training - step 5540 | epoch 0 | loss 3.5091 | lr 2.63e-03 | grad 0.1683 +2026-04-09 21:26:54 - INFO - parrotllm.training - step 5550 | epoch 0 | loss 3.5198 | lr 2.63e-03 | grad 0.1502 +2026-04-09 21:26:57 - INFO - parrotllm.training - step 5560 | epoch 0 | loss 3.6628 | lr 2.63e-03 | grad 0.1725 +2026-04-09 21:27:00 - INFO - parrotllm.training - step 5570 | epoch 0 | loss 3.6404 | lr 2.63e-03 | grad 0.1554 +2026-04-09 21:27:03 - INFO - parrotllm.training - step 5580 | epoch 0 | loss 3.5347 | lr 2.63e-03 | grad 0.1695 +2026-04-09 21:27:06 - INFO - parrotllm.training - step 5590 | epoch 0 | loss 3.5408 | lr 2.63e-03 | grad 0.1242 +2026-04-09 21:27:09 - INFO - parrotllm.training - step 5600 | epoch 0 | loss 3.5453 | lr 2.63e-03 | grad 0.1381 +2026-04-09 21:27:12 - INFO - parrotllm.training - step 5610 | epoch 0 | loss 3.6891 | lr 2.63e-03 | grad 0.1643 +2026-04-09 21:27:15 - INFO - parrotllm.training - step 5620 | epoch 0 | loss 3.6332 | lr 2.63e-03 | grad 0.1749 +2026-04-09 21:27:18 - INFO - parrotllm.training - step 5630 | epoch 0 | loss 3.5063 | lr 2.63e-03 | grad 0.1588 +2026-04-09 21:27:21 - INFO - parrotllm.training - step 5640 | epoch 0 | loss 3.6367 | lr 2.63e-03 | grad 0.1503 +2026-04-09 21:27:24 - INFO - parrotllm.training - step 5650 | epoch 0 | loss 3.5863 | lr 2.63e-03 | grad 0.1685 +2026-04-09 21:27:27 - INFO - parrotllm.training - step 5660 | epoch 0 | loss 3.6716 | lr 2.63e-03 | grad 0.1528 +2026-04-09 21:27:30 - INFO - parrotllm.training - step 5670 | epoch 0 | loss 3.5861 | lr 2.63e-03 | grad 0.1564 +2026-04-09 21:27:33 - INFO - parrotllm.training - step 5680 | epoch 0 | loss 3.5612 | lr 2.63e-03 | grad 0.1667 +2026-04-09 21:27:36 - INFO - parrotllm.training - step 5690 | epoch 0 | loss 3.6039 | lr 2.63e-03 | grad 0.1691 +2026-04-09 21:27:39 - INFO - parrotllm.training - step 5700 | epoch 0 | loss 3.5030 | lr 2.63e-03 | grad 0.1683 +2026-04-09 21:27:43 - INFO - parrotllm.training - step 5710 | epoch 0 | loss 3.6099 | lr 2.63e-03 | grad 0.1721 +2026-04-09 21:27:46 - INFO - parrotllm.training - step 5720 | epoch 0 | loss 3.6278 | lr 2.63e-03 | grad 0.1633 +2026-04-09 21:27:49 - INFO - parrotllm.training - step 5730 | epoch 0 | loss 3.5473 | lr 2.63e-03 | grad 0.1466 +2026-04-09 21:27:52 - INFO - parrotllm.training - step 5740 | epoch 0 | loss 3.5494 | lr 2.63e-03 | grad 0.1780 +2026-04-09 21:27:55 - INFO - parrotllm.training - step 5750 | epoch 0 | loss 3.6783 | lr 2.63e-03 | grad 0.1389 +2026-04-09 21:27:58 - INFO - parrotllm.training - step 5760 | epoch 0 | loss 3.4816 | lr 2.63e-03 | grad 0.1707 +2026-04-09 21:28:01 - INFO - parrotllm.training - step 5770 | epoch 0 | loss 3.5466 | lr 2.63e-03 | grad 0.1583 +2026-04-09 21:28:04 - INFO - parrotllm.training - step 5780 | epoch 0 | loss 3.5682 | lr 2.63e-03 | grad 0.1670 +2026-04-09 21:28:07 - INFO - parrotllm.training - step 5790 | epoch 0 | loss 3.5844 | lr 2.63e-03 | grad 0.1647 +2026-04-09 21:28:10 - INFO - parrotllm.training - step 5800 | epoch 0 | loss 3.5572 | lr 2.63e-03 | grad 0.1824 +2026-04-09 21:28:13 - INFO - parrotllm.training - step 5810 | epoch 0 | loss 3.6086 | lr 2.63e-03 | grad 0.1442 +2026-04-09 21:28:16 - INFO - parrotllm.training - step 5820 | epoch 0 | loss 3.6392 | lr 2.63e-03 | grad 0.1641 +2026-04-09 21:28:19 - INFO - parrotllm.training - step 5830 | epoch 0 | loss 3.7033 | lr 2.63e-03 | grad 0.1538 +2026-04-09 21:28:22 - INFO - parrotllm.training - step 5840 | epoch 0 | loss 3.5726 | lr 2.63e-03 | grad 0.1617 +2026-04-09 21:28:25 - INFO - parrotllm.training - step 5850 | epoch 0 | loss 3.6094 | lr 2.63e-03 | grad 0.2161 +2026-04-09 21:28:28 - INFO - parrotllm.training - step 5860 | epoch 0 | loss 3.5477 | lr 2.63e-03 | grad 0.1597 +2026-04-09 21:28:31 - INFO - parrotllm.training - step 5870 | epoch 0 | loss 3.6866 | lr 2.63e-03 | grad 0.1581 +2026-04-09 21:28:34 - INFO - parrotllm.training - step 5880 | epoch 0 | loss 3.5533 | lr 2.63e-03 | grad 0.1328 +2026-04-09 21:28:38 - INFO - parrotllm.training - step 5890 | epoch 0 | loss 3.5652 | lr 2.63e-03 | grad 0.2198 +2026-04-09 21:28:41 - INFO - parrotllm.training - step 5900 | epoch 0 | loss 3.6371 | lr 2.63e-03 | grad 0.1452 +2026-04-09 21:28:44 - INFO - parrotllm.training - step 5910 | epoch 0 | loss 3.6004 | lr 2.63e-03 | grad 0.1498 +2026-04-09 21:28:47 - INFO - parrotllm.training - step 5920 | epoch 0 | loss 3.6845 | lr 2.63e-03 | grad 0.1687 +2026-04-09 21:28:50 - INFO - parrotllm.training - step 5930 | epoch 0 | loss 3.5724 | lr 2.63e-03 | grad 0.1711 +2026-04-09 21:28:53 - INFO - parrotllm.training - step 5940 | epoch 0 | loss 3.5059 | lr 2.63e-03 | grad 0.1512 +2026-04-09 21:28:56 - INFO - parrotllm.training - step 5950 | epoch 0 | loss 3.6508 | lr 2.63e-03 | grad 0.1647 +2026-04-09 21:28:59 - INFO - parrotllm.training - step 5960 | epoch 0 | loss 3.5788 | lr 2.63e-03 | grad 0.1648 +2026-04-09 21:29:02 - INFO - parrotllm.training - step 5970 | epoch 0 | loss 3.5033 | lr 2.63e-03 | grad 0.1480 +2026-04-09 21:29:05 - INFO - parrotllm.training - step 5980 | epoch 0 | loss 3.6418 | lr 2.63e-03 | grad 0.1397 +2026-04-09 21:29:08 - INFO - parrotllm.training - step 5990 | epoch 0 | loss 3.6367 | lr 2.63e-03 | grad 0.1570 +2026-04-09 21:29:11 - INFO - parrotllm.training - step 6000 | epoch 0 | loss 3.5934 | lr 2.63e-03 | grad 0.2044 +2026-04-09 21:29:11 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:29:11 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:29:14 - INFO - parrotllm.training - Train: loss=3.5934, ppl=36.36 +2026-04-09 21:29:14 - INFO - parrotllm.training - Val: loss=3.5405, ppl=34.48 +2026-04-09 21:29:14 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:29:15 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5405_epoch_0000_step_0006000.pt +2026-04-09 21:29:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:29:19 - INFO - parrotllm.training - step 6010 | epoch 0 | loss 3.5797 | lr 2.63e-03 | grad 0.1603 +2026-04-09 21:29:22 - INFO - parrotllm.training - step 6020 | epoch 0 | loss 3.6287 | lr 2.63e-03 | grad 0.1396 +2026-04-09 21:29:26 - INFO - parrotllm.training - step 6030 | epoch 0 | loss 3.5525 | lr 2.63e-03 | grad 0.1700 +2026-04-09 21:29:29 - INFO - parrotllm.training - step 6040 | epoch 0 | loss 3.6948 | lr 2.63e-03 | grad 0.1651 +2026-04-09 21:29:32 - INFO - parrotllm.training - step 6050 | epoch 0 | loss 3.6443 | lr 2.63e-03 | grad 0.1646 +2026-04-09 21:29:35 - INFO - parrotllm.training - step 6060 | epoch 0 | loss 3.5706 | lr 2.63e-03 | grad 0.1583 +2026-04-09 21:29:38 - INFO - parrotllm.training - step 6070 | epoch 0 | loss 3.5115 | lr 2.63e-03 | grad 0.1555 +2026-04-09 21:29:41 - INFO - parrotllm.training - step 6080 | epoch 0 | loss 3.5771 | lr 2.63e-03 | grad 0.1405 +2026-04-09 21:29:44 - INFO - parrotllm.training - step 6090 | epoch 0 | loss 3.6340 | lr 2.63e-03 | grad 0.1762 +2026-04-09 21:29:47 - INFO - parrotllm.training - step 6100 | epoch 0 | loss 3.5857 | lr 2.63e-03 | grad 0.1463 +2026-04-09 21:29:50 - INFO - parrotllm.training - step 6110 | epoch 0 | loss 3.5955 | lr 2.63e-03 | grad 0.1900 +2026-04-09 21:29:53 - INFO - parrotllm.training - step 6120 | epoch 0 | loss 3.6351 | lr 2.63e-03 | grad 0.1515 +2026-04-09 21:29:56 - INFO - parrotllm.training - step 6130 | epoch 0 | loss 3.6876 | lr 2.63e-03 | grad 0.1687 +2026-04-09 21:29:59 - INFO - parrotllm.training - step 6140 | epoch 0 | loss 3.5824 | lr 2.63e-03 | grad 0.1536 +2026-04-09 21:30:02 - INFO - parrotllm.training - step 6150 | epoch 0 | loss 3.5949 | lr 2.63e-03 | grad 0.1686 +2026-04-09 21:30:05 - INFO - parrotllm.training - step 6160 | epoch 0 | loss 3.5613 | lr 2.63e-03 | grad 0.1481 +2026-04-09 21:30:08 - INFO - parrotllm.training - step 6170 | epoch 0 | loss 3.6582 | lr 2.63e-03 | grad 0.1602 +2026-04-09 21:30:11 - INFO - parrotllm.training - step 6180 | epoch 0 | loss 3.6262 | lr 2.63e-03 | grad 0.1529 +2026-04-09 21:30:14 - INFO - parrotllm.training - step 6190 | epoch 0 | loss 3.5837 | lr 2.63e-03 | grad 0.1788 +2026-04-09 21:30:18 - INFO - parrotllm.training - step 6200 | epoch 0 | loss 3.5272 | lr 2.63e-03 | grad 0.1656 +2026-04-09 21:30:21 - INFO - parrotllm.training - step 6210 | epoch 0 | loss 3.6379 | lr 2.63e-03 | grad 0.1508 +2026-04-09 21:30:24 - INFO - parrotllm.training - step 6220 | epoch 0 | loss 3.5448 | lr 2.63e-03 | grad 0.1423 +2026-04-09 21:30:27 - INFO - parrotllm.training - step 6230 | epoch 0 | loss 3.6085 | lr 2.63e-03 | grad 0.1495 +2026-04-09 21:30:30 - INFO - parrotllm.training - step 6240 | epoch 0 | loss 3.6700 | lr 2.63e-03 | grad 0.1447 +2026-04-09 21:30:33 - INFO - parrotllm.training - step 6250 | epoch 0 | loss 3.6191 | lr 2.63e-03 | grad 0.1627 +2026-04-09 21:30:36 - INFO - parrotllm.training - step 6260 | epoch 0 | loss 3.5452 | lr 2.63e-03 | grad 0.1403 +2026-04-09 21:30:39 - INFO - parrotllm.training - step 6270 | epoch 0 | loss 3.5896 | lr 2.63e-03 | grad 0.1732 +2026-04-09 21:30:42 - INFO - parrotllm.training - step 6280 | epoch 0 | loss 3.6146 | lr 2.63e-03 | grad 0.1553 +2026-04-09 21:30:45 - INFO - parrotllm.training - step 6290 | epoch 0 | loss 3.5596 | lr 2.63e-03 | grad 0.1679 +2026-04-09 21:30:48 - INFO - parrotllm.training - step 6300 | epoch 0 | loss 3.5818 | lr 2.63e-03 | grad 0.1939 +2026-04-09 21:30:51 - INFO - parrotllm.training - step 6310 | epoch 0 | loss 3.5734 | lr 2.63e-03 | grad 0.1527 +2026-04-09 21:30:54 - INFO - parrotllm.training - step 6320 | epoch 0 | loss 3.5739 | lr 2.63e-03 | grad 0.1581 +2026-04-09 21:30:57 - INFO - parrotllm.training - step 6330 | epoch 0 | loss 3.6568 | lr 2.63e-03 | grad 0.1554 +2026-04-09 21:31:00 - INFO - parrotllm.training - step 6340 | epoch 0 | loss 3.5575 | lr 2.63e-03 | grad 0.1537 +2026-04-09 21:31:04 - INFO - parrotllm.training - step 6350 | epoch 0 | loss 3.6500 | lr 2.63e-03 | grad 0.1684 +2026-04-09 21:31:07 - INFO - parrotllm.training - step 6360 | epoch 0 | loss 3.5207 | lr 2.63e-03 | grad 0.1580 +2026-04-09 21:31:10 - INFO - parrotllm.training - step 6370 | epoch 0 | loss 3.5543 | lr 2.63e-03 | grad 0.1646 +2026-04-09 21:31:13 - INFO - parrotllm.training - step 6380 | epoch 0 | loss 3.5102 | lr 2.63e-03 | grad 0.1629 +2026-04-09 21:31:16 - INFO - parrotllm.training - step 6390 | epoch 0 | loss 3.5432 | lr 2.63e-03 | grad 0.1556 +2026-04-09 21:31:19 - INFO - parrotllm.training - step 6400 | epoch 0 | loss 3.6211 | lr 2.63e-03 | grad 0.2002 +2026-04-09 21:31:22 - INFO - parrotllm.training - step 6410 | epoch 0 | loss 3.4928 | lr 2.63e-03 | grad 0.1823 +2026-04-09 21:31:25 - INFO - parrotllm.training - step 6420 | epoch 0 | loss 3.6502 | lr 2.63e-03 | grad 0.1509 +2026-04-09 21:31:28 - INFO - parrotllm.training - step 6430 | epoch 0 | loss 3.6320 | lr 2.63e-03 | grad 0.1429 +2026-04-09 21:31:31 - INFO - parrotllm.training - step 6440 | epoch 0 | loss 3.5982 | lr 2.63e-03 | grad 0.1721 +2026-04-09 21:31:34 - INFO - parrotllm.training - step 6450 | epoch 0 | loss 3.6105 | lr 2.63e-03 | grad 0.1677 +2026-04-09 21:31:37 - INFO - parrotllm.training - step 6460 | epoch 0 | loss 3.5630 | lr 2.63e-03 | grad 0.1792 +2026-04-09 21:31:40 - INFO - parrotllm.training - step 6470 | epoch 0 | loss 3.5757 | lr 2.63e-03 | grad 0.1531 +2026-04-09 21:31:43 - INFO - parrotllm.training - step 6480 | epoch 0 | loss 3.5319 | lr 2.63e-03 | grad 0.1587 +2026-04-09 21:31:46 - INFO - parrotllm.training - step 6490 | epoch 0 | loss 3.5799 | lr 2.63e-03 | grad 0.1412 +2026-04-09 21:31:49 - INFO - parrotllm.training - step 6500 | epoch 0 | loss 3.5687 | lr 2.63e-03 | grad 0.1278 +2026-04-09 21:31:49 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:31:49 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:31:53 - INFO - parrotllm.training - Train: loss=3.5687, ppl=35.47 +2026-04-09 21:31:53 - INFO - parrotllm.training - Val: loss=3.5287, ppl=34.08 +2026-04-09 21:31:53 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:31:53 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5287_epoch_0000_step_0006500.pt +2026-04-09 21:31:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:31:57 - INFO - parrotllm.training - step 6510 | epoch 0 | loss 3.5503 | lr 2.63e-03 | grad 0.1746 +2026-04-09 21:32:00 - INFO - parrotllm.training - step 6520 | epoch 0 | loss 3.5106 | lr 2.63e-03 | grad 0.1755 +2026-04-09 21:32:03 - INFO - parrotllm.training - step 6530 | epoch 0 | loss 3.6269 | lr 2.63e-03 | grad 0.1856 +2026-04-09 21:32:07 - INFO - parrotllm.training - step 6540 | epoch 0 | loss 3.5249 | lr 2.63e-03 | grad 0.1438 +2026-04-09 21:32:10 - INFO - parrotllm.training - step 6550 | epoch 0 | loss 3.5466 | lr 2.63e-03 | grad 0.1475 +2026-04-09 21:32:13 - INFO - parrotllm.training - step 6560 | epoch 0 | loss 3.5810 | lr 2.63e-03 | grad 0.1580 +2026-04-09 21:32:16 - INFO - parrotllm.training - step 6570 | epoch 0 | loss 3.4846 | lr 2.63e-03 | grad 0.1695 +2026-04-09 21:32:19 - INFO - parrotllm.training - step 6580 | epoch 0 | loss 3.5967 | lr 2.63e-03 | grad 0.1627 +2026-04-09 21:32:22 - INFO - parrotllm.training - step 6590 | epoch 0 | loss 3.5910 | lr 2.63e-03 | grad 0.1714 +2026-04-09 21:32:25 - INFO - parrotllm.training - step 6600 | epoch 0 | loss 3.5149 | lr 2.63e-03 | grad 0.1637 +2026-04-09 21:32:28 - INFO - parrotllm.training - step 6610 | epoch 0 | loss 3.4982 | lr 2.63e-03 | grad 0.1648 +2026-04-09 21:32:31 - INFO - parrotllm.training - step 6620 | epoch 0 | loss 3.5479 | lr 2.63e-03 | grad 0.1300 +2026-04-09 21:32:34 - INFO - parrotllm.training - step 6630 | epoch 0 | loss 3.6048 | lr 2.63e-03 | grad 0.1596 +2026-04-09 21:32:37 - INFO - parrotllm.training - step 6640 | epoch 0 | loss 3.5487 | lr 2.63e-03 | grad 0.1602 +2026-04-09 21:32:40 - INFO - parrotllm.training - step 6650 | epoch 0 | loss 3.5243 | lr 2.63e-03 | grad 0.1840 +2026-04-09 21:32:43 - INFO - parrotllm.training - step 6660 | epoch 0 | loss 3.5655 | lr 2.63e-03 | grad 0.1517 +2026-04-09 21:32:46 - INFO - parrotllm.training - step 6670 | epoch 0 | loss 3.5919 | lr 2.63e-03 | grad 0.1536 +2026-04-09 21:32:49 - INFO - parrotllm.training - step 6680 | epoch 0 | loss 3.5904 | lr 2.63e-03 | grad 0.1484 +2026-04-09 21:32:52 - INFO - parrotllm.training - step 6690 | epoch 0 | loss 3.5776 | lr 2.63e-03 | grad 0.1603 +2026-04-09 21:32:55 - INFO - parrotllm.training - step 6700 | epoch 0 | loss 3.5008 | lr 2.63e-03 | grad 0.1823 +2026-04-09 21:32:59 - INFO - parrotllm.training - step 6710 | epoch 0 | loss 3.5934 | lr 2.63e-03 | grad 0.1533 +2026-04-09 21:33:02 - INFO - parrotllm.training - step 6720 | epoch 0 | loss 3.4815 | lr 2.63e-03 | grad 0.1584 +2026-04-09 21:33:05 - INFO - parrotllm.training - step 6730 | epoch 0 | loss 3.6275 | lr 2.63e-03 | grad 0.1518 +2026-04-09 21:33:08 - INFO - parrotllm.training - step 6740 | epoch 0 | loss 3.5830 | lr 2.63e-03 | grad 0.1931 +2026-04-09 21:33:11 - INFO - parrotllm.training - step 6750 | epoch 0 | loss 3.6215 | lr 2.63e-03 | grad 0.1403 +2026-04-09 21:33:14 - INFO - parrotllm.training - step 6760 | epoch 0 | loss 3.5666 | lr 2.63e-03 | grad 0.1661 +2026-04-09 21:33:17 - INFO - parrotllm.training - step 6770 | epoch 0 | loss 3.6938 | lr 2.63e-03 | grad 0.1438 +2026-04-09 21:33:20 - INFO - parrotllm.training - step 6780 | epoch 0 | loss 3.5348 | lr 2.63e-03 | grad 0.1806 +2026-04-09 21:33:23 - INFO - parrotllm.training - step 6790 | epoch 0 | loss 3.5829 | lr 2.63e-03 | grad 0.1599 +2026-04-09 21:33:26 - INFO - parrotllm.training - step 6800 | epoch 0 | loss 3.5782 | lr 2.63e-03 | grad 0.1490 +2026-04-09 21:33:29 - INFO - parrotllm.training - step 6810 | epoch 0 | loss 3.5903 | lr 2.63e-03 | grad 0.1643 +2026-04-09 21:33:32 - INFO - parrotllm.training - step 6820 | epoch 0 | loss 3.6589 | lr 2.63e-03 | grad 0.1574 +2026-04-09 21:33:35 - INFO - parrotllm.training - step 6830 | epoch 0 | loss 3.6042 | lr 2.63e-03 | grad 0.1765 +2026-04-09 21:33:38 - INFO - parrotllm.training - step 6840 | epoch 0 | loss 3.5954 | lr 2.63e-03 | grad 0.1736 +2026-04-09 21:33:41 - INFO - parrotllm.training - step 6850 | epoch 0 | loss 3.5865 | lr 2.63e-03 | grad 0.1571 +2026-04-09 21:33:45 - INFO - parrotllm.training - step 6860 | epoch 0 | loss 3.6110 | lr 2.63e-03 | grad 0.1843 +2026-04-09 21:33:48 - INFO - parrotllm.training - step 6870 | epoch 0 | loss 3.6517 | lr 2.63e-03 | grad 0.1699 +2026-04-09 21:33:51 - INFO - parrotllm.training - step 6880 | epoch 0 | loss 3.6221 | lr 2.63e-03 | grad 0.1440 +2026-04-09 21:33:54 - INFO - parrotllm.training - step 6890 | epoch 0 | loss 3.4868 | lr 2.63e-03 | grad 0.1739 +2026-04-09 21:33:57 - INFO - parrotllm.training - step 6900 | epoch 0 | loss 3.4945 | lr 2.63e-03 | grad 0.1752 +2026-04-09 21:34:00 - INFO - parrotllm.training - step 6910 | epoch 0 | loss 3.5845 | lr 2.63e-03 | grad 0.1561 +2026-04-09 21:34:03 - INFO - parrotllm.training - step 6920 | epoch 0 | loss 3.5790 | lr 2.63e-03 | grad 0.1910 +2026-04-09 21:34:06 - INFO - parrotllm.training - step 6930 | epoch 0 | loss 3.5495 | lr 2.63e-03 | grad 0.1507 +2026-04-09 21:34:09 - INFO - parrotllm.training - step 6940 | epoch 0 | loss 3.5117 | lr 2.63e-03 | grad 0.1575 +2026-04-09 21:34:12 - INFO - parrotllm.training - step 6950 | epoch 0 | loss 3.5894 | lr 2.63e-03 | grad 0.1605 +2026-04-09 21:34:15 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 6959/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-09 21:34:15 - INFO - parrotllm.training - step 6960 | epoch 0 | loss 3.5148 | lr 2.63e-03 | grad 0.1750 +2026-04-09 21:34:19 - INFO - parrotllm.training - step 6970 | epoch 0 | loss 3.5154 | lr 2.63e-03 | grad 0.1706 +2026-04-09 21:34:22 - INFO - parrotllm.training - step 6980 | epoch 0 | loss 3.5697 | lr 2.63e-03 | grad 0.1533 +2026-04-09 21:34:25 - INFO - parrotllm.training - step 6990 | epoch 0 | loss 3.5958 | lr 2.63e-03 | grad 0.1366 +2026-04-09 21:34:28 - INFO - parrotllm.training - step 7000 | epoch 0 | loss 3.5975 | lr 2.63e-03 | grad 0.1843 +2026-04-09 21:34:28 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:34:28 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:34:31 - INFO - parrotllm.training - Train: loss=3.5975, ppl=36.51 +2026-04-09 21:34:31 - INFO - parrotllm.training - Val: loss=3.5255, ppl=33.97 +2026-04-09 21:34:31 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:34:32 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5255_epoch_0000_step_0007000.pt +2026-04-09 21:34:33 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:34:36 - INFO - parrotllm.training - step 7010 | epoch 0 | loss 3.6223 | lr 2.63e-03 | grad 0.1463 +2026-04-09 21:34:39 - INFO - parrotllm.training - step 7020 | epoch 0 | loss 3.5972 | lr 2.63e-03 | grad 0.1491 +2026-04-09 21:34:42 - INFO - parrotllm.training - step 7030 | epoch 0 | loss 3.5856 | lr 2.63e-03 | grad 0.1569 +2026-04-09 21:34:45 - INFO - parrotllm.training - step 7040 | epoch 0 | loss 3.5881 | lr 2.63e-03 | grad 0.1588 +2026-04-09 21:34:48 - INFO - parrotllm.training - step 7050 | epoch 0 | loss 3.6484 | lr 2.63e-03 | grad 0.1672 +2026-04-09 21:34:51 - INFO - parrotllm.training - step 7060 | epoch 0 | loss 3.6468 | lr 2.63e-03 | grad 0.1535 +2026-04-09 21:34:54 - INFO - parrotllm.training - step 7070 | epoch 0 | loss 3.5301 | lr 2.63e-03 | grad 0.1725 +2026-04-09 21:34:57 - INFO - parrotllm.training - step 7080 | epoch 0 | loss 3.5903 | lr 2.63e-03 | grad 0.1476 +2026-04-09 21:35:00 - INFO - parrotllm.training - step 7090 | epoch 0 | loss 3.7004 | lr 2.63e-03 | grad 0.1673 +2026-04-09 21:35:03 - INFO - parrotllm.training - step 7100 | epoch 0 | loss 3.5810 | lr 2.63e-03 | grad 0.1381 +2026-04-09 21:35:06 - INFO - parrotllm.training - step 7110 | epoch 0 | loss 3.5951 | lr 2.63e-03 | grad 0.1493 +2026-04-09 21:35:09 - INFO - parrotllm.training - step 7120 | epoch 0 | loss 3.6078 | lr 2.63e-03 | grad 0.1865 +2026-04-09 21:35:12 - INFO - parrotllm.training - step 7130 | epoch 0 | loss 3.6403 | lr 2.63e-03 | grad 0.1507 +2026-04-09 21:35:16 - INFO - parrotllm.training - step 7140 | epoch 0 | loss 3.6101 | lr 2.63e-03 | grad 0.1624 +2026-04-09 21:35:19 - INFO - parrotllm.training - step 7150 | epoch 0 | loss 3.5161 | lr 2.63e-03 | grad 0.1688 +2026-04-09 21:35:22 - INFO - parrotllm.training - step 7160 | epoch 0 | loss 3.5745 | lr 2.63e-03 | grad 0.1533 +2026-04-09 21:35:25 - INFO - parrotllm.training - step 7170 | epoch 0 | loss 3.5132 | lr 2.63e-03 | grad 0.1587 +2026-04-09 21:35:28 - INFO - parrotllm.training - step 7180 | epoch 0 | loss 3.5611 | lr 2.63e-03 | grad 0.1401 +2026-04-09 21:35:31 - INFO - parrotllm.training - step 7190 | epoch 0 | loss 3.5728 | lr 2.63e-03 | grad 0.1611 +2026-04-09 21:35:34 - INFO - parrotllm.training - step 7200 | epoch 0 | loss 3.6039 | lr 2.63e-03 | grad 0.1743 +2026-04-09 21:35:37 - INFO - parrotllm.training - step 7210 | epoch 0 | loss 3.5788 | lr 2.63e-03 | grad 0.1626 +2026-04-09 21:35:40 - INFO - parrotllm.training - step 7220 | epoch 0 | loss 3.5803 | lr 2.63e-03 | grad 0.1682 +2026-04-09 21:35:43 - INFO - parrotllm.training - step 7230 | epoch 0 | loss 3.6145 | lr 2.63e-03 | grad 0.1576 +2026-04-09 21:35:46 - INFO - parrotllm.training - step 7240 | epoch 0 | loss 3.4642 | lr 2.63e-03 | grad 0.2178 +2026-04-09 21:35:49 - INFO - parrotllm.training - step 7250 | epoch 0 | loss 3.4828 | lr 2.63e-03 | grad 0.1761 +2026-04-09 21:35:52 - INFO - parrotllm.training - step 7260 | epoch 0 | loss 3.5997 | lr 2.63e-03 | grad 0.1539 +2026-04-09 21:35:55 - INFO - parrotllm.training - step 7270 | epoch 0 | loss 3.5875 | lr 2.63e-03 | grad 0.1542 +2026-04-09 21:35:58 - INFO - parrotllm.training - step 7280 | epoch 0 | loss 3.4748 | lr 2.63e-03 | grad 0.1410 +2026-04-09 21:36:01 - INFO - parrotllm.training - step 7290 | epoch 0 | loss 3.6313 | lr 2.63e-03 | grad 0.1456 +2026-04-09 21:36:04 - INFO - parrotllm.training - step 7300 | epoch 0 | loss 3.5446 | lr 2.63e-03 | grad 0.1855 +2026-04-09 21:36:07 - INFO - parrotllm.training - step 7310 | epoch 0 | loss 3.6429 | lr 2.63e-03 | grad 0.1840 +2026-04-09 21:36:11 - INFO - parrotllm.training - step 7320 | epoch 0 | loss 3.4948 | lr 2.63e-03 | grad 0.1618 +2026-04-09 21:36:14 - INFO - parrotllm.training - step 7330 | epoch 0 | loss 3.4766 | lr 2.63e-03 | grad 0.1664 +2026-04-09 21:36:17 - INFO - parrotllm.training - step 7340 | epoch 0 | loss 3.5222 | lr 2.63e-03 | grad 0.1570 +2026-04-09 21:36:20 - INFO - parrotllm.training - step 7350 | epoch 0 | loss 3.5299 | lr 2.63e-03 | grad 0.1486 +2026-04-09 21:36:23 - INFO - parrotllm.training - step 7360 | epoch 0 | loss 3.5899 | lr 2.63e-03 | grad 0.1560 +2026-04-09 21:36:26 - INFO - parrotllm.training - step 7370 | epoch 0 | loss 3.6277 | lr 2.63e-03 | grad 0.1411 +2026-04-09 21:36:29 - INFO - parrotllm.training - step 7380 | epoch 0 | loss 3.5455 | lr 2.63e-03 | grad 0.1774 +2026-04-09 21:36:32 - INFO - parrotllm.training - step 7390 | epoch 0 | loss 3.5858 | lr 2.63e-03 | grad 0.1584 +2026-04-09 21:36:35 - INFO - parrotllm.training - step 7400 | epoch 0 | loss 3.4783 | lr 2.63e-03 | grad 0.1904 +2026-04-09 21:36:38 - INFO - parrotllm.training - step 7410 | epoch 0 | loss 3.5135 | lr 2.63e-03 | grad 0.1608 +2026-04-09 21:36:41 - INFO - parrotllm.training - step 7420 | epoch 0 | loss 3.5763 | lr 2.63e-03 | grad 0.1486 +2026-04-09 21:36:44 - INFO - parrotllm.training - step 7430 | epoch 0 | loss 3.6062 | lr 2.63e-03 | grad 0.1734 +2026-04-09 21:36:47 - INFO - parrotllm.training - step 7440 | epoch 0 | loss 3.5591 | lr 2.63e-03 | grad 0.1779 +2026-04-09 21:36:50 - INFO - parrotllm.training - step 7450 | epoch 0 | loss 3.5947 | lr 2.63e-03 | grad 0.1707 +2026-04-09 21:36:53 - INFO - parrotllm.training - step 7460 | epoch 0 | loss 3.5943 | lr 2.63e-03 | grad 0.1356 +2026-04-09 21:36:56 - INFO - parrotllm.training - step 7470 | epoch 0 | loss 3.5677 | lr 2.63e-03 | grad 0.1637 +2026-04-09 21:36:59 - INFO - parrotllm.training - step 7480 | epoch 0 | loss 3.5504 | lr 2.63e-03 | grad 0.1599 +2026-04-09 21:37:03 - INFO - parrotllm.training - step 7490 | epoch 0 | loss 3.6047 | lr 2.63e-03 | grad 0.1537 +2026-04-09 21:37:06 - INFO - parrotllm.training - step 7500 | epoch 0 | loss 3.4163 | lr 2.63e-03 | grad 0.1703 +2026-04-09 21:37:06 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:37:06 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:37:09 - INFO - parrotllm.training - Train: loss=3.4163, ppl=30.46 +2026-04-09 21:37:09 - INFO - parrotllm.training - Val: loss=3.5198, ppl=33.78 +2026-04-09 21:37:09 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:37:10 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5198_epoch_0000_step_0007500.pt +2026-04-09 21:37:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:37:13 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0007500.pt +2026-04-09 21:37:17 - INFO - parrotllm.training - step 7510 | epoch 0 | loss 3.5947 | lr 2.63e-03 | grad 0.1694 +2026-04-09 21:37:20 - INFO - parrotllm.training - step 7520 | epoch 0 | loss 3.4897 | lr 2.63e-03 | grad 0.1451 +2026-04-09 21:37:23 - INFO - parrotllm.training - step 7530 | epoch 0 | loss 3.5671 | lr 2.63e-03 | grad 0.1622 +2026-04-09 21:37:26 - INFO - parrotllm.training - step 7540 | epoch 0 | loss 3.5441 | lr 2.63e-03 | grad 0.1604 +2026-04-09 21:37:29 - INFO - parrotllm.training - step 7550 | epoch 0 | loss 3.5808 | lr 2.63e-03 | grad 0.1655 +2026-04-09 21:37:32 - INFO - parrotllm.training - step 7560 | epoch 0 | loss 3.5398 | lr 2.63e-03 | grad 0.1751 +2026-04-09 21:37:35 - INFO - parrotllm.training - step 7570 | epoch 0 | loss 3.5451 | lr 2.63e-03 | grad 0.1745 +2026-04-09 21:37:38 - INFO - parrotllm.training - step 7580 | epoch 0 | loss 3.4793 | lr 2.63e-03 | grad 0.1338 +2026-04-09 21:37:41 - INFO - parrotllm.training - step 7590 | epoch 0 | loss 3.4129 | lr 2.63e-03 | grad 0.1565 +2026-04-09 21:37:44 - INFO - parrotllm.training - step 7600 | epoch 0 | loss 3.6054 | lr 2.63e-03 | grad 0.1566 +2026-04-09 21:37:47 - INFO - parrotllm.training - step 7610 | epoch 0 | loss 3.7103 | lr 2.63e-03 | grad 0.1847 +2026-04-09 21:37:50 - INFO - parrotllm.training - step 7620 | epoch 0 | loss 3.5355 | lr 2.63e-03 | grad 0.1785 +2026-04-09 21:37:53 - INFO - parrotllm.training - step 7630 | epoch 0 | loss 3.5930 | lr 2.63e-03 | grad 0.1400 +2026-04-09 21:37:56 - INFO - parrotllm.training - step 7640 | epoch 0 | loss 3.4906 | lr 2.63e-03 | grad 0.1700 +2026-04-09 21:37:59 - INFO - parrotllm.training - step 7650 | epoch 0 | loss 3.5550 | lr 2.63e-03 | grad 0.1848 +2026-04-09 21:38:02 - INFO - parrotllm.training - step 7660 | epoch 0 | loss 3.5681 | lr 2.63e-03 | grad 0.1586 +2026-04-09 21:38:05 - INFO - parrotllm.training - step 7670 | epoch 0 | loss 3.5661 | lr 2.63e-03 | grad 0.1575 +2026-04-09 21:38:08 - INFO - parrotllm.training - step 7680 | epoch 0 | loss 3.6104 | lr 2.63e-03 | grad 0.1611 +2026-04-09 21:38:12 - INFO - parrotllm.training - step 7690 | epoch 0 | loss 3.5460 | lr 2.63e-03 | grad 0.1469 +2026-04-09 21:38:15 - INFO - parrotllm.training - step 7700 | epoch 0 | loss 3.5939 | lr 2.63e-03 | grad 0.1388 +2026-04-09 21:38:18 - INFO - parrotllm.training - step 7710 | epoch 0 | loss 3.6021 | lr 2.63e-03 | grad 0.1466 +2026-04-09 21:38:21 - INFO - parrotllm.training - step 7720 | epoch 0 | loss 3.5380 | lr 2.63e-03 | grad 0.1724 +2026-04-09 21:38:24 - INFO - parrotllm.training - step 7730 | epoch 0 | loss 3.5329 | lr 2.63e-03 | grad 0.1548 +2026-04-09 21:38:27 - INFO - parrotllm.training - step 7740 | epoch 0 | loss 3.5356 | lr 2.63e-03 | grad 0.1724 +2026-04-09 21:38:30 - INFO - parrotllm.training - step 7750 | epoch 0 | loss 3.5307 | lr 2.63e-03 | grad 0.1781 +2026-04-09 21:38:33 - INFO - parrotllm.training - step 7760 | epoch 0 | loss 3.5246 | lr 2.63e-03 | grad 0.1589 +2026-04-09 21:38:36 - INFO - parrotllm.training - step 7770 | epoch 0 | loss 3.6214 | lr 2.63e-03 | grad 0.1734 +2026-04-09 21:38:39 - INFO - parrotllm.training - step 7780 | epoch 0 | loss 3.6344 | lr 2.63e-03 | grad 0.1744 +2026-04-09 21:38:42 - INFO - parrotllm.training - step 7790 | epoch 0 | loss 3.7068 | lr 2.63e-03 | grad 0.1860 +2026-04-09 21:38:45 - INFO - parrotllm.training - step 7800 | epoch 0 | loss 3.5653 | lr 2.63e-03 | grad 0.1545 +2026-04-09 21:38:48 - INFO - parrotllm.training - step 7810 | epoch 0 | loss 3.6362 | lr 2.63e-03 | grad 0.1884 +2026-04-09 21:38:51 - INFO - parrotllm.training - step 7820 | epoch 0 | loss 3.5002 | lr 2.63e-03 | grad 0.1363 +2026-04-09 21:38:54 - INFO - parrotllm.training - step 7830 | epoch 0 | loss 3.5801 | lr 2.63e-03 | grad 0.1489 +2026-04-09 21:38:57 - INFO - parrotllm.training - step 7840 | epoch 0 | loss 3.5900 | lr 2.63e-03 | grad 0.1449 +2026-04-09 21:39:00 - INFO - parrotllm.training - step 7850 | epoch 0 | loss 3.5733 | lr 2.63e-03 | grad 0.1722 +2026-04-09 21:39:03 - INFO - parrotllm.training - step 7860 | epoch 0 | loss 3.5429 | lr 2.63e-03 | grad 0.1607 +2026-04-09 21:39:06 - INFO - parrotllm.training - step 7870 | epoch 0 | loss 3.6393 | lr 2.63e-03 | grad 0.1659 +2026-04-09 21:39:10 - INFO - parrotllm.training - step 7880 | epoch 0 | loss 3.5105 | lr 2.63e-03 | grad 0.1921 +2026-04-09 21:39:13 - INFO - parrotllm.training - step 7890 | epoch 0 | loss 3.5294 | lr 2.63e-03 | grad 0.1462 +2026-04-09 21:39:16 - INFO - parrotllm.training - step 7900 | epoch 0 | loss 3.5989 | lr 2.63e-03 | grad 0.1846 +2026-04-09 21:39:19 - INFO - parrotllm.training - step 7910 | epoch 0 | loss 3.4829 | lr 2.63e-03 | grad 0.1640 +2026-04-09 21:39:22 - INFO - parrotllm.training - step 7920 | epoch 0 | loss 3.5440 | lr 2.63e-03 | grad 0.1581 +2026-04-09 21:39:25 - INFO - parrotllm.training - step 7930 | epoch 0 | loss 3.5627 | lr 2.63e-03 | grad 0.1532 +2026-04-09 21:39:28 - INFO - parrotllm.training - step 7940 | epoch 0 | loss 3.5103 | lr 2.63e-03 | grad 0.1687 +2026-04-09 21:39:31 - INFO - parrotllm.training - step 7950 | epoch 0 | loss 3.5015 | lr 2.63e-03 | grad 0.1559 +2026-04-09 21:39:34 - INFO - parrotllm.training - step 7960 | epoch 0 | loss 3.6070 | lr 2.63e-03 | grad 0.1776 +2026-04-09 21:39:37 - INFO - parrotllm.training - step 7970 | epoch 0 | loss 3.6292 | lr 2.63e-03 | grad 0.1511 +2026-04-09 21:39:40 - INFO - parrotllm.training - step 7980 | epoch 0 | loss 3.6851 | lr 2.63e-03 | grad 0.1819 +2026-04-09 21:39:43 - INFO - parrotllm.training - step 7990 | epoch 0 | loss 3.4735 | lr 2.63e-03 | grad 0.1410 +2026-04-09 21:39:46 - INFO - parrotllm.training - step 8000 | epoch 0 | loss 3.5939 | lr 2.63e-03 | grad 0.1647 +2026-04-09 21:39:46 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:39:46 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:39:49 - INFO - parrotllm.training - Train: loss=3.5939, ppl=36.38 +2026-04-09 21:39:49 - INFO - parrotllm.training - Val: loss=3.5144, ppl=33.60 +2026-04-09 21:39:49 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:39:50 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5144_epoch_0000_step_0008000.pt +2026-04-09 21:39:51 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:39:54 - INFO - parrotllm.training - step 8010 | epoch 0 | loss 3.5305 | lr 2.63e-03 | grad 0.1739 +2026-04-09 21:39:57 - INFO - parrotllm.training - step 8020 | epoch 0 | loss 3.5879 | lr 2.63e-03 | grad 0.1680 +2026-04-09 21:40:00 - INFO - parrotllm.training - step 8030 | epoch 0 | loss 3.6355 | lr 2.63e-03 | grad 0.1600 +2026-04-09 21:40:03 - INFO - parrotllm.training - step 8040 | epoch 0 | loss 3.5663 | lr 2.63e-03 | grad 0.1616 +2026-04-09 21:40:06 - INFO - parrotllm.training - step 8050 | epoch 0 | loss 3.5939 | lr 2.63e-03 | grad 0.1748 +2026-04-09 21:40:09 - INFO - parrotllm.training - step 8060 | epoch 0 | loss 3.5711 | lr 2.63e-03 | grad 0.1960 +2026-04-09 21:40:12 - INFO - parrotllm.training - step 8070 | epoch 0 | loss 3.6066 | lr 2.63e-03 | grad 0.1649 +2026-04-09 21:40:15 - INFO - parrotllm.training - step 8080 | epoch 0 | loss 3.5026 | lr 2.63e-03 | grad 0.1485 +2026-04-09 21:40:19 - INFO - parrotllm.training - step 8090 | epoch 0 | loss 3.6028 | lr 2.63e-03 | grad 0.1684 +2026-04-09 21:40:22 - INFO - parrotllm.training - step 8100 | epoch 0 | loss 3.6028 | lr 2.63e-03 | grad 0.1651 +2026-04-09 21:40:25 - INFO - parrotllm.training - step 8110 | epoch 0 | loss 3.5833 | lr 2.63e-03 | grad 0.1703 +2026-04-09 21:40:28 - INFO - parrotllm.training - step 8120 | epoch 0 | loss 3.4358 | lr 2.63e-03 | grad 0.1537 +2026-04-09 21:40:31 - INFO - parrotllm.training - step 8130 | epoch 0 | loss 3.5500 | lr 2.63e-03 | grad 0.1444 +2026-04-09 21:40:34 - INFO - parrotllm.training - step 8140 | epoch 0 | loss 3.5377 | lr 2.63e-03 | grad 0.1797 +2026-04-09 21:40:37 - INFO - parrotllm.training - step 8150 | epoch 0 | loss 3.5682 | lr 2.63e-03 | grad 0.1736 +2026-04-09 21:40:40 - INFO - parrotllm.training - step 8160 | epoch 0 | loss 3.4937 | lr 2.63e-03 | grad 0.1653 +2026-04-09 21:40:43 - INFO - parrotllm.training - step 8170 | epoch 0 | loss 3.6304 | lr 2.63e-03 | grad 0.1513 +2026-04-09 21:40:46 - INFO - parrotllm.training - step 8180 | epoch 0 | loss 3.5664 | lr 2.63e-03 | grad 0.1629 +2026-04-09 21:40:49 - INFO - parrotllm.training - step 8190 | epoch 0 | loss 3.5560 | lr 2.63e-03 | grad 0.1703 +2026-04-09 21:40:52 - INFO - parrotllm.training - step 8200 | epoch 0 | loss 3.4866 | lr 2.63e-03 | grad 0.1471 +2026-04-09 21:40:55 - INFO - parrotllm.training - step 8210 | epoch 0 | loss 3.5730 | lr 2.63e-03 | grad 0.1538 +2026-04-09 21:40:58 - INFO - parrotllm.training - step 8220 | epoch 0 | loss 3.6682 | lr 2.63e-03 | grad 0.1731 +2026-04-09 21:41:01 - INFO - parrotllm.training - step 8230 | epoch 0 | loss 3.5499 | lr 2.63e-03 | grad 0.1928 +2026-04-09 21:41:04 - INFO - parrotllm.training - step 8240 | epoch 0 | loss 3.6685 | lr 2.63e-03 | grad 0.1543 +2026-04-09 21:41:07 - INFO - parrotllm.training - step 8250 | epoch 0 | loss 3.5022 | lr 2.63e-03 | grad 0.1679 +2026-04-09 21:41:10 - INFO - parrotllm.training - step 8260 | epoch 0 | loss 3.6107 | lr 2.63e-03 | grad 0.1743 +2026-04-09 21:41:13 - INFO - parrotllm.training - step 8270 | epoch 0 | loss 3.4412 | lr 2.63e-03 | grad 0.1681 +2026-04-09 21:41:16 - INFO - parrotllm.training - step 8280 | epoch 0 | loss 3.6062 | lr 2.63e-03 | grad 0.1721 +2026-04-09 21:41:19 - INFO - parrotllm.training - step 8290 | epoch 0 | loss 3.5392 | lr 2.63e-03 | grad 0.1660 +2026-04-09 21:41:22 - INFO - parrotllm.training - step 8300 | epoch 0 | loss 3.5851 | lr 2.63e-03 | grad 0.1484 +2026-04-09 21:41:26 - INFO - parrotllm.training - step 8310 | epoch 0 | loss 3.5260 | lr 2.63e-03 | grad 0.1714 +2026-04-09 21:41:29 - INFO - parrotllm.training - step 8320 | epoch 0 | loss 3.5036 | lr 2.63e-03 | grad 0.1803 +2026-04-09 21:41:32 - INFO - parrotllm.training - step 8330 | epoch 0 | loss 3.5469 | lr 2.63e-03 | grad 0.1529 +2026-04-09 21:41:35 - INFO - parrotllm.training - step 8340 | epoch 0 | loss 3.5393 | lr 2.63e-03 | grad 0.1625 +2026-04-09 21:41:38 - INFO - parrotllm.training - step 8350 | epoch 0 | loss 3.5450 | lr 2.63e-03 | grad 0.1647 +2026-04-09 21:41:41 - INFO - parrotllm.training - step 8360 | epoch 0 | loss 3.6494 | lr 2.63e-03 | grad 0.1636 +2026-04-09 21:41:44 - INFO - parrotllm.training - step 8370 | epoch 0 | loss 3.5658 | lr 2.63e-03 | grad 0.1751 +2026-04-09 21:41:47 - INFO - parrotllm.training - step 8380 | epoch 0 | loss 3.5862 | lr 2.63e-03 | grad 0.1762 +2026-04-09 21:41:50 - INFO - parrotllm.training - step 8390 | epoch 0 | loss 3.6025 | lr 2.63e-03 | grad 0.1631 +2026-04-09 21:41:53 - INFO - parrotllm.training - step 8400 | epoch 0 | loss 3.5450 | lr 2.63e-03 | grad 0.1502 +2026-04-09 21:41:56 - INFO - parrotllm.training - step 8410 | epoch 0 | loss 3.4556 | lr 2.63e-03 | grad 0.1512 +2026-04-09 21:41:59 - INFO - parrotllm.training - step 8420 | epoch 0 | loss 3.5683 | lr 2.63e-03 | grad 0.1585 +2026-04-09 21:42:02 - INFO - parrotllm.training - step 8430 | epoch 0 | loss 3.6708 | lr 2.63e-03 | grad 0.1592 +2026-04-09 21:42:05 - INFO - parrotllm.training - step 8440 | epoch 0 | loss 3.5200 | lr 2.63e-03 | grad 0.1702 +2026-04-09 21:42:08 - INFO - parrotllm.training - step 8450 | epoch 0 | loss 3.5266 | lr 2.63e-03 | grad 0.1911 +2026-04-09 21:42:11 - INFO - parrotllm.training - step 8460 | epoch 0 | loss 3.6244 | lr 2.63e-03 | grad 0.1575 +2026-04-09 21:42:14 - INFO - parrotllm.training - step 8470 | epoch 0 | loss 3.5167 | lr 2.63e-03 | grad 0.1560 +2026-04-09 21:42:17 - INFO - parrotllm.training - step 8480 | epoch 0 | loss 3.5099 | lr 2.63e-03 | grad 0.1848 +2026-04-09 21:42:20 - INFO - parrotllm.training - step 8490 | epoch 0 | loss 3.5869 | lr 2.63e-03 | grad 0.1699 +2026-04-09 21:42:23 - INFO - parrotllm.training - step 8500 | epoch 0 | loss 3.4905 | lr 2.63e-03 | grad 0.1535 +2026-04-09 21:42:23 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:42:23 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:42:26 - INFO - parrotllm.training - Train: loss=3.4905, ppl=32.80 +2026-04-09 21:42:26 - INFO - parrotllm.training - Val: loss=3.5091, ppl=33.42 +2026-04-09 21:42:26 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:42:27 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5091_epoch_0000_step_0008500.pt +2026-04-09 21:42:28 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:42:31 - INFO - parrotllm.training - step 8510 | epoch 0 | loss 3.6068 | lr 2.63e-03 | grad 0.1774 +2026-04-09 21:42:34 - INFO - parrotllm.training - step 8520 | epoch 0 | loss 3.4680 | lr 2.63e-03 | grad 0.1520 +2026-04-09 21:42:38 - INFO - parrotllm.training - step 8530 | epoch 0 | loss 3.5459 | lr 2.63e-03 | grad 0.1703 +2026-04-09 21:42:41 - INFO - parrotllm.training - step 8540 | epoch 0 | loss 3.5227 | lr 2.63e-03 | grad 0.1493 +2026-04-09 21:42:44 - INFO - parrotllm.training - step 8550 | epoch 0 | loss 3.5181 | lr 2.63e-03 | grad 0.1546 +2026-04-09 21:42:47 - INFO - parrotllm.training - step 8560 | epoch 0 | loss 3.4798 | lr 2.63e-03 | grad 0.1776 +2026-04-09 21:42:50 - INFO - parrotllm.training - step 8570 | epoch 0 | loss 3.5163 | lr 2.63e-03 | grad 0.1683 +2026-04-09 21:42:53 - INFO - parrotllm.training - step 8580 | epoch 0 | loss 3.5618 | lr 2.63e-03 | grad 0.1797 +2026-04-09 21:42:56 - INFO - parrotllm.training - step 8590 | epoch 0 | loss 3.5822 | lr 2.63e-03 | grad 0.1559 +2026-04-09 21:42:59 - INFO - parrotllm.training - step 8600 | epoch 0 | loss 3.5609 | lr 2.63e-03 | grad 0.1725 +2026-04-09 21:43:02 - INFO - parrotllm.training - step 8610 | epoch 0 | loss 3.5144 | lr 2.63e-03 | grad 0.1650 +2026-04-09 21:43:05 - INFO - parrotllm.training - step 8620 | epoch 0 | loss 3.5158 | lr 2.63e-03 | grad 0.1562 +2026-04-09 21:43:08 - INFO - parrotllm.training - step 8630 | epoch 0 | loss 3.5521 | lr 2.63e-03 | grad 0.1821 +2026-04-09 21:43:11 - INFO - parrotllm.training - step 8640 | epoch 0 | loss 3.4578 | lr 2.63e-03 | grad 0.1496 +2026-04-09 21:43:14 - INFO - parrotllm.training - step 8650 | epoch 0 | loss 3.3925 | lr 2.63e-03 | grad 0.1840 +2026-04-09 21:43:17 - INFO - parrotllm.training - step 8660 | epoch 0 | loss 3.4998 | lr 2.63e-03 | grad 0.1556 +2026-04-09 21:43:20 - INFO - parrotllm.training - step 8670 | epoch 0 | loss 3.4599 | lr 2.63e-03 | grad 0.1578 +2026-04-09 21:43:23 - INFO - parrotllm.training - step 8680 | epoch 0 | loss 3.6741 | lr 2.63e-03 | grad 0.1464 +2026-04-09 21:43:26 - INFO - parrotllm.training - step 8690 | epoch 0 | loss 3.5524 | lr 2.63e-03 | grad 0.1661 +2026-04-09 21:43:29 - INFO - parrotllm.training - step 8700 | epoch 0 | loss 3.5675 | lr 2.63e-03 | grad 0.1597 +2026-04-09 21:43:32 - INFO - parrotllm.training - step 8710 | epoch 0 | loss 3.5076 | lr 2.63e-03 | grad 0.1675 +2026-04-09 21:43:36 - INFO - parrotllm.training - step 8720 | epoch 0 | loss 3.5264 | lr 2.63e-03 | grad 0.1627 +2026-04-09 21:43:39 - INFO - parrotllm.training - step 8730 | epoch 0 | loss 3.5015 | lr 2.63e-03 | grad 0.1887 +2026-04-09 21:43:42 - INFO - parrotllm.training - step 8740 | epoch 0 | loss 3.5685 | lr 2.63e-03 | grad 0.1597 +2026-04-09 21:43:45 - INFO - parrotllm.training - step 8750 | epoch 0 | loss 3.6137 | lr 2.63e-03 | grad 0.1787 +2026-04-09 21:43:48 - INFO - parrotllm.training - step 8760 | epoch 0 | loss 3.4630 | lr 2.63e-03 | grad 0.1609 +2026-04-09 21:43:51 - INFO - parrotllm.training - step 8770 | epoch 0 | loss 3.5593 | lr 2.63e-03 | grad 0.1571 +2026-04-09 21:43:54 - INFO - parrotllm.training - step 8780 | epoch 0 | loss 3.5671 | lr 2.63e-03 | grad 0.1530 +2026-04-09 21:43:57 - INFO - parrotllm.training - step 8790 | epoch 0 | loss 3.5772 | lr 2.63e-03 | grad 0.1702 +2026-04-09 21:44:00 - INFO - parrotllm.training - step 8800 | epoch 0 | loss 3.5172 | lr 2.63e-03 | grad 0.1524 +2026-04-09 21:44:03 - INFO - parrotllm.training - step 8810 | epoch 0 | loss 3.5168 | lr 2.63e-03 | grad 0.1599 +2026-04-09 21:44:06 - INFO - parrotllm.training - step 8820 | epoch 0 | loss 3.4474 | lr 2.63e-03 | grad 0.1491 +2026-04-09 21:44:09 - INFO - parrotllm.training - step 8830 | epoch 0 | loss 3.5484 | lr 2.63e-03 | grad 0.1531 +2026-04-09 21:44:12 - INFO - parrotllm.training - step 8840 | epoch 0 | loss 3.4914 | lr 2.63e-03 | grad 0.1487 +2026-04-09 21:44:15 - INFO - parrotllm.training - step 8850 | epoch 0 | loss 3.5380 | lr 2.63e-03 | grad 0.1904 +2026-04-09 21:44:18 - INFO - parrotllm.training - step 8860 | epoch 0 | loss 3.4781 | lr 2.63e-03 | grad 0.1718 +2026-04-09 21:44:21 - INFO - parrotllm.training - step 8870 | epoch 0 | loss 3.6101 | lr 2.63e-03 | grad 0.1793 +2026-04-09 21:44:24 - INFO - parrotllm.training - step 8880 | epoch 0 | loss 3.5359 | lr 2.63e-03 | grad 0.1516 +2026-04-09 21:44:27 - INFO - parrotllm.training - step 8890 | epoch 0 | loss 3.5848 | lr 2.63e-03 | grad 0.1676 +2026-04-09 21:44:31 - INFO - parrotllm.training - step 8900 | epoch 0 | loss 3.6662 | lr 2.63e-03 | grad 0.1757 +2026-04-09 21:44:34 - INFO - parrotllm.training - step 8910 | epoch 0 | loss 3.5453 | lr 2.63e-03 | grad 0.1602 +2026-04-09 21:44:37 - INFO - parrotllm.training - step 8920 | epoch 0 | loss 3.5249 | lr 2.63e-03 | grad 0.1731 +2026-04-09 21:44:40 - INFO - parrotllm.training - step 8930 | epoch 0 | loss 3.5939 | lr 2.63e-03 | grad 0.1929 +2026-04-09 21:44:43 - INFO - parrotllm.training - step 8940 | epoch 0 | loss 3.5015 | lr 2.63e-03 | grad 0.1572 +2026-04-09 21:44:46 - INFO - parrotllm.training - step 8950 | epoch 0 | loss 3.4589 | lr 2.63e-03 | grad 0.1620 +2026-04-09 21:44:49 - INFO - parrotllm.training - step 8960 | epoch 0 | loss 3.6015 | lr 2.63e-03 | grad 0.1762 +2026-04-09 21:44:52 - INFO - parrotllm.training - step 8970 | epoch 0 | loss 3.5958 | lr 2.63e-03 | grad 0.1580 +2026-04-09 21:44:55 - INFO - parrotllm.training - step 8980 | epoch 0 | loss 3.5504 | lr 2.63e-03 | grad 0.1604 +2026-04-09 21:44:58 - INFO - parrotllm.training - step 8990 | epoch 0 | loss 3.4696 | lr 2.63e-03 | grad 0.1435 +2026-04-09 21:45:01 - INFO - parrotllm.training - step 9000 | epoch 0 | loss 3.5706 | lr 2.63e-03 | grad 0.1597 +2026-04-09 21:45:01 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:45:01 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:45:04 - INFO - parrotllm.training - Train: loss=3.5706, ppl=35.54 +2026-04-09 21:45:04 - INFO - parrotllm.training - Val: loss=3.5062, ppl=33.32 +2026-04-09 21:45:04 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:45:05 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5062_epoch_0000_step_0009000.pt +2026-04-09 21:45:06 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:45:09 - INFO - parrotllm.training - step 9010 | epoch 0 | loss 3.5392 | lr 2.63e-03 | grad 0.1624 +2026-04-09 21:45:12 - INFO - parrotllm.training - step 9020 | epoch 0 | loss 3.4660 | lr 2.63e-03 | grad 0.1552 +2026-04-09 21:45:15 - INFO - parrotllm.training - step 9030 | epoch 0 | loss 3.4246 | lr 2.63e-03 | grad 0.1699 +2026-04-09 21:45:18 - INFO - parrotllm.training - step 9040 | epoch 0 | loss 3.4896 | lr 2.63e-03 | grad 0.1923 +2026-04-09 21:45:21 - INFO - parrotllm.training - step 9050 | epoch 0 | loss 3.5160 | lr 2.63e-03 | grad 0.1544 +2026-04-09 21:45:24 - INFO - parrotllm.training - step 9060 | epoch 0 | loss 3.5055 | lr 2.63e-03 | grad 0.1521 +2026-04-09 21:45:27 - INFO - parrotllm.training - step 9070 | epoch 0 | loss 3.5146 | lr 2.63e-03 | grad 0.1490 +2026-04-09 21:45:30 - INFO - parrotllm.training - step 9080 | epoch 0 | loss 3.5234 | lr 2.63e-03 | grad 0.1515 +2026-04-09 21:45:33 - INFO - parrotllm.training - step 9090 | epoch 0 | loss 3.5512 | lr 2.63e-03 | grad 0.1659 +2026-04-09 21:45:37 - INFO - parrotllm.training - step 9100 | epoch 0 | loss 3.5035 | lr 2.63e-03 | grad 0.1695 +2026-04-09 21:45:40 - INFO - parrotllm.training - step 9110 | epoch 0 | loss 3.5600 | lr 2.63e-03 | grad 0.1518 +2026-04-09 21:45:43 - INFO - parrotllm.training - step 9120 | epoch 0 | loss 3.4888 | lr 2.63e-03 | grad 0.1664 +2026-04-09 21:45:46 - INFO - parrotllm.training - step 9130 | epoch 0 | loss 3.6030 | lr 2.63e-03 | grad 0.1782 +2026-04-09 21:45:49 - INFO - parrotllm.training - step 9140 | epoch 0 | loss 3.5702 | lr 2.63e-03 | grad 0.1681 +2026-04-09 21:45:52 - INFO - parrotllm.training - step 9150 | epoch 0 | loss 3.4045 | lr 2.63e-03 | grad 0.1629 +2026-04-09 21:45:55 - INFO - parrotllm.training - step 9160 | epoch 0 | loss 3.4893 | lr 2.63e-03 | grad 0.1593 +2026-04-09 21:45:58 - INFO - parrotllm.training - step 9170 | epoch 0 | loss 3.5867 | lr 2.63e-03 | grad 0.1455 +2026-04-09 21:46:01 - INFO - parrotllm.training - step 9180 | epoch 0 | loss 3.6748 | lr 2.63e-03 | grad 0.1510 +2026-04-09 21:46:04 - INFO - parrotllm.training - step 9190 | epoch 0 | loss 3.4990 | lr 2.63e-03 | grad 0.1738 +2026-04-09 21:46:07 - INFO - parrotllm.training - step 9200 | epoch 0 | loss 3.5785 | lr 2.63e-03 | grad 0.1777 +2026-04-09 21:46:10 - INFO - parrotllm.training - step 9210 | epoch 0 | loss 3.5723 | lr 2.63e-03 | grad 0.1594 +2026-04-09 21:46:13 - INFO - parrotllm.training - step 9220 | epoch 0 | loss 3.5914 | lr 2.63e-03 | grad 0.1394 +2026-04-09 21:46:16 - INFO - parrotllm.training - step 9230 | epoch 0 | loss 3.6425 | lr 2.63e-03 | grad 0.1676 +2026-04-09 21:46:19 - INFO - parrotllm.training - step 9240 | epoch 0 | loss 3.5871 | lr 2.63e-03 | grad 0.1954 +2026-04-09 21:46:22 - INFO - parrotllm.training - step 9250 | epoch 0 | loss 3.6874 | lr 2.63e-03 | grad 0.1466 +2026-04-09 21:46:25 - INFO - parrotllm.training - step 9260 | epoch 0 | loss 3.5503 | lr 2.63e-03 | grad 0.1754 +2026-04-09 21:46:28 - INFO - parrotllm.training - step 9270 | epoch 0 | loss 3.4197 | lr 2.63e-03 | grad 0.1821 +2026-04-09 21:46:32 - INFO - parrotllm.training - step 9280 | epoch 0 | loss 3.5014 | lr 2.63e-03 | grad 0.1533 +2026-04-09 21:46:35 - INFO - parrotllm.training - step 9290 | epoch 0 | loss 3.6275 | lr 2.63e-03 | grad 0.1629 +2026-04-09 21:46:38 - INFO - parrotllm.training - step 9300 | epoch 0 | loss 3.5148 | lr 2.63e-03 | grad 0.1570 +2026-04-09 21:46:41 - INFO - parrotllm.training - step 9310 | epoch 0 | loss 3.4640 | lr 2.63e-03 | grad 0.1572 +2026-04-09 21:46:44 - INFO - parrotllm.training - step 9320 | epoch 0 | loss 3.5544 | lr 2.63e-03 | grad 0.1355 +2026-04-09 21:46:47 - INFO - parrotllm.training - step 9330 | epoch 0 | loss 3.5198 | lr 2.63e-03 | grad 0.1568 +2026-04-09 21:46:50 - INFO - parrotllm.training - step 9340 | epoch 0 | loss 3.4945 | lr 2.63e-03 | grad 0.1876 +2026-04-09 21:46:53 - INFO - parrotllm.training - step 9350 | epoch 0 | loss 3.4998 | lr 2.63e-03 | grad 0.1659 +2026-04-09 21:46:56 - INFO - parrotllm.training - step 9360 | epoch 0 | loss 3.4764 | lr 2.63e-03 | grad 0.1629 +2026-04-09 21:46:59 - INFO - parrotllm.training - step 9370 | epoch 0 | loss 3.5530 | lr 2.63e-03 | grad 0.1805 +2026-04-09 21:47:02 - INFO - parrotllm.training - step 9380 | epoch 0 | loss 3.5712 | lr 2.63e-03 | grad 0.1584 +2026-04-09 21:47:05 - INFO - parrotllm.training - step 9390 | epoch 0 | loss 3.5085 | lr 2.63e-03 | grad 0.1694 +2026-04-09 21:47:08 - INFO - parrotllm.training - step 9400 | epoch 0 | loss 3.4877 | lr 2.63e-03 | grad 0.1930 +2026-04-09 21:47:11 - INFO - parrotllm.training - step 9410 | epoch 0 | loss 3.6212 | lr 2.63e-03 | grad 0.1590 +2026-04-09 21:47:14 - INFO - parrotllm.training - step 9420 | epoch 0 | loss 3.6139 | lr 2.63e-03 | grad 0.1641 +2026-04-09 21:47:17 - INFO - parrotllm.training - step 9430 | epoch 0 | loss 3.5640 | lr 2.63e-03 | grad 0.1564 +2026-04-09 21:47:20 - INFO - parrotllm.training - step 9440 | epoch 0 | loss 3.5239 | lr 2.63e-03 | grad 0.1535 +2026-04-09 21:47:23 - INFO - parrotllm.training - step 9450 | epoch 0 | loss 3.5478 | lr 2.63e-03 | grad 0.1618 +2026-04-09 21:47:26 - INFO - parrotllm.training - step 9460 | epoch 0 | loss 3.6054 | lr 2.63e-03 | grad 0.1714 +2026-04-09 21:47:29 - INFO - parrotllm.training - step 9470 | epoch 0 | loss 3.6106 | lr 2.63e-03 | grad 0.1877 +2026-04-09 21:47:33 - INFO - parrotllm.training - step 9480 | epoch 0 | loss 3.5094 | lr 2.63e-03 | grad 0.1695 +2026-04-09 21:47:36 - INFO - parrotllm.training - step 9490 | epoch 0 | loss 3.6163 | lr 2.63e-03 | grad 0.1648 +2026-04-09 21:47:39 - INFO - parrotllm.training - step 9500 | epoch 0 | loss 3.6975 | lr 2.63e-03 | grad 0.1575 +2026-04-09 21:47:39 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:47:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:47:42 - INFO - parrotllm.training - Train: loss=3.6975, ppl=40.35 +2026-04-09 21:47:42 - INFO - parrotllm.training - Val: loss=3.5021, ppl=33.18 +2026-04-09 21:47:42 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:47:43 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5021_epoch_0000_step_0009500.pt +2026-04-09 21:47:44 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:47:47 - INFO - parrotllm.training - step 9510 | epoch 0 | loss 3.6267 | lr 2.63e-03 | grad 0.1793 +2026-04-09 21:47:50 - INFO - parrotllm.training - step 9520 | epoch 0 | loss 3.5716 | lr 2.63e-03 | grad 0.1877 +2026-04-09 21:47:53 - INFO - parrotllm.training - step 9530 | epoch 0 | loss 3.6272 | lr 2.63e-03 | grad 0.1664 +2026-04-09 21:47:56 - INFO - parrotllm.training - step 9540 | epoch 0 | loss 3.5114 | lr 2.63e-03 | grad 0.1383 +2026-04-09 21:47:59 - INFO - parrotllm.training - step 9550 | epoch 0 | loss 3.5470 | lr 2.63e-03 | grad 0.1958 +2026-04-09 21:48:02 - INFO - parrotllm.training - step 9560 | epoch 0 | loss 3.6180 | lr 2.63e-03 | grad 0.1382 +2026-04-09 21:48:05 - INFO - parrotllm.training - step 9570 | epoch 0 | loss 3.5355 | lr 2.63e-03 | grad 0.1408 +2026-04-09 21:48:08 - INFO - parrotllm.training - step 9580 | epoch 0 | loss 3.5019 | lr 2.63e-03 | grad 0.1757 +2026-04-09 21:48:11 - INFO - parrotllm.training - step 9590 | epoch 0 | loss 3.5497 | lr 2.63e-03 | grad 0.1650 +2026-04-09 21:48:14 - INFO - parrotllm.training - step 9600 | epoch 0 | loss 3.5687 | lr 2.63e-03 | grad 0.1827 +2026-04-09 21:48:17 - INFO - parrotllm.training - step 9610 | epoch 0 | loss 3.5763 | lr 2.63e-03 | grad 0.1669 +2026-04-09 21:48:20 - INFO - parrotllm.training - step 9620 | epoch 0 | loss 3.4996 | lr 2.63e-03 | grad 0.1660 +2026-04-09 21:48:23 - INFO - parrotllm.training - step 9630 | epoch 0 | loss 3.5019 | lr 2.63e-03 | grad 0.1513 +2026-04-09 21:48:26 - INFO - parrotllm.training - step 9640 | epoch 0 | loss 3.5900 | lr 2.63e-03 | grad 0.1702 +2026-04-09 21:48:29 - INFO - parrotllm.training - step 9650 | epoch 0 | loss 3.6450 | lr 2.63e-03 | grad 0.1414 +2026-04-09 21:48:32 - INFO - parrotllm.training - step 9660 | epoch 0 | loss 3.5441 | lr 2.63e-03 | grad 0.1530 +2026-04-09 21:48:35 - INFO - parrotllm.training - step 9670 | epoch 0 | loss 3.5461 | lr 2.63e-03 | grad 0.1742 +2026-04-09 21:48:38 - INFO - parrotllm.training - step 9680 | epoch 0 | loss 3.4960 | lr 2.63e-03 | grad 0.1944 +2026-04-09 21:48:41 - INFO - parrotllm.training - step 9690 | epoch 0 | loss 3.4351 | lr 2.63e-03 | grad 0.1569 +2026-04-09 21:48:44 - INFO - parrotllm.training - step 9700 | epoch 0 | loss 3.4899 | lr 2.63e-03 | grad 0.1867 +2026-04-09 21:48:48 - INFO - parrotllm.training - step 9710 | epoch 0 | loss 3.5433 | lr 2.63e-03 | grad 0.1887 +2026-04-09 21:48:51 - INFO - parrotllm.training - step 9720 | epoch 0 | loss 3.5171 | lr 2.63e-03 | grad 0.1739 +2026-04-09 21:48:54 - INFO - parrotllm.training - step 9730 | epoch 0 | loss 3.5241 | lr 2.63e-03 | grad 0.1626 +2026-04-09 21:48:57 - INFO - parrotllm.training - step 9740 | epoch 0 | loss 3.5141 | lr 2.63e-03 | grad 0.1495 +2026-04-09 21:49:00 - INFO - parrotllm.training - step 9750 | epoch 0 | loss 3.4820 | lr 2.63e-03 | grad 0.1586 +2026-04-09 21:49:03 - INFO - parrotllm.training - step 9760 | epoch 0 | loss 3.5591 | lr 2.63e-03 | grad 0.1470 +2026-04-09 21:49:06 - INFO - parrotllm.training - step 9770 | epoch 0 | loss 3.5202 | lr 2.63e-03 | grad 0.1673 +2026-04-09 21:49:09 - INFO - parrotllm.training - step 9780 | epoch 0 | loss 3.5681 | lr 2.63e-03 | grad 0.1802 +2026-04-09 21:49:12 - INFO - parrotllm.training - step 9790 | epoch 0 | loss 3.5623 | lr 2.63e-03 | grad 0.1554 +2026-04-09 21:49:15 - INFO - parrotllm.training - step 9800 | epoch 0 | loss 3.5276 | lr 2.63e-03 | grad 0.1474 +2026-04-09 21:49:18 - INFO - parrotllm.training - step 9810 | epoch 0 | loss 3.5661 | lr 2.63e-03 | grad 0.1760 +2026-04-09 21:49:21 - INFO - parrotllm.training - step 9820 | epoch 0 | loss 3.4592 | lr 2.63e-03 | grad 0.1683 +2026-04-09 21:49:24 - INFO - parrotllm.training - step 9830 | epoch 0 | loss 3.4919 | lr 2.63e-03 | grad 0.1424 +2026-04-09 21:49:27 - INFO - parrotllm.training - step 9840 | epoch 0 | loss 3.5876 | lr 2.63e-03 | grad 0.1565 +2026-04-09 21:49:30 - INFO - parrotllm.training - step 9850 | epoch 0 | loss 3.5610 | lr 2.63e-03 | grad 0.1523 +2026-04-09 21:49:33 - INFO - parrotllm.training - step 9860 | epoch 0 | loss 3.5330 | lr 2.63e-03 | grad 0.1530 +2026-04-09 21:49:36 - INFO - parrotllm.training - step 9870 | epoch 0 | loss 3.6074 | lr 2.63e-03 | grad 0.1686 +2026-04-09 21:49:39 - INFO - parrotllm.training - step 9880 | epoch 0 | loss 3.5740 | lr 2.63e-03 | grad 0.1537 +2026-04-09 21:49:42 - INFO - parrotllm.training - step 9890 | epoch 0 | loss 3.6515 | lr 2.63e-03 | grad 0.1965 +2026-04-09 21:49:46 - INFO - parrotllm.training - step 9900 | epoch 0 | loss 3.4871 | lr 2.63e-03 | grad 0.1465 +2026-04-09 21:49:49 - INFO - parrotllm.training - step 9910 | epoch 0 | loss 3.5178 | lr 2.63e-03 | grad 0.1663 +2026-04-09 21:49:52 - INFO - parrotllm.training - step 9920 | epoch 0 | loss 3.5271 | lr 2.63e-03 | grad 0.1643 +2026-04-09 21:49:55 - INFO - parrotllm.training - step 9930 | epoch 0 | loss 3.5646 | lr 2.63e-03 | grad 0.1584 +2026-04-09 21:49:58 - INFO - parrotllm.training - step 9940 | epoch 0 | loss 3.4862 | lr 2.63e-03 | grad 0.1512 +2026-04-09 21:50:01 - INFO - parrotllm.training - step 9950 | epoch 0 | loss 3.6009 | lr 2.63e-03 | grad 0.1471 +2026-04-09 21:50:04 - INFO - parrotllm.training - step 9960 | epoch 0 | loss 3.5236 | lr 2.63e-03 | grad 0.1514 +2026-04-09 21:50:07 - INFO - parrotllm.training - step 9970 | epoch 0 | loss 3.5821 | lr 2.63e-03 | grad 0.1606 +2026-04-09 21:50:10 - INFO - parrotllm.training - step 9980 | epoch 0 | loss 3.5654 | lr 2.63e-03 | grad 0.1604 +2026-04-09 21:50:13 - INFO - parrotllm.training - step 9990 | epoch 0 | loss 3.5403 | lr 2.63e-03 | grad 0.1413 +2026-04-09 21:50:16 - INFO - parrotllm.training - step 10000 | epoch 0 | loss 3.4860 | lr 2.63e-03 | grad 0.1571 +2026-04-09 21:50:16 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:50:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:50:19 - INFO - parrotllm.training - Train: loss=3.4860, ppl=32.66 +2026-04-09 21:50:19 - INFO - parrotllm.training - Val: loss=3.4971, ppl=33.02 +2026-04-09 21:50:19 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:50:20 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4971_epoch_0000_step_0010000.pt +2026-04-09 21:50:21 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:50:22 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0010000.pt +2026-04-09 21:50:26 - INFO - parrotllm.training - step 10010 | epoch 0 | loss 3.6273 | lr 2.63e-03 | grad 0.1645 +2026-04-09 21:50:29 - INFO - parrotllm.training - step 10020 | epoch 0 | loss 3.5619 | lr 2.63e-03 | grad 0.1511 +2026-04-09 21:50:32 - INFO - parrotllm.training - step 10030 | epoch 0 | loss 3.5550 | lr 2.63e-03 | grad 0.2247 +2026-04-09 21:50:35 - INFO - parrotllm.training - step 10040 | epoch 0 | loss 3.4281 | lr 2.63e-03 | grad 0.1841 +2026-04-09 21:50:39 - INFO - parrotllm.training - step 10050 | epoch 0 | loss 3.5477 | lr 2.63e-03 | grad 0.1545 +2026-04-09 21:50:42 - INFO - parrotllm.training - step 10060 | epoch 0 | loss 3.6585 | lr 2.63e-03 | grad 0.1942 +2026-04-09 21:50:45 - INFO - parrotllm.training - step 10070 | epoch 0 | loss 3.4565 | lr 2.63e-03 | grad 0.1472 +2026-04-09 21:50:48 - INFO - parrotllm.training - step 10080 | epoch 0 | loss 3.4963 | lr 2.63e-03 | grad 0.1790 +2026-04-09 21:50:51 - INFO - parrotllm.training - step 10090 | epoch 0 | loss 3.5239 | lr 2.63e-03 | grad 0.1510 +2026-04-09 21:50:54 - INFO - parrotllm.training - step 10100 | epoch 0 | loss 3.6103 | lr 2.63e-03 | grad 0.1637 +2026-04-09 21:50:57 - INFO - parrotllm.training - step 10110 | epoch 0 | loss 3.5031 | lr 2.63e-03 | grad 0.1632 +2026-04-09 21:51:00 - INFO - parrotllm.training - step 10120 | epoch 0 | loss 3.5164 | lr 2.63e-03 | grad 0.1507 +2026-04-09 21:51:03 - INFO - parrotllm.training - step 10130 | epoch 0 | loss 3.5414 | lr 2.63e-03 | grad 0.1651 +2026-04-09 21:51:06 - INFO - parrotllm.training - step 10140 | epoch 0 | loss 3.5576 | lr 2.63e-03 | grad 0.1730 +2026-04-09 21:51:09 - INFO - parrotllm.training - step 10150 | epoch 0 | loss 3.5386 | lr 2.63e-03 | grad 0.1688 +2026-04-09 21:51:12 - INFO - parrotllm.training - step 10160 | epoch 0 | loss 3.4105 | lr 2.63e-03 | grad 0.1864 +2026-04-09 21:51:15 - INFO - parrotllm.training - step 10170 | epoch 0 | loss 3.5993 | lr 2.63e-03 | grad 0.1362 +2026-04-09 21:51:18 - INFO - parrotllm.training - step 10180 | epoch 0 | loss 3.5354 | lr 2.63e-03 | grad 0.1579 +2026-04-09 21:51:21 - INFO - parrotllm.training - step 10190 | epoch 0 | loss 3.4921 | lr 2.63e-03 | grad 0.2250 +2026-04-09 21:51:24 - INFO - parrotllm.training - step 10200 | epoch 0 | loss 3.5256 | lr 2.63e-03 | grad 0.1556 +2026-04-09 21:51:27 - INFO - parrotllm.training - step 10210 | epoch 0 | loss 3.6110 | lr 2.63e-03 | grad 0.2262 +2026-04-09 21:51:30 - INFO - parrotllm.training - step 10220 | epoch 0 | loss 3.6589 | lr 2.63e-03 | grad 0.1541 +2026-04-09 21:51:34 - INFO - parrotllm.training - step 10230 | epoch 0 | loss 3.5356 | lr 2.63e-03 | grad 0.1617 +2026-04-09 21:51:37 - INFO - parrotllm.training - step 10240 | epoch 0 | loss 3.4969 | lr 2.63e-03 | grad 0.1769 +2026-04-09 21:51:40 - INFO - parrotllm.training - step 10250 | epoch 0 | loss 3.5101 | lr 2.63e-03 | grad 0.1516 +2026-04-09 21:51:43 - INFO - parrotllm.training - step 10260 | epoch 0 | loss 3.6150 | lr 2.63e-03 | grad 0.1631 +2026-04-09 21:51:46 - INFO - parrotllm.training - step 10270 | epoch 0 | loss 3.5186 | lr 2.63e-03 | grad 0.1841 +2026-04-09 21:51:49 - INFO - parrotllm.training - step 10280 | epoch 0 | loss 3.4883 | lr 2.63e-03 | grad 0.1672 +2026-04-09 21:51:52 - INFO - parrotllm.training - step 10290 | epoch 0 | loss 3.5416 | lr 2.63e-03 | grad 0.1764 +2026-04-09 21:51:55 - INFO - parrotllm.training - step 10300 | epoch 0 | loss 3.4792 | lr 2.63e-03 | grad 0.1560 +2026-04-09 21:51:58 - INFO - parrotllm.training - step 10310 | epoch 0 | loss 3.5745 | lr 2.63e-03 | grad 0.1649 +2026-04-09 21:52:01 - INFO - parrotllm.training - step 10320 | epoch 0 | loss 3.5675 | lr 2.63e-03 | grad 0.1352 +2026-04-09 21:52:04 - INFO - parrotllm.training - step 10330 | epoch 0 | loss 3.5585 | lr 2.63e-03 | grad 0.1853 +2026-04-09 21:52:07 - INFO - parrotllm.training - step 10340 | epoch 0 | loss 3.5531 | lr 2.63e-03 | grad 0.1880 +2026-04-09 21:52:10 - INFO - parrotllm.training - step 10350 | epoch 0 | loss 3.5475 | lr 2.63e-03 | grad 0.1507 +2026-04-09 21:52:13 - INFO - parrotllm.training - step 10360 | epoch 0 | loss 3.4877 | lr 2.63e-03 | grad 0.1673 +2026-04-09 21:52:16 - INFO - parrotllm.training - step 10370 | epoch 0 | loss 3.5394 | lr 2.63e-03 | grad 0.1494 +2026-04-09 21:52:19 - INFO - parrotllm.training - step 10380 | epoch 0 | loss 3.5814 | lr 2.63e-03 | grad 0.1806 +2026-04-09 21:52:22 - INFO - parrotllm.training - step 10390 | epoch 0 | loss 3.5886 | lr 2.63e-03 | grad 0.1813 +2026-04-09 21:52:25 - INFO - parrotllm.training - step 10400 | epoch 0 | loss 3.6191 | lr 2.63e-03 | grad 0.1531 +2026-04-09 21:52:28 - INFO - parrotllm.training - step 10410 | epoch 0 | loss 3.6043 | lr 2.63e-03 | grad 0.1364 +2026-04-09 21:52:32 - INFO - parrotllm.training - step 10420 | epoch 0 | loss 3.4530 | lr 2.63e-03 | grad 0.1444 +2026-04-09 21:52:35 - INFO - parrotllm.training - step 10430 | epoch 0 | loss 3.4908 | lr 2.63e-03 | grad 0.1510 +2026-04-09 21:52:38 - INFO - parrotllm.training - step 10440 | epoch 0 | loss 3.4681 | lr 2.63e-03 | grad 0.1692 +2026-04-09 21:52:41 - INFO - parrotllm.training - step 10450 | epoch 0 | loss 3.4289 | lr 2.63e-03 | grad 0.1921 +2026-04-09 21:52:44 - INFO - parrotllm.training - step 10460 | epoch 0 | loss 3.5268 | lr 2.63e-03 | grad 0.1639 +2026-04-09 21:52:47 - INFO - parrotllm.training - step 10470 | epoch 0 | loss 3.4821 | lr 2.63e-03 | grad 0.1469 +2026-04-09 21:52:50 - INFO - parrotllm.training - step 10480 | epoch 0 | loss 3.4405 | lr 2.63e-03 | grad 0.1574 +2026-04-09 21:52:53 - INFO - parrotllm.training - step 10490 | epoch 0 | loss 3.4923 | lr 2.63e-03 | grad 0.1941 +2026-04-09 21:52:56 - INFO - parrotllm.training - step 10500 | epoch 0 | loss 3.5198 | lr 2.63e-03 | grad 0.1484 +2026-04-09 21:52:56 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:52:56 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:52:59 - INFO - parrotllm.training - Train: loss=3.5198, ppl=33.78 +2026-04-09 21:52:59 - INFO - parrotllm.training - Val: loss=3.4924, ppl=32.86 +2026-04-09 21:52:59 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 21:53:00 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4924_epoch_0000_step_0010500.pt +2026-04-09 21:53:01 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:53:04 - INFO - parrotllm.training - step 10510 | epoch 0 | loss 3.5555 | lr 2.63e-03 | grad 0.1541 +2026-04-09 21:53:07 - INFO - parrotllm.training - step 10520 | epoch 0 | loss 3.5788 | lr 2.63e-03 | grad 0.1750 +2026-04-09 21:53:10 - INFO - parrotllm.training - step 10530 | epoch 0 | loss 3.5995 | lr 2.63e-03 | grad 0.1714 +2026-04-09 21:53:13 - INFO - parrotllm.training - step 10540 | epoch 0 | loss 3.5156 | lr 2.63e-03 | grad 0.1700 +2026-04-09 21:53:16 - INFO - parrotllm.training - step 10550 | epoch 0 | loss 3.4825 | lr 2.63e-03 | grad 0.1865 +2026-04-09 21:53:19 - INFO - parrotllm.training - step 10560 | epoch 0 | loss 3.5024 | lr 2.63e-03 | grad 0.1553 +2026-04-09 21:53:22 - INFO - parrotllm.training - step 10570 | epoch 0 | loss 3.4692 | lr 2.63e-03 | grad 0.1882 +2026-04-09 21:53:25 - INFO - parrotllm.training - step 10580 | epoch 0 | loss 3.5729 | lr 2.63e-03 | grad 0.1942 +2026-04-09 21:53:28 - INFO - parrotllm.training - step 10590 | epoch 0 | loss 3.4818 | lr 2.63e-03 | grad 0.1501 +2026-04-09 21:53:31 - INFO - parrotllm.training - step 10600 | epoch 0 | loss 3.4915 | lr 2.63e-03 | grad 0.1363 +2026-04-09 21:53:34 - INFO - parrotllm.training - step 10610 | epoch 0 | loss 3.4997 | lr 2.63e-03 | grad 0.1454 +2026-04-09 21:53:37 - INFO - parrotllm.training - step 10620 | epoch 0 | loss 3.4307 | lr 2.63e-03 | grad 0.1949 +2026-04-09 21:53:40 - INFO - parrotllm.training - step 10630 | epoch 0 | loss 3.4535 | lr 2.63e-03 | grad 0.1761 +2026-04-09 21:53:44 - INFO - parrotllm.training - step 10640 | epoch 0 | loss 3.4981 | lr 2.63e-03 | grad 0.1931 +2026-04-09 21:53:47 - INFO - parrotllm.training - step 10650 | epoch 0 | loss 3.4733 | lr 2.63e-03 | grad 0.1792 +2026-04-09 21:53:50 - INFO - parrotllm.training - step 10660 | epoch 0 | loss 3.5835 | lr 2.63e-03 | grad 0.1468 +2026-04-09 21:53:53 - INFO - parrotllm.training - step 10670 | epoch 0 | loss 3.5854 | lr 2.63e-03 | grad 0.1504 +2026-04-09 21:53:56 - INFO - parrotllm.training - step 10680 | epoch 0 | loss 3.5405 | lr 2.63e-03 | grad 0.1636 +2026-04-09 21:53:59 - INFO - parrotllm.training - step 10690 | epoch 0 | loss 3.5875 | lr 2.63e-03 | grad 0.1934 +2026-04-09 21:54:02 - INFO - parrotllm.training - step 10700 | epoch 0 | loss 3.6333 | lr 2.63e-03 | grad 0.1552 +2026-04-09 21:54:05 - INFO - parrotllm.training - step 10710 | epoch 0 | loss 3.4571 | lr 2.63e-03 | grad 0.1574 +2026-04-09 21:54:08 - INFO - parrotllm.training - step 10720 | epoch 0 | loss 3.6571 | lr 2.63e-03 | grad 0.1761 +2026-04-09 21:54:11 - INFO - parrotllm.training - step 10730 | epoch 0 | loss 3.5568 | lr 2.63e-03 | grad 0.1749 +2026-04-09 21:54:14 - INFO - parrotllm.training - step 10740 | epoch 0 | loss 3.5732 | lr 2.63e-03 | grad 0.1984 +2026-04-09 21:54:17 - INFO - parrotllm.training - step 10750 | epoch 0 | loss 3.5429 | lr 2.63e-03 | grad 0.1706 +2026-04-09 21:54:20 - INFO - parrotllm.training - step 10760 | epoch 0 | loss 3.5846 | lr 2.63e-03 | grad 0.1456 +2026-04-09 21:54:21 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 10764/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-09 21:54:24 - INFO - parrotllm.training - step 10770 | epoch 0 | loss 3.5080 | lr 2.63e-03 | grad 0.1467 +2026-04-09 21:54:27 - INFO - parrotllm.training - step 10780 | epoch 0 | loss 3.5036 | lr 2.63e-03 | grad 0.1759 +2026-04-09 21:54:30 - INFO - parrotllm.training - step 10790 | epoch 0 | loss 3.6014 | lr 2.63e-03 | grad 0.1640 +2026-04-09 21:54:33 - INFO - parrotllm.training - step 10800 | epoch 0 | loss 3.5259 | lr 2.63e-03 | grad 0.1706 +2026-04-09 21:54:36 - INFO - parrotllm.training - step 10810 | epoch 0 | loss 3.5339 | lr 2.63e-03 | grad 0.1788 +2026-04-09 21:54:39 - INFO - parrotllm.training - step 10820 | epoch 0 | loss 3.5461 | lr 2.63e-03 | grad 0.1553 +2026-04-09 21:54:42 - INFO - parrotllm.training - step 10830 | epoch 0 | loss 3.4641 | lr 2.63e-03 | grad 0.1470 +2026-04-09 21:54:45 - INFO - parrotllm.training - step 10840 | epoch 0 | loss 3.6133 | lr 2.63e-03 | grad 0.1551 +2026-04-09 21:54:48 - INFO - parrotllm.training - step 10850 | epoch 0 | loss 3.5453 | lr 2.63e-03 | grad 0.2026 +2026-04-09 21:54:51 - INFO - parrotllm.training - step 10860 | epoch 0 | loss 3.4968 | lr 2.63e-03 | grad 0.1765 +2026-04-09 21:54:54 - INFO - parrotllm.training - step 10870 | epoch 0 | loss 3.5261 | lr 2.63e-03 | grad 0.1589 +2026-04-09 21:54:57 - INFO - parrotllm.training - step 10880 | epoch 0 | loss 3.5165 | lr 2.63e-03 | grad 0.1588 +2026-04-09 21:55:00 - INFO - parrotllm.training - step 10890 | epoch 0 | loss 3.5765 | lr 2.63e-03 | grad 0.1502 +2026-04-09 21:55:03 - INFO - parrotllm.training - step 10900 | epoch 0 | loss 3.4691 | lr 2.63e-03 | grad 0.1684 +2026-04-09 21:55:06 - INFO - parrotllm.training - step 10910 | epoch 0 | loss 3.5423 | lr 2.63e-03 | grad 0.1623 +2026-04-09 21:55:09 - INFO - parrotllm.training - step 10920 | epoch 0 | loss 3.5960 | lr 2.63e-03 | grad 0.1766 +2026-04-09 21:55:12 - INFO - parrotllm.training - step 10930 | epoch 0 | loss 3.4806 | lr 2.63e-03 | grad 0.1641 +2026-04-09 21:55:15 - INFO - parrotllm.training - step 10940 | epoch 0 | loss 3.5245 | lr 2.63e-03 | grad 0.1600 +2026-04-09 21:55:18 - INFO - parrotllm.training - step 10950 | epoch 0 | loss 3.4391 | lr 2.63e-03 | grad 0.1668 +2026-04-09 21:55:21 - INFO - parrotllm.training - step 10960 | epoch 0 | loss 3.4774 | lr 2.63e-03 | grad 0.1824 +2026-04-09 21:55:24 - INFO - parrotllm.training - step 10970 | epoch 0 | loss 3.5152 | lr 2.63e-03 | grad 0.1687 +2026-04-09 21:55:28 - INFO - parrotllm.training - step 10980 | epoch 0 | loss 3.4778 | lr 2.63e-03 | grad 0.2071 +2026-04-09 21:55:31 - INFO - parrotllm.training - step 10990 | epoch 0 | loss 3.6129 | lr 2.63e-03 | grad 0.1389 +2026-04-09 21:55:34 - INFO - parrotllm.training - step 11000 | epoch 0 | loss 3.5541 | lr 2.63e-03 | grad 0.1591 +2026-04-09 21:55:34 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:55:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:55:37 - INFO - parrotllm.training - Train: loss=3.5541, ppl=34.96 +2026-04-09 21:55:37 - INFO - parrotllm.training - Val: loss=3.4924, ppl=32.87 +2026-04-09 21:55:38 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4924_epoch_0000_step_0011000.pt +2026-04-09 21:55:39 - INFO - parrotllm.training - No validation improvement for 1/5 evaluation(s) (best=3.4924, min_delta=0.001000). +2026-04-09 21:55:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:55:42 - INFO - parrotllm.training - step 11010 | epoch 0 | loss 3.5078 | lr 2.63e-03 | grad 0.1658 +2026-04-09 21:55:45 - INFO - parrotllm.training - step 11020 | epoch 0 | loss 3.5488 | lr 2.63e-03 | grad 0.1778 +2026-04-09 21:55:48 - INFO - parrotllm.training - step 11030 | epoch 0 | loss 3.4419 | lr 2.63e-03 | grad 0.1613 +2026-04-09 21:55:51 - INFO - parrotllm.training - step 11040 | epoch 0 | loss 3.5483 | lr 2.63e-03 | grad 0.1668 +2026-04-09 21:55:54 - INFO - parrotllm.training - step 11050 | epoch 0 | loss 3.5886 | lr 2.63e-03 | grad 0.1442 +2026-04-09 21:55:57 - INFO - parrotllm.training - step 11060 | epoch 0 | loss 3.5897 | lr 2.63e-03 | grad 0.1706 +2026-04-09 21:56:00 - INFO - parrotllm.training - step 11070 | epoch 0 | loss 3.5259 | lr 2.63e-03 | grad 0.1802 +2026-04-09 21:56:03 - INFO - parrotllm.training - step 11080 | epoch 0 | loss 3.5719 | lr 2.63e-03 | grad 0.1678 +2026-04-09 21:56:06 - INFO - parrotllm.training - step 11090 | epoch 0 | loss 3.5479 | lr 2.63e-03 | grad 0.1571 +2026-04-09 21:56:09 - INFO - parrotllm.training - step 11100 | epoch 0 | loss 3.5317 | lr 2.63e-03 | grad 0.1688 +2026-04-09 21:56:12 - INFO - parrotllm.training - step 11110 | epoch 0 | loss 3.5819 | lr 2.63e-03 | grad 0.1576 +2026-04-09 21:56:15 - INFO - parrotllm.training - step 11120 | epoch 0 | loss 3.6476 | lr 2.63e-03 | grad 0.1485 +2026-04-09 21:56:18 - INFO - parrotllm.training - step 11130 | epoch 0 | loss 3.5109 | lr 2.63e-03 | grad 0.1864 +2026-04-09 21:56:21 - INFO - parrotllm.training - step 11140 | epoch 0 | loss 3.6072 | lr 2.63e-03 | grad 0.1486 +2026-04-09 21:56:24 - INFO - parrotllm.training - step 11150 | epoch 0 | loss 3.5733 | lr 2.63e-03 | grad 0.1866 +2026-04-09 21:56:28 - INFO - parrotllm.training - step 11160 | epoch 0 | loss 3.5716 | lr 2.63e-03 | grad 0.1758 +2026-04-09 21:56:31 - INFO - parrotllm.training - step 11170 | epoch 0 | loss 3.4687 | lr 2.63e-03 | grad 0.1922 +2026-04-09 21:56:34 - INFO - parrotllm.training - step 11180 | epoch 0 | loss 3.5204 | lr 2.63e-03 | grad 0.1666 +2026-04-09 21:56:37 - INFO - parrotllm.training - step 11190 | epoch 0 | loss 3.5042 | lr 2.63e-03 | grad 0.1618 +2026-04-09 21:56:40 - INFO - parrotllm.training - step 11200 | epoch 0 | loss 3.5009 | lr 2.63e-03 | grad 0.1586 +2026-04-09 21:56:43 - INFO - parrotllm.training - step 11210 | epoch 0 | loss 3.6181 | lr 2.63e-03 | grad 0.1570 +2026-04-09 21:56:46 - INFO - parrotllm.training - step 11220 | epoch 0 | loss 3.5438 | lr 2.63e-03 | grad 0.1661 +2026-04-09 21:56:49 - INFO - parrotllm.training - step 11230 | epoch 0 | loss 3.4288 | lr 2.63e-03 | grad 0.1657 +2026-04-09 21:56:52 - INFO - parrotllm.training - step 11240 | epoch 0 | loss 3.5274 | lr 2.63e-03 | grad 0.1760 +2026-04-09 21:56:55 - INFO - parrotllm.training - step 11250 | epoch 0 | loss 3.5576 | lr 2.63e-03 | grad 0.1774 +2026-04-09 21:56:58 - INFO - parrotllm.training - step 11260 | epoch 0 | loss 3.4928 | lr 2.63e-03 | grad 0.1570 +2026-04-09 21:57:01 - INFO - parrotllm.training - step 11270 | epoch 0 | loss 3.4791 | lr 2.63e-03 | grad 0.1736 +2026-04-09 21:57:04 - INFO - parrotllm.training - step 11280 | epoch 0 | loss 3.5593 | lr 2.63e-03 | grad 0.1608 +2026-04-09 21:57:07 - INFO - parrotllm.training - step 11290 | epoch 0 | loss 3.5711 | lr 2.63e-03 | grad 0.1554 +2026-04-09 21:57:10 - INFO - parrotllm.training - step 11300 | epoch 0 | loss 3.4792 | lr 2.63e-03 | grad 0.1642 +2026-04-09 21:57:13 - INFO - parrotllm.training - step 11310 | epoch 0 | loss 3.5444 | lr 2.63e-03 | grad 0.1761 +2026-04-09 21:57:16 - INFO - parrotllm.training - step 11320 | epoch 0 | loss 3.3952 | lr 2.63e-03 | grad 0.1545 +2026-04-09 21:57:19 - INFO - parrotllm.training - step 11330 | epoch 0 | loss 3.5131 | lr 2.63e-03 | grad 0.1659 +2026-04-09 21:57:22 - INFO - parrotllm.training - step 11340 | epoch 0 | loss 3.4462 | lr 2.63e-03 | grad 0.1483 +2026-04-09 21:57:25 - INFO - parrotllm.training - step 11350 | epoch 0 | loss 3.5654 | lr 2.63e-03 | grad 0.1502 +2026-04-09 21:57:28 - INFO - parrotllm.training - step 11360 | epoch 0 | loss 3.4288 | lr 2.63e-03 | grad 0.1930 +2026-04-09 21:57:32 - INFO - parrotllm.training - step 11370 | epoch 0 | loss 3.5620 | lr 2.63e-03 | grad 0.1600 +2026-04-09 21:57:35 - INFO - parrotllm.training - step 11380 | epoch 0 | loss 3.6071 | lr 2.63e-03 | grad 0.1834 +2026-04-09 21:57:38 - INFO - parrotllm.training - step 11390 | epoch 0 | loss 3.4952 | lr 2.63e-03 | grad 0.1945 +2026-04-09 21:57:41 - INFO - parrotllm.training - step 11400 | epoch 0 | loss 3.4736 | lr 2.63e-03 | grad 0.1551 +2026-04-09 21:57:44 - INFO - parrotllm.training - step 11410 | epoch 0 | loss 3.5923 | lr 2.63e-03 | grad 0.1942 +2026-04-09 21:57:47 - INFO - parrotllm.training - step 11420 | epoch 0 | loss 3.5103 | lr 2.63e-03 | grad 0.1733 +2026-04-09 21:57:50 - INFO - parrotllm.training - step 11430 | epoch 0 | loss 3.5570 | lr 2.63e-03 | grad 0.1950 +2026-04-09 21:57:53 - INFO - parrotllm.training - step 11440 | epoch 0 | loss 3.4458 | lr 2.63e-03 | grad 0.1822 +2026-04-09 21:57:56 - INFO - parrotllm.training - step 11450 | epoch 0 | loss 3.6128 | lr 2.63e-03 | grad 0.1838 +2026-04-09 21:57:59 - INFO - parrotllm.training - step 11460 | epoch 0 | loss 3.5443 | lr 2.63e-03 | grad 0.1897 +2026-04-09 21:58:02 - INFO - parrotllm.training - step 11470 | epoch 0 | loss 3.5968 | lr 2.63e-03 | grad 0.1536 +2026-04-09 21:58:05 - INFO - parrotllm.training - step 11480 | epoch 0 | loss 3.4879 | lr 2.63e-03 | grad 0.1746 +2026-04-09 21:58:08 - INFO - parrotllm.training - step 11490 | epoch 0 | loss 3.5765 | lr 2.63e-03 | grad 0.1393 +2026-04-09 21:58:11 - INFO - parrotllm.training - step 11500 | epoch 0 | loss 3.5378 | lr 2.63e-03 | grad 0.1975 +2026-04-09 21:58:11 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 21:58:11 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:58:14 - INFO - parrotllm.training - Train: loss=3.5378, ppl=34.39 +2026-04-09 21:58:14 - INFO - parrotllm.training - Val: loss=3.4946, ppl=32.94 +2026-04-09 21:58:15 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4946_epoch_0000_step_0011500.pt +2026-04-09 21:58:16 - INFO - parrotllm.training - No validation improvement for 2/5 evaluation(s) (best=3.4924, min_delta=0.001000). +2026-04-09 21:58:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 21:58:19 - INFO - parrotllm.training - step 11510 | epoch 0 | loss 3.5289 | lr 2.63e-03 | grad 0.1582 +2026-04-09 21:58:22 - INFO - parrotllm.training - step 11520 | epoch 0 | loss 3.5949 | lr 2.63e-03 | grad 0.1684 +2026-04-09 21:58:25 - INFO - parrotllm.training - step 11530 | epoch 0 | loss 3.5588 | lr 2.63e-03 | grad 0.1647 +2026-04-09 21:58:28 - INFO - parrotllm.training - step 11540 | epoch 0 | loss 3.5777 | lr 2.63e-03 | grad 0.1680 +2026-04-09 21:58:31 - INFO - parrotllm.training - step 11550 | epoch 0 | loss 3.4723 | lr 2.63e-03 | grad 0.1776 +2026-04-09 21:58:34 - INFO - parrotllm.training - step 11560 | epoch 0 | loss 3.6676 | lr 2.63e-03 | grad 0.1559 +2026-04-09 21:58:37 - INFO - parrotllm.training - step 11570 | epoch 0 | loss 3.5623 | lr 2.63e-03 | grad 0.1931 +2026-04-09 21:58:40 - INFO - parrotllm.training - step 11580 | epoch 0 | loss 3.6086 | lr 2.63e-03 | grad 0.1565 +2026-04-09 21:58:43 - INFO - parrotllm.training - step 11590 | epoch 0 | loss 3.5889 | lr 2.63e-03 | grad 0.1402 +2026-04-09 21:58:46 - INFO - parrotllm.training - step 11600 | epoch 0 | loss 3.5832 | lr 2.63e-03 | grad 0.1608 +2026-04-09 21:58:50 - INFO - parrotllm.training - step 11610 | epoch 0 | loss 3.5318 | lr 2.63e-03 | grad 0.1816 +2026-04-09 21:58:53 - INFO - parrotllm.training - step 11620 | epoch 0 | loss 3.5440 | lr 2.63e-03 | grad 0.1552 +2026-04-09 21:58:56 - INFO - parrotllm.training - step 11630 | epoch 0 | loss 3.5470 | lr 2.63e-03 | grad 0.1874 +2026-04-09 21:58:59 - INFO - parrotllm.training - step 11640 | epoch 0 | loss 3.5354 | lr 2.63e-03 | grad 0.1530 +2026-04-09 21:59:02 - INFO - parrotllm.training - step 11650 | epoch 0 | loss 3.5585 | lr 2.63e-03 | grad 0.1559 +2026-04-09 21:59:05 - INFO - parrotllm.training - step 11660 | epoch 0 | loss 3.4490 | lr 2.63e-03 | grad 0.1785 +2026-04-09 21:59:08 - INFO - parrotllm.training - step 11670 | epoch 0 | loss 3.4948 | lr 2.63e-03 | grad 0.1631 +2026-04-09 21:59:11 - INFO - parrotllm.training - step 11680 | epoch 0 | loss 3.4366 | lr 2.63e-03 | grad 0.1677 +2026-04-09 21:59:14 - INFO - parrotllm.training - step 11690 | epoch 0 | loss 3.6007 | lr 2.63e-03 | grad 0.1815 +2026-04-09 21:59:17 - INFO - parrotllm.training - step 11700 | epoch 0 | loss 3.4999 | lr 2.63e-03 | grad 0.1618 +2026-04-09 21:59:20 - INFO - parrotllm.training - step 11710 | epoch 0 | loss 3.5965 | lr 2.63e-03 | grad 0.1737 +2026-04-09 21:59:23 - INFO - parrotllm.training - step 11720 | epoch 0 | loss 3.5081 | lr 2.63e-03 | grad 0.1448 +2026-04-09 21:59:26 - INFO - parrotllm.training - step 11730 | epoch 0 | loss 3.5707 | lr 2.63e-03 | grad 0.1461 +2026-04-09 21:59:29 - INFO - parrotllm.training - step 11740 | epoch 0 | loss 3.5284 | lr 2.63e-03 | grad 0.1765 +2026-04-09 21:59:32 - INFO - parrotllm.training - step 11750 | epoch 0 | loss 3.4972 | lr 2.63e-03 | grad 0.1591 +2026-04-09 21:59:35 - INFO - parrotllm.training - step 11760 | epoch 0 | loss 3.4946 | lr 2.63e-03 | grad 0.1596 +2026-04-09 21:59:38 - INFO - parrotllm.training - step 11770 | epoch 0 | loss 3.6213 | lr 2.63e-03 | grad 0.1958 +2026-04-09 21:59:41 - INFO - parrotllm.training - step 11780 | epoch 0 | loss 3.6468 | lr 2.63e-03 | grad 0.1696 +2026-04-09 21:59:44 - INFO - parrotllm.training - step 11790 | epoch 0 | loss 3.4496 | lr 2.63e-03 | grad 0.1452 +2026-04-09 21:59:47 - INFO - parrotllm.training - step 11800 | epoch 0 | loss 3.5155 | lr 2.63e-03 | grad 0.1550 +2026-04-09 21:59:50 - INFO - parrotllm.training - step 11810 | epoch 0 | loss 3.5373 | lr 2.63e-03 | grad 0.1688 +2026-04-09 21:59:53 - INFO - parrotllm.training - step 11820 | epoch 0 | loss 3.4948 | lr 2.63e-03 | grad 0.1649 +2026-04-09 21:59:56 - INFO - parrotllm.training - step 11830 | epoch 0 | loss 3.5346 | lr 2.63e-03 | grad 0.1874 +2026-04-09 21:59:59 - INFO - parrotllm.training - step 11840 | epoch 0 | loss 3.6078 | lr 2.63e-03 | grad 0.1683 +2026-04-09 22:00:03 - INFO - parrotllm.training - step 11850 | epoch 0 | loss 3.6065 | lr 2.63e-03 | grad 0.1436 +2026-04-09 22:00:06 - INFO - parrotllm.training - step 11860 | epoch 0 | loss 3.5672 | lr 2.63e-03 | grad 0.1701 +2026-04-09 22:00:09 - INFO - parrotllm.training - step 11870 | epoch 0 | loss 3.4879 | lr 2.63e-03 | grad 0.1638 +2026-04-09 22:00:12 - INFO - parrotllm.training - step 11880 | epoch 0 | loss 3.5232 | lr 2.63e-03 | grad 0.1928 +2026-04-09 22:00:15 - INFO - parrotllm.training - step 11890 | epoch 0 | loss 3.5823 | lr 2.63e-03 | grad 0.1824 +2026-04-09 22:00:18 - INFO - parrotllm.training - step 11900 | epoch 0 | loss 3.6271 | lr 2.63e-03 | grad 0.1652 +2026-04-09 22:00:21 - INFO - parrotllm.training - step 11910 | epoch 0 | loss 3.4991 | lr 2.63e-03 | grad 0.1598 +2026-04-09 22:00:24 - INFO - parrotllm.training - step 11920 | epoch 0 | loss 3.5279 | lr 2.63e-03 | grad 0.1635 +2026-04-09 22:00:27 - INFO - parrotllm.training - step 11930 | epoch 0 | loss 3.5043 | lr 2.63e-03 | grad 0.1581 +2026-04-09 22:00:30 - INFO - parrotllm.training - step 11940 | epoch 0 | loss 3.5156 | lr 2.63e-03 | grad 0.1907 +2026-04-09 22:00:33 - INFO - parrotllm.training - step 11950 | epoch 0 | loss 3.4106 | lr 2.63e-03 | grad 0.1453 +2026-04-09 22:00:36 - INFO - parrotllm.training - step 11960 | epoch 0 | loss 3.5350 | lr 2.63e-03 | grad 0.1769 +2026-04-09 22:00:39 - INFO - parrotllm.training - step 11970 | epoch 0 | loss 3.5029 | lr 2.63e-03 | grad 0.1986 +2026-04-09 22:00:42 - INFO - parrotllm.training - step 11980 | epoch 0 | loss 3.6591 | lr 2.63e-03 | grad 0.1678 +2026-04-09 22:00:45 - INFO - parrotllm.training - step 11990 | epoch 0 | loss 3.5429 | lr 2.63e-03 | grad 0.1823 +2026-04-09 22:00:48 - INFO - parrotllm.training - step 12000 | epoch 0 | loss 3.5691 | lr 2.63e-03 | grad 0.1757 +2026-04-09 22:00:48 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:00:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:00:51 - INFO - parrotllm.training - Train: loss=3.5691, ppl=35.49 +2026-04-09 22:00:51 - INFO - parrotllm.training - Val: loss=3.4846, ppl=32.61 +2026-04-09 22:00:51 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 22:00:52 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4846_epoch_0000_step_0012000.pt +2026-04-09 22:00:53 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:00:56 - INFO - parrotllm.training - step 12010 | epoch 0 | loss 3.5557 | lr 2.63e-03 | grad 0.1405 +2026-04-09 22:00:59 - INFO - parrotllm.training - step 12020 | epoch 0 | loss 3.4809 | lr 2.63e-03 | grad 0.1826 +2026-04-09 22:01:02 - INFO - parrotllm.training - step 12030 | epoch 0 | loss 3.5482 | lr 2.63e-03 | grad 0.1629 +2026-04-09 22:01:05 - INFO - parrotllm.training - step 12040 | epoch 0 | loss 3.5003 | lr 2.63e-03 | grad 0.2053 +2026-04-09 22:01:08 - INFO - parrotllm.training - step 12050 | epoch 0 | loss 3.5562 | lr 2.63e-03 | grad 0.1695 +2026-04-09 22:01:11 - INFO - parrotllm.training - step 12060 | epoch 0 | loss 3.4875 | lr 2.63e-03 | grad 0.1754 +2026-04-09 22:01:14 - INFO - parrotllm.training - step 12070 | epoch 0 | loss 3.5311 | lr 2.63e-03 | grad 0.1643 +2026-04-09 22:01:17 - INFO - parrotllm.training - step 12080 | epoch 0 | loss 3.5179 | lr 2.63e-03 | grad 0.1541 +2026-04-09 22:01:21 - INFO - parrotllm.training - step 12090 | epoch 0 | loss 3.4401 | lr 2.63e-03 | grad 0.1723 +2026-04-09 22:01:24 - INFO - parrotllm.training - step 12100 | epoch 0 | loss 3.4641 | lr 2.63e-03 | grad 0.1579 +2026-04-09 22:01:27 - INFO - parrotllm.training - step 12110 | epoch 0 | loss 3.4184 | lr 2.63e-03 | grad 0.1620 +2026-04-09 22:01:30 - INFO - parrotllm.training - step 12120 | epoch 0 | loss 3.5463 | lr 2.63e-03 | grad 0.1682 +2026-04-09 22:01:33 - INFO - parrotllm.training - step 12130 | epoch 0 | loss 3.5753 | lr 2.63e-03 | grad 0.1511 +2026-04-09 22:01:36 - INFO - parrotllm.training - step 12140 | epoch 0 | loss 3.5352 | lr 2.63e-03 | grad 0.1711 +2026-04-09 22:01:39 - INFO - parrotllm.training - step 12150 | epoch 0 | loss 3.5253 | lr 2.63e-03 | grad 0.1846 +2026-04-09 22:01:42 - INFO - parrotllm.training - step 12160 | epoch 0 | loss 3.5204 | lr 2.63e-03 | grad 0.1466 +2026-04-09 22:01:45 - INFO - parrotllm.training - step 12170 | epoch 0 | loss 3.5568 | lr 2.63e-03 | grad 0.1669 +2026-04-09 22:01:48 - INFO - parrotllm.training - step 12180 | epoch 0 | loss 3.5494 | lr 2.63e-03 | grad 0.1628 +2026-04-09 22:01:51 - INFO - parrotllm.training - step 12190 | epoch 0 | loss 3.4572 | lr 2.63e-03 | grad 0.1651 +2026-04-09 22:01:54 - INFO - parrotllm.training - step 12200 | epoch 0 | loss 3.6376 | lr 2.63e-03 | grad 0.1605 +2026-04-09 22:01:57 - INFO - parrotllm.training - step 12210 | epoch 0 | loss 3.6500 | lr 2.63e-03 | grad 0.1655 +2026-04-09 22:02:00 - INFO - parrotllm.training - step 12220 | epoch 0 | loss 3.4765 | lr 2.63e-03 | grad 0.1512 +2026-04-09 22:02:03 - INFO - parrotllm.training - step 12230 | epoch 0 | loss 3.6121 | lr 2.63e-03 | grad 0.1487 +2026-04-09 22:02:06 - INFO - parrotllm.training - step 12240 | epoch 0 | loss 3.5429 | lr 2.63e-03 | grad 0.2091 +2026-04-09 22:02:09 - INFO - parrotllm.training - step 12250 | epoch 0 | loss 3.5134 | lr 2.63e-03 | grad 0.1474 +2026-04-09 22:02:12 - INFO - parrotllm.training - step 12260 | epoch 0 | loss 3.4822 | lr 2.63e-03 | grad 0.1592 +2026-04-09 22:02:15 - INFO - parrotllm.training - step 12270 | epoch 0 | loss 3.5505 | lr 2.63e-03 | grad 0.1646 +2026-04-09 22:02:18 - INFO - parrotllm.training - step 12280 | epoch 0 | loss 3.6272 | lr 2.63e-03 | grad 0.1627 +2026-04-09 22:02:21 - INFO - parrotllm.training - step 12290 | epoch 0 | loss 3.5241 | lr 2.63e-03 | grad 0.1547 +2026-04-09 22:02:24 - INFO - parrotllm.training - step 12300 | epoch 0 | loss 3.4977 | lr 2.63e-03 | grad 0.1569 +2026-04-09 22:02:28 - INFO - parrotllm.training - step 12310 | epoch 0 | loss 3.5626 | lr 2.63e-03 | grad 0.1742 +2026-04-09 22:02:31 - INFO - parrotllm.training - step 12320 | epoch 0 | loss 3.4951 | lr 2.63e-03 | grad 0.1966 +2026-04-09 22:02:34 - INFO - parrotllm.training - step 12330 | epoch 0 | loss 3.4624 | lr 2.63e-03 | grad 0.1589 +2026-04-09 22:02:37 - INFO - parrotllm.training - step 12340 | epoch 0 | loss 3.5221 | lr 2.63e-03 | grad 0.1938 +2026-04-09 22:02:40 - INFO - parrotllm.training - step 12350 | epoch 0 | loss 3.5131 | lr 2.63e-03 | grad 0.1493 +2026-04-09 22:02:43 - INFO - parrotllm.training - step 12360 | epoch 0 | loss 3.5130 | lr 2.63e-03 | grad 0.1677 +2026-04-09 22:02:46 - INFO - parrotllm.training - step 12370 | epoch 0 | loss 3.5456 | lr 2.63e-03 | grad 0.1692 +2026-04-09 22:02:49 - INFO - parrotllm.training - step 12380 | epoch 0 | loss 3.5419 | lr 2.63e-03 | grad 0.1550 +2026-04-09 22:02:52 - INFO - parrotllm.training - step 12390 | epoch 0 | loss 3.5460 | lr 2.63e-03 | grad 0.1715 +2026-04-09 22:02:55 - INFO - parrotllm.training - step 12400 | epoch 0 | loss 3.4749 | lr 2.63e-03 | grad 0.1756 +2026-04-09 22:02:58 - INFO - parrotllm.training - step 12410 | epoch 0 | loss 3.4924 | lr 2.63e-03 | grad 0.1612 +2026-04-09 22:03:01 - INFO - parrotllm.training - step 12420 | epoch 0 | loss 3.5045 | lr 2.63e-03 | grad 0.1689 +2026-04-09 22:03:04 - INFO - parrotllm.training - step 12430 | epoch 0 | loss 3.3862 | lr 2.63e-03 | grad 0.1487 +2026-04-09 22:03:07 - INFO - parrotllm.training - step 12440 | epoch 0 | loss 3.5421 | lr 2.63e-03 | grad 0.1810 +2026-04-09 22:03:10 - INFO - parrotllm.training - step 12450 | epoch 0 | loss 3.5706 | lr 2.63e-03 | grad 0.1457 +2026-04-09 22:03:13 - INFO - parrotllm.training - step 12460 | epoch 0 | loss 3.5270 | lr 2.63e-03 | grad 0.1620 +2026-04-09 22:03:16 - INFO - parrotllm.training - step 12470 | epoch 0 | loss 3.4352 | lr 2.63e-03 | grad 0.1588 +2026-04-09 22:03:19 - INFO - parrotllm.training - step 12480 | epoch 0 | loss 3.5881 | lr 2.63e-03 | grad 0.1810 +2026-04-09 22:03:22 - INFO - parrotllm.training - step 12490 | epoch 0 | loss 3.4858 | lr 2.63e-03 | grad 0.1421 +2026-04-09 22:03:26 - INFO - parrotllm.training - step 12500 | epoch 0 | loss 3.5438 | lr 2.63e-03 | grad 0.1455 +2026-04-09 22:03:26 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:03:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:03:29 - INFO - parrotllm.training - Train: loss=3.5438, ppl=34.60 +2026-04-09 22:03:29 - INFO - parrotllm.training - Val: loss=3.4811, ppl=32.50 +2026-04-09 22:03:29 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 22:03:30 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4811_epoch_0000_step_0012500.pt +2026-04-09 22:03:31 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:03:32 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0012500.pt +2026-04-09 22:03:36 - INFO - parrotllm.training - step 12510 | epoch 0 | loss 3.5518 | lr 2.63e-03 | grad 0.1852 +2026-04-09 22:03:39 - INFO - parrotllm.training - step 12520 | epoch 0 | loss 3.5150 | lr 2.63e-03 | grad 0.1682 +2026-04-09 22:03:42 - INFO - parrotllm.training - step 12530 | epoch 0 | loss 3.5050 | lr 2.63e-03 | grad 0.2228 +2026-04-09 22:03:45 - INFO - parrotllm.training - step 12540 | epoch 0 | loss 3.4776 | lr 2.63e-03 | grad 0.1554 +2026-04-09 22:03:48 - INFO - parrotllm.training - step 12550 | epoch 0 | loss 3.5101 | lr 2.63e-03 | grad 0.1572 +2026-04-09 22:03:51 - INFO - parrotllm.training - step 12560 | epoch 0 | loss 3.4187 | lr 2.63e-03 | grad 0.1775 +2026-04-09 22:03:54 - INFO - parrotllm.training - step 12570 | epoch 0 | loss 3.4542 | lr 2.63e-03 | grad 0.2035 +2026-04-09 22:03:57 - INFO - parrotllm.training - step 12580 | epoch 0 | loss 3.5173 | lr 2.63e-03 | grad 0.1576 +2026-04-09 22:04:00 - INFO - parrotllm.training - step 12590 | epoch 0 | loss 3.5047 | lr 2.63e-03 | grad 0.1638 +2026-04-09 22:04:03 - INFO - parrotllm.training - step 12600 | epoch 0 | loss 3.4464 | lr 2.63e-03 | grad 0.2001 +2026-04-09 22:04:06 - INFO - parrotllm.training - step 12610 | epoch 0 | loss 3.4008 | lr 2.63e-03 | grad 0.1518 +2026-04-09 22:04:09 - INFO - parrotllm.training - step 12620 | epoch 0 | loss 3.6048 | lr 2.63e-03 | grad 0.1525 +2026-04-09 22:04:12 - INFO - parrotllm.training - step 12630 | epoch 0 | loss 3.5347 | lr 2.63e-03 | grad 0.1761 +2026-04-09 22:04:15 - INFO - parrotllm.training - step 12640 | epoch 0 | loss 3.3776 | lr 2.63e-03 | grad 0.1707 +2026-04-09 22:04:18 - INFO - parrotllm.training - step 12650 | epoch 0 | loss 3.5580 | lr 2.63e-03 | grad 0.1654 +2026-04-09 22:04:21 - INFO - parrotllm.training - step 12660 | epoch 0 | loss 3.5457 | lr 2.63e-03 | grad 0.1629 +2026-04-09 22:04:25 - INFO - parrotllm.training - step 12670 | epoch 0 | loss 3.5243 | lr 2.63e-03 | grad 0.1695 +2026-04-09 22:04:28 - INFO - parrotllm.training - step 12680 | epoch 0 | loss 3.5510 | lr 2.63e-03 | grad 0.1715 +2026-04-09 22:04:31 - INFO - parrotllm.training - step 12690 | epoch 0 | loss 3.4528 | lr 2.63e-03 | grad 0.1909 +2026-04-09 22:04:34 - INFO - parrotllm.training - step 12700 | epoch 0 | loss 3.5236 | lr 2.63e-03 | grad 0.1716 +2026-04-09 22:04:37 - INFO - parrotllm.training - step 12710 | epoch 0 | loss 3.5017 | lr 2.63e-03 | grad 0.1684 +2026-04-09 22:04:40 - INFO - parrotllm.training - step 12720 | epoch 0 | loss 3.5055 | lr 2.63e-03 | grad 0.1487 +2026-04-09 22:04:43 - INFO - parrotllm.training - step 12730 | epoch 0 | loss 3.5448 | lr 2.63e-03 | grad 0.1730 +2026-04-09 22:04:46 - INFO - parrotllm.training - step 12740 | epoch 0 | loss 3.4980 | lr 2.63e-03 | grad 0.1506 +2026-04-09 22:04:49 - INFO - parrotllm.training - step 12750 | epoch 0 | loss 3.5892 | lr 2.63e-03 | grad 0.1739 +2026-04-09 22:04:52 - INFO - parrotllm.training - step 12760 | epoch 0 | loss 3.5126 | lr 2.63e-03 | grad 0.1497 +2026-04-09 22:04:55 - INFO - parrotllm.training - step 12770 | epoch 0 | loss 3.4663 | lr 2.63e-03 | grad 0.1555 +2026-04-09 22:04:58 - INFO - parrotllm.training - step 12780 | epoch 0 | loss 3.4841 | lr 2.63e-03 | grad 0.1764 +2026-04-09 22:05:01 - INFO - parrotllm.training - step 12790 | epoch 0 | loss 3.5207 | lr 2.63e-03 | grad 0.1591 +2026-04-09 22:05:04 - INFO - parrotllm.training - step 12800 | epoch 0 | loss 3.5223 | lr 2.63e-03 | grad 0.1691 +2026-04-09 22:05:07 - INFO - parrotllm.training - step 12810 | epoch 0 | loss 3.3938 | lr 2.63e-03 | grad 0.1561 +2026-04-09 22:05:10 - INFO - parrotllm.training - step 12820 | epoch 0 | loss 3.6035 | lr 2.63e-03 | grad 0.2119 +2026-04-09 22:05:13 - INFO - parrotllm.training - step 12830 | epoch 0 | loss 3.5251 | lr 2.63e-03 | grad 0.1810 +2026-04-09 22:05:16 - INFO - parrotllm.training - step 12840 | epoch 0 | loss 3.4637 | lr 2.63e-03 | grad 0.1682 +2026-04-09 22:05:19 - INFO - parrotllm.training - step 12850 | epoch 0 | loss 3.5502 | lr 2.63e-03 | grad 0.1519 +2026-04-09 22:05:23 - INFO - parrotllm.training - step 12860 | epoch 0 | loss 3.5991 | lr 2.63e-03 | grad 0.1527 +2026-04-09 22:05:26 - INFO - parrotllm.training - step 12870 | epoch 0 | loss 3.5025 | lr 2.63e-03 | grad 0.2147 +2026-04-09 22:05:29 - INFO - parrotllm.training - step 12880 | epoch 0 | loss 3.4576 | lr 2.63e-03 | grad 0.1539 +2026-04-09 22:05:32 - INFO - parrotllm.training - step 12890 | epoch 0 | loss 3.5223 | lr 2.63e-03 | grad 0.1603 +2026-04-09 22:05:35 - INFO - parrotllm.training - step 12900 | epoch 0 | loss 3.4158 | lr 2.63e-03 | grad 0.1539 +2026-04-09 22:05:38 - INFO - parrotllm.training - step 12910 | epoch 0 | loss 3.5092 | lr 2.63e-03 | grad 0.1656 +2026-04-09 22:05:41 - INFO - parrotllm.training - step 12920 | epoch 0 | loss 3.5554 | lr 2.63e-03 | grad 0.2268 +2026-04-09 22:05:44 - INFO - parrotllm.training - step 12930 | epoch 0 | loss 3.6217 | lr 2.63e-03 | grad 0.1658 +2026-04-09 22:05:47 - INFO - parrotllm.training - step 12940 | epoch 0 | loss 3.4753 | lr 2.63e-03 | grad 0.1491 +2026-04-09 22:05:50 - INFO - parrotllm.training - step 12950 | epoch 0 | loss 3.6156 | lr 2.63e-03 | grad 0.1714 +2026-04-09 22:05:53 - INFO - parrotllm.training - step 12960 | epoch 0 | loss 3.4882 | lr 2.63e-03 | grad 0.1494 +2026-04-09 22:05:56 - INFO - parrotllm.training - step 12970 | epoch 0 | loss 3.5290 | lr 2.63e-03 | grad 0.1860 +2026-04-09 22:05:59 - INFO - parrotllm.training - step 12980 | epoch 0 | loss 3.4570 | lr 2.63e-03 | grad 0.1785 +2026-04-09 22:06:02 - INFO - parrotllm.training - step 12990 | epoch 0 | loss 3.5543 | lr 2.63e-03 | grad 0.1565 +2026-04-09 22:06:05 - INFO - parrotllm.training - step 13000 | epoch 0 | loss 3.5078 | lr 2.63e-03 | grad 0.1790 +2026-04-09 22:06:05 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:06:05 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:06:08 - INFO - parrotllm.training - Train: loss=3.5078, ppl=33.37 +2026-04-09 22:06:08 - INFO - parrotllm.training - Val: loss=3.4784, ppl=32.41 +2026-04-09 22:06:08 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 22:06:09 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4784_epoch_0000_step_0013000.pt +2026-04-09 22:06:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:06:13 - INFO - parrotllm.training - step 13010 | epoch 0 | loss 3.5390 | lr 2.63e-03 | grad 0.1687 +2026-04-09 22:06:17 - INFO - parrotllm.training - step 13020 | epoch 0 | loss 3.4866 | lr 2.63e-03 | grad 0.1938 +2026-04-09 22:06:20 - INFO - parrotllm.training - step 13030 | epoch 0 | loss 3.5071 | lr 2.63e-03 | grad 0.1863 +2026-04-09 22:06:23 - INFO - parrotllm.training - step 13040 | epoch 0 | loss 3.5845 | lr 2.63e-03 | grad 0.1798 +2026-04-09 22:06:26 - INFO - parrotllm.training - step 13050 | epoch 0 | loss 3.5021 | lr 2.63e-03 | grad 0.1529 +2026-04-09 22:06:29 - INFO - parrotllm.training - step 13060 | epoch 0 | loss 3.4350 | lr 2.63e-03 | grad 0.1665 +2026-04-09 22:06:32 - INFO - parrotllm.training - step 13070 | epoch 0 | loss 3.4629 | lr 2.63e-03 | grad 0.1620 +2026-04-09 22:06:35 - INFO - parrotllm.training - step 13080 | epoch 0 | loss 3.5214 | lr 2.63e-03 | grad 0.1932 +2026-04-09 22:06:38 - INFO - parrotllm.training - step 13090 | epoch 0 | loss 3.5976 | lr 2.63e-03 | grad 0.2145 +2026-04-09 22:06:41 - INFO - parrotllm.training - step 13100 | epoch 0 | loss 3.5678 | lr 2.63e-03 | grad 0.1952 +2026-04-09 22:06:44 - INFO - parrotllm.training - step 13110 | epoch 0 | loss 3.4571 | lr 2.63e-03 | grad 0.1877 +2026-04-09 22:06:47 - INFO - parrotllm.training - step 13120 | epoch 0 | loss 3.5193 | lr 2.63e-03 | grad 0.2037 +2026-04-09 22:06:50 - INFO - parrotllm.training - step 13130 | epoch 0 | loss 3.5790 | lr 2.63e-03 | grad 0.1632 +2026-04-09 22:06:53 - INFO - parrotllm.training - step 13140 | epoch 0 | loss 3.5049 | lr 2.63e-03 | grad 0.1616 +2026-04-09 22:06:56 - INFO - parrotllm.training - step 13150 | epoch 0 | loss 3.4649 | lr 2.63e-03 | grad 0.1810 +2026-04-09 22:06:59 - INFO - parrotllm.training - step 13160 | epoch 0 | loss 3.6571 | lr 2.63e-03 | grad 0.1451 +2026-04-09 22:07:02 - INFO - parrotllm.training - step 13170 | epoch 0 | loss 3.5043 | lr 2.63e-03 | grad 0.1739 +2026-04-09 22:07:05 - INFO - parrotllm.training - step 13180 | epoch 0 | loss 3.4531 | lr 2.63e-03 | grad 0.1571 +2026-04-09 22:07:08 - INFO - parrotllm.training - step 13190 | epoch 0 | loss 3.5830 | lr 2.63e-03 | grad 0.1592 +2026-04-09 22:07:11 - INFO - parrotllm.training - step 13200 | epoch 0 | loss 3.5685 | lr 2.63e-03 | grad 0.1589 +2026-04-09 22:07:15 - INFO - parrotllm.training - step 13210 | epoch 0 | loss 3.4598 | lr 2.63e-03 | grad 0.1517 +2026-04-09 22:07:18 - INFO - parrotllm.training - step 13220 | epoch 0 | loss 3.6273 | lr 2.63e-03 | grad 0.1678 +2026-04-09 22:07:21 - INFO - parrotllm.training - step 13230 | epoch 0 | loss 3.5059 | lr 2.63e-03 | grad 0.1526 +2026-04-09 22:07:24 - INFO - parrotllm.training - step 13240 | epoch 0 | loss 3.5932 | lr 2.63e-03 | grad 0.1557 +2026-04-09 22:07:27 - INFO - parrotllm.training - step 13250 | epoch 0 | loss 3.5076 | lr 2.63e-03 | grad 0.1845 +2026-04-09 22:07:30 - INFO - parrotllm.training - step 13260 | epoch 0 | loss 3.4893 | lr 2.63e-03 | grad 0.1597 +2026-04-09 22:07:33 - INFO - parrotllm.training - step 13270 | epoch 0 | loss 3.4543 | lr 2.63e-03 | grad 0.1722 +2026-04-09 22:07:36 - INFO - parrotllm.training - step 13280 | epoch 0 | loss 3.5164 | lr 2.63e-03 | grad 0.1529 +2026-04-09 22:07:39 - INFO - parrotllm.training - step 13290 | epoch 0 | loss 3.5009 | lr 2.63e-03 | grad 0.2076 +2026-04-09 22:07:42 - INFO - parrotllm.training - step 13300 | epoch 0 | loss 3.5163 | lr 2.63e-03 | grad 0.1625 +2026-04-09 22:07:45 - INFO - parrotllm.training - step 13310 | epoch 0 | loss 3.4382 | lr 2.63e-03 | grad 0.1716 +2026-04-09 22:07:48 - INFO - parrotllm.training - step 13320 | epoch 0 | loss 3.5702 | lr 2.63e-03 | grad 0.1921 +2026-04-09 22:07:51 - INFO - parrotllm.training - step 13330 | epoch 0 | loss 3.5059 | lr 2.63e-03 | grad 0.1481 +2026-04-09 22:07:54 - INFO - parrotllm.training - step 13340 | epoch 0 | loss 3.5263 | lr 2.63e-03 | grad 0.1684 +2026-04-09 22:07:57 - INFO - parrotllm.training - step 13350 | epoch 0 | loss 3.4689 | lr 2.63e-03 | grad 0.1908 +2026-04-09 22:08:00 - INFO - parrotllm.training - step 13360 | epoch 0 | loss 3.4807 | lr 2.63e-03 | grad 0.1798 +2026-04-09 22:08:03 - INFO - parrotllm.training - step 13370 | epoch 0 | loss 3.4767 | lr 2.63e-03 | grad 0.1669 +2026-04-09 22:08:06 - INFO - parrotllm.training - step 13380 | epoch 0 | loss 3.5238 | lr 2.63e-03 | grad 0.1936 +2026-04-09 22:08:10 - INFO - parrotllm.training - step 13390 | epoch 0 | loss 3.5551 | lr 2.63e-03 | grad 0.1716 +2026-04-09 22:08:13 - INFO - parrotllm.training - step 13400 | epoch 0 | loss 3.4546 | lr 2.63e-03 | grad 0.1627 +2026-04-09 22:08:16 - INFO - parrotllm.training - step 13410 | epoch 0 | loss 3.4654 | lr 2.63e-03 | grad 0.1726 +2026-04-09 22:08:19 - INFO - parrotllm.training - step 13420 | epoch 0 | loss 3.5354 | lr 2.63e-03 | grad 0.1726 +2026-04-09 22:08:22 - INFO - parrotllm.training - step 13430 | epoch 0 | loss 3.4833 | lr 2.63e-03 | grad 0.1567 +2026-04-09 22:08:25 - INFO - parrotllm.training - step 13440 | epoch 0 | loss 3.5143 | lr 2.63e-03 | grad 0.1885 +2026-04-09 22:08:28 - INFO - parrotllm.training - step 13450 | epoch 0 | loss 3.6069 | lr 2.63e-03 | grad 0.1557 +2026-04-09 22:08:31 - INFO - parrotllm.training - step 13460 | epoch 0 | loss 3.5613 | lr 2.63e-03 | grad 0.1634 +2026-04-09 22:08:34 - INFO - parrotllm.training - step 13470 | epoch 0 | loss 3.5687 | lr 2.63e-03 | grad 0.1459 +2026-04-09 22:08:37 - INFO - parrotllm.training - step 13480 | epoch 0 | loss 3.4364 | lr 2.63e-03 | grad 0.1865 +2026-04-09 22:08:40 - INFO - parrotllm.training - step 13490 | epoch 0 | loss 3.5199 | lr 2.63e-03 | grad 0.2088 +2026-04-09 22:08:43 - INFO - parrotllm.training - step 13500 | epoch 0 | loss 3.5645 | lr 2.63e-03 | grad 0.1768 +2026-04-09 22:08:43 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:08:43 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:08:46 - INFO - parrotllm.training - Train: loss=3.5645, ppl=35.32 +2026-04-09 22:08:46 - INFO - parrotllm.training - Val: loss=3.4834, ppl=32.57 +2026-04-09 22:08:47 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4834_epoch_0000_step_0013500.pt +2026-04-09 22:08:48 - INFO - parrotllm.training - No validation improvement for 1/5 evaluation(s) (best=3.4784, min_delta=0.001000). +2026-04-09 22:08:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:08:51 - INFO - parrotllm.training - step 13510 | epoch 0 | loss 3.5190 | lr 2.63e-03 | grad 0.1804 +2026-04-09 22:08:54 - INFO - parrotllm.training - step 13520 | epoch 0 | loss 3.5293 | lr 2.63e-03 | grad 0.1602 +2026-04-09 22:08:57 - INFO - parrotllm.training - step 13530 | epoch 0 | loss 3.5480 | lr 2.63e-03 | grad 0.1695 +2026-04-09 22:09:01 - INFO - parrotllm.training - step 13540 | epoch 0 | loss 3.4942 | lr 2.63e-03 | grad 0.1645 +2026-04-09 22:09:04 - INFO - parrotllm.training - step 13550 | epoch 0 | loss 3.5821 | lr 2.63e-03 | grad 0.1911 +2026-04-09 22:09:07 - INFO - parrotllm.training - step 13560 | epoch 0 | loss 3.5578 | lr 2.63e-03 | grad 0.1449 +2026-04-09 22:09:10 - INFO - parrotllm.training - step 13570 | epoch 0 | loss 3.5187 | lr 2.63e-03 | grad 0.1726 +2026-04-09 22:09:13 - INFO - parrotllm.training - step 13580 | epoch 0 | loss 3.5742 | lr 2.63e-03 | grad 0.1911 +2026-04-09 22:09:16 - INFO - parrotllm.training - step 13590 | epoch 0 | loss 3.4332 | lr 2.63e-03 | grad 0.1798 +2026-04-09 22:09:19 - INFO - parrotllm.training - step 13600 | epoch 0 | loss 3.4818 | lr 2.63e-03 | grad 0.1442 +2026-04-09 22:09:22 - INFO - parrotllm.training - step 13610 | epoch 0 | loss 3.5416 | lr 2.63e-03 | grad 0.1770 +2026-04-09 22:09:25 - INFO - parrotllm.training - step 13620 | epoch 0 | loss 3.5144 | lr 2.63e-03 | grad 0.1936 +2026-04-09 22:09:28 - INFO - parrotllm.training - step 13630 | epoch 0 | loss 3.5167 | lr 2.63e-03 | grad 0.1584 +2026-04-09 22:09:31 - INFO - parrotllm.training - step 13640 | epoch 0 | loss 3.5271 | lr 2.63e-03 | grad 0.1594 +2026-04-09 22:09:34 - INFO - parrotllm.training - step 13650 | epoch 0 | loss 3.4352 | lr 2.63e-03 | grad 0.1628 +2026-04-09 22:09:37 - INFO - parrotllm.training - step 13660 | epoch 0 | loss 3.5770 | lr 2.63e-03 | grad 0.1587 +2026-04-09 22:09:40 - INFO - parrotllm.training - step 13670 | epoch 0 | loss 3.5044 | lr 2.63e-03 | grad 0.1658 +2026-04-09 22:09:43 - INFO - parrotllm.training - step 13680 | epoch 0 | loss 3.5292 | lr 2.63e-03 | grad 0.1516 +2026-04-09 22:09:46 - INFO - parrotllm.training - step 13690 | epoch 0 | loss 3.4521 | lr 2.63e-03 | grad 0.1744 +2026-04-09 22:09:49 - INFO - parrotllm.training - step 13700 | epoch 0 | loss 3.6451 | lr 2.63e-03 | grad 0.1629 +2026-04-09 22:09:52 - INFO - parrotllm.training - step 13710 | epoch 0 | loss 3.4536 | lr 2.63e-03 | grad 0.1822 +2026-04-09 22:09:55 - INFO - parrotllm.training - step 13720 | epoch 0 | loss 3.5182 | lr 2.63e-03 | grad 0.1863 +2026-04-09 22:09:58 - INFO - parrotllm.training - step 13730 | epoch 0 | loss 3.6023 | lr 2.63e-03 | grad 0.1714 +2026-04-09 22:10:02 - INFO - parrotllm.training - step 13740 | epoch 0 | loss 3.5153 | lr 2.63e-03 | grad 0.1418 +2026-04-09 22:10:05 - INFO - parrotllm.training - step 13750 | epoch 0 | loss 3.5647 | lr 2.63e-03 | grad 0.1852 +2026-04-09 22:10:08 - INFO - parrotllm.training - step 13760 | epoch 0 | loss 3.5343 | lr 2.63e-03 | grad 0.1571 +2026-04-09 22:10:11 - INFO - parrotllm.training - step 13770 | epoch 0 | loss 3.5251 | lr 2.63e-03 | grad 0.1684 +2026-04-09 22:10:14 - INFO - parrotllm.training - step 13780 | epoch 0 | loss 3.5319 | lr 2.63e-03 | grad 0.1663 +2026-04-09 22:10:17 - INFO - parrotllm.training - step 13790 | epoch 0 | loss 3.4714 | lr 2.63e-03 | grad 0.1685 +2026-04-09 22:10:20 - INFO - parrotllm.training - step 13800 | epoch 0 | loss 3.5163 | lr 2.63e-03 | grad 0.1856 +2026-04-09 22:10:23 - INFO - parrotllm.training - step 13810 | epoch 0 | loss 3.4777 | lr 2.63e-03 | grad 0.1484 +2026-04-09 22:10:26 - INFO - parrotllm.training - step 13820 | epoch 0 | loss 3.4152 | lr 2.63e-03 | grad 0.1560 +2026-04-09 22:10:29 - INFO - parrotllm.training - step 13830 | epoch 0 | loss 3.6243 | lr 2.63e-03 | grad 0.1739 +2026-04-09 22:10:32 - INFO - parrotllm.training - step 13840 | epoch 0 | loss 3.6094 | lr 2.63e-03 | grad 0.1638 +2026-04-09 22:10:35 - INFO - parrotllm.training - step 13850 | epoch 0 | loss 3.5211 | lr 2.63e-03 | grad 0.1723 +2026-04-09 22:10:38 - INFO - parrotllm.training - step 13860 | epoch 0 | loss 3.4280 | lr 2.63e-03 | grad 0.1745 +2026-04-09 22:10:41 - INFO - parrotllm.training - step 13870 | epoch 0 | loss 3.3924 | lr 2.63e-03 | grad 0.2027 +2026-04-09 22:10:44 - INFO - parrotllm.training - step 13880 | epoch 0 | loss 3.4215 | lr 2.63e-03 | grad 0.1387 +2026-04-09 22:10:47 - INFO - parrotllm.training - step 13890 | epoch 0 | loss 3.5681 | lr 2.63e-03 | grad 0.1805 +2026-04-09 22:10:50 - INFO - parrotllm.training - step 13900 | epoch 0 | loss 3.5399 | lr 2.63e-03 | grad 0.1748 +2026-04-09 22:10:53 - INFO - parrotllm.training - step 13910 | epoch 0 | loss 3.4893 | lr 2.63e-03 | grad 0.1552 +2026-04-09 22:10:56 - INFO - parrotllm.training - step 13920 | epoch 0 | loss 3.4691 | lr 2.63e-03 | grad 0.1631 +2026-04-09 22:10:59 - INFO - parrotllm.training - step 13930 | epoch 0 | loss 3.4471 | lr 2.63e-03 | grad 0.1855 +2026-04-09 22:11:03 - INFO - parrotllm.training - step 13940 | epoch 0 | loss 3.5262 | lr 2.63e-03 | grad 0.1624 +2026-04-09 22:11:06 - INFO - parrotllm.training - step 13950 | epoch 0 | loss 3.4908 | lr 2.63e-03 | grad 0.2136 +2026-04-09 22:11:09 - INFO - parrotllm.training - step 13960 | epoch 0 | loss 3.5215 | lr 2.63e-03 | grad 0.1960 +2026-04-09 22:11:12 - INFO - parrotllm.training - step 13970 | epoch 0 | loss 3.5309 | lr 2.63e-03 | grad 0.1673 +2026-04-09 22:11:15 - INFO - parrotllm.training - step 13980 | epoch 0 | loss 3.4832 | lr 2.63e-03 | grad 0.1661 +2026-04-09 22:11:18 - INFO - parrotllm.training - step 13990 | epoch 0 | loss 3.4791 | lr 2.63e-03 | grad 0.1704 +2026-04-09 22:11:21 - INFO - parrotllm.training - step 14000 | epoch 0 | loss 3.5606 | lr 2.63e-03 | grad 0.1875 +2026-04-09 22:11:21 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:11:21 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:11:24 - INFO - parrotllm.training - Train: loss=3.5606, ppl=35.19 +2026-04-09 22:11:24 - INFO - parrotllm.training - Val: loss=3.4781, ppl=32.40 +2026-04-09 22:11:25 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4781_epoch_0000_step_0014000.pt +2026-04-09 22:11:26 - INFO - parrotllm.training - No validation improvement for 2/5 evaluation(s) (best=3.4784, min_delta=0.001000). +2026-04-09 22:11:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:11:29 - INFO - parrotllm.training - step 14010 | epoch 0 | loss 3.4983 | lr 2.63e-03 | grad 0.1643 +2026-04-09 22:11:32 - INFO - parrotllm.training - step 14020 | epoch 0 | loss 3.5601 | lr 2.63e-03 | grad 0.1569 +2026-04-09 22:11:35 - INFO - parrotllm.training - step 14030 | epoch 0 | loss 3.4687 | lr 2.63e-03 | grad 0.1455 +2026-04-09 22:11:38 - INFO - parrotllm.training - step 14040 | epoch 0 | loss 3.6364 | lr 2.63e-03 | grad 0.1614 +2026-04-09 22:11:41 - INFO - parrotllm.training - step 14050 | epoch 0 | loss 3.4674 | lr 2.63e-03 | grad 0.1638 +2026-04-09 22:11:44 - INFO - parrotllm.training - step 14060 | epoch 0 | loss 3.5365 | lr 2.63e-03 | grad 0.1573 +2026-04-09 22:11:47 - INFO - parrotllm.training - step 14070 | epoch 0 | loss 3.5019 | lr 2.63e-03 | grad 0.1775 +2026-04-09 22:11:50 - INFO - parrotllm.training - step 14080 | epoch 0 | loss 3.4712 | lr 2.63e-03 | grad 0.1587 +2026-04-09 22:11:53 - INFO - parrotllm.training - step 14090 | epoch 0 | loss 3.5125 | lr 2.63e-03 | grad 0.1776 +2026-04-09 22:11:56 - INFO - parrotllm.training - step 14100 | epoch 0 | loss 3.5495 | lr 2.63e-03 | grad 0.1691 +2026-04-09 22:11:59 - INFO - parrotllm.training - step 14110 | epoch 0 | loss 3.5570 | lr 2.63e-03 | grad 0.1891 +2026-04-09 22:12:02 - INFO - parrotllm.training - step 14120 | epoch 0 | loss 3.5102 | lr 2.63e-03 | grad 0.1811 +2026-04-09 22:12:05 - INFO - parrotllm.training - step 14130 | epoch 0 | loss 3.5366 | lr 2.63e-03 | grad 0.1699 +2026-04-09 22:12:09 - INFO - parrotllm.training - step 14140 | epoch 0 | loss 3.5101 | lr 2.63e-03 | grad 0.1754 +2026-04-09 22:12:12 - INFO - parrotllm.training - step 14150 | epoch 0 | loss 3.5504 | lr 2.63e-03 | grad 0.1628 +2026-04-09 22:12:15 - INFO - parrotllm.training - step 14160 | epoch 0 | loss 3.5502 | lr 2.63e-03 | grad 0.2041 +2026-04-09 22:12:18 - INFO - parrotllm.training - step 14170 | epoch 0 | loss 3.5831 | lr 2.63e-03 | grad 0.1574 +2026-04-09 22:12:21 - INFO - parrotllm.training - step 14180 | epoch 0 | loss 3.5771 | lr 2.63e-03 | grad 0.1753 +2026-04-09 22:12:24 - INFO - parrotllm.training - step 14190 | epoch 0 | loss 3.5734 | lr 2.63e-03 | grad 0.1654 +2026-04-09 22:12:27 - INFO - parrotllm.training - step 14200 | epoch 0 | loss 3.4744 | lr 2.63e-03 | grad 0.1779 +2026-04-09 22:12:30 - INFO - parrotllm.training - step 14210 | epoch 0 | loss 3.4577 | lr 2.63e-03 | grad 0.1582 +2026-04-09 22:12:33 - INFO - parrotllm.training - step 14220 | epoch 0 | loss 3.4213 | lr 2.63e-03 | grad 0.2020 +2026-04-09 22:12:36 - INFO - parrotllm.training - step 14230 | epoch 0 | loss 3.4211 | lr 2.63e-03 | grad 0.1640 +2026-04-09 22:12:39 - INFO - parrotllm.training - step 14240 | epoch 0 | loss 3.4751 | lr 2.63e-03 | grad 0.1863 +2026-04-09 22:12:42 - INFO - parrotllm.training - step 14250 | epoch 0 | loss 3.5087 | lr 2.63e-03 | grad 0.1688 +2026-04-09 22:12:45 - INFO - parrotllm.training - step 14260 | epoch 0 | loss 3.4902 | lr 2.63e-03 | grad 0.1461 +2026-04-09 22:12:48 - INFO - parrotllm.training - step 14270 | epoch 0 | loss 3.5347 | lr 2.63e-03 | grad 0.1714 +2026-04-09 22:12:51 - INFO - parrotllm.training - step 14280 | epoch 0 | loss 3.4123 | lr 2.63e-03 | grad 0.1535 +2026-04-09 22:12:54 - INFO - parrotllm.training - step 14290 | epoch 0 | loss 3.5250 | lr 2.63e-03 | grad 0.1878 +2026-04-09 22:12:57 - INFO - parrotllm.training - step 14300 | epoch 0 | loss 3.5594 | lr 2.63e-03 | grad 0.1575 +2026-04-09 22:13:00 - INFO - parrotllm.training - step 14310 | epoch 0 | loss 3.5636 | lr 2.63e-03 | grad 0.1481 +2026-04-09 22:13:03 - INFO - parrotllm.training - step 14320 | epoch 0 | loss 3.4879 | lr 2.63e-03 | grad 0.1551 +2026-04-09 22:13:07 - INFO - parrotllm.training - step 14330 | epoch 0 | loss 3.5392 | lr 2.63e-03 | grad 0.1517 +2026-04-09 22:13:10 - INFO - parrotllm.training - step 14340 | epoch 0 | loss 3.6071 | lr 2.63e-03 | grad 0.1555 +2026-04-09 22:13:13 - INFO - parrotllm.training - step 14350 | epoch 0 | loss 3.5095 | lr 2.63e-03 | grad 0.1795 +2026-04-09 22:13:16 - INFO - parrotllm.training - step 14360 | epoch 0 | loss 3.5011 | lr 2.63e-03 | grad 0.1659 +2026-04-09 22:13:19 - INFO - parrotllm.training - step 14370 | epoch 0 | loss 3.4814 | lr 2.63e-03 | grad 0.1539 +2026-04-09 22:13:20 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 14377/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-09 22:13:22 - INFO - parrotllm.training - step 14380 | epoch 0 | loss 3.4802 | lr 2.63e-03 | grad 0.1761 +2026-04-09 22:13:25 - INFO - parrotllm.training - step 14390 | epoch 0 | loss 3.5103 | lr 2.63e-03 | grad 0.1696 +2026-04-09 22:13:28 - INFO - parrotllm.training - step 14400 | epoch 0 | loss 3.4679 | lr 2.63e-03 | grad 0.2131 +2026-04-09 22:13:31 - INFO - parrotllm.training - step 14410 | epoch 0 | loss 3.4827 | lr 2.63e-03 | grad 0.1731 +2026-04-09 22:13:34 - INFO - parrotllm.training - step 14420 | epoch 0 | loss 3.5314 | lr 2.63e-03 | grad 0.1617 +2026-04-09 22:13:37 - INFO - parrotllm.training - step 14430 | epoch 0 | loss 3.5153 | lr 2.63e-03 | grad 0.1870 +2026-04-09 22:13:40 - INFO - parrotllm.training - step 14440 | epoch 0 | loss 3.5263 | lr 2.63e-03 | grad 0.1487 +2026-04-09 22:13:43 - INFO - parrotllm.training - step 14450 | epoch 0 | loss 3.5759 | lr 2.63e-03 | grad 0.1828 +2026-04-09 22:13:47 - INFO - parrotllm.training - step 14460 | epoch 0 | loss 3.5798 | lr 2.63e-03 | grad 0.1969 +2026-04-09 22:13:50 - INFO - parrotllm.training - step 14470 | epoch 0 | loss 3.4471 | lr 2.63e-03 | grad 0.1729 +2026-04-09 22:13:53 - INFO - parrotllm.training - step 14480 | epoch 0 | loss 3.5958 | lr 2.63e-03 | grad 0.1659 +2026-04-09 22:13:56 - INFO - parrotllm.training - step 14490 | epoch 0 | loss 3.6024 | lr 2.63e-03 | grad 0.1760 +2026-04-09 22:13:59 - INFO - parrotllm.training - step 14500 | epoch 0 | loss 3.5915 | lr 2.63e-03 | grad 0.1614 +2026-04-09 22:13:59 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:13:59 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:14:02 - INFO - parrotllm.training - Train: loss=3.5915, ppl=36.29 +2026-04-09 22:14:02 - INFO - parrotllm.training - Val: loss=3.4766, ppl=32.35 +2026-04-09 22:14:02 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 22:14:03 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4766_epoch_0000_step_0014500.pt +2026-04-09 22:14:04 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:14:07 - INFO - parrotllm.training - step 14510 | epoch 0 | loss 3.5492 | lr 2.63e-03 | grad 0.1524 +2026-04-09 22:14:10 - INFO - parrotllm.training - step 14520 | epoch 0 | loss 3.4155 | lr 2.63e-03 | grad 0.1761 +2026-04-09 22:14:13 - INFO - parrotllm.training - step 14530 | epoch 0 | loss 3.4875 | lr 2.63e-03 | grad 0.1749 +2026-04-09 22:14:16 - INFO - parrotllm.training - step 14540 | epoch 0 | loss 3.5567 | lr 2.63e-03 | grad 0.1608 +2026-04-09 22:14:19 - INFO - parrotllm.training - step 14550 | epoch 0 | loss 3.5335 | lr 2.63e-03 | grad 0.1754 +2026-04-09 22:14:22 - INFO - parrotllm.training - step 14560 | epoch 0 | loss 3.5270 | lr 2.63e-03 | grad 0.1667 +2026-04-09 22:14:25 - INFO - parrotllm.training - step 14570 | epoch 0 | loss 3.4588 | lr 2.63e-03 | grad 0.1670 +2026-04-09 22:14:28 - INFO - parrotllm.training - step 14580 | epoch 0 | loss 3.4990 | lr 2.63e-03 | grad 0.1922 +2026-04-09 22:14:31 - INFO - parrotllm.training - step 14590 | epoch 0 | loss 3.5744 | lr 2.63e-03 | grad 0.2113 +2026-04-09 22:14:34 - INFO - parrotllm.training - step 14600 | epoch 0 | loss 3.5113 | lr 2.63e-03 | grad 0.1751 +2026-04-09 22:14:37 - INFO - parrotllm.training - step 14610 | epoch 0 | loss 3.4960 | lr 2.63e-03 | grad 0.1651 +2026-04-09 22:14:40 - INFO - parrotllm.training - step 14620 | epoch 0 | loss 3.5177 | lr 2.63e-03 | grad 0.1824 +2026-04-09 22:14:43 - INFO - parrotllm.training - step 14630 | epoch 0 | loss 3.5574 | lr 2.63e-03 | grad 0.1748 +2026-04-09 22:14:46 - INFO - parrotllm.training - step 14640 | epoch 0 | loss 3.4925 | lr 2.63e-03 | grad 0.1617 +2026-04-09 22:14:50 - INFO - parrotllm.training - step 14650 | epoch 0 | loss 3.6366 | lr 2.63e-03 | grad 0.1671 +2026-04-09 22:14:53 - INFO - parrotllm.training - step 14660 | epoch 0 | loss 3.5507 | lr 2.63e-03 | grad 0.1568 +2026-04-09 22:14:56 - INFO - parrotllm.training - step 14670 | epoch 0 | loss 3.5132 | lr 2.63e-03 | grad 0.1596 +2026-04-09 22:14:59 - INFO - parrotllm.training - step 14680 | epoch 0 | loss 3.6318 | lr 2.63e-03 | grad 0.1632 +2026-04-09 22:15:02 - INFO - parrotllm.training - step 14690 | epoch 0 | loss 3.5085 | lr 2.63e-03 | grad 0.1685 +2026-04-09 22:15:05 - INFO - parrotllm.training - step 14700 | epoch 0 | loss 3.5241 | lr 2.63e-03 | grad 0.1618 +2026-04-09 22:15:08 - INFO - parrotllm.training - step 14710 | epoch 0 | loss 3.4515 | lr 2.63e-03 | grad 0.1565 +2026-04-09 22:15:11 - INFO - parrotllm.training - step 14720 | epoch 0 | loss 3.4958 | lr 2.63e-03 | grad 0.1551 +2026-04-09 22:15:14 - INFO - parrotllm.training - step 14730 | epoch 0 | loss 3.5736 | lr 2.63e-03 | grad 0.1630 +2026-04-09 22:15:17 - INFO - parrotllm.training - step 14740 | epoch 0 | loss 3.5167 | lr 2.63e-03 | grad 0.1547 +2026-04-09 22:15:20 - INFO - parrotllm.training - step 14750 | epoch 0 | loss 3.5078 | lr 2.63e-03 | grad 0.1684 +2026-04-09 22:15:23 - INFO - parrotllm.training - step 14760 | epoch 0 | loss 3.5202 | lr 2.63e-03 | grad 0.1678 +2026-04-09 22:15:26 - INFO - parrotllm.training - step 14770 | epoch 0 | loss 3.5606 | lr 2.63e-03 | grad 0.1947 +2026-04-09 22:15:29 - INFO - parrotllm.training - step 14780 | epoch 0 | loss 3.4888 | lr 2.63e-03 | grad 0.1870 +2026-04-09 22:15:32 - INFO - parrotllm.training - step 14790 | epoch 0 | loss 3.5794 | lr 2.63e-03 | grad 0.1558 +2026-04-09 22:15:35 - INFO - parrotllm.training - step 14800 | epoch 0 | loss 3.5909 | lr 2.63e-03 | grad 0.1547 +2026-04-09 22:15:38 - INFO - parrotllm.training - step 14810 | epoch 0 | loss 3.4051 | lr 2.63e-03 | grad 0.1648 +2026-04-09 22:15:41 - INFO - parrotllm.training - step 14820 | epoch 0 | loss 3.4786 | lr 2.63e-03 | grad 0.1548 +2026-04-09 22:15:44 - INFO - parrotllm.training - step 14830 | epoch 0 | loss 3.4938 | lr 2.63e-03 | grad 0.1763 +2026-04-09 22:15:47 - INFO - parrotllm.training - step 14840 | epoch 0 | loss 3.4939 | lr 2.63e-03 | grad 0.1723 +2026-04-09 22:15:51 - INFO - parrotllm.training - step 14850 | epoch 0 | loss 3.5178 | lr 2.63e-03 | grad 0.1624 +2026-04-09 22:15:54 - INFO - parrotllm.training - step 14860 | epoch 0 | loss 3.5890 | lr 2.63e-03 | grad 0.1540 +2026-04-09 22:15:57 - INFO - parrotllm.training - step 14870 | epoch 0 | loss 3.5601 | lr 2.63e-03 | grad 0.1929 +2026-04-09 22:16:00 - INFO - parrotllm.training - step 14880 | epoch 0 | loss 3.5073 | lr 2.63e-03 | grad 0.1823 +2026-04-09 22:16:03 - INFO - parrotllm.training - step 14890 | epoch 0 | loss 3.5385 | lr 2.63e-03 | grad 0.1508 +2026-04-09 22:16:06 - INFO - parrotllm.training - step 14900 | epoch 0 | loss 3.5162 | lr 2.63e-03 | grad 0.1485 +2026-04-09 22:16:09 - INFO - parrotllm.training - step 14910 | epoch 0 | loss 3.4108 | lr 2.63e-03 | grad 0.1812 +2026-04-09 22:16:12 - INFO - parrotllm.training - step 14920 | epoch 0 | loss 3.4764 | lr 2.63e-03 | grad 0.1496 +2026-04-09 22:16:15 - INFO - parrotllm.training - step 14930 | epoch 0 | loss 3.5585 | lr 2.63e-03 | grad 0.1775 +2026-04-09 22:16:18 - INFO - parrotllm.training - step 14940 | epoch 0 | loss 3.5734 | lr 2.63e-03 | grad 0.1835 +2026-04-09 22:16:21 - INFO - parrotllm.training - step 14950 | epoch 0 | loss 3.4131 | lr 2.63e-03 | grad 0.2042 +2026-04-09 22:16:24 - INFO - parrotllm.training - step 14960 | epoch 0 | loss 3.4727 | lr 2.63e-03 | grad 0.1812 +2026-04-09 22:16:27 - INFO - parrotllm.training - step 14970 | epoch 0 | loss 3.5094 | lr 2.63e-03 | grad 0.1533 +2026-04-09 22:16:30 - INFO - parrotllm.training - step 14980 | epoch 0 | loss 3.5703 | lr 2.63e-03 | grad 0.1607 +2026-04-09 22:16:33 - INFO - parrotllm.training - step 14990 | epoch 0 | loss 3.5183 | lr 2.63e-03 | grad 0.1659 +2026-04-09 22:16:36 - INFO - parrotllm.training - step 15000 | epoch 0 | loss 3.5401 | lr 2.63e-03 | grad 0.1631 +2026-04-09 22:16:36 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:16:36 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:16:39 - INFO - parrotllm.training - Train: loss=3.5401, ppl=34.47 +2026-04-09 22:16:39 - INFO - parrotllm.training - Val: loss=3.4733, ppl=32.24 +2026-04-09 22:16:39 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 22:16:40 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4733_epoch_0000_step_0015000.pt +2026-04-09 22:16:41 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:16:42 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0015000.pt +2026-04-09 22:16:46 - INFO - parrotllm.training - step 15010 | epoch 0 | loss 3.5580 | lr 2.63e-03 | grad 0.1555 +2026-04-09 22:16:49 - INFO - parrotllm.training - step 15020 | epoch 0 | loss 3.5136 | lr 2.63e-03 | grad 0.1986 +2026-04-09 22:16:52 - INFO - parrotllm.training - step 15030 | epoch 0 | loss 3.5655 | lr 2.63e-03 | grad 0.1895 +2026-04-09 22:16:55 - INFO - parrotllm.training - step 15040 | epoch 0 | loss 3.5182 | lr 2.63e-03 | grad 0.1972 +2026-04-09 22:16:58 - INFO - parrotllm.training - step 15050 | epoch 0 | loss 3.4552 | lr 2.63e-03 | grad 0.1713 +2026-04-09 22:17:01 - INFO - parrotllm.training - step 15060 | epoch 0 | loss 3.5579 | lr 2.63e-03 | grad 0.1536 +2026-04-09 22:17:04 - INFO - parrotllm.training - step 15070 | epoch 0 | loss 3.4950 | lr 2.63e-03 | grad 0.1536 +2026-04-09 22:17:07 - INFO - parrotllm.training - step 15080 | epoch 0 | loss 3.3710 | lr 2.63e-03 | grad 0.1435 +2026-04-09 22:17:11 - INFO - parrotllm.training - step 15090 | epoch 0 | loss 3.4316 | lr 2.63e-03 | grad 0.1594 +2026-04-09 22:17:14 - INFO - parrotllm.training - step 15100 | epoch 0 | loss 3.4667 | lr 2.63e-03 | grad 0.1495 +2026-04-09 22:17:17 - INFO - parrotllm.training - step 15110 | epoch 0 | loss 3.5285 | lr 2.63e-03 | grad 0.1621 +2026-04-09 22:17:20 - INFO - parrotllm.training - step 15120 | epoch 0 | loss 3.5245 | lr 2.63e-03 | grad 0.2013 +2026-04-09 22:17:23 - INFO - parrotllm.training - step 15130 | epoch 0 | loss 3.5473 | lr 2.63e-03 | grad 0.1630 +2026-04-09 22:17:26 - INFO - parrotllm.training - step 15140 | epoch 0 | loss 3.4474 | lr 2.63e-03 | grad 0.1668 +2026-04-09 22:17:29 - INFO - parrotllm.training - step 15150 | epoch 0 | loss 3.5876 | lr 2.63e-03 | grad 0.1641 +2026-04-09 22:17:32 - INFO - parrotllm.training - step 15160 | epoch 0 | loss 3.6375 | lr 2.63e-03 | grad 0.1626 +2026-04-09 22:17:35 - INFO - parrotllm.training - step 15170 | epoch 0 | loss 3.4821 | lr 2.63e-03 | grad 0.1636 +2026-04-09 22:17:38 - INFO - parrotllm.training - step 15180 | epoch 0 | loss 3.4911 | lr 2.63e-03 | grad 0.1780 +2026-04-09 22:17:41 - INFO - parrotllm.training - step 15190 | epoch 0 | loss 3.6076 | lr 2.63e-03 | grad 0.1934 +2026-04-09 22:17:44 - INFO - parrotllm.training - step 15200 | epoch 0 | loss 3.5362 | lr 2.63e-03 | grad 0.1604 +2026-04-09 22:17:47 - INFO - parrotllm.training - step 15210 | epoch 0 | loss 3.5418 | lr 2.63e-03 | grad 0.1995 +2026-04-09 22:17:50 - INFO - parrotllm.training - step 15220 | epoch 0 | loss 3.4615 | lr 2.63e-03 | grad 0.1760 +2026-04-09 22:17:53 - INFO - parrotllm.training - step 15230 | epoch 0 | loss 3.4896 | lr 2.63e-03 | grad 0.1829 +2026-04-09 22:17:56 - INFO - parrotllm.training - step 15240 | epoch 0 | loss 3.4411 | lr 2.63e-03 | grad 0.1566 +2026-04-09 22:17:59 - INFO - parrotllm.training - step 15250 | epoch 0 | loss 3.5277 | lr 2.63e-03 | grad 0.1892 +2026-04-09 22:18:02 - INFO - parrotllm.training - step 15260 | epoch 0 | loss 3.5670 | lr 2.63e-03 | grad 0.1527 +2026-04-09 22:18:05 - INFO - parrotllm.training - step 15270 | epoch 0 | loss 3.5427 | lr 2.63e-03 | grad 0.1684 +2026-04-09 22:18:08 - INFO - parrotllm.training - step 15280 | epoch 0 | loss 3.4260 | lr 2.63e-03 | grad 0.1565 +2026-04-09 22:18:11 - INFO - parrotllm.training - step 15290 | epoch 0 | loss 3.4952 | lr 2.63e-03 | grad 0.1612 +2026-04-09 22:18:15 - INFO - parrotllm.training - step 15300 | epoch 0 | loss 3.4648 | lr 2.63e-03 | grad 0.1724 +2026-04-09 22:18:18 - INFO - parrotllm.training - step 15310 | epoch 0 | loss 3.4988 | lr 2.63e-03 | grad 0.1742 +2026-04-09 22:18:21 - INFO - parrotllm.training - step 15320 | epoch 0 | loss 3.4424 | lr 2.63e-03 | grad 0.1944 +2026-04-09 22:18:24 - INFO - parrotllm.training - step 15330 | epoch 0 | loss 3.4489 | lr 2.63e-03 | grad 0.2203 +2026-04-09 22:18:27 - INFO - parrotllm.training - step 15340 | epoch 0 | loss 3.5327 | lr 2.63e-03 | grad 0.1572 +2026-04-09 22:18:30 - INFO - parrotllm.training - step 15350 | epoch 0 | loss 3.5252 | lr 2.63e-03 | grad 0.1621 +2026-04-09 22:18:33 - INFO - parrotllm.training - step 15360 | epoch 0 | loss 3.5280 | lr 2.63e-03 | grad 0.1616 +2026-04-09 22:18:36 - INFO - parrotllm.training - step 15370 | epoch 0 | loss 3.5440 | lr 2.63e-03 | grad 0.1664 +2026-04-09 22:18:39 - INFO - parrotllm.training - step 15380 | epoch 0 | loss 3.5261 | lr 2.63e-03 | grad 0.1771 +2026-04-09 22:18:42 - INFO - parrotllm.training - step 15390 | epoch 0 | loss 3.4258 | lr 2.63e-03 | grad 0.1791 +2026-04-09 22:18:45 - INFO - parrotllm.training - step 15400 | epoch 0 | loss 3.5287 | lr 2.63e-03 | grad 0.1872 +2026-04-09 22:18:48 - INFO - parrotllm.training - step 15410 | epoch 0 | loss 3.5454 | lr 2.63e-03 | grad 0.1852 +2026-04-09 22:18:51 - INFO - parrotllm.training - step 15420 | epoch 0 | loss 3.4487 | lr 2.63e-03 | grad 0.1549 +2026-04-09 22:18:54 - INFO - parrotllm.training - step 15430 | epoch 0 | loss 3.4758 | lr 2.63e-03 | grad 0.1526 +2026-04-09 22:18:57 - INFO - parrotllm.training - step 15440 | epoch 0 | loss 3.4332 | lr 2.63e-03 | grad 0.1638 +2026-04-09 22:19:00 - INFO - parrotllm.training - step 15450 | epoch 0 | loss 3.5771 | lr 2.63e-03 | grad 0.1829 +2026-04-09 22:19:03 - INFO - parrotllm.training - step 15460 | epoch 0 | loss 3.5065 | lr 2.63e-03 | grad 0.1471 +2026-04-09 22:19:06 - INFO - parrotllm.training - step 15470 | epoch 0 | loss 3.4599 | lr 2.63e-03 | grad 0.1779 +2026-04-09 22:19:09 - INFO - parrotllm.training - step 15480 | epoch 0 | loss 3.5470 | lr 2.63e-03 | grad 0.1664 +2026-04-09 22:19:12 - INFO - parrotllm.training - step 15490 | epoch 0 | loss 3.6032 | lr 2.63e-03 | grad 0.2023 +2026-04-09 22:19:15 - INFO - parrotllm.training - step 15500 | epoch 0 | loss 3.6361 | lr 2.63e-03 | grad 0.2035 +2026-04-09 22:19:15 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:19:15 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:19:19 - INFO - parrotllm.training - Train: loss=3.6361, ppl=37.94 +2026-04-09 22:19:19 - INFO - parrotllm.training - Val: loss=3.4744, ppl=32.28 +2026-04-09 22:19:20 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4744_epoch_0000_step_0015500.pt +2026-04-09 22:19:21 - INFO - parrotllm.training - No validation improvement for 1/5 evaluation(s) (best=3.4733, min_delta=0.001000). +2026-04-09 22:19:21 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:19:24 - INFO - parrotllm.training - step 15510 | epoch 0 | loss 3.4843 | lr 2.63e-03 | grad 0.2062 +2026-04-09 22:19:27 - INFO - parrotllm.training - step 15520 | epoch 0 | loss 3.5131 | lr 2.63e-03 | grad 0.1565 +2026-04-09 22:19:30 - INFO - parrotllm.training - step 15530 | epoch 0 | loss 3.5191 | lr 2.63e-03 | grad 0.1596 +2026-04-09 22:19:33 - INFO - parrotllm.training - step 15540 | epoch 0 | loss 3.4758 | lr 2.63e-03 | grad 0.1638 +2026-04-09 22:19:36 - INFO - parrotllm.training - step 15550 | epoch 0 | loss 3.4508 | lr 2.63e-03 | grad 0.1825 +2026-04-09 22:19:39 - INFO - parrotllm.training - step 15560 | epoch 0 | loss 3.4830 | lr 2.63e-03 | grad 0.1951 +2026-04-09 22:19:42 - INFO - parrotllm.training - step 15570 | epoch 0 | loss 3.5979 | lr 2.63e-03 | grad 0.1633 +2026-04-09 22:19:45 - INFO - parrotllm.training - step 15580 | epoch 0 | loss 3.5672 | lr 2.63e-03 | grad 0.1962 +2026-04-09 22:19:48 - INFO - parrotllm.training - step 15590 | epoch 0 | loss 3.5337 | lr 2.63e-03 | grad 0.1644 +2026-04-09 22:19:51 - INFO - parrotllm.training - step 15600 | epoch 0 | loss 3.4800 | lr 2.63e-03 | grad 0.1991 +2026-04-09 22:19:54 - INFO - parrotllm.training - step 15610 | epoch 0 | loss 3.4893 | lr 2.63e-03 | grad 0.1702 +2026-04-09 22:19:57 - INFO - parrotllm.training - step 15620 | epoch 0 | loss 3.4866 | lr 2.63e-03 | grad 0.1888 +2026-04-09 22:20:00 - INFO - parrotllm.training - step 15630 | epoch 0 | loss 3.4553 | lr 2.63e-03 | grad 0.1804 +2026-04-09 22:20:03 - INFO - parrotllm.training - step 15640 | epoch 0 | loss 3.5579 | lr 2.63e-03 | grad 0.1744 +2026-04-09 22:20:06 - INFO - parrotllm.training - step 15650 | epoch 0 | loss 3.4038 | lr 2.63e-03 | grad 0.1632 +2026-04-09 22:20:09 - INFO - parrotllm.training - step 15660 | epoch 0 | loss 3.5074 | lr 2.63e-03 | grad 0.1520 +2026-04-09 22:20:12 - INFO - parrotllm.training - step 15670 | epoch 0 | loss 3.4848 | lr 2.63e-03 | grad 0.1744 +2026-04-09 22:20:15 - INFO - parrotllm.training - step 15680 | epoch 0 | loss 3.4917 | lr 2.63e-03 | grad 0.1972 +2026-04-09 22:20:18 - INFO - parrotllm.training - step 15690 | epoch 0 | loss 3.5286 | lr 2.63e-03 | grad 0.1654 +2026-04-09 22:20:22 - INFO - parrotllm.training - step 15700 | epoch 0 | loss 3.5316 | lr 2.63e-03 | grad 0.1837 +2026-04-09 22:20:25 - INFO - parrotllm.training - step 15710 | epoch 0 | loss 3.5428 | lr 2.63e-03 | grad 0.1611 +2026-04-09 22:20:28 - INFO - parrotllm.training - step 15720 | epoch 0 | loss 3.4237 | lr 2.63e-03 | grad 0.1740 +2026-04-09 22:20:31 - INFO - parrotllm.training - step 15730 | epoch 0 | loss 3.4401 | lr 2.63e-03 | grad 0.1739 +2026-04-09 22:20:34 - INFO - parrotllm.training - step 15740 | epoch 0 | loss 3.5745 | lr 2.63e-03 | grad 0.1742 +2026-04-09 22:20:37 - INFO - parrotllm.training - step 15750 | epoch 0 | loss 3.5503 | lr 2.63e-03 | grad 0.1589 +2026-04-09 22:20:40 - INFO - parrotllm.training - step 15760 | epoch 0 | loss 3.5564 | lr 2.63e-03 | grad 0.1523 +2026-04-09 22:20:43 - INFO - parrotllm.training - step 15770 | epoch 0 | loss 3.5374 | lr 2.63e-03 | grad 0.1601 +2026-04-09 22:20:46 - INFO - parrotllm.training - step 15780 | epoch 0 | loss 3.3844 | lr 2.63e-03 | grad 0.1410 +2026-04-09 22:20:49 - INFO - parrotllm.training - step 15790 | epoch 0 | loss 3.6277 | lr 2.63e-03 | grad 0.1610 +2026-04-09 22:20:52 - INFO - parrotllm.training - step 15800 | epoch 0 | loss 3.5324 | lr 2.63e-03 | grad 0.2443 +2026-04-09 22:20:55 - INFO - parrotllm.training - step 15810 | epoch 0 | loss 3.6098 | lr 2.63e-03 | grad 0.1847 +2026-04-09 22:20:58 - INFO - parrotllm.training - step 15820 | epoch 0 | loss 3.6232 | lr 2.63e-03 | grad 0.1813 +2026-04-09 22:21:01 - INFO - parrotllm.training - step 15830 | epoch 0 | loss 3.5913 | lr 2.63e-03 | grad 0.1545 +2026-04-09 22:21:04 - INFO - parrotllm.training - step 15840 | epoch 0 | loss 3.5090 | lr 2.63e-03 | grad 0.1461 +2026-04-09 22:21:07 - INFO - parrotllm.training - step 15850 | epoch 0 | loss 3.5173 | lr 2.63e-03 | grad 0.1750 +2026-04-09 22:21:10 - INFO - parrotllm.training - step 15860 | epoch 0 | loss 3.4051 | lr 2.63e-03 | grad 0.1608 +2026-04-09 22:21:13 - INFO - parrotllm.training - step 15870 | epoch 0 | loss 3.5696 | lr 2.63e-03 | grad 0.1704 +2026-04-09 22:21:16 - INFO - parrotllm.training - step 15880 | epoch 0 | loss 3.5273 | lr 2.63e-03 | grad 0.1645 +2026-04-09 22:21:19 - INFO - parrotllm.training - step 15890 | epoch 0 | loss 3.5780 | lr 2.63e-03 | grad 0.1739 +2026-04-09 22:21:23 - INFO - parrotllm.training - step 15900 | epoch 0 | loss 3.5904 | lr 2.63e-03 | grad 0.1744 +2026-04-09 22:21:26 - INFO - parrotllm.training - step 15910 | epoch 0 | loss 3.4461 | lr 2.63e-03 | grad 0.1721 +2026-04-09 22:21:29 - INFO - parrotllm.training - step 15920 | epoch 0 | loss 3.5863 | lr 2.63e-03 | grad 0.1597 +2026-04-09 22:21:32 - INFO - parrotllm.training - step 15930 | epoch 0 | loss 3.4829 | lr 2.63e-03 | grad 0.1804 +2026-04-09 22:21:35 - INFO - parrotllm.training - step 15940 | epoch 0 | loss 3.5631 | lr 2.63e-03 | grad 0.1764 +2026-04-09 22:21:38 - INFO - parrotllm.training - step 15950 | epoch 0 | loss 3.5356 | lr 2.63e-03 | grad 0.1635 +2026-04-09 22:21:41 - INFO - parrotllm.training - step 15960 | epoch 0 | loss 3.6130 | lr 2.63e-03 | grad 0.1781 +2026-04-09 22:21:44 - INFO - parrotllm.training - step 15970 | epoch 0 | loss 3.5066 | lr 2.63e-03 | grad 0.1876 +2026-04-09 22:21:47 - INFO - parrotllm.training - step 15980 | epoch 0 | loss 3.5199 | lr 2.63e-03 | grad 0.1386 +2026-04-09 22:21:50 - INFO - parrotllm.training - step 15990 | epoch 0 | loss 3.5007 | lr 2.63e-03 | grad 0.1823 +2026-04-09 22:21:53 - INFO - parrotllm.training - step 16000 | epoch 0 | loss 3.3914 | lr 2.63e-03 | grad 0.1590 +2026-04-09 22:21:53 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:21:53 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:21:56 - INFO - parrotllm.training - Train: loss=3.3914, ppl=29.71 +2026-04-09 22:21:56 - INFO - parrotllm.training - Val: loss=3.4682, ppl=32.08 +2026-04-09 22:21:56 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 22:21:57 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4682_epoch_0000_step_0016000.pt +2026-04-09 22:21:58 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:22:01 - INFO - parrotllm.training - step 16010 | epoch 0 | loss 3.4801 | lr 2.63e-03 | grad 0.1524 +2026-04-09 22:22:04 - INFO - parrotllm.training - step 16020 | epoch 0 | loss 3.5545 | lr 2.63e-03 | grad 0.1558 +2026-04-09 22:22:07 - INFO - parrotllm.training - step 16030 | epoch 0 | loss 3.4975 | lr 2.63e-03 | grad 0.1782 +2026-04-09 22:22:10 - INFO - parrotllm.training - step 16040 | epoch 0 | loss 3.5799 | lr 2.63e-03 | grad 0.1857 +2026-04-09 22:22:13 - INFO - parrotllm.training - step 16050 | epoch 0 | loss 3.4280 | lr 2.63e-03 | grad 0.1827 +2026-04-09 22:22:16 - INFO - parrotllm.training - step 16060 | epoch 0 | loss 3.4189 | lr 2.63e-03 | grad 0.1672 +2026-04-09 22:22:19 - INFO - parrotllm.training - step 16070 | epoch 0 | loss 3.4376 | lr 2.63e-03 | grad 0.2178 +2026-04-09 22:22:22 - INFO - parrotllm.training - step 16080 | epoch 0 | loss 3.5352 | lr 2.63e-03 | grad 0.1819 +2026-04-09 22:22:25 - INFO - parrotllm.training - step 16090 | epoch 0 | loss 3.5626 | lr 2.63e-03 | grad 0.1729 +2026-04-09 22:22:29 - INFO - parrotllm.training - step 16100 | epoch 0 | loss 3.5059 | lr 2.63e-03 | grad 0.1521 +2026-04-09 22:22:32 - INFO - parrotllm.training - step 16110 | epoch 0 | loss 3.6048 | lr 2.63e-03 | grad 0.1502 +2026-04-09 22:22:35 - INFO - parrotllm.training - step 16120 | epoch 0 | loss 3.4724 | lr 2.63e-03 | grad 0.1718 +2026-04-09 22:22:38 - INFO - parrotllm.training - step 16130 | epoch 0 | loss 3.5357 | lr 2.63e-03 | grad 0.1651 +2026-04-09 22:22:41 - INFO - parrotllm.training - step 16140 | epoch 0 | loss 3.5954 | lr 2.63e-03 | grad 0.1602 +2026-04-09 22:22:44 - INFO - parrotllm.training - step 16150 | epoch 0 | loss 3.4904 | lr 2.63e-03 | grad 0.1949 +2026-04-09 22:22:47 - INFO - parrotllm.training - step 16160 | epoch 0 | loss 3.5080 | lr 2.63e-03 | grad 0.2018 +2026-04-09 22:22:50 - INFO - parrotllm.training - step 16170 | epoch 0 | loss 3.5694 | lr 2.63e-03 | grad 0.1793 +2026-04-09 22:22:53 - INFO - parrotllm.training - step 16180 | epoch 0 | loss 3.4973 | lr 2.63e-03 | grad 0.1775 +2026-04-09 22:22:56 - INFO - parrotllm.training - step 16190 | epoch 0 | loss 3.6136 | lr 2.63e-03 | grad 0.1488 +2026-04-09 22:22:59 - INFO - parrotllm.training - step 16200 | epoch 0 | loss 3.4879 | lr 2.63e-03 | grad 0.1941 +2026-04-09 22:23:02 - INFO - parrotllm.training - step 16210 | epoch 0 | loss 3.5007 | lr 2.63e-03 | grad 0.1902 +2026-04-09 22:23:05 - INFO - parrotllm.training - step 16220 | epoch 0 | loss 3.4682 | lr 2.63e-03 | grad 0.1898 +2026-04-09 22:23:08 - INFO - parrotllm.training - step 16230 | epoch 0 | loss 3.5696 | lr 2.63e-03 | grad 0.1611 +2026-04-09 22:23:11 - INFO - parrotllm.training - step 16240 | epoch 0 | loss 3.4092 | lr 2.63e-03 | grad 0.1701 +2026-04-09 22:23:14 - INFO - parrotllm.training - step 16250 | epoch 0 | loss 3.5141 | lr 2.63e-03 | grad 0.1841 +2026-04-09 22:23:17 - INFO - parrotllm.training - step 16260 | epoch 0 | loss 3.4230 | lr 2.63e-03 | grad 0.2041 +2026-04-09 22:23:20 - INFO - parrotllm.training - step 16270 | epoch 0 | loss 3.4695 | lr 2.63e-03 | grad 0.1550 +2026-04-09 22:23:23 - INFO - parrotllm.training - step 16280 | epoch 0 | loss 3.5227 | lr 2.63e-03 | grad 0.1909 +2026-04-09 22:23:26 - INFO - parrotllm.training - step 16290 | epoch 0 | loss 3.5288 | lr 2.63e-03 | grad 0.1723 +2026-04-09 22:23:30 - INFO - parrotllm.training - step 16300 | epoch 0 | loss 3.5806 | lr 2.63e-03 | grad 0.1845 +2026-04-09 22:23:33 - INFO - parrotllm.training - step 16310 | epoch 0 | loss 3.4860 | lr 2.63e-03 | grad 0.1701 +2026-04-09 22:23:36 - INFO - parrotllm.training - step 16320 | epoch 0 | loss 3.4680 | lr 2.63e-03 | grad 0.1651 +2026-04-09 22:23:39 - INFO - parrotllm.training - step 16330 | epoch 0 | loss 3.5436 | lr 2.63e-03 | grad 0.2275 +2026-04-09 22:23:42 - INFO - parrotllm.training - step 16340 | epoch 0 | loss 3.4796 | lr 2.63e-03 | grad 0.1583 +2026-04-09 22:23:45 - INFO - parrotllm.training - step 16350 | epoch 0 | loss 3.5799 | lr 2.63e-03 | grad 0.1627 +2026-04-09 22:23:48 - INFO - parrotllm.training - step 16360 | epoch 0 | loss 3.6291 | lr 2.63e-03 | grad 0.1606 +2026-04-09 22:23:51 - INFO - parrotllm.training - step 16370 | epoch 0 | loss 3.4394 | lr 2.63e-03 | grad 0.1601 +2026-04-09 22:23:54 - INFO - parrotllm.training - step 16380 | epoch 0 | loss 3.6953 | lr 2.63e-03 | grad 0.1734 +2026-04-09 22:23:57 - INFO - parrotllm.training - step 16390 | epoch 0 | loss 3.5484 | lr 2.63e-03 | grad 0.1489 +2026-04-09 22:24:00 - INFO - parrotllm.training - step 16400 | epoch 0 | loss 3.5293 | lr 2.63e-03 | grad 0.1924 +2026-04-09 22:24:03 - INFO - parrotllm.training - step 16410 | epoch 0 | loss 3.5281 | lr 2.63e-03 | grad 0.1690 +2026-04-09 22:24:06 - INFO - parrotllm.training - step 16420 | epoch 0 | loss 3.4531 | lr 2.63e-03 | grad 0.1679 +2026-04-09 22:24:09 - INFO - parrotllm.training - step 16430 | epoch 0 | loss 3.4900 | lr 2.63e-03 | grad 0.1833 +2026-04-09 22:24:12 - INFO - parrotllm.training - step 16440 | epoch 0 | loss 3.4084 | lr 2.63e-03 | grad 0.1524 +2026-04-09 22:24:15 - INFO - parrotllm.training - step 16450 | epoch 0 | loss 3.5682 | lr 2.63e-03 | grad 0.1637 +2026-04-09 22:24:18 - INFO - parrotllm.training - step 16460 | epoch 0 | loss 3.6295 | lr 2.63e-03 | grad 0.1533 +2026-04-09 22:24:21 - INFO - parrotllm.training - step 16470 | epoch 0 | loss 3.5730 | lr 2.63e-03 | grad 0.2062 +2026-04-09 22:24:24 - INFO - parrotllm.training - step 16480 | epoch 0 | loss 3.4832 | lr 2.63e-03 | grad 0.1959 +2026-04-09 22:24:28 - INFO - parrotllm.training - step 16490 | epoch 0 | loss 3.4453 | lr 2.63e-03 | grad 0.1566 +2026-04-09 22:24:31 - INFO - parrotllm.training - step 16500 | epoch 0 | loss 3.6219 | lr 2.63e-03 | grad 0.1788 +2026-04-09 22:24:31 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:24:31 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:24:34 - INFO - parrotllm.training - Train: loss=3.6219, ppl=37.41 +2026-04-09 22:24:34 - INFO - parrotllm.training - Val: loss=3.4703, ppl=32.15 +2026-04-09 22:24:35 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4703_epoch_0000_step_0016500.pt +2026-04-09 22:24:36 - INFO - parrotllm.training - No validation improvement for 1/5 evaluation(s) (best=3.4682, min_delta=0.001000). +2026-04-09 22:24:36 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:24:39 - INFO - parrotllm.training - step 16510 | epoch 0 | loss 3.5947 | lr 2.63e-03 | grad 0.1563 +2026-04-09 22:24:42 - INFO - parrotllm.training - step 16520 | epoch 0 | loss 3.4795 | lr 2.63e-03 | grad 0.1750 +2026-04-09 22:24:45 - INFO - parrotllm.training - step 16530 | epoch 0 | loss 3.5206 | lr 2.63e-03 | grad 0.1745 +2026-04-09 22:24:48 - INFO - parrotllm.training - step 16540 | epoch 0 | loss 3.4268 | lr 2.63e-03 | grad 0.1810 +2026-04-09 22:24:51 - INFO - parrotllm.training - step 16550 | epoch 0 | loss 3.6204 | lr 2.63e-03 | grad 0.1816 +2026-04-09 22:24:54 - INFO - parrotllm.training - step 16560 | epoch 0 | loss 3.4690 | lr 2.63e-03 | grad 0.1717 +2026-04-09 22:24:57 - INFO - parrotllm.training - step 16570 | epoch 0 | loss 3.5614 | lr 2.63e-03 | grad 0.1861 +2026-04-09 22:25:00 - INFO - parrotllm.training - step 16580 | epoch 0 | loss 3.6275 | lr 2.63e-03 | grad 0.1619 +2026-04-09 22:25:03 - INFO - parrotllm.training - step 16590 | epoch 0 | loss 3.6386 | lr 2.63e-03 | grad 0.1724 +2026-04-09 22:25:06 - INFO - parrotllm.training - step 16600 | epoch 0 | loss 3.4852 | lr 2.63e-03 | grad 0.1660 +2026-04-09 22:25:09 - INFO - parrotllm.training - step 16610 | epoch 0 | loss 3.4745 | lr 2.63e-03 | grad 0.1877 +2026-04-09 22:25:12 - INFO - parrotllm.training - step 16620 | epoch 0 | loss 3.5308 | lr 2.63e-03 | grad 0.1448 +2026-04-09 22:25:15 - INFO - parrotllm.training - step 16630 | epoch 0 | loss 3.4692 | lr 2.63e-03 | grad 0.2021 +2026-04-09 22:25:18 - INFO - parrotllm.training - step 16640 | epoch 0 | loss 3.5131 | lr 2.63e-03 | grad 0.1673 +2026-04-09 22:25:21 - INFO - parrotllm.training - step 16650 | epoch 0 | loss 3.5193 | lr 2.63e-03 | grad 0.1812 +2026-04-09 22:25:24 - INFO - parrotllm.training - step 16660 | epoch 0 | loss 3.4967 | lr 2.63e-03 | grad 0.1910 +2026-04-09 22:25:27 - INFO - parrotllm.training - step 16670 | epoch 0 | loss 3.5542 | lr 2.63e-03 | grad 0.1561 +2026-04-09 22:25:31 - INFO - parrotllm.training - step 16680 | epoch 0 | loss 3.4795 | lr 2.63e-03 | grad 0.1623 +2026-04-09 22:25:34 - INFO - parrotllm.training - step 16690 | epoch 0 | loss 3.4973 | lr 2.63e-03 | grad 0.1549 +2026-04-09 22:25:37 - INFO - parrotllm.training - step 16700 | epoch 0 | loss 3.4777 | lr 2.63e-03 | grad 0.1704 +2026-04-09 22:25:40 - INFO - parrotllm.training - step 16710 | epoch 0 | loss 3.4185 | lr 2.63e-03 | grad 0.2106 +2026-04-09 22:25:43 - INFO - parrotllm.training - step 16720 | epoch 0 | loss 3.5775 | lr 2.63e-03 | grad 0.1799 +2026-04-09 22:25:46 - INFO - parrotllm.training - step 16730 | epoch 0 | loss 3.4920 | lr 2.63e-03 | grad 0.1686 +2026-04-09 22:25:49 - INFO - parrotllm.training - step 16740 | epoch 0 | loss 3.4896 | lr 2.63e-03 | grad 0.1798 +2026-04-09 22:25:52 - INFO - parrotllm.training - step 16750 | epoch 0 | loss 3.4917 | lr 2.63e-03 | grad 0.1824 +2026-04-09 22:25:55 - INFO - parrotllm.training - step 16760 | epoch 0 | loss 3.4661 | lr 2.63e-03 | grad 0.1873 +2026-04-09 22:25:58 - INFO - parrotllm.training - step 16770 | epoch 0 | loss 3.4193 | lr 2.63e-03 | grad 0.1670 +2026-04-09 22:26:01 - INFO - parrotllm.training - step 16780 | epoch 0 | loss 3.5052 | lr 2.63e-03 | grad 0.1584 +2026-04-09 22:26:04 - INFO - parrotllm.training - step 16790 | epoch 0 | loss 3.4099 | lr 2.63e-03 | grad 0.1536 +2026-04-09 22:26:07 - INFO - parrotllm.training - step 16800 | epoch 0 | loss 3.4860 | lr 2.63e-03 | grad 0.1717 +2026-04-09 22:26:10 - INFO - parrotllm.training - step 16810 | epoch 0 | loss 3.4692 | lr 2.63e-03 | grad 0.1602 +2026-04-09 22:26:13 - INFO - parrotllm.training - step 16820 | epoch 0 | loss 3.4777 | lr 2.63e-03 | grad 0.1655 +2026-04-09 22:26:16 - INFO - parrotllm.training - step 16830 | epoch 0 | loss 3.4599 | lr 2.63e-03 | grad 0.1673 +2026-04-09 22:26:19 - INFO - parrotllm.training - step 16840 | epoch 0 | loss 3.5268 | lr 2.63e-03 | grad 0.1488 +2026-04-09 22:26:22 - INFO - parrotllm.training - step 16850 | epoch 0 | loss 3.4768 | lr 2.63e-03 | grad 0.1762 +2026-04-09 22:26:25 - INFO - parrotllm.training - step 16860 | epoch 0 | loss 3.4910 | lr 2.63e-03 | grad 0.1969 +2026-04-09 22:26:29 - INFO - parrotllm.training - step 16870 | epoch 0 | loss 3.4699 | lr 2.63e-03 | grad 0.1661 +2026-04-09 22:26:32 - INFO - parrotllm.training - step 16880 | epoch 0 | loss 3.3567 | lr 2.63e-03 | grad 0.1636 +2026-04-09 22:26:35 - INFO - parrotllm.training - step 16890 | epoch 0 | loss 3.5237 | lr 2.63e-03 | grad 0.1698 +2026-04-09 22:26:38 - INFO - parrotllm.training - step 16900 | epoch 0 | loss 3.5111 | lr 2.63e-03 | grad 0.1695 +2026-04-09 22:26:41 - INFO - parrotllm.training - step 16910 | epoch 0 | loss 3.4059 | lr 2.63e-03 | grad 0.2140 +2026-04-09 22:26:44 - INFO - parrotllm.training - step 16920 | epoch 0 | loss 3.4879 | lr 2.63e-03 | grad 0.1814 +2026-04-09 22:26:47 - INFO - parrotllm.training - step 16930 | epoch 0 | loss 3.5134 | lr 2.63e-03 | grad 0.1895 +2026-04-09 22:26:50 - INFO - parrotllm.training - step 16940 | epoch 0 | loss 3.4511 | lr 2.63e-03 | grad 0.2099 +2026-04-09 22:26:53 - INFO - parrotllm.training - step 16950 | epoch 0 | loss 3.4984 | lr 2.63e-03 | grad 0.1610 +2026-04-09 22:26:56 - INFO - parrotllm.training - step 16960 | epoch 0 | loss 3.5023 | lr 2.63e-03 | grad 0.2034 +2026-04-09 22:26:59 - INFO - parrotllm.training - step 16970 | epoch 0 | loss 3.5552 | lr 2.63e-03 | grad 0.1564 +2026-04-09 22:27:02 - INFO - parrotllm.training - step 16980 | epoch 0 | loss 3.5884 | lr 2.63e-03 | grad 0.1713 +2026-04-09 22:27:05 - INFO - parrotllm.training - step 16990 | epoch 0 | loss 3.5069 | lr 2.63e-03 | grad 0.1901 +2026-04-09 22:27:08 - INFO - parrotllm.training - step 17000 | epoch 0 | loss 3.5336 | lr 2.63e-03 | grad 0.1679 +2026-04-09 22:27:08 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:27:08 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:27:11 - INFO - parrotllm.training - Train: loss=3.5336, ppl=34.25 +2026-04-09 22:27:11 - INFO - parrotllm.training - Val: loss=3.4680, ppl=32.07 +2026-04-09 22:27:12 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4680_epoch_0000_step_0017000.pt +2026-04-09 22:27:13 - INFO - parrotllm.training - No validation improvement for 2/5 evaluation(s) (best=3.4682, min_delta=0.001000). +2026-04-09 22:27:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:27:16 - INFO - parrotllm.training - step 17010 | epoch 0 | loss 3.4455 | lr 2.63e-03 | grad 0.1721 +2026-04-09 22:27:20 - INFO - parrotllm.training - step 17020 | epoch 0 | loss 3.4610 | lr 2.63e-03 | grad 0.1785 +2026-04-09 22:27:23 - INFO - parrotllm.training - step 17030 | epoch 0 | loss 3.5408 | lr 2.63e-03 | grad 0.1692 +2026-04-09 22:27:26 - INFO - parrotllm.training - step 17040 | epoch 0 | loss 3.5565 | lr 2.63e-03 | grad 0.1834 +2026-04-09 22:27:29 - INFO - parrotllm.training - step 17050 | epoch 0 | loss 3.5147 | lr 2.63e-03 | grad 0.1762 +2026-04-09 22:27:32 - INFO - parrotllm.training - step 17060 | epoch 0 | loss 3.5110 | lr 2.63e-03 | grad 0.1731 +2026-04-09 22:27:35 - INFO - parrotllm.training - step 17070 | epoch 0 | loss 3.5373 | lr 2.63e-03 | grad 0.1448 +2026-04-09 22:27:38 - INFO - parrotllm.training - step 17080 | epoch 0 | loss 3.5400 | lr 2.63e-03 | grad 0.1612 +2026-04-09 22:27:41 - INFO - parrotllm.training - step 17090 | epoch 0 | loss 3.5227 | lr 2.63e-03 | grad 0.1606 +2026-04-09 22:27:44 - INFO - parrotllm.training - step 17100 | epoch 0 | loss 3.5180 | lr 2.63e-03 | grad 0.1811 +2026-04-09 22:27:47 - INFO - parrotllm.training - step 17110 | epoch 0 | loss 3.5408 | lr 2.63e-03 | grad 0.1844 +2026-04-09 22:27:50 - INFO - parrotllm.training - step 17120 | epoch 0 | loss 3.4359 | lr 2.63e-03 | grad 0.1567 +2026-04-09 22:27:53 - INFO - parrotllm.training - step 17130 | epoch 0 | loss 3.4559 | lr 2.63e-03 | grad 0.1761 +2026-04-09 22:27:56 - INFO - parrotllm.training - step 17140 | epoch 0 | loss 3.4778 | lr 2.63e-03 | grad 0.1795 +2026-04-09 22:27:59 - INFO - parrotllm.training - step 17150 | epoch 0 | loss 3.4985 | lr 2.63e-03 | grad 0.1889 +2026-04-09 22:28:02 - INFO - parrotllm.training - step 17160 | epoch 0 | loss 3.5089 | lr 2.63e-03 | grad 0.1858 +2026-04-09 22:28:05 - INFO - parrotllm.training - step 17170 | epoch 0 | loss 3.5772 | lr 2.63e-03 | grad 0.1722 +2026-04-09 22:28:06 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 17174/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-09 22:28:09 - INFO - parrotllm.training - step 17180 | epoch 0 | loss 3.4151 | lr 2.63e-03 | grad 0.1862 +2026-04-09 22:28:12 - INFO - parrotllm.training - step 17190 | epoch 0 | loss 3.5216 | lr 2.63e-03 | grad 0.1547 +2026-04-09 22:28:15 - INFO - parrotllm.training - step 17200 | epoch 0 | loss 3.4793 | lr 2.63e-03 | grad 0.1826 +2026-04-09 22:28:18 - INFO - parrotllm.training - step 17210 | epoch 0 | loss 3.4799 | lr 2.63e-03 | grad 0.2142 +2026-04-09 22:28:21 - INFO - parrotllm.training - step 17220 | epoch 0 | loss 3.5845 | lr 2.63e-03 | grad 0.1606 +2026-04-09 22:28:24 - INFO - parrotllm.training - step 17230 | epoch 0 | loss 3.5118 | lr 2.63e-03 | grad 0.1719 +2026-04-09 22:28:27 - INFO - parrotllm.training - step 17240 | epoch 0 | loss 3.4722 | lr 2.63e-03 | grad 0.1817 +2026-04-09 22:28:30 - INFO - parrotllm.training - step 17250 | epoch 0 | loss 3.4399 | lr 2.63e-03 | grad 0.1967 +2026-04-09 22:28:33 - INFO - parrotllm.training - step 17260 | epoch 0 | loss 3.5777 | lr 2.63e-03 | grad 0.1636 +2026-04-09 22:28:36 - INFO - parrotllm.training - step 17270 | epoch 0 | loss 3.3964 | lr 2.63e-03 | grad 0.1605 +2026-04-09 22:28:39 - INFO - parrotllm.training - step 17280 | epoch 0 | loss 3.5057 | lr 2.63e-03 | grad 0.1864 +2026-04-09 22:28:42 - INFO - parrotllm.training - step 17290 | epoch 0 | loss 3.4663 | lr 2.63e-03 | grad 0.1675 +2026-04-09 22:28:45 - INFO - parrotllm.training - step 17300 | epoch 0 | loss 3.4342 | lr 2.63e-03 | grad 0.1807 +2026-04-09 22:28:48 - INFO - parrotllm.training - step 17310 | epoch 0 | loss 3.5000 | lr 2.63e-03 | grad 0.1564 +2026-04-09 22:28:51 - INFO - parrotllm.training - step 17320 | epoch 0 | loss 3.4047 | lr 2.63e-03 | grad 0.1800 +2026-04-09 22:28:54 - INFO - parrotllm.training - step 17330 | epoch 0 | loss 3.4375 | lr 2.63e-03 | grad 0.1884 +2026-04-09 22:28:57 - INFO - parrotllm.training - step 17340 | epoch 0 | loss 3.4729 | lr 2.63e-03 | grad 0.2069 +2026-04-09 22:29:00 - INFO - parrotllm.training - step 17350 | epoch 0 | loss 3.6161 | lr 2.63e-03 | grad 0.1594 +2026-04-09 22:29:04 - INFO - parrotllm.training - step 17360 | epoch 0 | loss 3.5637 | lr 2.63e-03 | grad 0.1814 +2026-04-09 22:29:07 - INFO - parrotllm.training - step 17370 | epoch 0 | loss 3.4897 | lr 2.63e-03 | grad 0.1723 +2026-04-09 22:29:10 - INFO - parrotllm.training - step 17380 | epoch 0 | loss 3.4919 | lr 2.63e-03 | grad 0.1632 +2026-04-09 22:29:13 - INFO - parrotllm.training - step 17390 | epoch 0 | loss 3.7105 | lr 2.63e-03 | grad 0.1673 +2026-04-09 22:29:16 - INFO - parrotllm.training - step 17400 | epoch 0 | loss 3.4330 | lr 2.63e-03 | grad 0.1614 +2026-04-09 22:29:19 - INFO - parrotllm.training - step 17410 | epoch 0 | loss 3.5772 | lr 2.63e-03 | grad 0.1675 +2026-04-09 22:29:22 - INFO - parrotllm.training - step 17420 | epoch 0 | loss 3.5485 | lr 2.63e-03 | grad 0.1662 +2026-04-09 22:29:25 - INFO - parrotllm.training - step 17430 | epoch 0 | loss 3.5027 | lr 2.63e-03 | grad 0.1723 +2026-04-09 22:29:28 - INFO - parrotllm.training - step 17440 | epoch 0 | loss 3.4991 | lr 2.63e-03 | grad 0.1646 +2026-04-09 22:29:31 - INFO - parrotllm.training - step 17450 | epoch 0 | loss 3.4355 | lr 2.63e-03 | grad 0.1839 +2026-04-09 22:29:34 - INFO - parrotllm.training - step 17460 | epoch 0 | loss 3.3605 | lr 2.63e-03 | grad 0.1787 +2026-04-09 22:29:37 - INFO - parrotllm.training - step 17470 | epoch 0 | loss 3.4962 | lr 2.63e-03 | grad 0.1715 +2026-04-09 22:29:40 - INFO - parrotllm.training - step 17480 | epoch 0 | loss 3.6124 | lr 2.63e-03 | grad 0.1808 +2026-04-09 22:29:43 - INFO - parrotllm.training - step 17490 | epoch 0 | loss 3.5004 | lr 2.63e-03 | grad 0.1662 +2026-04-09 22:29:46 - INFO - parrotllm.training - step 17500 | epoch 0 | loss 3.5291 | lr 2.63e-03 | grad 0.1572 +2026-04-09 22:29:46 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:29:46 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:29:49 - INFO - parrotllm.training - Train: loss=3.5291, ppl=34.09 +2026-04-09 22:29:49 - INFO - parrotllm.training - Val: loss=3.4645, ppl=31.96 +2026-04-09 22:29:49 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 22:29:50 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4645_epoch_0000_step_0017500.pt +2026-04-09 22:29:51 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:29:52 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0017500.pt +2026-04-09 22:29:56 - INFO - parrotllm.training - step 17510 | epoch 0 | loss 3.4765 | lr 2.63e-03 | grad 0.1733 +2026-04-09 22:30:00 - INFO - parrotllm.training - step 17520 | epoch 0 | loss 3.4257 | lr 2.63e-03 | grad 0.1940 +2026-04-09 22:30:03 - INFO - parrotllm.training - step 17530 | epoch 0 | loss 3.4052 | lr 2.63e-03 | grad 0.2163 +2026-04-09 22:30:06 - INFO - parrotllm.training - step 17540 | epoch 0 | loss 3.4848 | lr 2.63e-03 | grad 0.1770 +2026-04-09 22:30:09 - INFO - parrotllm.training - step 17550 | epoch 0 | loss 3.5353 | lr 2.63e-03 | grad 0.1728 +2026-04-09 22:30:12 - INFO - parrotllm.training - step 17560 | epoch 0 | loss 3.6050 | lr 2.63e-03 | grad 0.1743 +2026-04-09 22:30:15 - INFO - parrotllm.training - step 17570 | epoch 0 | loss 3.4131 | lr 2.63e-03 | grad 0.1575 +2026-04-09 22:30:18 - INFO - parrotllm.training - step 17580 | epoch 0 | loss 3.4595 | lr 2.63e-03 | grad 0.1707 +2026-04-09 22:30:21 - INFO - parrotllm.training - step 17590 | epoch 0 | loss 3.4841 | lr 2.63e-03 | grad 0.1694 +2026-04-09 22:30:24 - INFO - parrotllm.training - step 17600 | epoch 0 | loss 3.5039 | lr 2.63e-03 | grad 0.1597 +2026-04-09 22:30:27 - INFO - parrotllm.training - step 17610 | epoch 0 | loss 3.5585 | lr 2.63e-03 | grad 0.1958 +2026-04-09 22:30:30 - INFO - parrotllm.training - step 17620 | epoch 0 | loss 3.4475 | lr 2.63e-03 | grad 0.1587 +2026-04-09 22:30:33 - INFO - parrotllm.training - step 17630 | epoch 0 | loss 3.5418 | lr 2.63e-03 | grad 0.1469 +2026-04-09 22:30:36 - INFO - parrotllm.training - step 17640 | epoch 0 | loss 3.5094 | lr 2.63e-03 | grad 0.1710 +2026-04-09 22:30:39 - INFO - parrotllm.training - step 17650 | epoch 0 | loss 3.5091 | lr 2.63e-03 | grad 0.2027 +2026-04-09 22:30:42 - INFO - parrotllm.training - step 17660 | epoch 0 | loss 3.4374 | lr 2.63e-03 | grad 0.1686 +2026-04-09 22:30:45 - INFO - parrotllm.training - step 17670 | epoch 0 | loss 3.5011 | lr 2.63e-03 | grad 0.1546 +2026-04-09 22:30:48 - INFO - parrotllm.training - step 17680 | epoch 0 | loss 3.4063 | lr 2.63e-03 | grad 0.1475 +2026-04-09 22:30:51 - INFO - parrotllm.training - step 17690 | epoch 0 | loss 3.4040 | lr 2.63e-03 | grad 0.1750 +2026-04-09 22:30:54 - INFO - parrotllm.training - step 17700 | epoch 0 | loss 3.4510 | lr 2.63e-03 | grad 0.1488 +2026-04-09 22:30:58 - INFO - parrotllm.training - step 17710 | epoch 0 | loss 3.5119 | lr 2.63e-03 | grad 0.1720 +2026-04-09 22:31:01 - INFO - parrotllm.training - step 17720 | epoch 0 | loss 3.4584 | lr 2.63e-03 | grad 0.1651 +2026-04-09 22:31:04 - INFO - parrotllm.training - step 17730 | epoch 0 | loss 3.4060 | lr 2.63e-03 | grad 0.1679 +2026-04-09 22:31:07 - INFO - parrotllm.training - step 17740 | epoch 0 | loss 3.5617 | lr 2.63e-03 | grad 0.1661 +2026-04-09 22:31:10 - INFO - parrotllm.training - step 17750 | epoch 0 | loss 3.4832 | lr 2.63e-03 | grad 0.1835 +2026-04-09 22:31:13 - INFO - parrotllm.training - step 17760 | epoch 0 | loss 3.3902 | lr 2.63e-03 | grad 0.1971 +2026-04-09 22:31:16 - INFO - parrotllm.training - step 17770 | epoch 0 | loss 3.5370 | lr 2.63e-03 | grad 0.1843 +2026-04-09 22:31:19 - INFO - parrotllm.training - step 17780 | epoch 0 | loss 3.5639 | lr 2.63e-03 | grad 0.1681 +2026-04-09 22:31:22 - INFO - parrotllm.training - step 17790 | epoch 0 | loss 3.4585 | lr 2.63e-03 | grad 0.1722 +2026-04-09 22:31:25 - INFO - parrotllm.training - step 17800 | epoch 0 | loss 3.5092 | lr 2.63e-03 | grad 0.1925 +2026-04-09 22:31:28 - INFO - parrotllm.training - step 17810 | epoch 0 | loss 3.5531 | lr 2.63e-03 | grad 0.1834 +2026-04-09 22:31:31 - INFO - parrotllm.training - step 17820 | epoch 0 | loss 3.5440 | lr 2.63e-03 | grad 0.1682 +2026-04-09 22:31:34 - INFO - parrotllm.training - step 17830 | epoch 0 | loss 3.5646 | lr 2.63e-03 | grad 0.1670 +2026-04-09 22:31:37 - INFO - parrotllm.training - step 17840 | epoch 0 | loss 3.4171 | lr 2.63e-03 | grad 0.1621 +2026-04-09 22:31:40 - INFO - parrotllm.training - step 17850 | epoch 0 | loss 3.4316 | lr 2.63e-03 | grad 0.1586 +2026-04-09 22:31:43 - INFO - parrotllm.training - step 17860 | epoch 0 | loss 3.5648 | lr 2.63e-03 | grad 0.1682 +2026-04-09 22:31:46 - INFO - parrotllm.training - step 17870 | epoch 0 | loss 3.5711 | lr 2.63e-03 | grad 0.1913 +2026-04-09 22:31:49 - INFO - parrotllm.training - step 17880 | epoch 0 | loss 3.4504 | lr 2.63e-03 | grad 0.1551 +2026-04-09 22:31:52 - INFO - parrotllm.training - step 17890 | epoch 0 | loss 3.4968 | lr 2.63e-03 | grad 0.1930 +2026-04-09 22:31:55 - INFO - parrotllm.training - step 17900 | epoch 0 | loss 3.4185 | lr 2.63e-03 | grad 0.1749 +2026-04-09 22:31:59 - INFO - parrotllm.training - step 17910 | epoch 0 | loss 3.5105 | lr 2.63e-03 | grad 0.1933 +2026-04-09 22:32:02 - INFO - parrotllm.training - step 17920 | epoch 0 | loss 3.4402 | lr 2.63e-03 | grad 0.1626 +2026-04-09 22:32:05 - INFO - parrotllm.training - step 17930 | epoch 0 | loss 3.5735 | lr 2.63e-03 | grad 0.1923 +2026-04-09 22:32:08 - INFO - parrotllm.training - step 17940 | epoch 0 | loss 3.6029 | lr 2.63e-03 | grad 0.1667 +2026-04-09 22:32:11 - INFO - parrotllm.training - step 17950 | epoch 0 | loss 3.4711 | lr 2.63e-03 | grad 0.1584 +2026-04-09 22:32:14 - INFO - parrotllm.training - step 17960 | epoch 0 | loss 3.5594 | lr 2.63e-03 | grad 0.1649 +2026-04-09 22:32:17 - INFO - parrotllm.training - step 17970 | epoch 0 | loss 3.4352 | lr 2.63e-03 | grad 0.1885 +2026-04-09 22:32:20 - INFO - parrotllm.training - step 17980 | epoch 0 | loss 3.5283 | lr 2.63e-03 | grad 0.1826 +2026-04-09 22:32:23 - INFO - parrotllm.training - step 17990 | epoch 0 | loss 3.4295 | lr 2.63e-03 | grad 0.2059 +2026-04-09 22:32:26 - INFO - parrotllm.training - step 18000 | epoch 0 | loss 3.5094 | lr 2.63e-03 | grad 0.1750 +2026-04-09 22:32:26 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:32:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:32:29 - INFO - parrotllm.training - Train: loss=3.5094, ppl=33.43 +2026-04-09 22:32:29 - INFO - parrotllm.training - Val: loss=3.4646, ppl=31.96 +2026-04-09 22:32:30 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4646_epoch_0000_step_0018000.pt +2026-04-09 22:32:31 - INFO - parrotllm.training - No validation improvement for 1/5 evaluation(s) (best=3.4645, min_delta=0.001000). +2026-04-09 22:32:31 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:32:34 - INFO - parrotllm.training - step 18010 | epoch 0 | loss 3.4281 | lr 2.63e-03 | grad 0.1822 +2026-04-09 22:32:37 - INFO - parrotllm.training - step 18020 | epoch 0 | loss 3.5215 | lr 2.63e-03 | grad 0.1732 +2026-04-09 22:32:40 - INFO - parrotllm.training - step 18030 | epoch 0 | loss 3.4074 | lr 2.63e-03 | grad 0.1676 +2026-04-09 22:32:43 - INFO - parrotllm.training - step 18040 | epoch 0 | loss 3.4712 | lr 2.63e-03 | grad 0.1736 +2026-04-09 22:32:46 - INFO - parrotllm.training - step 18050 | epoch 0 | loss 3.4208 | lr 2.63e-03 | grad 0.1725 +2026-04-09 22:32:49 - INFO - parrotllm.training - step 18060 | epoch 0 | loss 3.4986 | lr 2.63e-03 | grad 0.1742 +2026-04-09 22:32:52 - INFO - parrotllm.training - step 18070 | epoch 0 | loss 3.4657 | lr 2.63e-03 | grad 0.1706 +2026-04-09 22:32:56 - INFO - parrotllm.training - step 18080 | epoch 0 | loss 3.4116 | lr 2.63e-03 | grad 0.1795 +2026-04-09 22:32:59 - INFO - parrotllm.training - step 18090 | epoch 0 | loss 3.4793 | lr 2.63e-03 | grad 0.1568 +2026-04-09 22:33:02 - INFO - parrotllm.training - step 18100 | epoch 0 | loss 3.4888 | lr 2.63e-03 | grad 0.1723 +2026-04-09 22:33:05 - INFO - parrotllm.training - step 18110 | epoch 0 | loss 3.5074 | lr 2.63e-03 | grad 0.1497 +2026-04-09 22:33:08 - INFO - parrotllm.training - step 18120 | epoch 0 | loss 3.4977 | lr 2.63e-03 | grad 0.1562 +2026-04-09 22:33:11 - INFO - parrotllm.training - step 18130 | epoch 0 | loss 3.6201 | lr 2.63e-03 | grad 0.1790 +2026-04-09 22:33:14 - INFO - parrotllm.training - step 18140 | epoch 0 | loss 3.4185 | lr 2.63e-03 | grad 0.1672 +2026-04-09 22:33:17 - INFO - parrotllm.training - step 18150 | epoch 0 | loss 3.4773 | lr 2.63e-03 | grad 0.1734 +2026-04-09 22:33:20 - INFO - parrotllm.training - step 18160 | epoch 0 | loss 3.4687 | lr 2.63e-03 | grad 0.1906 +2026-04-09 22:33:23 - INFO - parrotllm.training - step 18170 | epoch 0 | loss 3.3986 | lr 2.63e-03 | grad 0.1626 +2026-04-09 22:33:26 - INFO - parrotllm.training - step 18180 | epoch 0 | loss 3.5873 | lr 2.63e-03 | grad 0.1503 +2026-04-09 22:33:29 - INFO - parrotllm.training - step 18190 | epoch 0 | loss 3.5090 | lr 2.63e-03 | grad 0.1722 +2026-04-09 22:33:32 - INFO - parrotllm.training - step 18200 | epoch 0 | loss 3.5020 | lr 2.63e-03 | grad 0.1691 +2026-04-09 22:33:35 - INFO - parrotllm.training - step 18210 | epoch 0 | loss 3.4627 | lr 2.63e-03 | grad 0.1688 +2026-04-09 22:33:38 - INFO - parrotllm.training - step 18220 | epoch 0 | loss 3.3716 | lr 2.63e-03 | grad 0.2109 +2026-04-09 22:33:41 - INFO - parrotllm.training - step 18230 | epoch 0 | loss 3.5133 | lr 2.63e-03 | grad 0.1740 +2026-04-09 22:33:44 - INFO - parrotllm.training - step 18240 | epoch 0 | loss 3.5796 | lr 2.63e-03 | grad 0.1647 +2026-04-09 22:33:47 - INFO - parrotllm.training - step 18250 | epoch 0 | loss 3.4739 | lr 2.63e-03 | grad 0.1857 +2026-04-09 22:33:50 - INFO - parrotllm.training - step 18260 | epoch 0 | loss 3.5275 | lr 2.63e-03 | grad 0.1571 +2026-04-09 22:33:53 - INFO - parrotllm.training - step 18270 | epoch 0 | loss 3.5693 | lr 2.63e-03 | grad 0.1540 +2026-04-09 22:33:56 - INFO - parrotllm.training - step 18280 | epoch 0 | loss 3.5600 | lr 2.63e-03 | grad 0.1577 +2026-04-09 22:34:00 - INFO - parrotllm.training - step 18290 | epoch 0 | loss 3.4254 | lr 2.63e-03 | grad 0.1969 +2026-04-09 22:34:03 - INFO - parrotllm.training - step 18300 | epoch 0 | loss 3.5905 | lr 2.63e-03 | grad 0.1790 +2026-04-09 22:34:06 - INFO - parrotllm.training - step 18310 | epoch 0 | loss 3.5058 | lr 2.63e-03 | grad 0.1713 +2026-04-09 22:34:09 - INFO - parrotllm.training - step 18320 | epoch 0 | loss 3.5606 | lr 2.63e-03 | grad 0.1946 +2026-04-09 22:34:12 - INFO - parrotllm.training - step 18330 | epoch 0 | loss 3.4628 | lr 2.63e-03 | grad 0.1886 +2026-04-09 22:34:15 - INFO - parrotllm.training - step 18340 | epoch 0 | loss 3.4458 | lr 2.63e-03 | grad 0.1581 +2026-04-09 22:34:18 - INFO - parrotllm.training - step 18350 | epoch 0 | loss 3.5305 | lr 2.63e-03 | grad 0.1622 +2026-04-09 22:34:21 - INFO - parrotllm.training - step 18360 | epoch 0 | loss 3.4900 | lr 2.63e-03 | grad 0.1855 +2026-04-09 22:34:24 - INFO - parrotllm.training - step 18370 | epoch 0 | loss 3.4594 | lr 2.63e-03 | grad 0.1598 +2026-04-09 22:34:27 - INFO - parrotllm.training - step 18380 | epoch 0 | loss 3.5025 | lr 2.63e-03 | grad 0.1888 +2026-04-09 22:34:30 - INFO - parrotllm.training - step 18390 | epoch 0 | loss 3.3797 | lr 2.63e-03 | grad 0.2250 +2026-04-09 22:34:33 - INFO - parrotllm.training - step 18400 | epoch 0 | loss 3.5203 | lr 2.63e-03 | grad 0.1906 +2026-04-09 22:34:36 - INFO - parrotllm.training - step 18410 | epoch 0 | loss 3.5993 | lr 2.63e-03 | grad 0.2096 +2026-04-09 22:34:39 - INFO - parrotllm.training - step 18420 | epoch 0 | loss 3.5452 | lr 2.63e-03 | grad 0.1815 +2026-04-09 22:34:42 - INFO - parrotllm.training - step 18430 | epoch 0 | loss 3.5323 | lr 2.63e-03 | grad 0.2102 +2026-04-09 22:34:45 - INFO - parrotllm.training - step 18440 | epoch 0 | loss 3.5367 | lr 2.63e-03 | grad 0.1565 +2026-04-09 22:34:48 - INFO - parrotllm.training - step 18450 | epoch 0 | loss 3.5746 | lr 2.63e-03 | grad 0.1679 +2026-04-09 22:34:51 - INFO - parrotllm.training - step 18460 | epoch 0 | loss 3.6099 | lr 2.63e-03 | grad 0.1525 +2026-04-09 22:34:54 - INFO - parrotllm.training - step 18470 | epoch 0 | loss 3.3808 | lr 2.63e-03 | grad 0.1710 +2026-04-09 22:34:58 - INFO - parrotllm.training - step 18480 | epoch 0 | loss 3.5186 | lr 2.63e-03 | grad 0.1593 +2026-04-09 22:35:01 - INFO - parrotllm.training - step 18490 | epoch 0 | loss 3.4791 | lr 2.63e-03 | grad 0.1769 +2026-04-09 22:35:04 - INFO - parrotllm.training - step 18500 | epoch 0 | loss 3.5300 | lr 2.63e-03 | grad 0.1782 +2026-04-09 22:35:04 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:35:04 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:35:07 - INFO - parrotllm.training - Train: loss=3.5300, ppl=34.12 +2026-04-09 22:35:07 - INFO - parrotllm.training - Val: loss=3.4646, ppl=31.96 +2026-04-09 22:35:08 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4646_epoch_0000_step_0018500.pt +2026-04-09 22:35:09 - INFO - parrotllm.training - No validation improvement for 2/5 evaluation(s) (best=3.4645, min_delta=0.001000). +2026-04-09 22:35:09 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:35:12 - INFO - parrotllm.training - step 18510 | epoch 0 | loss 3.5004 | lr 2.63e-03 | grad 0.1684 +2026-04-09 22:35:15 - INFO - parrotllm.training - step 18520 | epoch 0 | loss 3.4357 | lr 2.63e-03 | grad 0.1996 +2026-04-09 22:35:18 - INFO - parrotllm.training - step 18530 | epoch 0 | loss 3.5415 | lr 2.63e-03 | grad 0.1712 +2026-04-09 22:35:21 - INFO - parrotllm.training - step 18540 | epoch 0 | loss 3.5262 | lr 2.63e-03 | grad 0.1682 +2026-04-09 22:35:24 - INFO - parrotllm.training - step 18550 | epoch 0 | loss 3.5236 | lr 2.63e-03 | grad 0.1639 +2026-04-09 22:35:27 - INFO - parrotllm.training - step 18560 | epoch 0 | loss 3.3812 | lr 2.63e-03 | grad 0.1702 +2026-04-09 22:35:30 - INFO - parrotllm.training - step 18570 | epoch 0 | loss 3.4879 | lr 2.63e-03 | grad 0.1441 +2026-04-09 22:35:33 - INFO - parrotllm.training - step 18580 | epoch 0 | loss 3.5334 | lr 2.63e-03 | grad 0.1555 +2026-04-09 22:35:36 - INFO - parrotllm.training - step 18590 | epoch 0 | loss 3.5168 | lr 2.63e-03 | grad 0.1518 +2026-04-09 22:35:39 - INFO - parrotllm.training - step 18600 | epoch 0 | loss 3.3721 | lr 2.63e-03 | grad 0.1934 +2026-04-09 22:35:42 - INFO - parrotllm.training - step 18610 | epoch 0 | loss 3.4523 | lr 2.63e-03 | grad 0.1781 +2026-04-09 22:35:45 - INFO - parrotllm.training - step 18620 | epoch 0 | loss 3.4975 | lr 2.63e-03 | grad 0.2020 +2026-04-09 22:35:48 - INFO - parrotllm.training - step 18630 | epoch 0 | loss 3.4334 | lr 2.63e-03 | grad 0.1569 +2026-04-09 22:35:51 - INFO - parrotllm.training - step 18640 | epoch 0 | loss 3.5042 | lr 2.63e-03 | grad 0.1659 +2026-04-09 22:35:55 - INFO - parrotllm.training - step 18650 | epoch 0 | loss 3.5672 | lr 2.63e-03 | grad 0.2349 +2026-04-09 22:35:58 - INFO - parrotllm.training - step 18660 | epoch 0 | loss 3.4879 | lr 2.63e-03 | grad 0.1570 +2026-04-09 22:36:01 - INFO - parrotllm.training - step 18670 | epoch 0 | loss 3.5316 | lr 2.63e-03 | grad 0.1841 +2026-04-09 22:36:04 - INFO - parrotllm.training - step 18680 | epoch 0 | loss 3.5116 | lr 2.63e-03 | grad 0.1655 +2026-04-09 22:36:07 - INFO - parrotllm.training - step 18690 | epoch 0 | loss 3.5208 | lr 2.63e-03 | grad 0.1755 +2026-04-09 22:36:10 - INFO - parrotllm.training - step 18700 | epoch 0 | loss 3.4013 | lr 2.63e-03 | grad 0.1784 +2026-04-09 22:36:13 - INFO - parrotllm.training - step 18710 | epoch 0 | loss 3.5144 | lr 2.63e-03 | grad 0.1799 +2026-04-09 22:36:16 - INFO - parrotllm.training - step 18720 | epoch 0 | loss 3.4367 | lr 2.63e-03 | grad 0.1902 +2026-04-09 22:36:19 - INFO - parrotllm.training - step 18730 | epoch 0 | loss 3.5271 | lr 2.63e-03 | grad 0.1854 +2026-04-09 22:36:22 - INFO - parrotllm.training - step 18740 | epoch 0 | loss 3.5159 | lr 2.63e-03 | grad 0.1883 +2026-04-09 22:36:25 - INFO - parrotllm.training - step 18750 | epoch 0 | loss 3.4763 | lr 2.63e-03 | grad 0.1650 +2026-04-09 22:36:28 - INFO - parrotllm.training - step 18760 | epoch 0 | loss 3.4474 | lr 2.63e-03 | grad 0.2100 +2026-04-09 22:36:31 - INFO - parrotllm.training - step 18770 | epoch 0 | loss 3.5499 | lr 2.63e-03 | grad 0.1922 +2026-04-09 22:36:34 - INFO - parrotllm.training - step 18780 | epoch 0 | loss 3.5954 | lr 2.63e-03 | grad 0.1965 +2026-04-09 22:36:37 - INFO - parrotllm.training - step 18790 | epoch 0 | loss 3.4558 | lr 2.63e-03 | grad 0.1830 +2026-04-09 22:36:40 - INFO - parrotllm.training - step 18800 | epoch 0 | loss 3.4355 | lr 2.63e-03 | grad 0.1781 +2026-04-09 22:36:43 - INFO - parrotllm.training - step 18810 | epoch 0 | loss 3.5320 | lr 2.63e-03 | grad 0.2230 +2026-04-09 22:36:46 - INFO - parrotllm.training - step 18820 | epoch 0 | loss 3.5293 | lr 2.63e-03 | grad 0.1742 +2026-04-09 22:36:49 - INFO - parrotllm.training - step 18830 | epoch 0 | loss 3.4817 | lr 2.63e-03 | grad 0.1720 +2026-04-09 22:36:52 - INFO - parrotllm.training - step 18840 | epoch 0 | loss 3.5699 | lr 2.63e-03 | grad 0.1780 +2026-04-09 22:36:55 - INFO - parrotllm.training - step 18850 | epoch 0 | loss 3.4734 | lr 2.63e-03 | grad 0.1714 +2026-04-09 22:36:59 - INFO - parrotllm.training - step 18860 | epoch 0 | loss 3.5104 | lr 2.63e-03 | grad 0.1630 +2026-04-09 22:37:02 - INFO - parrotllm.training - step 18870 | epoch 0 | loss 3.4304 | lr 2.63e-03 | grad 0.1811 +2026-04-09 22:37:05 - INFO - parrotllm.training - step 18880 | epoch 0 | loss 3.5781 | lr 2.63e-03 | grad 0.1763 +2026-04-09 22:37:08 - INFO - parrotllm.training - step 18890 | epoch 0 | loss 3.5210 | lr 2.63e-03 | grad 0.1811 +2026-04-09 22:37:11 - INFO - parrotllm.training - step 18900 | epoch 0 | loss 3.3972 | lr 2.63e-03 | grad 0.1898 +2026-04-09 22:37:14 - INFO - parrotllm.training - step 18910 | epoch 0 | loss 3.4816 | lr 2.63e-03 | grad 0.1829 +2026-04-09 22:37:17 - INFO - parrotllm.training - step 18920 | epoch 0 | loss 3.5161 | lr 2.63e-03 | grad 0.1973 +2026-04-09 22:37:20 - INFO - parrotllm.training - step 18930 | epoch 0 | loss 3.4886 | lr 2.63e-03 | grad 0.1937 +2026-04-09 22:37:23 - INFO - parrotllm.training - step 18940 | epoch 0 | loss 3.4949 | lr 2.63e-03 | grad 0.2033 +2026-04-09 22:37:26 - INFO - parrotllm.training - step 18950 | epoch 0 | loss 3.5323 | lr 2.63e-03 | grad 0.1767 +2026-04-09 22:37:29 - INFO - parrotllm.training - step 18960 | epoch 0 | loss 3.4896 | lr 2.63e-03 | grad 0.1643 +2026-04-09 22:37:32 - INFO - parrotllm.training - step 18970 | epoch 0 | loss 3.5119 | lr 2.63e-03 | grad 0.1581 +2026-04-09 22:37:35 - INFO - parrotllm.training - step 18980 | epoch 0 | loss 3.6246 | lr 2.63e-03 | grad 0.1606 +2026-04-09 22:37:38 - INFO - parrotllm.training - step 18990 | epoch 0 | loss 3.4495 | lr 2.63e-03 | grad 0.1577 +2026-04-09 22:37:41 - INFO - parrotllm.training - step 19000 | epoch 0 | loss 3.5509 | lr 2.63e-03 | grad 0.1688 +2026-04-09 22:37:41 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:37:41 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:37:44 - INFO - parrotllm.training - Train: loss=3.5509, ppl=34.85 +2026-04-09 22:37:44 - INFO - parrotllm.training - Val: loss=3.4605, ppl=31.83 +2026-04-09 22:37:44 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 22:37:45 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4605_epoch_0000_step_0019000.pt +2026-04-09 22:37:46 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:37:49 - INFO - parrotllm.training - step 19010 | epoch 0 | loss 3.4659 | lr 2.63e-03 | grad 0.1821 +2026-04-09 22:37:52 - INFO - parrotllm.training - step 19020 | epoch 0 | loss 3.5409 | lr 2.63e-03 | grad 0.1507 +2026-04-09 22:37:56 - INFO - parrotllm.training - step 19030 | epoch 0 | loss 3.4291 | lr 2.63e-03 | grad 0.1807 +2026-04-09 22:37:59 - INFO - parrotllm.training - step 19040 | epoch 0 | loss 3.4358 | lr 2.63e-03 | grad 0.1969 +2026-04-09 22:38:02 - INFO - parrotllm.training - step 19050 | epoch 0 | loss 3.4801 | lr 2.63e-03 | grad 0.1759 +2026-04-09 22:38:05 - INFO - parrotllm.training - step 19060 | epoch 0 | loss 3.4793 | lr 2.63e-03 | grad 0.1720 +2026-04-09 22:38:08 - INFO - parrotllm.training - step 19070 | epoch 0 | loss 3.5721 | lr 2.63e-03 | grad 0.1884 +2026-04-09 22:38:11 - INFO - parrotllm.training - step 19080 | epoch 0 | loss 3.4538 | lr 2.63e-03 | grad 0.1657 +2026-04-09 22:38:14 - INFO - parrotllm.training - step 19090 | epoch 0 | loss 3.4452 | lr 2.63e-03 | grad 0.1606 +2026-04-09 22:38:17 - INFO - parrotllm.training - step 19100 | epoch 0 | loss 3.4322 | lr 2.63e-03 | grad 0.1892 +2026-04-09 22:38:20 - INFO - parrotllm.training - step 19110 | epoch 0 | loss 3.5474 | lr 2.63e-03 | grad 0.1742 +2026-04-09 22:38:23 - INFO - parrotllm.training - step 19120 | epoch 0 | loss 3.4693 | lr 2.63e-03 | grad 0.1547 +2026-04-09 22:38:26 - INFO - parrotllm.training - step 19130 | epoch 0 | loss 3.5285 | lr 2.63e-03 | grad 0.1914 +2026-04-09 22:38:29 - INFO - parrotllm.training - step 19140 | epoch 0 | loss 3.4677 | lr 2.63e-03 | grad 0.1977 +2026-04-09 22:38:32 - INFO - parrotllm.training - step 19150 | epoch 0 | loss 3.4820 | lr 2.63e-03 | grad 0.1772 +2026-04-09 22:38:35 - INFO - parrotllm.training - step 19160 | epoch 0 | loss 3.4612 | lr 2.63e-03 | grad 0.1483 +2026-04-09 22:38:38 - INFO - parrotllm.training - step 19170 | epoch 0 | loss 3.5548 | lr 2.63e-03 | grad 0.1739 +2026-04-09 22:38:41 - INFO - parrotllm.training - step 19180 | epoch 0 | loss 3.5121 | lr 2.63e-03 | grad 0.1905 +2026-04-09 22:38:44 - INFO - parrotllm.training - step 19190 | epoch 0 | loss 3.5144 | lr 2.63e-03 | grad 0.1554 +2026-04-09 22:38:47 - INFO - parrotllm.training - step 19200 | epoch 0 | loss 3.5078 | lr 2.63e-03 | grad 0.1582 +2026-04-09 22:38:50 - INFO - parrotllm.training - step 19210 | epoch 0 | loss 3.4913 | lr 2.63e-03 | grad 0.1926 +2026-04-09 22:38:53 - INFO - parrotllm.training - step 19220 | epoch 0 | loss 3.5248 | lr 2.63e-03 | grad 0.2077 +2026-04-09 22:38:56 - INFO - parrotllm.training - step 19230 | epoch 0 | loss 3.5302 | lr 2.63e-03 | grad 0.1577 +2026-04-09 22:39:00 - INFO - parrotllm.training - step 19240 | epoch 0 | loss 3.5183 | lr 2.63e-03 | grad 0.2006 +2026-04-09 22:39:03 - INFO - parrotllm.training - step 19250 | epoch 0 | loss 3.4983 | lr 2.63e-03 | grad 0.2012 +2026-04-09 22:39:06 - INFO - parrotllm.training - step 19260 | epoch 0 | loss 3.5471 | lr 2.63e-03 | grad 0.1615 +2026-04-09 22:39:09 - INFO - parrotllm.training - step 19270 | epoch 0 | loss 3.5469 | lr 2.63e-03 | grad 0.1463 +2026-04-09 22:39:12 - INFO - parrotllm.training - step 19280 | epoch 0 | loss 3.4623 | lr 2.63e-03 | grad 0.1880 +2026-04-09 22:39:15 - INFO - parrotllm.training - step 19290 | epoch 0 | loss 3.4286 | lr 2.63e-03 | grad 0.1986 +2026-04-09 22:39:18 - INFO - parrotllm.training - step 19300 | epoch 0 | loss 3.6153 | lr 2.63e-03 | grad 0.1597 +2026-04-09 22:39:21 - INFO - parrotllm.training - step 19310 | epoch 0 | loss 3.4094 | lr 2.63e-03 | grad 0.1659 +2026-04-09 22:39:24 - INFO - parrotllm.training - step 19320 | epoch 0 | loss 3.5220 | lr 2.63e-03 | grad 0.1637 +2026-04-09 22:39:27 - INFO - parrotllm.training - step 19330 | epoch 0 | loss 3.5062 | lr 2.63e-03 | grad 0.1744 +2026-04-09 22:39:30 - INFO - parrotllm.training - step 19340 | epoch 0 | loss 3.4640 | lr 2.63e-03 | grad 0.1621 +2026-04-09 22:39:33 - INFO - parrotllm.training - step 19350 | epoch 0 | loss 3.4774 | lr 2.63e-03 | grad 0.1876 +2026-04-09 22:39:36 - INFO - parrotllm.training - step 19360 | epoch 0 | loss 3.5583 | lr 2.63e-03 | grad 0.1786 +2026-04-09 22:39:39 - INFO - parrotllm.training - step 19370 | epoch 0 | loss 3.5052 | lr 2.63e-03 | grad 0.1881 +2026-04-09 22:39:42 - INFO - parrotllm.training - step 19380 | epoch 0 | loss 3.5591 | lr 2.63e-03 | grad 0.1997 +2026-04-09 22:39:45 - INFO - parrotllm.training - step 19390 | epoch 0 | loss 3.4593 | lr 2.63e-03 | grad 0.1635 +2026-04-09 22:39:48 - INFO - parrotllm.training - step 19400 | epoch 0 | loss 3.3562 | lr 2.63e-03 | grad 0.1921 +2026-04-09 22:39:51 - INFO - parrotllm.training - step 19410 | epoch 0 | loss 3.2991 | lr 2.63e-03 | grad 0.1775 +2026-04-09 22:39:54 - INFO - parrotllm.training - step 19420 | epoch 0 | loss 3.4021 | lr 2.63e-03 | grad 0.2334 +2026-04-09 22:39:58 - INFO - parrotllm.training - step 19430 | epoch 0 | loss 3.5609 | lr 2.63e-03 | grad 0.1664 +2026-04-09 22:40:01 - INFO - parrotllm.training - step 19440 | epoch 0 | loss 3.4237 | lr 2.63e-03 | grad 0.1705 +2026-04-09 22:40:04 - INFO - parrotllm.training - step 19450 | epoch 0 | loss 3.5659 | lr 2.63e-03 | grad 0.1651 +2026-04-09 22:40:07 - INFO - parrotllm.training - step 19460 | epoch 0 | loss 3.4874 | lr 2.63e-03 | grad 0.1793 +2026-04-09 22:40:10 - INFO - parrotllm.training - step 19470 | epoch 0 | loss 3.5018 | lr 2.63e-03 | grad 0.2055 +2026-04-09 22:40:13 - INFO - parrotllm.training - step 19480 | epoch 0 | loss 3.5400 | lr 2.63e-03 | grad 0.1509 +2026-04-09 22:40:16 - INFO - parrotllm.training - step 19490 | epoch 0 | loss 3.4187 | lr 2.63e-03 | grad 0.1709 +2026-04-09 22:40:19 - INFO - parrotllm.training - step 19500 | epoch 0 | loss 3.5181 | lr 2.63e-03 | grad 0.1690 +2026-04-09 22:40:19 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:40:19 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:40:22 - INFO - parrotllm.training - Train: loss=3.5181, ppl=33.72 +2026-04-09 22:40:22 - INFO - parrotllm.training - Val: loss=3.4561, ppl=31.69 +2026-04-09 22:40:22 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 22:40:23 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4561_epoch_0000_step_0019500.pt +2026-04-09 22:40:24 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:40:27 - INFO - parrotllm.training - step 19510 | epoch 0 | loss 3.5231 | lr 2.63e-03 | grad 0.1632 +2026-04-09 22:40:30 - INFO - parrotllm.training - step 19520 | epoch 0 | loss 3.5321 | lr 2.63e-03 | grad 0.1613 +2026-04-09 22:40:33 - INFO - parrotllm.training - step 19530 | epoch 0 | loss 3.4341 | lr 2.63e-03 | grad 0.1944 +2026-04-09 22:40:36 - INFO - parrotllm.training - step 19540 | epoch 0 | loss 3.5473 | lr 2.63e-03 | grad 0.1629 +2026-04-09 22:40:39 - INFO - parrotllm.training - step 19550 | epoch 0 | loss 3.5447 | lr 2.63e-03 | grad 0.1757 +2026-04-09 22:40:42 - INFO - parrotllm.training - step 19560 | epoch 0 | loss 3.5303 | lr 2.63e-03 | grad 0.1800 +2026-04-09 22:40:46 - INFO - parrotllm.training - step 19570 | epoch 0 | loss 3.4458 | lr 2.63e-03 | grad 0.1603 +2026-04-09 22:40:49 - INFO - parrotllm.training - step 19580 | epoch 0 | loss 3.5473 | lr 2.63e-03 | grad 0.1726 +2026-04-09 22:40:52 - INFO - parrotllm.training - step 19590 | epoch 0 | loss 3.4686 | lr 2.63e-03 | grad 0.1900 +2026-04-09 22:40:55 - INFO - parrotllm.training - step 19600 | epoch 0 | loss 3.3647 | lr 2.63e-03 | grad 0.1981 +2026-04-09 22:40:58 - INFO - parrotllm.training - step 19610 | epoch 0 | loss 3.5103 | lr 2.63e-03 | grad 0.1726 +2026-04-09 22:41:01 - INFO - parrotllm.training - step 19620 | epoch 0 | loss 3.4781 | lr 2.63e-03 | grad 0.1898 +2026-04-09 22:41:04 - INFO - parrotllm.training - step 19630 | epoch 0 | loss 3.5600 | lr 2.63e-03 | grad 0.2014 +2026-04-09 22:41:07 - INFO - parrotllm.training - step 19640 | epoch 0 | loss 3.3791 | lr 2.63e-03 | grad 0.1781 +2026-04-09 22:41:10 - INFO - parrotllm.training - step 19650 | epoch 0 | loss 3.5559 | lr 2.63e-03 | grad 0.1880 +2026-04-09 22:41:13 - INFO - parrotllm.training - step 19660 | epoch 0 | loss 3.4966 | lr 2.63e-03 | grad 0.1817 +2026-04-09 22:41:16 - INFO - parrotllm.training - step 19670 | epoch 0 | loss 3.4761 | lr 2.63e-03 | grad 0.1889 +2026-04-09 22:41:19 - INFO - parrotllm.training - step 19680 | epoch 0 | loss 3.4918 | lr 2.63e-03 | grad 0.1694 +2026-04-09 22:41:22 - INFO - parrotllm.training - step 19690 | epoch 0 | loss 3.4384 | lr 2.63e-03 | grad 0.2388 +2026-04-09 22:41:25 - INFO - parrotllm.training - step 19700 | epoch 0 | loss 3.4108 | lr 2.63e-03 | grad 0.2062 +2026-04-09 22:41:28 - INFO - parrotllm.training - step 19710 | epoch 0 | loss 3.4909 | lr 2.63e-03 | grad 0.1738 +2026-04-09 22:41:31 - INFO - parrotllm.training - step 19720 | epoch 0 | loss 3.4361 | lr 2.63e-03 | grad 0.1798 +2026-04-09 22:41:34 - INFO - parrotllm.training - step 19730 | epoch 0 | loss 3.5798 | lr 2.63e-03 | grad 0.1726 +2026-04-09 22:41:37 - INFO - parrotllm.training - step 19740 | epoch 0 | loss 3.5124 | lr 2.63e-03 | grad 0.2369 +2026-04-09 22:41:40 - INFO - parrotllm.training - step 19750 | epoch 0 | loss 3.4234 | lr 2.63e-03 | grad 0.1669 +2026-04-09 22:41:43 - INFO - parrotllm.training - step 19760 | epoch 0 | loss 3.4692 | lr 2.63e-03 | grad 0.1563 +2026-04-09 22:41:47 - INFO - parrotllm.training - step 19770 | epoch 0 | loss 3.4828 | lr 2.63e-03 | grad 0.2490 +2026-04-09 22:41:50 - INFO - parrotllm.training - step 19780 | epoch 0 | loss 3.5432 | lr 2.63e-03 | grad 0.1661 +2026-04-09 22:41:53 - INFO - parrotllm.training - step 19790 | epoch 0 | loss 3.4266 | lr 2.63e-03 | grad 0.1809 +2026-04-09 22:41:56 - INFO - parrotllm.training - step 19800 | epoch 0 | loss 3.3776 | lr 2.63e-03 | grad 0.1599 +2026-04-09 22:41:59 - INFO - parrotllm.training - step 19810 | epoch 0 | loss 3.5452 | lr 2.63e-03 | grad 0.1750 +2026-04-09 22:42:02 - INFO - parrotllm.training - step 19820 | epoch 0 | loss 3.3980 | lr 2.63e-03 | grad 0.2398 +2026-04-09 22:42:05 - INFO - parrotllm.training - step 19830 | epoch 0 | loss 3.4555 | lr 2.63e-03 | grad 0.1540 +2026-04-09 22:42:08 - INFO - parrotllm.training - step 19840 | epoch 0 | loss 3.5248 | lr 2.63e-03 | grad 0.1703 +2026-04-09 22:42:11 - INFO - parrotllm.training - step 19850 | epoch 0 | loss 3.5113 | lr 2.63e-03 | grad 0.2061 +2026-04-09 22:42:14 - INFO - parrotllm.training - step 19860 | epoch 0 | loss 3.5130 | lr 2.63e-03 | grad 0.1785 +2026-04-09 22:42:17 - INFO - parrotllm.training - step 19870 | epoch 0 | loss 3.5029 | lr 2.63e-03 | grad 0.1660 +2026-04-09 22:42:20 - INFO - parrotllm.training - step 19880 | epoch 0 | loss 3.4636 | lr 2.63e-03 | grad 0.1842 +2026-04-09 22:42:23 - INFO - parrotllm.training - step 19890 | epoch 0 | loss 3.4825 | lr 2.63e-03 | grad 0.1608 +2026-04-09 22:42:26 - INFO - parrotllm.training - step 19900 | epoch 0 | loss 3.4850 | lr 2.63e-03 | grad 0.2076 +2026-04-09 22:42:29 - INFO - parrotllm.training - step 19910 | epoch 0 | loss 3.5702 | lr 2.63e-03 | grad 0.1715 +2026-04-09 22:42:32 - INFO - parrotllm.training - step 19920 | epoch 0 | loss 3.4314 | lr 2.63e-03 | grad 0.1844 +2026-04-09 22:42:35 - INFO - parrotllm.training - step 19930 | epoch 0 | loss 3.5551 | lr 2.63e-03 | grad 0.1656 +2026-04-09 22:42:38 - INFO - parrotllm.training - step 19940 | epoch 0 | loss 3.4703 | lr 2.63e-03 | grad 0.1535 +2026-04-09 22:42:41 - INFO - parrotllm.training - step 19950 | epoch 0 | loss 3.4368 | lr 2.63e-03 | grad 0.1540 +2026-04-09 22:42:45 - INFO - parrotllm.training - step 19960 | epoch 0 | loss 3.5486 | lr 2.63e-03 | grad 0.1581 +2026-04-09 22:42:48 - INFO - parrotllm.training - step 19970 | epoch 0 | loss 3.4584 | lr 2.63e-03 | grad 0.1805 +2026-04-09 22:42:51 - INFO - parrotllm.training - step 19980 | epoch 0 | loss 3.5671 | lr 2.63e-03 | grad 0.1798 +2026-04-09 22:42:54 - INFO - parrotllm.training - step 19990 | epoch 0 | loss 3.4342 | lr 2.63e-03 | grad 0.1713 +2026-04-09 22:42:57 - INFO - parrotllm.training - step 20000 | epoch 0 | loss 3.4970 | lr 2.63e-03 | grad 0.1900 +2026-04-09 22:42:57 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:42:57 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:43:00 - INFO - parrotllm.training - Train: loss=3.4970, ppl=33.02 +2026-04-09 22:43:00 - INFO - parrotllm.training - Val: loss=3.4581, ppl=31.76 +2026-04-09 22:43:01 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4581_epoch_0000_step_0020000.pt +2026-04-09 22:43:02 - INFO - parrotllm.training - No validation improvement for 1/5 evaluation(s) (best=3.4561, min_delta=0.001000). +2026-04-09 22:43:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:43:03 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0020000.pt +2026-04-09 22:43:07 - INFO - parrotllm.training - step 20010 | epoch 0 | loss 3.5363 | lr 2.63e-03 | grad 0.1862 +2026-04-09 22:43:11 - INFO - parrotllm.training - step 20020 | epoch 0 | loss 3.5236 | lr 2.63e-03 | grad 0.1693 +2026-04-09 22:43:14 - INFO - parrotllm.training - step 20030 | epoch 0 | loss 3.4813 | lr 2.63e-03 | grad 0.1913 +2026-04-09 22:43:17 - INFO - parrotllm.training - step 20040 | epoch 0 | loss 3.4655 | lr 2.63e-03 | grad 0.2229 +2026-04-09 22:43:20 - INFO - parrotllm.training - step 20050 | epoch 0 | loss 3.4885 | lr 2.63e-03 | grad 0.1699 +2026-04-09 22:43:23 - INFO - parrotllm.training - step 20060 | epoch 0 | loss 3.5356 | lr 2.63e-03 | grad 0.1928 +2026-04-09 22:43:26 - INFO - parrotllm.training - step 20070 | epoch 0 | loss 3.5494 | lr 2.63e-03 | grad 0.1589 +2026-04-09 22:43:29 - INFO - parrotllm.training - step 20080 | epoch 0 | loss 3.4855 | lr 2.63e-03 | grad 0.1668 +2026-04-09 22:43:32 - INFO - parrotllm.training - step 20090 | epoch 0 | loss 3.5815 | lr 2.63e-03 | grad 0.1772 +2026-04-09 22:43:35 - INFO - parrotllm.training - step 20100 | epoch 0 | loss 3.5813 | lr 2.63e-03 | grad 0.1654 +2026-04-09 22:43:38 - INFO - parrotllm.training - step 20110 | epoch 0 | loss 3.4845 | lr 2.63e-03 | grad 0.1540 +2026-04-09 22:43:41 - INFO - parrotllm.training - step 20120 | epoch 0 | loss 3.5218 | lr 2.63e-03 | grad 0.1721 +2026-04-09 22:43:44 - INFO - parrotllm.training - step 20130 | epoch 0 | loss 3.5864 | lr 2.63e-03 | grad 0.2047 +2026-04-09 22:43:47 - INFO - parrotllm.training - step 20140 | epoch 0 | loss 3.5506 | lr 2.63e-03 | grad 0.1861 +2026-04-09 22:43:50 - INFO - parrotllm.training - step 20150 | epoch 0 | loss 3.4767 | lr 2.63e-03 | grad 0.1571 +2026-04-09 22:43:53 - INFO - parrotllm.training - step 20160 | epoch 0 | loss 3.5657 | lr 2.63e-03 | grad 0.2251 +2026-04-09 22:43:56 - INFO - parrotllm.training - step 20170 | epoch 0 | loss 3.4687 | lr 2.63e-03 | grad 0.1426 +2026-04-09 22:43:59 - INFO - parrotllm.training - step 20180 | epoch 0 | loss 3.4184 | lr 2.63e-03 | grad 0.1927 +2026-04-09 22:44:02 - INFO - parrotllm.training - step 20190 | epoch 0 | loss 3.5428 | lr 2.63e-03 | grad 0.1593 +2026-04-09 22:44:06 - INFO - parrotllm.training - step 20200 | epoch 0 | loss 3.4510 | lr 2.63e-03 | grad 0.1685 +2026-04-09 22:44:09 - INFO - parrotllm.training - step 20210 | epoch 0 | loss 3.5003 | lr 2.63e-03 | grad 0.1683 +2026-04-09 22:44:12 - INFO - parrotllm.training - step 20220 | epoch 0 | loss 3.5693 | lr 2.63e-03 | grad 0.1727 +2026-04-09 22:44:15 - INFO - parrotllm.training - step 20230 | epoch 0 | loss 3.4384 | lr 2.63e-03 | grad 0.1705 +2026-04-09 22:44:18 - INFO - parrotllm.training - step 20240 | epoch 0 | loss 3.5258 | lr 2.63e-03 | grad 0.1849 +2026-04-09 22:44:21 - INFO - parrotllm.training - step 20250 | epoch 0 | loss 3.3213 | lr 2.63e-03 | grad 0.1926 +2026-04-09 22:44:24 - INFO - parrotllm.training - step 20260 | epoch 0 | loss 3.6108 | lr 2.63e-03 | grad 0.1967 +2026-04-09 22:44:27 - INFO - parrotllm.training - step 20270 | epoch 0 | loss 3.5115 | lr 2.63e-03 | grad 0.1636 +2026-04-09 22:44:30 - INFO - parrotllm.training - step 20280 | epoch 0 | loss 3.4904 | lr 2.63e-03 | grad 0.1613 +2026-04-09 22:44:33 - INFO - parrotllm.training - step 20290 | epoch 0 | loss 3.5345 | lr 2.63e-03 | grad 0.1888 +2026-04-09 22:44:36 - INFO - parrotllm.training - step 20300 | epoch 0 | loss 3.5180 | lr 2.63e-03 | grad 0.2059 +2026-04-09 22:44:39 - INFO - parrotllm.training - step 20310 | epoch 0 | loss 3.3976 | lr 2.63e-03 | grad 0.1835 +2026-04-09 22:44:42 - INFO - parrotllm.training - step 20320 | epoch 0 | loss 3.4797 | lr 2.63e-03 | grad 0.1681 +2026-04-09 22:44:45 - INFO - parrotllm.training - step 20330 | epoch 0 | loss 3.4189 | lr 2.63e-03 | grad 0.2179 +2026-04-09 22:44:48 - INFO - parrotllm.training - step 20340 | epoch 0 | loss 3.4148 | lr 2.63e-03 | grad 0.1496 +2026-04-09 22:44:51 - INFO - parrotllm.training - step 20350 | epoch 0 | loss 3.3198 | lr 2.63e-03 | grad 0.1808 +2026-04-09 22:44:54 - INFO - parrotllm.training - step 20360 | epoch 0 | loss 3.5732 | lr 2.63e-03 | grad 0.1670 +2026-04-09 22:44:57 - INFO - parrotllm.training - step 20370 | epoch 0 | loss 3.6016 | lr 2.63e-03 | grad 0.1587 +2026-04-09 22:45:00 - INFO - parrotllm.training - step 20380 | epoch 0 | loss 3.5869 | lr 2.63e-03 | grad 0.1868 +2026-04-09 22:45:03 - INFO - parrotllm.training - step 20390 | epoch 0 | loss 3.4502 | lr 2.63e-03 | grad 0.2241 +2026-04-09 22:45:07 - INFO - parrotllm.training - step 20400 | epoch 0 | loss 3.4589 | lr 2.63e-03 | grad 0.1744 +2026-04-09 22:45:10 - INFO - parrotllm.training - step 20410 | epoch 0 | loss 3.5214 | lr 2.63e-03 | grad 0.1705 +2026-04-09 22:45:13 - INFO - parrotllm.training - step 20420 | epoch 0 | loss 3.4498 | lr 2.63e-03 | grad 0.1873 +2026-04-09 22:45:16 - INFO - parrotllm.training - step 20430 | epoch 0 | loss 3.5197 | lr 2.63e-03 | grad 0.1689 +2026-04-09 22:45:19 - INFO - parrotllm.training - step 20440 | epoch 0 | loss 3.5418 | lr 2.63e-03 | grad 0.1794 +2026-04-09 22:45:22 - INFO - parrotllm.training - step 20450 | epoch 0 | loss 3.4585 | lr 2.63e-03 | grad 0.1662 +2026-04-09 22:45:25 - INFO - parrotllm.training - step 20460 | epoch 0 | loss 3.4713 | lr 2.63e-03 | grad 0.1676 +2026-04-09 22:45:28 - INFO - parrotllm.training - step 20470 | epoch 0 | loss 3.5765 | lr 2.63e-03 | grad 0.2039 +2026-04-09 22:45:31 - INFO - parrotllm.training - step 20480 | epoch 0 | loss 3.4984 | lr 2.63e-03 | grad 0.1781 +2026-04-09 22:45:34 - INFO - parrotllm.training - step 20490 | epoch 0 | loss 3.5518 | lr 2.63e-03 | grad 0.1750 +2026-04-09 22:45:37 - INFO - parrotllm.training - step 20500 | epoch 0 | loss 3.5255 | lr 2.63e-03 | grad 0.1679 +2026-04-09 22:45:37 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:45:37 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:45:40 - INFO - parrotllm.training - Train: loss=3.5255, ppl=33.97 +2026-04-09 22:45:40 - INFO - parrotllm.training - Val: loss=3.4581, ppl=31.76 +2026-04-09 22:45:41 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4581_epoch_0000_step_0020500.pt +2026-04-09 22:45:42 - INFO - parrotllm.training - No validation improvement for 2/5 evaluation(s) (best=3.4561, min_delta=0.001000). +2026-04-09 22:45:42 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:45:45 - INFO - parrotllm.training - step 20510 | epoch 0 | loss 3.5741 | lr 2.63e-03 | grad 0.1812 +2026-04-09 22:45:48 - INFO - parrotllm.training - step 20520 | epoch 0 | loss 3.6040 | lr 2.63e-03 | grad 0.1600 +2026-04-09 22:45:51 - INFO - parrotllm.training - step 20530 | epoch 0 | loss 3.5659 | lr 2.63e-03 | grad 0.1827 +2026-04-09 22:45:54 - INFO - parrotllm.training - step 20540 | epoch 0 | loss 3.4516 | lr 2.63e-03 | grad 0.1651 +2026-04-09 22:45:57 - INFO - parrotllm.training - step 20550 | epoch 0 | loss 3.4486 | lr 2.63e-03 | grad 0.1719 +2026-04-09 22:46:00 - INFO - parrotllm.training - step 20560 | epoch 0 | loss 3.6109 | lr 2.63e-03 | grad 0.1713 +2026-04-09 22:46:03 - INFO - parrotllm.training - step 20570 | epoch 0 | loss 3.4244 | lr 2.63e-03 | grad 0.1912 +2026-04-09 22:46:07 - INFO - parrotllm.training - step 20580 | epoch 0 | loss 3.4696 | lr 2.63e-03 | grad 0.1584 +2026-04-09 22:46:10 - INFO - parrotllm.training - step 20590 | epoch 0 | loss 3.4654 | lr 2.63e-03 | grad 0.1873 +2026-04-09 22:46:13 - INFO - parrotllm.training - step 20600 | epoch 0 | loss 3.4377 | lr 2.63e-03 | grad 0.1627 +2026-04-09 22:46:16 - INFO - parrotllm.training - step 20610 | epoch 0 | loss 3.5236 | lr 2.63e-03 | grad 0.1657 +2026-04-09 22:46:19 - INFO - parrotllm.training - step 20620 | epoch 0 | loss 3.4725 | lr 2.63e-03 | grad 0.2044 +2026-04-09 22:46:22 - INFO - parrotllm.training - step 20630 | epoch 0 | loss 3.5560 | lr 2.63e-03 | grad 0.1703 +2026-04-09 22:46:25 - INFO - parrotllm.training - step 20640 | epoch 0 | loss 3.5261 | lr 2.63e-03 | grad 0.1925 +2026-04-09 22:46:28 - INFO - parrotllm.training - step 20650 | epoch 0 | loss 3.5934 | lr 2.63e-03 | grad 0.1645 +2026-04-09 22:46:31 - INFO - parrotllm.training - step 20660 | epoch 0 | loss 3.5684 | lr 2.63e-03 | grad 0.2032 +2026-04-09 22:46:34 - INFO - parrotllm.training - step 20670 | epoch 0 | loss 3.5317 | lr 2.63e-03 | grad 0.1984 +2026-04-09 22:46:37 - INFO - parrotllm.training - step 20680 | epoch 0 | loss 3.5100 | lr 2.63e-03 | grad 0.1935 +2026-04-09 22:46:40 - INFO - parrotllm.training - step 20690 | epoch 0 | loss 3.4626 | lr 2.63e-03 | grad 0.1495 +2026-04-09 22:46:43 - INFO - parrotllm.training - step 20700 | epoch 0 | loss 3.4901 | lr 2.63e-03 | grad 0.1597 +2026-04-09 22:46:46 - INFO - parrotllm.training - step 20710 | epoch 0 | loss 3.3811 | lr 2.63e-03 | grad 0.1660 +2026-04-09 22:46:49 - INFO - parrotllm.training - step 20720 | epoch 0 | loss 3.5059 | lr 2.63e-03 | grad 0.1711 +2026-04-09 22:46:52 - INFO - parrotllm.training - step 20730 | epoch 0 | loss 3.4606 | lr 2.63e-03 | grad 0.1706 +2026-04-09 22:46:55 - INFO - parrotllm.training - step 20740 | epoch 0 | loss 3.5203 | lr 2.63e-03 | grad 0.2158 +2026-04-09 22:46:58 - INFO - parrotllm.training - step 20750 | epoch 0 | loss 3.4897 | lr 2.63e-03 | grad 0.1871 +2026-04-09 22:47:01 - INFO - parrotllm.training - step 20760 | epoch 0 | loss 3.4358 | lr 2.63e-03 | grad 0.1672 +2026-04-09 22:47:04 - INFO - parrotllm.training - step 20770 | epoch 0 | loss 3.5280 | lr 2.63e-03 | grad 0.1725 +2026-04-09 22:47:08 - INFO - parrotllm.training - step 20780 | epoch 0 | loss 3.3773 | lr 2.63e-03 | grad 0.1894 +2026-04-09 22:47:11 - INFO - parrotllm.training - step 20790 | epoch 0 | loss 3.5158 | lr 2.63e-03 | grad 0.1896 +2026-04-09 22:47:14 - INFO - parrotllm.training - step 20800 | epoch 0 | loss 3.4510 | lr 2.63e-03 | grad 0.1874 +2026-04-09 22:47:17 - INFO - parrotllm.training - step 20810 | epoch 0 | loss 3.5241 | lr 2.63e-03 | grad 0.1533 +2026-04-09 22:47:20 - INFO - parrotllm.training - step 20820 | epoch 0 | loss 3.4285 | lr 2.63e-03 | grad 0.1958 +2026-04-09 22:47:23 - INFO - parrotllm.training - step 20830 | epoch 0 | loss 3.4628 | lr 2.63e-03 | grad 0.2078 +2026-04-09 22:47:26 - INFO - parrotllm.training - step 20840 | epoch 0 | loss 3.5218 | lr 2.63e-03 | grad 0.1837 +2026-04-09 22:47:29 - INFO - parrotllm.training - step 20850 | epoch 0 | loss 3.5414 | lr 2.63e-03 | grad 0.1623 +2026-04-09 22:47:32 - INFO - parrotllm.training - step 20860 | epoch 0 | loss 3.6292 | lr 2.63e-03 | grad 0.1774 +2026-04-09 22:47:35 - INFO - parrotllm.training - step 20870 | epoch 0 | loss 3.4119 | lr 2.63e-03 | grad 0.2058 +2026-04-09 22:47:38 - INFO - parrotllm.training - step 20880 | epoch 0 | loss 3.4666 | lr 2.63e-03 | grad 0.1924 +2026-04-09 22:47:41 - INFO - parrotllm.training - step 20890 | epoch 0 | loss 3.4476 | lr 2.63e-03 | grad 0.1705 +2026-04-09 22:47:44 - INFO - parrotllm.training - step 20900 | epoch 0 | loss 3.4630 | lr 2.63e-03 | grad 0.1713 +2026-04-09 22:47:47 - INFO - parrotllm.training - step 20910 | epoch 0 | loss 3.5098 | lr 2.63e-03 | grad 0.1699 +2026-04-09 22:47:50 - INFO - parrotllm.training - step 20920 | epoch 0 | loss 3.4234 | lr 2.63e-03 | grad 0.1782 +2026-04-09 22:47:53 - INFO - parrotllm.training - step 20930 | epoch 0 | loss 3.5189 | lr 2.63e-03 | grad 0.1682 +2026-04-09 22:47:56 - INFO - parrotllm.training - step 20940 | epoch 0 | loss 3.5358 | lr 2.63e-03 | grad 0.1804 +2026-04-09 22:47:59 - INFO - parrotllm.training - step 20950 | epoch 0 | loss 3.4747 | lr 2.63e-03 | grad 0.1981 +2026-04-09 22:48:02 - INFO - parrotllm.training - step 20960 | epoch 0 | loss 3.4990 | lr 2.63e-03 | grad 0.1636 +2026-04-09 22:48:06 - INFO - parrotllm.training - step 20970 | epoch 0 | loss 3.5080 | lr 2.63e-03 | grad 0.1890 +2026-04-09 22:48:09 - INFO - parrotllm.training - step 20980 | epoch 0 | loss 3.4838 | lr 2.63e-03 | grad 0.1485 +2026-04-09 22:48:12 - INFO - parrotllm.training - step 20990 | epoch 0 | loss 3.5334 | lr 2.63e-03 | grad 0.1539 +2026-04-09 22:48:15 - INFO - parrotllm.training - step 21000 | epoch 0 | loss 3.4254 | lr 2.63e-03 | grad 0.2044 +2026-04-09 22:48:15 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:48:15 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:48:18 - INFO - parrotllm.training - Train: loss=3.4254, ppl=30.73 +2026-04-09 22:48:18 - INFO - parrotllm.training - Val: loss=3.4616, ppl=31.87 +2026-04-09 22:48:19 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4616_epoch_0000_step_0021000.pt +2026-04-09 22:48:20 - INFO - parrotllm.training - No validation improvement for 3/5 evaluation(s) (best=3.4561, min_delta=0.001000). +2026-04-09 22:48:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:48:23 - INFO - parrotllm.training - step 21010 | epoch 0 | loss 3.6115 | lr 2.63e-03 | grad 0.1771 +2026-04-09 22:48:26 - INFO - parrotllm.training - step 21020 | epoch 0 | loss 3.3980 | lr 2.63e-03 | grad 0.1787 +2026-04-09 22:48:29 - INFO - parrotllm.training - step 21030 | epoch 0 | loss 3.5239 | lr 2.63e-03 | grad 0.1736 +2026-04-09 22:48:32 - INFO - parrotllm.training - step 21040 | epoch 0 | loss 3.4110 | lr 2.63e-03 | grad 0.1940 +2026-04-09 22:48:35 - INFO - parrotllm.training - step 21050 | epoch 0 | loss 3.5045 | lr 2.63e-03 | grad 0.1849 +2026-04-09 22:48:38 - INFO - parrotllm.training - step 21060 | epoch 0 | loss 3.4588 | lr 2.63e-03 | grad 0.2089 +2026-04-09 22:48:41 - INFO - parrotllm.training - step 21070 | epoch 0 | loss 3.5475 | lr 2.63e-03 | grad 0.1838 +2026-04-09 22:48:44 - INFO - parrotllm.training - step 21080 | epoch 0 | loss 3.5231 | lr 2.63e-03 | grad 0.1746 +2026-04-09 22:48:47 - INFO - parrotllm.training - step 21090 | epoch 0 | loss 3.3801 | lr 2.63e-03 | grad 0.1800 +2026-04-09 22:48:50 - INFO - parrotllm.training - step 21100 | epoch 0 | loss 3.5012 | lr 2.63e-03 | grad 0.1436 +2026-04-09 22:48:53 - INFO - parrotllm.training - step 21110 | epoch 0 | loss 3.5176 | lr 2.63e-03 | grad 0.1641 +2026-04-09 22:48:56 - INFO - parrotllm.training - step 21120 | epoch 0 | loss 3.4197 | lr 2.63e-03 | grad 0.1837 +2026-04-09 22:48:59 - INFO - parrotllm.training - step 21130 | epoch 0 | loss 3.4819 | lr 2.63e-03 | grad 0.1590 +2026-04-09 22:49:02 - INFO - parrotllm.training - step 21140 | epoch 0 | loss 3.5889 | lr 2.63e-03 | grad 0.1544 +2026-04-09 22:49:06 - INFO - parrotllm.training - step 21150 | epoch 0 | loss 3.4823 | lr 2.63e-03 | grad 0.1974 +2026-04-09 22:49:09 - INFO - parrotllm.training - step 21160 | epoch 0 | loss 3.5552 | lr 2.63e-03 | grad 0.2185 +2026-04-09 22:49:12 - INFO - parrotllm.training - step 21170 | epoch 0 | loss 3.5174 | lr 2.63e-03 | grad 0.2178 +2026-04-09 22:49:15 - INFO - parrotllm.training - step 21180 | epoch 0 | loss 3.5868 | lr 2.63e-03 | grad 0.1667 +2026-04-09 22:49:18 - INFO - parrotllm.training - step 21190 | epoch 0 | loss 3.4244 | lr 2.63e-03 | grad 0.1629 +2026-04-09 22:49:21 - INFO - parrotllm.training - step 21200 | epoch 0 | loss 3.4682 | lr 2.63e-03 | grad 0.1758 +2026-04-09 22:49:23 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 21211/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-09 22:49:24 - INFO - parrotllm.training - step 21210 | epoch 0 | loss 3.4729 | lr 2.63e-03 | grad 0.1838 +2026-04-09 22:49:27 - INFO - parrotllm.training - step 21220 | epoch 0 | loss 3.4242 | lr 2.63e-03 | grad 0.1561 +2026-04-09 22:49:30 - INFO - parrotllm.training - step 21230 | epoch 0 | loss 3.5266 | lr 2.63e-03 | grad 0.1644 +2026-04-09 22:49:33 - INFO - parrotllm.training - step 21240 | epoch 0 | loss 3.4469 | lr 2.63e-03 | grad 0.1884 +2026-04-09 22:49:36 - INFO - parrotllm.training - step 21250 | epoch 0 | loss 3.5258 | lr 2.63e-03 | grad 0.1891 +2026-04-09 22:49:39 - INFO - parrotllm.training - step 21260 | epoch 0 | loss 3.3980 | lr 2.63e-03 | grad 0.1758 +2026-04-09 22:49:42 - INFO - parrotllm.training - step 21270 | epoch 0 | loss 3.4651 | lr 2.63e-03 | grad 0.1941 +2026-04-09 22:49:45 - INFO - parrotllm.training - step 21280 | epoch 0 | loss 3.4980 | lr 2.63e-03 | grad 0.2126 +2026-04-09 22:49:49 - INFO - parrotllm.training - step 21290 | epoch 0 | loss 3.5172 | lr 2.63e-03 | grad 0.1858 +2026-04-09 22:49:52 - INFO - parrotllm.training - step 21300 | epoch 0 | loss 3.5529 | lr 2.63e-03 | grad 0.1528 +2026-04-09 22:49:55 - INFO - parrotllm.training - step 21310 | epoch 0 | loss 3.4240 | lr 2.63e-03 | grad 0.1441 +2026-04-09 22:49:58 - INFO - parrotllm.training - step 21320 | epoch 0 | loss 3.5904 | lr 2.63e-03 | grad 0.1546 +2026-04-09 22:50:01 - INFO - parrotllm.training - step 21330 | epoch 0 | loss 3.3811 | lr 2.63e-03 | grad 0.1733 +2026-04-09 22:50:04 - INFO - parrotllm.training - step 21340 | epoch 0 | loss 3.5269 | lr 2.63e-03 | grad 0.1647 +2026-04-09 22:50:07 - INFO - parrotllm.training - step 21350 | epoch 0 | loss 3.5496 | lr 2.63e-03 | grad 0.1956 +2026-04-09 22:50:10 - INFO - parrotllm.training - step 21360 | epoch 0 | loss 3.4438 | lr 2.63e-03 | grad 0.1665 +2026-04-09 22:50:13 - INFO - parrotllm.training - step 21370 | epoch 0 | loss 3.5179 | lr 2.63e-03 | grad 0.1696 +2026-04-09 22:50:16 - INFO - parrotllm.training - step 21380 | epoch 0 | loss 3.4630 | lr 2.63e-03 | grad 0.1828 +2026-04-09 22:50:19 - INFO - parrotllm.training - step 21390 | epoch 0 | loss 3.4722 | lr 2.63e-03 | grad 0.1671 +2026-04-09 22:50:22 - INFO - parrotllm.training - step 21400 | epoch 0 | loss 3.4562 | lr 2.63e-03 | grad 0.1698 +2026-04-09 22:50:25 - INFO - parrotllm.training - step 21410 | epoch 0 | loss 3.5280 | lr 2.63e-03 | grad 0.1848 +2026-04-09 22:50:28 - INFO - parrotllm.training - step 21420 | epoch 0 | loss 3.5658 | lr 2.63e-03 | grad 0.1802 +2026-04-09 22:50:31 - INFO - parrotllm.training - step 21430 | epoch 0 | loss 3.5065 | lr 2.63e-03 | grad 0.1599 +2026-04-09 22:50:34 - INFO - parrotllm.training - step 21440 | epoch 0 | loss 3.5084 | lr 2.63e-03 | grad 0.2109 +2026-04-09 22:50:37 - INFO - parrotllm.training - step 21450 | epoch 0 | loss 3.5413 | lr 2.63e-03 | grad 0.2010 +2026-04-09 22:50:40 - INFO - parrotllm.training - step 21460 | epoch 0 | loss 3.4265 | lr 2.63e-03 | grad 0.1812 +2026-04-09 22:50:43 - INFO - parrotllm.training - step 21470 | epoch 0 | loss 3.4776 | lr 2.63e-03 | grad 0.2051 +2026-04-09 22:50:47 - INFO - parrotllm.training - step 21480 | epoch 0 | loss 3.5567 | lr 2.63e-03 | grad 0.1654 +2026-04-09 22:50:50 - INFO - parrotllm.training - step 21490 | epoch 0 | loss 3.4692 | lr 2.63e-03 | grad 0.1561 +2026-04-09 22:50:53 - INFO - parrotllm.training - step 21500 | epoch 0 | loss 3.4678 | lr 2.63e-03 | grad 0.2151 +2026-04-09 22:50:53 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:50:53 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:50:56 - INFO - parrotllm.training - Train: loss=3.4678, ppl=32.07 +2026-04-09 22:50:56 - INFO - parrotllm.training - Val: loss=3.4555, ppl=31.68 +2026-04-09 22:50:57 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4555_epoch_0000_step_0021500.pt +2026-04-09 22:50:58 - INFO - parrotllm.training - No validation improvement for 4/5 evaluation(s) (best=3.4561, min_delta=0.001000). +2026-04-09 22:50:58 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:51:01 - INFO - parrotllm.training - step 21510 | epoch 0 | loss 3.5341 | lr 2.63e-03 | grad 0.1626 +2026-04-09 22:51:04 - INFO - parrotllm.training - step 21520 | epoch 0 | loss 3.4474 | lr 2.63e-03 | grad 0.1741 +2026-04-09 22:51:07 - INFO - parrotllm.training - step 21530 | epoch 0 | loss 3.3651 | lr 2.63e-03 | grad 0.1702 +2026-04-09 22:51:10 - INFO - parrotllm.training - step 21540 | epoch 0 | loss 3.6099 | lr 2.63e-03 | grad 0.1864 +2026-04-09 22:51:13 - INFO - parrotllm.training - step 21550 | epoch 0 | loss 3.3786 | lr 2.63e-03 | grad 0.1791 +2026-04-09 22:51:16 - INFO - parrotllm.training - step 21560 | epoch 0 | loss 3.5383 | lr 2.63e-03 | grad 0.1688 +2026-04-09 22:51:19 - INFO - parrotllm.training - step 21570 | epoch 0 | loss 3.5217 | lr 2.63e-03 | grad 0.1910 +2026-04-09 22:51:22 - INFO - parrotllm.training - step 21580 | epoch 0 | loss 3.4966 | lr 2.63e-03 | grad 0.1971 +2026-04-09 22:51:25 - INFO - parrotllm.training - step 21590 | epoch 0 | loss 3.4687 | lr 2.63e-03 | grad 0.2023 +2026-04-09 22:51:28 - INFO - parrotllm.training - step 21600 | epoch 0 | loss 3.4812 | lr 2.63e-03 | grad 0.1885 +2026-04-09 22:51:31 - INFO - parrotllm.training - step 21610 | epoch 0 | loss 3.6036 | lr 2.63e-03 | grad 0.1628 +2026-04-09 22:51:34 - INFO - parrotllm.training - step 21620 | epoch 0 | loss 3.4712 | lr 2.63e-03 | grad 0.1779 +2026-04-09 22:51:37 - INFO - parrotllm.training - step 21630 | epoch 0 | loss 3.5469 | lr 2.63e-03 | grad 0.2259 +2026-04-09 22:51:40 - INFO - parrotllm.training - step 21640 | epoch 0 | loss 3.4257 | lr 2.63e-03 | grad 0.1912 +2026-04-09 22:51:43 - INFO - parrotllm.training - step 21650 | epoch 0 | loss 3.4659 | lr 2.63e-03 | grad 0.1972 +2026-04-09 22:51:46 - INFO - parrotllm.training - step 21660 | epoch 0 | loss 3.5532 | lr 2.63e-03 | grad 0.1735 +2026-04-09 22:51:49 - INFO - parrotllm.training - step 21670 | epoch 0 | loss 3.5192 | lr 2.63e-03 | grad 0.1721 +2026-04-09 22:51:53 - INFO - parrotllm.training - step 21680 | epoch 0 | loss 3.5869 | lr 2.63e-03 | grad 0.1683 +2026-04-09 22:51:56 - INFO - parrotllm.training - step 21690 | epoch 0 | loss 3.5164 | lr 2.63e-03 | grad 0.1796 +2026-04-09 22:51:59 - INFO - parrotllm.training - step 21700 | epoch 0 | loss 3.4848 | lr 2.63e-03 | grad 0.1809 +2026-04-09 22:52:00 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 21710/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-09 22:52:02 - INFO - parrotllm.training - step 21710 | epoch 0 | loss 3.3752 | lr 2.63e-03 | grad 0.1631 +2026-04-09 22:52:05 - INFO - parrotllm.training - step 21720 | epoch 0 | loss 3.6414 | lr 2.63e-03 | grad 0.1771 +2026-04-09 22:52:08 - INFO - parrotllm.training - step 21730 | epoch 0 | loss 3.4336 | lr 2.63e-03 | grad 0.1732 +2026-04-09 22:52:11 - INFO - parrotllm.training - step 21740 | epoch 0 | loss 3.5019 | lr 2.63e-03 | grad 0.1635 +2026-04-09 22:52:14 - INFO - parrotllm.training - step 21750 | epoch 0 | loss 3.4521 | lr 2.63e-03 | grad 0.1977 +2026-04-09 22:52:17 - INFO - parrotllm.training - step 21760 | epoch 0 | loss 3.5318 | lr 2.63e-03 | grad 0.2029 +2026-04-09 22:52:20 - INFO - parrotllm.training - step 21770 | epoch 0 | loss 3.4487 | lr 2.63e-03 | grad 0.1854 +2026-04-09 22:52:23 - INFO - parrotllm.training - step 21780 | epoch 0 | loss 3.4228 | lr 2.63e-03 | grad 0.1942 +2026-04-09 22:52:26 - INFO - parrotllm.training - step 21790 | epoch 0 | loss 3.5684 | lr 2.63e-03 | grad 0.1728 +2026-04-09 22:52:29 - INFO - parrotllm.training - step 21800 | epoch 0 | loss 3.4768 | lr 2.63e-03 | grad 0.1883 +2026-04-09 22:52:32 - INFO - parrotllm.training - step 21810 | epoch 0 | loss 3.4389 | lr 2.63e-03 | grad 0.1631 +2026-04-09 22:52:36 - INFO - parrotllm.training - step 21820 | epoch 0 | loss 3.5528 | lr 2.63e-03 | grad 0.1684 +2026-04-09 22:52:39 - INFO - parrotllm.training - step 21830 | epoch 0 | loss 3.4470 | lr 2.63e-03 | grad 0.1657 +2026-04-09 22:52:42 - INFO - parrotllm.training - step 21840 | epoch 0 | loss 3.4675 | lr 2.63e-03 | grad 0.1820 +2026-04-09 22:52:45 - INFO - parrotllm.training - step 21850 | epoch 0 | loss 3.5249 | lr 2.63e-03 | grad 0.1987 +2026-04-09 22:52:48 - INFO - parrotllm.training - step 21860 | epoch 0 | loss 3.5560 | lr 2.63e-03 | grad 0.1762 +2026-04-09 22:52:51 - INFO - parrotllm.training - step 21870 | epoch 0 | loss 3.4274 | lr 2.63e-03 | grad 0.1685 +2026-04-09 22:52:54 - INFO - parrotllm.training - step 21880 | epoch 0 | loss 3.5111 | lr 2.63e-03 | grad 0.1633 +2026-04-09 22:52:57 - INFO - parrotllm.training - step 21890 | epoch 0 | loss 3.4307 | lr 2.63e-03 | grad 0.2000 +2026-04-09 22:53:00 - INFO - parrotllm.training - step 21900 | epoch 0 | loss 3.5435 | lr 2.63e-03 | grad 0.1917 +2026-04-09 22:53:03 - INFO - parrotllm.training - step 21910 | epoch 0 | loss 3.5128 | lr 2.63e-03 | grad 0.1700 +2026-04-09 22:53:06 - INFO - parrotllm.training - step 21920 | epoch 0 | loss 3.5074 | lr 2.63e-03 | grad 0.1610 +2026-04-09 22:53:09 - INFO - parrotllm.training - step 21930 | epoch 0 | loss 3.4552 | lr 2.63e-03 | grad 0.1565 +2026-04-09 22:53:12 - INFO - parrotllm.training - step 21940 | epoch 0 | loss 3.5495 | lr 2.63e-03 | grad 0.1985 +2026-04-09 22:53:15 - INFO - parrotllm.training - step 21950 | epoch 0 | loss 3.4408 | lr 2.63e-03 | grad 0.1890 +2026-04-09 22:53:18 - INFO - parrotllm.training - step 21960 | epoch 0 | loss 3.5688 | lr 2.63e-03 | grad 0.1936 +2026-04-09 22:53:21 - INFO - parrotllm.training - step 21970 | epoch 0 | loss 3.6145 | lr 2.63e-03 | grad 0.1833 +2026-04-09 22:53:24 - INFO - parrotllm.training - step 21980 | epoch 0 | loss 3.5650 | lr 2.63e-03 | grad 0.1909 +2026-04-09 22:53:27 - INFO - parrotllm.training - step 21990 | epoch 0 | loss 3.4643 | lr 2.63e-03 | grad 0.1944 +2026-04-09 22:53:30 - INFO - parrotllm.training - step 22000 | epoch 0 | loss 3.5467 | lr 2.63e-03 | grad 0.1816 +2026-04-09 22:53:30 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:53:30 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:53:33 - INFO - parrotllm.training - Train: loss=3.5467, ppl=34.70 +2026-04-09 22:53:33 - INFO - parrotllm.training - Val: loss=3.4510, ppl=31.53 +2026-04-09 22:53:33 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-09 22:53:34 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4510_epoch_0000_step_0022000.pt +2026-04-09 22:53:35 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:53:38 - INFO - parrotllm.training - step 22010 | epoch 0 | loss 3.6080 | lr 2.63e-03 | grad 0.1894 +2026-04-09 22:53:41 - INFO - parrotllm.training - step 22020 | epoch 0 | loss 3.5368 | lr 2.63e-03 | grad 0.1885 +2026-04-09 22:53:45 - INFO - parrotllm.training - step 22030 | epoch 0 | loss 3.4379 | lr 2.63e-03 | grad 0.1942 +2026-04-09 22:53:48 - INFO - parrotllm.training - step 22040 | epoch 0 | loss 3.6120 | lr 2.63e-03 | grad 0.1943 +2026-04-09 22:53:51 - INFO - parrotllm.training - step 22050 | epoch 0 | loss 3.5117 | lr 2.63e-03 | grad 0.2165 +2026-04-09 22:53:54 - INFO - parrotllm.training - step 22060 | epoch 0 | loss 3.4617 | lr 2.63e-03 | grad 0.1739 +2026-04-09 22:53:57 - INFO - parrotllm.training - step 22070 | epoch 0 | loss 3.4756 | lr 2.63e-03 | grad 0.1630 +2026-04-09 22:54:00 - INFO - parrotllm.training - step 22080 | epoch 0 | loss 3.4752 | lr 2.63e-03 | grad 0.1615 +2026-04-09 22:54:03 - INFO - parrotllm.training - step 22090 | epoch 0 | loss 3.6171 | lr 2.63e-03 | grad 0.1810 +2026-04-09 22:54:06 - INFO - parrotllm.training - step 22100 | epoch 0 | loss 3.5750 | lr 2.63e-03 | grad 0.1777 +2026-04-09 22:54:09 - INFO - parrotllm.training - step 22110 | epoch 0 | loss 3.3755 | lr 2.63e-03 | grad 0.1564 +2026-04-09 22:54:12 - INFO - parrotllm.training - step 22120 | epoch 0 | loss 3.4738 | lr 2.63e-03 | grad 0.1815 +2026-04-09 22:54:15 - INFO - parrotllm.training - step 22130 | epoch 0 | loss 3.6277 | lr 2.63e-03 | grad 0.1642 +2026-04-09 22:54:18 - INFO - parrotllm.training - step 22140 | epoch 0 | loss 3.5134 | lr 2.63e-03 | grad 0.1884 +2026-04-09 22:54:21 - INFO - parrotllm.training - step 22150 | epoch 0 | loss 3.5855 | lr 2.63e-03 | grad 0.2025 +2026-04-09 22:54:24 - INFO - parrotllm.training - step 22160 | epoch 0 | loss 3.4428 | lr 2.63e-03 | grad 0.1807 +2026-04-09 22:54:27 - INFO - parrotllm.training - step 22170 | epoch 0 | loss 3.5239 | lr 2.63e-03 | grad 0.1717 +2026-04-09 22:54:30 - INFO - parrotllm.training - step 22180 | epoch 0 | loss 3.4587 | lr 2.63e-03 | grad 0.1785 +2026-04-09 22:54:33 - INFO - parrotllm.training - step 22190 | epoch 0 | loss 3.6333 | lr 2.63e-03 | grad 0.1809 +2026-04-09 22:54:36 - INFO - parrotllm.training - step 22200 | epoch 0 | loss 3.6224 | lr 2.63e-03 | grad 0.2020 +2026-04-09 22:54:39 - INFO - parrotllm.training - step 22210 | epoch 0 | loss 3.4544 | lr 2.63e-03 | grad 0.1715 +2026-04-09 22:54:42 - INFO - parrotllm.training - step 22220 | epoch 0 | loss 3.6391 | lr 2.63e-03 | grad 0.1675 +2026-04-09 22:54:45 - INFO - parrotllm.training - step 22230 | epoch 0 | loss 3.4765 | lr 2.63e-03 | grad 0.1662 +2026-04-09 22:54:49 - INFO - parrotllm.training - step 22240 | epoch 0 | loss 3.5390 | lr 2.63e-03 | grad 0.2079 +2026-04-09 22:54:52 - INFO - parrotllm.training - step 22250 | epoch 0 | loss 3.4579 | lr 2.63e-03 | grad 0.2158 +2026-04-09 22:54:55 - INFO - parrotllm.training - step 22260 | epoch 0 | loss 3.5432 | lr 2.63e-03 | grad 0.1686 +2026-04-09 22:54:58 - INFO - parrotllm.training - step 22270 | epoch 0 | loss 3.5272 | lr 2.63e-03 | grad 0.2026 +2026-04-09 22:55:01 - INFO - parrotllm.training - step 22280 | epoch 0 | loss 3.3973 | lr 2.63e-03 | grad 0.1701 +2026-04-09 22:55:04 - INFO - parrotllm.training - step 22290 | epoch 0 | loss 3.4519 | lr 2.63e-03 | grad 0.1826 +2026-04-09 22:55:07 - INFO - parrotllm.training - step 22300 | epoch 0 | loss 3.5444 | lr 2.63e-03 | grad 0.1810 +2026-04-09 22:55:10 - INFO - parrotllm.training - step 22310 | epoch 0 | loss 3.5021 | lr 2.63e-03 | grad 0.1774 +2026-04-09 22:55:13 - INFO - parrotllm.training - step 22320 | epoch 0 | loss 3.5932 | lr 2.63e-03 | grad 0.1759 +2026-04-09 22:55:16 - INFO - parrotllm.training - step 22330 | epoch 0 | loss 3.4289 | lr 2.63e-03 | grad 0.1524 +2026-04-09 22:55:19 - INFO - parrotllm.training - step 22340 | epoch 0 | loss 3.4824 | lr 2.63e-03 | grad 0.1806 +2026-04-09 22:55:22 - INFO - parrotllm.training - step 22350 | epoch 0 | loss 3.4117 | lr 2.63e-03 | grad 0.1621 +2026-04-09 22:55:25 - INFO - parrotllm.training - step 22360 | epoch 0 | loss 3.4960 | lr 2.63e-03 | grad 0.1605 +2026-04-09 22:55:28 - INFO - parrotllm.training - step 22370 | epoch 0 | loss 3.5589 | lr 2.63e-03 | grad 0.2084 +2026-04-09 22:55:31 - INFO - parrotllm.training - step 22380 | epoch 0 | loss 3.3616 | lr 2.63e-03 | grad 0.1715 +2026-04-09 22:55:34 - INFO - parrotllm.training - step 22390 | epoch 0 | loss 3.5329 | lr 2.63e-03 | grad 0.1699 +2026-04-09 22:55:37 - INFO - parrotllm.training - step 22400 | epoch 0 | loss 3.5313 | lr 2.63e-03 | grad 0.1853 +2026-04-09 22:55:40 - INFO - parrotllm.training - step 22410 | epoch 0 | loss 3.4600 | lr 2.63e-03 | grad 0.1792 +2026-04-09 22:55:43 - INFO - parrotllm.training - step 22420 | epoch 0 | loss 3.5046 | lr 2.63e-03 | grad 0.1988 +2026-04-09 22:55:46 - INFO - parrotllm.training - step 22430 | epoch 0 | loss 3.4270 | lr 2.63e-03 | grad 0.1906 +2026-04-09 22:55:49 - INFO - parrotllm.training - step 22440 | epoch 0 | loss 3.4970 | lr 2.63e-03 | grad 0.2094 +2026-04-09 22:55:53 - INFO - parrotllm.training - step 22450 | epoch 0 | loss 3.5662 | lr 2.63e-03 | grad 0.1577 +2026-04-09 22:55:56 - INFO - parrotllm.training - step 22460 | epoch 0 | loss 3.5585 | lr 2.63e-03 | grad 0.1811 +2026-04-09 22:55:59 - INFO - parrotllm.training - step 22470 | epoch 0 | loss 3.5686 | lr 2.63e-03 | grad 0.1557 +2026-04-09 22:56:02 - INFO - parrotllm.training - step 22480 | epoch 0 | loss 3.5043 | lr 2.63e-03 | grad 0.1698 +2026-04-09 22:56:05 - INFO - parrotllm.training - step 22490 | epoch 0 | loss 3.4570 | lr 2.63e-03 | grad 0.2180 +2026-04-09 22:56:08 - INFO - parrotllm.training - step 22500 | epoch 0 | loss 3.4062 | lr 2.63e-03 | grad 0.1697 +2026-04-09 22:56:08 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:56:08 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:56:11 - INFO - parrotllm.training - Train: loss=3.4062, ppl=30.15 +2026-04-09 22:56:11 - INFO - parrotllm.training - Val: loss=3.4518, ppl=31.56 +2026-04-09 22:56:12 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4518_epoch_0000_step_0022500.pt +2026-04-09 22:56:13 - INFO - parrotllm.training - No validation improvement for 1/5 evaluation(s) (best=3.4510, min_delta=0.001000). +2026-04-09 22:56:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:56:14 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0022500.pt +2026-04-09 22:56:18 - INFO - parrotllm.training - step 22510 | epoch 0 | loss 3.5433 | lr 2.63e-03 | grad 0.1676 +2026-04-09 22:56:21 - INFO - parrotllm.training - step 22520 | epoch 0 | loss 3.6046 | lr 2.63e-03 | grad 0.1805 +2026-04-09 22:56:24 - INFO - parrotllm.training - step 22530 | epoch 0 | loss 3.5082 | lr 2.63e-03 | grad 0.1626 +2026-04-09 22:56:28 - INFO - parrotllm.training - step 22540 | epoch 0 | loss 3.3786 | lr 2.63e-03 | grad 0.1820 +2026-04-09 22:56:31 - INFO - parrotllm.training - step 22550 | epoch 0 | loss 3.5018 | lr 2.63e-03 | grad 0.1740 +2026-04-09 22:56:34 - INFO - parrotllm.training - step 22560 | epoch 0 | loss 3.5075 | lr 2.63e-03 | grad 0.1955 +2026-04-09 22:56:37 - INFO - parrotllm.training - step 22570 | epoch 0 | loss 3.4175 | lr 2.63e-03 | grad 0.1777 +2026-04-09 22:56:40 - INFO - parrotllm.training - step 22580 | epoch 0 | loss 3.4662 | lr 2.63e-03 | grad 0.1732 +2026-04-09 22:56:43 - INFO - parrotllm.training - step 22590 | epoch 0 | loss 3.4961 | lr 2.63e-03 | grad 0.2030 +2026-04-09 22:56:46 - INFO - parrotllm.training - step 22600 | epoch 0 | loss 3.4572 | lr 2.63e-03 | grad 0.1534 +2026-04-09 22:56:49 - INFO - parrotllm.training - step 22610 | epoch 0 | loss 3.4353 | lr 2.63e-03 | grad 0.1705 +2026-04-09 22:56:52 - INFO - parrotllm.training - step 22620 | epoch 0 | loss 3.4769 | lr 2.63e-03 | grad 0.1652 +2026-04-09 22:56:55 - INFO - parrotllm.training - step 22630 | epoch 0 | loss 3.4831 | lr 2.63e-03 | grad 0.1864 +2026-04-09 22:56:58 - INFO - parrotllm.training - step 22640 | epoch 0 | loss 3.4398 | lr 2.63e-03 | grad 0.1855 +2026-04-09 22:57:01 - INFO - parrotllm.training - step 22650 | epoch 0 | loss 3.4625 | lr 2.63e-03 | grad 0.1805 +2026-04-09 22:57:04 - INFO - parrotllm.training - step 22660 | epoch 0 | loss 3.4746 | lr 2.63e-03 | grad 0.1710 +2026-04-09 22:57:07 - INFO - parrotllm.training - step 22670 | epoch 0 | loss 3.4888 | lr 2.63e-03 | grad 0.2000 +2026-04-09 22:57:10 - INFO - parrotllm.training - step 22680 | epoch 0 | loss 3.4201 | lr 2.63e-03 | grad 0.1669 +2026-04-09 22:57:13 - INFO - parrotllm.training - step 22690 | epoch 0 | loss 3.5707 | lr 2.63e-03 | grad 0.1761 +2026-04-09 22:57:16 - INFO - parrotllm.training - step 22700 | epoch 0 | loss 3.4515 | lr 2.63e-03 | grad 0.1689 +2026-04-09 22:57:19 - INFO - parrotllm.training - step 22710 | epoch 0 | loss 3.4465 | lr 2.63e-03 | grad 0.1512 +2026-04-09 22:57:22 - INFO - parrotllm.training - step 22720 | epoch 0 | loss 3.5199 | lr 2.63e-03 | grad 0.1749 +2026-04-09 22:57:25 - INFO - parrotllm.training - step 22730 | epoch 0 | loss 3.5917 | lr 2.63e-03 | grad 0.1996 +2026-04-09 22:57:28 - INFO - parrotllm.training - step 22740 | epoch 0 | loss 3.4477 | lr 2.63e-03 | grad 0.1800 +2026-04-09 22:57:31 - INFO - parrotllm.training - step 22750 | epoch 0 | loss 3.4961 | lr 2.63e-03 | grad 0.1644 +2026-04-09 22:57:35 - INFO - parrotllm.training - step 22760 | epoch 0 | loss 3.4151 | lr 2.63e-03 | grad 0.2193 +2026-04-09 22:57:38 - INFO - parrotllm.training - step 22770 | epoch 0 | loss 3.5053 | lr 2.63e-03 | grad 0.1475 +2026-04-09 22:57:41 - INFO - parrotllm.training - step 22780 | epoch 0 | loss 3.5794 | lr 2.63e-03 | grad 0.1661 +2026-04-09 22:57:44 - INFO - parrotllm.training - step 22790 | epoch 0 | loss 3.5618 | lr 2.63e-03 | grad 0.1921 +2026-04-09 22:57:47 - INFO - parrotllm.training - step 22800 | epoch 0 | loss 3.5053 | lr 2.63e-03 | grad 0.1884 +2026-04-09 22:57:50 - INFO - parrotllm.training - step 22810 | epoch 0 | loss 3.5548 | lr 2.63e-03 | grad 0.1623 +2026-04-09 22:57:53 - INFO - parrotllm.training - step 22820 | epoch 0 | loss 3.5011 | lr 2.63e-03 | grad 0.1684 +2026-04-09 22:57:56 - INFO - parrotllm.training - step 22830 | epoch 0 | loss 3.4774 | lr 2.63e-03 | grad 0.1962 +2026-04-09 22:57:59 - INFO - parrotllm.training - step 22840 | epoch 0 | loss 3.5140 | lr 2.63e-03 | grad 0.1597 +2026-04-09 22:58:02 - INFO - parrotllm.training - step 22850 | epoch 0 | loss 3.4794 | lr 2.63e-03 | grad 0.1460 +2026-04-09 22:58:05 - INFO - parrotllm.training - step 22860 | epoch 0 | loss 3.3576 | lr 2.63e-03 | grad 0.1745 +2026-04-09 22:58:08 - INFO - parrotllm.training - step 22870 | epoch 0 | loss 3.4631 | lr 2.63e-03 | grad 0.1706 +2026-04-09 22:58:11 - INFO - parrotllm.training - step 22880 | epoch 0 | loss 3.4576 | lr 2.63e-03 | grad 0.1708 +2026-04-09 22:58:14 - INFO - parrotllm.training - step 22890 | epoch 0 | loss 3.4273 | lr 2.63e-03 | grad 0.1751 +2026-04-09 22:58:17 - INFO - parrotllm.training - step 22900 | epoch 0 | loss 3.4114 | lr 2.63e-03 | grad 0.1966 +2026-04-09 22:58:20 - INFO - parrotllm.training - step 22910 | epoch 0 | loss 3.4231 | lr 2.63e-03 | grad 0.1716 +2026-04-09 22:58:23 - INFO - parrotllm.training - step 22920 | epoch 0 | loss 3.4366 | lr 2.63e-03 | grad 0.1732 +2026-04-09 22:58:26 - INFO - parrotllm.training - step 22930 | epoch 0 | loss 3.4865 | lr 2.63e-03 | grad 0.1819 +2026-04-09 22:58:29 - INFO - parrotllm.training - step 22940 | epoch 0 | loss 3.4898 | lr 2.63e-03 | grad 0.1682 +2026-04-09 22:58:32 - INFO - parrotllm.training - step 22950 | epoch 0 | loss 3.4919 | lr 2.63e-03 | grad 0.1700 +2026-04-09 22:58:36 - INFO - parrotllm.training - step 22960 | epoch 0 | loss 3.5087 | lr 2.63e-03 | grad 0.1596 +2026-04-09 22:58:39 - INFO - parrotllm.training - step 22970 | epoch 0 | loss 3.5225 | lr 2.63e-03 | grad 0.2370 +2026-04-09 22:58:42 - INFO - parrotllm.training - step 22980 | epoch 0 | loss 3.5062 | lr 2.63e-03 | grad 0.1544 +2026-04-09 22:58:45 - INFO - parrotllm.training - step 22990 | epoch 0 | loss 3.5202 | lr 2.63e-03 | grad 0.1560 +2026-04-09 22:58:48 - INFO - parrotllm.training - step 23000 | epoch 0 | loss 3.5791 | lr 2.63e-03 | grad 0.1656 +2026-04-09 22:58:48 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 22:58:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:58:51 - INFO - parrotllm.training - Train: loss=3.5791, ppl=35.84 +2026-04-09 22:58:51 - INFO - parrotllm.training - Val: loss=3.4519, ppl=31.56 +2026-04-09 22:58:52 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4519_epoch_0000_step_0023000.pt +2026-04-09 22:58:53 - INFO - parrotllm.training - No validation improvement for 2/5 evaluation(s) (best=3.4510, min_delta=0.001000). +2026-04-09 22:58:53 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 22:58:56 - INFO - parrotllm.training - step 23010 | epoch 0 | loss 3.5813 | lr 2.63e-03 | grad 0.1720 +2026-04-09 22:58:59 - INFO - parrotllm.training - step 23020 | epoch 0 | loss 3.5169 | lr 2.63e-03 | grad 0.1896 +2026-04-09 22:59:02 - INFO - parrotllm.training - step 23030 | epoch 0 | loss 3.4324 | lr 2.63e-03 | grad 0.1729 +2026-04-09 22:59:05 - INFO - parrotllm.training - step 23040 | epoch 0 | loss 3.4862 | lr 2.63e-03 | grad 0.1851 +2026-04-09 22:59:08 - INFO - parrotllm.training - step 23050 | epoch 0 | loss 3.6026 | lr 2.63e-03 | grad 0.1956 +2026-04-09 22:59:11 - INFO - parrotllm.training - step 23060 | epoch 0 | loss 3.4439 | lr 2.63e-03 | grad 0.1876 +2026-04-09 22:59:14 - INFO - parrotllm.training - step 23070 | epoch 0 | loss 3.4572 | lr 2.63e-03 | grad 0.2321 +2026-04-09 22:59:17 - INFO - parrotllm.training - step 23080 | epoch 0 | loss 3.4194 | lr 2.63e-03 | grad 0.1616 +2026-04-09 22:59:20 - INFO - parrotllm.training - step 23090 | epoch 0 | loss 3.4544 | lr 2.63e-03 | grad 0.1557 +2026-04-09 22:59:23 - INFO - parrotllm.training - step 23100 | epoch 0 | loss 3.4054 | lr 2.63e-03 | grad 0.1655 +2026-04-09 22:59:26 - INFO - parrotllm.training - step 23110 | epoch 0 | loss 3.5729 | lr 2.63e-03 | grad 0.1857 +2026-04-09 22:59:29 - INFO - parrotllm.training - step 23120 | epoch 0 | loss 3.4683 | lr 2.63e-03 | grad 0.1820 +2026-04-09 22:59:32 - INFO - parrotllm.training - step 23130 | epoch 0 | loss 3.5993 | lr 2.63e-03 | grad 0.2277 +2026-04-09 22:59:36 - INFO - parrotllm.training - step 23140 | epoch 0 | loss 3.4919 | lr 2.63e-03 | grad 0.2041 +2026-04-09 22:59:39 - INFO - parrotllm.training - step 23150 | epoch 0 | loss 3.4994 | lr 2.63e-03 | grad 0.1654 +2026-04-09 22:59:42 - INFO - parrotllm.training - step 23160 | epoch 0 | loss 3.5193 | lr 2.63e-03 | grad 0.1975 +2026-04-09 22:59:45 - INFO - parrotllm.training - step 23170 | epoch 0 | loss 3.4775 | lr 2.63e-03 | grad 0.1788 +2026-04-09 22:59:48 - INFO - parrotllm.training - step 23180 | epoch 0 | loss 3.5030 | lr 2.63e-03 | grad 0.1922 +2026-04-09 22:59:51 - INFO - parrotllm.training - step 23190 | epoch 0 | loss 3.5242 | lr 2.63e-03 | grad 0.1879 +2026-04-09 22:59:54 - INFO - parrotllm.training - step 23200 | epoch 0 | loss 3.4815 | lr 2.63e-03 | grad 0.1868 +2026-04-09 22:59:57 - INFO - parrotllm.training - step 23210 | epoch 0 | loss 3.5088 | lr 2.63e-03 | grad 0.1614 +2026-04-09 23:00:00 - INFO - parrotllm.training - step 23220 | epoch 0 | loss 3.5104 | lr 2.63e-03 | grad 0.1482 +2026-04-09 23:00:03 - INFO - parrotllm.training - step 23230 | epoch 0 | loss 3.5327 | lr 2.63e-03 | grad 0.1743 +2026-04-09 23:00:06 - INFO - parrotllm.training - step 23240 | epoch 0 | loss 3.4354 | lr 2.63e-03 | grad 0.1624 +2026-04-09 23:00:09 - INFO - parrotllm.training - step 23250 | epoch 0 | loss 3.4941 | lr 2.63e-03 | grad 0.1440 +2026-04-09 23:00:12 - INFO - parrotllm.training - step 23260 | epoch 0 | loss 3.5246 | lr 2.63e-03 | grad 0.1622 +2026-04-09 23:00:15 - INFO - parrotllm.training - step 23270 | epoch 0 | loss 3.4481 | lr 2.63e-03 | grad 0.1614 +2026-04-09 23:00:18 - INFO - parrotllm.training - step 23280 | epoch 0 | loss 3.5577 | lr 2.63e-03 | grad 0.1657 +2026-04-09 23:00:21 - INFO - parrotllm.training - step 23290 | epoch 0 | loss 3.4985 | lr 2.63e-03 | grad 0.1736 +2026-04-09 23:00:24 - INFO - parrotllm.training - step 23300 | epoch 0 | loss 3.5591 | lr 2.63e-03 | grad 0.2030 +2026-04-09 23:00:27 - INFO - parrotllm.training - step 23310 | epoch 0 | loss 3.4532 | lr 2.63e-03 | grad 0.1497 +2026-04-09 23:00:30 - INFO - parrotllm.training - step 23320 | epoch 0 | loss 3.4670 | lr 2.63e-03 | grad 0.2068 +2026-04-09 23:00:33 - INFO - parrotllm.training - step 23330 | epoch 0 | loss 3.4563 | lr 2.63e-03 | grad 0.1796 +2026-04-09 23:00:36 - INFO - parrotllm.training - step 23340 | epoch 0 | loss 3.4759 | lr 2.63e-03 | grad 0.1614 +2026-04-09 23:00:40 - INFO - parrotllm.training - step 23350 | epoch 0 | loss 3.5689 | lr 2.63e-03 | grad 0.1690 +2026-04-09 23:00:43 - INFO - parrotllm.training - step 23360 | epoch 0 | loss 3.5793 | lr 2.63e-03 | grad 0.2174 +2026-04-09 23:00:46 - INFO - parrotllm.training - step 23370 | epoch 0 | loss 3.4865 | lr 2.63e-03 | grad 0.1738 +2026-04-09 23:00:49 - INFO - parrotllm.training - step 23380 | epoch 0 | loss 3.4199 | lr 2.63e-03 | grad 0.1655 +2026-04-09 23:00:52 - INFO - parrotllm.training - step 23390 | epoch 0 | loss 3.5397 | lr 2.63e-03 | grad 0.1695 +2026-04-09 23:00:55 - INFO - parrotllm.training - step 23400 | epoch 0 | loss 3.4128 | lr 2.63e-03 | grad 0.1711 +2026-04-09 23:00:58 - INFO - parrotllm.training - step 23410 | epoch 0 | loss 3.4367 | lr 2.63e-03 | grad 0.1753 +2026-04-09 23:01:01 - INFO - parrotllm.training - step 23420 | epoch 0 | loss 3.4161 | lr 2.63e-03 | grad 0.1711 +2026-04-09 23:01:04 - INFO - parrotllm.training - step 23430 | epoch 0 | loss 3.5712 | lr 2.63e-03 | grad 0.1548 +2026-04-09 23:01:07 - INFO - parrotllm.training - step 23440 | epoch 0 | loss 3.4829 | lr 2.63e-03 | grad 0.1735 +2026-04-09 23:01:10 - INFO - parrotllm.training - step 23450 | epoch 0 | loss 3.4780 | lr 2.63e-03 | grad 0.1646 +2026-04-09 23:01:13 - INFO - parrotllm.training - step 23460 | epoch 0 | loss 3.5027 | lr 2.63e-03 | grad 0.2000 +2026-04-09 23:01:16 - INFO - parrotllm.training - step 23470 | epoch 0 | loss 3.5326 | lr 2.63e-03 | grad 0.1840 +2026-04-09 23:01:19 - INFO - parrotllm.training - step 23480 | epoch 0 | loss 3.4682 | lr 2.63e-03 | grad 0.1555 +2026-04-09 23:01:22 - INFO - parrotllm.training - step 23490 | epoch 0 | loss 3.4063 | lr 2.63e-03 | grad 0.1682 +2026-04-09 23:01:25 - INFO - parrotllm.training - step 23500 | epoch 0 | loss 3.5057 | lr 2.63e-03 | grad 0.1617 +2026-04-09 23:01:25 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 23:01:25 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 23:01:28 - INFO - parrotllm.training - Train: loss=3.5057, ppl=33.30 +2026-04-09 23:01:28 - INFO - parrotllm.training - Val: loss=3.4510, ppl=31.53 +2026-04-09 23:01:29 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4510_epoch_0000_step_0023500.pt +2026-04-09 23:01:30 - INFO - parrotllm.training - No validation improvement for 3/5 evaluation(s) (best=3.4510, min_delta=0.001000). +2026-04-09 23:01:30 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 23:01:33 - INFO - parrotllm.training - step 23510 | epoch 0 | loss 3.5012 | lr 2.63e-03 | grad 0.1721 +2026-04-09 23:01:36 - INFO - parrotllm.training - step 23520 | epoch 0 | loss 3.5722 | lr 2.63e-03 | grad 0.1724 +2026-04-09 23:01:40 - INFO - parrotllm.training - step 23530 | epoch 0 | loss 3.4923 | lr 2.63e-03 | grad 0.1589 +2026-04-09 23:01:43 - INFO - parrotllm.training - step 23540 | epoch 0 | loss 3.5884 | lr 2.63e-03 | grad 0.1777 +2026-04-09 23:01:46 - INFO - parrotllm.training - step 23550 | epoch 0 | loss 3.4045 | lr 2.63e-03 | grad 0.2208 +2026-04-09 23:01:49 - INFO - parrotllm.training - step 23560 | epoch 0 | loss 3.4533 | lr 2.63e-03 | grad 0.1718 +2026-04-09 23:01:52 - INFO - parrotllm.training - step 23570 | epoch 0 | loss 3.5263 | lr 2.63e-03 | grad 0.1745 +2026-04-09 23:01:55 - INFO - parrotllm.training - step 23580 | epoch 0 | loss 3.5322 | lr 2.63e-03 | grad 0.1513 +2026-04-09 23:01:58 - INFO - parrotllm.training - step 23590 | epoch 0 | loss 3.4483 | lr 2.63e-03 | grad 0.1702 +2026-04-09 23:02:01 - INFO - parrotllm.training - step 23600 | epoch 0 | loss 3.4934 | lr 2.63e-03 | grad 0.1521 +2026-04-09 23:02:04 - INFO - parrotllm.training - step 23610 | epoch 0 | loss 3.4735 | lr 2.63e-03 | grad 0.1692 +2026-04-09 23:02:07 - INFO - parrotllm.training - step 23620 | epoch 0 | loss 3.5405 | lr 2.63e-03 | grad 0.1759 +2026-04-09 23:02:10 - INFO - parrotllm.training - step 23630 | epoch 0 | loss 3.3903 | lr 2.63e-03 | grad 0.1626 +2026-04-09 23:02:13 - INFO - parrotllm.training - step 23640 | epoch 0 | loss 3.4578 | lr 2.63e-03 | grad 0.1569 +2026-04-09 23:02:16 - INFO - parrotllm.training - step 23650 | epoch 0 | loss 3.4369 | lr 2.63e-03 | grad 0.1556 +2026-04-09 23:02:19 - INFO - parrotllm.training - step 23660 | epoch 0 | loss 3.4303 | lr 2.63e-03 | grad 0.1658 +2026-04-09 23:02:22 - INFO - parrotllm.training - step 23670 | epoch 0 | loss 3.4662 | lr 2.63e-03 | grad 0.1925 +2026-04-09 23:02:25 - INFO - parrotllm.training - step 23680 | epoch 0 | loss 3.4430 | lr 2.63e-03 | grad 0.1643 +2026-04-09 23:02:28 - INFO - parrotllm.training - step 23690 | epoch 0 | loss 3.5928 | lr 2.63e-03 | grad 0.1535 +2026-04-09 23:02:31 - INFO - parrotllm.training - step 23700 | epoch 0 | loss 3.5656 | lr 2.63e-03 | grad 0.1816 +2026-04-09 23:02:34 - INFO - parrotllm.training - step 23710 | epoch 0 | loss 3.5110 | lr 2.63e-03 | grad 0.1728 +2026-04-09 23:02:37 - INFO - parrotllm.training - step 23720 | epoch 0 | loss 3.4355 | lr 2.63e-03 | grad 0.1512 +2026-04-09 23:02:40 - INFO - parrotllm.training - step 23730 | epoch 0 | loss 3.4132 | lr 2.63e-03 | grad 0.1742 +2026-04-09 23:02:44 - INFO - parrotllm.training - step 23740 | epoch 0 | loss 3.5256 | lr 2.63e-03 | grad 0.2001 +2026-04-09 23:02:47 - INFO - parrotllm.training - step 23750 | epoch 0 | loss 3.5541 | lr 2.63e-03 | grad 0.1774 +2026-04-09 23:02:50 - INFO - parrotllm.training - step 23760 | epoch 0 | loss 3.5188 | lr 2.63e-03 | grad 0.1635 +2026-04-09 23:02:53 - INFO - parrotllm.training - step 23770 | epoch 0 | loss 3.4570 | lr 2.63e-03 | grad 0.1686 +2026-04-09 23:02:56 - INFO - parrotllm.training - step 23780 | epoch 0 | loss 3.5527 | lr 2.63e-03 | grad 0.1777 +2026-04-09 23:02:59 - INFO - parrotllm.training - step 23790 | epoch 0 | loss 3.5017 | lr 2.63e-03 | grad 0.1469 +2026-04-09 23:03:02 - INFO - parrotllm.training - step 23800 | epoch 0 | loss 3.6312 | lr 2.63e-03 | grad 0.1859 +2026-04-09 23:03:05 - INFO - parrotllm.training - step 23810 | epoch 0 | loss 3.5014 | lr 2.63e-03 | grad 0.1841 +2026-04-09 23:03:08 - INFO - parrotllm.training - step 23820 | epoch 0 | loss 3.5077 | lr 2.63e-03 | grad 0.1510 +2026-04-09 23:03:11 - INFO - parrotllm.training - step 23830 | epoch 0 | loss 3.5807 | lr 2.63e-03 | grad 0.1998 +2026-04-09 23:03:14 - INFO - parrotllm.training - step 23840 | epoch 0 | loss 3.5045 | lr 2.63e-03 | grad 0.1732 +2026-04-09 23:03:17 - INFO - parrotllm.training - step 23850 | epoch 0 | loss 3.5230 | lr 2.63e-03 | grad 0.1613 +2026-04-09 23:03:20 - INFO - parrotllm.training - step 23860 | epoch 0 | loss 3.4769 | lr 2.63e-03 | grad 0.2013 +2026-04-09 23:03:23 - INFO - parrotllm.training - step 23870 | epoch 0 | loss 3.4859 | lr 2.63e-03 | grad 0.1526 +2026-04-09 23:03:26 - INFO - parrotllm.training - step 23880 | epoch 0 | loss 3.5065 | lr 2.63e-03 | grad 0.1701 +2026-04-09 23:03:29 - INFO - parrotllm.training - step 23890 | epoch 0 | loss 3.5005 | lr 2.63e-03 | grad 0.1734 +2026-04-09 23:03:32 - INFO - parrotllm.training - step 23900 | epoch 0 | loss 3.4615 | lr 2.63e-03 | grad 0.2058 +2026-04-09 23:03:35 - INFO - parrotllm.training - step 23910 | epoch 0 | loss 3.4734 | lr 2.63e-03 | grad 0.1862 +2026-04-09 23:03:38 - INFO - parrotllm.training - step 23920 | epoch 0 | loss 3.4492 | lr 2.63e-03 | grad 0.1750 +2026-04-09 23:03:41 - INFO - parrotllm.training - step 23930 | epoch 0 | loss 3.5396 | lr 2.63e-03 | grad 0.1697 +2026-04-09 23:03:45 - INFO - parrotllm.training - step 23940 | epoch 0 | loss 3.5166 | lr 2.63e-03 | grad 0.1776 +2026-04-09 23:03:48 - INFO - parrotllm.training - step 23950 | epoch 0 | loss 3.4423 | lr 2.63e-03 | grad 0.1625 +2026-04-09 23:03:51 - INFO - parrotllm.training - step 23960 | epoch 0 | loss 3.6085 | lr 2.63e-03 | grad 0.1809 +2026-04-09 23:03:52 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 23972/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-09 23:03:54 - INFO - parrotllm.training - step 23970 | epoch 0 | loss 3.5035 | lr 2.63e-03 | grad 0.1666 +2026-04-09 23:03:57 - INFO - parrotllm.training - step 23980 | epoch 0 | loss 3.4529 | lr 2.63e-03 | grad 0.1839 +2026-04-09 23:04:00 - INFO - parrotllm.training - step 23990 | epoch 0 | loss 3.4943 | lr 2.63e-03 | grad 0.1763 +2026-04-09 23:04:03 - INFO - parrotllm.training - step 24000 | epoch 0 | loss 3.5342 | lr 2.63e-03 | grad 0.1892 +2026-04-09 23:04:03 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 23:04:03 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 23:04:06 - INFO - parrotllm.training - Train: loss=3.5342, ppl=34.27 +2026-04-09 23:04:06 - INFO - parrotllm.training - Val: loss=3.4516, ppl=31.55 +2026-04-09 23:04:07 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4516_epoch_0000_step_0024000.pt +2026-04-09 23:04:08 - INFO - parrotllm.training - No validation improvement for 4/5 evaluation(s) (best=3.4510, min_delta=0.001000). +2026-04-09 23:04:08 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 23:04:11 - INFO - parrotllm.training - step 24010 | epoch 0 | loss 3.3657 | lr 2.63e-03 | grad 0.1597 +2026-04-09 23:04:14 - INFO - parrotllm.training - step 24020 | epoch 0 | loss 3.5579 | lr 2.63e-03 | grad 0.1795 +2026-04-09 23:04:17 - INFO - parrotllm.training - step 24030 | epoch 0 | loss 3.4615 | lr 2.63e-03 | grad 0.1806 +2026-04-09 23:04:21 - INFO - parrotllm.training - step 24040 | epoch 0 | loss 3.5299 | lr 2.63e-03 | grad 0.1689 +2026-04-09 23:04:24 - INFO - parrotllm.training - step 24050 | epoch 0 | loss 3.5911 | lr 2.63e-03 | grad 0.1552 +2026-04-09 23:04:27 - INFO - parrotllm.training - step 24060 | epoch 0 | loss 3.4971 | lr 2.63e-03 | grad 0.1912 +2026-04-09 23:04:30 - INFO - parrotllm.training - step 24070 | epoch 0 | loss 3.4895 | lr 2.63e-03 | grad 0.2404 +2026-04-09 23:04:33 - INFO - parrotllm.training - step 24080 | epoch 0 | loss 3.4687 | lr 2.63e-03 | grad 0.2162 +2026-04-09 23:04:36 - INFO - parrotllm.training - step 24090 | epoch 0 | loss 3.5964 | lr 2.63e-03 | grad 0.1532 +2026-04-09 23:04:39 - INFO - parrotllm.training - step 24100 | epoch 0 | loss 3.6041 | lr 2.63e-03 | grad 0.1936 +2026-04-09 23:04:42 - INFO - parrotllm.training - step 24110 | epoch 0 | loss 3.3971 | lr 2.63e-03 | grad 0.1818 +2026-04-09 23:04:45 - INFO - parrotllm.training - step 24120 | epoch 0 | loss 3.5442 | lr 2.63e-03 | grad 0.1718 +2026-04-09 23:04:48 - INFO - parrotllm.training - step 24130 | epoch 0 | loss 3.4704 | lr 2.63e-03 | grad 0.1669 +2026-04-09 23:04:51 - INFO - parrotllm.training - step 24140 | epoch 0 | loss 3.4739 | lr 2.63e-03 | grad 0.1810 +2026-04-09 23:04:54 - INFO - parrotllm.training - step 24150 | epoch 0 | loss 3.5821 | lr 2.63e-03 | grad 0.1895 +2026-04-09 23:04:57 - INFO - parrotllm.training - step 24160 | epoch 0 | loss 3.4715 | lr 2.63e-03 | grad 0.1748 +2026-04-09 23:05:00 - INFO - parrotllm.training - step 24170 | epoch 0 | loss 3.5610 | lr 2.63e-03 | grad 0.1490 +2026-04-09 23:05:03 - INFO - parrotllm.training - step 24180 | epoch 0 | loss 3.5261 | lr 2.63e-03 | grad 0.1892 +2026-04-09 23:05:06 - INFO - parrotllm.training - step 24190 | epoch 0 | loss 3.5711 | lr 2.63e-03 | grad 0.1687 +2026-04-09 23:05:09 - INFO - parrotllm.training - step 24200 | epoch 0 | loss 3.4033 | lr 2.63e-03 | grad 0.1579 +2026-04-09 23:05:12 - INFO - parrotllm.training - step 24210 | epoch 0 | loss 3.5237 | lr 2.63e-03 | grad 0.1933 +2026-04-09 23:05:15 - INFO - parrotllm.training - step 24220 | epoch 0 | loss 3.5269 | lr 2.63e-03 | grad 0.2195 +2026-04-09 23:05:18 - INFO - parrotllm.training - step 24230 | epoch 0 | loss 3.5662 | lr 2.63e-03 | grad 0.1605 +2026-04-09 23:05:22 - INFO - parrotllm.training - step 24240 | epoch 0 | loss 3.5347 | lr 2.63e-03 | grad 0.1662 +2026-04-09 23:05:25 - INFO - parrotllm.training - step 24250 | epoch 0 | loss 3.5690 | lr 2.63e-03 | grad 0.1689 +2026-04-09 23:05:28 - INFO - parrotllm.training - step 24260 | epoch 0 | loss 3.4386 | lr 2.63e-03 | grad 0.1814 +2026-04-09 23:05:31 - INFO - parrotllm.training - step 24270 | epoch 0 | loss 3.4701 | lr 2.63e-03 | grad 0.1894 +2026-04-09 23:05:34 - INFO - parrotllm.training - step 24280 | epoch 0 | loss 3.5976 | lr 2.63e-03 | grad 0.1673 +2026-04-09 23:05:37 - INFO - parrotllm.training - step 24290 | epoch 0 | loss 3.4569 | lr 2.63e-03 | grad 0.1651 +2026-04-09 23:05:40 - INFO - parrotllm.training - step 24300 | epoch 0 | loss 3.5263 | lr 2.63e-03 | grad 0.1906 +2026-04-09 23:05:43 - INFO - parrotllm.training - step 24310 | epoch 0 | loss 3.4181 | lr 2.63e-03 | grad 0.1764 +2026-04-09 23:05:46 - INFO - parrotllm.training - step 24320 | epoch 0 | loss 3.5145 | lr 2.63e-03 | grad 0.1717 +2026-04-09 23:05:49 - INFO - parrotllm.training - step 24330 | epoch 0 | loss 3.5353 | lr 2.63e-03 | grad 0.1529 +2026-04-09 23:05:52 - INFO - parrotllm.training - step 24340 | epoch 0 | loss 3.5206 | lr 2.63e-03 | grad 0.1838 +2026-04-09 23:05:55 - INFO - parrotllm.training - step 24350 | epoch 0 | loss 3.5150 | lr 2.63e-03 | grad 0.1645 +2026-04-09 23:05:58 - INFO - parrotllm.training - step 24360 | epoch 0 | loss 3.5364 | lr 2.63e-03 | grad 0.1521 +2026-04-09 23:06:01 - INFO - parrotllm.training - step 24370 | epoch 0 | loss 3.6168 | lr 2.63e-03 | grad 0.1942 +2026-04-09 23:06:04 - INFO - parrotllm.training - step 24380 | epoch 0 | loss 3.4136 | lr 2.63e-03 | grad 0.1975 +2026-04-09 23:06:07 - INFO - parrotllm.training - step 24390 | epoch 0 | loss 3.5797 | lr 2.63e-03 | grad 0.1790 +2026-04-09 23:06:10 - INFO - parrotllm.training - step 24400 | epoch 0 | loss 3.4918 | lr 2.63e-03 | grad 0.1491 +2026-04-09 23:06:13 - INFO - parrotllm.training - step 24410 | epoch 0 | loss 3.4631 | lr 2.63e-03 | grad 0.2299 +2026-04-09 23:06:17 - INFO - parrotllm.training - step 24420 | epoch 0 | loss 3.5904 | lr 2.63e-03 | grad 0.1819 +2026-04-09 23:06:20 - INFO - parrotllm.training - step 24430 | epoch 0 | loss 3.4825 | lr 2.63e-03 | grad 0.1723 +2026-04-09 23:06:23 - INFO - parrotllm.training - step 24440 | epoch 0 | loss 3.5137 | lr 2.63e-03 | grad 0.1934 +2026-04-09 23:06:26 - INFO - parrotllm.training - step 24450 | epoch 0 | loss 3.5138 | lr 2.63e-03 | grad 0.1748 +2026-04-09 23:06:29 - INFO - parrotllm.training - step 24460 | epoch 0 | loss 3.5477 | lr 2.63e-03 | grad 0.1773 +2026-04-09 23:06:32 - INFO - parrotllm.training - step 24470 | epoch 0 | loss 3.5478 | lr 2.63e-03 | grad 0.1750 +2026-04-09 23:06:35 - INFO - parrotllm.training - step 24480 | epoch 0 | loss 3.4320 | lr 2.63e-03 | grad 0.1679 +2026-04-09 23:06:38 - INFO - parrotllm.training - step 24490 | epoch 0 | loss 3.5099 | lr 2.63e-03 | grad 0.1597 +2026-04-09 23:06:41 - INFO - parrotllm.training - step 24500 | epoch 0 | loss 3.4888 | lr 2.63e-03 | grad 0.1679 +2026-04-09 23:06:41 - INFO - parrotllm.training - Starting evaluation... +2026-04-09 23:06:41 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 23:06:44 - INFO - parrotllm.training - Train: loss=3.4888, ppl=32.75 +2026-04-09 23:06:44 - INFO - parrotllm.training - Val: loss=3.4520, ppl=31.56 +2026-04-09 23:06:45 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4520_epoch_0000_step_0024500.pt +2026-04-09 23:06:46 - INFO - parrotllm.training - No validation improvement for 5/5 evaluation(s) (best=3.4510, min_delta=0.001000). +2026-04-09 23:06:46 - INFO - parrotllm.training - Early stopping triggered. +2026-04-09 23:06:46 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-09 23:06:47 - INFO - parrotllm.training - Saved final checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0024500.pt +2026-04-09 23:06:47 - INFO - parrotllm.training - +Loss Curve: + Loss + ^ + 10.91 |* + |* + |* + |* + |* + |* + |** + 3.25 | ************************************************* + +--------------------------------------------------> step + 1 24500 +2026-04-09 23:06:47 - INFO - parrotllm.training - Stopped early after 24500 step(s): validation loss stopped improving +2026-04-09 23:06:47 - INFO - parrotllm.training - Best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4510_epoch_0000_step_0022000.pt +2026-04-09 23:06:47 - INFO - parrotllm.training - +============================================================ +TRAINING COMPLETE +============================================================ + Epochs: 1 + Total steps: 24500 + Total time: 2.16 hours + Best validation loss: 3.4510 + Run directory: runs/big_run/exp_c_8b/run_20260409_205717 +============================================================ +2026-04-09 23:06:47 - INFO - parrotllm.training - Uploading completed run directory to Hugging Face: repo=ParrotLabs/Preprocessed