diff --git "a/runs/big_run/exp_c_8b/run_20260410_044337/train.log" "b/runs/big_run/exp_c_8b/run_20260410_044337/train.log" new file mode 100644--- /dev/null +++ "b/runs/big_run/exp_c_8b/run_20260410_044337/train.log" @@ -0,0 +1,11183 @@ +2026-04-10 04:43:38 - INFO - parrotllm - Logging initialised -> runs/big_run/exp_c_8b/run_20260410_044337/train.log +2026-04-10 04:43:38 - INFO - parrotllm.training - device=cuda:0 | rank=0 | world_size=8 | distributed=yes +2026-04-10 04:43:38 - INFO - parrotllm.training - checkpoint directory=/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints | keep_last=10 | keep_best=10 +2026-04-10 04:43:38 - INFO - parrotllm.training - train windows=random-overlapping | tokens=8083182257 | valid_starts=8083181233 | sampled_windows_per_epoch=8083181233 +2026-04-10 04:43:38 - INFO - parrotllm.training - val windows=sequential | tokens=81648305 | stride=512 | windows=159468 +2026-04-10 04:43:39 - INFO - parrotllm.training - compiling model with torch.compile... +2026-04-10 04:43:40 - INFO - parrotllm.training - +============================================================ +MODEL ARCHITECTURE SUMMARY +============================================================ + +Configuration: + Vocab size: 50258 + Block size (context): 1024 + Layers: 14 + Heads: 6 + Embedding dim: 384 + FFN hidden dim: 768 + Dropout: 0.03 + Bias: False + +Parameters (unique, weight-tied layers counted once): + Total: 39,966,592 (39.97M) + Trainable: 39,966,592 (39.97M) + Non-trainable: 0 + Non-embedding: 39,966,592 (39.97M) + Position embeddings: 0 (RoPE: 0 learned params) + Size (MB): 159.87 + +Layer-wise breakdown (torchinfo): + Note: torchinfo double-counts weight-tied layers (tok_emb/lm_head). +============================================================================================================================================ +Layer (type:depth-idx) Input Shape Output Shape Param # Trainable +============================================================================================================================================ +ParrotLLM [32, 1024] [32, 1024, 50258] -- True +├─Embedding: 1-1 [32, 1024] [32, 1024, 384] 19,299,072 True +├─Dropout: 1-2 [32, 1024, 384] [32, 1024, 384] -- -- +├─ModuleList: 1-3 -- -- -- True +│ └─TransformerBlock: 2-1 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-1 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-2 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-3 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-4 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-5 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-6 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-2 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-7 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-8 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-9 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-10 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-11 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-12 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-3 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-13 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-14 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-15 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-16 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-17 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-18 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-4 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-19 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-20 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-21 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-22 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-23 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-24 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-5 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-25 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-26 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-27 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-28 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-29 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-30 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-6 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-31 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-32 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-33 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-34 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-35 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-36 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-7 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-37 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-38 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-39 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-40 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-41 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-42 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-8 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-43 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-44 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-45 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-46 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-47 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-48 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-9 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-49 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-50 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-51 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-52 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-53 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-54 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-10 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-55 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-56 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-57 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-58 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-59 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-60 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-11 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-61 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-62 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-63 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-64 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-65 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-66 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-12 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-67 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-68 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-69 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-70 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-71 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-72 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-13 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-73 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-74 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-75 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-76 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-77 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-78 [32, 1024, 384] [32, 1024, 384] 384 True +│ └─TransformerBlock: 2-14 [32, 1024, 384] [32, 1024, 384] -- True +│ │ └─RMSNorm: 3-79 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─MultiHeadAttention: 3-80 [32, 1024, 384] [32, 1024, 384] 589,952 True +│ │ └─RMSNorm: 3-81 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─RMSNorm: 3-82 [32, 1024, 384] [32, 1024, 384] 384 True +│ │ └─SwiGLUMLP: 3-83 [32, 1024, 384] [32, 1024, 384] 884,736 True +│ │ └─RMSNorm: 3-84 [32, 1024, 384] [32, 1024, 384] 384 True +├─RMSNorm: 1-4 [32, 1024, 384] [32, 1024, 384] 384 True +├─Linear: 1-5 [32, 1024, 384] [32, 1024, 50258] 19,299,072 True +============================================================================================================================================ +Total params: 59,265,664 +Trainable params: 59,265,664 +Non-trainable params: 0 +Total mult-adds (Units.GIGABYTES): 1.90 +============================================================================================================================================ +Input size (MB): 0.26 +Forward/backward pass size (MB): 34515.45 +Params size (MB): 237.06 +Estimated Total Size (MB): 34752.78 +============================================================================================================================================ +============================================================ +2026-04-10 04:43:40 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 04:43:43 - INFO - parrotllm.training - Initial val: loss=10.8942, ppl=53862.54 +2026-04-10 04:43:43 - INFO - parrotllm.training - +============================================================ +Starting training... + Steps per epoch (approx): 31574926 + Max steps: 122070 +============================================================ +2026-04-10 04:43:43 - INFO - parrotllm.training - LR schedule=cosine | decay_ratio=0.0 | z_loss_coeff=0.00013887413811122497 +2026-04-10 04:43:43 - INFO - parrotllm.training - Early stopping enabled | patience=15 evals | min_delta=0.001000 +2026-04-10 04:44:24 - INFO - parrotllm.training - step 1 | epoch 0 | loss 10.9060 | lr 1.00e-06 | grad 6.9485 +2026-04-10 04:44:27 - INFO - parrotllm.training - step 10 | epoch 0 | loss 10.5752 | lr 1.00e-05 | grad 5.4858 +2026-04-10 04:44:30 - INFO - parrotllm.training - step 20 | epoch 0 | loss 10.2148 | lr 2.00e-05 | grad 1.7432 +2026-04-10 04:44:33 - INFO - parrotllm.training - step 30 | epoch 0 | loss 10.0884 | lr 3.00e-05 | grad 1.6584 +2026-04-10 04:44:36 - INFO - parrotllm.training - step 40 | epoch 0 | loss 9.8609 | lr 4.00e-05 | grad 1.7262 +2026-04-10 04:44:39 - INFO - parrotllm.training - step 50 | epoch 0 | loss 9.5716 | lr 5.00e-05 | grad 1.6190 +2026-04-10 04:44:42 - INFO - parrotllm.training - step 60 | epoch 0 | loss 9.3381 | lr 6.00e-05 | grad 1.5620 +2026-04-10 04:44:45 - INFO - parrotllm.training - step 70 | epoch 0 | loss 8.9515 | lr 7.00e-05 | grad 1.5648 +2026-04-10 04:44:48 - INFO - parrotllm.training - step 80 | epoch 0 | loss 8.7403 | lr 8.00e-05 | grad 1.5118 +2026-04-10 04:44:51 - INFO - parrotllm.training - step 90 | epoch 0 | loss 8.3734 | lr 9.00e-05 | grad 1.4650 +2026-04-10 04:44:54 - INFO - parrotllm.training - step 100 | epoch 0 | loss 8.0735 | lr 1.00e-04 | grad 1.4211 +2026-04-10 04:44:57 - INFO - parrotllm.training - step 110 | epoch 0 | loss 7.8374 | lr 1.10e-04 | grad 1.2840 +2026-04-10 04:45:00 - INFO - parrotllm.training - step 120 | epoch 0 | loss 7.5992 | lr 1.20e-04 | grad 1.4933 +2026-04-10 04:45:03 - INFO - parrotllm.training - step 130 | epoch 0 | loss 7.3564 | lr 1.30e-04 | grad 1.0862 +2026-04-10 04:45:06 - INFO - parrotllm.training - step 140 | epoch 0 | loss 7.1657 | lr 1.40e-04 | grad 1.1230 +2026-04-10 04:45:09 - INFO - parrotllm.training - step 150 | epoch 0 | loss 7.0142 | lr 1.50e-04 | grad 0.6880 +2026-04-10 04:45:12 - INFO - parrotllm.training - step 160 | epoch 0 | loss 6.8135 | lr 1.60e-04 | grad 1.1055 +2026-04-10 04:45:16 - INFO - parrotllm.training - step 170 | epoch 0 | loss 6.7167 | lr 1.70e-04 | grad 1.1968 +2026-04-10 04:45:19 - INFO - parrotllm.training - step 180 | epoch 0 | loss 6.6445 | lr 1.80e-04 | grad 1.0525 +2026-04-10 04:45:22 - INFO - parrotllm.training - step 190 | epoch 0 | loss 6.6612 | lr 1.90e-04 | grad 1.1572 +2026-04-10 04:45:25 - INFO - parrotllm.training - step 200 | epoch 0 | loss 6.4939 | lr 2.00e-04 | grad 0.9375 +2026-04-10 04:45:28 - INFO - parrotllm.training - step 210 | epoch 0 | loss 6.3287 | lr 2.10e-04 | grad 1.5911 +2026-04-10 04:45:31 - INFO - parrotllm.training - step 220 | epoch 0 | loss 6.2696 | lr 2.20e-04 | grad 1.0732 +2026-04-10 04:45:34 - INFO - parrotllm.training - step 230 | epoch 0 | loss 6.2800 | lr 2.30e-04 | grad 1.0999 +2026-04-10 04:45:37 - INFO - parrotllm.training - step 240 | epoch 0 | loss 6.1595 | lr 2.40e-04 | grad 1.0262 +2026-04-10 04:45:40 - INFO - parrotllm.training - step 250 | epoch 0 | loss 6.1347 | lr 2.50e-04 | grad 1.1109 +2026-04-10 04:45:43 - INFO - parrotllm.training - step 260 | epoch 0 | loss 5.9887 | lr 2.60e-04 | grad 1.2016 +2026-04-10 04:45:46 - INFO - parrotllm.training - step 270 | epoch 0 | loss 5.9565 | lr 2.70e-04 | grad 1.0974 +2026-04-10 04:45:49 - INFO - parrotllm.training - step 280 | epoch 0 | loss 6.0187 | lr 2.80e-04 | grad 1.2621 +2026-04-10 04:45:52 - INFO - parrotllm.training - step 290 | epoch 0 | loss 5.7943 | lr 2.90e-04 | grad 1.2380 +2026-04-10 04:45:56 - INFO - parrotllm.training - step 300 | epoch 0 | loss 5.8018 | lr 3.00e-04 | grad 1.4612 +2026-04-10 04:45:59 - INFO - parrotllm.training - step 310 | epoch 0 | loss 5.8612 | lr 3.10e-04 | grad 1.1734 +2026-04-10 04:46:02 - INFO - parrotllm.training - step 320 | epoch 0 | loss 5.8039 | lr 3.20e-04 | grad 1.1831 +2026-04-10 04:46:05 - INFO - parrotllm.training - step 330 | epoch 0 | loss 5.7854 | lr 3.30e-04 | grad 1.3627 +2026-04-10 04:46:08 - INFO - parrotllm.training - step 340 | epoch 0 | loss 5.6960 | lr 3.40e-04 | grad 1.0980 +2026-04-10 04:46:11 - INFO - parrotllm.training - step 350 | epoch 0 | loss 5.6594 | lr 3.50e-04 | grad 0.8521 +2026-04-10 04:46:14 - INFO - parrotllm.training - step 360 | epoch 0 | loss 5.5579 | lr 3.60e-04 | grad 0.9856 +2026-04-10 04:46:17 - INFO - parrotllm.training - step 370 | epoch 0 | loss 5.4931 | lr 3.70e-04 | grad 1.2643 +2026-04-10 04:46:20 - INFO - parrotllm.training - step 380 | epoch 0 | loss 5.4182 | lr 3.80e-04 | grad 1.3353 +2026-04-10 04:46:23 - INFO - parrotllm.training - step 390 | epoch 0 | loss 5.4034 | lr 3.90e-04 | grad 1.0236 +2026-04-10 04:46:26 - INFO - parrotllm.training - step 400 | epoch 0 | loss 5.3503 | lr 4.00e-04 | grad 0.8446 +2026-04-10 04:46:29 - INFO - parrotllm.training - step 410 | epoch 0 | loss 5.4306 | lr 4.10e-04 | grad 0.7889 +2026-04-10 04:46:32 - INFO - parrotllm.training - step 420 | epoch 0 | loss 5.3204 | lr 4.20e-04 | grad 0.9367 +2026-04-10 04:46:35 - INFO - parrotllm.training - step 430 | epoch 0 | loss 5.4133 | lr 4.30e-04 | grad 0.8504 +2026-04-10 04:46:38 - INFO - parrotllm.training - step 440 | epoch 0 | loss 5.3873 | lr 4.40e-04 | grad 1.0055 +2026-04-10 04:46:41 - INFO - parrotllm.training - step 450 | epoch 0 | loss 5.2047 | lr 4.50e-04 | grad 1.1108 +2026-04-10 04:46:44 - INFO - parrotllm.training - step 460 | epoch 0 | loss 5.2313 | lr 4.60e-04 | grad 0.8630 +2026-04-10 04:46:47 - INFO - parrotllm.training - step 470 | epoch 0 | loss 5.2379 | lr 4.70e-04 | grad 0.8398 +2026-04-10 04:46:51 - INFO - parrotllm.training - step 480 | epoch 0 | loss 5.2360 | lr 4.80e-04 | grad 0.8206 +2026-04-10 04:46:54 - INFO - parrotllm.training - step 490 | epoch 0 | loss 5.2540 | lr 4.90e-04 | grad 0.8039 +2026-04-10 04:46:57 - INFO - parrotllm.training - step 500 | epoch 0 | loss 5.1229 | lr 5.00e-04 | grad 0.8288 +2026-04-10 04:46:57 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 04:46:57 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 04:47:00 - INFO - parrotllm.training - Train: loss=5.1229, ppl=167.81 +2026-04-10 04:47:00 - INFO - parrotllm.training - Val: loss=5.0357, ppl=153.81 +2026-04-10 04:47:00 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 04:47:01 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_5p0357_epoch_0000_step_0000500.pt +2026-04-10 04:47:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 04:47:05 - INFO - parrotllm.training - step 510 | epoch 0 | loss 4.9943 | lr 5.10e-04 | grad 0.8654 +2026-04-10 04:47:08 - INFO - parrotllm.training - step 520 | epoch 0 | loss 5.0711 | lr 5.20e-04 | grad 1.0222 +2026-04-10 04:47:11 - INFO - parrotllm.training - step 530 | epoch 0 | loss 5.0886 | lr 5.30e-04 | grad 1.0545 +2026-04-10 04:47:14 - INFO - parrotllm.training - step 540 | epoch 0 | loss 5.1380 | lr 5.40e-04 | grad 0.7180 +2026-04-10 04:47:17 - INFO - parrotllm.training - step 550 | epoch 0 | loss 5.0016 | lr 5.50e-04 | grad 0.5987 +2026-04-10 04:47:20 - INFO - parrotllm.training - step 560 | epoch 0 | loss 4.8642 | lr 5.60e-04 | grad 0.8007 +2026-04-10 04:47:23 - INFO - parrotllm.training - step 570 | epoch 0 | loss 4.8686 | lr 5.70e-04 | grad 0.8010 +2026-04-10 04:47:26 - INFO - parrotllm.training - step 580 | epoch 0 | loss 4.9222 | lr 5.80e-04 | grad 0.9675 +2026-04-10 04:47:29 - INFO - parrotllm.training - step 590 | epoch 0 | loss 4.8375 | lr 5.90e-04 | grad 0.5031 +2026-04-10 04:47:32 - INFO - parrotllm.training - step 600 | epoch 0 | loss 4.9127 | lr 6.00e-04 | grad 0.8218 +2026-04-10 04:47:35 - INFO - parrotllm.training - step 610 | epoch 0 | loss 4.8995 | lr 6.10e-04 | grad 0.5995 +2026-04-10 04:47:38 - INFO - parrotllm.training - step 620 | epoch 0 | loss 4.7514 | lr 6.20e-04 | grad 0.6606 +2026-04-10 04:47:41 - INFO - parrotllm.training - step 630 | epoch 0 | loss 4.8201 | lr 6.30e-04 | grad 0.6743 +2026-04-10 04:47:45 - INFO - parrotllm.training - step 640 | epoch 0 | loss 4.6899 | lr 6.40e-04 | grad 0.4933 +2026-04-10 04:47:48 - INFO - parrotllm.training - step 650 | epoch 0 | loss 4.7207 | lr 6.50e-04 | grad 0.7326 +2026-04-10 04:47:51 - INFO - parrotllm.training - step 660 | epoch 0 | loss 4.7166 | lr 6.60e-04 | grad 0.7420 +2026-04-10 04:47:54 - INFO - parrotllm.training - step 670 | epoch 0 | loss 4.7742 | lr 6.70e-04 | grad 0.4936 +2026-04-10 04:47:57 - INFO - parrotllm.training - step 680 | epoch 0 | loss 4.7250 | lr 6.80e-04 | grad 0.6226 +2026-04-10 04:48:00 - INFO - parrotllm.training - step 690 | epoch 0 | loss 4.6218 | lr 6.90e-04 | grad 0.6557 +2026-04-10 04:48:03 - INFO - parrotllm.training - step 700 | epoch 0 | loss 4.5708 | lr 7.00e-04 | grad 0.5691 +2026-04-10 04:48:06 - INFO - parrotllm.training - step 710 | epoch 0 | loss 4.6273 | lr 7.10e-04 | grad 0.8229 +2026-04-10 04:48:09 - INFO - parrotllm.training - step 720 | epoch 0 | loss 4.5982 | lr 7.20e-04 | grad 0.6212 +2026-04-10 04:48:12 - INFO - parrotllm.training - step 730 | epoch 0 | loss 4.6164 | lr 7.30e-04 | grad 0.4906 +2026-04-10 04:48:15 - INFO - parrotllm.training - step 740 | epoch 0 | loss 4.5818 | lr 7.40e-04 | grad 0.6957 +2026-04-10 04:48:18 - INFO - parrotllm.training - step 750 | epoch 0 | loss 4.5017 | lr 7.50e-04 | grad 0.5000 +2026-04-10 04:48:21 - INFO - parrotllm.training - step 760 | epoch 0 | loss 4.5811 | lr 7.60e-04 | grad 0.6365 +2026-04-10 04:48:24 - INFO - parrotllm.training - step 770 | epoch 0 | loss 4.5011 | lr 7.70e-04 | grad 0.7582 +2026-04-10 04:48:27 - INFO - parrotllm.training - step 780 | epoch 0 | loss 4.6018 | lr 7.80e-04 | grad 0.4515 +2026-04-10 04:48:30 - INFO - parrotllm.training - step 790 | epoch 0 | loss 4.5975 | lr 7.90e-04 | grad 0.5041 +2026-04-10 04:48:33 - INFO - parrotllm.training - step 800 | epoch 0 | loss 4.5529 | lr 8.00e-04 | grad 0.5545 +2026-04-10 04:48:36 - INFO - parrotllm.training - step 810 | epoch 0 | loss 4.3512 | lr 8.10e-04 | grad 0.6117 +2026-04-10 04:48:40 - INFO - parrotllm.training - step 820 | epoch 0 | loss 4.3782 | lr 8.20e-04 | grad 0.4906 +2026-04-10 04:48:43 - INFO - parrotllm.training - step 830 | epoch 0 | loss 4.4750 | lr 8.30e-04 | grad 0.5168 +2026-04-10 04:48:46 - INFO - parrotllm.training - step 840 | epoch 0 | loss 4.4363 | lr 8.40e-04 | grad 0.4748 +2026-04-10 04:48:49 - INFO - parrotllm.training - step 850 | epoch 0 | loss 4.4589 | lr 8.50e-04 | grad 0.5506 +2026-04-10 04:48:52 - INFO - parrotllm.training - step 860 | epoch 0 | loss 4.5185 | lr 8.60e-04 | grad 0.4574 +2026-04-10 04:48:55 - INFO - parrotllm.training - step 870 | epoch 0 | loss 4.4637 | lr 8.70e-04 | grad 0.4439 +2026-04-10 04:48:58 - INFO - parrotllm.training - step 880 | epoch 0 | loss 4.3121 | lr 8.80e-04 | grad 0.5110 +2026-04-10 04:49:01 - INFO - parrotllm.training - step 890 | epoch 0 | loss 4.4120 | lr 8.90e-04 | grad 0.5022 +2026-04-10 04:49:04 - INFO - parrotllm.training - step 900 | epoch 0 | loss 4.4646 | lr 9.00e-04 | grad 0.4707 +2026-04-10 04:49:07 - INFO - parrotllm.training - step 910 | epoch 0 | loss 4.3489 | lr 9.10e-04 | grad 0.3983 +2026-04-10 04:49:10 - INFO - parrotllm.training - step 920 | epoch 0 | loss 4.3665 | lr 9.20e-04 | grad 0.5186 +2026-04-10 04:49:13 - INFO - parrotllm.training - step 930 | epoch 0 | loss 4.3304 | lr 9.30e-04 | grad 0.4908 +2026-04-10 04:49:16 - INFO - parrotllm.training - step 940 | epoch 0 | loss 4.2716 | lr 9.40e-04 | grad 0.4917 +2026-04-10 04:49:19 - INFO - parrotllm.training - step 950 | epoch 0 | loss 4.1942 | lr 9.50e-04 | grad 0.3961 +2026-04-10 04:49:22 - INFO - parrotllm.training - step 960 | epoch 0 | loss 4.2838 | lr 9.60e-04 | grad 0.4960 +2026-04-10 04:49:25 - INFO - parrotllm.training - step 970 | epoch 0 | loss 4.3490 | lr 9.70e-04 | grad 0.4506 +2026-04-10 04:49:28 - INFO - parrotllm.training - step 980 | epoch 0 | loss 4.2449 | lr 9.80e-04 | grad 0.3963 +2026-04-10 04:49:31 - INFO - parrotllm.training - step 990 | epoch 0 | loss 4.2546 | lr 9.90e-04 | grad 0.3983 +2026-04-10 04:49:35 - INFO - parrotllm.training - step 1000 | epoch 0 | loss 4.2183 | lr 1.00e-03 | grad 0.3921 +2026-04-10 04:49:35 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 04:49:35 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 04:49:38 - INFO - parrotllm.training - Train: loss=4.2183, ppl=67.92 +2026-04-10 04:49:38 - INFO - parrotllm.training - Val: loss=4.1572, ppl=63.89 +2026-04-10 04:49:38 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 04:49:38 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_4p1572_epoch_0000_step_0001000.pt +2026-04-10 04:49:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 04:49:42 - INFO - parrotllm.training - step 1010 | epoch 0 | loss 4.1723 | lr 1.01e-03 | grad 0.4209 +2026-04-10 04:49:45 - INFO - parrotllm.training - step 1020 | epoch 0 | loss 4.2617 | lr 1.02e-03 | grad 0.3565 +2026-04-10 04:49:48 - INFO - parrotllm.training - step 1030 | epoch 0 | loss 4.2861 | lr 1.03e-03 | grad 0.3755 +2026-04-10 04:49:51 - INFO - parrotllm.training - step 1040 | epoch 0 | loss 4.1141 | lr 1.04e-03 | grad 0.4240 +2026-04-10 04:49:54 - INFO - parrotllm.training - step 1050 | epoch 0 | loss 4.2139 | lr 1.05e-03 | grad 0.5972 +2026-04-10 04:49:57 - INFO - parrotllm.training - step 1060 | epoch 0 | loss 4.2191 | lr 1.06e-03 | grad 0.4055 +2026-04-10 04:50:00 - INFO - parrotllm.training - step 1070 | epoch 0 | loss 4.1973 | lr 1.07e-03 | grad 0.5353 +2026-04-10 04:50:04 - INFO - parrotllm.training - step 1080 | epoch 0 | loss 4.1032 | lr 1.08e-03 | grad 0.3829 +2026-04-10 04:50:07 - INFO - parrotllm.training - step 1090 | epoch 0 | loss 4.1706 | lr 1.09e-03 | grad 0.3697 +2026-04-10 04:50:10 - INFO - parrotllm.training - step 1100 | epoch 0 | loss 4.1431 | lr 1.10e-03 | grad 0.3683 +2026-04-10 04:50:13 - INFO - parrotllm.training - step 1110 | epoch 0 | loss 4.0387 | lr 1.11e-03 | grad 0.4808 +2026-04-10 04:50:16 - INFO - parrotllm.training - step 1120 | epoch 0 | loss 4.0924 | lr 1.12e-03 | grad 0.3215 +2026-04-10 04:50:19 - INFO - parrotllm.training - step 1130 | epoch 0 | loss 4.1511 | lr 1.13e-03 | grad 0.3576 +2026-04-10 04:50:22 - INFO - parrotllm.training - step 1140 | epoch 0 | loss 4.1988 | lr 1.14e-03 | grad 0.3696 +2026-04-10 04:50:25 - INFO - parrotllm.training - step 1150 | epoch 0 | loss 4.1387 | lr 1.15e-03 | grad 0.3822 +2026-04-10 04:50:28 - INFO - parrotllm.training - step 1160 | epoch 0 | loss 4.1581 | lr 1.16e-03 | grad 0.4169 +2026-04-10 04:50:31 - INFO - parrotllm.training - step 1170 | epoch 0 | loss 4.0991 | lr 1.17e-03 | grad 0.3861 +2026-04-10 04:50:34 - INFO - parrotllm.training - step 1180 | epoch 0 | loss 4.1479 | lr 1.18e-03 | grad 0.3350 +2026-04-10 04:50:37 - INFO - parrotllm.training - step 1190 | epoch 0 | loss 4.0180 | lr 1.19e-03 | grad 0.3589 +2026-04-10 04:50:40 - INFO - parrotllm.training - step 1200 | epoch 0 | loss 4.1466 | lr 1.20e-03 | grad 0.3126 +2026-04-10 04:50:43 - INFO - parrotllm.training - step 1210 | epoch 0 | loss 4.1714 | lr 1.21e-03 | grad 0.3834 +2026-04-10 04:50:46 - INFO - parrotllm.training - step 1220 | epoch 0 | loss 4.0119 | lr 1.22e-03 | grad 0.3510 +2026-04-10 04:50:49 - INFO - parrotllm.training - step 1230 | epoch 0 | loss 4.1283 | lr 1.23e-03 | grad 0.3098 +2026-04-10 04:50:52 - INFO - parrotllm.training - step 1240 | epoch 0 | loss 4.0418 | lr 1.24e-03 | grad 0.3464 +2026-04-10 04:50:56 - INFO - parrotllm.training - step 1250 | epoch 0 | loss 4.1016 | lr 1.25e-03 | grad 0.2993 +2026-04-10 04:50:59 - INFO - parrotllm.training - step 1260 | epoch 0 | loss 3.9906 | lr 1.26e-03 | grad 0.3916 +2026-04-10 04:51:02 - INFO - parrotllm.training - step 1270 | epoch 0 | loss 4.0465 | lr 1.27e-03 | grad 0.3397 +2026-04-10 04:51:05 - INFO - parrotllm.training - step 1280 | epoch 0 | loss 4.1436 | lr 1.28e-03 | grad 0.3048 +2026-04-10 04:51:08 - INFO - parrotllm.training - step 1290 | epoch 0 | loss 4.1168 | lr 1.29e-03 | grad 0.3337 +2026-04-10 04:51:11 - INFO - parrotllm.training - step 1300 | epoch 0 | loss 4.0372 | lr 1.30e-03 | grad 0.3038 +2026-04-10 04:51:14 - INFO - parrotllm.training - step 1310 | epoch 0 | loss 4.0569 | lr 1.31e-03 | grad 0.3885 +2026-04-10 04:51:17 - INFO - parrotllm.training - step 1320 | epoch 0 | loss 4.1448 | lr 1.32e-03 | grad 0.3448 +2026-04-10 04:51:20 - INFO - parrotllm.training - step 1330 | epoch 0 | loss 4.0612 | lr 1.33e-03 | grad 0.3062 +2026-04-10 04:51:23 - INFO - parrotllm.training - step 1340 | epoch 0 | loss 4.0183 | lr 1.34e-03 | grad 0.3420 +2026-04-10 04:51:26 - INFO - parrotllm.training - step 1350 | epoch 0 | loss 3.9662 | lr 1.35e-03 | grad 0.3083 +2026-04-10 04:51:29 - INFO - parrotllm.training - step 1360 | epoch 0 | loss 3.9693 | lr 1.36e-03 | grad 0.3283 +2026-04-10 04:51:32 - INFO - parrotllm.training - step 1370 | epoch 0 | loss 3.9823 | lr 1.37e-03 | grad 0.2861 +2026-04-10 04:51:35 - INFO - parrotllm.training - step 1380 | epoch 0 | loss 4.0511 | lr 1.38e-03 | grad 0.2932 +2026-04-10 04:51:38 - INFO - parrotllm.training - step 1390 | epoch 0 | loss 4.1239 | lr 1.39e-03 | grad 0.2961 +2026-04-10 04:51:41 - INFO - parrotllm.training - step 1400 | epoch 0 | loss 3.9574 | lr 1.40e-03 | grad 0.3306 +2026-04-10 04:51:44 - INFO - parrotllm.training - step 1410 | epoch 0 | loss 4.0071 | lr 1.41e-03 | grad 0.3677 +2026-04-10 04:51:47 - INFO - parrotllm.training - step 1420 | epoch 0 | loss 4.1290 | lr 1.42e-03 | grad 0.2945 +2026-04-10 04:51:51 - INFO - parrotllm.training - step 1430 | epoch 0 | loss 4.0558 | lr 1.43e-03 | grad 0.2611 +2026-04-10 04:51:54 - INFO - parrotllm.training - step 1440 | epoch 0 | loss 3.9244 | lr 1.44e-03 | grad 0.3125 +2026-04-10 04:51:57 - INFO - parrotllm.training - step 1450 | epoch 0 | loss 4.0237 | lr 1.45e-03 | grad 0.3493 +2026-04-10 04:52:00 - INFO - parrotllm.training - step 1460 | epoch 0 | loss 4.0275 | lr 1.46e-03 | grad 0.2649 +2026-04-10 04:52:03 - INFO - parrotllm.training - step 1470 | epoch 0 | loss 4.0127 | lr 1.47e-03 | grad 0.3133 +2026-04-10 04:52:06 - INFO - parrotllm.training - step 1480 | epoch 0 | loss 4.0248 | lr 1.48e-03 | grad 0.3212 +2026-04-10 04:52:09 - INFO - parrotllm.training - step 1490 | epoch 0 | loss 4.0056 | lr 1.49e-03 | grad 0.2817 +2026-04-10 04:52:12 - INFO - parrotllm.training - step 1500 | epoch 0 | loss 3.9132 | lr 1.50e-03 | grad 0.2868 +2026-04-10 04:52:12 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 04:52:12 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 04:52:15 - INFO - parrotllm.training - Train: loss=3.9132, ppl=50.06 +2026-04-10 04:52:15 - INFO - parrotllm.training - Val: loss=3.9069, ppl=49.74 +2026-04-10 04:52:15 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 04:52:16 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p9069_epoch_0000_step_0001500.pt +2026-04-10 04:52:17 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 04:52:20 - INFO - parrotllm.training - step 1510 | epoch 0 | loss 4.1191 | lr 1.50e-03 | grad 0.3409 +2026-04-10 04:52:23 - INFO - parrotllm.training - step 1520 | epoch 0 | loss 4.0255 | lr 1.50e-03 | grad 0.2606 +2026-04-10 04:52:26 - INFO - parrotllm.training - step 1530 | epoch 0 | loss 3.9488 | lr 1.50e-03 | grad 0.3600 +2026-04-10 04:52:29 - INFO - parrotllm.training - step 1540 | epoch 0 | loss 4.0256 | lr 1.50e-03 | grad 0.2576 +2026-04-10 04:52:32 - INFO - parrotllm.training - step 1550 | epoch 0 | loss 3.9716 | lr 1.50e-03 | grad 0.2740 +2026-04-10 04:52:35 - INFO - parrotllm.training - step 1560 | epoch 0 | loss 3.9937 | lr 1.50e-03 | grad 0.2716 +2026-04-10 04:52:38 - INFO - parrotllm.training - step 1570 | epoch 0 | loss 4.0103 | lr 1.50e-03 | grad 0.2546 +2026-04-10 04:52:41 - INFO - parrotllm.training - step 1580 | epoch 0 | loss 3.9714 | lr 1.50e-03 | grad 0.2764 +2026-04-10 04:52:44 - INFO - parrotllm.training - step 1590 | epoch 0 | loss 3.9616 | lr 1.50e-03 | grad 0.2796 +2026-04-10 04:52:47 - INFO - parrotllm.training - step 1600 | epoch 0 | loss 3.9106 | lr 1.50e-03 | grad 0.3663 +2026-04-10 04:52:50 - INFO - parrotllm.training - step 1610 | epoch 0 | loss 4.0125 | lr 1.50e-03 | grad 0.2543 +2026-04-10 04:52:53 - INFO - parrotllm.training - step 1620 | epoch 0 | loss 3.8742 | lr 1.50e-03 | grad 0.2387 +2026-04-10 04:52:56 - INFO - parrotllm.training - step 1630 | epoch 0 | loss 3.8906 | lr 1.50e-03 | grad 0.2497 +2026-04-10 04:52:59 - INFO - parrotllm.training - step 1640 | epoch 0 | loss 3.8676 | lr 1.50e-03 | grad 0.2614 +2026-04-10 04:53:02 - INFO - parrotllm.training - step 1650 | epoch 0 | loss 3.9785 | lr 1.50e-03 | grad 0.2504 +2026-04-10 04:53:05 - INFO - parrotllm.training - step 1660 | epoch 0 | loss 3.8825 | lr 1.50e-03 | grad 0.2773 +2026-04-10 04:53:08 - INFO - parrotllm.training - step 1670 | epoch 0 | loss 3.9980 | lr 1.50e-03 | grad 0.2446 +2026-04-10 04:53:11 - INFO - parrotllm.training - step 1680 | epoch 0 | loss 3.9707 | lr 1.50e-03 | grad 0.2202 +2026-04-10 04:53:15 - INFO - parrotllm.training - step 1690 | epoch 0 | loss 3.8491 | lr 1.50e-03 | grad 0.2803 +2026-04-10 04:53:18 - INFO - parrotllm.training - step 1700 | epoch 0 | loss 4.0247 | lr 1.50e-03 | grad 0.2269 +2026-04-10 04:53:21 - INFO - parrotllm.training - step 1710 | epoch 0 | loss 3.8366 | lr 1.50e-03 | grad 0.2933 +2026-04-10 04:53:24 - INFO - parrotllm.training - step 1720 | epoch 0 | loss 3.9075 | lr 1.50e-03 | grad 0.2593 +2026-04-10 04:53:27 - INFO - parrotllm.training - step 1730 | epoch 0 | loss 3.8033 | lr 1.50e-03 | grad 0.2451 +2026-04-10 04:53:30 - INFO - parrotllm.training - step 1740 | epoch 0 | loss 3.9065 | lr 1.50e-03 | grad 0.2474 +2026-04-10 04:53:33 - INFO - parrotllm.training - step 1750 | epoch 0 | loss 3.8530 | lr 1.50e-03 | grad 0.2584 +2026-04-10 04:53:36 - INFO - parrotllm.training - step 1760 | epoch 0 | loss 3.7851 | lr 1.50e-03 | grad 0.2270 +2026-04-10 04:53:39 - INFO - parrotllm.training - step 1770 | epoch 0 | loss 3.8062 | lr 1.50e-03 | grad 0.2487 +2026-04-10 04:53:42 - INFO - parrotllm.training - step 1780 | epoch 0 | loss 3.8224 | lr 1.50e-03 | grad 0.2425 +2026-04-10 04:53:45 - INFO - parrotllm.training - step 1790 | epoch 0 | loss 3.9372 | lr 1.50e-03 | grad 0.2276 +2026-04-10 04:53:48 - INFO - parrotllm.training - step 1800 | epoch 0 | loss 3.8173 | lr 1.50e-03 | grad 0.2340 +2026-04-10 04:53:51 - INFO - parrotllm.training - step 1810 | epoch 0 | loss 3.9225 | lr 1.50e-03 | grad 0.2270 +2026-04-10 04:53:54 - INFO - parrotllm.training - step 1820 | epoch 0 | loss 3.9024 | lr 1.50e-03 | grad 0.2123 +2026-04-10 04:53:57 - INFO - parrotllm.training - step 1830 | epoch 0 | loss 3.8410 | lr 1.50e-03 | grad 0.2393 +2026-04-10 04:54:00 - INFO - parrotllm.training - step 1840 | epoch 0 | loss 3.9375 | lr 1.50e-03 | grad 0.2840 +2026-04-10 04:54:03 - INFO - parrotllm.training - step 1850 | epoch 0 | loss 3.9601 | lr 1.50e-03 | grad 0.2525 +2026-04-10 04:54:06 - INFO - parrotllm.training - step 1860 | epoch 0 | loss 3.9455 | lr 1.50e-03 | grad 0.2006 +2026-04-10 04:54:09 - INFO - parrotllm.training - step 1870 | epoch 0 | loss 3.9371 | lr 1.50e-03 | grad 0.2074 +2026-04-10 04:54:12 - INFO - parrotllm.training - step 1880 | epoch 0 | loss 3.8891 | lr 1.50e-03 | grad 0.2473 +2026-04-10 04:54:15 - INFO - parrotllm.training - step 1890 | epoch 0 | loss 3.7992 | lr 1.50e-03 | grad 0.2185 +2026-04-10 04:54:19 - INFO - parrotllm.training - step 1900 | epoch 0 | loss 3.9235 | lr 1.50e-03 | grad 0.2207 +2026-04-10 04:54:22 - INFO - parrotllm.training - step 1910 | epoch 0 | loss 4.0034 | lr 1.50e-03 | grad 0.2648 +2026-04-10 04:54:25 - INFO - parrotllm.training - step 1920 | epoch 0 | loss 3.7978 | lr 1.50e-03 | grad 0.2028 +2026-04-10 04:54:28 - INFO - parrotllm.training - step 1930 | epoch 0 | loss 3.8834 | lr 1.50e-03 | grad 0.2447 +2026-04-10 04:54:31 - INFO - parrotllm.training - step 1940 | epoch 0 | loss 3.7384 | lr 1.50e-03 | grad 0.2166 +2026-04-10 04:54:34 - INFO - parrotllm.training - step 1950 | epoch 0 | loss 3.7295 | lr 1.50e-03 | grad 0.2427 +2026-04-10 04:54:37 - INFO - parrotllm.training - step 1960 | epoch 0 | loss 3.9042 | lr 1.50e-03 | grad 0.2266 +2026-04-10 04:54:40 - INFO - parrotllm.training - step 1970 | epoch 0 | loss 3.8449 | lr 1.50e-03 | grad 0.2206 +2026-04-10 04:54:43 - INFO - parrotllm.training - step 1980 | epoch 0 | loss 3.8602 | lr 1.50e-03 | grad 0.2249 +2026-04-10 04:54:46 - INFO - parrotllm.training - step 1990 | epoch 0 | loss 3.9165 | lr 1.50e-03 | grad 0.2305 +2026-04-10 04:54:49 - INFO - parrotllm.training - step 2000 | epoch 0 | loss 3.8687 | lr 1.50e-03 | grad 0.2221 +2026-04-10 04:54:49 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 04:54:49 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 04:54:52 - INFO - parrotllm.training - Train: loss=3.8687, ppl=47.88 +2026-04-10 04:54:52 - INFO - parrotllm.training - Val: loss=3.7745, ppl=43.58 +2026-04-10 04:54:52 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 04:54:53 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p7745_epoch_0000_step_0002000.pt +2026-04-10 04:54:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 04:54:57 - INFO - parrotllm.training - step 2010 | epoch 0 | loss 3.7757 | lr 1.50e-03 | grad 0.2260 +2026-04-10 04:55:00 - INFO - parrotllm.training - step 2020 | epoch 0 | loss 3.8574 | lr 1.50e-03 | grad 0.2278 +2026-04-10 04:55:03 - INFO - parrotllm.training - step 2030 | epoch 0 | loss 3.8779 | lr 1.50e-03 | grad 0.2211 +2026-04-10 04:55:06 - INFO - parrotllm.training - step 2040 | epoch 0 | loss 3.9266 | lr 1.50e-03 | grad 0.2179 +2026-04-10 04:55:09 - INFO - parrotllm.training - step 2050 | epoch 0 | loss 3.8949 | lr 1.50e-03 | grad 0.2127 +2026-04-10 04:55:12 - INFO - parrotllm.training - step 2060 | epoch 0 | loss 3.7985 | lr 1.50e-03 | grad 0.2325 +2026-04-10 04:55:15 - INFO - parrotllm.training - step 2070 | epoch 0 | loss 3.8426 | lr 1.50e-03 | grad 0.2502 +2026-04-10 04:55:18 - INFO - parrotllm.training - step 2080 | epoch 0 | loss 3.8228 | lr 1.50e-03 | grad 0.1906 +2026-04-10 04:55:21 - INFO - parrotllm.training - step 2090 | epoch 0 | loss 3.8874 | lr 1.50e-03 | grad 0.2361 +2026-04-10 04:55:24 - INFO - parrotllm.training - step 2100 | epoch 0 | loss 3.9731 | lr 1.50e-03 | grad 0.2229 +2026-04-10 04:55:27 - INFO - parrotllm.training - step 2110 | epoch 0 | loss 3.8146 | lr 1.50e-03 | grad 0.2210 +2026-04-10 04:55:30 - INFO - parrotllm.training - step 2120 | epoch 0 | loss 3.8099 | lr 1.50e-03 | grad 0.2163 +2026-04-10 04:55:33 - INFO - parrotllm.training - step 2130 | epoch 0 | loss 3.7762 | lr 1.50e-03 | grad 0.1937 +2026-04-10 04:55:36 - INFO - parrotllm.training - step 2140 | epoch 0 | loss 3.9414 | lr 1.50e-03 | grad 0.2408 +2026-04-10 04:55:39 - INFO - parrotllm.training - step 2150 | epoch 0 | loss 3.7547 | lr 1.50e-03 | grad 0.1930 +2026-04-10 04:55:42 - INFO - parrotllm.training - step 2160 | epoch 0 | loss 3.7804 | lr 1.50e-03 | grad 0.2398 +2026-04-10 04:55:45 - INFO - parrotllm.training - step 2170 | epoch 0 | loss 3.7146 | lr 1.50e-03 | grad 0.2238 +2026-04-10 04:55:48 - INFO - parrotllm.training - step 2180 | epoch 0 | loss 3.8114 | lr 1.50e-03 | grad 0.2127 +2026-04-10 04:55:52 - INFO - parrotllm.training - step 2190 | epoch 0 | loss 3.8376 | lr 1.50e-03 | grad 0.1976 +2026-04-10 04:55:55 - INFO - parrotllm.training - step 2200 | epoch 0 | loss 3.7832 | lr 1.50e-03 | grad 0.2137 +2026-04-10 04:55:58 - INFO - parrotllm.training - step 2210 | epoch 0 | loss 3.7847 | lr 1.50e-03 | grad 0.1999 +2026-04-10 04:56:01 - INFO - parrotllm.training - step 2220 | epoch 0 | loss 3.7669 | lr 1.50e-03 | grad 0.1868 +2026-04-10 04:56:04 - INFO - parrotllm.training - step 2230 | epoch 0 | loss 3.9349 | lr 1.50e-03 | grad 0.2327 +2026-04-10 04:56:07 - INFO - parrotllm.training - step 2240 | epoch 0 | loss 3.7337 | lr 1.50e-03 | grad 0.2413 +2026-04-10 04:56:10 - INFO - parrotllm.training - step 2250 | epoch 0 | loss 3.8312 | lr 1.50e-03 | grad 0.1864 +2026-04-10 04:56:13 - INFO - parrotllm.training - step 2260 | epoch 0 | loss 3.6226 | lr 1.50e-03 | grad 0.1986 +2026-04-10 04:56:16 - INFO - parrotllm.training - step 2270 | epoch 0 | loss 3.7924 | lr 1.50e-03 | grad 0.2110 +2026-04-10 04:56:19 - INFO - parrotllm.training - step 2280 | epoch 0 | loss 3.7714 | lr 1.50e-03 | grad 0.2428 +2026-04-10 04:56:22 - INFO - parrotllm.training - step 2290 | epoch 0 | loss 3.8866 | lr 1.50e-03 | grad 0.1969 +2026-04-10 04:56:25 - INFO - parrotllm.training - step 2300 | epoch 0 | loss 3.7779 | lr 1.50e-03 | grad 0.2094 +2026-04-10 04:56:28 - INFO - parrotllm.training - step 2310 | epoch 0 | loss 3.6897 | lr 1.50e-03 | grad 0.1975 +2026-04-10 04:56:31 - INFO - parrotllm.training - step 2320 | epoch 0 | loss 3.8036 | lr 1.50e-03 | grad 0.2105 +2026-04-10 04:56:34 - INFO - parrotllm.training - step 2330 | epoch 0 | loss 3.6945 | lr 1.50e-03 | grad 0.2017 +2026-04-10 04:56:37 - INFO - parrotllm.training - step 2340 | epoch 0 | loss 3.7841 | lr 1.50e-03 | grad 0.2122 +2026-04-10 04:56:40 - INFO - parrotllm.training - step 2350 | epoch 0 | loss 3.8579 | lr 1.50e-03 | grad 0.2028 +2026-04-10 04:56:43 - INFO - parrotllm.training - step 2360 | epoch 0 | loss 3.7870 | lr 1.50e-03 | grad 0.2117 +2026-04-10 04:56:46 - INFO - parrotllm.training - step 2370 | epoch 0 | loss 3.8520 | lr 1.50e-03 | grad 0.2162 +2026-04-10 04:56:50 - INFO - parrotllm.training - step 2380 | epoch 0 | loss 3.8468 | lr 1.50e-03 | grad 0.1947 +2026-04-10 04:56:53 - INFO - parrotllm.training - step 2390 | epoch 0 | loss 3.7474 | lr 1.50e-03 | grad 0.2020 +2026-04-10 04:56:56 - INFO - parrotllm.training - step 2400 | epoch 0 | loss 3.7359 | lr 1.50e-03 | grad 0.1888 +2026-04-10 04:56:59 - INFO - parrotllm.training - step 2410 | epoch 0 | loss 3.7866 | lr 1.50e-03 | grad 0.2120 +2026-04-10 04:57:02 - INFO - parrotllm.training - step 2420 | epoch 0 | loss 3.7995 | lr 1.50e-03 | grad 0.1930 +2026-04-10 04:57:05 - INFO - parrotllm.training - step 2430 | epoch 0 | loss 3.7826 | lr 1.50e-03 | grad 0.1972 +2026-04-10 04:57:08 - INFO - parrotllm.training - step 2440 | epoch 0 | loss 3.7841 | lr 1.50e-03 | grad 0.2226 +2026-04-10 04:57:11 - INFO - parrotllm.training - step 2450 | epoch 0 | loss 3.8388 | lr 1.50e-03 | grad 0.2198 +2026-04-10 04:57:14 - INFO - parrotllm.training - step 2460 | epoch 0 | loss 3.7126 | lr 1.50e-03 | grad 0.2114 +2026-04-10 04:57:17 - INFO - parrotllm.training - step 2470 | epoch 0 | loss 3.8446 | lr 1.50e-03 | grad 0.2360 +2026-04-10 04:57:20 - INFO - parrotllm.training - step 2480 | epoch 0 | loss 3.7456 | lr 1.50e-03 | grad 0.1883 +2026-04-10 04:57:23 - INFO - parrotllm.training - step 2490 | epoch 0 | loss 3.7578 | lr 1.50e-03 | grad 0.2062 +2026-04-10 04:57:26 - INFO - parrotllm.training - step 2500 | epoch 0 | loss 3.6908 | lr 1.50e-03 | grad 0.2117 +2026-04-10 04:57:26 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 04:57:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 04:57:29 - INFO - parrotllm.training - Train: loss=3.6908, ppl=40.08 +2026-04-10 04:57:29 - INFO - parrotllm.training - Val: loss=3.6987, ppl=40.39 +2026-04-10 04:57:29 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 04:57:30 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p6987_epoch_0000_step_0002500.pt +2026-04-10 04:57:31 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 04:57:32 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0002500.pt +2026-04-10 04:57:36 - INFO - parrotllm.training - step 2510 | epoch 0 | loss 3.8023 | lr 1.50e-03 | grad 0.2337 +2026-04-10 04:57:39 - INFO - parrotllm.training - step 2520 | epoch 0 | loss 3.6838 | lr 1.50e-03 | grad 0.2043 +2026-04-10 04:57:42 - INFO - parrotllm.training - step 2530 | epoch 0 | loss 3.7726 | lr 1.50e-03 | grad 0.1833 +2026-04-10 04:57:45 - INFO - parrotllm.training - step 2540 | epoch 0 | loss 3.7002 | lr 1.50e-03 | grad 0.2152 +2026-04-10 04:57:48 - INFO - parrotllm.training - step 2550 | epoch 0 | loss 3.7466 | lr 1.50e-03 | grad 0.2117 +2026-04-10 04:57:51 - INFO - parrotllm.training - step 2560 | epoch 0 | loss 3.6904 | lr 1.50e-03 | grad 0.2072 +2026-04-10 04:57:54 - INFO - parrotllm.training - step 2570 | epoch 0 | loss 3.7674 | lr 1.50e-03 | grad 0.2069 +2026-04-10 04:57:57 - INFO - parrotllm.training - step 2580 | epoch 0 | loss 3.7761 | lr 1.50e-03 | grad 0.2479 +2026-04-10 04:58:00 - INFO - parrotllm.training - step 2590 | epoch 0 | loss 3.9147 | lr 1.50e-03 | grad 0.2183 +2026-04-10 04:58:03 - INFO - parrotllm.training - step 2600 | epoch 0 | loss 3.7610 | lr 1.50e-03 | grad 0.2067 +2026-04-10 04:58:06 - INFO - parrotllm.training - step 2610 | epoch 0 | loss 3.7212 | lr 1.50e-03 | grad 0.2211 +2026-04-10 04:58:09 - INFO - parrotllm.training - step 2620 | epoch 0 | loss 3.6065 | lr 1.50e-03 | grad 0.1837 +2026-04-10 04:58:12 - INFO - parrotllm.training - step 2630 | epoch 0 | loss 3.8170 | lr 1.50e-03 | grad 0.2230 +2026-04-10 04:58:16 - INFO - parrotllm.training - step 2640 | epoch 0 | loss 3.6976 | lr 1.50e-03 | grad 0.2300 +2026-04-10 04:58:19 - INFO - parrotllm.training - step 2650 | epoch 0 | loss 3.8463 | lr 1.50e-03 | grad 0.1912 +2026-04-10 04:58:22 - INFO - parrotllm.training - step 2660 | epoch 0 | loss 3.7168 | lr 1.50e-03 | grad 0.1994 +2026-04-10 04:58:25 - INFO - parrotllm.training - step 2670 | epoch 0 | loss 3.7950 | lr 1.50e-03 | grad 0.2011 +2026-04-10 04:58:28 - INFO - parrotllm.training - step 2680 | epoch 0 | loss 3.7069 | lr 1.50e-03 | grad 0.2143 +2026-04-10 04:58:31 - INFO - parrotllm.training - step 2690 | epoch 0 | loss 3.7621 | lr 1.50e-03 | grad 0.2033 +2026-04-10 04:58:34 - INFO - parrotllm.training - step 2700 | epoch 0 | loss 3.7406 | lr 1.50e-03 | grad 0.2049 +2026-04-10 04:58:37 - INFO - parrotllm.training - step 2710 | epoch 0 | loss 3.7231 | lr 1.50e-03 | grad 0.1929 +2026-04-10 04:58:40 - INFO - parrotllm.training - step 2720 | epoch 0 | loss 3.7778 | lr 1.50e-03 | grad 0.2128 +2026-04-10 04:58:43 - INFO - parrotllm.training - step 2730 | epoch 0 | loss 3.7607 | lr 1.50e-03 | grad 0.1815 +2026-04-10 04:58:46 - INFO - parrotllm.training - step 2740 | epoch 0 | loss 3.6836 | lr 1.50e-03 | grad 0.2179 +2026-04-10 04:58:49 - INFO - parrotllm.training - step 2750 | epoch 0 | loss 3.7106 | lr 1.50e-03 | grad 0.1971 +2026-04-10 04:58:52 - INFO - parrotllm.training - step 2760 | epoch 0 | loss 3.8182 | lr 1.50e-03 | grad 0.1803 +2026-04-10 04:58:55 - INFO - parrotllm.training - step 2770 | epoch 0 | loss 3.7602 | lr 1.50e-03 | grad 0.2116 +2026-04-10 04:58:58 - INFO - parrotllm.training - step 2780 | epoch 0 | loss 3.7838 | lr 1.50e-03 | grad 0.2045 +2026-04-10 04:59:01 - INFO - parrotllm.training - step 2790 | epoch 0 | loss 3.8146 | lr 1.50e-03 | grad 0.2154 +2026-04-10 04:59:04 - INFO - parrotllm.training - step 2800 | epoch 0 | loss 3.7210 | lr 1.50e-03 | grad 0.1875 +2026-04-10 04:59:07 - INFO - parrotllm.training - step 2810 | epoch 0 | loss 3.7269 | lr 1.50e-03 | grad 0.2104 +2026-04-10 04:59:10 - INFO - parrotllm.training - step 2820 | epoch 0 | loss 3.6847 | lr 1.50e-03 | grad 0.1800 +2026-04-10 04:59:13 - INFO - parrotllm.training - step 2830 | epoch 0 | loss 3.7418 | lr 1.50e-03 | grad 0.2108 +2026-04-10 04:59:17 - INFO - parrotllm.training - step 2840 | epoch 0 | loss 3.7878 | lr 1.50e-03 | grad 0.2016 +2026-04-10 04:59:20 - INFO - parrotllm.training - step 2850 | epoch 0 | loss 3.6676 | lr 1.50e-03 | grad 0.2092 +2026-04-10 04:59:23 - INFO - parrotllm.training - step 2860 | epoch 0 | loss 3.6727 | lr 1.50e-03 | grad 0.1951 +2026-04-10 04:59:26 - INFO - parrotllm.training - step 2870 | epoch 0 | loss 3.7485 | lr 1.50e-03 | grad 0.2129 +2026-04-10 04:59:29 - INFO - parrotllm.training - step 2880 | epoch 0 | loss 3.8494 | lr 1.50e-03 | grad 0.1741 +2026-04-10 04:59:32 - INFO - parrotllm.training - step 2890 | epoch 0 | loss 3.7752 | lr 1.50e-03 | grad 0.2149 +2026-04-10 04:59:35 - INFO - parrotllm.training - step 2900 | epoch 0 | loss 3.7344 | lr 1.50e-03 | grad 0.2056 +2026-04-10 04:59:38 - INFO - parrotllm.training - step 2910 | epoch 0 | loss 3.6691 | lr 1.50e-03 | grad 0.1877 +2026-04-10 04:59:41 - INFO - parrotllm.training - step 2920 | epoch 0 | loss 3.6354 | lr 1.50e-03 | grad 0.1800 +2026-04-10 04:59:44 - INFO - parrotllm.training - step 2930 | epoch 0 | loss 3.7899 | lr 1.50e-03 | grad 0.1926 +2026-04-10 04:59:47 - INFO - parrotllm.training - step 2940 | epoch 0 | loss 3.6943 | lr 1.50e-03 | grad 0.2079 +2026-04-10 04:59:50 - INFO - parrotllm.training - step 2950 | epoch 0 | loss 3.6809 | lr 1.50e-03 | grad 0.2310 +2026-04-10 04:59:53 - INFO - parrotllm.training - step 2960 | epoch 0 | loss 3.7747 | lr 1.50e-03 | grad 0.2230 +2026-04-10 04:59:56 - INFO - parrotllm.training - step 2970 | epoch 0 | loss 3.6995 | lr 1.50e-03 | grad 0.1955 +2026-04-10 04:59:59 - INFO - parrotllm.training - step 2980 | epoch 0 | loss 3.7500 | lr 1.50e-03 | grad 0.1807 +2026-04-10 05:00:02 - INFO - parrotllm.training - step 2990 | epoch 0 | loss 3.7209 | lr 1.50e-03 | grad 0.1802 +2026-04-10 05:00:05 - INFO - parrotllm.training - step 3000 | epoch 0 | loss 3.7927 | lr 1.50e-03 | grad 0.2188 +2026-04-10 05:00:05 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:00:05 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:00:08 - INFO - parrotllm.training - Train: loss=3.7927, ppl=44.38 +2026-04-10 05:00:08 - INFO - parrotllm.training - Val: loss=3.6508, ppl=38.50 +2026-04-10 05:00:08 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:00:09 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p6508_epoch_0000_step_0003000.pt +2026-04-10 05:00:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:00:13 - INFO - parrotllm.training - step 3010 | epoch 0 | loss 3.7905 | lr 1.50e-03 | grad 0.1879 +2026-04-10 05:00:16 - INFO - parrotllm.training - step 3020 | epoch 0 | loss 3.7633 | lr 1.50e-03 | grad 0.1986 +2026-04-10 05:00:19 - INFO - parrotllm.training - step 3030 | epoch 0 | loss 3.7025 | lr 1.50e-03 | grad 0.2086 +2026-04-10 05:00:22 - INFO - parrotllm.training - step 3040 | epoch 0 | loss 3.7920 | lr 1.50e-03 | grad 0.1758 +2026-04-10 05:00:25 - INFO - parrotllm.training - step 3050 | epoch 0 | loss 3.7077 | lr 1.50e-03 | grad 0.1967 +2026-04-10 05:00:28 - INFO - parrotllm.training - step 3060 | epoch 0 | loss 3.6909 | lr 1.50e-03 | grad 0.1996 +2026-04-10 05:00:31 - INFO - parrotllm.training - step 3070 | epoch 0 | loss 3.6413 | lr 1.50e-03 | grad 0.1848 +2026-04-10 05:00:34 - INFO - parrotllm.training - step 3080 | epoch 0 | loss 3.6964 | lr 1.50e-03 | grad 0.1987 +2026-04-10 05:00:37 - INFO - parrotllm.training - step 3090 | epoch 0 | loss 3.6404 | lr 1.50e-03 | grad 0.1841 +2026-04-10 05:00:41 - INFO - parrotllm.training - step 3100 | epoch 0 | loss 3.6258 | lr 1.50e-03 | grad 0.2061 +2026-04-10 05:00:44 - INFO - parrotllm.training - step 3110 | epoch 0 | loss 3.8082 | lr 1.50e-03 | grad 0.1914 +2026-04-10 05:00:47 - INFO - parrotllm.training - step 3120 | epoch 0 | loss 3.7043 | lr 1.50e-03 | grad 0.1985 +2026-04-10 05:00:50 - INFO - parrotllm.training - step 3130 | epoch 0 | loss 3.6262 | lr 1.50e-03 | grad 0.1892 +2026-04-10 05:00:53 - INFO - parrotllm.training - step 3140 | epoch 0 | loss 3.7670 | lr 1.50e-03 | grad 0.1932 +2026-04-10 05:00:56 - INFO - parrotllm.training - step 3150 | epoch 0 | loss 3.6496 | lr 1.50e-03 | grad 0.1974 +2026-04-10 05:00:59 - INFO - parrotllm.training - step 3160 | epoch 0 | loss 3.6944 | lr 1.50e-03 | grad 0.1925 +2026-04-10 05:01:02 - INFO - parrotllm.training - step 3170 | epoch 0 | loss 3.6549 | lr 1.50e-03 | grad 0.1789 +2026-04-10 05:01:05 - INFO - parrotllm.training - step 3180 | epoch 0 | loss 3.7091 | lr 1.50e-03 | grad 0.1824 +2026-04-10 05:01:08 - INFO - parrotllm.training - step 3190 | epoch 0 | loss 3.6967 | lr 1.50e-03 | grad 0.1901 +2026-04-10 05:01:11 - INFO - parrotllm.training - step 3200 | epoch 0 | loss 3.6769 | lr 1.50e-03 | grad 0.2006 +2026-04-10 05:01:14 - INFO - parrotllm.training - step 3210 | epoch 0 | loss 3.6684 | lr 1.50e-03 | grad 0.1917 +2026-04-10 05:01:17 - INFO - parrotllm.training - step 3220 | epoch 0 | loss 3.7876 | lr 1.50e-03 | grad 0.1641 +2026-04-10 05:01:20 - INFO - parrotllm.training - step 3230 | epoch 0 | loss 3.6541 | lr 1.50e-03 | grad 0.1921 +2026-04-10 05:01:23 - INFO - parrotllm.training - step 3240 | epoch 0 | loss 3.6717 | lr 1.50e-03 | grad 0.2338 +2026-04-10 05:01:26 - INFO - parrotllm.training - step 3250 | epoch 0 | loss 3.8635 | lr 1.50e-03 | grad 0.1936 +2026-04-10 05:01:29 - INFO - parrotllm.training - step 3260 | epoch 0 | loss 3.7747 | lr 1.50e-03 | grad 0.1959 +2026-04-10 05:01:32 - INFO - parrotllm.training - step 3270 | epoch 0 | loss 3.6558 | lr 1.50e-03 | grad 0.1902 +2026-04-10 05:01:35 - INFO - parrotllm.training - step 3280 | epoch 0 | loss 3.6899 | lr 1.50e-03 | grad 0.1942 +2026-04-10 05:01:38 - INFO - parrotllm.training - step 3290 | epoch 0 | loss 3.8078 | lr 1.50e-03 | grad 0.2039 +2026-04-10 05:01:41 - INFO - parrotllm.training - step 3300 | epoch 0 | loss 3.7583 | lr 1.50e-03 | grad 0.2106 +2026-04-10 05:01:45 - INFO - parrotllm.training - step 3310 | epoch 0 | loss 3.7530 | lr 1.50e-03 | grad 0.1932 +2026-04-10 05:01:48 - INFO - parrotllm.training - step 3320 | epoch 0 | loss 3.7318 | lr 1.50e-03 | grad 0.1691 +2026-04-10 05:01:51 - INFO - parrotllm.training - step 3330 | epoch 0 | loss 3.6766 | lr 1.50e-03 | grad 0.2012 +2026-04-10 05:01:54 - INFO - parrotllm.training - step 3340 | epoch 0 | loss 3.6966 | lr 1.50e-03 | grad 0.1756 +2026-04-10 05:01:57 - INFO - parrotllm.training - step 3350 | epoch 0 | loss 3.6580 | lr 1.50e-03 | grad 0.2469 +2026-04-10 05:02:00 - INFO - parrotllm.training - step 3360 | epoch 0 | loss 3.6967 | lr 1.50e-03 | grad 0.1990 +2026-04-10 05:02:03 - INFO - parrotllm.training - step 3370 | epoch 0 | loss 3.8199 | lr 1.50e-03 | grad 0.1789 +2026-04-10 05:02:06 - INFO - parrotllm.training - step 3380 | epoch 0 | loss 3.6522 | lr 1.50e-03 | grad 0.2149 +2026-04-10 05:02:09 - INFO - parrotllm.training - step 3390 | epoch 0 | loss 3.8127 | lr 1.50e-03 | grad 0.1786 +2026-04-10 05:02:12 - INFO - parrotllm.training - step 3400 | epoch 0 | loss 3.6925 | lr 1.50e-03 | grad 0.1983 +2026-04-10 05:02:15 - INFO - parrotllm.training - step 3410 | epoch 0 | loss 3.6884 | lr 1.50e-03 | grad 0.1976 +2026-04-10 05:02:18 - INFO - parrotllm.training - step 3420 | epoch 0 | loss 3.6972 | lr 1.50e-03 | grad 0.1871 +2026-04-10 05:02:21 - INFO - parrotllm.training - step 3430 | epoch 0 | loss 3.7011 | lr 1.50e-03 | grad 0.1926 +2026-04-10 05:02:24 - INFO - parrotllm.training - step 3440 | epoch 0 | loss 3.7340 | lr 1.50e-03 | grad 0.1919 +2026-04-10 05:02:27 - INFO - parrotllm.training - step 3450 | epoch 0 | loss 3.7139 | lr 1.50e-03 | grad 0.2227 +2026-04-10 05:02:30 - INFO - parrotllm.training - step 3460 | epoch 0 | loss 3.7541 | lr 1.50e-03 | grad 0.2148 +2026-04-10 05:02:33 - INFO - parrotllm.training - step 3470 | epoch 0 | loss 3.5622 | lr 1.50e-03 | grad 0.1871 +2026-04-10 05:02:36 - INFO - parrotllm.training - step 3480 | epoch 0 | loss 3.6899 | lr 1.50e-03 | grad 0.1966 +2026-04-10 05:02:39 - INFO - parrotllm.training - step 3490 | epoch 0 | loss 3.6443 | lr 1.50e-03 | grad 0.1816 +2026-04-10 05:02:42 - INFO - parrotllm.training - step 3500 | epoch 0 | loss 3.7406 | lr 1.50e-03 | grad 0.1799 +2026-04-10 05:02:42 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:02:42 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:02:46 - INFO - parrotllm.training - Train: loss=3.7406, ppl=42.12 +2026-04-10 05:02:46 - INFO - parrotllm.training - Val: loss=3.6186, ppl=37.28 +2026-04-10 05:02:46 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:02:46 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p6186_epoch_0000_step_0003500.pt +2026-04-10 05:02:47 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:02:50 - INFO - parrotllm.training - step 3510 | epoch 0 | loss 3.7866 | lr 1.50e-03 | grad 0.2112 +2026-04-10 05:02:53 - INFO - parrotllm.training - step 3520 | epoch 0 | loss 3.6036 | lr 1.50e-03 | grad 0.1678 +2026-04-10 05:02:56 - INFO - parrotllm.training - step 3530 | epoch 0 | loss 3.6582 | lr 1.50e-03 | grad 0.1809 +2026-04-10 05:03:00 - INFO - parrotllm.training - step 3540 | epoch 0 | loss 3.7322 | lr 1.50e-03 | grad 0.2446 +2026-04-10 05:03:03 - INFO - parrotllm.training - step 3550 | epoch 0 | loss 3.5628 | lr 1.50e-03 | grad 0.1900 +2026-04-10 05:03:06 - INFO - parrotllm.training - step 3560 | epoch 0 | loss 3.7737 | lr 1.50e-03 | grad 0.1931 +2026-04-10 05:03:09 - INFO - parrotllm.training - step 3570 | epoch 0 | loss 3.7060 | lr 1.50e-03 | grad 0.2341 +2026-04-10 05:03:12 - INFO - parrotllm.training - step 3580 | epoch 0 | loss 3.6678 | lr 1.50e-03 | grad 0.1789 +2026-04-10 05:03:15 - INFO - parrotllm.training - step 3590 | epoch 0 | loss 3.7237 | lr 1.50e-03 | grad 0.1936 +2026-04-10 05:03:18 - INFO - parrotllm.training - step 3600 | epoch 0 | loss 3.6483 | lr 1.50e-03 | grad 0.2213 +2026-04-10 05:03:21 - INFO - parrotllm.training - step 3610 | epoch 0 | loss 3.7443 | lr 1.50e-03 | grad 0.2079 +2026-04-10 05:03:24 - INFO - parrotllm.training - step 3620 | epoch 0 | loss 3.6348 | lr 1.50e-03 | grad 0.2166 +2026-04-10 05:03:27 - INFO - parrotllm.training - step 3630 | epoch 0 | loss 3.6428 | lr 1.50e-03 | grad 0.1950 +2026-04-10 05:03:30 - INFO - parrotllm.training - step 3640 | epoch 0 | loss 3.7005 | lr 1.50e-03 | grad 0.2067 +2026-04-10 05:03:33 - INFO - parrotllm.training - step 3650 | epoch 0 | loss 3.7095 | lr 1.50e-03 | grad 0.2065 +2026-04-10 05:03:36 - INFO - parrotllm.training - step 3660 | epoch 0 | loss 3.7869 | lr 1.50e-03 | grad 0.1850 +2026-04-10 05:03:39 - INFO - parrotllm.training - step 3670 | epoch 0 | loss 3.6740 | lr 1.50e-03 | grad 0.2002 +2026-04-10 05:03:42 - INFO - parrotllm.training - step 3680 | epoch 0 | loss 3.6487 | lr 1.50e-03 | grad 0.1848 +2026-04-10 05:03:45 - INFO - parrotllm.training - step 3690 | epoch 0 | loss 3.6072 | lr 1.50e-03 | grad 0.1738 +2026-04-10 05:03:48 - INFO - parrotllm.training - step 3700 | epoch 0 | loss 3.6322 | lr 1.50e-03 | grad 0.2070 +2026-04-10 05:03:51 - INFO - parrotllm.training - step 3710 | epoch 0 | loss 3.7283 | lr 1.50e-03 | grad 0.1770 +2026-04-10 05:03:54 - INFO - parrotllm.training - step 3720 | epoch 0 | loss 3.6929 | lr 1.50e-03 | grad 0.2026 +2026-04-10 05:03:57 - INFO - parrotllm.training - step 3730 | epoch 0 | loss 3.7183 | lr 1.50e-03 | grad 0.2303 +2026-04-10 05:04:00 - INFO - parrotllm.training - step 3740 | epoch 0 | loss 3.6277 | lr 1.50e-03 | grad 0.2128 +2026-04-10 05:04:03 - INFO - parrotllm.training - step 3750 | epoch 0 | loss 3.7910 | lr 1.50e-03 | grad 0.1734 +2026-04-10 05:04:07 - INFO - parrotllm.training - step 3760 | epoch 0 | loss 3.7181 | lr 1.50e-03 | grad 0.2140 +2026-04-10 05:04:10 - INFO - parrotllm.training - step 3770 | epoch 0 | loss 3.6305 | lr 1.50e-03 | grad 0.1838 +2026-04-10 05:04:13 - INFO - parrotllm.training - step 3780 | epoch 0 | loss 3.6798 | lr 1.50e-03 | grad 0.2144 +2026-04-10 05:04:16 - INFO - parrotllm.training - step 3790 | epoch 0 | loss 3.7330 | lr 1.50e-03 | grad 0.1812 +2026-04-10 05:04:19 - INFO - parrotllm.training - step 3800 | epoch 0 | loss 3.7228 | lr 1.50e-03 | grad 0.1822 +2026-04-10 05:04:22 - INFO - parrotllm.training - step 3810 | epoch 0 | loss 3.5582 | lr 1.50e-03 | grad 0.1897 +2026-04-10 05:04:25 - INFO - parrotllm.training - step 3820 | epoch 0 | loss 3.6435 | lr 1.50e-03 | grad 0.1798 +2026-04-10 05:04:28 - INFO - parrotllm.training - step 3830 | epoch 0 | loss 3.6748 | lr 1.50e-03 | grad 0.1777 +2026-04-10 05:04:31 - INFO - parrotllm.training - step 3840 | epoch 0 | loss 3.6536 | lr 1.50e-03 | grad 0.1990 +2026-04-10 05:04:34 - INFO - parrotllm.training - step 3850 | epoch 0 | loss 3.5939 | lr 1.50e-03 | grad 0.1926 +2026-04-10 05:04:37 - INFO - parrotllm.training - step 3860 | epoch 0 | loss 3.7215 | lr 1.50e-03 | grad 0.1797 +2026-04-10 05:04:40 - INFO - parrotllm.training - step 3870 | epoch 0 | loss 3.6114 | lr 1.50e-03 | grad 0.2101 +2026-04-10 05:04:43 - INFO - parrotllm.training - step 3880 | epoch 0 | loss 3.5493 | lr 1.50e-03 | grad 0.2074 +2026-04-10 05:04:46 - INFO - parrotllm.training - step 3890 | epoch 0 | loss 3.6137 | lr 1.50e-03 | grad 0.1951 +2026-04-10 05:04:49 - INFO - parrotllm.training - step 3900 | epoch 0 | loss 3.6590 | lr 1.50e-03 | grad 0.1773 +2026-04-10 05:04:52 - INFO - parrotllm.training - step 3910 | epoch 0 | loss 3.6687 | lr 1.50e-03 | grad 0.1783 +2026-04-10 05:04:55 - INFO - parrotllm.training - step 3920 | epoch 0 | loss 3.6770 | lr 1.50e-03 | grad 0.2238 +2026-04-10 05:04:58 - INFO - parrotllm.training - step 3930 | epoch 0 | loss 3.7297 | lr 1.50e-03 | grad 0.1664 +2026-04-10 05:05:01 - INFO - parrotllm.training - step 3940 | epoch 0 | loss 3.7856 | lr 1.50e-03 | grad 0.1820 +2026-04-10 05:05:04 - INFO - parrotllm.training - step 3950 | epoch 0 | loss 3.5622 | lr 1.50e-03 | grad 0.1887 +2026-04-10 05:05:07 - INFO - parrotllm.training - step 3960 | epoch 0 | loss 3.7252 | lr 1.50e-03 | grad 0.1954 +2026-04-10 05:05:11 - INFO - parrotllm.training - step 3970 | epoch 0 | loss 3.7388 | lr 1.50e-03 | grad 0.1797 +2026-04-10 05:05:14 - INFO - parrotllm.training - step 3980 | epoch 0 | loss 3.6836 | lr 1.50e-03 | grad 0.1866 +2026-04-10 05:05:17 - INFO - parrotllm.training - step 3990 | epoch 0 | loss 3.6094 | lr 1.50e-03 | grad 0.2276 +2026-04-10 05:05:20 - INFO - parrotllm.training - step 4000 | epoch 0 | loss 3.6975 | lr 1.50e-03 | grad 0.2223 +2026-04-10 05:05:20 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:05:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:05:23 - INFO - parrotllm.training - Train: loss=3.6975, ppl=40.35 +2026-04-10 05:05:23 - INFO - parrotllm.training - Val: loss=3.5925, ppl=36.33 +2026-04-10 05:05:23 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:05:24 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p5925_epoch_0000_step_0004000.pt +2026-04-10 05:05:25 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:05:28 - INFO - parrotllm.training - step 4010 | epoch 0 | loss 3.6874 | lr 1.50e-03 | grad 0.1910 +2026-04-10 05:05:31 - INFO - parrotllm.training - step 4020 | epoch 0 | loss 3.6077 | lr 1.50e-03 | grad 0.1714 +2026-04-10 05:05:34 - INFO - parrotllm.training - step 4030 | epoch 0 | loss 3.7023 | lr 1.50e-03 | grad 0.2085 +2026-04-10 05:05:37 - INFO - parrotllm.training - step 4040 | epoch 0 | loss 3.6792 | lr 1.50e-03 | grad 0.1886 +2026-04-10 05:05:40 - INFO - parrotllm.training - step 4050 | epoch 0 | loss 3.5227 | lr 1.50e-03 | grad 0.1963 +2026-04-10 05:05:43 - INFO - parrotllm.training - step 4060 | epoch 0 | loss 3.6889 | lr 1.50e-03 | grad 0.1703 +2026-04-10 05:05:46 - INFO - parrotllm.training - step 4070 | epoch 0 | loss 3.6701 | lr 1.50e-03 | grad 0.1740 +2026-04-10 05:05:49 - INFO - parrotllm.training - step 4080 | epoch 0 | loss 3.6886 | lr 1.50e-03 | grad 0.1958 +2026-04-10 05:05:52 - INFO - parrotllm.training - step 4090 | epoch 0 | loss 3.7420 | lr 1.50e-03 | grad 0.2123 +2026-04-10 05:05:55 - INFO - parrotllm.training - step 4100 | epoch 0 | loss 3.6196 | lr 1.50e-03 | grad 0.2206 +2026-04-10 05:05:58 - INFO - parrotllm.training - step 4110 | epoch 0 | loss 3.6052 | lr 1.50e-03 | grad 0.1966 +2026-04-10 05:06:01 - INFO - parrotllm.training - step 4120 | epoch 0 | loss 3.6203 | lr 1.50e-03 | grad 0.1922 +2026-04-10 05:06:04 - INFO - parrotllm.training - step 4130 | epoch 0 | loss 3.6762 | lr 1.50e-03 | grad 0.1922 +2026-04-10 05:06:07 - INFO - parrotllm.training - step 4140 | epoch 0 | loss 3.6635 | lr 1.50e-03 | grad 0.1707 +2026-04-10 05:06:10 - INFO - parrotllm.training - step 4150 | epoch 0 | loss 3.6370 | lr 1.50e-03 | grad 0.1840 +2026-04-10 05:06:13 - INFO - parrotllm.training - step 4160 | epoch 0 | loss 3.7257 | lr 1.50e-03 | grad 0.1956 +2026-04-10 05:06:16 - INFO - parrotllm.training - step 4170 | epoch 0 | loss 3.6841 | lr 1.50e-03 | grad 0.2326 +2026-04-10 05:06:19 - INFO - parrotllm.training - step 4180 | epoch 0 | loss 3.6566 | lr 1.50e-03 | grad 0.1936 +2026-04-10 05:06:22 - INFO - parrotllm.training - step 4190 | epoch 0 | loss 3.7104 | lr 1.50e-03 | grad 0.1923 +2026-04-10 05:06:25 - INFO - parrotllm.training - step 4200 | epoch 0 | loss 3.6298 | lr 1.50e-03 | grad 0.1857 +2026-04-10 05:06:29 - INFO - parrotllm.training - step 4210 | epoch 0 | loss 3.7005 | lr 1.50e-03 | grad 0.1984 +2026-04-10 05:06:32 - INFO - parrotllm.training - step 4220 | epoch 0 | loss 3.5857 | lr 1.50e-03 | grad 0.1839 +2026-04-10 05:06:35 - INFO - parrotllm.training - step 4230 | epoch 0 | loss 3.7291 | lr 1.50e-03 | grad 0.2205 +2026-04-10 05:06:38 - INFO - parrotllm.training - step 4240 | epoch 0 | loss 3.6873 | lr 1.50e-03 | grad 0.1997 +2026-04-10 05:06:41 - INFO - parrotllm.training - step 4250 | epoch 0 | loss 3.5797 | lr 1.50e-03 | grad 0.1902 +2026-04-10 05:06:44 - INFO - parrotllm.training - step 4260 | epoch 0 | loss 3.6034 | lr 1.50e-03 | grad 0.1807 +2026-04-10 05:06:47 - INFO - parrotllm.training - step 4270 | epoch 0 | loss 3.6198 | lr 1.50e-03 | grad 0.1697 +2026-04-10 05:06:50 - INFO - parrotllm.training - step 4280 | epoch 0 | loss 3.6456 | lr 1.50e-03 | grad 0.1951 +2026-04-10 05:06:53 - INFO - parrotllm.training - step 4290 | epoch 0 | loss 3.6455 | lr 1.50e-03 | grad 0.1811 +2026-04-10 05:06:56 - INFO - parrotllm.training - step 4300 | epoch 0 | loss 3.6668 | lr 1.50e-03 | grad 0.2017 +2026-04-10 05:06:59 - INFO - parrotllm.training - step 4310 | epoch 0 | loss 3.6975 | lr 1.50e-03 | grad 0.1999 +2026-04-10 05:07:02 - INFO - parrotllm.training - step 4320 | epoch 0 | loss 3.6458 | lr 1.50e-03 | grad 0.2042 +2026-04-10 05:07:05 - INFO - parrotllm.training - step 4330 | epoch 0 | loss 3.6412 | lr 1.50e-03 | grad 0.1926 +2026-04-10 05:07:08 - INFO - parrotllm.training - step 4340 | epoch 0 | loss 3.5365 | lr 1.50e-03 | grad 0.2087 +2026-04-10 05:07:11 - INFO - parrotllm.training - step 4350 | epoch 0 | loss 3.5842 | lr 1.50e-03 | grad 0.2012 +2026-04-10 05:07:14 - INFO - parrotllm.training - step 4360 | epoch 0 | loss 3.6892 | lr 1.50e-03 | grad 0.2157 +2026-04-10 05:07:17 - INFO - parrotllm.training - step 4370 | epoch 0 | loss 3.6210 | lr 1.50e-03 | grad 0.1794 +2026-04-10 05:07:20 - INFO - parrotllm.training - step 4380 | epoch 0 | loss 3.6566 | lr 1.50e-03 | grad 0.1933 +2026-04-10 05:07:23 - INFO - parrotllm.training - step 4390 | epoch 0 | loss 3.6703 | lr 1.50e-03 | grad 0.2336 +2026-04-10 05:07:26 - INFO - parrotllm.training - step 4400 | epoch 0 | loss 3.6718 | lr 1.50e-03 | grad 0.1955 +2026-04-10 05:07:30 - INFO - parrotllm.training - step 4410 | epoch 0 | loss 3.6539 | lr 1.50e-03 | grad 0.1795 +2026-04-10 05:07:33 - INFO - parrotllm.training - step 4420 | epoch 0 | loss 3.7252 | lr 1.50e-03 | grad 0.1968 +2026-04-10 05:07:36 - INFO - parrotllm.training - step 4430 | epoch 0 | loss 3.7067 | lr 1.50e-03 | grad 0.2038 +2026-04-10 05:07:39 - INFO - parrotllm.training - step 4440 | epoch 0 | loss 3.5953 | lr 1.50e-03 | grad 0.1821 +2026-04-10 05:07:42 - INFO - parrotllm.training - step 4450 | epoch 0 | loss 3.5755 | lr 1.50e-03 | grad 0.1922 +2026-04-10 05:07:45 - INFO - parrotllm.training - step 4460 | epoch 0 | loss 3.6077 | lr 1.50e-03 | grad 0.2024 +2026-04-10 05:07:48 - INFO - parrotllm.training - step 4470 | epoch 0 | loss 3.6905 | lr 1.50e-03 | grad 0.1679 +2026-04-10 05:07:51 - INFO - parrotllm.training - step 4480 | epoch 0 | loss 3.5590 | lr 1.50e-03 | grad 0.1849 +2026-04-10 05:07:54 - INFO - parrotllm.training - step 4490 | epoch 0 | loss 3.6846 | lr 1.50e-03 | grad 0.2023 +2026-04-10 05:07:57 - INFO - parrotllm.training - step 4500 | epoch 0 | loss 3.7499 | lr 1.50e-03 | grad 0.1717 +2026-04-10 05:07:57 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:07:57 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:08:00 - INFO - parrotllm.training - Train: loss=3.7499, ppl=42.52 +2026-04-10 05:08:00 - INFO - parrotllm.training - Val: loss=3.5712, ppl=35.56 +2026-04-10 05:08:00 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:08:01 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p5712_epoch_0000_step_0004500.pt +2026-04-10 05:08:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:08:05 - INFO - parrotllm.training - step 4510 | epoch 0 | loss 3.6717 | lr 1.50e-03 | grad 0.1783 +2026-04-10 05:08:08 - INFO - parrotllm.training - step 4520 | epoch 0 | loss 3.7466 | lr 1.50e-03 | grad 0.1885 +2026-04-10 05:08:11 - INFO - parrotllm.training - step 4530 | epoch 0 | loss 3.6356 | lr 1.50e-03 | grad 0.1783 +2026-04-10 05:08:14 - INFO - parrotllm.training - step 4540 | epoch 0 | loss 3.5742 | lr 1.50e-03 | grad 0.1749 +2026-04-10 05:08:17 - INFO - parrotllm.training - step 4550 | epoch 0 | loss 3.6907 | lr 1.50e-03 | grad 0.2123 +2026-04-10 05:08:20 - INFO - parrotllm.training - step 4560 | epoch 0 | loss 3.7077 | lr 1.50e-03 | grad 0.1734 +2026-04-10 05:08:23 - INFO - parrotllm.training - step 4570 | epoch 0 | loss 3.6420 | lr 1.50e-03 | grad 0.1879 +2026-04-10 05:08:26 - INFO - parrotllm.training - step 4580 | epoch 0 | loss 3.6021 | lr 1.50e-03 | grad 0.2058 +2026-04-10 05:08:29 - INFO - parrotllm.training - step 4590 | epoch 0 | loss 3.7387 | lr 1.50e-03 | grad 0.1969 +2026-04-10 05:08:32 - INFO - parrotllm.training - step 4600 | epoch 0 | loss 3.5738 | lr 1.50e-03 | grad 0.1998 +2026-04-10 05:08:35 - INFO - parrotllm.training - step 4610 | epoch 0 | loss 3.5681 | lr 1.50e-03 | grad 0.1735 +2026-04-10 05:08:38 - INFO - parrotllm.training - step 4620 | epoch 0 | loss 3.7108 | lr 1.50e-03 | grad 0.2261 +2026-04-10 05:08:41 - INFO - parrotllm.training - step 4630 | epoch 0 | loss 3.6147 | lr 1.50e-03 | grad 0.1787 +2026-04-10 05:08:44 - INFO - parrotllm.training - step 4640 | epoch 0 | loss 3.6238 | lr 1.50e-03 | grad 0.1800 +2026-04-10 05:08:48 - INFO - parrotllm.training - step 4650 | epoch 0 | loss 3.6101 | lr 1.50e-03 | grad 0.2018 +2026-04-10 05:08:51 - INFO - parrotllm.training - step 4660 | epoch 0 | loss 3.5931 | lr 1.50e-03 | grad 0.1759 +2026-04-10 05:08:54 - INFO - parrotllm.training - step 4670 | epoch 0 | loss 3.6657 | lr 1.50e-03 | grad 0.1925 +2026-04-10 05:08:57 - INFO - parrotllm.training - step 4680 | epoch 0 | loss 3.6849 | lr 1.50e-03 | grad 0.1944 +2026-04-10 05:09:00 - INFO - parrotllm.training - step 4690 | epoch 0 | loss 3.6875 | lr 1.50e-03 | grad 0.1852 +2026-04-10 05:09:03 - INFO - parrotllm.training - step 4700 | epoch 0 | loss 3.6302 | lr 1.50e-03 | grad 0.2080 +2026-04-10 05:09:06 - INFO - parrotllm.training - step 4710 | epoch 0 | loss 3.6084 | lr 1.50e-03 | grad 0.1985 +2026-04-10 05:09:09 - INFO - parrotllm.training - step 4720 | epoch 0 | loss 3.7594 | lr 1.50e-03 | grad 0.1679 +2026-04-10 05:09:12 - INFO - parrotllm.training - step 4730 | epoch 0 | loss 3.5361 | lr 1.50e-03 | grad 0.2007 +2026-04-10 05:09:15 - INFO - parrotllm.training - step 4740 | epoch 0 | loss 3.6335 | lr 1.50e-03 | grad 0.2022 +2026-04-10 05:09:18 - INFO - parrotllm.training - step 4750 | epoch 0 | loss 3.7470 | lr 1.50e-03 | grad 0.1961 +2026-04-10 05:09:21 - INFO - parrotllm.training - step 4760 | epoch 0 | loss 3.6592 | lr 1.50e-03 | grad 0.1840 +2026-04-10 05:09:24 - INFO - parrotllm.training - step 4770 | epoch 0 | loss 3.5561 | lr 1.50e-03 | grad 0.1726 +2026-04-10 05:09:27 - INFO - parrotllm.training - step 4780 | epoch 0 | loss 3.5935 | lr 1.50e-03 | grad 0.1979 +2026-04-10 05:09:30 - INFO - parrotllm.training - step 4790 | epoch 0 | loss 3.5956 | lr 1.50e-03 | grad 0.1942 +2026-04-10 05:09:33 - INFO - parrotllm.training - step 4800 | epoch 0 | loss 3.6247 | lr 1.50e-03 | grad 0.1621 +2026-04-10 05:09:36 - INFO - parrotllm.training - step 4810 | epoch 0 | loss 3.5341 | lr 1.50e-03 | grad 0.2025 +2026-04-10 05:09:39 - INFO - parrotllm.training - step 4820 | epoch 0 | loss 3.5986 | lr 1.50e-03 | grad 0.1952 +2026-04-10 05:09:42 - INFO - parrotllm.training - step 4830 | epoch 0 | loss 3.6868 | lr 1.50e-03 | grad 0.2000 +2026-04-10 05:09:45 - INFO - parrotllm.training - step 4840 | epoch 0 | loss 3.5749 | lr 1.50e-03 | grad 0.1932 +2026-04-10 05:09:48 - INFO - parrotllm.training - step 4850 | epoch 0 | loss 3.5581 | lr 1.50e-03 | grad 0.1670 +2026-04-10 05:09:52 - INFO - parrotllm.training - step 4860 | epoch 0 | loss 3.5847 | lr 1.50e-03 | grad 0.1935 +2026-04-10 05:09:55 - INFO - parrotllm.training - step 4870 | epoch 0 | loss 3.6461 | lr 1.50e-03 | grad 0.2020 +2026-04-10 05:09:58 - INFO - parrotllm.training - step 4880 | epoch 0 | loss 3.5444 | lr 1.50e-03 | grad 0.2187 +2026-04-10 05:10:01 - INFO - parrotllm.training - step 4890 | epoch 0 | loss 3.6317 | lr 1.50e-03 | grad 0.1894 +2026-04-10 05:10:04 - INFO - parrotllm.training - step 4900 | epoch 0 | loss 3.5882 | lr 1.50e-03 | grad 0.1943 +2026-04-10 05:10:07 - INFO - parrotllm.training - step 4910 | epoch 0 | loss 3.6049 | lr 1.50e-03 | grad 0.1757 +2026-04-10 05:10:10 - INFO - parrotllm.training - step 4920 | epoch 0 | loss 3.7642 | lr 1.50e-03 | grad 0.1918 +2026-04-10 05:10:13 - INFO - parrotllm.training - step 4930 | epoch 0 | loss 3.5876 | lr 1.50e-03 | grad 0.1981 +2026-04-10 05:10:16 - INFO - parrotllm.training - step 4940 | epoch 0 | loss 3.6498 | lr 1.50e-03 | grad 0.2181 +2026-04-10 05:10:19 - INFO - parrotllm.training - step 4950 | epoch 0 | loss 3.7547 | lr 1.50e-03 | grad 0.2072 +2026-04-10 05:10:22 - INFO - parrotllm.training - step 4960 | epoch 0 | loss 3.5765 | lr 1.50e-03 | grad 0.1824 +2026-04-10 05:10:25 - INFO - parrotllm.training - step 4970 | epoch 0 | loss 3.6275 | lr 1.50e-03 | grad 0.2229 +2026-04-10 05:10:28 - INFO - parrotllm.training - step 4980 | epoch 0 | loss 3.6363 | lr 1.50e-03 | grad 0.2182 +2026-04-10 05:10:31 - INFO - parrotllm.training - step 4990 | epoch 0 | loss 3.5728 | lr 1.50e-03 | grad 0.1941 +2026-04-10 05:10:34 - INFO - parrotllm.training - step 5000 | epoch 0 | loss 3.6880 | lr 1.50e-03 | grad 0.2132 +2026-04-10 05:10:34 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:10:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:10:37 - INFO - parrotllm.training - Train: loss=3.6880, ppl=39.97 +2026-04-10 05:10:37 - INFO - parrotllm.training - Val: loss=3.5577, ppl=35.08 +2026-04-10 05:10:37 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:10:38 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p5577_epoch_0000_step_0005000.pt +2026-04-10 05:10:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:10:40 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0005000.pt +2026-04-10 05:10:44 - INFO - parrotllm.training - step 5010 | epoch 0 | loss 3.7004 | lr 1.50e-03 | grad 0.1676 +2026-04-10 05:10:47 - INFO - parrotllm.training - step 5020 | epoch 0 | loss 3.6914 | lr 1.50e-03 | grad 0.1936 +2026-04-10 05:10:50 - INFO - parrotllm.training - step 5030 | epoch 0 | loss 3.6399 | lr 1.50e-03 | grad 0.2013 +2026-04-10 05:10:53 - INFO - parrotllm.training - step 5040 | epoch 0 | loss 3.4923 | lr 1.50e-03 | grad 0.1810 +2026-04-10 05:10:56 - INFO - parrotllm.training - step 5050 | epoch 0 | loss 3.6092 | lr 1.50e-03 | grad 0.1757 +2026-04-10 05:10:59 - INFO - parrotllm.training - step 5060 | epoch 0 | loss 3.6424 | lr 1.50e-03 | grad 0.1829 +2026-04-10 05:11:03 - INFO - parrotllm.training - step 5070 | epoch 0 | loss 3.5533 | lr 1.50e-03 | grad 0.1925 +2026-04-10 05:11:06 - INFO - parrotllm.training - step 5080 | epoch 0 | loss 3.5768 | lr 1.50e-03 | grad 0.1846 +2026-04-10 05:11:09 - INFO - parrotllm.training - step 5090 | epoch 0 | loss 3.5620 | lr 1.50e-03 | grad 0.1793 +2026-04-10 05:11:12 - INFO - parrotllm.training - step 5100 | epoch 0 | loss 3.7331 | lr 1.50e-03 | grad 0.1839 +2026-04-10 05:11:15 - INFO - parrotllm.training - step 5110 | epoch 0 | loss 3.6108 | lr 1.50e-03 | grad 0.1881 +2026-04-10 05:11:18 - INFO - parrotllm.training - step 5120 | epoch 0 | loss 3.6079 | lr 1.50e-03 | grad 0.2031 +2026-04-10 05:11:21 - INFO - parrotllm.training - step 5130 | epoch 0 | loss 3.6668 | lr 1.50e-03 | grad 0.1895 +2026-04-10 05:11:24 - INFO - parrotllm.training - step 5140 | epoch 0 | loss 3.6734 | lr 1.50e-03 | grad 0.2051 +2026-04-10 05:11:27 - INFO - parrotllm.training - step 5150 | epoch 0 | loss 3.5187 | lr 1.50e-03 | grad 0.1856 +2026-04-10 05:11:30 - INFO - parrotllm.training - step 5160 | epoch 0 | loss 3.5305 | lr 1.50e-03 | grad 0.1917 +2026-04-10 05:11:33 - INFO - parrotllm.training - step 5170 | epoch 0 | loss 3.5251 | lr 1.50e-03 | grad 0.1882 +2026-04-10 05:11:36 - INFO - parrotllm.training - step 5180 | epoch 0 | loss 3.5845 | lr 1.50e-03 | grad 0.2079 +2026-04-10 05:11:39 - INFO - parrotllm.training - step 5190 | epoch 0 | loss 3.5835 | lr 1.50e-03 | grad 0.1791 +2026-04-10 05:11:42 - INFO - parrotllm.training - step 5200 | epoch 0 | loss 3.6155 | lr 1.50e-03 | grad 0.2035 +2026-04-10 05:11:45 - INFO - parrotllm.training - step 5210 | epoch 0 | loss 3.5812 | lr 1.50e-03 | grad 0.1941 +2026-04-10 05:11:48 - INFO - parrotllm.training - step 5220 | epoch 0 | loss 3.5857 | lr 1.50e-03 | grad 0.1810 +2026-04-10 05:11:51 - INFO - parrotllm.training - step 5230 | epoch 0 | loss 3.6438 | lr 1.50e-03 | grad 0.1885 +2026-04-10 05:11:54 - INFO - parrotllm.training - step 5240 | epoch 0 | loss 3.6036 | lr 1.50e-03 | grad 0.1886 +2026-04-10 05:11:58 - INFO - parrotllm.training - step 5250 | epoch 0 | loss 3.5624 | lr 1.50e-03 | grad 0.1992 +2026-04-10 05:12:01 - INFO - parrotllm.training - step 5260 | epoch 0 | loss 3.6035 | lr 1.50e-03 | grad 0.1981 +2026-04-10 05:12:04 - INFO - parrotllm.training - step 5270 | epoch 0 | loss 3.6450 | lr 1.50e-03 | grad 0.1749 +2026-04-10 05:12:07 - INFO - parrotllm.training - step 5280 | epoch 0 | loss 3.5706 | lr 1.50e-03 | grad 0.2066 +2026-04-10 05:12:10 - INFO - parrotllm.training - step 5290 | epoch 0 | loss 3.5098 | lr 1.50e-03 | grad 0.1668 +2026-04-10 05:12:13 - INFO - parrotllm.training - step 5300 | epoch 0 | loss 3.5922 | lr 1.50e-03 | grad 0.1757 +2026-04-10 05:12:16 - INFO - parrotllm.training - step 5310 | epoch 0 | loss 3.6019 | lr 1.50e-03 | grad 0.1800 +2026-04-10 05:12:19 - INFO - parrotllm.training - step 5320 | epoch 0 | loss 3.5133 | lr 1.50e-03 | grad 0.1870 +2026-04-10 05:12:22 - INFO - parrotllm.training - step 5330 | epoch 0 | loss 3.6227 | lr 1.50e-03 | grad 0.1928 +2026-04-10 05:12:25 - INFO - parrotllm.training - step 5340 | epoch 0 | loss 3.6071 | lr 1.50e-03 | grad 0.1738 +2026-04-10 05:12:28 - INFO - parrotllm.training - step 5350 | epoch 0 | loss 3.6044 | lr 1.50e-03 | grad 0.1895 +2026-04-10 05:12:31 - INFO - parrotllm.training - step 5360 | epoch 0 | loss 3.6057 | lr 1.50e-03 | grad 0.2000 +2026-04-10 05:12:34 - INFO - parrotllm.training - step 5370 | epoch 0 | loss 3.6380 | lr 1.50e-03 | grad 0.1998 +2026-04-10 05:12:37 - INFO - parrotllm.training - step 5380 | epoch 0 | loss 3.6481 | lr 1.50e-03 | grad 0.1803 +2026-04-10 05:12:40 - INFO - parrotllm.training - step 5390 | epoch 0 | loss 3.5418 | lr 1.50e-03 | grad 0.1933 +2026-04-10 05:12:43 - INFO - parrotllm.training - step 5400 | epoch 0 | loss 3.6549 | lr 1.50e-03 | grad 0.1714 +2026-04-10 05:12:46 - INFO - parrotllm.training - step 5410 | epoch 0 | loss 3.5929 | lr 1.50e-03 | grad 0.2181 +2026-04-10 05:12:49 - INFO - parrotllm.training - step 5420 | epoch 0 | loss 3.5377 | lr 1.50e-03 | grad 0.1970 +2026-04-10 05:12:52 - INFO - parrotllm.training - step 5430 | epoch 0 | loss 3.6428 | lr 1.50e-03 | grad 0.1834 +2026-04-10 05:12:56 - INFO - parrotllm.training - step 5440 | epoch 0 | loss 3.6634 | lr 1.50e-03 | grad 0.2019 +2026-04-10 05:12:59 - INFO - parrotllm.training - step 5450 | epoch 0 | loss 3.6426 | lr 1.50e-03 | grad 0.1944 +2026-04-10 05:13:02 - INFO - parrotllm.training - step 5460 | epoch 0 | loss 3.6215 | lr 1.50e-03 | grad 0.1858 +2026-04-10 05:13:05 - INFO - parrotllm.training - step 5470 | epoch 0 | loss 3.6696 | lr 1.50e-03 | grad 0.1720 +2026-04-10 05:13:08 - INFO - parrotllm.training - step 5480 | epoch 0 | loss 3.5236 | lr 1.50e-03 | grad 0.2158 +2026-04-10 05:13:11 - INFO - parrotllm.training - step 5490 | epoch 0 | loss 3.6286 | lr 1.50e-03 | grad 0.1807 +2026-04-10 05:13:14 - INFO - parrotllm.training - step 5500 | epoch 0 | loss 3.6471 | lr 1.50e-03 | grad 0.2052 +2026-04-10 05:13:14 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:13:14 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:13:17 - INFO - parrotllm.training - Train: loss=3.6471, ppl=38.36 +2026-04-10 05:13:17 - INFO - parrotllm.training - Val: loss=3.5401, ppl=34.47 +2026-04-10 05:13:17 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:13:18 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p5401_epoch_0000_step_0005500.pt +2026-04-10 05:13:19 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:13:22 - INFO - parrotllm.training - step 5510 | epoch 0 | loss 3.5425 | lr 1.50e-03 | grad 0.2011 +2026-04-10 05:13:25 - INFO - parrotllm.training - step 5520 | epoch 0 | loss 3.5758 | lr 1.50e-03 | grad 0.1809 +2026-04-10 05:13:28 - INFO - parrotllm.training - step 5530 | epoch 0 | loss 3.6184 | lr 1.50e-03 | grad 0.1780 +2026-04-10 05:13:31 - INFO - parrotllm.training - step 5540 | epoch 0 | loss 3.5127 | lr 1.50e-03 | grad 0.1714 +2026-04-10 05:13:34 - INFO - parrotllm.training - step 5550 | epoch 0 | loss 3.5313 | lr 1.50e-03 | grad 0.2028 +2026-04-10 05:13:37 - INFO - parrotllm.training - step 5560 | epoch 0 | loss 3.6801 | lr 1.50e-03 | grad 0.2071 +2026-04-10 05:13:40 - INFO - parrotllm.training - step 5570 | epoch 0 | loss 3.6468 | lr 1.50e-03 | grad 0.2192 +2026-04-10 05:13:43 - INFO - parrotllm.training - step 5580 | epoch 0 | loss 3.5504 | lr 1.50e-03 | grad 0.1839 +2026-04-10 05:13:46 - INFO - parrotllm.training - step 5590 | epoch 0 | loss 3.5496 | lr 1.50e-03 | grad 0.1631 +2026-04-10 05:13:49 - INFO - parrotllm.training - step 5600 | epoch 0 | loss 3.5629 | lr 1.50e-03 | grad 0.1851 +2026-04-10 05:13:52 - INFO - parrotllm.training - step 5610 | epoch 0 | loss 3.7054 | lr 1.50e-03 | grad 0.1730 +2026-04-10 05:13:55 - INFO - parrotllm.training - step 5620 | epoch 0 | loss 3.6424 | lr 1.50e-03 | grad 0.1992 +2026-04-10 05:13:58 - INFO - parrotllm.training - step 5630 | epoch 0 | loss 3.5128 | lr 1.50e-03 | grad 0.1850 +2026-04-10 05:14:01 - INFO - parrotllm.training - step 5640 | epoch 0 | loss 3.6565 | lr 1.50e-03 | grad 0.1787 +2026-04-10 05:14:04 - INFO - parrotllm.training - step 5650 | epoch 0 | loss 3.6060 | lr 1.50e-03 | grad 0.2187 +2026-04-10 05:14:08 - INFO - parrotllm.training - step 5660 | epoch 0 | loss 3.6853 | lr 1.50e-03 | grad 0.1923 +2026-04-10 05:14:11 - INFO - parrotllm.training - step 5670 | epoch 0 | loss 3.6000 | lr 1.50e-03 | grad 0.1860 +2026-04-10 05:14:14 - INFO - parrotllm.training - step 5680 | epoch 0 | loss 3.5601 | lr 1.50e-03 | grad 0.2018 +2026-04-10 05:14:17 - INFO - parrotllm.training - step 5690 | epoch 0 | loss 3.6109 | lr 1.50e-03 | grad 0.1977 +2026-04-10 05:14:20 - INFO - parrotllm.training - step 5700 | epoch 0 | loss 3.5085 | lr 1.50e-03 | grad 0.1911 +2026-04-10 05:14:23 - INFO - parrotllm.training - step 5710 | epoch 0 | loss 3.6265 | lr 1.50e-03 | grad 0.2274 +2026-04-10 05:14:26 - INFO - parrotllm.training - step 5720 | epoch 0 | loss 3.6356 | lr 1.50e-03 | grad 0.1683 +2026-04-10 05:14:29 - INFO - parrotllm.training - step 5730 | epoch 0 | loss 3.5620 | lr 1.50e-03 | grad 0.2053 +2026-04-10 05:14:32 - INFO - parrotllm.training - step 5740 | epoch 0 | loss 3.5652 | lr 1.50e-03 | grad 0.1986 +2026-04-10 05:14:35 - INFO - parrotllm.training - step 5750 | epoch 0 | loss 3.6845 | lr 1.50e-03 | grad 0.1825 +2026-04-10 05:14:38 - INFO - parrotllm.training - step 5760 | epoch 0 | loss 3.4944 | lr 1.50e-03 | grad 0.1980 +2026-04-10 05:14:41 - INFO - parrotllm.training - step 5770 | epoch 0 | loss 3.5567 | lr 1.50e-03 | grad 0.1694 +2026-04-10 05:14:44 - INFO - parrotllm.training - step 5780 | epoch 0 | loss 3.5734 | lr 1.50e-03 | grad 0.1930 +2026-04-10 05:14:47 - INFO - parrotllm.training - step 5790 | epoch 0 | loss 3.5945 | lr 1.50e-03 | grad 0.1825 +2026-04-10 05:14:50 - INFO - parrotllm.training - step 5800 | epoch 0 | loss 3.5725 | lr 1.50e-03 | grad 0.2059 +2026-04-10 05:14:53 - INFO - parrotllm.training - step 5810 | epoch 0 | loss 3.6130 | lr 1.50e-03 | grad 0.1890 +2026-04-10 05:14:56 - INFO - parrotllm.training - step 5820 | epoch 0 | loss 3.6484 | lr 1.50e-03 | grad 0.1828 +2026-04-10 05:14:59 - INFO - parrotllm.training - step 5830 | epoch 0 | loss 3.7121 | lr 1.50e-03 | grad 0.1947 +2026-04-10 05:15:03 - INFO - parrotllm.training - step 5840 | epoch 0 | loss 3.5914 | lr 1.50e-03 | grad 0.1828 +2026-04-10 05:15:06 - INFO - parrotllm.training - step 5850 | epoch 0 | loss 3.6257 | lr 1.50e-03 | grad 0.2088 +2026-04-10 05:15:09 - INFO - parrotllm.training - step 5860 | epoch 0 | loss 3.5617 | lr 1.50e-03 | grad 0.2034 +2026-04-10 05:15:12 - INFO - parrotllm.training - step 5870 | epoch 0 | loss 3.7046 | lr 1.50e-03 | grad 0.1989 +2026-04-10 05:15:15 - INFO - parrotllm.training - step 5880 | epoch 0 | loss 3.5716 | lr 1.50e-03 | grad 0.1821 +2026-04-10 05:15:18 - INFO - parrotllm.training - step 5890 | epoch 0 | loss 3.5674 | lr 1.50e-03 | grad 0.2219 +2026-04-10 05:15:21 - INFO - parrotllm.training - step 5900 | epoch 0 | loss 3.6491 | lr 1.50e-03 | grad 0.1932 +2026-04-10 05:15:24 - INFO - parrotllm.training - step 5910 | epoch 0 | loss 3.6065 | lr 1.50e-03 | grad 0.1720 +2026-04-10 05:15:27 - INFO - parrotllm.training - step 5920 | epoch 0 | loss 3.6957 | lr 1.50e-03 | grad 0.1985 +2026-04-10 05:15:30 - INFO - parrotllm.training - step 5930 | epoch 0 | loss 3.5815 | lr 1.50e-03 | grad 0.2175 +2026-04-10 05:15:33 - INFO - parrotllm.training - step 5940 | epoch 0 | loss 3.5165 | lr 1.50e-03 | grad 0.1949 +2026-04-10 05:15:36 - INFO - parrotllm.training - step 5950 | epoch 0 | loss 3.6534 | lr 1.50e-03 | grad 0.2124 +2026-04-10 05:15:39 - INFO - parrotllm.training - step 5960 | epoch 0 | loss 3.5888 | lr 1.50e-03 | grad 0.1766 +2026-04-10 05:15:42 - INFO - parrotllm.training - step 5970 | epoch 0 | loss 3.5092 | lr 1.50e-03 | grad 0.1944 +2026-04-10 05:15:45 - INFO - parrotllm.training - step 5980 | epoch 0 | loss 3.6554 | lr 1.50e-03 | grad 0.1970 +2026-04-10 05:15:48 - INFO - parrotllm.training - step 5990 | epoch 0 | loss 3.6505 | lr 1.50e-03 | grad 0.1983 +2026-04-10 05:15:51 - INFO - parrotllm.training - step 6000 | epoch 0 | loss 3.6082 | lr 1.50e-03 | grad 0.2053 +2026-04-10 05:15:51 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:15:51 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:15:54 - INFO - parrotllm.training - Train: loss=3.6082, ppl=36.90 +2026-04-10 05:15:54 - INFO - parrotllm.training - Val: loss=3.5300, ppl=34.12 +2026-04-10 05:15:54 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:15:55 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p5300_epoch_0000_step_0006000.pt +2026-04-10 05:15:56 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:15:59 - INFO - parrotllm.training - step 6010 | epoch 0 | loss 3.5907 | lr 1.50e-03 | grad 0.1964 +2026-04-10 05:16:02 - INFO - parrotllm.training - step 6020 | epoch 0 | loss 3.6362 | lr 1.50e-03 | grad 0.1672 +2026-04-10 05:16:05 - INFO - parrotllm.training - step 6030 | epoch 0 | loss 3.5577 | lr 1.50e-03 | grad 0.1992 +2026-04-10 05:16:08 - INFO - parrotllm.training - step 6040 | epoch 0 | loss 3.7013 | lr 1.50e-03 | grad 0.1899 +2026-04-10 05:16:11 - INFO - parrotllm.training - step 6050 | epoch 0 | loss 3.6568 | lr 1.50e-03 | grad 0.1955 +2026-04-10 05:16:14 - INFO - parrotllm.training - step 6060 | epoch 0 | loss 3.5904 | lr 1.50e-03 | grad 0.2242 +2026-04-10 05:16:18 - INFO - parrotllm.training - step 6070 | epoch 0 | loss 3.5301 | lr 1.50e-03 | grad 0.1723 +2026-04-10 05:16:21 - INFO - parrotllm.training - step 6080 | epoch 0 | loss 3.5867 | lr 1.50e-03 | grad 0.1855 +2026-04-10 05:16:24 - INFO - parrotllm.training - step 6090 | epoch 0 | loss 3.6379 | lr 1.50e-03 | grad 0.2031 +2026-04-10 05:16:27 - INFO - parrotllm.training - step 6100 | epoch 0 | loss 3.5949 | lr 1.50e-03 | grad 0.2055 +2026-04-10 05:16:30 - INFO - parrotllm.training - step 6110 | epoch 0 | loss 3.6043 | lr 1.50e-03 | grad 0.2121 +2026-04-10 05:16:33 - INFO - parrotllm.training - step 6120 | epoch 0 | loss 3.6341 | lr 1.50e-03 | grad 0.1841 +2026-04-10 05:16:36 - INFO - parrotllm.training - step 6130 | epoch 0 | loss 3.6955 | lr 1.50e-03 | grad 0.1716 +2026-04-10 05:16:39 - INFO - parrotllm.training - step 6140 | epoch 0 | loss 3.5858 | lr 1.50e-03 | grad 0.2020 +2026-04-10 05:16:42 - INFO - parrotllm.training - step 6150 | epoch 0 | loss 3.6020 | lr 1.50e-03 | grad 0.1861 +2026-04-10 05:16:45 - INFO - parrotllm.training - step 6160 | epoch 0 | loss 3.5677 | lr 1.50e-03 | grad 0.2132 +2026-04-10 05:16:48 - INFO - parrotllm.training - step 6170 | epoch 0 | loss 3.6673 | lr 1.50e-03 | grad 0.1873 +2026-04-10 05:16:51 - INFO - parrotllm.training - step 6180 | epoch 0 | loss 3.6316 | lr 1.49e-03 | grad 0.1930 +2026-04-10 05:16:54 - INFO - parrotllm.training - step 6190 | epoch 0 | loss 3.5914 | lr 1.49e-03 | grad 0.1871 +2026-04-10 05:16:57 - INFO - parrotllm.training - step 6200 | epoch 0 | loss 3.5360 | lr 1.49e-03 | grad 0.1906 +2026-04-10 05:17:00 - INFO - parrotllm.training - step 6210 | epoch 0 | loss 3.6413 | lr 1.49e-03 | grad 0.1886 +2026-04-10 05:17:03 - INFO - parrotllm.training - step 6220 | epoch 0 | loss 3.5546 | lr 1.49e-03 | grad 0.1733 +2026-04-10 05:17:06 - INFO - parrotllm.training - step 6230 | epoch 0 | loss 3.6094 | lr 1.49e-03 | grad 0.1907 +2026-04-10 05:17:09 - INFO - parrotllm.training - step 6240 | epoch 0 | loss 3.6769 | lr 1.49e-03 | grad 0.1909 +2026-04-10 05:17:12 - INFO - parrotllm.training - step 6250 | epoch 0 | loss 3.6342 | lr 1.49e-03 | grad 0.1885 +2026-04-10 05:17:15 - INFO - parrotllm.training - step 6260 | epoch 0 | loss 3.5598 | lr 1.49e-03 | grad 0.1834 +2026-04-10 05:17:18 - INFO - parrotllm.training - step 6270 | epoch 0 | loss 3.5895 | lr 1.49e-03 | grad 0.1930 +2026-04-10 05:17:21 - INFO - parrotllm.training - step 6280 | epoch 0 | loss 3.6295 | lr 1.49e-03 | grad 0.1877 +2026-04-10 05:17:25 - INFO - parrotllm.training - step 6290 | epoch 0 | loss 3.5677 | lr 1.49e-03 | grad 0.1731 +2026-04-10 05:17:28 - INFO - parrotllm.training - step 6300 | epoch 0 | loss 3.5792 | lr 1.49e-03 | grad 0.2162 +2026-04-10 05:17:31 - INFO - parrotllm.training - step 6310 | epoch 0 | loss 3.5748 | lr 1.49e-03 | grad 0.1747 +2026-04-10 05:17:34 - INFO - parrotllm.training - step 6320 | epoch 0 | loss 3.5851 | lr 1.49e-03 | grad 0.1736 +2026-04-10 05:17:37 - INFO - parrotllm.training - step 6330 | epoch 0 | loss 3.6620 | lr 1.49e-03 | grad 0.1782 +2026-04-10 05:17:40 - INFO - parrotllm.training - step 6340 | epoch 0 | loss 3.5654 | lr 1.49e-03 | grad 0.2049 +2026-04-10 05:17:43 - INFO - parrotllm.training - step 6350 | epoch 0 | loss 3.6600 | lr 1.49e-03 | grad 0.1926 +2026-04-10 05:17:46 - INFO - parrotllm.training - step 6360 | epoch 0 | loss 3.5341 | lr 1.49e-03 | grad 0.1741 +2026-04-10 05:17:48 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 6368/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 05:17:49 - INFO - parrotllm.training - step 6370 | epoch 0 | loss 3.6052 | lr 1.49e-03 | grad 0.1853 +2026-04-10 05:17:52 - INFO - parrotllm.training - step 6380 | epoch 0 | loss 3.6090 | lr 1.49e-03 | grad 0.1802 +2026-04-10 05:17:55 - INFO - parrotllm.training - step 6390 | epoch 0 | loss 3.5770 | lr 1.49e-03 | grad 0.2015 +2026-04-10 05:17:58 - INFO - parrotllm.training - step 6400 | epoch 0 | loss 3.6740 | lr 1.49e-03 | grad 0.2112 +2026-04-10 05:18:01 - INFO - parrotllm.training - step 6410 | epoch 0 | loss 3.5395 | lr 1.49e-03 | grad 0.1757 +2026-04-10 05:18:04 - INFO - parrotllm.training - step 6420 | epoch 0 | loss 3.6482 | lr 1.49e-03 | grad 0.1918 +2026-04-10 05:18:07 - INFO - parrotllm.training - step 6430 | epoch 0 | loss 3.5779 | lr 1.49e-03 | grad 0.1917 +2026-04-10 05:18:11 - INFO - parrotllm.training - step 6440 | epoch 0 | loss 3.5609 | lr 1.49e-03 | grad 0.2757 +2026-04-10 05:18:14 - INFO - parrotllm.training - step 6450 | epoch 0 | loss 3.5234 | lr 1.49e-03 | grad 0.2074 +2026-04-10 05:18:17 - INFO - parrotllm.training - step 6460 | epoch 0 | loss 3.6597 | lr 1.49e-03 | grad 0.1928 +2026-04-10 05:18:20 - INFO - parrotllm.training - step 6470 | epoch 0 | loss 3.5573 | lr 1.49e-03 | grad 0.1875 +2026-04-10 05:18:23 - INFO - parrotllm.training - step 6480 | epoch 0 | loss 3.5595 | lr 1.49e-03 | grad 0.1898 +2026-04-10 05:18:26 - INFO - parrotllm.training - step 6490 | epoch 0 | loss 3.6350 | lr 1.49e-03 | grad 0.1777 +2026-04-10 05:18:29 - INFO - parrotllm.training - step 6500 | epoch 0 | loss 3.5432 | lr 1.49e-03 | grad 0.1741 +2026-04-10 05:18:29 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:18:29 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:18:32 - INFO - parrotllm.training - Train: loss=3.5432, ppl=34.58 +2026-04-10 05:18:32 - INFO - parrotllm.training - Val: loss=3.5179, ppl=33.71 +2026-04-10 05:18:32 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:18:33 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p5179_epoch_0000_step_0006500.pt +2026-04-10 05:18:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:18:37 - INFO - parrotllm.training - step 6510 | epoch 0 | loss 3.6185 | lr 1.49e-03 | grad 0.1965 +2026-04-10 05:18:40 - INFO - parrotllm.training - step 6520 | epoch 0 | loss 3.6305 | lr 1.49e-03 | grad 0.1753 +2026-04-10 05:18:43 - INFO - parrotllm.training - step 6530 | epoch 0 | loss 3.6303 | lr 1.49e-03 | grad 0.1756 +2026-04-10 05:18:46 - INFO - parrotllm.training - step 6540 | epoch 0 | loss 3.5344 | lr 1.49e-03 | grad 0.1752 +2026-04-10 05:18:49 - INFO - parrotllm.training - step 6550 | epoch 0 | loss 3.5599 | lr 1.49e-03 | grad 0.1992 +2026-04-10 05:18:52 - INFO - parrotllm.training - step 6560 | epoch 0 | loss 3.6123 | lr 1.49e-03 | grad 0.2066 +2026-04-10 05:18:55 - INFO - parrotllm.training - step 6570 | epoch 0 | loss 3.4927 | lr 1.49e-03 | grad 0.1768 +2026-04-10 05:18:58 - INFO - parrotllm.training - step 6580 | epoch 0 | loss 3.5086 | lr 1.49e-03 | grad 0.1889 +2026-04-10 05:19:01 - INFO - parrotllm.training - step 6590 | epoch 0 | loss 3.4158 | lr 1.49e-03 | grad 0.1834 +2026-04-10 05:19:04 - INFO - parrotllm.training - step 6600 | epoch 0 | loss 3.5952 | lr 1.49e-03 | grad 0.2251 +2026-04-10 05:19:07 - INFO - parrotllm.training - step 6610 | epoch 0 | loss 3.5166 | lr 1.49e-03 | grad 0.1817 +2026-04-10 05:19:10 - INFO - parrotllm.training - step 6620 | epoch 0 | loss 3.6246 | lr 1.49e-03 | grad 0.1769 +2026-04-10 05:19:13 - INFO - parrotllm.training - step 6630 | epoch 0 | loss 3.6390 | lr 1.49e-03 | grad 0.1917 +2026-04-10 05:19:16 - INFO - parrotllm.training - step 6640 | epoch 0 | loss 3.5865 | lr 1.49e-03 | grad 0.1793 +2026-04-10 05:19:19 - INFO - parrotllm.training - step 6650 | epoch 0 | loss 3.6824 | lr 1.49e-03 | grad 0.1874 +2026-04-10 05:19:22 - INFO - parrotllm.training - step 6660 | epoch 0 | loss 3.5683 | lr 1.49e-03 | grad 0.1797 +2026-04-10 05:19:25 - INFO - parrotllm.training - step 6670 | epoch 0 | loss 3.6385 | lr 1.49e-03 | grad 0.2068 +2026-04-10 05:19:28 - INFO - parrotllm.training - step 6680 | epoch 0 | loss 3.6019 | lr 1.49e-03 | grad 0.1734 +2026-04-10 05:19:31 - INFO - parrotllm.training - step 6690 | epoch 0 | loss 3.3894 | lr 1.49e-03 | grad 0.1888 +2026-04-10 05:19:35 - INFO - parrotllm.training - step 6700 | epoch 0 | loss 3.6268 | lr 1.49e-03 | grad 0.1942 +2026-04-10 05:19:38 - INFO - parrotllm.training - step 6710 | epoch 0 | loss 3.6278 | lr 1.49e-03 | grad 0.1886 +2026-04-10 05:19:41 - INFO - parrotllm.training - step 6720 | epoch 0 | loss 3.6478 | lr 1.49e-03 | grad 0.2291 +2026-04-10 05:19:44 - INFO - parrotllm.training - step 6730 | epoch 0 | loss 3.6348 | lr 1.49e-03 | grad 0.2020 +2026-04-10 05:19:47 - INFO - parrotllm.training - step 6740 | epoch 0 | loss 3.5569 | lr 1.49e-03 | grad 0.1990 +2026-04-10 05:19:50 - INFO - parrotllm.training - step 6750 | epoch 0 | loss 3.6125 | lr 1.49e-03 | grad 0.1703 +2026-04-10 05:19:53 - INFO - parrotllm.training - step 6760 | epoch 0 | loss 3.6621 | lr 1.49e-03 | grad 0.1906 +2026-04-10 05:19:56 - INFO - parrotllm.training - step 6770 | epoch 0 | loss 3.6040 | lr 1.49e-03 | grad 0.2089 +2026-04-10 05:19:59 - INFO - parrotllm.training - step 6780 | epoch 0 | loss 3.5892 | lr 1.49e-03 | grad 0.1710 +2026-04-10 05:20:02 - INFO - parrotllm.training - step 6790 | epoch 0 | loss 3.5998 | lr 1.49e-03 | grad 0.1832 +2026-04-10 05:20:05 - INFO - parrotllm.training - step 6800 | epoch 0 | loss 3.5787 | lr 1.49e-03 | grad 0.2074 +2026-04-10 05:20:08 - INFO - parrotllm.training - step 6810 | epoch 0 | loss 3.5107 | lr 1.49e-03 | grad 0.1890 +2026-04-10 05:20:11 - INFO - parrotllm.training - step 6820 | epoch 0 | loss 3.6027 | lr 1.49e-03 | grad 0.1968 +2026-04-10 05:20:14 - INFO - parrotllm.training - step 6830 | epoch 0 | loss 3.6137 | lr 1.49e-03 | grad 0.2103 +2026-04-10 05:20:17 - INFO - parrotllm.training - step 6840 | epoch 0 | loss 3.5445 | lr 1.49e-03 | grad 0.1759 +2026-04-10 05:20:20 - INFO - parrotllm.training - step 6850 | epoch 0 | loss 3.4700 | lr 1.49e-03 | grad 0.1965 +2026-04-10 05:20:23 - INFO - parrotllm.training - step 6860 | epoch 0 | loss 3.6237 | lr 1.49e-03 | grad 0.1960 +2026-04-10 05:20:26 - INFO - parrotllm.training - step 6870 | epoch 0 | loss 3.6263 | lr 1.49e-03 | grad 0.1934 +2026-04-10 05:20:29 - INFO - parrotllm.training - step 6880 | epoch 0 | loss 3.5383 | lr 1.49e-03 | grad 0.2057 +2026-04-10 05:20:32 - INFO - parrotllm.training - step 6890 | epoch 0 | loss 3.6392 | lr 1.49e-03 | grad 0.1836 +2026-04-10 05:20:35 - INFO - parrotllm.training - step 6900 | epoch 0 | loss 3.5710 | lr 1.49e-03 | grad 0.1891 +2026-04-10 05:20:38 - INFO - parrotllm.training - step 6910 | epoch 0 | loss 3.5704 | lr 1.49e-03 | grad 0.1744 +2026-04-10 05:20:41 - INFO - parrotllm.training - step 6920 | epoch 0 | loss 3.5232 | lr 1.49e-03 | grad 0.1632 +2026-04-10 05:20:45 - INFO - parrotllm.training - step 6930 | epoch 0 | loss 3.5099 | lr 1.49e-03 | grad 0.1866 +2026-04-10 05:20:48 - INFO - parrotllm.training - step 6940 | epoch 0 | loss 3.5348 | lr 1.49e-03 | grad 0.2040 +2026-04-10 05:20:51 - INFO - parrotllm.training - step 6950 | epoch 0 | loss 3.5288 | lr 1.49e-03 | grad 0.1809 +2026-04-10 05:20:54 - INFO - parrotllm.training - step 6960 | epoch 0 | loss 3.5211 | lr 1.49e-03 | grad 0.1853 +2026-04-10 05:20:57 - INFO - parrotllm.training - step 6970 | epoch 0 | loss 3.5207 | lr 1.49e-03 | grad 0.1962 +2026-04-10 05:21:00 - INFO - parrotllm.training - step 6980 | epoch 0 | loss 3.5715 | lr 1.49e-03 | grad 0.1665 +2026-04-10 05:21:03 - INFO - parrotllm.training - step 6990 | epoch 0 | loss 3.6126 | lr 1.49e-03 | grad 0.1902 +2026-04-10 05:21:06 - INFO - parrotllm.training - step 7000 | epoch 0 | loss 3.6081 | lr 1.49e-03 | grad 0.2165 +2026-04-10 05:21:06 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:21:06 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:21:09 - INFO - parrotllm.training - Train: loss=3.6081, ppl=36.90 +2026-04-10 05:21:09 - INFO - parrotllm.training - Val: loss=3.5104, ppl=33.46 +2026-04-10 05:21:09 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:21:10 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p5104_epoch_0000_step_0007000.pt +2026-04-10 05:21:11 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:21:14 - INFO - parrotllm.training - step 7010 | epoch 0 | loss 3.6342 | lr 1.49e-03 | grad 0.1839 +2026-04-10 05:21:17 - INFO - parrotllm.training - step 7020 | epoch 0 | loss 3.6078 | lr 1.49e-03 | grad 0.1965 +2026-04-10 05:21:20 - INFO - parrotllm.training - step 7030 | epoch 0 | loss 3.5922 | lr 1.49e-03 | grad 0.1802 +2026-04-10 05:21:23 - INFO - parrotllm.training - step 7040 | epoch 0 | loss 3.5910 | lr 1.49e-03 | grad 0.1832 +2026-04-10 05:21:26 - INFO - parrotllm.training - step 7050 | epoch 0 | loss 3.6590 | lr 1.49e-03 | grad 0.2040 +2026-04-10 05:21:29 - INFO - parrotllm.training - step 7060 | epoch 0 | loss 3.6595 | lr 1.49e-03 | grad 0.1915 +2026-04-10 05:21:32 - INFO - parrotllm.training - step 7070 | epoch 0 | loss 3.5309 | lr 1.49e-03 | grad 0.1996 +2026-04-10 05:21:35 - INFO - parrotllm.training - step 7080 | epoch 0 | loss 3.5952 | lr 1.49e-03 | grad 0.1800 +2026-04-10 05:21:38 - INFO - parrotllm.training - step 7090 | epoch 0 | loss 3.7118 | lr 1.49e-03 | grad 0.1906 +2026-04-10 05:21:41 - INFO - parrotllm.training - step 7100 | epoch 0 | loss 3.5790 | lr 1.49e-03 | grad 0.1882 +2026-04-10 05:21:44 - INFO - parrotllm.training - step 7110 | epoch 0 | loss 3.6024 | lr 1.49e-03 | grad 0.1684 +2026-04-10 05:21:47 - INFO - parrotllm.training - step 7120 | epoch 0 | loss 3.6199 | lr 1.49e-03 | grad 0.2074 +2026-04-10 05:21:50 - INFO - parrotllm.training - step 7130 | epoch 0 | loss 3.6523 | lr 1.49e-03 | grad 0.1656 +2026-04-10 05:21:53 - INFO - parrotllm.training - step 7140 | epoch 0 | loss 3.6118 | lr 1.49e-03 | grad 0.1965 +2026-04-10 05:21:56 - INFO - parrotllm.training - step 7150 | epoch 0 | loss 3.5251 | lr 1.49e-03 | grad 0.2054 +2026-04-10 05:21:59 - INFO - parrotllm.training - step 7160 | epoch 0 | loss 3.5846 | lr 1.49e-03 | grad 0.1832 +2026-04-10 05:22:02 - INFO - parrotllm.training - step 7170 | epoch 0 | loss 3.5215 | lr 1.49e-03 | grad 0.2188 +2026-04-10 05:22:06 - INFO - parrotllm.training - step 7180 | epoch 0 | loss 3.5642 | lr 1.49e-03 | grad 0.2181 +2026-04-10 05:22:09 - INFO - parrotllm.training - step 7190 | epoch 0 | loss 3.5696 | lr 1.49e-03 | grad 0.1854 +2026-04-10 05:22:12 - INFO - parrotllm.training - step 7200 | epoch 0 | loss 3.6083 | lr 1.49e-03 | grad 0.2055 +2026-04-10 05:22:15 - INFO - parrotllm.training - step 7210 | epoch 0 | loss 3.5863 | lr 1.49e-03 | grad 0.1868 +2026-04-10 05:22:18 - INFO - parrotllm.training - step 7220 | epoch 0 | loss 3.5889 | lr 1.49e-03 | grad 0.1871 +2026-04-10 05:22:21 - INFO - parrotllm.training - step 7230 | epoch 0 | loss 3.6310 | lr 1.49e-03 | grad 0.1794 +2026-04-10 05:22:24 - INFO - parrotllm.training - step 7240 | epoch 0 | loss 3.4816 | lr 1.49e-03 | grad 0.2139 +2026-04-10 05:22:27 - INFO - parrotllm.training - step 7250 | epoch 0 | loss 3.4848 | lr 1.49e-03 | grad 0.1808 +2026-04-10 05:22:30 - INFO - parrotllm.training - step 7260 | epoch 0 | loss 3.6065 | lr 1.49e-03 | grad 0.1980 +2026-04-10 05:22:33 - INFO - parrotllm.training - step 7270 | epoch 0 | loss 3.6021 | lr 1.49e-03 | grad 0.1839 +2026-04-10 05:22:36 - INFO - parrotllm.training - step 7280 | epoch 0 | loss 3.4878 | lr 1.49e-03 | grad 0.1839 +2026-04-10 05:22:39 - INFO - parrotllm.training - step 7290 | epoch 0 | loss 3.6442 | lr 1.49e-03 | grad 0.1787 +2026-04-10 05:22:42 - INFO - parrotllm.training - step 7300 | epoch 0 | loss 3.5439 | lr 1.49e-03 | grad 0.1783 +2026-04-10 05:22:45 - INFO - parrotllm.training - step 7310 | epoch 0 | loss 3.6452 | lr 1.49e-03 | grad 0.2171 +2026-04-10 05:22:48 - INFO - parrotllm.training - step 7320 | epoch 0 | loss 3.5059 | lr 1.49e-03 | grad 0.2058 +2026-04-10 05:22:51 - INFO - parrotllm.training - step 7330 | epoch 0 | loss 3.4854 | lr 1.49e-03 | grad 0.1895 +2026-04-10 05:22:54 - INFO - parrotllm.training - step 7340 | epoch 0 | loss 3.5192 | lr 1.49e-03 | grad 0.1900 +2026-04-10 05:22:57 - INFO - parrotllm.training - step 7350 | epoch 0 | loss 3.5352 | lr 1.49e-03 | grad 0.1857 +2026-04-10 05:23:00 - INFO - parrotllm.training - step 7360 | epoch 0 | loss 3.5905 | lr 1.49e-03 | grad 0.1870 +2026-04-10 05:23:03 - INFO - parrotllm.training - step 7370 | epoch 0 | loss 3.6345 | lr 1.49e-03 | grad 0.1763 +2026-04-10 05:23:06 - INFO - parrotllm.training - step 7380 | epoch 0 | loss 3.5476 | lr 1.49e-03 | grad 0.2005 +2026-04-10 05:23:09 - INFO - parrotllm.training - step 7390 | epoch 0 | loss 3.5972 | lr 1.49e-03 | grad 0.2212 +2026-04-10 05:23:12 - INFO - parrotllm.training - step 7400 | epoch 0 | loss 3.4840 | lr 1.49e-03 | grad 0.2303 +2026-04-10 05:23:16 - INFO - parrotllm.training - step 7410 | epoch 0 | loss 3.5179 | lr 1.49e-03 | grad 0.1762 +2026-04-10 05:23:19 - INFO - parrotllm.training - step 7420 | epoch 0 | loss 3.5728 | lr 1.49e-03 | grad 0.1715 +2026-04-10 05:23:22 - INFO - parrotllm.training - step 7430 | epoch 0 | loss 3.6148 | lr 1.49e-03 | grad 0.2000 +2026-04-10 05:23:25 - INFO - parrotllm.training - step 7440 | epoch 0 | loss 3.5568 | lr 1.49e-03 | grad 0.1959 +2026-04-10 05:23:28 - INFO - parrotllm.training - step 7450 | epoch 0 | loss 3.5978 | lr 1.49e-03 | grad 0.1910 +2026-04-10 05:23:31 - INFO - parrotllm.training - step 7460 | epoch 0 | loss 3.5993 | lr 1.49e-03 | grad 0.1733 +2026-04-10 05:23:34 - INFO - parrotllm.training - step 7470 | epoch 0 | loss 3.5732 | lr 1.49e-03 | grad 0.1750 +2026-04-10 05:23:37 - INFO - parrotllm.training - step 7480 | epoch 0 | loss 3.5580 | lr 1.49e-03 | grad 0.1917 +2026-04-10 05:23:40 - INFO - parrotllm.training - step 7490 | epoch 0 | loss 3.6117 | lr 1.49e-03 | grad 0.2021 +2026-04-10 05:23:43 - INFO - parrotllm.training - step 7500 | epoch 0 | loss 3.4316 | lr 1.49e-03 | grad 0.2350 +2026-04-10 05:23:43 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:23:43 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:23:46 - INFO - parrotllm.training - Train: loss=3.4316, ppl=30.93 +2026-04-10 05:23:46 - INFO - parrotllm.training - Val: loss=3.5052, ppl=33.29 +2026-04-10 05:23:46 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:23:47 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p5052_epoch_0000_step_0007500.pt +2026-04-10 05:23:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:23:49 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0007500.pt +2026-04-10 05:23:53 - INFO - parrotllm.training - step 7510 | epoch 0 | loss 3.6015 | lr 1.49e-03 | grad 0.2086 +2026-04-10 05:23:56 - INFO - parrotllm.training - step 7520 | epoch 0 | loss 3.4981 | lr 1.49e-03 | grad 0.1622 +2026-04-10 05:23:59 - INFO - parrotllm.training - step 7530 | epoch 0 | loss 3.5750 | lr 1.49e-03 | grad 0.1948 +2026-04-10 05:24:02 - INFO - parrotllm.training - step 7540 | epoch 0 | loss 3.5500 | lr 1.49e-03 | grad 0.1856 +2026-04-10 05:24:05 - INFO - parrotllm.training - step 7550 | epoch 0 | loss 3.5956 | lr 1.49e-03 | grad 0.1719 +2026-04-10 05:24:08 - INFO - parrotllm.training - step 7560 | epoch 0 | loss 3.5469 | lr 1.49e-03 | grad 0.2042 +2026-04-10 05:24:11 - INFO - parrotllm.training - step 7570 | epoch 0 | loss 3.5514 | lr 1.49e-03 | grad 0.1790 +2026-04-10 05:24:14 - INFO - parrotllm.training - step 7580 | epoch 0 | loss 3.4860 | lr 1.49e-03 | grad 0.1985 +2026-04-10 05:24:17 - INFO - parrotllm.training - step 7590 | epoch 0 | loss 3.4163 | lr 1.49e-03 | grad 0.1754 +2026-04-10 05:24:21 - INFO - parrotllm.training - step 7600 | epoch 0 | loss 3.6092 | lr 1.49e-03 | grad 0.1718 +2026-04-10 05:24:24 - INFO - parrotllm.training - step 7610 | epoch 0 | loss 3.7136 | lr 1.49e-03 | grad 0.1958 +2026-04-10 05:24:27 - INFO - parrotllm.training - step 7620 | epoch 0 | loss 3.5393 | lr 1.49e-03 | grad 0.1964 +2026-04-10 05:24:30 - INFO - parrotllm.training - step 7630 | epoch 0 | loss 3.5991 | lr 1.49e-03 | grad 0.1737 +2026-04-10 05:24:33 - INFO - parrotllm.training - step 7640 | epoch 0 | loss 3.4950 | lr 1.49e-03 | grad 0.2007 +2026-04-10 05:24:36 - INFO - parrotllm.training - step 7650 | epoch 0 | loss 3.5520 | lr 1.49e-03 | grad 0.2046 +2026-04-10 05:24:39 - INFO - parrotllm.training - step 7660 | epoch 0 | loss 3.5638 | lr 1.49e-03 | grad 0.2161 +2026-04-10 05:24:42 - INFO - parrotllm.training - step 7670 | epoch 0 | loss 3.5724 | lr 1.49e-03 | grad 0.1745 +2026-04-10 05:24:45 - INFO - parrotllm.training - step 7680 | epoch 0 | loss 3.6017 | lr 1.49e-03 | grad 0.1816 +2026-04-10 05:24:48 - INFO - parrotllm.training - step 7690 | epoch 0 | loss 3.5510 | lr 1.49e-03 | grad 0.1873 +2026-04-10 05:24:51 - INFO - parrotllm.training - step 7700 | epoch 0 | loss 3.5986 | lr 1.49e-03 | grad 0.2015 +2026-04-10 05:24:54 - INFO - parrotllm.training - step 7710 | epoch 0 | loss 3.6141 | lr 1.49e-03 | grad 0.1814 +2026-04-10 05:24:57 - INFO - parrotllm.training - step 7720 | epoch 0 | loss 3.5518 | lr 1.49e-03 | grad 0.2031 +2026-04-10 05:25:00 - INFO - parrotllm.training - step 7730 | epoch 0 | loss 3.5360 | lr 1.49e-03 | grad 0.1857 +2026-04-10 05:25:03 - INFO - parrotllm.training - step 7740 | epoch 0 | loss 3.5311 | lr 1.49e-03 | grad 0.2009 +2026-04-10 05:25:06 - INFO - parrotllm.training - step 7750 | epoch 0 | loss 3.5372 | lr 1.49e-03 | grad 0.2178 +2026-04-10 05:25:09 - INFO - parrotllm.training - step 7760 | epoch 0 | loss 3.5236 | lr 1.49e-03 | grad 0.1826 +2026-04-10 05:25:12 - INFO - parrotllm.training - step 7770 | epoch 0 | loss 3.6268 | lr 1.49e-03 | grad 0.1880 +2026-04-10 05:25:15 - INFO - parrotllm.training - step 7780 | epoch 0 | loss 3.6337 | lr 1.49e-03 | grad 0.1974 +2026-04-10 05:25:18 - INFO - parrotllm.training - step 7790 | epoch 0 | loss 3.7078 | lr 1.49e-03 | grad 0.1893 +2026-04-10 05:25:22 - INFO - parrotllm.training - step 7800 | epoch 0 | loss 3.5739 | lr 1.49e-03 | grad 0.1921 +2026-04-10 05:25:25 - INFO - parrotllm.training - step 7810 | epoch 0 | loss 3.6399 | lr 1.49e-03 | grad 0.1851 +2026-04-10 05:25:28 - INFO - parrotllm.training - step 7820 | epoch 0 | loss 3.5075 | lr 1.49e-03 | grad 0.1759 +2026-04-10 05:25:31 - INFO - parrotllm.training - step 7830 | epoch 0 | loss 3.5814 | lr 1.49e-03 | grad 0.1792 +2026-04-10 05:25:34 - INFO - parrotllm.training - step 7840 | epoch 0 | loss 3.5979 | lr 1.49e-03 | grad 0.1845 +2026-04-10 05:25:37 - INFO - parrotllm.training - step 7850 | epoch 0 | loss 3.5785 | lr 1.49e-03 | grad 0.1737 +2026-04-10 05:25:40 - INFO - parrotllm.training - step 7860 | epoch 0 | loss 3.5461 | lr 1.49e-03 | grad 0.1726 +2026-04-10 05:25:43 - INFO - parrotllm.training - step 7870 | epoch 0 | loss 3.6482 | lr 1.49e-03 | grad 0.1932 +2026-04-10 05:25:46 - INFO - parrotllm.training - step 7880 | epoch 0 | loss 3.5182 | lr 1.49e-03 | grad 0.1973 +2026-04-10 05:25:49 - INFO - parrotllm.training - step 7890 | epoch 0 | loss 3.5292 | lr 1.49e-03 | grad 0.1893 +2026-04-10 05:25:52 - INFO - parrotllm.training - step 7900 | epoch 0 | loss 3.6084 | lr 1.49e-03 | grad 0.2096 +2026-04-10 05:25:55 - INFO - parrotllm.training - step 7910 | epoch 0 | loss 3.4770 | lr 1.49e-03 | grad 0.2270 +2026-04-10 05:25:58 - INFO - parrotllm.training - step 7920 | epoch 0 | loss 3.5480 | lr 1.49e-03 | grad 0.1758 +2026-04-10 05:26:01 - INFO - parrotllm.training - step 7930 | epoch 0 | loss 3.5716 | lr 1.49e-03 | grad 0.1795 +2026-04-10 05:26:04 - INFO - parrotllm.training - step 7940 | epoch 0 | loss 3.5103 | lr 1.49e-03 | grad 0.1887 +2026-04-10 05:26:07 - INFO - parrotllm.training - step 7950 | epoch 0 | loss 3.5016 | lr 1.49e-03 | grad 0.1977 +2026-04-10 05:26:10 - INFO - parrotllm.training - step 7960 | epoch 0 | loss 3.6154 | lr 1.49e-03 | grad 0.2503 +2026-04-10 05:26:13 - INFO - parrotllm.training - step 7970 | epoch 0 | loss 3.6379 | lr 1.49e-03 | grad 0.1860 +2026-04-10 05:26:16 - INFO - parrotllm.training - step 7980 | epoch 0 | loss 3.6816 | lr 1.49e-03 | grad 0.2012 +2026-04-10 05:26:19 - INFO - parrotllm.training - step 7990 | epoch 0 | loss 3.4775 | lr 1.49e-03 | grad 0.2134 +2026-04-10 05:26:22 - INFO - parrotllm.training - step 8000 | epoch 0 | loss 3.5934 | lr 1.49e-03 | grad 0.1884 +2026-04-10 05:26:22 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:26:22 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:26:26 - INFO - parrotllm.training - Train: loss=3.5934, ppl=36.36 +2026-04-10 05:26:26 - INFO - parrotllm.training - Val: loss=3.4969, ppl=33.01 +2026-04-10 05:26:26 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:26:27 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4969_epoch_0000_step_0008000.pt +2026-04-10 05:26:28 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:26:31 - INFO - parrotllm.training - step 8010 | epoch 0 | loss 3.5290 | lr 1.49e-03 | grad 0.1867 +2026-04-10 05:26:34 - INFO - parrotllm.training - step 8020 | epoch 0 | loss 3.5826 | lr 1.49e-03 | grad 0.1833 +2026-04-10 05:26:37 - INFO - parrotllm.training - step 8030 | epoch 0 | loss 3.6410 | lr 1.49e-03 | grad 0.1834 +2026-04-10 05:26:40 - INFO - parrotllm.training - step 8040 | epoch 0 | loss 3.5567 | lr 1.49e-03 | grad 0.1939 +2026-04-10 05:26:43 - INFO - parrotllm.training - step 8050 | epoch 0 | loss 3.6098 | lr 1.49e-03 | grad 0.1765 +2026-04-10 05:26:46 - INFO - parrotllm.training - step 8060 | epoch 0 | loss 3.5712 | lr 1.49e-03 | grad 0.2039 +2026-04-10 05:26:49 - INFO - parrotllm.training - step 8070 | epoch 0 | loss 3.5992 | lr 1.49e-03 | grad 0.1873 +2026-04-10 05:26:52 - INFO - parrotllm.training - step 8080 | epoch 0 | loss 3.5085 | lr 1.49e-03 | grad 0.1858 +2026-04-10 05:26:55 - INFO - parrotllm.training - step 8090 | epoch 0 | loss 3.6058 | lr 1.49e-03 | grad 0.1862 +2026-04-10 05:26:58 - INFO - parrotllm.training - step 8100 | epoch 0 | loss 3.5947 | lr 1.49e-03 | grad 0.1789 +2026-04-10 05:27:01 - INFO - parrotllm.training - step 8110 | epoch 0 | loss 3.5902 | lr 1.49e-03 | grad 0.2127 +2026-04-10 05:27:04 - INFO - parrotllm.training - step 8120 | epoch 0 | loss 3.4375 | lr 1.49e-03 | grad 0.1890 +2026-04-10 05:27:07 - INFO - parrotllm.training - step 8130 | epoch 0 | loss 3.5504 | lr 1.49e-03 | grad 0.1800 +2026-04-10 05:27:10 - INFO - parrotllm.training - step 8140 | epoch 0 | loss 3.5418 | lr 1.49e-03 | grad 0.1994 +2026-04-10 05:27:13 - INFO - parrotllm.training - step 8150 | epoch 0 | loss 3.5659 | lr 1.49e-03 | grad 0.1779 +2026-04-10 05:27:16 - INFO - parrotllm.training - step 8160 | epoch 0 | loss 3.5018 | lr 1.49e-03 | grad 0.1997 +2026-04-10 05:27:19 - INFO - parrotllm.training - step 8170 | epoch 0 | loss 3.6307 | lr 1.49e-03 | grad 0.1917 +2026-04-10 05:27:22 - INFO - parrotllm.training - step 8180 | epoch 0 | loss 3.5733 | lr 1.49e-03 | grad 0.1904 +2026-04-10 05:27:25 - INFO - parrotllm.training - step 8190 | epoch 0 | loss 3.5624 | lr 1.49e-03 | grad 0.1851 +2026-04-10 05:27:28 - INFO - parrotllm.training - step 8200 | epoch 0 | loss 3.4996 | lr 1.49e-03 | grad 0.1855 +2026-04-10 05:27:32 - INFO - parrotllm.training - step 8210 | epoch 0 | loss 3.5709 | lr 1.49e-03 | grad 0.1801 +2026-04-10 05:27:35 - INFO - parrotllm.training - step 8220 | epoch 0 | loss 3.6635 | lr 1.49e-03 | grad 0.1841 +2026-04-10 05:27:38 - INFO - parrotllm.training - step 8230 | epoch 0 | loss 3.5482 | lr 1.49e-03 | grad 0.2220 +2026-04-10 05:27:41 - INFO - parrotllm.training - step 8240 | epoch 0 | loss 3.6638 | lr 1.49e-03 | grad 0.1944 +2026-04-10 05:27:44 - INFO - parrotllm.training - step 8250 | epoch 0 | loss 3.4983 | lr 1.49e-03 | grad 0.1870 +2026-04-10 05:27:47 - INFO - parrotllm.training - step 8260 | epoch 0 | loss 3.6171 | lr 1.49e-03 | grad 0.1873 +2026-04-10 05:27:50 - INFO - parrotllm.training - step 8270 | epoch 0 | loss 3.4502 | lr 1.49e-03 | grad 0.1723 +2026-04-10 05:27:53 - INFO - parrotllm.training - step 8280 | epoch 0 | loss 3.6017 | lr 1.49e-03 | grad 0.2077 +2026-04-10 05:27:56 - INFO - parrotllm.training - step 8290 | epoch 0 | loss 3.5288 | lr 1.49e-03 | grad 0.1701 +2026-04-10 05:27:59 - INFO - parrotllm.training - step 8300 | epoch 0 | loss 3.5885 | lr 1.49e-03 | grad 0.1902 +2026-04-10 05:28:02 - INFO - parrotllm.training - step 8310 | epoch 0 | loss 3.5208 | lr 1.49e-03 | grad 0.1832 +2026-04-10 05:28:05 - INFO - parrotllm.training - step 8320 | epoch 0 | loss 3.4973 | lr 1.49e-03 | grad 0.1807 +2026-04-10 05:28:08 - INFO - parrotllm.training - step 8330 | epoch 0 | loss 3.5432 | lr 1.49e-03 | grad 0.1850 +2026-04-10 05:28:11 - INFO - parrotllm.training - step 8340 | epoch 0 | loss 3.5335 | lr 1.49e-03 | grad 0.2056 +2026-04-10 05:28:14 - INFO - parrotllm.training - step 8350 | epoch 0 | loss 3.5471 | lr 1.49e-03 | grad 0.1917 +2026-04-10 05:28:17 - INFO - parrotllm.training - step 8360 | epoch 0 | loss 3.6595 | lr 1.49e-03 | grad 0.1841 +2026-04-10 05:28:20 - INFO - parrotllm.training - step 8370 | epoch 0 | loss 3.5613 | lr 1.49e-03 | grad 0.1968 +2026-04-10 05:28:23 - INFO - parrotllm.training - step 8380 | epoch 0 | loss 3.5877 | lr 1.49e-03 | grad 0.2066 +2026-04-10 05:28:26 - INFO - parrotllm.training - step 8390 | epoch 0 | loss 3.6120 | lr 1.49e-03 | grad 0.2087 +2026-04-10 05:28:29 - INFO - parrotllm.training - step 8400 | epoch 0 | loss 3.5558 | lr 1.49e-03 | grad 0.1760 +2026-04-10 05:28:32 - INFO - parrotllm.training - step 8410 | epoch 0 | loss 3.4572 | lr 1.49e-03 | grad 0.1941 +2026-04-10 05:28:35 - INFO - parrotllm.training - step 8420 | epoch 0 | loss 3.5669 | lr 1.49e-03 | grad 0.2030 +2026-04-10 05:28:39 - INFO - parrotllm.training - step 8430 | epoch 0 | loss 3.6698 | lr 1.49e-03 | grad 0.2017 +2026-04-10 05:28:42 - INFO - parrotllm.training - step 8440 | epoch 0 | loss 3.5238 | lr 1.49e-03 | grad 0.2130 +2026-04-10 05:28:45 - INFO - parrotllm.training - step 8450 | epoch 0 | loss 3.5300 | lr 1.49e-03 | grad 0.1998 +2026-04-10 05:28:48 - INFO - parrotllm.training - step 8460 | epoch 0 | loss 3.6287 | lr 1.49e-03 | grad 0.2076 +2026-04-10 05:28:51 - INFO - parrotllm.training - step 8470 | epoch 0 | loss 3.5053 | lr 1.49e-03 | grad 0.1863 +2026-04-10 05:28:54 - INFO - parrotllm.training - step 8480 | epoch 0 | loss 3.5082 | lr 1.49e-03 | grad 0.2116 +2026-04-10 05:28:57 - INFO - parrotllm.training - step 8490 | epoch 0 | loss 3.5885 | lr 1.49e-03 | grad 0.2317 +2026-04-10 05:29:00 - INFO - parrotllm.training - step 8500 | epoch 0 | loss 3.5015 | lr 1.49e-03 | grad 0.1956 +2026-04-10 05:29:00 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:29:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:29:03 - INFO - parrotllm.training - Train: loss=3.5015, ppl=33.17 +2026-04-10 05:29:03 - INFO - parrotllm.training - Val: loss=3.4916, ppl=32.84 +2026-04-10 05:29:03 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:29:04 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4916_epoch_0000_step_0008500.pt +2026-04-10 05:29:05 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:29:08 - INFO - parrotllm.training - step 8510 | epoch 0 | loss 3.6108 | lr 1.49e-03 | grad 0.1896 +2026-04-10 05:29:11 - INFO - parrotllm.training - step 8520 | epoch 0 | loss 3.4692 | lr 1.49e-03 | grad 0.1881 +2026-04-10 05:29:14 - INFO - parrotllm.training - step 8530 | epoch 0 | loss 3.5451 | lr 1.49e-03 | grad 0.1723 +2026-04-10 05:29:17 - INFO - parrotllm.training - step 8540 | epoch 0 | loss 3.5169 | lr 1.49e-03 | grad 0.1846 +2026-04-10 05:29:20 - INFO - parrotllm.training - step 8550 | epoch 0 | loss 3.5240 | lr 1.49e-03 | grad 0.1713 +2026-04-10 05:29:23 - INFO - parrotllm.training - step 8560 | epoch 0 | loss 3.4917 | lr 1.49e-03 | grad 0.1959 +2026-04-10 05:29:26 - INFO - parrotllm.training - step 8570 | epoch 0 | loss 3.5180 | lr 1.49e-03 | grad 0.2015 +2026-04-10 05:29:29 - INFO - parrotllm.training - step 8580 | epoch 0 | loss 3.5665 | lr 1.49e-03 | grad 0.2077 +2026-04-10 05:29:32 - INFO - parrotllm.training - step 8590 | epoch 0 | loss 3.5798 | lr 1.49e-03 | grad 0.1836 +2026-04-10 05:29:35 - INFO - parrotllm.training - step 8600 | epoch 0 | loss 3.5530 | lr 1.49e-03 | grad 0.1762 +2026-04-10 05:29:38 - INFO - parrotllm.training - step 8610 | epoch 0 | loss 3.5283 | lr 1.49e-03 | grad 0.1815 +2026-04-10 05:29:41 - INFO - parrotllm.training - step 8620 | epoch 0 | loss 3.5262 | lr 1.49e-03 | grad 0.2117 +2026-04-10 05:29:44 - INFO - parrotllm.training - step 8630 | epoch 0 | loss 3.5494 | lr 1.49e-03 | grad 0.2137 +2026-04-10 05:29:47 - INFO - parrotllm.training - step 8640 | epoch 0 | loss 3.4642 | lr 1.49e-03 | grad 0.1844 +2026-04-10 05:29:50 - INFO - parrotllm.training - step 8650 | epoch 0 | loss 3.3881 | lr 1.49e-03 | grad 0.1926 +2026-04-10 05:29:53 - INFO - parrotllm.training - step 8660 | epoch 0 | loss 3.5076 | lr 1.49e-03 | grad 0.1773 +2026-04-10 05:29:56 - INFO - parrotllm.training - step 8670 | epoch 0 | loss 3.4596 | lr 1.49e-03 | grad 0.1852 +2026-04-10 05:30:00 - INFO - parrotllm.training - step 8680 | epoch 0 | loss 3.6825 | lr 1.49e-03 | grad 0.1962 +2026-04-10 05:30:03 - INFO - parrotllm.training - step 8690 | epoch 0 | loss 3.5471 | lr 1.49e-03 | grad 0.1888 +2026-04-10 05:30:06 - INFO - parrotllm.training - step 8700 | epoch 0 | loss 3.5648 | lr 1.49e-03 | grad 0.1965 +2026-04-10 05:30:09 - INFO - parrotllm.training - step 8710 | epoch 0 | loss 3.5142 | lr 1.49e-03 | grad 0.2018 +2026-04-10 05:30:12 - INFO - parrotllm.training - step 8720 | epoch 0 | loss 3.5238 | lr 1.49e-03 | grad 0.1839 +2026-04-10 05:30:15 - INFO - parrotllm.training - step 8730 | epoch 0 | loss 3.5067 | lr 1.49e-03 | grad 0.1988 +2026-04-10 05:30:18 - INFO - parrotllm.training - step 8740 | epoch 0 | loss 3.5563 | lr 1.49e-03 | grad 0.1709 +2026-04-10 05:30:21 - INFO - parrotllm.training - step 8750 | epoch 0 | loss 3.6156 | lr 1.49e-03 | grad 0.1753 +2026-04-10 05:30:24 - INFO - parrotllm.training - step 8760 | epoch 0 | loss 3.4677 | lr 1.49e-03 | grad 0.1867 +2026-04-10 05:30:27 - INFO - parrotllm.training - step 8770 | epoch 0 | loss 3.5560 | lr 1.49e-03 | grad 0.1975 +2026-04-10 05:30:30 - INFO - parrotllm.training - step 8780 | epoch 0 | loss 3.5673 | lr 1.49e-03 | grad 0.1872 +2026-04-10 05:30:33 - INFO - parrotllm.training - step 8790 | epoch 0 | loss 3.5850 | lr 1.49e-03 | grad 0.2142 +2026-04-10 05:30:36 - INFO - parrotllm.training - step 8800 | epoch 0 | loss 3.5186 | lr 1.49e-03 | grad 0.1920 +2026-04-10 05:30:39 - INFO - parrotllm.training - step 8810 | epoch 0 | loss 3.5167 | lr 1.49e-03 | grad 0.1903 +2026-04-10 05:30:42 - INFO - parrotllm.training - step 8820 | epoch 0 | loss 3.4419 | lr 1.49e-03 | grad 0.1892 +2026-04-10 05:30:45 - INFO - parrotllm.training - step 8830 | epoch 0 | loss 3.5474 | lr 1.49e-03 | grad 0.1825 +2026-04-10 05:30:48 - INFO - parrotllm.training - step 8840 | epoch 0 | loss 3.4851 | lr 1.49e-03 | grad 0.1888 +2026-04-10 05:30:51 - INFO - parrotllm.training - step 8850 | epoch 0 | loss 3.5334 | lr 1.49e-03 | grad 0.2336 +2026-04-10 05:30:54 - INFO - parrotllm.training - step 8860 | epoch 0 | loss 3.4826 | lr 1.49e-03 | grad 0.2148 +2026-04-10 05:30:57 - INFO - parrotllm.training - step 8870 | epoch 0 | loss 3.6078 | lr 1.49e-03 | grad 0.1790 +2026-04-10 05:31:00 - INFO - parrotllm.training - step 8880 | epoch 0 | loss 3.5410 | lr 1.49e-03 | grad 0.1789 +2026-04-10 05:31:03 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 8891/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 05:31:04 - INFO - parrotllm.training - step 8890 | epoch 0 | loss 3.5016 | lr 1.49e-03 | grad 0.1912 +2026-04-10 05:31:07 - INFO - parrotllm.training - step 8900 | epoch 0 | loss 3.5284 | lr 1.49e-03 | grad 0.1875 +2026-04-10 05:31:10 - INFO - parrotllm.training - step 8910 | epoch 0 | loss 3.6671 | lr 1.49e-03 | grad 0.1924 +2026-04-10 05:31:13 - INFO - parrotllm.training - step 8920 | epoch 0 | loss 3.5735 | lr 1.49e-03 | grad 0.2235 +2026-04-10 05:31:16 - INFO - parrotllm.training - step 8930 | epoch 0 | loss 3.5812 | lr 1.49e-03 | grad 0.1840 +2026-04-10 05:31:19 - INFO - parrotllm.training - step 8940 | epoch 0 | loss 3.6006 | lr 1.49e-03 | grad 0.2056 +2026-04-10 05:31:22 - INFO - parrotllm.training - step 8950 | epoch 0 | loss 3.4839 | lr 1.49e-03 | grad 0.2131 +2026-04-10 05:31:25 - INFO - parrotllm.training - step 8960 | epoch 0 | loss 3.5499 | lr 1.49e-03 | grad 0.1916 +2026-04-10 05:31:28 - INFO - parrotllm.training - step 8970 | epoch 0 | loss 3.5424 | lr 1.49e-03 | grad 0.2112 +2026-04-10 05:31:31 - INFO - parrotllm.training - step 8980 | epoch 0 | loss 3.4642 | lr 1.49e-03 | grad 0.2148 +2026-04-10 05:31:34 - INFO - parrotllm.training - step 8990 | epoch 0 | loss 3.5945 | lr 1.49e-03 | grad 0.1734 +2026-04-10 05:31:37 - INFO - parrotllm.training - step 9000 | epoch 0 | loss 3.5376 | lr 1.49e-03 | grad 0.2210 +2026-04-10 05:31:37 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:31:37 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:31:40 - INFO - parrotllm.training - Train: loss=3.5376, ppl=34.38 +2026-04-10 05:31:40 - INFO - parrotllm.training - Val: loss=3.4892, ppl=32.76 +2026-04-10 05:31:40 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:31:41 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4892_epoch_0000_step_0009000.pt +2026-04-10 05:31:42 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:31:45 - INFO - parrotllm.training - step 9010 | epoch 0 | loss 3.5275 | lr 1.49e-03 | grad 0.1705 +2026-04-10 05:31:48 - INFO - parrotllm.training - step 9020 | epoch 0 | loss 3.5059 | lr 1.49e-03 | grad 0.2170 +2026-04-10 05:31:51 - INFO - parrotllm.training - step 9030 | epoch 0 | loss 3.5877 | lr 1.49e-03 | grad 0.2162 +2026-04-10 05:31:54 - INFO - parrotllm.training - step 9040 | epoch 0 | loss 3.5916 | lr 1.49e-03 | grad 0.1898 +2026-04-10 05:31:57 - INFO - parrotllm.training - step 9050 | epoch 0 | loss 3.5189 | lr 1.49e-03 | grad 0.1800 +2026-04-10 05:32:00 - INFO - parrotllm.training - step 9060 | epoch 0 | loss 3.5602 | lr 1.49e-03 | grad 0.1862 +2026-04-10 05:32:03 - INFO - parrotllm.training - step 9070 | epoch 0 | loss 3.5474 | lr 1.49e-03 | grad 0.1896 +2026-04-10 05:32:06 - INFO - parrotllm.training - step 9080 | epoch 0 | loss 3.5278 | lr 1.49e-03 | grad 0.1832 +2026-04-10 05:32:10 - INFO - parrotllm.training - step 9090 | epoch 0 | loss 3.4429 | lr 1.49e-03 | grad 0.1707 +2026-04-10 05:32:13 - INFO - parrotllm.training - step 9100 | epoch 0 | loss 3.5645 | lr 1.49e-03 | grad 0.1882 +2026-04-10 05:32:16 - INFO - parrotllm.training - step 9110 | epoch 0 | loss 3.6532 | lr 1.49e-03 | grad 0.1921 +2026-04-10 05:32:19 - INFO - parrotllm.training - step 9120 | epoch 0 | loss 3.6627 | lr 1.49e-03 | grad 0.1781 +2026-04-10 05:32:22 - INFO - parrotllm.training - step 9130 | epoch 0 | loss 3.5629 | lr 1.49e-03 | grad 0.1997 +2026-04-10 05:32:25 - INFO - parrotllm.training - step 9140 | epoch 0 | loss 3.4181 | lr 1.49e-03 | grad 0.1761 +2026-04-10 05:32:28 - INFO - parrotllm.training - step 9150 | epoch 0 | loss 3.5480 | lr 1.49e-03 | grad 0.1976 +2026-04-10 05:32:31 - INFO - parrotllm.training - step 9160 | epoch 0 | loss 3.4607 | lr 1.49e-03 | grad 0.2092 +2026-04-10 05:32:34 - INFO - parrotllm.training - step 9170 | epoch 0 | loss 3.6134 | lr 1.49e-03 | grad 0.1939 +2026-04-10 05:32:37 - INFO - parrotllm.training - step 9180 | epoch 0 | loss 3.6111 | lr 1.49e-03 | grad 0.2035 +2026-04-10 05:32:40 - INFO - parrotllm.training - step 9190 | epoch 0 | loss 3.6261 | lr 1.49e-03 | grad 0.1856 +2026-04-10 05:32:43 - INFO - parrotllm.training - step 9200 | epoch 0 | loss 3.4563 | lr 1.49e-03 | grad 0.1806 +2026-04-10 05:32:46 - INFO - parrotllm.training - step 9210 | epoch 0 | loss 3.6170 | lr 1.49e-03 | grad 0.1768 +2026-04-10 05:32:49 - INFO - parrotllm.training - step 9220 | epoch 0 | loss 3.5524 | lr 1.49e-03 | grad 0.1746 +2026-04-10 05:32:52 - INFO - parrotllm.training - step 9230 | epoch 0 | loss 3.6921 | lr 1.49e-03 | grad 0.2030 +2026-04-10 05:32:55 - INFO - parrotllm.training - step 9240 | epoch 0 | loss 3.5878 | lr 1.49e-03 | grad 0.1862 +2026-04-10 05:32:58 - INFO - parrotllm.training - step 9250 | epoch 0 | loss 3.6175 | lr 1.49e-03 | grad 0.1914 +2026-04-10 05:33:01 - INFO - parrotllm.training - step 9260 | epoch 0 | loss 3.5811 | lr 1.49e-03 | grad 0.1848 +2026-04-10 05:33:04 - INFO - parrotllm.training - step 9270 | epoch 0 | loss 3.6040 | lr 1.49e-03 | grad 0.2013 +2026-04-10 05:33:07 - INFO - parrotllm.training - step 9280 | epoch 0 | loss 3.5274 | lr 1.49e-03 | grad 0.1875 +2026-04-10 05:33:10 - INFO - parrotllm.training - step 9290 | epoch 0 | loss 3.5288 | lr 1.49e-03 | grad 0.1700 +2026-04-10 05:33:14 - INFO - parrotllm.training - step 9300 | epoch 0 | loss 3.5293 | lr 1.49e-03 | grad 0.2166 +2026-04-10 05:33:17 - INFO - parrotllm.training - step 9310 | epoch 0 | loss 3.4316 | lr 1.49e-03 | grad 0.1862 +2026-04-10 05:33:20 - INFO - parrotllm.training - step 9320 | epoch 0 | loss 3.4665 | lr 1.49e-03 | grad 0.1979 +2026-04-10 05:33:23 - INFO - parrotllm.training - step 9330 | epoch 0 | loss 3.5978 | lr 1.49e-03 | grad 0.1772 +2026-04-10 05:33:26 - INFO - parrotllm.training - step 9340 | epoch 0 | loss 3.6956 | lr 1.49e-03 | grad 0.2072 +2026-04-10 05:33:29 - INFO - parrotllm.training - step 9350 | epoch 0 | loss 3.6082 | lr 1.49e-03 | grad 0.2068 +2026-04-10 05:33:32 - INFO - parrotllm.training - step 9360 | epoch 0 | loss 3.4819 | lr 1.49e-03 | grad 0.2055 +2026-04-10 05:33:35 - INFO - parrotllm.training - step 9370 | epoch 0 | loss 3.5444 | lr 1.49e-03 | grad 0.1967 +2026-04-10 05:33:38 - INFO - parrotllm.training - step 9380 | epoch 0 | loss 3.5488 | lr 1.49e-03 | grad 0.1917 +2026-04-10 05:33:41 - INFO - parrotllm.training - step 9390 | epoch 0 | loss 3.5235 | lr 1.49e-03 | grad 0.2478 +2026-04-10 05:33:44 - INFO - parrotllm.training - step 9400 | epoch 0 | loss 3.5156 | lr 1.49e-03 | grad 0.2388 +2026-04-10 05:33:47 - INFO - parrotllm.training - step 9410 | epoch 0 | loss 3.5997 | lr 1.49e-03 | grad 0.1847 +2026-04-10 05:33:50 - INFO - parrotllm.training - step 9420 | epoch 0 | loss 3.5066 | lr 1.49e-03 | grad 0.1883 +2026-04-10 05:33:53 - INFO - parrotllm.training - step 9430 | epoch 0 | loss 3.6473 | lr 1.49e-03 | grad 0.2061 +2026-04-10 05:33:56 - INFO - parrotllm.training - step 9440 | epoch 0 | loss 3.6510 | lr 1.49e-03 | grad 0.1960 +2026-04-10 05:33:59 - INFO - parrotllm.training - step 9450 | epoch 0 | loss 3.5351 | lr 1.49e-03 | grad 0.1785 +2026-04-10 05:34:02 - INFO - parrotllm.training - step 9460 | epoch 0 | loss 3.5023 | lr 1.49e-03 | grad 0.1871 +2026-04-10 05:34:05 - INFO - parrotllm.training - step 9470 | epoch 0 | loss 3.5919 | lr 1.49e-03 | grad 0.1978 +2026-04-10 05:34:08 - INFO - parrotllm.training - step 9480 | epoch 0 | loss 3.4641 | lr 1.49e-03 | grad 0.1969 +2026-04-10 05:34:11 - INFO - parrotllm.training - step 9490 | epoch 0 | loss 3.5714 | lr 1.49e-03 | grad 0.1710 +2026-04-10 05:34:14 - INFO - parrotllm.training - step 9500 | epoch 0 | loss 3.4679 | lr 1.49e-03 | grad 0.1812 +2026-04-10 05:34:14 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:34:14 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:34:18 - INFO - parrotllm.training - Train: loss=3.4679, ppl=32.07 +2026-04-10 05:34:18 - INFO - parrotllm.training - Val: loss=3.4820, ppl=32.52 +2026-04-10 05:34:18 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:34:19 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4820_epoch_0000_step_0009500.pt +2026-04-10 05:34:19 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:34:22 - INFO - parrotllm.training - step 9510 | epoch 0 | loss 3.5073 | lr 1.49e-03 | grad 0.1811 +2026-04-10 05:34:26 - INFO - parrotllm.training - step 9520 | epoch 0 | loss 3.4366 | lr 1.49e-03 | grad 0.1983 +2026-04-10 05:34:29 - INFO - parrotllm.training - step 9530 | epoch 0 | loss 3.5315 | lr 1.49e-03 | grad 0.2115 +2026-04-10 05:34:32 - INFO - parrotllm.training - step 9540 | epoch 0 | loss 3.6094 | lr 1.49e-03 | grad 0.1938 +2026-04-10 05:34:35 - INFO - parrotllm.training - step 9550 | epoch 0 | loss 3.6263 | lr 1.49e-03 | grad 0.1963 +2026-04-10 05:34:38 - INFO - parrotllm.training - step 9560 | epoch 0 | loss 3.4948 | lr 1.49e-03 | grad 0.1860 +2026-04-10 05:34:41 - INFO - parrotllm.training - step 9570 | epoch 0 | loss 3.5778 | lr 1.49e-03 | grad 0.1945 +2026-04-10 05:34:44 - INFO - parrotllm.training - step 9580 | epoch 0 | loss 3.6582 | lr 1.49e-03 | grad 0.2048 +2026-04-10 05:34:47 - INFO - parrotllm.training - step 9590 | epoch 0 | loss 3.5573 | lr 1.49e-03 | grad 0.1863 +2026-04-10 05:34:50 - INFO - parrotllm.training - step 9600 | epoch 0 | loss 3.5703 | lr 1.49e-03 | grad 0.1845 +2026-04-10 05:34:53 - INFO - parrotllm.training - step 9610 | epoch 0 | loss 3.5941 | lr 1.48e-03 | grad 0.1968 +2026-04-10 05:34:56 - INFO - parrotllm.training - step 9620 | epoch 0 | loss 3.5709 | lr 1.48e-03 | grad 0.1976 +2026-04-10 05:34:59 - INFO - parrotllm.training - step 9630 | epoch 0 | loss 3.4333 | lr 1.48e-03 | grad 0.1904 +2026-04-10 05:35:02 - INFO - parrotllm.training - step 9640 | epoch 0 | loss 3.4382 | lr 1.48e-03 | grad 0.1856 +2026-04-10 05:35:05 - INFO - parrotllm.training - step 9650 | epoch 0 | loss 3.4766 | lr 1.48e-03 | grad 0.1866 +2026-04-10 05:35:08 - INFO - parrotllm.training - step 9660 | epoch 0 | loss 3.6215 | lr 1.48e-03 | grad 0.1850 +2026-04-10 05:35:11 - INFO - parrotllm.training - step 9670 | epoch 0 | loss 3.4909 | lr 1.48e-03 | grad 0.1800 +2026-04-10 05:35:14 - INFO - parrotllm.training - step 9680 | epoch 0 | loss 3.5997 | lr 1.48e-03 | grad 0.1705 +2026-04-10 05:35:17 - INFO - parrotllm.training - step 9690 | epoch 0 | loss 3.6261 | lr 1.48e-03 | grad 0.1820 +2026-04-10 05:35:20 - INFO - parrotllm.training - step 9700 | epoch 0 | loss 3.5384 | lr 1.48e-03 | grad 0.2000 +2026-04-10 05:35:23 - INFO - parrotllm.training - step 9710 | epoch 0 | loss 3.4696 | lr 1.48e-03 | grad 0.1872 +2026-04-10 05:35:26 - INFO - parrotllm.training - step 9720 | epoch 0 | loss 3.5390 | lr 1.48e-03 | grad 0.2163 +2026-04-10 05:35:29 - INFO - parrotllm.training - step 9730 | epoch 0 | loss 3.6166 | lr 1.48e-03 | grad 0.1752 +2026-04-10 05:35:33 - INFO - parrotllm.training - step 9740 | epoch 0 | loss 3.5645 | lr 1.48e-03 | grad 0.2023 +2026-04-10 05:35:36 - INFO - parrotllm.training - step 9750 | epoch 0 | loss 3.5221 | lr 1.48e-03 | grad 0.1821 +2026-04-10 05:35:39 - INFO - parrotllm.training - step 9760 | epoch 0 | loss 3.5200 | lr 1.48e-03 | grad 0.1662 +2026-04-10 05:35:42 - INFO - parrotllm.training - step 9770 | epoch 0 | loss 3.5229 | lr 1.48e-03 | grad 0.1786 +2026-04-10 05:35:45 - INFO - parrotllm.training - step 9780 | epoch 0 | loss 3.5827 | lr 1.48e-03 | grad 0.1965 +2026-04-10 05:35:48 - INFO - parrotllm.training - step 9790 | epoch 0 | loss 3.6696 | lr 1.48e-03 | grad 0.1757 +2026-04-10 05:35:51 - INFO - parrotllm.training - step 9800 | epoch 0 | loss 3.4836 | lr 1.48e-03 | grad 0.1933 +2026-04-10 05:35:54 - INFO - parrotllm.training - step 9810 | epoch 0 | loss 3.4956 | lr 1.48e-03 | grad 0.1796 +2026-04-10 05:35:57 - INFO - parrotllm.training - step 9820 | epoch 0 | loss 3.5700 | lr 1.48e-03 | grad 0.1942 +2026-04-10 05:36:00 - INFO - parrotllm.training - step 9830 | epoch 0 | loss 3.5184 | lr 1.48e-03 | grad 0.1764 +2026-04-10 05:36:03 - INFO - parrotllm.training - step 9840 | epoch 0 | loss 3.5506 | lr 1.48e-03 | grad 0.1645 +2026-04-10 05:36:06 - INFO - parrotllm.training - step 9850 | epoch 0 | loss 3.5528 | lr 1.48e-03 | grad 0.2263 +2026-04-10 05:36:09 - INFO - parrotllm.training - step 9860 | epoch 0 | loss 3.5051 | lr 1.48e-03 | grad 0.1909 +2026-04-10 05:36:12 - INFO - parrotllm.training - step 9870 | epoch 0 | loss 3.5326 | lr 1.48e-03 | grad 0.1880 +2026-04-10 05:36:15 - INFO - parrotllm.training - step 9880 | epoch 0 | loss 3.5225 | lr 1.48e-03 | grad 0.2350 +2026-04-10 05:36:18 - INFO - parrotllm.training - step 9890 | epoch 0 | loss 3.5648 | lr 1.48e-03 | grad 0.2542 +2026-04-10 05:36:21 - INFO - parrotllm.training - step 9900 | epoch 0 | loss 3.5158 | lr 1.48e-03 | grad 0.1916 +2026-04-10 05:36:24 - INFO - parrotllm.training - step 9910 | epoch 0 | loss 3.5118 | lr 1.48e-03 | grad 0.2049 +2026-04-10 05:36:27 - INFO - parrotllm.training - step 9920 | epoch 0 | loss 3.6073 | lr 1.48e-03 | grad 0.1909 +2026-04-10 05:36:30 - INFO - parrotllm.training - step 9930 | epoch 0 | loss 3.5407 | lr 1.48e-03 | grad 0.2143 +2026-04-10 05:36:33 - INFO - parrotllm.training - step 9940 | epoch 0 | loss 3.4899 | lr 1.48e-03 | grad 0.2276 +2026-04-10 05:36:37 - INFO - parrotllm.training - step 9950 | epoch 0 | loss 3.5153 | lr 1.48e-03 | grad 0.1803 +2026-04-10 05:36:40 - INFO - parrotllm.training - step 9960 | epoch 0 | loss 3.5247 | lr 1.48e-03 | grad 0.1871 +2026-04-10 05:36:43 - INFO - parrotllm.training - step 9970 | epoch 0 | loss 3.5920 | lr 1.48e-03 | grad 0.1945 +2026-04-10 05:36:46 - INFO - parrotllm.training - step 9980 | epoch 0 | loss 3.5292 | lr 1.48e-03 | grad 0.2141 +2026-04-10 05:36:49 - INFO - parrotllm.training - step 9990 | epoch 0 | loss 3.5383 | lr 1.48e-03 | grad 0.1974 +2026-04-10 05:36:52 - INFO - parrotllm.training - step 10000 | epoch 0 | loss 3.5015 | lr 1.48e-03 | grad 0.2289 +2026-04-10 05:36:52 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:36:52 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:36:55 - INFO - parrotllm.training - Train: loss=3.5015, ppl=33.16 +2026-04-10 05:36:55 - INFO - parrotllm.training - Val: loss=3.4769, ppl=32.36 +2026-04-10 05:36:55 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:36:56 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4769_epoch_0000_step_0010000.pt +2026-04-10 05:36:57 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:36:58 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0010000.pt +2026-04-10 05:37:02 - INFO - parrotllm.training - step 10010 | epoch 0 | loss 3.5413 | lr 1.48e-03 | grad 0.1796 +2026-04-10 05:37:05 - INFO - parrotllm.training - step 10020 | epoch 0 | loss 3.5541 | lr 1.48e-03 | grad 0.2113 +2026-04-10 05:37:08 - INFO - parrotllm.training - step 10030 | epoch 0 | loss 3.5767 | lr 1.48e-03 | grad 0.1927 +2026-04-10 05:37:11 - INFO - parrotllm.training - step 10040 | epoch 0 | loss 3.6037 | lr 1.48e-03 | grad 0.1931 +2026-04-10 05:37:15 - INFO - parrotllm.training - step 10050 | epoch 0 | loss 3.6006 | lr 1.48e-03 | grad 0.1825 +2026-04-10 05:37:18 - INFO - parrotllm.training - step 10060 | epoch 0 | loss 3.6364 | lr 1.48e-03 | grad 0.1980 +2026-04-10 05:37:21 - INFO - parrotllm.training - step 10070 | epoch 0 | loss 3.5823 | lr 1.48e-03 | grad 0.1746 +2026-04-10 05:37:24 - INFO - parrotllm.training - step 10080 | epoch 0 | loss 3.6187 | lr 1.48e-03 | grad 0.2136 +2026-04-10 05:37:27 - INFO - parrotllm.training - step 10090 | epoch 0 | loss 3.5355 | lr 1.48e-03 | grad 0.1964 +2026-04-10 05:37:30 - INFO - parrotllm.training - step 10100 | epoch 0 | loss 3.6522 | lr 1.48e-03 | grad 0.1980 +2026-04-10 05:37:33 - INFO - parrotllm.training - step 10110 | epoch 0 | loss 3.5045 | lr 1.48e-03 | grad 0.1955 +2026-04-10 05:37:36 - INFO - parrotllm.training - step 10120 | epoch 0 | loss 3.6649 | lr 1.48e-03 | grad 0.2006 +2026-04-10 05:37:39 - INFO - parrotllm.training - step 10130 | epoch 0 | loss 3.6032 | lr 1.48e-03 | grad 0.1871 +2026-04-10 05:37:42 - INFO - parrotllm.training - step 10140 | epoch 0 | loss 3.4950 | lr 1.48e-03 | grad 0.1769 +2026-04-10 05:37:45 - INFO - parrotllm.training - step 10150 | epoch 0 | loss 3.5743 | lr 1.48e-03 | grad 0.1851 +2026-04-10 05:37:48 - INFO - parrotllm.training - step 10160 | epoch 0 | loss 3.4044 | lr 1.48e-03 | grad 0.1955 +2026-04-10 05:37:51 - INFO - parrotllm.training - step 10170 | epoch 0 | loss 3.4665 | lr 1.48e-03 | grad 0.1784 +2026-04-10 05:37:54 - INFO - parrotllm.training - step 10180 | epoch 0 | loss 3.5883 | lr 1.48e-03 | grad 0.1739 +2026-04-10 05:37:57 - INFO - parrotllm.training - step 10190 | epoch 0 | loss 3.5463 | lr 1.48e-03 | grad 0.2509 +2026-04-10 05:38:00 - INFO - parrotllm.training - step 10200 | epoch 0 | loss 3.5269 | lr 1.48e-03 | grad 0.1818 +2026-04-10 05:38:03 - INFO - parrotllm.training - step 10210 | epoch 0 | loss 3.4741 | lr 1.48e-03 | grad 0.1979 +2026-04-10 05:38:06 - INFO - parrotllm.training - step 10220 | epoch 0 | loss 3.5177 | lr 1.48e-03 | grad 0.1834 +2026-04-10 05:38:09 - INFO - parrotllm.training - step 10230 | epoch 0 | loss 3.5787 | lr 1.48e-03 | grad 0.2018 +2026-04-10 05:38:12 - INFO - parrotllm.training - step 10240 | epoch 0 | loss 3.4598 | lr 1.48e-03 | grad 0.2236 +2026-04-10 05:38:15 - INFO - parrotllm.training - step 10250 | epoch 0 | loss 3.4417 | lr 1.48e-03 | grad 0.1716 +2026-04-10 05:38:18 - INFO - parrotllm.training - step 10260 | epoch 0 | loss 3.5576 | lr 1.48e-03 | grad 0.2198 +2026-04-10 05:38:22 - INFO - parrotllm.training - step 10270 | epoch 0 | loss 3.4669 | lr 1.48e-03 | grad 0.1796 +2026-04-10 05:38:25 - INFO - parrotllm.training - step 10280 | epoch 0 | loss 3.4340 | lr 1.48e-03 | grad 0.1898 +2026-04-10 05:38:28 - INFO - parrotllm.training - step 10290 | epoch 0 | loss 3.5262 | lr 1.48e-03 | grad 0.2026 +2026-04-10 05:38:31 - INFO - parrotllm.training - step 10300 | epoch 0 | loss 3.5860 | lr 1.48e-03 | grad 0.2047 +2026-04-10 05:38:34 - INFO - parrotllm.training - step 10310 | epoch 0 | loss 3.5408 | lr 1.48e-03 | grad 0.2118 +2026-04-10 05:38:37 - INFO - parrotllm.training - step 10320 | epoch 0 | loss 3.4852 | lr 1.48e-03 | grad 0.1689 +2026-04-10 05:38:40 - INFO - parrotllm.training - step 10330 | epoch 0 | loss 3.5050 | lr 1.48e-03 | grad 0.1881 +2026-04-10 05:38:43 - INFO - parrotllm.training - step 10340 | epoch 0 | loss 3.6621 | lr 1.48e-03 | grad 0.2075 +2026-04-10 05:38:46 - INFO - parrotllm.training - step 10350 | epoch 0 | loss 3.6619 | lr 1.48e-03 | grad 0.1864 +2026-04-10 05:38:49 - INFO - parrotllm.training - step 10360 | epoch 0 | loss 3.5426 | lr 1.48e-03 | grad 0.1825 +2026-04-10 05:38:52 - INFO - parrotllm.training - step 10370 | epoch 0 | loss 3.5854 | lr 1.48e-03 | grad 0.1802 +2026-04-10 05:38:55 - INFO - parrotllm.training - step 10380 | epoch 0 | loss 3.4426 | lr 1.48e-03 | grad 0.1940 +2026-04-10 05:38:58 - INFO - parrotllm.training - step 10390 | epoch 0 | loss 3.6264 | lr 1.48e-03 | grad 0.2012 +2026-04-10 05:39:01 - INFO - parrotllm.training - step 10400 | epoch 0 | loss 3.4254 | lr 1.48e-03 | grad 0.1719 +2026-04-10 05:39:04 - INFO - parrotllm.training - step 10410 | epoch 0 | loss 3.6139 | lr 1.48e-03 | grad 0.1919 +2026-04-10 05:39:07 - INFO - parrotllm.training - step 10420 | epoch 0 | loss 3.5854 | lr 1.48e-03 | grad 0.2094 +2026-04-10 05:39:10 - INFO - parrotllm.training - step 10430 | epoch 0 | loss 3.3830 | lr 1.48e-03 | grad 0.1753 +2026-04-10 05:39:13 - INFO - parrotllm.training - step 10440 | epoch 0 | loss 3.6066 | lr 1.48e-03 | grad 0.1727 +2026-04-10 05:39:16 - INFO - parrotllm.training - step 10450 | epoch 0 | loss 3.5263 | lr 1.48e-03 | grad 0.1867 +2026-04-10 05:39:19 - INFO - parrotllm.training - step 10460 | epoch 0 | loss 3.6080 | lr 1.48e-03 | grad 0.2081 +2026-04-10 05:39:23 - INFO - parrotllm.training - step 10470 | epoch 0 | loss 3.4787 | lr 1.48e-03 | grad 0.1723 +2026-04-10 05:39:26 - INFO - parrotllm.training - step 10480 | epoch 0 | loss 3.6232 | lr 1.48e-03 | grad 0.1986 +2026-04-10 05:39:29 - INFO - parrotllm.training - step 10490 | epoch 0 | loss 3.4758 | lr 1.48e-03 | grad 0.1780 +2026-04-10 05:39:32 - INFO - parrotllm.training - step 10500 | epoch 0 | loss 3.5020 | lr 1.48e-03 | grad 0.1821 +2026-04-10 05:39:32 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:39:32 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:39:35 - INFO - parrotllm.training - Train: loss=3.5020, ppl=33.18 +2026-04-10 05:39:35 - INFO - parrotllm.training - Val: loss=3.4730, ppl=32.23 +2026-04-10 05:39:35 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:39:36 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4730_epoch_0000_step_0010500.pt +2026-04-10 05:39:37 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:39:40 - INFO - parrotllm.training - step 10510 | epoch 0 | loss 3.5092 | lr 1.48e-03 | grad 0.2041 +2026-04-10 05:39:43 - INFO - parrotllm.training - step 10520 | epoch 0 | loss 3.5882 | lr 1.48e-03 | grad 0.1764 +2026-04-10 05:39:46 - INFO - parrotllm.training - step 10530 | epoch 0 | loss 3.4784 | lr 1.48e-03 | grad 0.1997 +2026-04-10 05:39:49 - INFO - parrotllm.training - step 10540 | epoch 0 | loss 3.6162 | lr 1.48e-03 | grad 0.1773 +2026-04-10 05:39:52 - INFO - parrotllm.training - step 10550 | epoch 0 | loss 3.6089 | lr 1.48e-03 | grad 0.1876 +2026-04-10 05:39:55 - INFO - parrotllm.training - step 10560 | epoch 0 | loss 3.4944 | lr 1.48e-03 | grad 0.2007 +2026-04-10 05:39:58 - INFO - parrotllm.training - step 10570 | epoch 0 | loss 3.5423 | lr 1.48e-03 | grad 0.1870 +2026-04-10 05:40:01 - INFO - parrotllm.training - step 10580 | epoch 0 | loss 3.5738 | lr 1.48e-03 | grad 0.1855 +2026-04-10 05:40:04 - INFO - parrotllm.training - step 10590 | epoch 0 | loss 3.4734 | lr 1.48e-03 | grad 0.2170 +2026-04-10 05:40:07 - INFO - parrotllm.training - step 10600 | epoch 0 | loss 3.5195 | lr 1.48e-03 | grad 0.2015 +2026-04-10 05:40:10 - INFO - parrotllm.training - step 10610 | epoch 0 | loss 3.4648 | lr 1.48e-03 | grad 0.2288 +2026-04-10 05:40:13 - INFO - parrotllm.training - step 10620 | epoch 0 | loss 3.5055 | lr 1.48e-03 | grad 0.1870 +2026-04-10 05:40:16 - INFO - parrotllm.training - step 10630 | epoch 0 | loss 3.4687 | lr 1.48e-03 | grad 0.1724 +2026-04-10 05:40:19 - INFO - parrotllm.training - step 10640 | epoch 0 | loss 3.5387 | lr 1.48e-03 | grad 0.1880 +2026-04-10 05:40:22 - INFO - parrotllm.training - step 10650 | epoch 0 | loss 3.6084 | lr 1.48e-03 | grad 0.1869 +2026-04-10 05:40:25 - INFO - parrotllm.training - step 10660 | epoch 0 | loss 3.6198 | lr 1.48e-03 | grad 0.1821 +2026-04-10 05:40:28 - INFO - parrotllm.training - step 10670 | epoch 0 | loss 3.5738 | lr 1.48e-03 | grad 0.1985 +2026-04-10 05:40:31 - INFO - parrotllm.training - step 10680 | epoch 0 | loss 3.4663 | lr 1.48e-03 | grad 0.1963 +2026-04-10 05:40:35 - INFO - parrotllm.training - step 10690 | epoch 0 | loss 3.4731 | lr 1.48e-03 | grad 0.1828 +2026-04-10 05:40:38 - INFO - parrotllm.training - step 10700 | epoch 0 | loss 3.5674 | lr 1.48e-03 | grad 0.1805 +2026-04-10 05:40:41 - INFO - parrotllm.training - step 10710 | epoch 0 | loss 3.4930 | lr 1.48e-03 | grad 0.1665 +2026-04-10 05:40:44 - INFO - parrotllm.training - step 10720 | epoch 0 | loss 3.6162 | lr 1.48e-03 | grad 0.2319 +2026-04-10 05:40:47 - INFO - parrotllm.training - step 10730 | epoch 0 | loss 3.5795 | lr 1.48e-03 | grad 0.2063 +2026-04-10 05:40:50 - INFO - parrotllm.training - step 10740 | epoch 0 | loss 3.5227 | lr 1.48e-03 | grad 0.1812 +2026-04-10 05:40:53 - INFO - parrotllm.training - step 10750 | epoch 0 | loss 3.5947 | lr 1.48e-03 | grad 0.1938 +2026-04-10 05:40:56 - INFO - parrotllm.training - step 10760 | epoch 0 | loss 3.5072 | lr 1.48e-03 | grad 0.1845 +2026-04-10 05:40:59 - INFO - parrotllm.training - step 10770 | epoch 0 | loss 3.4999 | lr 1.48e-03 | grad 0.2121 +2026-04-10 05:41:02 - INFO - parrotllm.training - step 10780 | epoch 0 | loss 3.4970 | lr 1.48e-03 | grad 0.1891 +2026-04-10 05:41:05 - INFO - parrotllm.training - step 10790 | epoch 0 | loss 3.6019 | lr 1.48e-03 | grad 0.2044 +2026-04-10 05:41:08 - INFO - parrotllm.training - step 10800 | epoch 0 | loss 3.5237 | lr 1.48e-03 | grad 0.2052 +2026-04-10 05:41:11 - INFO - parrotllm.training - step 10810 | epoch 0 | loss 3.5280 | lr 1.48e-03 | grad 0.2151 +2026-04-10 05:41:14 - INFO - parrotllm.training - step 10820 | epoch 0 | loss 3.5562 | lr 1.48e-03 | grad 0.1760 +2026-04-10 05:41:17 - INFO - parrotllm.training - step 10830 | epoch 0 | loss 3.4576 | lr 1.48e-03 | grad 0.1757 +2026-04-10 05:41:20 - INFO - parrotllm.training - step 10840 | epoch 0 | loss 3.6061 | lr 1.48e-03 | grad 0.2167 +2026-04-10 05:41:23 - INFO - parrotllm.training - step 10850 | epoch 0 | loss 3.5315 | lr 1.48e-03 | grad 0.2012 +2026-04-10 05:41:26 - INFO - parrotllm.training - step 10860 | epoch 0 | loss 3.5058 | lr 1.48e-03 | grad 0.1769 +2026-04-10 05:41:29 - INFO - parrotllm.training - step 10870 | epoch 0 | loss 3.5202 | lr 1.48e-03 | grad 0.1690 +2026-04-10 05:41:32 - INFO - parrotllm.training - step 10880 | epoch 0 | loss 3.5072 | lr 1.48e-03 | grad 0.1976 +2026-04-10 05:41:35 - INFO - parrotllm.training - step 10890 | epoch 0 | loss 3.5738 | lr 1.48e-03 | grad 0.1913 +2026-04-10 05:41:38 - INFO - parrotllm.training - step 10900 | epoch 0 | loss 3.4670 | lr 1.48e-03 | grad 0.2322 +2026-04-10 05:41:42 - INFO - parrotllm.training - step 10910 | epoch 0 | loss 3.5316 | lr 1.48e-03 | grad 0.1800 +2026-04-10 05:41:45 - INFO - parrotllm.training - step 10920 | epoch 0 | loss 3.5895 | lr 1.48e-03 | grad 0.1893 +2026-04-10 05:41:48 - INFO - parrotllm.training - step 10930 | epoch 0 | loss 3.4814 | lr 1.48e-03 | grad 0.1831 +2026-04-10 05:41:50 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 10941/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 05:41:51 - INFO - parrotllm.training - step 10940 | epoch 0 | loss 3.4964 | lr 1.48e-03 | grad 0.1854 +2026-04-10 05:41:54 - INFO - parrotllm.training - step 10950 | epoch 0 | loss 3.6258 | lr 1.48e-03 | grad 0.1901 +2026-04-10 05:41:57 - INFO - parrotllm.training - step 10960 | epoch 0 | loss 3.6096 | lr 1.48e-03 | grad 0.1768 +2026-04-10 05:42:00 - INFO - parrotllm.training - step 10970 | epoch 0 | loss 3.5057 | lr 1.48e-03 | grad 0.1977 +2026-04-10 05:42:03 - INFO - parrotllm.training - step 10980 | epoch 0 | loss 3.5613 | lr 1.48e-03 | grad 0.1878 +2026-04-10 05:42:06 - INFO - parrotllm.training - step 10990 | epoch 0 | loss 3.4779 | lr 1.48e-03 | grad 0.1856 +2026-04-10 05:42:09 - INFO - parrotllm.training - step 11000 | epoch 0 | loss 3.5664 | lr 1.48e-03 | grad 0.1897 +2026-04-10 05:42:09 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:42:09 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:42:12 - INFO - parrotllm.training - Train: loss=3.5664, ppl=35.39 +2026-04-10 05:42:12 - INFO - parrotllm.training - Val: loss=3.4704, ppl=32.15 +2026-04-10 05:42:12 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:42:13 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4704_epoch_0000_step_0011000.pt +2026-04-10 05:42:14 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:42:17 - INFO - parrotllm.training - step 11010 | epoch 0 | loss 3.3739 | lr 1.48e-03 | grad 0.1588 +2026-04-10 05:42:20 - INFO - parrotllm.training - step 11020 | epoch 0 | loss 3.5154 | lr 1.48e-03 | grad 0.1949 +2026-04-10 05:42:23 - INFO - parrotllm.training - step 11030 | epoch 0 | loss 3.5684 | lr 1.48e-03 | grad 0.2082 +2026-04-10 05:42:27 - INFO - parrotllm.training - step 11040 | epoch 0 | loss 3.5814 | lr 1.48e-03 | grad 0.1932 +2026-04-10 05:42:30 - INFO - parrotllm.training - step 11050 | epoch 0 | loss 3.4602 | lr 1.48e-03 | grad 0.1823 +2026-04-10 05:42:33 - INFO - parrotllm.training - step 11060 | epoch 0 | loss 3.5549 | lr 1.48e-03 | grad 0.1775 +2026-04-10 05:42:36 - INFO - parrotllm.training - step 11070 | epoch 0 | loss 3.5273 | lr 1.48e-03 | grad 0.1977 +2026-04-10 05:42:39 - INFO - parrotllm.training - step 11080 | epoch 0 | loss 3.5431 | lr 1.48e-03 | grad 0.2083 +2026-04-10 05:42:42 - INFO - parrotllm.training - step 11090 | epoch 0 | loss 3.5158 | lr 1.48e-03 | grad 0.2026 +2026-04-10 05:42:45 - INFO - parrotllm.training - step 11100 | epoch 0 | loss 3.4532 | lr 1.48e-03 | grad 0.1894 +2026-04-10 05:42:48 - INFO - parrotllm.training - step 11110 | epoch 0 | loss 3.6978 | lr 1.48e-03 | grad 0.1815 +2026-04-10 05:42:51 - INFO - parrotllm.training - step 11120 | epoch 0 | loss 3.6253 | lr 1.48e-03 | grad 0.1779 +2026-04-10 05:42:54 - INFO - parrotllm.training - step 11130 | epoch 0 | loss 3.5815 | lr 1.48e-03 | grad 0.1868 +2026-04-10 05:42:57 - INFO - parrotllm.training - step 11140 | epoch 0 | loss 3.4045 | lr 1.48e-03 | grad 0.2025 +2026-04-10 05:43:00 - INFO - parrotllm.training - step 11150 | epoch 0 | loss 3.5632 | lr 1.48e-03 | grad 0.2052 +2026-04-10 05:43:03 - INFO - parrotllm.training - step 11160 | epoch 0 | loss 3.4398 | lr 1.48e-03 | grad 0.1797 +2026-04-10 05:43:06 - INFO - parrotllm.training - step 11170 | epoch 0 | loss 3.5311 | lr 1.48e-03 | grad 0.1843 +2026-04-10 05:43:09 - INFO - parrotllm.training - step 11180 | epoch 0 | loss 3.5246 | lr 1.48e-03 | grad 0.1840 +2026-04-10 05:43:12 - INFO - parrotllm.training - step 11190 | epoch 0 | loss 3.5142 | lr 1.48e-03 | grad 0.1658 +2026-04-10 05:43:15 - INFO - parrotllm.training - step 11200 | epoch 0 | loss 3.6288 | lr 1.48e-03 | grad 0.1766 +2026-04-10 05:43:18 - INFO - parrotllm.training - step 11210 | epoch 0 | loss 3.5494 | lr 1.48e-03 | grad 0.1977 +2026-04-10 05:43:21 - INFO - parrotllm.training - step 11220 | epoch 0 | loss 3.3709 | lr 1.48e-03 | grad 0.1963 +2026-04-10 05:43:24 - INFO - parrotllm.training - step 11230 | epoch 0 | loss 3.5888 | lr 1.48e-03 | grad 0.1766 +2026-04-10 05:43:27 - INFO - parrotllm.training - step 11240 | epoch 0 | loss 3.5172 | lr 1.48e-03 | grad 0.2243 +2026-04-10 05:43:30 - INFO - parrotllm.training - step 11250 | epoch 0 | loss 3.5217 | lr 1.48e-03 | grad 0.2110 +2026-04-10 05:43:34 - INFO - parrotllm.training - step 11260 | epoch 0 | loss 3.5228 | lr 1.48e-03 | grad 0.1914 +2026-04-10 05:43:37 - INFO - parrotllm.training - step 11270 | epoch 0 | loss 3.5431 | lr 1.48e-03 | grad 0.2109 +2026-04-10 05:43:40 - INFO - parrotllm.training - step 11280 | epoch 0 | loss 3.6358 | lr 1.48e-03 | grad 0.1775 +2026-04-10 05:43:43 - INFO - parrotllm.training - step 11290 | epoch 0 | loss 3.4459 | lr 1.48e-03 | grad 0.2174 +2026-04-10 05:43:46 - INFO - parrotllm.training - step 11300 | epoch 0 | loss 3.4936 | lr 1.48e-03 | grad 0.1699 +2026-04-10 05:43:49 - INFO - parrotllm.training - step 11310 | epoch 0 | loss 3.4124 | lr 1.48e-03 | grad 0.1713 +2026-04-10 05:43:52 - INFO - parrotllm.training - step 11320 | epoch 0 | loss 3.5362 | lr 1.48e-03 | grad 0.1786 +2026-04-10 05:43:55 - INFO - parrotllm.training - step 11330 | epoch 0 | loss 3.6196 | lr 1.48e-03 | grad 0.1956 +2026-04-10 05:43:58 - INFO - parrotllm.training - step 11340 | epoch 0 | loss 3.3621 | lr 1.48e-03 | grad 0.1988 +2026-04-10 05:44:01 - INFO - parrotllm.training - step 11350 | epoch 0 | loss 3.4685 | lr 1.48e-03 | grad 0.2161 +2026-04-10 05:44:04 - INFO - parrotllm.training - step 11360 | epoch 0 | loss 3.6402 | lr 1.48e-03 | grad 0.2077 +2026-04-10 05:44:07 - INFO - parrotllm.training - step 11370 | epoch 0 | loss 3.5633 | lr 1.48e-03 | grad 0.2150 +2026-04-10 05:44:10 - INFO - parrotllm.training - step 11380 | epoch 0 | loss 3.4559 | lr 1.48e-03 | grad 0.1732 +2026-04-10 05:44:13 - INFO - parrotllm.training - step 11390 | epoch 0 | loss 3.5546 | lr 1.48e-03 | grad 0.1846 +2026-04-10 05:44:16 - INFO - parrotllm.training - step 11400 | epoch 0 | loss 3.4176 | lr 1.48e-03 | grad 0.1904 +2026-04-10 05:44:19 - INFO - parrotllm.training - step 11410 | epoch 0 | loss 3.5955 | lr 1.48e-03 | grad 0.1977 +2026-04-10 05:44:22 - INFO - parrotllm.training - step 11420 | epoch 0 | loss 3.4717 | lr 1.48e-03 | grad 0.1971 +2026-04-10 05:44:25 - INFO - parrotllm.training - step 11430 | epoch 0 | loss 3.4712 | lr 1.48e-03 | grad 0.1963 +2026-04-10 05:44:28 - INFO - parrotllm.training - step 11440 | epoch 0 | loss 3.5915 | lr 1.48e-03 | grad 0.1813 +2026-04-10 05:44:31 - INFO - parrotllm.training - step 11450 | epoch 0 | loss 3.5387 | lr 1.48e-03 | grad 0.1944 +2026-04-10 05:44:34 - INFO - parrotllm.training - step 11460 | epoch 0 | loss 3.5729 | lr 1.48e-03 | grad 0.1728 +2026-04-10 05:44:37 - INFO - parrotllm.training - step 11470 | epoch 0 | loss 3.4716 | lr 1.48e-03 | grad 0.1933 +2026-04-10 05:44:41 - INFO - parrotllm.training - step 11480 | epoch 0 | loss 3.3996 | lr 1.48e-03 | grad 0.2047 +2026-04-10 05:44:44 - INFO - parrotllm.training - step 11490 | epoch 0 | loss 3.6208 | lr 1.48e-03 | grad 0.1937 +2026-04-10 05:44:47 - INFO - parrotllm.training - step 11500 | epoch 0 | loss 3.4642 | lr 1.48e-03 | grad 0.1918 +2026-04-10 05:44:47 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:44:47 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:44:50 - INFO - parrotllm.training - Train: loss=3.4642, ppl=31.95 +2026-04-10 05:44:50 - INFO - parrotllm.training - Val: loss=3.4683, ppl=32.08 +2026-04-10 05:44:50 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:44:51 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4683_epoch_0000_step_0011500.pt +2026-04-10 05:44:52 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:44:55 - INFO - parrotllm.training - step 11510 | epoch 0 | loss 3.4342 | lr 1.48e-03 | grad 0.2078 +2026-04-10 05:44:58 - INFO - parrotllm.training - step 11520 | epoch 0 | loss 3.5186 | lr 1.48e-03 | grad 0.1737 +2026-04-10 05:45:01 - INFO - parrotllm.training - step 11530 | epoch 0 | loss 3.4904 | lr 1.48e-03 | grad 0.1732 +2026-04-10 05:45:04 - INFO - parrotllm.training - step 11540 | epoch 0 | loss 3.5769 | lr 1.48e-03 | grad 0.2116 +2026-04-10 05:45:07 - INFO - parrotllm.training - step 11550 | epoch 0 | loss 3.4862 | lr 1.48e-03 | grad 0.2135 +2026-04-10 05:45:10 - INFO - parrotllm.training - step 11560 | epoch 0 | loss 3.4156 | lr 1.48e-03 | grad 0.1812 +2026-04-10 05:45:13 - INFO - parrotllm.training - step 11570 | epoch 0 | loss 3.5617 | lr 1.48e-03 | grad 0.2135 +2026-04-10 05:45:16 - INFO - parrotllm.training - step 11580 | epoch 0 | loss 3.5090 | lr 1.48e-03 | grad 0.2298 +2026-04-10 05:45:19 - INFO - parrotllm.training - step 11590 | epoch 0 | loss 3.4167 | lr 1.48e-03 | grad 0.1930 +2026-04-10 05:45:22 - INFO - parrotllm.training - step 11600 | epoch 0 | loss 3.4828 | lr 1.48e-03 | grad 0.1684 +2026-04-10 05:45:25 - INFO - parrotllm.training - step 11610 | epoch 0 | loss 3.4606 | lr 1.48e-03 | grad 0.2146 +2026-04-10 05:45:28 - INFO - parrotllm.training - step 11620 | epoch 0 | loss 3.4848 | lr 1.48e-03 | grad 0.1957 +2026-04-10 05:45:31 - INFO - parrotllm.training - step 11630 | epoch 0 | loss 3.5358 | lr 1.48e-03 | grad 0.2140 +2026-04-10 05:45:34 - INFO - parrotllm.training - step 11640 | epoch 0 | loss 3.4856 | lr 1.48e-03 | grad 0.1864 +2026-04-10 05:45:37 - INFO - parrotllm.training - step 11650 | epoch 0 | loss 3.4885 | lr 1.48e-03 | grad 0.1889 +2026-04-10 05:45:40 - INFO - parrotllm.training - step 11660 | epoch 0 | loss 3.5644 | lr 1.48e-03 | grad 0.1812 +2026-04-10 05:45:43 - INFO - parrotllm.training - step 11670 | epoch 0 | loss 3.6181 | lr 1.48e-03 | grad 0.2094 +2026-04-10 05:45:46 - INFO - parrotllm.training - step 11680 | epoch 0 | loss 3.5176 | lr 1.48e-03 | grad 0.1876 +2026-04-10 05:45:49 - INFO - parrotllm.training - step 11690 | epoch 0 | loss 3.5715 | lr 1.48e-03 | grad 0.1700 +2026-04-10 05:45:53 - INFO - parrotllm.training - step 11700 | epoch 0 | loss 3.4921 | lr 1.48e-03 | grad 0.2080 +2026-04-10 05:45:56 - INFO - parrotllm.training - step 11710 | epoch 0 | loss 3.5758 | lr 1.48e-03 | grad 0.1895 +2026-04-10 05:45:59 - INFO - parrotllm.training - step 11720 | epoch 0 | loss 3.4625 | lr 1.48e-03 | grad 0.1878 +2026-04-10 05:46:02 - INFO - parrotllm.training - step 11730 | epoch 0 | loss 3.4780 | lr 1.48e-03 | grad 0.1898 +2026-04-10 05:46:05 - INFO - parrotllm.training - step 11740 | epoch 0 | loss 3.5558 | lr 1.48e-03 | grad 0.1840 +2026-04-10 05:46:08 - INFO - parrotllm.training - step 11750 | epoch 0 | loss 3.5855 | lr 1.48e-03 | grad 0.1768 +2026-04-10 05:46:11 - INFO - parrotllm.training - step 11760 | epoch 0 | loss 3.4845 | lr 1.48e-03 | grad 0.1738 +2026-04-10 05:46:14 - INFO - parrotllm.training - step 11770 | epoch 0 | loss 3.4788 | lr 1.48e-03 | grad 0.2167 +2026-04-10 05:46:17 - INFO - parrotllm.training - step 11780 | epoch 0 | loss 3.4776 | lr 1.48e-03 | grad 0.1982 +2026-04-10 05:46:20 - INFO - parrotllm.training - step 11790 | epoch 0 | loss 3.4707 | lr 1.48e-03 | grad 0.2270 +2026-04-10 05:46:23 - INFO - parrotllm.training - step 11800 | epoch 0 | loss 3.5569 | lr 1.48e-03 | grad 0.1851 +2026-04-10 05:46:26 - INFO - parrotllm.training - step 11810 | epoch 0 | loss 3.5288 | lr 1.48e-03 | grad 0.2083 +2026-04-10 05:46:29 - INFO - parrotllm.training - step 11820 | epoch 0 | loss 3.5933 | lr 1.48e-03 | grad 0.2034 +2026-04-10 05:46:32 - INFO - parrotllm.training - step 11830 | epoch 0 | loss 3.5587 | lr 1.48e-03 | grad 0.1962 +2026-04-10 05:46:35 - INFO - parrotllm.training - step 11840 | epoch 0 | loss 3.5544 | lr 1.48e-03 | grad 0.2005 +2026-04-10 05:46:38 - INFO - parrotllm.training - step 11850 | epoch 0 | loss 3.5207 | lr 1.48e-03 | grad 0.1868 +2026-04-10 05:46:41 - INFO - parrotllm.training - step 11860 | epoch 0 | loss 3.4620 | lr 1.48e-03 | grad 0.1988 +2026-04-10 05:46:44 - INFO - parrotllm.training - step 11870 | epoch 0 | loss 3.5682 | lr 1.48e-03 | grad 0.1969 +2026-04-10 05:46:47 - INFO - parrotllm.training - step 11880 | epoch 0 | loss 3.4606 | lr 1.48e-03 | grad 0.1965 +2026-04-10 05:46:51 - INFO - parrotllm.training - step 11890 | epoch 0 | loss 3.4559 | lr 1.48e-03 | grad 0.1669 +2026-04-10 05:46:54 - INFO - parrotllm.training - step 11900 | epoch 0 | loss 3.5418 | lr 1.48e-03 | grad 0.1773 +2026-04-10 05:46:57 - INFO - parrotllm.training - step 11910 | epoch 0 | loss 3.6094 | lr 1.48e-03 | grad 0.2208 +2026-04-10 05:47:00 - INFO - parrotllm.training - step 11920 | epoch 0 | loss 3.4850 | lr 1.48e-03 | grad 0.1772 +2026-04-10 05:47:03 - INFO - parrotllm.training - step 11930 | epoch 0 | loss 3.3887 | lr 1.48e-03 | grad 0.1988 +2026-04-10 05:47:06 - INFO - parrotllm.training - step 11940 | epoch 0 | loss 3.5563 | lr 1.48e-03 | grad 0.2208 +2026-04-10 05:47:09 - INFO - parrotllm.training - step 11950 | epoch 0 | loss 3.4633 | lr 1.48e-03 | grad 0.2454 +2026-04-10 05:47:12 - INFO - parrotllm.training - step 11960 | epoch 0 | loss 3.4958 | lr 1.48e-03 | grad 0.1883 +2026-04-10 05:47:15 - INFO - parrotllm.training - step 11970 | epoch 0 | loss 3.5816 | lr 1.48e-03 | grad 0.1923 +2026-04-10 05:47:18 - INFO - parrotllm.training - step 11980 | epoch 0 | loss 3.5116 | lr 1.47e-03 | grad 0.1814 +2026-04-10 05:47:21 - INFO - parrotllm.training - step 11990 | epoch 0 | loss 3.6258 | lr 1.47e-03 | grad 0.2048 +2026-04-10 05:47:24 - INFO - parrotllm.training - step 12000 | epoch 0 | loss 3.4646 | lr 1.47e-03 | grad 0.2244 +2026-04-10 05:47:24 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:47:24 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:47:27 - INFO - parrotllm.training - Train: loss=3.4646, ppl=31.96 +2026-04-10 05:47:27 - INFO - parrotllm.training - Val: loss=3.4613, ppl=31.86 +2026-04-10 05:47:27 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:47:28 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4613_epoch_0000_step_0012000.pt +2026-04-10 05:47:29 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:47:32 - INFO - parrotllm.training - step 12010 | epoch 0 | loss 3.5979 | lr 1.47e-03 | grad 0.1961 +2026-04-10 05:47:35 - INFO - parrotllm.training - step 12020 | epoch 0 | loss 3.5754 | lr 1.47e-03 | grad 0.1951 +2026-04-10 05:47:38 - INFO - parrotllm.training - step 12030 | epoch 0 | loss 3.5011 | lr 1.47e-03 | grad 0.1891 +2026-04-10 05:47:41 - INFO - parrotllm.training - step 12040 | epoch 0 | loss 3.4217 | lr 1.47e-03 | grad 0.1795 +2026-04-10 05:47:44 - INFO - parrotllm.training - step 12050 | epoch 0 | loss 3.5254 | lr 1.47e-03 | grad 0.1870 +2026-04-10 05:47:47 - INFO - parrotllm.training - step 12060 | epoch 0 | loss 3.6679 | lr 1.47e-03 | grad 0.2005 +2026-04-10 05:47:50 - INFO - parrotllm.training - step 12070 | epoch 0 | loss 3.4791 | lr 1.47e-03 | grad 0.2075 +2026-04-10 05:47:53 - INFO - parrotllm.training - step 12080 | epoch 0 | loss 3.4469 | lr 1.47e-03 | grad 0.2006 +2026-04-10 05:47:56 - INFO - parrotllm.training - step 12090 | epoch 0 | loss 3.4942 | lr 1.47e-03 | grad 0.2025 +2026-04-10 05:47:59 - INFO - parrotllm.training - step 12100 | epoch 0 | loss 3.4751 | lr 1.47e-03 | grad 0.1892 +2026-04-10 05:48:03 - INFO - parrotllm.training - step 12110 | epoch 0 | loss 3.5346 | lr 1.47e-03 | grad 0.1855 +2026-04-10 05:48:06 - INFO - parrotllm.training - step 12120 | epoch 0 | loss 3.5075 | lr 1.47e-03 | grad 0.2004 +2026-04-10 05:48:09 - INFO - parrotllm.training - step 12130 | epoch 0 | loss 3.3764 | lr 1.47e-03 | grad 0.1975 +2026-04-10 05:48:12 - INFO - parrotllm.training - step 12140 | epoch 0 | loss 3.4260 | lr 1.47e-03 | grad 0.1891 +2026-04-10 05:48:15 - INFO - parrotllm.training - step 12150 | epoch 0 | loss 3.4894 | lr 1.47e-03 | grad 0.1756 +2026-04-10 05:48:18 - INFO - parrotllm.training - step 12160 | epoch 0 | loss 3.5562 | lr 1.47e-03 | grad 0.1883 +2026-04-10 05:48:21 - INFO - parrotllm.training - step 12170 | epoch 0 | loss 3.5476 | lr 1.47e-03 | grad 0.1985 +2026-04-10 05:48:24 - INFO - parrotllm.training - step 12180 | epoch 0 | loss 3.4534 | lr 1.47e-03 | grad 0.1793 +2026-04-10 05:48:27 - INFO - parrotllm.training - step 12190 | epoch 0 | loss 3.5410 | lr 1.47e-03 | grad 0.1978 +2026-04-10 05:48:30 - INFO - parrotllm.training - step 12200 | epoch 0 | loss 3.5275 | lr 1.47e-03 | grad 0.1988 +2026-04-10 05:48:33 - INFO - parrotllm.training - step 12210 | epoch 0 | loss 3.5781 | lr 1.47e-03 | grad 0.1739 +2026-04-10 05:48:36 - INFO - parrotllm.training - step 12220 | epoch 0 | loss 3.5087 | lr 1.47e-03 | grad 0.1940 +2026-04-10 05:48:39 - INFO - parrotllm.training - step 12230 | epoch 0 | loss 3.4802 | lr 1.47e-03 | grad 0.2007 +2026-04-10 05:48:42 - INFO - parrotllm.training - step 12240 | epoch 0 | loss 3.4925 | lr 1.47e-03 | grad 0.1886 +2026-04-10 05:48:45 - INFO - parrotllm.training - step 12250 | epoch 0 | loss 3.3999 | lr 1.47e-03 | grad 0.1794 +2026-04-10 05:48:48 - INFO - parrotllm.training - step 12260 | epoch 0 | loss 3.4277 | lr 1.47e-03 | grad 0.1811 +2026-04-10 05:48:51 - INFO - parrotllm.training - step 12270 | epoch 0 | loss 3.5096 | lr 1.47e-03 | grad 0.1922 +2026-04-10 05:48:54 - INFO - parrotllm.training - step 12280 | epoch 0 | loss 3.5492 | lr 1.47e-03 | grad 0.1985 +2026-04-10 05:48:57 - INFO - parrotllm.training - step 12290 | epoch 0 | loss 3.6039 | lr 1.47e-03 | grad 0.2261 +2026-04-10 05:49:01 - INFO - parrotllm.training - step 12300 | epoch 0 | loss 3.5884 | lr 1.47e-03 | grad 0.2010 +2026-04-10 05:49:04 - INFO - parrotllm.training - step 12310 | epoch 0 | loss 3.4456 | lr 1.47e-03 | grad 0.2200 +2026-04-10 05:49:07 - INFO - parrotllm.training - step 12320 | epoch 0 | loss 3.4499 | lr 1.47e-03 | grad 0.1635 +2026-04-10 05:49:10 - INFO - parrotllm.training - step 12330 | epoch 0 | loss 3.4833 | lr 1.47e-03 | grad 0.1834 +2026-04-10 05:49:13 - INFO - parrotllm.training - step 12340 | epoch 0 | loss 3.6279 | lr 1.47e-03 | grad 0.1861 +2026-04-10 05:49:16 - INFO - parrotllm.training - step 12350 | epoch 0 | loss 3.4309 | lr 1.47e-03 | grad 0.2017 +2026-04-10 05:49:19 - INFO - parrotllm.training - step 12360 | epoch 0 | loss 3.5275 | lr 1.47e-03 | grad 0.1970 +2026-04-10 05:49:22 - INFO - parrotllm.training - step 12370 | epoch 0 | loss 3.5682 | lr 1.47e-03 | grad 0.1818 +2026-04-10 05:49:25 - INFO - parrotllm.training - step 12380 | epoch 0 | loss 3.4183 | lr 1.47e-03 | grad 0.2158 +2026-04-10 05:49:28 - INFO - parrotllm.training - step 12390 | epoch 0 | loss 3.5442 | lr 1.47e-03 | grad 0.1819 +2026-04-10 05:49:31 - INFO - parrotllm.training - step 12400 | epoch 0 | loss 3.5471 | lr 1.47e-03 | grad 0.1934 +2026-04-10 05:49:34 - INFO - parrotllm.training - step 12410 | epoch 0 | loss 3.5480 | lr 1.47e-03 | grad 0.2064 +2026-04-10 05:49:37 - INFO - parrotllm.training - step 12420 | epoch 0 | loss 3.5136 | lr 1.47e-03 | grad 0.1761 +2026-04-10 05:49:40 - INFO - parrotllm.training - step 12430 | epoch 0 | loss 3.5708 | lr 1.47e-03 | grad 0.1770 +2026-04-10 05:49:43 - INFO - parrotllm.training - step 12440 | epoch 0 | loss 3.5494 | lr 1.47e-03 | grad 0.2032 +2026-04-10 05:49:46 - INFO - parrotllm.training - step 12450 | epoch 0 | loss 3.4543 | lr 1.47e-03 | grad 0.1854 +2026-04-10 05:49:49 - INFO - parrotllm.training - step 12460 | epoch 0 | loss 3.4102 | lr 1.47e-03 | grad 0.1921 +2026-04-10 05:49:52 - INFO - parrotllm.training - step 12470 | epoch 0 | loss 3.5392 | lr 1.47e-03 | grad 0.2394 +2026-04-10 05:49:55 - INFO - parrotllm.training - step 12480 | epoch 0 | loss 3.4760 | lr 1.47e-03 | grad 0.1980 +2026-04-10 05:49:58 - INFO - parrotllm.training - step 12490 | epoch 0 | loss 3.5571 | lr 1.47e-03 | grad 0.1771 +2026-04-10 05:50:02 - INFO - parrotllm.training - step 12500 | epoch 0 | loss 3.4969 | lr 1.47e-03 | grad 0.1844 +2026-04-10 05:50:02 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:50:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:50:05 - INFO - parrotllm.training - Train: loss=3.4969, ppl=33.01 +2026-04-10 05:50:05 - INFO - parrotllm.training - Val: loss=3.4590, ppl=31.79 +2026-04-10 05:50:05 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:50:06 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4590_epoch_0000_step_0012500.pt +2026-04-10 05:50:07 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:50:08 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0012500.pt +2026-04-10 05:50:11 - INFO - parrotllm.training - step 12510 | epoch 0 | loss 3.5498 | lr 1.47e-03 | grad 0.1886 +2026-04-10 05:50:15 - INFO - parrotllm.training - step 12520 | epoch 0 | loss 3.4615 | lr 1.47e-03 | grad 0.2533 +2026-04-10 05:50:18 - INFO - parrotllm.training - step 12530 | epoch 0 | loss 3.5824 | lr 1.47e-03 | grad 0.1702 +2026-04-10 05:50:21 - INFO - parrotllm.training - step 12540 | epoch 0 | loss 3.5113 | lr 1.47e-03 | grad 0.1819 +2026-04-10 05:50:24 - INFO - parrotllm.training - step 12550 | epoch 0 | loss 3.4286 | lr 1.47e-03 | grad 0.1966 +2026-04-10 05:50:27 - INFO - parrotllm.training - step 12560 | epoch 0 | loss 3.5817 | lr 1.47e-03 | grad 0.1931 +2026-04-10 05:50:30 - INFO - parrotllm.training - step 12570 | epoch 0 | loss 3.5659 | lr 1.47e-03 | grad 0.1798 +2026-04-10 05:50:33 - INFO - parrotllm.training - step 12580 | epoch 0 | loss 3.5794 | lr 1.47e-03 | grad 0.1948 +2026-04-10 05:50:36 - INFO - parrotllm.training - step 12590 | epoch 0 | loss 3.5656 | lr 1.47e-03 | grad 0.2281 +2026-04-10 05:50:39 - INFO - parrotllm.training - step 12600 | epoch 0 | loss 3.3677 | lr 1.47e-03 | grad 0.2222 +2026-04-10 05:50:42 - INFO - parrotllm.training - step 12610 | epoch 0 | loss 3.5603 | lr 1.47e-03 | grad 0.1905 +2026-04-10 05:50:46 - INFO - parrotllm.training - step 12620 | epoch 0 | loss 3.5396 | lr 1.47e-03 | grad 0.1820 +2026-04-10 05:50:49 - INFO - parrotllm.training - step 12630 | epoch 0 | loss 3.4532 | lr 1.47e-03 | grad 0.1988 +2026-04-10 05:50:52 - INFO - parrotllm.training - step 12640 | epoch 0 | loss 3.5650 | lr 1.47e-03 | grad 0.1805 +2026-04-10 05:50:55 - INFO - parrotllm.training - step 12650 | epoch 0 | loss 3.4996 | lr 1.47e-03 | grad 0.2021 +2026-04-10 05:50:58 - INFO - parrotllm.training - step 12660 | epoch 0 | loss 3.5521 | lr 1.47e-03 | grad 0.1919 +2026-04-10 05:51:01 - INFO - parrotllm.training - step 12670 | epoch 0 | loss 3.5622 | lr 1.47e-03 | grad 0.2022 +2026-04-10 05:51:04 - INFO - parrotllm.training - step 12680 | epoch 0 | loss 3.4707 | lr 1.47e-03 | grad 0.1949 +2026-04-10 05:51:07 - INFO - parrotllm.training - step 12690 | epoch 0 | loss 3.6138 | lr 1.47e-03 | grad 0.1684 +2026-04-10 05:51:10 - INFO - parrotllm.training - step 12700 | epoch 0 | loss 3.5266 | lr 1.47e-03 | grad 0.2098 +2026-04-10 05:51:13 - INFO - parrotllm.training - step 12710 | epoch 0 | loss 3.4984 | lr 1.47e-03 | grad 0.2096 +2026-04-10 05:51:16 - INFO - parrotllm.training - step 12720 | epoch 0 | loss 3.4155 | lr 1.47e-03 | grad 0.2069 +2026-04-10 05:51:19 - INFO - parrotllm.training - step 12730 | epoch 0 | loss 3.4111 | lr 1.47e-03 | grad 0.1736 +2026-04-10 05:51:22 - INFO - parrotllm.training - step 12740 | epoch 0 | loss 3.4814 | lr 1.47e-03 | grad 0.2075 +2026-04-10 05:51:25 - INFO - parrotllm.training - step 12750 | epoch 0 | loss 3.5674 | lr 1.47e-03 | grad 0.2129 +2026-04-10 05:51:28 - INFO - parrotllm.training - step 12760 | epoch 0 | loss 3.5418 | lr 1.47e-03 | grad 0.2323 +2026-04-10 05:51:31 - INFO - parrotllm.training - step 12770 | epoch 0 | loss 3.4652 | lr 1.47e-03 | grad 0.1787 +2026-04-10 05:51:34 - INFO - parrotllm.training - step 12780 | epoch 0 | loss 3.5256 | lr 1.47e-03 | grad 0.1821 +2026-04-10 05:51:37 - INFO - parrotllm.training - step 12790 | epoch 0 | loss 3.4913 | lr 1.47e-03 | grad 0.1962 +2026-04-10 05:51:41 - INFO - parrotllm.training - step 12800 | epoch 0 | loss 3.4312 | lr 1.47e-03 | grad 0.1751 +2026-04-10 05:51:44 - INFO - parrotllm.training - step 12810 | epoch 0 | loss 3.5342 | lr 1.47e-03 | grad 0.1840 +2026-04-10 05:51:47 - INFO - parrotllm.training - step 12820 | epoch 0 | loss 3.4938 | lr 1.47e-03 | grad 0.2119 +2026-04-10 05:51:50 - INFO - parrotllm.training - step 12830 | epoch 0 | loss 3.6490 | lr 1.47e-03 | grad 0.1941 +2026-04-10 05:51:53 - INFO - parrotllm.training - step 12840 | epoch 0 | loss 3.5208 | lr 1.47e-03 | grad 0.2260 +2026-04-10 05:51:56 - INFO - parrotllm.training - step 12850 | epoch 0 | loss 3.4719 | lr 1.47e-03 | grad 0.1893 +2026-04-10 05:51:59 - INFO - parrotllm.training - step 12860 | epoch 0 | loss 3.4715 | lr 1.47e-03 | grad 0.2027 +2026-04-10 05:52:02 - INFO - parrotllm.training - step 12870 | epoch 0 | loss 3.5333 | lr 1.47e-03 | grad 0.2430 +2026-04-10 05:52:05 - INFO - parrotllm.training - step 12880 | epoch 0 | loss 3.6563 | lr 1.47e-03 | grad 0.2383 +2026-04-10 05:52:08 - INFO - parrotllm.training - step 12890 | epoch 0 | loss 3.5065 | lr 1.47e-03 | grad 0.1739 +2026-04-10 05:52:11 - INFO - parrotllm.training - step 12900 | epoch 0 | loss 3.5568 | lr 1.47e-03 | grad 0.1960 +2026-04-10 05:52:14 - INFO - parrotllm.training - step 12910 | epoch 0 | loss 3.5284 | lr 1.47e-03 | grad 0.2052 +2026-04-10 05:52:17 - INFO - parrotllm.training - step 12920 | epoch 0 | loss 3.5082 | lr 1.47e-03 | grad 0.1877 +2026-04-10 05:52:20 - INFO - parrotllm.training - step 12930 | epoch 0 | loss 3.4517 | lr 1.47e-03 | grad 0.2310 +2026-04-10 05:52:23 - INFO - parrotllm.training - step 12940 | epoch 0 | loss 3.6846 | lr 1.47e-03 | grad 0.1947 +2026-04-10 05:52:26 - INFO - parrotllm.training - step 12950 | epoch 0 | loss 3.5105 | lr 1.47e-03 | grad 0.2008 +2026-04-10 05:52:29 - INFO - parrotllm.training - step 12960 | epoch 0 | loss 3.6471 | lr 1.47e-03 | grad 0.1878 +2026-04-10 05:52:32 - INFO - parrotllm.training - step 12970 | epoch 0 | loss 3.6385 | lr 1.47e-03 | grad 0.1730 +2026-04-10 05:52:35 - INFO - parrotllm.training - step 12980 | epoch 0 | loss 3.5481 | lr 1.47e-03 | grad 0.2383 +2026-04-10 05:52:38 - INFO - parrotllm.training - step 12990 | epoch 0 | loss 3.4686 | lr 1.47e-03 | grad 0.1974 +2026-04-10 05:52:42 - INFO - parrotllm.training - step 13000 | epoch 0 | loss 3.5588 | lr 1.47e-03 | grad 0.1757 +2026-04-10 05:52:42 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:52:42 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:52:45 - INFO - parrotllm.training - Train: loss=3.5588, ppl=35.12 +2026-04-10 05:52:45 - INFO - parrotllm.training - Val: loss=3.4560, ppl=31.69 +2026-04-10 05:52:45 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:52:46 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4560_epoch_0000_step_0013000.pt +2026-04-10 05:52:47 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:52:50 - INFO - parrotllm.training - step 13010 | epoch 0 | loss 3.5508 | lr 1.47e-03 | grad 0.1983 +2026-04-10 05:52:53 - INFO - parrotllm.training - step 13020 | epoch 0 | loss 3.5845 | lr 1.47e-03 | grad 0.2053 +2026-04-10 05:52:56 - INFO - parrotllm.training - step 13030 | epoch 0 | loss 3.4695 | lr 1.47e-03 | grad 0.1947 +2026-04-10 05:52:59 - INFO - parrotllm.training - step 13040 | epoch 0 | loss 3.6053 | lr 1.47e-03 | grad 0.1879 +2026-04-10 05:53:02 - INFO - parrotllm.training - step 13050 | epoch 0 | loss 3.5092 | lr 1.47e-03 | grad 0.2085 +2026-04-10 05:53:05 - INFO - parrotllm.training - step 13060 | epoch 0 | loss 3.4940 | lr 1.47e-03 | grad 0.1798 +2026-04-10 05:53:08 - INFO - parrotllm.training - step 13070 | epoch 0 | loss 3.5570 | lr 1.47e-03 | grad 0.1854 +2026-04-10 05:53:11 - INFO - parrotllm.training - step 13080 | epoch 0 | loss 3.4840 | lr 1.47e-03 | grad 0.2130 +2026-04-10 05:53:14 - INFO - parrotllm.training - step 13090 | epoch 0 | loss 3.5648 | lr 1.47e-03 | grad 0.1791 +2026-04-10 05:53:17 - INFO - parrotllm.training - step 13100 | epoch 0 | loss 3.6019 | lr 1.47e-03 | grad 0.2176 +2026-04-10 05:53:20 - INFO - parrotllm.training - step 13110 | epoch 0 | loss 3.5831 | lr 1.47e-03 | grad 0.1807 +2026-04-10 05:53:23 - INFO - parrotllm.training - step 13120 | epoch 0 | loss 3.4416 | lr 1.47e-03 | grad 0.1727 +2026-04-10 05:53:26 - INFO - parrotllm.training - step 13130 | epoch 0 | loss 3.5601 | lr 1.47e-03 | grad 0.1930 +2026-04-10 05:53:29 - INFO - parrotllm.training - step 13140 | epoch 0 | loss 3.5487 | lr 1.47e-03 | grad 0.1845 +2026-04-10 05:53:32 - INFO - parrotllm.training - step 13150 | epoch 0 | loss 3.5375 | lr 1.47e-03 | grad 0.1735 +2026-04-10 05:53:35 - INFO - parrotllm.training - step 13160 | epoch 0 | loss 3.5776 | lr 1.47e-03 | grad 0.2095 +2026-04-10 05:53:38 - INFO - parrotllm.training - step 13170 | epoch 0 | loss 3.4914 | lr 1.47e-03 | grad 0.1977 +2026-04-10 05:53:41 - INFO - parrotllm.training - step 13180 | epoch 0 | loss 3.5166 | lr 1.47e-03 | grad 0.1889 +2026-04-10 05:53:44 - INFO - parrotllm.training - step 13190 | epoch 0 | loss 3.6329 | lr 1.47e-03 | grad 0.1967 +2026-04-10 05:53:47 - INFO - parrotllm.training - step 13200 | epoch 0 | loss 3.4050 | lr 1.47e-03 | grad 0.1814 +2026-04-10 05:53:51 - INFO - parrotllm.training - step 13210 | epoch 0 | loss 3.4998 | lr 1.47e-03 | grad 0.2084 +2026-04-10 05:53:54 - INFO - parrotllm.training - step 13220 | epoch 0 | loss 3.5843 | lr 1.47e-03 | grad 0.1984 +2026-04-10 05:53:54 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 13225/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 05:53:57 - INFO - parrotllm.training - step 13230 | epoch 0 | loss 3.4167 | lr 1.47e-03 | grad 0.2273 +2026-04-10 05:54:00 - INFO - parrotllm.training - step 13240 | epoch 0 | loss 3.5421 | lr 1.47e-03 | grad 0.2035 +2026-04-10 05:54:03 - INFO - parrotllm.training - step 13250 | epoch 0 | loss 3.5172 | lr 1.47e-03 | grad 0.1854 +2026-04-10 05:54:06 - INFO - parrotllm.training - step 13260 | epoch 0 | loss 3.6400 | lr 1.47e-03 | grad 0.2393 +2026-04-10 05:54:09 - INFO - parrotllm.training - step 13270 | epoch 0 | loss 3.4412 | lr 1.47e-03 | grad 0.1815 +2026-04-10 05:54:12 - INFO - parrotllm.training - step 13280 | epoch 0 | loss 3.5062 | lr 1.47e-03 | grad 0.1854 +2026-04-10 05:54:15 - INFO - parrotllm.training - step 13290 | epoch 0 | loss 3.6217 | lr 1.47e-03 | grad 0.2069 +2026-04-10 05:54:18 - INFO - parrotllm.training - step 13300 | epoch 0 | loss 3.5384 | lr 1.47e-03 | grad 0.1984 +2026-04-10 05:54:21 - INFO - parrotllm.training - step 13310 | epoch 0 | loss 3.5907 | lr 1.47e-03 | grad 0.2033 +2026-04-10 05:54:24 - INFO - parrotllm.training - step 13320 | epoch 0 | loss 3.5162 | lr 1.47e-03 | grad 0.2026 +2026-04-10 05:54:27 - INFO - parrotllm.training - step 13330 | epoch 0 | loss 3.4862 | lr 1.47e-03 | grad 0.1990 +2026-04-10 05:54:30 - INFO - parrotllm.training - step 13340 | epoch 0 | loss 3.4630 | lr 1.47e-03 | grad 0.2012 +2026-04-10 05:54:33 - INFO - parrotllm.training - step 13350 | epoch 0 | loss 3.5664 | lr 1.47e-03 | grad 0.2031 +2026-04-10 05:54:37 - INFO - parrotllm.training - step 13360 | epoch 0 | loss 3.5487 | lr 1.47e-03 | grad 0.2108 +2026-04-10 05:54:40 - INFO - parrotllm.training - step 13370 | epoch 0 | loss 3.3963 | lr 1.47e-03 | grad 0.1958 +2026-04-10 05:54:43 - INFO - parrotllm.training - step 13380 | epoch 0 | loss 3.5717 | lr 1.47e-03 | grad 0.2087 +2026-04-10 05:54:46 - INFO - parrotllm.training - step 13390 | epoch 0 | loss 3.5166 | lr 1.47e-03 | grad 0.2118 +2026-04-10 05:54:49 - INFO - parrotllm.training - step 13400 | epoch 0 | loss 3.4990 | lr 1.47e-03 | grad 0.2060 +2026-04-10 05:54:52 - INFO - parrotllm.training - step 13410 | epoch 0 | loss 3.5582 | lr 1.47e-03 | grad 0.1785 +2026-04-10 05:54:55 - INFO - parrotllm.training - step 13420 | epoch 0 | loss 3.5262 | lr 1.47e-03 | grad 0.2183 +2026-04-10 05:54:58 - INFO - parrotllm.training - step 13430 | epoch 0 | loss 3.4974 | lr 1.47e-03 | grad 0.1962 +2026-04-10 05:55:01 - INFO - parrotllm.training - step 13440 | epoch 0 | loss 3.4767 | lr 1.47e-03 | grad 0.1872 +2026-04-10 05:55:04 - INFO - parrotllm.training - step 13450 | epoch 0 | loss 3.4766 | lr 1.47e-03 | grad 0.1843 +2026-04-10 05:55:07 - INFO - parrotllm.training - step 13460 | epoch 0 | loss 3.5476 | lr 1.47e-03 | grad 0.1921 +2026-04-10 05:55:10 - INFO - parrotllm.training - step 13470 | epoch 0 | loss 3.5287 | lr 1.47e-03 | grad 0.1705 +2026-04-10 05:55:13 - INFO - parrotllm.training - step 13480 | epoch 0 | loss 3.5525 | lr 1.47e-03 | grad 0.1844 +2026-04-10 05:55:16 - INFO - parrotllm.training - step 13490 | epoch 0 | loss 3.4466 | lr 1.47e-03 | grad 0.2163 +2026-04-10 05:55:19 - INFO - parrotllm.training - step 13500 | epoch 0 | loss 3.5313 | lr 1.47e-03 | grad 0.1969 +2026-04-10 05:55:19 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:55:19 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:55:22 - INFO - parrotllm.training - Train: loss=3.5313, ppl=34.17 +2026-04-10 05:55:22 - INFO - parrotllm.training - Val: loss=3.4560, ppl=31.69 +2026-04-10 05:55:23 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4560_epoch_0000_step_0013500.pt +2026-04-10 05:55:24 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4560, min_delta=0.001000). +2026-04-10 05:55:24 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:55:27 - INFO - parrotllm.training - step 13510 | epoch 0 | loss 3.4727 | lr 1.47e-03 | grad 0.2256 +2026-04-10 05:55:30 - INFO - parrotllm.training - step 13520 | epoch 0 | loss 3.3966 | lr 1.47e-03 | grad 0.2184 +2026-04-10 05:55:33 - INFO - parrotllm.training - step 13530 | epoch 0 | loss 3.5150 | lr 1.47e-03 | grad 0.1797 +2026-04-10 05:55:37 - INFO - parrotllm.training - step 13540 | epoch 0 | loss 3.3777 | lr 1.47e-03 | grad 0.1645 +2026-04-10 05:55:40 - INFO - parrotllm.training - step 13550 | epoch 0 | loss 3.5528 | lr 1.47e-03 | grad 0.1946 +2026-04-10 05:55:43 - INFO - parrotllm.training - step 13560 | epoch 0 | loss 3.5603 | lr 1.47e-03 | grad 0.1843 +2026-04-10 05:55:46 - INFO - parrotllm.training - step 13570 | epoch 0 | loss 3.5632 | lr 1.47e-03 | grad 0.1990 +2026-04-10 05:55:49 - INFO - parrotllm.training - step 13580 | epoch 0 | loss 3.4969 | lr 1.47e-03 | grad 0.2053 +2026-04-10 05:55:52 - INFO - parrotllm.training - step 13590 | epoch 0 | loss 3.5619 | lr 1.47e-03 | grad 0.1992 +2026-04-10 05:55:55 - INFO - parrotllm.training - step 13600 | epoch 0 | loss 3.4031 | lr 1.47e-03 | grad 0.1756 +2026-04-10 05:55:58 - INFO - parrotllm.training - step 13610 | epoch 0 | loss 3.4680 | lr 1.47e-03 | grad 0.1848 +2026-04-10 05:56:01 - INFO - parrotllm.training - step 13620 | epoch 0 | loss 3.5419 | lr 1.47e-03 | grad 0.1969 +2026-04-10 05:56:04 - INFO - parrotllm.training - step 13630 | epoch 0 | loss 3.4806 | lr 1.47e-03 | grad 0.1720 +2026-04-10 05:56:07 - INFO - parrotllm.training - step 13640 | epoch 0 | loss 3.4824 | lr 1.47e-03 | grad 0.2029 +2026-04-10 05:56:10 - INFO - parrotllm.training - step 13650 | epoch 0 | loss 3.5315 | lr 1.47e-03 | grad 0.1968 +2026-04-10 05:56:13 - INFO - parrotllm.training - step 13660 | epoch 0 | loss 3.5600 | lr 1.47e-03 | grad 0.2105 +2026-04-10 05:56:16 - INFO - parrotllm.training - step 13670 | epoch 0 | loss 3.5764 | lr 1.47e-03 | grad 0.1967 +2026-04-10 05:56:19 - INFO - parrotllm.training - step 13680 | epoch 0 | loss 3.4701 | lr 1.47e-03 | grad 0.1997 +2026-04-10 05:56:22 - INFO - parrotllm.training - step 13690 | epoch 0 | loss 3.5471 | lr 1.47e-03 | grad 0.2151 +2026-04-10 05:56:25 - INFO - parrotllm.training - step 13700 | epoch 0 | loss 3.5734 | lr 1.47e-03 | grad 0.1878 +2026-04-10 05:56:28 - INFO - parrotllm.training - step 13710 | epoch 0 | loss 3.5337 | lr 1.47e-03 | grad 0.2004 +2026-04-10 05:56:31 - INFO - parrotllm.training - step 13720 | epoch 0 | loss 3.4992 | lr 1.47e-03 | grad 0.1779 +2026-04-10 05:56:34 - INFO - parrotllm.training - step 13730 | epoch 0 | loss 3.4932 | lr 1.47e-03 | grad 0.1880 +2026-04-10 05:56:37 - INFO - parrotllm.training - step 13740 | epoch 0 | loss 3.3536 | lr 1.47e-03 | grad 0.2062 +2026-04-10 05:56:40 - INFO - parrotllm.training - step 13750 | epoch 0 | loss 3.4457 | lr 1.47e-03 | grad 0.2309 +2026-04-10 05:56:43 - INFO - parrotllm.training - step 13760 | epoch 0 | loss 3.5157 | lr 1.47e-03 | grad 0.1930 +2026-04-10 05:56:47 - INFO - parrotllm.training - step 13770 | epoch 0 | loss 3.5419 | lr 1.47e-03 | grad 0.1843 +2026-04-10 05:56:50 - INFO - parrotllm.training - step 13780 | epoch 0 | loss 3.5164 | lr 1.47e-03 | grad 0.1927 +2026-04-10 05:56:53 - INFO - parrotllm.training - step 13790 | epoch 0 | loss 3.4786 | lr 1.47e-03 | grad 0.1724 +2026-04-10 05:56:56 - INFO - parrotllm.training - step 13800 | epoch 0 | loss 3.4567 | lr 1.47e-03 | grad 0.1641 +2026-04-10 05:56:59 - INFO - parrotllm.training - step 13810 | epoch 0 | loss 3.5558 | lr 1.47e-03 | grad 0.1889 +2026-04-10 05:57:02 - INFO - parrotllm.training - step 13820 | epoch 0 | loss 3.5035 | lr 1.47e-03 | grad 0.1764 +2026-04-10 05:57:05 - INFO - parrotllm.training - step 13830 | epoch 0 | loss 3.4936 | lr 1.47e-03 | grad 0.1921 +2026-04-10 05:57:08 - INFO - parrotllm.training - step 13840 | epoch 0 | loss 3.6406 | lr 1.47e-03 | grad 0.2011 +2026-04-10 05:57:11 - INFO - parrotllm.training - step 13850 | epoch 0 | loss 3.5941 | lr 1.47e-03 | grad 0.2323 +2026-04-10 05:57:14 - INFO - parrotllm.training - step 13860 | epoch 0 | loss 3.4978 | lr 1.47e-03 | grad 0.1785 +2026-04-10 05:57:17 - INFO - parrotllm.training - step 13870 | epoch 0 | loss 3.4515 | lr 1.47e-03 | grad 0.1755 +2026-04-10 05:57:20 - INFO - parrotllm.training - step 13880 | epoch 0 | loss 3.5558 | lr 1.47e-03 | grad 0.1779 +2026-04-10 05:57:23 - INFO - parrotllm.training - step 13890 | epoch 0 | loss 3.5725 | lr 1.47e-03 | grad 0.1885 +2026-04-10 05:57:26 - INFO - parrotllm.training - step 13900 | epoch 0 | loss 3.5255 | lr 1.47e-03 | grad 0.1855 +2026-04-10 05:57:29 - INFO - parrotllm.training - step 13910 | epoch 0 | loss 3.5335 | lr 1.47e-03 | grad 0.1893 +2026-04-10 05:57:32 - INFO - parrotllm.training - step 13920 | epoch 0 | loss 3.5441 | lr 1.46e-03 | grad 0.2156 +2026-04-10 05:57:35 - INFO - parrotllm.training - step 13930 | epoch 0 | loss 3.5319 | lr 1.46e-03 | grad 0.2056 +2026-04-10 05:57:38 - INFO - parrotllm.training - step 13940 | epoch 0 | loss 3.5363 | lr 1.46e-03 | grad 0.1738 +2026-04-10 05:57:41 - INFO - parrotllm.training - step 13950 | epoch 0 | loss 3.5274 | lr 1.46e-03 | grad 0.2033 +2026-04-10 05:57:44 - INFO - parrotllm.training - step 13960 | epoch 0 | loss 3.4924 | lr 1.46e-03 | grad 0.1727 +2026-04-10 05:57:47 - INFO - parrotllm.training - step 13970 | epoch 0 | loss 3.5190 | lr 1.46e-03 | grad 0.1777 +2026-04-10 05:57:50 - INFO - parrotllm.training - step 13980 | epoch 0 | loss 3.5546 | lr 1.46e-03 | grad 0.1969 +2026-04-10 05:57:54 - INFO - parrotllm.training - step 13990 | epoch 0 | loss 3.5329 | lr 1.46e-03 | grad 0.2016 +2026-04-10 05:57:57 - INFO - parrotllm.training - step 14000 | epoch 0 | loss 3.5296 | lr 1.46e-03 | grad 0.1776 +2026-04-10 05:57:57 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 05:57:57 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:58:00 - INFO - parrotllm.training - Train: loss=3.5296, ppl=34.11 +2026-04-10 05:58:00 - INFO - parrotllm.training - Val: loss=3.4515, ppl=31.55 +2026-04-10 05:58:00 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 05:58:01 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4515_epoch_0000_step_0014000.pt +2026-04-10 05:58:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 05:58:05 - INFO - parrotllm.training - step 14010 | epoch 0 | loss 3.4786 | lr 1.46e-03 | grad 0.2055 +2026-04-10 05:58:08 - INFO - parrotllm.training - step 14020 | epoch 0 | loss 3.5140 | lr 1.46e-03 | grad 0.1792 +2026-04-10 05:58:11 - INFO - parrotllm.training - step 14030 | epoch 0 | loss 3.4463 | lr 1.46e-03 | grad 0.1865 +2026-04-10 05:58:14 - INFO - parrotllm.training - step 14040 | epoch 0 | loss 3.5764 | lr 1.46e-03 | grad 0.2059 +2026-04-10 05:58:17 - INFO - parrotllm.training - step 14050 | epoch 0 | loss 3.4620 | lr 1.46e-03 | grad 0.1788 +2026-04-10 05:58:20 - INFO - parrotllm.training - step 14060 | epoch 0 | loss 3.5400 | lr 1.46e-03 | grad 0.1837 +2026-04-10 05:58:23 - INFO - parrotllm.training - step 14070 | epoch 0 | loss 3.4267 | lr 1.46e-03 | grad 0.1932 +2026-04-10 05:58:26 - INFO - parrotllm.training - step 14080 | epoch 0 | loss 3.5182 | lr 1.46e-03 | grad 0.1934 +2026-04-10 05:58:29 - INFO - parrotllm.training - step 14090 | epoch 0 | loss 3.5327 | lr 1.46e-03 | grad 0.2105 +2026-04-10 05:58:32 - INFO - parrotllm.training - step 14100 | epoch 0 | loss 3.4832 | lr 1.46e-03 | grad 0.2221 +2026-04-10 05:58:35 - INFO - parrotllm.training - step 14110 | epoch 0 | loss 3.5975 | lr 1.46e-03 | grad 0.1900 +2026-04-10 05:58:38 - INFO - parrotllm.training - step 14120 | epoch 0 | loss 3.5262 | lr 1.46e-03 | grad 0.1942 +2026-04-10 05:58:41 - INFO - parrotllm.training - step 14130 | epoch 0 | loss 3.4439 | lr 1.46e-03 | grad 0.2136 +2026-04-10 05:58:44 - INFO - parrotllm.training - step 14140 | epoch 0 | loss 3.5204 | lr 1.46e-03 | grad 0.1894 +2026-04-10 05:58:47 - INFO - parrotllm.training - step 14150 | epoch 0 | loss 3.5604 | lr 1.46e-03 | grad 0.1960 +2026-04-10 05:58:50 - INFO - parrotllm.training - step 14160 | epoch 0 | loss 3.4118 | lr 1.46e-03 | grad 0.2173 +2026-04-10 05:58:53 - INFO - parrotllm.training - step 14170 | epoch 0 | loss 3.5078 | lr 1.46e-03 | grad 0.1738 +2026-04-10 05:58:56 - INFO - parrotllm.training - step 14180 | epoch 0 | loss 3.4054 | lr 1.46e-03 | grad 0.2088 +2026-04-10 05:58:59 - INFO - parrotllm.training - step 14190 | epoch 0 | loss 3.5186 | lr 1.46e-03 | grad 0.2039 +2026-04-10 05:59:02 - INFO - parrotllm.training - step 14200 | epoch 0 | loss 3.5855 | lr 1.46e-03 | grad 0.1765 +2026-04-10 05:59:06 - INFO - parrotllm.training - step 14210 | epoch 0 | loss 3.6114 | lr 1.46e-03 | grad 0.1967 +2026-04-10 05:59:09 - INFO - parrotllm.training - step 14220 | epoch 0 | loss 3.5104 | lr 1.46e-03 | grad 0.2078 +2026-04-10 05:59:12 - INFO - parrotllm.training - step 14230 | epoch 0 | loss 3.5562 | lr 1.46e-03 | grad 0.2390 +2026-04-10 05:59:15 - INFO - parrotllm.training - step 14240 | epoch 0 | loss 3.4349 | lr 1.46e-03 | grad 0.1989 +2026-04-10 05:59:18 - INFO - parrotllm.training - step 14250 | epoch 0 | loss 3.4421 | lr 1.46e-03 | grad 0.2311 +2026-04-10 05:59:21 - INFO - parrotllm.training - step 14260 | epoch 0 | loss 3.5894 | lr 1.46e-03 | grad 0.1752 +2026-04-10 05:59:24 - INFO - parrotllm.training - step 14270 | epoch 0 | loss 3.5072 | lr 1.46e-03 | grad 0.1881 +2026-04-10 05:59:27 - INFO - parrotllm.training - step 14280 | epoch 0 | loss 3.4700 | lr 1.46e-03 | grad 0.2019 +2026-04-10 05:59:30 - INFO - parrotllm.training - step 14290 | epoch 0 | loss 3.5650 | lr 1.46e-03 | grad 0.2002 +2026-04-10 05:59:33 - INFO - parrotllm.training - step 14300 | epoch 0 | loss 3.4966 | lr 1.46e-03 | grad 0.1877 +2026-04-10 05:59:36 - INFO - parrotllm.training - step 14310 | epoch 0 | loss 3.6141 | lr 1.46e-03 | grad 0.1914 +2026-04-10 05:59:39 - INFO - parrotllm.training - step 14320 | epoch 0 | loss 3.5773 | lr 1.46e-03 | grad 0.2020 +2026-04-10 05:59:42 - INFO - parrotllm.training - step 14330 | epoch 0 | loss 3.5108 | lr 1.46e-03 | grad 0.1955 +2026-04-10 05:59:45 - INFO - parrotllm.training - step 14340 | epoch 0 | loss 3.5068 | lr 1.46e-03 | grad 0.2171 +2026-04-10 05:59:48 - INFO - parrotllm.training - step 14350 | epoch 0 | loss 3.5364 | lr 1.46e-03 | grad 0.1817 +2026-04-10 05:59:51 - INFO - parrotllm.training - step 14360 | epoch 0 | loss 3.4859 | lr 1.46e-03 | grad 0.1936 +2026-04-10 05:59:54 - INFO - parrotllm.training - step 14370 | epoch 0 | loss 3.5268 | lr 1.46e-03 | grad 0.1862 +2026-04-10 05:59:57 - INFO - parrotllm.training - step 14380 | epoch 0 | loss 3.4826 | lr 1.46e-03 | grad 0.2064 +2026-04-10 06:00:00 - INFO - parrotllm.training - step 14390 | epoch 0 | loss 3.4863 | lr 1.46e-03 | grad 0.2133 +2026-04-10 06:00:03 - INFO - parrotllm.training - step 14400 | epoch 0 | loss 3.5554 | lr 1.46e-03 | grad 0.1951 +2026-04-10 06:00:07 - INFO - parrotllm.training - step 14410 | epoch 0 | loss 3.5145 | lr 1.46e-03 | grad 0.2024 +2026-04-10 06:00:10 - INFO - parrotllm.training - step 14420 | epoch 0 | loss 3.5636 | lr 1.46e-03 | grad 0.1815 +2026-04-10 06:00:13 - INFO - parrotllm.training - step 14430 | epoch 0 | loss 3.6068 | lr 1.46e-03 | grad 0.1766 +2026-04-10 06:00:16 - INFO - parrotllm.training - step 14440 | epoch 0 | loss 3.5847 | lr 1.46e-03 | grad 0.1816 +2026-04-10 06:00:19 - INFO - parrotllm.training - step 14450 | epoch 0 | loss 3.5622 | lr 1.46e-03 | grad 0.2050 +2026-04-10 06:00:22 - INFO - parrotllm.training - step 14460 | epoch 0 | loss 3.5720 | lr 1.46e-03 | grad 0.1909 +2026-04-10 06:00:25 - INFO - parrotllm.training - step 14470 | epoch 0 | loss 3.4462 | lr 1.46e-03 | grad 0.1916 +2026-04-10 06:00:28 - INFO - parrotllm.training - step 14480 | epoch 0 | loss 3.5062 | lr 1.46e-03 | grad 0.2058 +2026-04-10 06:00:31 - INFO - parrotllm.training - step 14490 | epoch 0 | loss 3.4980 | lr 1.46e-03 | grad 0.2085 +2026-04-10 06:00:34 - INFO - parrotllm.training - step 14500 | epoch 0 | loss 3.4158 | lr 1.46e-03 | grad 0.1930 +2026-04-10 06:00:34 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:00:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:00:37 - INFO - parrotllm.training - Train: loss=3.4158, ppl=30.44 +2026-04-10 06:00:37 - INFO - parrotllm.training - Val: loss=3.4498, ppl=31.49 +2026-04-10 06:00:37 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:00:38 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4498_epoch_0000_step_0014500.pt +2026-04-10 06:00:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:00:42 - INFO - parrotllm.training - step 14510 | epoch 0 | loss 3.4483 | lr 1.46e-03 | grad 0.1910 +2026-04-10 06:00:45 - INFO - parrotllm.training - step 14520 | epoch 0 | loss 3.5337 | lr 1.46e-03 | grad 0.2225 +2026-04-10 06:00:48 - INFO - parrotllm.training - step 14530 | epoch 0 | loss 3.4669 | lr 1.46e-03 | grad 0.2032 +2026-04-10 06:00:51 - INFO - parrotllm.training - step 14540 | epoch 0 | loss 3.5202 | lr 1.46e-03 | grad 0.2005 +2026-04-10 06:00:54 - INFO - parrotllm.training - step 14550 | epoch 0 | loss 3.4559 | lr 1.46e-03 | grad 0.1840 +2026-04-10 06:00:57 - INFO - parrotllm.training - step 14560 | epoch 0 | loss 3.6150 | lr 1.46e-03 | grad 0.1814 +2026-04-10 06:01:01 - INFO - parrotllm.training - step 14570 | epoch 0 | loss 3.5658 | lr 1.46e-03 | grad 0.2164 +2026-04-10 06:01:04 - INFO - parrotllm.training - step 14580 | epoch 0 | loss 3.5869 | lr 1.46e-03 | grad 0.1992 +2026-04-10 06:01:07 - INFO - parrotllm.training - step 14590 | epoch 0 | loss 3.4694 | lr 1.46e-03 | grad 0.2294 +2026-04-10 06:01:10 - INFO - parrotllm.training - step 14600 | epoch 0 | loss 3.4860 | lr 1.46e-03 | grad 0.2180 +2026-04-10 06:01:13 - INFO - parrotllm.training - step 14610 | epoch 0 | loss 3.5358 | lr 1.46e-03 | grad 0.2184 +2026-04-10 06:01:16 - INFO - parrotllm.training - step 14620 | epoch 0 | loss 3.4720 | lr 1.46e-03 | grad 0.1940 +2026-04-10 06:01:19 - INFO - parrotllm.training - step 14630 | epoch 0 | loss 3.5123 | lr 1.46e-03 | grad 0.2150 +2026-04-10 06:01:22 - INFO - parrotllm.training - step 14640 | epoch 0 | loss 3.5058 | lr 1.46e-03 | grad 0.1819 +2026-04-10 06:01:25 - INFO - parrotllm.training - step 14650 | epoch 0 | loss 3.5241 | lr 1.46e-03 | grad 0.1971 +2026-04-10 06:01:28 - INFO - parrotllm.training - step 14660 | epoch 0 | loss 3.5029 | lr 1.46e-03 | grad 0.1837 +2026-04-10 06:01:31 - INFO - parrotllm.training - step 14670 | epoch 0 | loss 3.5075 | lr 1.46e-03 | grad 0.2083 +2026-04-10 06:01:34 - INFO - parrotllm.training - step 14680 | epoch 0 | loss 3.5401 | lr 1.46e-03 | grad 0.2191 +2026-04-10 06:01:37 - INFO - parrotllm.training - step 14690 | epoch 0 | loss 3.4200 | lr 1.46e-03 | grad 0.1810 +2026-04-10 06:01:40 - INFO - parrotllm.training - step 14700 | epoch 0 | loss 3.6155 | lr 1.46e-03 | grad 0.2004 +2026-04-10 06:01:43 - INFO - parrotllm.training - step 14710 | epoch 0 | loss 3.5008 | lr 1.46e-03 | grad 0.2107 +2026-04-10 06:01:46 - INFO - parrotllm.training - step 14720 | epoch 0 | loss 3.6004 | lr 1.46e-03 | grad 0.2066 +2026-04-10 06:01:49 - INFO - parrotllm.training - step 14730 | epoch 0 | loss 3.5865 | lr 1.46e-03 | grad 0.2196 +2026-04-10 06:01:52 - INFO - parrotllm.training - step 14740 | epoch 0 | loss 3.4799 | lr 1.46e-03 | grad 0.2142 +2026-04-10 06:01:55 - INFO - parrotllm.training - step 14750 | epoch 0 | loss 3.4833 | lr 1.46e-03 | grad 0.1976 +2026-04-10 06:01:58 - INFO - parrotllm.training - step 14760 | epoch 0 | loss 3.5279 | lr 1.46e-03 | grad 0.2033 +2026-04-10 06:02:02 - INFO - parrotllm.training - step 14770 | epoch 0 | loss 3.4364 | lr 1.46e-03 | grad 0.2240 +2026-04-10 06:02:05 - INFO - parrotllm.training - step 14780 | epoch 0 | loss 3.5040 | lr 1.46e-03 | grad 0.1927 +2026-04-10 06:02:08 - INFO - parrotllm.training - step 14790 | epoch 0 | loss 3.4777 | lr 1.46e-03 | grad 0.2097 +2026-04-10 06:02:11 - INFO - parrotllm.training - step 14800 | epoch 0 | loss 3.3451 | lr 1.46e-03 | grad 0.2061 +2026-04-10 06:02:14 - INFO - parrotllm.training - step 14810 | epoch 0 | loss 3.4577 | lr 1.46e-03 | grad 0.2077 +2026-04-10 06:02:17 - INFO - parrotllm.training - step 14820 | epoch 0 | loss 3.4709 | lr 1.46e-03 | grad 0.2091 +2026-04-10 06:02:20 - INFO - parrotllm.training - step 14830 | epoch 0 | loss 3.4944 | lr 1.46e-03 | grad 0.1950 +2026-04-10 06:02:23 - INFO - parrotllm.training - step 14840 | epoch 0 | loss 3.4967 | lr 1.46e-03 | grad 0.2353 +2026-04-10 06:02:26 - INFO - parrotllm.training - step 14850 | epoch 0 | loss 3.4915 | lr 1.46e-03 | grad 0.2083 +2026-04-10 06:02:29 - INFO - parrotllm.training - step 14860 | epoch 0 | loss 3.4203 | lr 1.46e-03 | grad 0.1915 +2026-04-10 06:02:32 - INFO - parrotllm.training - step 14870 | epoch 0 | loss 3.4569 | lr 1.46e-03 | grad 0.2087 +2026-04-10 06:02:35 - INFO - parrotllm.training - step 14880 | epoch 0 | loss 3.4661 | lr 1.46e-03 | grad 0.1995 +2026-04-10 06:02:38 - INFO - parrotllm.training - step 14890 | epoch 0 | loss 3.4230 | lr 1.46e-03 | grad 0.1908 +2026-04-10 06:02:41 - INFO - parrotllm.training - step 14900 | epoch 0 | loss 3.4503 | lr 1.46e-03 | grad 0.1686 +2026-04-10 06:02:44 - INFO - parrotllm.training - step 14910 | epoch 0 | loss 3.4056 | lr 1.46e-03 | grad 0.1761 +2026-04-10 06:02:47 - INFO - parrotllm.training - step 14920 | epoch 0 | loss 3.5058 | lr 1.46e-03 | grad 0.2179 +2026-04-10 06:02:50 - INFO - parrotllm.training - step 14930 | epoch 0 | loss 3.5754 | lr 1.46e-03 | grad 0.2082 +2026-04-10 06:02:53 - INFO - parrotllm.training - step 14940 | epoch 0 | loss 3.5056 | lr 1.46e-03 | grad 0.2107 +2026-04-10 06:02:56 - INFO - parrotllm.training - step 14950 | epoch 0 | loss 3.4475 | lr 1.46e-03 | grad 0.1832 +2026-04-10 06:02:59 - INFO - parrotllm.training - step 14960 | epoch 0 | loss 3.5860 | lr 1.46e-03 | grad 0.1889 +2026-04-10 06:03:03 - INFO - parrotllm.training - step 14970 | epoch 0 | loss 3.5127 | lr 1.46e-03 | grad 0.1967 +2026-04-10 06:03:06 - INFO - parrotllm.training - step 14980 | epoch 0 | loss 3.5296 | lr 1.46e-03 | grad 0.1861 +2026-04-10 06:03:09 - INFO - parrotllm.training - step 14990 | epoch 0 | loss 3.4914 | lr 1.46e-03 | grad 0.1919 +2026-04-10 06:03:12 - INFO - parrotllm.training - step 15000 | epoch 0 | loss 3.4573 | lr 1.46e-03 | grad 0.1871 +2026-04-10 06:03:12 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:03:12 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:03:15 - INFO - parrotllm.training - Train: loss=3.4573, ppl=31.73 +2026-04-10 06:03:15 - INFO - parrotllm.training - Val: loss=3.4472, ppl=31.41 +2026-04-10 06:03:15 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:03:16 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4472_epoch_0000_step_0015000.pt +2026-04-10 06:03:17 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:03:18 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0015000.pt +2026-04-10 06:03:22 - INFO - parrotllm.training - step 15010 | epoch 0 | loss 3.4291 | lr 1.46e-03 | grad 0.1798 +2026-04-10 06:03:25 - INFO - parrotllm.training - step 15020 | epoch 0 | loss 3.5139 | lr 1.46e-03 | grad 0.2324 +2026-04-10 06:03:28 - INFO - parrotllm.training - step 15030 | epoch 0 | loss 3.5546 | lr 1.46e-03 | grad 0.2202 +2026-04-10 06:03:31 - INFO - parrotllm.training - step 15040 | epoch 0 | loss 3.5235 | lr 1.46e-03 | grad 0.1918 +2026-04-10 06:03:34 - INFO - parrotllm.training - step 15050 | epoch 0 | loss 3.4758 | lr 1.46e-03 | grad 0.1809 +2026-04-10 06:03:37 - INFO - parrotllm.training - step 15060 | epoch 0 | loss 3.3923 | lr 1.46e-03 | grad 0.2033 +2026-04-10 06:03:40 - INFO - parrotllm.training - step 15070 | epoch 0 | loss 3.5639 | lr 1.46e-03 | grad 0.1980 +2026-04-10 06:03:43 - INFO - parrotllm.training - step 15080 | epoch 0 | loss 3.4719 | lr 1.46e-03 | grad 0.1870 +2026-04-10 06:03:46 - INFO - parrotllm.training - step 15090 | epoch 0 | loss 3.5971 | lr 1.46e-03 | grad 0.2334 +2026-04-10 06:03:49 - INFO - parrotllm.training - step 15100 | epoch 0 | loss 3.4872 | lr 1.46e-03 | grad 0.1997 +2026-04-10 06:03:52 - INFO - parrotllm.training - step 15110 | epoch 0 | loss 3.4532 | lr 1.46e-03 | grad 0.1973 +2026-04-10 06:03:55 - INFO - parrotllm.training - step 15120 | epoch 0 | loss 3.5229 | lr 1.46e-03 | grad 0.2168 +2026-04-10 06:03:59 - INFO - parrotllm.training - step 15130 | epoch 0 | loss 3.4931 | lr 1.46e-03 | grad 0.2036 +2026-04-10 06:04:02 - INFO - parrotllm.training - step 15140 | epoch 0 | loss 3.5015 | lr 1.46e-03 | grad 0.2422 +2026-04-10 06:04:05 - INFO - parrotllm.training - step 15150 | epoch 0 | loss 3.5668 | lr 1.46e-03 | grad 0.1867 +2026-04-10 06:04:08 - INFO - parrotllm.training - step 15160 | epoch 0 | loss 3.4755 | lr 1.46e-03 | grad 0.1946 +2026-04-10 06:04:11 - INFO - parrotllm.training - step 15170 | epoch 0 | loss 3.5023 | lr 1.46e-03 | grad 0.1926 +2026-04-10 06:04:14 - INFO - parrotllm.training - step 15180 | epoch 0 | loss 3.4282 | lr 1.46e-03 | grad 0.1858 +2026-04-10 06:04:17 - INFO - parrotllm.training - step 15190 | epoch 0 | loss 3.5157 | lr 1.46e-03 | grad 0.2319 +2026-04-10 06:04:20 - INFO - parrotllm.training - step 15200 | epoch 0 | loss 3.4584 | lr 1.46e-03 | grad 0.2063 +2026-04-10 06:04:23 - INFO - parrotllm.training - step 15210 | epoch 0 | loss 3.5256 | lr 1.46e-03 | grad 0.2012 +2026-04-10 06:04:26 - INFO - parrotllm.training - step 15220 | epoch 0 | loss 3.5088 | lr 1.46e-03 | grad 0.1870 +2026-04-10 06:04:29 - INFO - parrotllm.training - step 15230 | epoch 0 | loss 3.5040 | lr 1.46e-03 | grad 0.1949 +2026-04-10 06:04:32 - INFO - parrotllm.training - step 15240 | epoch 0 | loss 3.4467 | lr 1.46e-03 | grad 0.2012 +2026-04-10 06:04:35 - INFO - parrotllm.training - step 15250 | epoch 0 | loss 3.5535 | lr 1.46e-03 | grad 0.2137 +2026-04-10 06:04:37 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 15261/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 06:04:38 - INFO - parrotllm.training - step 15260 | epoch 0 | loss 3.6063 | lr 1.46e-03 | grad 0.1859 +2026-04-10 06:04:42 - INFO - parrotllm.training - step 15270 | epoch 0 | loss 3.3967 | lr 1.46e-03 | grad 0.1930 +2026-04-10 06:04:45 - INFO - parrotllm.training - step 15280 | epoch 0 | loss 3.6541 | lr 1.46e-03 | grad 0.1865 +2026-04-10 06:04:48 - INFO - parrotllm.training - step 15290 | epoch 0 | loss 3.4901 | lr 1.46e-03 | grad 0.1900 +2026-04-10 06:04:51 - INFO - parrotllm.training - step 15300 | epoch 0 | loss 3.3884 | lr 1.46e-03 | grad 0.1933 +2026-04-10 06:04:54 - INFO - parrotllm.training - step 15310 | epoch 0 | loss 3.4906 | lr 1.46e-03 | grad 0.2045 +2026-04-10 06:04:57 - INFO - parrotllm.training - step 15320 | epoch 0 | loss 3.5009 | lr 1.46e-03 | grad 0.2259 +2026-04-10 06:05:00 - INFO - parrotllm.training - step 15330 | epoch 0 | loss 3.5238 | lr 1.46e-03 | grad 0.2057 +2026-04-10 06:05:03 - INFO - parrotllm.training - step 15340 | epoch 0 | loss 3.5442 | lr 1.46e-03 | grad 0.2507 +2026-04-10 06:05:06 - INFO - parrotllm.training - step 15350 | epoch 0 | loss 3.4913 | lr 1.46e-03 | grad 0.2038 +2026-04-10 06:05:09 - INFO - parrotllm.training - step 15360 | epoch 0 | loss 3.6223 | lr 1.46e-03 | grad 0.1760 +2026-04-10 06:05:12 - INFO - parrotllm.training - step 15370 | epoch 0 | loss 3.5628 | lr 1.46e-03 | grad 0.1827 +2026-04-10 06:05:15 - INFO - parrotllm.training - step 15380 | epoch 0 | loss 3.3969 | lr 1.46e-03 | grad 0.1799 +2026-04-10 06:05:18 - INFO - parrotllm.training - step 15390 | epoch 0 | loss 3.4664 | lr 1.46e-03 | grad 0.2173 +2026-04-10 06:05:21 - INFO - parrotllm.training - step 15400 | epoch 0 | loss 3.5472 | lr 1.46e-03 | grad 0.1892 +2026-04-10 06:05:24 - INFO - parrotllm.training - step 15410 | epoch 0 | loss 3.4425 | lr 1.46e-03 | grad 0.1900 +2026-04-10 06:05:27 - INFO - parrotllm.training - step 15420 | epoch 0 | loss 3.3570 | lr 1.46e-03 | grad 0.2040 +2026-04-10 06:05:30 - INFO - parrotllm.training - step 15430 | epoch 0 | loss 3.4878 | lr 1.46e-03 | grad 0.2011 +2026-04-10 06:05:33 - INFO - parrotllm.training - step 15440 | epoch 0 | loss 3.4954 | lr 1.46e-03 | grad 0.1802 +2026-04-10 06:05:36 - INFO - parrotllm.training - step 15450 | epoch 0 | loss 3.5295 | lr 1.46e-03 | grad 0.1896 +2026-04-10 06:05:39 - INFO - parrotllm.training - step 15460 | epoch 0 | loss 3.4969 | lr 1.46e-03 | grad 0.1880 +2026-04-10 06:05:43 - INFO - parrotllm.training - step 15470 | epoch 0 | loss 3.4525 | lr 1.46e-03 | grad 0.2120 +2026-04-10 06:05:46 - INFO - parrotllm.training - step 15480 | epoch 0 | loss 3.4209 | lr 1.46e-03 | grad 0.2160 +2026-04-10 06:05:49 - INFO - parrotllm.training - step 15490 | epoch 0 | loss 3.5043 | lr 1.46e-03 | grad 0.2088 +2026-04-10 06:05:52 - INFO - parrotllm.training - step 15500 | epoch 0 | loss 3.5780 | lr 1.46e-03 | grad 0.2316 +2026-04-10 06:05:52 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:05:52 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:05:55 - INFO - parrotllm.training - Train: loss=3.5780, ppl=35.80 +2026-04-10 06:05:55 - INFO - parrotllm.training - Val: loss=3.4443, ppl=31.32 +2026-04-10 06:05:55 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:05:56 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4443_epoch_0000_step_0015500.pt +2026-04-10 06:05:57 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:06:00 - INFO - parrotllm.training - step 15510 | epoch 0 | loss 3.5539 | lr 1.46e-03 | grad 0.2301 +2026-04-10 06:06:03 - INFO - parrotllm.training - step 15520 | epoch 0 | loss 3.4871 | lr 1.46e-03 | grad 0.1831 +2026-04-10 06:06:06 - INFO - parrotllm.training - step 15530 | epoch 0 | loss 3.4449 | lr 1.46e-03 | grad 0.1969 +2026-04-10 06:06:09 - INFO - parrotllm.training - step 15540 | epoch 0 | loss 3.5377 | lr 1.46e-03 | grad 0.1794 +2026-04-10 06:06:12 - INFO - parrotllm.training - step 15550 | epoch 0 | loss 3.5022 | lr 1.46e-03 | grad 0.1965 +2026-04-10 06:06:15 - INFO - parrotllm.training - step 15560 | epoch 0 | loss 3.5877 | lr 1.46e-03 | grad 0.1971 +2026-04-10 06:06:18 - INFO - parrotllm.training - step 15570 | epoch 0 | loss 3.4893 | lr 1.46e-03 | grad 0.1857 +2026-04-10 06:06:21 - INFO - parrotllm.training - step 15580 | epoch 0 | loss 3.5064 | lr 1.46e-03 | grad 0.1887 +2026-04-10 06:06:24 - INFO - parrotllm.training - step 15590 | epoch 0 | loss 3.4740 | lr 1.46e-03 | grad 0.1995 +2026-04-10 06:06:27 - INFO - parrotllm.training - step 15600 | epoch 0 | loss 3.4258 | lr 1.45e-03 | grad 0.2018 +2026-04-10 06:06:30 - INFO - parrotllm.training - step 15610 | epoch 0 | loss 3.4423 | lr 1.45e-03 | grad 0.1795 +2026-04-10 06:06:33 - INFO - parrotllm.training - step 15620 | epoch 0 | loss 3.5223 | lr 1.45e-03 | grad 0.1992 +2026-04-10 06:06:36 - INFO - parrotllm.training - step 15630 | epoch 0 | loss 3.5224 | lr 1.45e-03 | grad 0.2122 +2026-04-10 06:06:39 - INFO - parrotllm.training - step 15640 | epoch 0 | loss 3.5399 | lr 1.45e-03 | grad 0.2289 +2026-04-10 06:06:42 - INFO - parrotllm.training - step 15650 | epoch 0 | loss 3.4971 | lr 1.45e-03 | grad 0.2200 +2026-04-10 06:06:45 - INFO - parrotllm.training - step 15660 | epoch 0 | loss 3.4525 | lr 1.45e-03 | grad 0.2158 +2026-04-10 06:06:49 - INFO - parrotllm.training - step 15670 | epoch 0 | loss 3.4892 | lr 1.45e-03 | grad 0.1817 +2026-04-10 06:06:52 - INFO - parrotllm.training - step 15680 | epoch 0 | loss 3.5271 | lr 1.45e-03 | grad 0.2455 +2026-04-10 06:06:55 - INFO - parrotllm.training - step 15690 | epoch 0 | loss 3.4664 | lr 1.45e-03 | grad 0.1895 +2026-04-10 06:06:58 - INFO - parrotllm.training - step 15700 | epoch 0 | loss 3.4470 | lr 1.45e-03 | grad 0.1955 +2026-04-10 06:07:01 - INFO - parrotllm.training - step 15710 | epoch 0 | loss 3.5127 | lr 1.45e-03 | grad 0.2191 +2026-04-10 06:07:04 - INFO - parrotllm.training - step 15720 | epoch 0 | loss 3.4346 | lr 1.45e-03 | grad 0.1990 +2026-04-10 06:07:07 - INFO - parrotllm.training - step 15730 | epoch 0 | loss 3.5039 | lr 1.45e-03 | grad 0.2114 +2026-04-10 06:07:10 - INFO - parrotllm.training - step 15740 | epoch 0 | loss 3.5078 | lr 1.45e-03 | grad 0.1908 +2026-04-10 06:07:13 - INFO - parrotllm.training - step 15750 | epoch 0 | loss 3.5169 | lr 1.45e-03 | grad 0.1964 +2026-04-10 06:07:16 - INFO - parrotllm.training - step 15760 | epoch 0 | loss 3.5357 | lr 1.45e-03 | grad 0.1953 +2026-04-10 06:07:19 - INFO - parrotllm.training - step 15770 | epoch 0 | loss 3.3649 | lr 1.45e-03 | grad 0.1963 +2026-04-10 06:07:22 - INFO - parrotllm.training - step 15780 | epoch 0 | loss 3.4959 | lr 1.45e-03 | grad 0.2055 +2026-04-10 06:07:25 - INFO - parrotllm.training - step 15790 | epoch 0 | loss 3.5781 | lr 1.45e-03 | grad 0.2179 +2026-04-10 06:07:28 - INFO - parrotllm.training - step 15800 | epoch 0 | loss 3.4949 | lr 1.45e-03 | grad 0.1865 +2026-04-10 06:07:31 - INFO - parrotllm.training - step 15810 | epoch 0 | loss 3.4947 | lr 1.45e-03 | grad 0.2212 +2026-04-10 06:07:34 - INFO - parrotllm.training - step 15820 | epoch 0 | loss 3.4399 | lr 1.45e-03 | grad 0.1740 +2026-04-10 06:07:37 - INFO - parrotllm.training - step 15830 | epoch 0 | loss 3.5255 | lr 1.45e-03 | grad 0.1792 +2026-04-10 06:07:40 - INFO - parrotllm.training - step 15840 | epoch 0 | loss 3.4093 | lr 1.45e-03 | grad 0.2181 +2026-04-10 06:07:43 - INFO - parrotllm.training - step 15850 | epoch 0 | loss 3.5037 | lr 1.45e-03 | grad 0.2131 +2026-04-10 06:07:47 - INFO - parrotllm.training - step 15860 | epoch 0 | loss 3.3900 | lr 1.45e-03 | grad 0.1901 +2026-04-10 06:07:50 - INFO - parrotllm.training - step 15870 | epoch 0 | loss 3.4968 | lr 1.45e-03 | grad 0.2019 +2026-04-10 06:07:53 - INFO - parrotllm.training - step 15880 | epoch 0 | loss 3.4891 | lr 1.45e-03 | grad 0.2024 +2026-04-10 06:07:56 - INFO - parrotllm.training - step 15890 | epoch 0 | loss 3.5674 | lr 1.45e-03 | grad 0.1877 +2026-04-10 06:07:59 - INFO - parrotllm.training - step 15900 | epoch 0 | loss 3.4923 | lr 1.45e-03 | grad 0.1801 +2026-04-10 06:08:02 - INFO - parrotllm.training - step 15910 | epoch 0 | loss 3.4617 | lr 1.45e-03 | grad 0.2490 +2026-04-10 06:08:05 - INFO - parrotllm.training - step 15920 | epoch 0 | loss 3.4816 | lr 1.45e-03 | grad 0.2296 +2026-04-10 06:08:08 - INFO - parrotllm.training - step 15930 | epoch 0 | loss 3.4749 | lr 1.45e-03 | grad 0.2359 +2026-04-10 06:08:11 - INFO - parrotllm.training - step 15940 | epoch 0 | loss 3.4828 | lr 1.45e-03 | grad 0.1988 +2026-04-10 06:08:14 - INFO - parrotllm.training - step 15950 | epoch 0 | loss 3.4782 | lr 1.45e-03 | grad 0.2019 +2026-04-10 06:08:17 - INFO - parrotllm.training - step 15960 | epoch 0 | loss 3.4650 | lr 1.45e-03 | grad 0.2012 +2026-04-10 06:08:20 - INFO - parrotllm.training - step 15970 | epoch 0 | loss 3.5323 | lr 1.45e-03 | grad 0.2394 +2026-04-10 06:08:23 - INFO - parrotllm.training - step 15980 | epoch 0 | loss 3.4674 | lr 1.45e-03 | grad 0.1697 +2026-04-10 06:08:26 - INFO - parrotllm.training - step 15990 | epoch 0 | loss 3.5304 | lr 1.45e-03 | grad 0.1852 +2026-04-10 06:08:29 - INFO - parrotllm.training - step 16000 | epoch 0 | loss 3.5807 | lr 1.45e-03 | grad 0.2082 +2026-04-10 06:08:29 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:08:29 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:08:32 - INFO - parrotllm.training - Train: loss=3.5807, ppl=35.90 +2026-04-10 06:08:32 - INFO - parrotllm.training - Val: loss=3.4413, ppl=31.23 +2026-04-10 06:08:32 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:08:33 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4413_epoch_0000_step_0016000.pt +2026-04-10 06:08:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:08:37 - INFO - parrotllm.training - step 16010 | epoch 0 | loss 3.4859 | lr 1.45e-03 | grad 0.1905 +2026-04-10 06:08:40 - INFO - parrotllm.training - step 16020 | epoch 0 | loss 3.5997 | lr 1.45e-03 | grad 0.1938 +2026-04-10 06:08:43 - INFO - parrotllm.training - step 16030 | epoch 0 | loss 3.4980 | lr 1.45e-03 | grad 0.1999 +2026-04-10 06:08:46 - INFO - parrotllm.training - step 16040 | epoch 0 | loss 3.4511 | lr 1.45e-03 | grad 0.2090 +2026-04-10 06:08:50 - INFO - parrotllm.training - step 16050 | epoch 0 | loss 3.4872 | lr 1.45e-03 | grad 0.2096 +2026-04-10 06:08:53 - INFO - parrotllm.training - step 16060 | epoch 0 | loss 3.4613 | lr 1.45e-03 | grad 0.2293 +2026-04-10 06:08:56 - INFO - parrotllm.training - step 16070 | epoch 0 | loss 3.5360 | lr 1.45e-03 | grad 0.2141 +2026-04-10 06:08:59 - INFO - parrotllm.training - step 16080 | epoch 0 | loss 3.4505 | lr 1.45e-03 | grad 0.2022 +2026-04-10 06:09:02 - INFO - parrotllm.training - step 16090 | epoch 0 | loss 3.6487 | lr 1.45e-03 | grad 0.2059 +2026-04-10 06:09:05 - INFO - parrotllm.training - step 16100 | epoch 0 | loss 3.4944 | lr 1.45e-03 | grad 0.1844 +2026-04-10 06:09:08 - INFO - parrotllm.training - step 16110 | epoch 0 | loss 3.5111 | lr 1.45e-03 | grad 0.1939 +2026-04-10 06:09:11 - INFO - parrotllm.training - step 16120 | epoch 0 | loss 3.4609 | lr 1.45e-03 | grad 0.1890 +2026-04-10 06:09:14 - INFO - parrotllm.training - step 16130 | epoch 0 | loss 3.4592 | lr 1.45e-03 | grad 0.1843 +2026-04-10 06:09:17 - INFO - parrotllm.training - step 16140 | epoch 0 | loss 3.4591 | lr 1.45e-03 | grad 0.1872 +2026-04-10 06:09:20 - INFO - parrotllm.training - step 16150 | epoch 0 | loss 3.4748 | lr 1.45e-03 | grad 0.1905 +2026-04-10 06:09:23 - INFO - parrotllm.training - step 16160 | epoch 0 | loss 3.4797 | lr 1.45e-03 | grad 0.2138 +2026-04-10 06:09:26 - INFO - parrotllm.training - step 16170 | epoch 0 | loss 3.5078 | lr 1.45e-03 | grad 0.2269 +2026-04-10 06:09:29 - INFO - parrotllm.training - step 16180 | epoch 0 | loss 3.4557 | lr 1.45e-03 | grad 0.1955 +2026-04-10 06:09:32 - INFO - parrotllm.training - step 16190 | epoch 0 | loss 3.4705 | lr 1.45e-03 | grad 0.2221 +2026-04-10 06:09:35 - INFO - parrotllm.training - step 16200 | epoch 0 | loss 3.5605 | lr 1.45e-03 | grad 0.1944 +2026-04-10 06:09:38 - INFO - parrotllm.training - step 16210 | epoch 0 | loss 3.4377 | lr 1.45e-03 | grad 0.2185 +2026-04-10 06:09:41 - INFO - parrotllm.training - step 16220 | epoch 0 | loss 3.5164 | lr 1.45e-03 | grad 0.1893 +2026-04-10 06:09:44 - INFO - parrotllm.training - step 16230 | epoch 0 | loss 3.4376 | lr 1.45e-03 | grad 0.2279 +2026-04-10 06:09:48 - INFO - parrotllm.training - step 16240 | epoch 0 | loss 3.5665 | lr 1.45e-03 | grad 0.2058 +2026-04-10 06:09:51 - INFO - parrotllm.training - step 16250 | epoch 0 | loss 3.4799 | lr 1.45e-03 | grad 0.2056 +2026-04-10 06:09:54 - INFO - parrotllm.training - step 16260 | epoch 0 | loss 3.4949 | lr 1.45e-03 | grad 0.2065 +2026-04-10 06:09:57 - INFO - parrotllm.training - step 16270 | epoch 0 | loss 3.5418 | lr 1.45e-03 | grad 0.2218 +2026-04-10 06:10:00 - INFO - parrotllm.training - step 16280 | epoch 0 | loss 3.5573 | lr 1.45e-03 | grad 0.1890 +2026-04-10 06:10:03 - INFO - parrotllm.training - step 16290 | epoch 0 | loss 3.5380 | lr 1.45e-03 | grad 0.2058 +2026-04-10 06:10:06 - INFO - parrotllm.training - step 16300 | epoch 0 | loss 3.4725 | lr 1.45e-03 | grad 0.1793 +2026-04-10 06:10:09 - INFO - parrotllm.training - step 16310 | epoch 0 | loss 3.4947 | lr 1.45e-03 | grad 0.1829 +2026-04-10 06:10:12 - INFO - parrotllm.training - step 16320 | epoch 0 | loss 3.4281 | lr 1.45e-03 | grad 0.2040 +2026-04-10 06:10:15 - INFO - parrotllm.training - step 16330 | epoch 0 | loss 3.5397 | lr 1.45e-03 | grad 0.2338 +2026-04-10 06:10:18 - INFO - parrotllm.training - step 16340 | epoch 0 | loss 3.5634 | lr 1.45e-03 | grad 0.1992 +2026-04-10 06:10:21 - INFO - parrotllm.training - step 16350 | epoch 0 | loss 3.4702 | lr 1.45e-03 | grad 0.1827 +2026-04-10 06:10:24 - INFO - parrotllm.training - step 16360 | epoch 0 | loss 3.4743 | lr 1.45e-03 | grad 0.1805 +2026-04-10 06:10:27 - INFO - parrotllm.training - step 16370 | epoch 0 | loss 3.5357 | lr 1.45e-03 | grad 0.2045 +2026-04-10 06:10:30 - INFO - parrotllm.training - step 16380 | epoch 0 | loss 3.4620 | lr 1.45e-03 | grad 0.1888 +2026-04-10 06:10:33 - INFO - parrotllm.training - step 16390 | epoch 0 | loss 3.4488 | lr 1.45e-03 | grad 0.1844 +2026-04-10 06:10:36 - INFO - parrotllm.training - step 16400 | epoch 0 | loss 3.4546 | lr 1.45e-03 | grad 0.2208 +2026-04-10 06:10:39 - INFO - parrotllm.training - step 16410 | epoch 0 | loss 3.5319 | lr 1.45e-03 | grad 0.1836 +2026-04-10 06:10:42 - INFO - parrotllm.training - step 16420 | epoch 0 | loss 3.4897 | lr 1.45e-03 | grad 0.1843 +2026-04-10 06:10:45 - INFO - parrotllm.training - step 16430 | epoch 0 | loss 3.5342 | lr 1.45e-03 | grad 0.1992 +2026-04-10 06:10:49 - INFO - parrotllm.training - step 16440 | epoch 0 | loss 3.4234 | lr 1.45e-03 | grad 0.2002 +2026-04-10 06:10:52 - INFO - parrotllm.training - step 16450 | epoch 0 | loss 3.4962 | lr 1.45e-03 | grad 0.1927 +2026-04-10 06:10:55 - INFO - parrotllm.training - step 16460 | epoch 0 | loss 3.4350 | lr 1.45e-03 | grad 0.2008 +2026-04-10 06:10:58 - INFO - parrotllm.training - step 16470 | epoch 0 | loss 3.5444 | lr 1.45e-03 | grad 0.2020 +2026-04-10 06:11:01 - INFO - parrotllm.training - step 16480 | epoch 0 | loss 3.5699 | lr 1.45e-03 | grad 0.1851 +2026-04-10 06:11:04 - INFO - parrotllm.training - step 16490 | epoch 0 | loss 3.4745 | lr 1.45e-03 | grad 0.2180 +2026-04-10 06:11:07 - INFO - parrotllm.training - step 16500 | epoch 0 | loss 3.5624 | lr 1.45e-03 | grad 0.2145 +2026-04-10 06:11:07 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:11:07 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:11:10 - INFO - parrotllm.training - Train: loss=3.5624, ppl=35.25 +2026-04-10 06:11:10 - INFO - parrotllm.training - Val: loss=3.4409, ppl=31.22 +2026-04-10 06:11:11 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4409_epoch_0000_step_0016500.pt +2026-04-10 06:11:12 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4413, min_delta=0.001000). +2026-04-10 06:11:12 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:11:15 - INFO - parrotllm.training - step 16510 | epoch 0 | loss 3.4944 | lr 1.45e-03 | grad 0.2168 +2026-04-10 06:11:18 - INFO - parrotllm.training - step 16520 | epoch 0 | loss 3.4619 | lr 1.45e-03 | grad 0.1851 +2026-04-10 06:11:21 - INFO - parrotllm.training - step 16530 | epoch 0 | loss 3.5074 | lr 1.45e-03 | grad 0.2062 +2026-04-10 06:11:24 - INFO - parrotllm.training - step 16540 | epoch 0 | loss 3.4910 | lr 1.45e-03 | grad 0.1978 +2026-04-10 06:11:27 - INFO - parrotllm.training - step 16550 | epoch 0 | loss 3.4874 | lr 1.45e-03 | grad 0.2179 +2026-04-10 06:11:30 - INFO - parrotllm.training - step 16560 | epoch 0 | loss 3.5640 | lr 1.45e-03 | grad 0.2052 +2026-04-10 06:11:33 - INFO - parrotllm.training - step 16570 | epoch 0 | loss 3.4723 | lr 1.45e-03 | grad 0.1794 +2026-04-10 06:11:36 - INFO - parrotllm.training - step 16580 | epoch 0 | loss 3.4778 | lr 1.45e-03 | grad 0.1868 +2026-04-10 06:11:40 - INFO - parrotllm.training - step 16590 | epoch 0 | loss 3.5164 | lr 1.45e-03 | grad 0.1938 +2026-04-10 06:11:43 - INFO - parrotllm.training - step 16600 | epoch 0 | loss 3.4150 | lr 1.45e-03 | grad 0.2031 +2026-04-10 06:11:46 - INFO - parrotllm.training - step 16610 | epoch 0 | loss 3.5416 | lr 1.45e-03 | grad 0.1914 +2026-04-10 06:11:49 - INFO - parrotllm.training - step 16620 | epoch 0 | loss 3.5243 | lr 1.45e-03 | grad 0.2046 +2026-04-10 06:11:52 - INFO - parrotllm.training - step 16630 | epoch 0 | loss 3.4592 | lr 1.45e-03 | grad 0.2148 +2026-04-10 06:11:55 - INFO - parrotllm.training - step 16640 | epoch 0 | loss 3.4435 | lr 1.45e-03 | grad 0.2018 +2026-04-10 06:11:58 - INFO - parrotllm.training - step 16650 | epoch 0 | loss 3.5264 | lr 1.45e-03 | grad 0.1913 +2026-04-10 06:12:01 - INFO - parrotllm.training - step 16660 | epoch 0 | loss 3.4730 | lr 1.45e-03 | grad 0.2404 +2026-04-10 06:12:04 - INFO - parrotllm.training - step 16670 | epoch 0 | loss 3.4144 | lr 1.45e-03 | grad 0.1952 +2026-04-10 06:12:07 - INFO - parrotllm.training - step 16680 | epoch 0 | loss 3.6361 | lr 1.45e-03 | grad 0.1829 +2026-04-10 06:12:10 - INFO - parrotllm.training - step 16690 | epoch 0 | loss 3.4424 | lr 1.45e-03 | grad 0.2086 +2026-04-10 06:12:13 - INFO - parrotllm.training - step 16700 | epoch 0 | loss 3.4715 | lr 1.45e-03 | grad 0.2114 +2026-04-10 06:12:16 - INFO - parrotllm.training - step 16710 | epoch 0 | loss 3.5218 | lr 1.45e-03 | grad 0.2240 +2026-04-10 06:12:19 - INFO - parrotllm.training - step 16720 | epoch 0 | loss 3.3534 | lr 1.45e-03 | grad 0.1932 +2026-04-10 06:12:22 - INFO - parrotllm.training - step 16730 | epoch 0 | loss 3.4856 | lr 1.45e-03 | grad 0.1804 +2026-04-10 06:12:25 - INFO - parrotllm.training - step 16740 | epoch 0 | loss 3.4508 | lr 1.45e-03 | grad 0.2018 +2026-04-10 06:12:28 - INFO - parrotllm.training - step 16750 | epoch 0 | loss 3.4125 | lr 1.45e-03 | grad 0.1866 +2026-04-10 06:12:31 - INFO - parrotllm.training - step 16760 | epoch 0 | loss 3.5739 | lr 1.45e-03 | grad 0.2215 +2026-04-10 06:12:35 - INFO - parrotllm.training - step 16770 | epoch 0 | loss 3.5192 | lr 1.45e-03 | grad 0.1852 +2026-04-10 06:12:38 - INFO - parrotllm.training - step 16780 | epoch 0 | loss 3.5053 | lr 1.45e-03 | grad 0.1899 +2026-04-10 06:12:41 - INFO - parrotllm.training - step 16790 | epoch 0 | loss 3.4979 | lr 1.45e-03 | grad 0.1878 +2026-04-10 06:12:44 - INFO - parrotllm.training - step 16800 | epoch 0 | loss 3.4719 | lr 1.45e-03 | grad 0.1984 +2026-04-10 06:12:47 - INFO - parrotllm.training - step 16810 | epoch 0 | loss 3.5050 | lr 1.45e-03 | grad 0.1927 +2026-04-10 06:12:50 - INFO - parrotllm.training - step 16820 | epoch 0 | loss 3.4463 | lr 1.45e-03 | grad 0.2024 +2026-04-10 06:12:53 - INFO - parrotllm.training - step 16830 | epoch 0 | loss 3.5857 | lr 1.45e-03 | grad 0.1645 +2026-04-10 06:12:56 - INFO - parrotllm.training - step 16840 | epoch 0 | loss 3.6000 | lr 1.45e-03 | grad 0.2095 +2026-04-10 06:12:59 - INFO - parrotllm.training - step 16850 | epoch 0 | loss 3.5758 | lr 1.45e-03 | grad 0.1739 +2026-04-10 06:13:02 - INFO - parrotllm.training - step 16860 | epoch 0 | loss 3.6698 | lr 1.45e-03 | grad 0.2332 +2026-04-10 06:13:05 - INFO - parrotllm.training - step 16870 | epoch 0 | loss 3.4673 | lr 1.45e-03 | grad 0.1943 +2026-04-10 06:13:08 - INFO - parrotllm.training - step 16880 | epoch 0 | loss 3.4705 | lr 1.45e-03 | grad 0.2073 +2026-04-10 06:13:11 - INFO - parrotllm.training - step 16890 | epoch 0 | loss 3.5634 | lr 1.45e-03 | grad 0.1976 +2026-04-10 06:13:14 - INFO - parrotllm.training - step 16900 | epoch 0 | loss 3.5096 | lr 1.45e-03 | grad 0.2096 +2026-04-10 06:13:17 - INFO - parrotllm.training - step 16910 | epoch 0 | loss 3.4670 | lr 1.45e-03 | grad 0.2295 +2026-04-10 06:13:20 - INFO - parrotllm.training - step 16920 | epoch 0 | loss 3.5402 | lr 1.45e-03 | grad 0.2247 +2026-04-10 06:13:23 - INFO - parrotllm.training - step 16930 | epoch 0 | loss 3.4805 | lr 1.45e-03 | grad 0.2002 +2026-04-10 06:13:26 - INFO - parrotllm.training - step 16940 | epoch 0 | loss 3.5613 | lr 1.45e-03 | grad 0.1979 +2026-04-10 06:13:29 - INFO - parrotllm.training - step 16950 | epoch 0 | loss 3.4551 | lr 1.45e-03 | grad 0.1799 +2026-04-10 06:13:33 - INFO - parrotllm.training - step 16960 | epoch 0 | loss 3.4993 | lr 1.45e-03 | grad 0.1950 +2026-04-10 06:13:36 - INFO - parrotllm.training - step 16970 | epoch 0 | loss 3.4777 | lr 1.45e-03 | grad 0.1960 +2026-04-10 06:13:39 - INFO - parrotllm.training - step 16980 | epoch 0 | loss 3.3883 | lr 1.45e-03 | grad 0.2228 +2026-04-10 06:13:42 - INFO - parrotllm.training - step 16990 | epoch 0 | loss 3.5929 | lr 1.45e-03 | grad 0.1831 +2026-04-10 06:13:45 - INFO - parrotllm.training - step 17000 | epoch 0 | loss 3.4910 | lr 1.45e-03 | grad 0.2273 +2026-04-10 06:13:45 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:13:45 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:13:48 - INFO - parrotllm.training - Train: loss=3.4910, ppl=32.82 +2026-04-10 06:13:48 - INFO - parrotllm.training - Val: loss=3.4409, ppl=31.22 +2026-04-10 06:13:49 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4409_epoch_0000_step_0017000.pt +2026-04-10 06:13:50 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.4413, min_delta=0.001000). +2026-04-10 06:13:50 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:13:53 - INFO - parrotllm.training - step 17010 | epoch 0 | loss 3.6069 | lr 1.45e-03 | grad 0.2451 +2026-04-10 06:13:56 - INFO - parrotllm.training - step 17020 | epoch 0 | loss 3.5486 | lr 1.45e-03 | grad 0.1972 +2026-04-10 06:13:59 - INFO - parrotllm.training - step 17030 | epoch 0 | loss 3.5435 | lr 1.45e-03 | grad 0.1873 +2026-04-10 06:14:02 - INFO - parrotllm.training - step 17040 | epoch 0 | loss 3.5308 | lr 1.45e-03 | grad 0.1989 +2026-04-10 06:14:05 - INFO - parrotllm.training - step 17050 | epoch 0 | loss 3.5190 | lr 1.45e-03 | grad 0.1887 +2026-04-10 06:14:08 - INFO - parrotllm.training - step 17060 | epoch 0 | loss 3.5760 | lr 1.45e-03 | grad 0.2132 +2026-04-10 06:14:11 - INFO - parrotllm.training - step 17070 | epoch 0 | loss 3.5135 | lr 1.45e-03 | grad 0.1916 +2026-04-10 06:14:14 - INFO - parrotllm.training - step 17080 | epoch 0 | loss 3.5552 | lr 1.45e-03 | grad 0.1924 +2026-04-10 06:14:17 - INFO - parrotllm.training - step 17090 | epoch 0 | loss 3.4556 | lr 1.45e-03 | grad 0.1939 +2026-04-10 06:14:20 - INFO - parrotllm.training - step 17100 | epoch 0 | loss 3.5742 | lr 1.45e-03 | grad 0.2334 +2026-04-10 06:14:23 - INFO - parrotllm.training - step 17110 | epoch 0 | loss 3.4898 | lr 1.44e-03 | grad 0.2718 +2026-04-10 06:14:26 - INFO - parrotllm.training - step 17120 | epoch 0 | loss 3.4827 | lr 1.44e-03 | grad 0.2447 +2026-04-10 06:14:29 - INFO - parrotllm.training - step 17130 | epoch 0 | loss 3.4645 | lr 1.44e-03 | grad 0.2337 +2026-04-10 06:14:33 - INFO - parrotllm.training - step 17140 | epoch 0 | loss 3.5850 | lr 1.44e-03 | grad 0.2228 +2026-04-10 06:14:36 - INFO - parrotllm.training - step 17150 | epoch 0 | loss 3.5335 | lr 1.44e-03 | grad 0.1853 +2026-04-10 06:14:39 - INFO - parrotllm.training - step 17160 | epoch 0 | loss 3.4947 | lr 1.44e-03 | grad 0.1749 +2026-04-10 06:14:42 - INFO - parrotllm.training - step 17170 | epoch 0 | loss 3.6137 | lr 1.44e-03 | grad 0.2192 +2026-04-10 06:14:45 - INFO - parrotllm.training - step 17180 | epoch 0 | loss 3.5431 | lr 1.44e-03 | grad 0.1974 +2026-04-10 06:14:48 - INFO - parrotllm.training - step 17190 | epoch 0 | loss 3.5566 | lr 1.44e-03 | grad 0.2011 +2026-04-10 06:14:51 - INFO - parrotllm.training - step 17200 | epoch 0 | loss 3.5281 | lr 1.44e-03 | grad 0.2006 +2026-04-10 06:14:54 - INFO - parrotllm.training - step 17210 | epoch 0 | loss 3.4776 | lr 1.44e-03 | grad 0.1980 +2026-04-10 06:14:57 - INFO - parrotllm.training - step 17220 | epoch 0 | loss 3.5446 | lr 1.44e-03 | grad 0.1978 +2026-04-10 06:15:00 - INFO - parrotllm.training - step 17230 | epoch 0 | loss 3.3939 | lr 1.44e-03 | grad 0.1933 +2026-04-10 06:15:03 - INFO - parrotllm.training - step 17240 | epoch 0 | loss 3.4515 | lr 1.44e-03 | grad 0.2242 +2026-04-10 06:15:06 - INFO - parrotllm.training - step 17250 | epoch 0 | loss 3.5267 | lr 1.44e-03 | grad 0.2482 +2026-04-10 06:15:09 - INFO - parrotllm.training - step 17260 | epoch 0 | loss 3.4555 | lr 1.44e-03 | grad 0.2125 +2026-04-10 06:15:12 - INFO - parrotllm.training - step 17270 | epoch 0 | loss 3.4663 | lr 1.44e-03 | grad 0.2091 +2026-04-10 06:15:15 - INFO - parrotllm.training - step 17280 | epoch 0 | loss 3.3684 | lr 1.44e-03 | grad 0.1869 +2026-04-10 06:15:18 - INFO - parrotllm.training - step 17290 | epoch 0 | loss 3.4671 | lr 1.44e-03 | grad 0.1864 +2026-04-10 06:15:21 - INFO - parrotllm.training - step 17300 | epoch 0 | loss 3.5709 | lr 1.44e-03 | grad 0.2488 +2026-04-10 06:15:24 - INFO - parrotllm.training - step 17310 | epoch 0 | loss 3.5264 | lr 1.44e-03 | grad 0.1919 +2026-04-10 06:15:28 - INFO - parrotllm.training - step 17320 | epoch 0 | loss 3.4681 | lr 1.44e-03 | grad 0.1841 +2026-04-10 06:15:31 - INFO - parrotllm.training - step 17330 | epoch 0 | loss 3.4276 | lr 1.44e-03 | grad 0.2009 +2026-04-10 06:15:34 - INFO - parrotllm.training - step 17340 | epoch 0 | loss 3.5182 | lr 1.44e-03 | grad 0.2398 +2026-04-10 06:15:37 - INFO - parrotllm.training - step 17350 | epoch 0 | loss 3.5072 | lr 1.44e-03 | grad 0.1994 +2026-04-10 06:15:40 - INFO - parrotllm.training - step 17360 | epoch 0 | loss 3.4150 | lr 1.44e-03 | grad 0.2392 +2026-04-10 06:15:43 - INFO - parrotllm.training - step 17370 | epoch 0 | loss 3.4946 | lr 1.44e-03 | grad 0.2411 +2026-04-10 06:15:44 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 17380/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 06:15:46 - INFO - parrotllm.training - step 17380 | epoch 0 | loss 3.4074 | lr 1.44e-03 | grad 0.1822 +2026-04-10 06:15:49 - INFO - parrotllm.training - step 17390 | epoch 0 | loss 3.4898 | lr 1.44e-03 | grad 0.2044 +2026-04-10 06:15:52 - INFO - parrotllm.training - step 17400 | epoch 0 | loss 3.5100 | lr 1.44e-03 | grad 0.2012 +2026-04-10 06:15:55 - INFO - parrotllm.training - step 17410 | epoch 0 | loss 3.4256 | lr 1.44e-03 | grad 0.1985 +2026-04-10 06:15:58 - INFO - parrotllm.training - step 17420 | epoch 0 | loss 3.4963 | lr 1.44e-03 | grad 0.2031 +2026-04-10 06:16:01 - INFO - parrotllm.training - step 17430 | epoch 0 | loss 3.5551 | lr 1.44e-03 | grad 0.1894 +2026-04-10 06:16:04 - INFO - parrotllm.training - step 17440 | epoch 0 | loss 3.5082 | lr 1.44e-03 | grad 0.2257 +2026-04-10 06:16:08 - INFO - parrotllm.training - step 17450 | epoch 0 | loss 3.4889 | lr 1.44e-03 | grad 0.1926 +2026-04-10 06:16:11 - INFO - parrotllm.training - step 17460 | epoch 0 | loss 3.5366 | lr 1.44e-03 | grad 0.1817 +2026-04-10 06:16:14 - INFO - parrotllm.training - step 17470 | epoch 0 | loss 3.5845 | lr 1.44e-03 | grad 0.2118 +2026-04-10 06:16:17 - INFO - parrotllm.training - step 17480 | epoch 0 | loss 3.3682 | lr 1.44e-03 | grad 0.1990 +2026-04-10 06:16:20 - INFO - parrotllm.training - step 17490 | epoch 0 | loss 3.5120 | lr 1.44e-03 | grad 0.2200 +2026-04-10 06:16:23 - INFO - parrotllm.training - step 17500 | epoch 0 | loss 3.5603 | lr 1.44e-03 | grad 0.1824 +2026-04-10 06:16:23 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:16:23 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:16:26 - INFO - parrotllm.training - Train: loss=3.5603, ppl=35.17 +2026-04-10 06:16:26 - INFO - parrotllm.training - Val: loss=3.4349, ppl=31.03 +2026-04-10 06:16:26 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:16:27 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4349_epoch_0000_step_0017500.pt +2026-04-10 06:16:28 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:16:29 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0017500.pt +2026-04-10 06:16:33 - INFO - parrotllm.training - step 17510 | epoch 0 | loss 3.4810 | lr 1.44e-03 | grad 0.2018 +2026-04-10 06:16:36 - INFO - parrotllm.training - step 17520 | epoch 0 | loss 3.3577 | lr 1.44e-03 | grad 0.2124 +2026-04-10 06:16:39 - INFO - parrotllm.training - step 17530 | epoch 0 | loss 3.4493 | lr 1.44e-03 | grad 0.1962 +2026-04-10 06:16:42 - INFO - parrotllm.training - step 17540 | epoch 0 | loss 3.5859 | lr 1.44e-03 | grad 0.1835 +2026-04-10 06:16:45 - INFO - parrotllm.training - step 17550 | epoch 0 | loss 3.5098 | lr 1.44e-03 | grad 0.2092 +2026-04-10 06:16:48 - INFO - parrotllm.training - step 17560 | epoch 0 | loss 3.5224 | lr 1.44e-03 | grad 0.1890 +2026-04-10 06:16:51 - INFO - parrotllm.training - step 17570 | epoch 0 | loss 3.5106 | lr 1.44e-03 | grad 0.1862 +2026-04-10 06:16:54 - INFO - parrotllm.training - step 17580 | epoch 0 | loss 3.5948 | lr 1.44e-03 | grad 0.2103 +2026-04-10 06:16:58 - INFO - parrotllm.training - step 17590 | epoch 0 | loss 3.5803 | lr 1.44e-03 | grad 0.2084 +2026-04-10 06:17:01 - INFO - parrotllm.training - step 17600 | epoch 0 | loss 3.4154 | lr 1.44e-03 | grad 0.1968 +2026-04-10 06:17:04 - INFO - parrotllm.training - step 17610 | epoch 0 | loss 3.5676 | lr 1.44e-03 | grad 0.1924 +2026-04-10 06:17:07 - INFO - parrotllm.training - step 17620 | epoch 0 | loss 3.4052 | lr 1.44e-03 | grad 0.1900 +2026-04-10 06:17:10 - INFO - parrotllm.training - step 17630 | epoch 0 | loss 3.5157 | lr 1.44e-03 | grad 0.2102 +2026-04-10 06:17:13 - INFO - parrotllm.training - step 17640 | epoch 0 | loss 3.5348 | lr 1.44e-03 | grad 0.2028 +2026-04-10 06:17:16 - INFO - parrotllm.training - step 17650 | epoch 0 | loss 3.6353 | lr 1.44e-03 | grad 0.1951 +2026-04-10 06:17:19 - INFO - parrotllm.training - step 17660 | epoch 0 | loss 3.4785 | lr 1.44e-03 | grad 0.1955 +2026-04-10 06:17:22 - INFO - parrotllm.training - step 17670 | epoch 0 | loss 3.4643 | lr 1.44e-03 | grad 0.1860 +2026-04-10 06:17:25 - INFO - parrotllm.training - step 17680 | epoch 0 | loss 3.5333 | lr 1.44e-03 | grad 0.1781 +2026-04-10 06:17:28 - INFO - parrotllm.training - step 17690 | epoch 0 | loss 3.4906 | lr 1.44e-03 | grad 0.2133 +2026-04-10 06:17:31 - INFO - parrotllm.training - step 17700 | epoch 0 | loss 3.4094 | lr 1.44e-03 | grad 0.1896 +2026-04-10 06:17:34 - INFO - parrotllm.training - step 17710 | epoch 0 | loss 3.4920 | lr 1.44e-03 | grad 0.2029 +2026-04-10 06:17:37 - INFO - parrotllm.training - step 17720 | epoch 0 | loss 3.4573 | lr 1.44e-03 | grad 0.2045 +2026-04-10 06:17:40 - INFO - parrotllm.training - step 17730 | epoch 0 | loss 3.5807 | lr 1.44e-03 | grad 0.1861 +2026-04-10 06:17:43 - INFO - parrotllm.training - step 17740 | epoch 0 | loss 3.5738 | lr 1.44e-03 | grad 0.2314 +2026-04-10 06:17:46 - INFO - parrotllm.training - step 17750 | epoch 0 | loss 3.5229 | lr 1.44e-03 | grad 0.1953 +2026-04-10 06:17:49 - INFO - parrotllm.training - step 17760 | epoch 0 | loss 3.4521 | lr 1.44e-03 | grad 0.2214 +2026-04-10 06:17:53 - INFO - parrotllm.training - step 17770 | epoch 0 | loss 3.4504 | lr 1.44e-03 | grad 0.2092 +2026-04-10 06:17:56 - INFO - parrotllm.training - step 17780 | epoch 0 | loss 3.3766 | lr 1.44e-03 | grad 0.1854 +2026-04-10 06:17:59 - INFO - parrotllm.training - step 17790 | epoch 0 | loss 3.5251 | lr 1.44e-03 | grad 0.1931 +2026-04-10 06:18:02 - INFO - parrotllm.training - step 17800 | epoch 0 | loss 3.4434 | lr 1.44e-03 | grad 0.2196 +2026-04-10 06:18:05 - INFO - parrotllm.training - step 17810 | epoch 0 | loss 3.4837 | lr 1.44e-03 | grad 0.1914 +2026-04-10 06:18:08 - INFO - parrotllm.training - step 17820 | epoch 0 | loss 3.4049 | lr 1.44e-03 | grad 0.2102 +2026-04-10 06:18:11 - INFO - parrotllm.training - step 17830 | epoch 0 | loss 3.5018 | lr 1.44e-03 | grad 0.1944 +2026-04-10 06:18:14 - INFO - parrotllm.training - step 17840 | epoch 0 | loss 3.4134 | lr 1.44e-03 | grad 0.2000 +2026-04-10 06:18:17 - INFO - parrotllm.training - step 17850 | epoch 0 | loss 3.4872 | lr 1.44e-03 | grad 0.2199 +2026-04-10 06:18:20 - INFO - parrotllm.training - step 17860 | epoch 0 | loss 3.4871 | lr 1.44e-03 | grad 0.1791 +2026-04-10 06:18:23 - INFO - parrotllm.training - step 17870 | epoch 0 | loss 3.5487 | lr 1.44e-03 | grad 0.1826 +2026-04-10 06:18:26 - INFO - parrotllm.training - step 17880 | epoch 0 | loss 3.5009 | lr 1.44e-03 | grad 0.1778 +2026-04-10 06:18:29 - INFO - parrotllm.training - step 17890 | epoch 0 | loss 3.4966 | lr 1.44e-03 | grad 0.1861 +2026-04-10 06:18:32 - INFO - parrotllm.training - step 17900 | epoch 0 | loss 3.4950 | lr 1.44e-03 | grad 0.1955 +2026-04-10 06:18:35 - INFO - parrotllm.training - step 17910 | epoch 0 | loss 3.3784 | lr 1.44e-03 | grad 0.1913 +2026-04-10 06:18:38 - INFO - parrotllm.training - step 17920 | epoch 0 | loss 3.4218 | lr 1.44e-03 | grad 0.2287 +2026-04-10 06:18:41 - INFO - parrotllm.training - step 17930 | epoch 0 | loss 3.4916 | lr 1.44e-03 | grad 0.1771 +2026-04-10 06:18:44 - INFO - parrotllm.training - step 17940 | epoch 0 | loss 3.5228 | lr 1.44e-03 | grad 0.1835 +2026-04-10 06:18:47 - INFO - parrotllm.training - step 17950 | epoch 0 | loss 3.4492 | lr 1.44e-03 | grad 0.2148 +2026-04-10 06:18:50 - INFO - parrotllm.training - step 17960 | epoch 0 | loss 3.4584 | lr 1.44e-03 | grad 0.2044 +2026-04-10 06:18:54 - INFO - parrotllm.training - step 17970 | epoch 0 | loss 3.4708 | lr 1.44e-03 | grad 0.2155 +2026-04-10 06:18:57 - INFO - parrotllm.training - step 17980 | epoch 0 | loss 3.4625 | lr 1.44e-03 | grad 0.2005 +2026-04-10 06:19:00 - INFO - parrotllm.training - step 17990 | epoch 0 | loss 3.5122 | lr 1.44e-03 | grad 0.1857 +2026-04-10 06:19:03 - INFO - parrotllm.training - step 18000 | epoch 0 | loss 3.4708 | lr 1.44e-03 | grad 0.2089 +2026-04-10 06:19:03 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:19:03 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:19:06 - INFO - parrotllm.training - Train: loss=3.4708, ppl=32.16 +2026-04-10 06:19:06 - INFO - parrotllm.training - Val: loss=3.4351, ppl=31.04 +2026-04-10 06:19:07 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4351_epoch_0000_step_0018000.pt +2026-04-10 06:19:08 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4349, min_delta=0.001000). +2026-04-10 06:19:08 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:19:11 - INFO - parrotllm.training - step 18010 | epoch 0 | loss 3.4909 | lr 1.44e-03 | grad 0.1828 +2026-04-10 06:19:14 - INFO - parrotllm.training - step 18020 | epoch 0 | loss 3.4242 | lr 1.44e-03 | grad 0.2115 +2026-04-10 06:19:17 - INFO - parrotllm.training - step 18030 | epoch 0 | loss 3.4188 | lr 1.44e-03 | grad 0.1942 +2026-04-10 06:19:20 - INFO - parrotllm.training - step 18040 | epoch 0 | loss 3.4263 | lr 1.44e-03 | grad 0.1976 +2026-04-10 06:19:23 - INFO - parrotllm.training - step 18050 | epoch 0 | loss 3.5238 | lr 1.44e-03 | grad 0.1991 +2026-04-10 06:19:26 - INFO - parrotllm.training - step 18060 | epoch 0 | loss 3.4298 | lr 1.44e-03 | grad 0.1950 +2026-04-10 06:19:29 - INFO - parrotllm.training - step 18070 | epoch 0 | loss 3.4492 | lr 1.44e-03 | grad 0.1946 +2026-04-10 06:19:32 - INFO - parrotllm.training - step 18080 | epoch 0 | loss 3.3754 | lr 1.44e-03 | grad 0.1970 +2026-04-10 06:19:35 - INFO - parrotllm.training - step 18090 | epoch 0 | loss 3.5244 | lr 1.44e-03 | grad 0.1953 +2026-04-10 06:19:38 - INFO - parrotllm.training - step 18100 | epoch 0 | loss 3.6290 | lr 1.44e-03 | grad 0.2131 +2026-04-10 06:19:41 - INFO - parrotllm.training - step 18110 | epoch 0 | loss 3.5267 | lr 1.44e-03 | grad 0.1914 +2026-04-10 06:19:44 - INFO - parrotllm.training - step 18120 | epoch 0 | loss 3.5962 | lr 1.44e-03 | grad 0.1866 +2026-04-10 06:19:47 - INFO - parrotllm.training - step 18130 | epoch 0 | loss 3.4807 | lr 1.44e-03 | grad 0.1927 +2026-04-10 06:19:50 - INFO - parrotllm.training - step 18140 | epoch 0 | loss 3.4361 | lr 1.44e-03 | grad 0.2008 +2026-04-10 06:19:53 - INFO - parrotllm.training - step 18150 | epoch 0 | loss 3.5012 | lr 1.44e-03 | grad 0.2059 +2026-04-10 06:19:57 - INFO - parrotllm.training - step 18160 | epoch 0 | loss 3.4922 | lr 1.44e-03 | grad 0.1901 +2026-04-10 06:20:00 - INFO - parrotllm.training - step 18170 | epoch 0 | loss 3.4297 | lr 1.44e-03 | grad 0.1864 +2026-04-10 06:20:03 - INFO - parrotllm.training - step 18180 | epoch 0 | loss 3.5156 | lr 1.44e-03 | grad 0.2007 +2026-04-10 06:20:06 - INFO - parrotllm.training - step 18190 | epoch 0 | loss 3.5199 | lr 1.44e-03 | grad 0.1741 +2026-04-10 06:20:09 - INFO - parrotllm.training - step 18200 | epoch 0 | loss 3.5047 | lr 1.44e-03 | grad 0.1942 +2026-04-10 06:20:12 - INFO - parrotllm.training - step 18210 | epoch 0 | loss 3.4869 | lr 1.44e-03 | grad 0.1851 +2026-04-10 06:20:15 - INFO - parrotllm.training - step 18220 | epoch 0 | loss 3.5414 | lr 1.44e-03 | grad 0.1907 +2026-04-10 06:20:18 - INFO - parrotllm.training - step 18230 | epoch 0 | loss 3.4654 | lr 1.44e-03 | grad 0.2286 +2026-04-10 06:20:21 - INFO - parrotllm.training - step 18240 | epoch 0 | loss 3.4993 | lr 1.44e-03 | grad 0.1952 +2026-04-10 06:20:24 - INFO - parrotllm.training - step 18250 | epoch 0 | loss 3.4622 | lr 1.44e-03 | grad 0.1948 +2026-04-10 06:20:27 - INFO - parrotllm.training - step 18260 | epoch 0 | loss 3.4650 | lr 1.44e-03 | grad 0.2001 +2026-04-10 06:20:30 - INFO - parrotllm.training - step 18270 | epoch 0 | loss 3.4065 | lr 1.44e-03 | grad 0.1819 +2026-04-10 06:20:33 - INFO - parrotllm.training - step 18280 | epoch 0 | loss 3.5481 | lr 1.44e-03 | grad 0.1994 +2026-04-10 06:20:36 - INFO - parrotllm.training - step 18290 | epoch 0 | loss 3.3555 | lr 1.44e-03 | grad 0.1962 +2026-04-10 06:20:39 - INFO - parrotllm.training - step 18300 | epoch 0 | loss 3.5254 | lr 1.44e-03 | grad 0.1796 +2026-04-10 06:20:42 - INFO - parrotllm.training - step 18310 | epoch 0 | loss 3.4230 | lr 1.44e-03 | grad 0.2111 +2026-04-10 06:20:45 - INFO - parrotllm.training - step 18320 | epoch 0 | loss 3.4787 | lr 1.44e-03 | grad 0.2023 +2026-04-10 06:20:48 - INFO - parrotllm.training - step 18330 | epoch 0 | loss 3.3891 | lr 1.44e-03 | grad 0.1811 +2026-04-10 06:20:51 - INFO - parrotllm.training - step 18340 | epoch 0 | loss 3.4846 | lr 1.44e-03 | grad 0.2519 +2026-04-10 06:20:54 - INFO - parrotllm.training - step 18350 | epoch 0 | loss 3.6063 | lr 1.44e-03 | grad 0.2175 +2026-04-10 06:20:58 - INFO - parrotllm.training - step 18360 | epoch 0 | loss 3.4926 | lr 1.44e-03 | grad 0.1934 +2026-04-10 06:21:01 - INFO - parrotllm.training - step 18370 | epoch 0 | loss 3.3942 | lr 1.44e-03 | grad 0.2111 +2026-04-10 06:21:04 - INFO - parrotllm.training - step 18380 | epoch 0 | loss 3.5576 | lr 1.44e-03 | grad 0.2115 +2026-04-10 06:21:07 - INFO - parrotllm.training - step 18390 | epoch 0 | loss 3.6396 | lr 1.44e-03 | grad 0.2099 +2026-04-10 06:21:10 - INFO - parrotllm.training - step 18400 | epoch 0 | loss 3.5302 | lr 1.44e-03 | grad 0.1980 +2026-04-10 06:21:13 - INFO - parrotllm.training - step 18410 | epoch 0 | loss 3.4459 | lr 1.44e-03 | grad 0.2128 +2026-04-10 06:21:16 - INFO - parrotllm.training - step 18420 | epoch 0 | loss 3.4940 | lr 1.44e-03 | grad 0.1879 +2026-04-10 06:21:19 - INFO - parrotllm.training - step 18430 | epoch 0 | loss 3.5532 | lr 1.44e-03 | grad 0.1902 +2026-04-10 06:21:22 - INFO - parrotllm.training - step 18440 | epoch 0 | loss 3.4344 | lr 1.44e-03 | grad 0.2389 +2026-04-10 06:21:25 - INFO - parrotllm.training - step 18450 | epoch 0 | loss 3.4292 | lr 1.44e-03 | grad 0.1942 +2026-04-10 06:21:28 - INFO - parrotllm.training - step 18460 | epoch 0 | loss 3.5477 | lr 1.44e-03 | grad 0.1908 +2026-04-10 06:21:31 - INFO - parrotllm.training - step 18470 | epoch 0 | loss 3.4499 | lr 1.44e-03 | grad 0.2064 +2026-04-10 06:21:34 - INFO - parrotllm.training - step 18480 | epoch 0 | loss 3.4803 | lr 1.44e-03 | grad 0.2031 +2026-04-10 06:21:37 - INFO - parrotllm.training - step 18490 | epoch 0 | loss 3.6050 | lr 1.43e-03 | grad 0.1809 +2026-04-10 06:21:40 - INFO - parrotllm.training - step 18500 | epoch 0 | loss 3.4865 | lr 1.43e-03 | grad 0.2098 +2026-04-10 06:21:40 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:21:40 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:21:43 - INFO - parrotllm.training - Train: loss=3.4865, ppl=32.67 +2026-04-10 06:21:43 - INFO - parrotllm.training - Val: loss=3.4334, ppl=30.98 +2026-04-10 06:21:43 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:21:44 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4334_epoch_0000_step_0018500.pt +2026-04-10 06:21:45 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:21:48 - INFO - parrotllm.training - step 18510 | epoch 0 | loss 3.5224 | lr 1.43e-03 | grad 0.2131 +2026-04-10 06:21:51 - INFO - parrotllm.training - step 18520 | epoch 0 | loss 3.4280 | lr 1.43e-03 | grad 0.2396 +2026-04-10 06:21:54 - INFO - parrotllm.training - step 18530 | epoch 0 | loss 3.6403 | lr 1.43e-03 | grad 0.1904 +2026-04-10 06:21:57 - INFO - parrotllm.training - step 18540 | epoch 0 | loss 3.4836 | lr 1.43e-03 | grad 0.1938 +2026-04-10 06:22:01 - INFO - parrotllm.training - step 18550 | epoch 0 | loss 3.5222 | lr 1.43e-03 | grad 0.2015 +2026-04-10 06:22:04 - INFO - parrotllm.training - step 18560 | epoch 0 | loss 3.4014 | lr 1.43e-03 | grad 0.2230 +2026-04-10 06:22:07 - INFO - parrotllm.training - step 18570 | epoch 0 | loss 3.4038 | lr 1.43e-03 | grad 0.1894 +2026-04-10 06:22:10 - INFO - parrotllm.training - step 18580 | epoch 0 | loss 3.5439 | lr 1.43e-03 | grad 0.2282 +2026-04-10 06:22:13 - INFO - parrotllm.training - step 18590 | epoch 0 | loss 3.5344 | lr 1.43e-03 | grad 0.2080 +2026-04-10 06:22:16 - INFO - parrotllm.training - step 18600 | epoch 0 | loss 3.5772 | lr 1.43e-03 | grad 0.2000 +2026-04-10 06:22:19 - INFO - parrotllm.training - step 18610 | epoch 0 | loss 3.4909 | lr 1.43e-03 | grad 0.2426 +2026-04-10 06:22:22 - INFO - parrotllm.training - step 18620 | epoch 0 | loss 3.5194 | lr 1.43e-03 | grad 0.1794 +2026-04-10 06:22:25 - INFO - parrotllm.training - step 18630 | epoch 0 | loss 3.5117 | lr 1.43e-03 | grad 0.1998 +2026-04-10 06:22:28 - INFO - parrotllm.training - step 18640 | epoch 0 | loss 3.5431 | lr 1.43e-03 | grad 0.2136 +2026-04-10 06:22:31 - INFO - parrotllm.training - step 18650 | epoch 0 | loss 3.4473 | lr 1.43e-03 | grad 0.1923 +2026-04-10 06:22:34 - INFO - parrotllm.training - step 18660 | epoch 0 | loss 3.5685 | lr 1.43e-03 | grad 0.1957 +2026-04-10 06:22:37 - INFO - parrotllm.training - step 18670 | epoch 0 | loss 3.5274 | lr 1.43e-03 | grad 0.2257 +2026-04-10 06:22:40 - INFO - parrotllm.training - step 18680 | epoch 0 | loss 3.4724 | lr 1.43e-03 | grad 0.2111 +2026-04-10 06:22:43 - INFO - parrotllm.training - step 18690 | epoch 0 | loss 3.4870 | lr 1.43e-03 | grad 0.1724 +2026-04-10 06:22:46 - INFO - parrotllm.training - step 18700 | epoch 0 | loss 3.4687 | lr 1.43e-03 | grad 0.2335 +2026-04-10 06:22:49 - INFO - parrotllm.training - step 18710 | epoch 0 | loss 3.4962 | lr 1.43e-03 | grad 0.1781 +2026-04-10 06:22:52 - INFO - parrotllm.training - step 18720 | epoch 0 | loss 3.4538 | lr 1.43e-03 | grad 0.2046 +2026-04-10 06:22:56 - INFO - parrotllm.training - step 18730 | epoch 0 | loss 3.3803 | lr 1.43e-03 | grad 0.2063 +2026-04-10 06:22:59 - INFO - parrotllm.training - step 18740 | epoch 0 | loss 3.5093 | lr 1.43e-03 | grad 0.1981 +2026-04-10 06:23:02 - INFO - parrotllm.training - step 18750 | epoch 0 | loss 3.3838 | lr 1.43e-03 | grad 0.1890 +2026-04-10 06:23:05 - INFO - parrotllm.training - step 18760 | epoch 0 | loss 3.5925 | lr 1.43e-03 | grad 0.1968 +2026-04-10 06:23:08 - INFO - parrotllm.training - step 18770 | epoch 0 | loss 3.5157 | lr 1.43e-03 | grad 0.2513 +2026-04-10 06:23:11 - INFO - parrotllm.training - step 18780 | epoch 0 | loss 3.5063 | lr 1.43e-03 | grad 0.2187 +2026-04-10 06:23:14 - INFO - parrotllm.training - step 18790 | epoch 0 | loss 3.4739 | lr 1.43e-03 | grad 0.1973 +2026-04-10 06:23:17 - INFO - parrotllm.training - step 18800 | epoch 0 | loss 3.5266 | lr 1.43e-03 | grad 0.1921 +2026-04-10 06:23:20 - INFO - parrotllm.training - step 18810 | epoch 0 | loss 3.5297 | lr 1.43e-03 | grad 0.2650 +2026-04-10 06:23:23 - INFO - parrotllm.training - step 18820 | epoch 0 | loss 3.4519 | lr 1.43e-03 | grad 0.2100 +2026-04-10 06:23:26 - INFO - parrotllm.training - step 18830 | epoch 0 | loss 3.5028 | lr 1.43e-03 | grad 0.2054 +2026-04-10 06:23:29 - INFO - parrotllm.training - step 18840 | epoch 0 | loss 3.5073 | lr 1.43e-03 | grad 0.2117 +2026-04-10 06:23:32 - INFO - parrotllm.training - step 18850 | epoch 0 | loss 3.5774 | lr 1.43e-03 | grad 0.1915 +2026-04-10 06:23:35 - INFO - parrotllm.training - step 18860 | epoch 0 | loss 3.6016 | lr 1.43e-03 | grad 0.2255 +2026-04-10 06:23:38 - INFO - parrotllm.training - step 18870 | epoch 0 | loss 3.4698 | lr 1.43e-03 | grad 0.2167 +2026-04-10 06:23:41 - INFO - parrotllm.training - step 18880 | epoch 0 | loss 3.5240 | lr 1.43e-03 | grad 0.1977 +2026-04-10 06:23:44 - INFO - parrotllm.training - step 18890 | epoch 0 | loss 3.4625 | lr 1.43e-03 | grad 0.1996 +2026-04-10 06:23:47 - INFO - parrotllm.training - step 18900 | epoch 0 | loss 3.4553 | lr 1.43e-03 | grad 0.1889 +2026-04-10 06:23:50 - INFO - parrotllm.training - step 18910 | epoch 0 | loss 3.5106 | lr 1.43e-03 | grad 0.1993 +2026-04-10 06:23:54 - INFO - parrotllm.training - step 18920 | epoch 0 | loss 3.5269 | lr 1.43e-03 | grad 0.2302 +2026-04-10 06:23:57 - INFO - parrotllm.training - step 18930 | epoch 0 | loss 3.4344 | lr 1.43e-03 | grad 0.2007 +2026-04-10 06:24:00 - INFO - parrotllm.training - step 18940 | epoch 0 | loss 3.5589 | lr 1.43e-03 | grad 0.2060 +2026-04-10 06:24:03 - INFO - parrotllm.training - step 18950 | epoch 0 | loss 3.5394 | lr 1.43e-03 | grad 0.2136 +2026-04-10 06:24:06 - INFO - parrotllm.training - step 18960 | epoch 0 | loss 3.5081 | lr 1.43e-03 | grad 0.2498 +2026-04-10 06:24:09 - INFO - parrotllm.training - step 18970 | epoch 0 | loss 3.3908 | lr 1.43e-03 | grad 0.1892 +2026-04-10 06:24:12 - INFO - parrotllm.training - step 18980 | epoch 0 | loss 3.4046 | lr 1.43e-03 | grad 0.2228 +2026-04-10 06:24:15 - INFO - parrotllm.training - step 18990 | epoch 0 | loss 3.4700 | lr 1.43e-03 | grad 0.1976 +2026-04-10 06:24:18 - INFO - parrotllm.training - step 19000 | epoch 0 | loss 3.4772 | lr 1.43e-03 | grad 0.1813 +2026-04-10 06:24:18 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:24:18 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:24:21 - INFO - parrotllm.training - Train: loss=3.4772, ppl=32.37 +2026-04-10 06:24:21 - INFO - parrotllm.training - Val: loss=3.4322, ppl=30.94 +2026-04-10 06:24:21 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:24:22 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4322_epoch_0000_step_0019000.pt +2026-04-10 06:24:23 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:24:26 - INFO - parrotllm.training - step 19010 | epoch 0 | loss 3.4331 | lr 1.43e-03 | grad 0.2305 +2026-04-10 06:24:29 - INFO - parrotllm.training - step 19020 | epoch 0 | loss 3.5912 | lr 1.43e-03 | grad 0.1809 +2026-04-10 06:24:32 - INFO - parrotllm.training - step 19030 | epoch 0 | loss 3.5605 | lr 1.43e-03 | grad 0.2238 +2026-04-10 06:24:35 - INFO - parrotllm.training - step 19040 | epoch 0 | loss 3.5373 | lr 1.43e-03 | grad 0.1862 +2026-04-10 06:24:38 - INFO - parrotllm.training - step 19050 | epoch 0 | loss 3.5063 | lr 1.43e-03 | grad 0.1956 +2026-04-10 06:24:41 - INFO - parrotllm.training - step 19060 | epoch 0 | loss 3.5143 | lr 1.43e-03 | grad 0.2046 +2026-04-10 06:24:44 - INFO - parrotllm.training - step 19070 | epoch 0 | loss 3.5131 | lr 1.43e-03 | grad 0.1941 +2026-04-10 06:24:47 - INFO - parrotllm.training - step 19080 | epoch 0 | loss 3.5038 | lr 1.43e-03 | grad 0.2148 +2026-04-10 06:24:50 - INFO - parrotllm.training - step 19090 | epoch 0 | loss 3.5122 | lr 1.43e-03 | grad 0.1996 +2026-04-10 06:24:53 - INFO - parrotllm.training - step 19100 | epoch 0 | loss 3.4532 | lr 1.43e-03 | grad 0.1693 +2026-04-10 06:24:56 - INFO - parrotllm.training - step 19110 | epoch 0 | loss 3.5072 | lr 1.43e-03 | grad 0.2477 +2026-04-10 06:24:59 - INFO - parrotllm.training - step 19120 | epoch 0 | loss 3.4910 | lr 1.43e-03 | grad 0.2084 +2026-04-10 06:25:03 - INFO - parrotllm.training - step 19130 | epoch 0 | loss 3.5200 | lr 1.43e-03 | grad 0.2304 +2026-04-10 06:25:06 - INFO - parrotllm.training - step 19140 | epoch 0 | loss 3.3948 | lr 1.43e-03 | grad 0.2171 +2026-04-10 06:25:09 - INFO - parrotllm.training - step 19150 | epoch 0 | loss 3.5068 | lr 1.43e-03 | grad 0.2047 +2026-04-10 06:25:12 - INFO - parrotllm.training - step 19160 | epoch 0 | loss 3.5469 | lr 1.43e-03 | grad 0.2282 +2026-04-10 06:25:15 - INFO - parrotllm.training - step 19170 | epoch 0 | loss 3.5439 | lr 1.43e-03 | grad 0.1937 +2026-04-10 06:25:18 - INFO - parrotllm.training - step 19180 | epoch 0 | loss 3.5034 | lr 1.43e-03 | grad 0.2774 +2026-04-10 06:25:21 - INFO - parrotllm.training - step 19190 | epoch 0 | loss 3.5465 | lr 1.43e-03 | grad 0.1942 +2026-04-10 06:25:24 - INFO - parrotllm.training - step 19200 | epoch 0 | loss 3.3826 | lr 1.43e-03 | grad 0.2029 +2026-04-10 06:25:27 - INFO - parrotllm.training - step 19210 | epoch 0 | loss 3.4389 | lr 1.43e-03 | grad 0.1991 +2026-04-10 06:25:30 - INFO - parrotllm.training - step 19220 | epoch 0 | loss 3.3547 | lr 1.43e-03 | grad 0.1924 +2026-04-10 06:25:33 - INFO - parrotllm.training - step 19230 | epoch 0 | loss 3.4889 | lr 1.43e-03 | grad 0.2195 +2026-04-10 06:25:36 - INFO - parrotllm.training - step 19240 | epoch 0 | loss 3.4648 | lr 1.43e-03 | grad 0.1752 +2026-04-10 06:25:39 - INFO - parrotllm.training - step 19250 | epoch 0 | loss 3.5018 | lr 1.43e-03 | grad 0.2218 +2026-04-10 06:25:42 - INFO - parrotllm.training - step 19260 | epoch 0 | loss 3.4402 | lr 1.43e-03 | grad 0.1968 +2026-04-10 06:25:45 - INFO - parrotllm.training - step 19270 | epoch 0 | loss 3.4295 | lr 1.43e-03 | grad 0.2232 +2026-04-10 06:25:48 - INFO - parrotllm.training - step 19280 | epoch 0 | loss 3.4681 | lr 1.43e-03 | grad 0.2216 +2026-04-10 06:25:51 - INFO - parrotllm.training - step 19290 | epoch 0 | loss 3.4056 | lr 1.43e-03 | grad 0.2562 +2026-04-10 06:25:54 - INFO - parrotllm.training - step 19300 | epoch 0 | loss 3.5090 | lr 1.43e-03 | grad 0.1870 +2026-04-10 06:25:57 - INFO - parrotllm.training - step 19310 | epoch 0 | loss 3.3619 | lr 1.43e-03 | grad 0.2276 +2026-04-10 06:26:01 - INFO - parrotllm.training - step 19320 | epoch 0 | loss 3.4518 | lr 1.43e-03 | grad 0.2214 +2026-04-10 06:26:04 - INFO - parrotllm.training - step 19330 | epoch 0 | loss 3.5400 | lr 1.43e-03 | grad 0.2054 +2026-04-10 06:26:07 - INFO - parrotllm.training - step 19340 | epoch 0 | loss 3.4866 | lr 1.43e-03 | grad 0.1729 +2026-04-10 06:26:10 - INFO - parrotllm.training - step 19350 | epoch 0 | loss 3.4659 | lr 1.43e-03 | grad 0.2030 +2026-04-10 06:26:13 - INFO - parrotllm.training - step 19360 | epoch 0 | loss 3.5120 | lr 1.43e-03 | grad 0.1903 +2026-04-10 06:26:16 - INFO - parrotllm.training - step 19370 | epoch 0 | loss 3.5093 | lr 1.43e-03 | grad 0.2009 +2026-04-10 06:26:19 - INFO - parrotllm.training - step 19380 | epoch 0 | loss 3.4301 | lr 1.43e-03 | grad 0.2077 +2026-04-10 06:26:22 - INFO - parrotllm.training - step 19390 | epoch 0 | loss 3.5613 | lr 1.43e-03 | grad 0.1847 +2026-04-10 06:26:25 - INFO - parrotllm.training - step 19400 | epoch 0 | loss 3.5199 | lr 1.43e-03 | grad 0.1864 +2026-04-10 06:26:28 - INFO - parrotllm.training - step 19410 | epoch 0 | loss 3.4768 | lr 1.43e-03 | grad 0.1779 +2026-04-10 06:26:31 - INFO - parrotllm.training - step 19420 | epoch 0 | loss 3.6041 | lr 1.43e-03 | grad 0.2033 +2026-04-10 06:26:34 - INFO - parrotllm.training - step 19430 | epoch 0 | loss 3.4927 | lr 1.43e-03 | grad 0.2141 +2026-04-10 06:26:37 - INFO - parrotllm.training - step 19440 | epoch 0 | loss 3.4905 | lr 1.43e-03 | grad 0.2161 +2026-04-10 06:26:40 - INFO - parrotllm.training - step 19450 | epoch 0 | loss 3.5000 | lr 1.43e-03 | grad 0.2209 +2026-04-10 06:26:42 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 19463/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 06:26:44 - INFO - parrotllm.training - step 19460 | epoch 0 | loss 3.3441 | lr 1.43e-03 | grad 0.2359 +2026-04-10 06:26:47 - INFO - parrotllm.training - step 19470 | epoch 0 | loss 3.4602 | lr 1.43e-03 | grad 0.2005 +2026-04-10 06:26:50 - INFO - parrotllm.training - step 19480 | epoch 0 | loss 3.4764 | lr 1.43e-03 | grad 0.1999 +2026-04-10 06:26:53 - INFO - parrotllm.training - step 19490 | epoch 0 | loss 3.5171 | lr 1.43e-03 | grad 0.1886 +2026-04-10 06:26:56 - INFO - parrotllm.training - step 19500 | epoch 0 | loss 3.4654 | lr 1.43e-03 | grad 0.2494 +2026-04-10 06:26:56 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:26:56 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:26:59 - INFO - parrotllm.training - Train: loss=3.4654, ppl=31.99 +2026-04-10 06:26:59 - INFO - parrotllm.training - Val: loss=3.4298, ppl=30.87 +2026-04-10 06:26:59 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:27:00 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4298_epoch_0000_step_0019500.pt +2026-04-10 06:27:01 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:27:04 - INFO - parrotllm.training - step 19510 | epoch 0 | loss 3.4824 | lr 1.43e-03 | grad 0.1896 +2026-04-10 06:27:07 - INFO - parrotllm.training - step 19520 | epoch 0 | loss 3.5523 | lr 1.43e-03 | grad 0.2106 +2026-04-10 06:27:10 - INFO - parrotllm.training - step 19530 | epoch 0 | loss 3.4935 | lr 1.43e-03 | grad 0.2381 +2026-04-10 06:27:13 - INFO - parrotllm.training - step 19540 | epoch 0 | loss 3.5621 | lr 1.43e-03 | grad 0.2129 +2026-04-10 06:27:16 - INFO - parrotllm.training - step 19550 | epoch 0 | loss 3.5180 | lr 1.43e-03 | grad 0.1878 +2026-04-10 06:27:19 - INFO - parrotllm.training - step 19560 | epoch 0 | loss 3.4169 | lr 1.43e-03 | grad 0.1843 +2026-04-10 06:27:22 - INFO - parrotllm.training - step 19570 | epoch 0 | loss 3.4734 | lr 1.43e-03 | grad 0.2134 +2026-04-10 06:27:25 - INFO - parrotllm.training - step 19580 | epoch 0 | loss 3.6197 | lr 1.43e-03 | grad 0.2095 +2026-04-10 06:27:28 - INFO - parrotllm.training - step 19590 | epoch 0 | loss 3.4134 | lr 1.43e-03 | grad 0.2177 +2026-04-10 06:27:31 - INFO - parrotllm.training - step 19600 | epoch 0 | loss 3.5195 | lr 1.43e-03 | grad 0.2170 +2026-04-10 06:27:34 - INFO - parrotllm.training - step 19610 | epoch 0 | loss 3.5450 | lr 1.43e-03 | grad 0.1987 +2026-04-10 06:27:37 - INFO - parrotllm.training - step 19620 | epoch 0 | loss 3.3656 | lr 1.43e-03 | grad 0.2023 +2026-04-10 06:27:40 - INFO - parrotllm.training - step 19630 | epoch 0 | loss 3.4983 | lr 1.43e-03 | grad 0.1971 +2026-04-10 06:27:43 - INFO - parrotllm.training - step 19640 | epoch 0 | loss 3.5746 | lr 1.43e-03 | grad 0.2191 +2026-04-10 06:27:47 - INFO - parrotllm.training - step 19650 | epoch 0 | loss 3.4951 | lr 1.43e-03 | grad 0.2270 +2026-04-10 06:27:50 - INFO - parrotllm.training - step 19660 | epoch 0 | loss 3.4467 | lr 1.43e-03 | grad 0.2020 +2026-04-10 06:27:53 - INFO - parrotllm.training - step 19670 | epoch 0 | loss 3.5539 | lr 1.43e-03 | grad 0.2024 +2026-04-10 06:27:56 - INFO - parrotllm.training - step 19680 | epoch 0 | loss 3.4779 | lr 1.43e-03 | grad 0.2204 +2026-04-10 06:27:59 - INFO - parrotllm.training - step 19690 | epoch 0 | loss 3.4976 | lr 1.43e-03 | grad 0.1879 +2026-04-10 06:28:02 - INFO - parrotllm.training - step 19700 | epoch 0 | loss 3.4541 | lr 1.43e-03 | grad 0.2133 +2026-04-10 06:28:05 - INFO - parrotllm.training - step 19710 | epoch 0 | loss 3.4880 | lr 1.43e-03 | grad 0.1838 +2026-04-10 06:28:08 - INFO - parrotllm.training - step 19720 | epoch 0 | loss 3.5200 | lr 1.43e-03 | grad 0.1865 +2026-04-10 06:28:11 - INFO - parrotllm.training - step 19730 | epoch 0 | loss 3.4899 | lr 1.43e-03 | grad 0.2064 +2026-04-10 06:28:14 - INFO - parrotllm.training - step 19740 | epoch 0 | loss 3.5226 | lr 1.43e-03 | grad 0.1903 +2026-04-10 06:28:17 - INFO - parrotllm.training - step 19750 | epoch 0 | loss 3.4707 | lr 1.43e-03 | grad 0.1873 +2026-04-10 06:28:20 - INFO - parrotllm.training - step 19760 | epoch 0 | loss 3.4197 | lr 1.43e-03 | grad 0.1994 +2026-04-10 06:28:23 - INFO - parrotllm.training - step 19770 | epoch 0 | loss 3.4554 | lr 1.42e-03 | grad 0.1933 +2026-04-10 06:28:26 - INFO - parrotllm.training - step 19780 | epoch 0 | loss 3.4590 | lr 1.42e-03 | grad 0.1971 +2026-04-10 06:28:29 - INFO - parrotllm.training - step 19790 | epoch 0 | loss 3.5003 | lr 1.42e-03 | grad 0.1900 +2026-04-10 06:28:32 - INFO - parrotllm.training - step 19800 | epoch 0 | loss 3.5871 | lr 1.42e-03 | grad 0.2212 +2026-04-10 06:28:35 - INFO - parrotllm.training - step 19810 | epoch 0 | loss 3.5152 | lr 1.42e-03 | grad 0.2264 +2026-04-10 06:28:38 - INFO - parrotllm.training - step 19820 | epoch 0 | loss 3.4594 | lr 1.42e-03 | grad 0.1947 +2026-04-10 06:28:41 - INFO - parrotllm.training - step 19830 | epoch 0 | loss 3.5916 | lr 1.42e-03 | grad 0.1780 +2026-04-10 06:28:45 - INFO - parrotllm.training - step 19840 | epoch 0 | loss 3.5436 | lr 1.42e-03 | grad 0.2125 +2026-04-10 06:28:48 - INFO - parrotllm.training - step 19850 | epoch 0 | loss 3.5412 | lr 1.42e-03 | grad 0.2294 +2026-04-10 06:28:51 - INFO - parrotllm.training - step 19860 | epoch 0 | loss 3.5044 | lr 1.42e-03 | grad 0.2045 +2026-04-10 06:28:54 - INFO - parrotllm.training - step 19870 | epoch 0 | loss 3.4854 | lr 1.42e-03 | grad 0.2024 +2026-04-10 06:28:57 - INFO - parrotllm.training - step 19880 | epoch 0 | loss 3.4548 | lr 1.42e-03 | grad 0.2214 +2026-04-10 06:29:00 - INFO - parrotllm.training - step 19890 | epoch 0 | loss 3.5857 | lr 1.42e-03 | grad 0.2262 +2026-04-10 06:29:03 - INFO - parrotllm.training - step 19900 | epoch 0 | loss 3.5075 | lr 1.42e-03 | grad 0.2203 +2026-04-10 06:29:06 - INFO - parrotllm.training - step 19910 | epoch 0 | loss 3.4210 | lr 1.42e-03 | grad 0.2340 +2026-04-10 06:29:09 - INFO - parrotllm.training - step 19920 | epoch 0 | loss 3.3894 | lr 1.42e-03 | grad 0.2057 +2026-04-10 06:29:12 - INFO - parrotllm.training - step 19930 | epoch 0 | loss 3.5043 | lr 1.42e-03 | grad 0.2130 +2026-04-10 06:29:15 - INFO - parrotllm.training - step 19940 | epoch 0 | loss 3.6057 | lr 1.42e-03 | grad 0.1881 +2026-04-10 06:29:18 - INFO - parrotllm.training - step 19950 | epoch 0 | loss 3.5040 | lr 1.42e-03 | grad 0.2035 +2026-04-10 06:29:21 - INFO - parrotllm.training - step 19960 | epoch 0 | loss 3.4408 | lr 1.42e-03 | grad 0.2178 +2026-04-10 06:29:24 - INFO - parrotllm.training - step 19970 | epoch 0 | loss 3.4848 | lr 1.42e-03 | grad 0.1802 +2026-04-10 06:29:27 - INFO - parrotllm.training - step 19980 | epoch 0 | loss 3.5201 | lr 1.42e-03 | grad 0.2019 +2026-04-10 06:29:30 - INFO - parrotllm.training - step 19990 | epoch 0 | loss 3.5894 | lr 1.42e-03 | grad 0.1949 +2026-04-10 06:29:33 - INFO - parrotllm.training - step 20000 | epoch 0 | loss 3.4171 | lr 1.42e-03 | grad 0.1836 +2026-04-10 06:29:33 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:29:33 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:29:36 - INFO - parrotllm.training - Train: loss=3.4171, ppl=30.48 +2026-04-10 06:29:36 - INFO - parrotllm.training - Val: loss=3.4313, ppl=30.92 +2026-04-10 06:29:37 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4313_epoch_0000_step_0020000.pt +2026-04-10 06:29:38 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4298, min_delta=0.001000). +2026-04-10 06:29:38 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:29:39 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0020000.pt +2026-04-10 06:29:43 - INFO - parrotllm.training - step 20010 | epoch 0 | loss 3.6604 | lr 1.42e-03 | grad 0.2192 +2026-04-10 06:29:47 - INFO - parrotllm.training - step 20020 | epoch 0 | loss 3.5488 | lr 1.42e-03 | grad 0.1920 +2026-04-10 06:29:50 - INFO - parrotllm.training - step 20030 | epoch 0 | loss 3.4326 | lr 1.42e-03 | grad 0.1967 +2026-04-10 06:29:53 - INFO - parrotllm.training - step 20040 | epoch 0 | loss 3.5485 | lr 1.42e-03 | grad 0.2009 +2026-04-10 06:29:56 - INFO - parrotllm.training - step 20050 | epoch 0 | loss 3.5327 | lr 1.42e-03 | grad 0.2092 +2026-04-10 06:29:59 - INFO - parrotllm.training - step 20060 | epoch 0 | loss 3.5506 | lr 1.42e-03 | grad 0.2151 +2026-04-10 06:30:02 - INFO - parrotllm.training - step 20070 | epoch 0 | loss 3.5743 | lr 1.42e-03 | grad 0.2152 +2026-04-10 06:30:05 - INFO - parrotllm.training - step 20080 | epoch 0 | loss 3.4752 | lr 1.42e-03 | grad 0.1901 +2026-04-10 06:30:08 - INFO - parrotllm.training - step 20090 | epoch 0 | loss 3.4017 | lr 1.42e-03 | grad 0.2066 +2026-04-10 06:30:11 - INFO - parrotllm.training - step 20100 | epoch 0 | loss 3.4724 | lr 1.42e-03 | grad 0.1932 +2026-04-10 06:30:14 - INFO - parrotllm.training - step 20110 | epoch 0 | loss 3.4227 | lr 1.42e-03 | grad 0.1969 +2026-04-10 06:30:17 - INFO - parrotllm.training - step 20120 | epoch 0 | loss 3.5235 | lr 1.42e-03 | grad 0.1967 +2026-04-10 06:30:20 - INFO - parrotllm.training - step 20130 | epoch 0 | loss 3.6460 | lr 1.42e-03 | grad 0.1990 +2026-04-10 06:30:23 - INFO - parrotllm.training - step 20140 | epoch 0 | loss 3.4553 | lr 1.42e-03 | grad 0.2024 +2026-04-10 06:30:26 - INFO - parrotllm.training - step 20150 | epoch 0 | loss 3.4812 | lr 1.42e-03 | grad 0.2027 +2026-04-10 06:30:29 - INFO - parrotllm.training - step 20160 | epoch 0 | loss 3.5628 | lr 1.42e-03 | grad 0.1880 +2026-04-10 06:30:32 - INFO - parrotllm.training - step 20170 | epoch 0 | loss 3.5929 | lr 1.42e-03 | grad 0.1945 +2026-04-10 06:30:35 - INFO - parrotllm.training - step 20180 | epoch 0 | loss 3.4785 | lr 1.42e-03 | grad 0.2030 +2026-04-10 06:30:38 - INFO - parrotllm.training - step 20190 | epoch 0 | loss 3.5042 | lr 1.42e-03 | grad 0.2073 +2026-04-10 06:30:41 - INFO - parrotllm.training - step 20200 | epoch 0 | loss 3.4748 | lr 1.42e-03 | grad 0.1796 +2026-04-10 06:30:45 - INFO - parrotllm.training - step 20210 | epoch 0 | loss 3.4783 | lr 1.42e-03 | grad 0.1799 +2026-04-10 06:30:48 - INFO - parrotllm.training - step 20220 | epoch 0 | loss 3.4003 | lr 1.42e-03 | grad 0.2020 +2026-04-10 06:30:51 - INFO - parrotllm.training - step 20230 | epoch 0 | loss 3.4551 | lr 1.42e-03 | grad 0.2192 +2026-04-10 06:30:54 - INFO - parrotllm.training - step 20240 | epoch 0 | loss 3.5025 | lr 1.42e-03 | grad 0.2199 +2026-04-10 06:30:57 - INFO - parrotllm.training - step 20250 | epoch 0 | loss 3.4383 | lr 1.42e-03 | grad 0.2077 +2026-04-10 06:31:00 - INFO - parrotllm.training - step 20260 | epoch 0 | loss 3.6186 | lr 1.42e-03 | grad 0.2127 +2026-04-10 06:31:03 - INFO - parrotllm.training - step 20270 | epoch 0 | loss 3.5413 | lr 1.42e-03 | grad 0.1969 +2026-04-10 06:31:06 - INFO - parrotllm.training - step 20280 | epoch 0 | loss 3.4836 | lr 1.42e-03 | grad 0.1928 +2026-04-10 06:31:09 - INFO - parrotllm.training - step 20290 | epoch 0 | loss 3.4793 | lr 1.42e-03 | grad 0.2091 +2026-04-10 06:31:12 - INFO - parrotllm.training - step 20300 | epoch 0 | loss 3.5049 | lr 1.42e-03 | grad 0.2054 +2026-04-10 06:31:15 - INFO - parrotllm.training - step 20310 | epoch 0 | loss 3.4389 | lr 1.42e-03 | grad 0.1932 +2026-04-10 06:31:18 - INFO - parrotllm.training - step 20320 | epoch 0 | loss 3.2460 | lr 1.42e-03 | grad 0.1991 +2026-04-10 06:31:21 - INFO - parrotllm.training - step 20330 | epoch 0 | loss 3.5000 | lr 1.42e-03 | grad 0.1899 +2026-04-10 06:31:24 - INFO - parrotllm.training - step 20340 | epoch 0 | loss 3.4804 | lr 1.42e-03 | grad 0.1935 +2026-04-10 06:31:27 - INFO - parrotllm.training - step 20350 | epoch 0 | loss 3.4686 | lr 1.42e-03 | grad 0.1813 +2026-04-10 06:31:30 - INFO - parrotllm.training - step 20360 | epoch 0 | loss 3.5148 | lr 1.42e-03 | grad 0.2067 +2026-04-10 06:31:33 - INFO - parrotllm.training - step 20370 | epoch 0 | loss 3.2790 | lr 1.42e-03 | grad 0.1802 +2026-04-10 06:31:36 - INFO - parrotllm.training - step 20380 | epoch 0 | loss 3.5159 | lr 1.42e-03 | grad 0.2105 +2026-04-10 06:31:39 - INFO - parrotllm.training - step 20390 | epoch 0 | loss 3.5267 | lr 1.42e-03 | grad 0.1936 +2026-04-10 06:31:42 - INFO - parrotllm.training - step 20400 | epoch 0 | loss 3.4560 | lr 1.42e-03 | grad 0.2507 +2026-04-10 06:31:45 - INFO - parrotllm.training - step 20410 | epoch 0 | loss 3.5591 | lr 1.42e-03 | grad 0.2000 +2026-04-10 06:31:49 - INFO - parrotllm.training - step 20420 | epoch 0 | loss 3.3995 | lr 1.42e-03 | grad 0.2134 +2026-04-10 06:31:52 - INFO - parrotllm.training - step 20430 | epoch 0 | loss 3.5351 | lr 1.42e-03 | grad 0.2196 +2026-04-10 06:31:55 - INFO - parrotllm.training - step 20440 | epoch 0 | loss 3.3879 | lr 1.42e-03 | grad 0.1877 +2026-04-10 06:31:58 - INFO - parrotllm.training - step 20450 | epoch 0 | loss 3.3290 | lr 1.42e-03 | grad 0.2052 +2026-04-10 06:32:01 - INFO - parrotllm.training - step 20460 | epoch 0 | loss 3.4115 | lr 1.42e-03 | grad 0.2003 +2026-04-10 06:32:04 - INFO - parrotllm.training - step 20470 | epoch 0 | loss 3.4878 | lr 1.42e-03 | grad 0.1938 +2026-04-10 06:32:07 - INFO - parrotllm.training - step 20480 | epoch 0 | loss 3.5649 | lr 1.42e-03 | grad 0.2249 +2026-04-10 06:32:10 - INFO - parrotllm.training - step 20490 | epoch 0 | loss 3.4704 | lr 1.42e-03 | grad 0.2047 +2026-04-10 06:32:13 - INFO - parrotllm.training - step 20500 | epoch 0 | loss 3.5684 | lr 1.42e-03 | grad 0.1994 +2026-04-10 06:32:13 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:32:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:32:16 - INFO - parrotllm.training - Train: loss=3.5684, ppl=35.46 +2026-04-10 06:32:16 - INFO - parrotllm.training - Val: loss=3.4279, ppl=30.81 +2026-04-10 06:32:16 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:32:17 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4279_epoch_0000_step_0020500.pt +2026-04-10 06:32:18 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:32:21 - INFO - parrotllm.training - step 20510 | epoch 0 | loss 3.3853 | lr 1.42e-03 | grad 0.2375 +2026-04-10 06:32:24 - INFO - parrotllm.training - step 20520 | epoch 0 | loss 3.4744 | lr 1.42e-03 | grad 0.2287 +2026-04-10 06:32:27 - INFO - parrotllm.training - step 20530 | epoch 0 | loss 3.5209 | lr 1.42e-03 | grad 0.1945 +2026-04-10 06:32:30 - INFO - parrotllm.training - step 20540 | epoch 0 | loss 3.4563 | lr 1.42e-03 | grad 0.1924 +2026-04-10 06:32:33 - INFO - parrotllm.training - step 20550 | epoch 0 | loss 3.4676 | lr 1.42e-03 | grad 0.1919 +2026-04-10 06:32:36 - INFO - parrotllm.training - step 20560 | epoch 0 | loss 3.4045 | lr 1.42e-03 | grad 0.2306 +2026-04-10 06:32:39 - INFO - parrotllm.training - step 20570 | epoch 0 | loss 3.5060 | lr 1.42e-03 | grad 0.1984 +2026-04-10 06:32:42 - INFO - parrotllm.training - step 20580 | epoch 0 | loss 3.4763 | lr 1.42e-03 | grad 0.1976 +2026-04-10 06:32:46 - INFO - parrotllm.training - step 20590 | epoch 0 | loss 3.6100 | lr 1.42e-03 | grad 0.2264 +2026-04-10 06:32:49 - INFO - parrotllm.training - step 20600 | epoch 0 | loss 3.3648 | lr 1.42e-03 | grad 0.1996 +2026-04-10 06:32:52 - INFO - parrotllm.training - step 20610 | epoch 0 | loss 3.6055 | lr 1.42e-03 | grad 0.1766 +2026-04-10 06:32:55 - INFO - parrotllm.training - step 20620 | epoch 0 | loss 3.3383 | lr 1.42e-03 | grad 0.1836 +2026-04-10 06:32:58 - INFO - parrotllm.training - step 20630 | epoch 0 | loss 3.4994 | lr 1.42e-03 | grad 0.2141 +2026-04-10 06:33:01 - INFO - parrotllm.training - step 20640 | epoch 0 | loss 3.4677 | lr 1.42e-03 | grad 0.1775 +2026-04-10 06:33:04 - INFO - parrotllm.training - step 20650 | epoch 0 | loss 3.4601 | lr 1.42e-03 | grad 0.2267 +2026-04-10 06:33:07 - INFO - parrotllm.training - step 20660 | epoch 0 | loss 3.4970 | lr 1.42e-03 | grad 0.1785 +2026-04-10 06:33:10 - INFO - parrotllm.training - step 20670 | epoch 0 | loss 3.4498 | lr 1.42e-03 | grad 0.2024 +2026-04-10 06:33:13 - INFO - parrotllm.training - step 20680 | epoch 0 | loss 3.3950 | lr 1.42e-03 | grad 0.2493 +2026-04-10 06:33:16 - INFO - parrotllm.training - step 20690 | epoch 0 | loss 3.5370 | lr 1.42e-03 | grad 0.1910 +2026-04-10 06:33:19 - INFO - parrotllm.training - step 20700 | epoch 0 | loss 3.4116 | lr 1.42e-03 | grad 0.1928 +2026-04-10 06:33:22 - INFO - parrotllm.training - step 20710 | epoch 0 | loss 3.5032 | lr 1.42e-03 | grad 0.2229 +2026-04-10 06:33:25 - INFO - parrotllm.training - step 20720 | epoch 0 | loss 3.5533 | lr 1.42e-03 | grad 0.1921 +2026-04-10 06:33:28 - INFO - parrotllm.training - step 20730 | epoch 0 | loss 3.5027 | lr 1.42e-03 | grad 0.1914 +2026-04-10 06:33:31 - INFO - parrotllm.training - step 20740 | epoch 0 | loss 3.4499 | lr 1.42e-03 | grad 0.2034 +2026-04-10 06:33:34 - INFO - parrotllm.training - step 20750 | epoch 0 | loss 3.5002 | lr 1.42e-03 | grad 0.2024 +2026-04-10 06:33:37 - INFO - parrotllm.training - step 20760 | epoch 0 | loss 3.4541 | lr 1.42e-03 | grad 0.1913 +2026-04-10 06:33:40 - INFO - parrotllm.training - step 20770 | epoch 0 | loss 3.5071 | lr 1.42e-03 | grad 0.2076 +2026-04-10 06:33:43 - INFO - parrotllm.training - step 20780 | epoch 0 | loss 3.5140 | lr 1.42e-03 | grad 0.2200 +2026-04-10 06:33:47 - INFO - parrotllm.training - step 20790 | epoch 0 | loss 3.4233 | lr 1.42e-03 | grad 0.2118 +2026-04-10 06:33:50 - INFO - parrotllm.training - step 20800 | epoch 0 | loss 3.4055 | lr 1.42e-03 | grad 0.2212 +2026-04-10 06:33:53 - INFO - parrotllm.training - step 20810 | epoch 0 | loss 3.4716 | lr 1.42e-03 | grad 0.1959 +2026-04-10 06:33:56 - INFO - parrotllm.training - step 20820 | epoch 0 | loss 3.3923 | lr 1.42e-03 | grad 0.2359 +2026-04-10 06:33:59 - INFO - parrotllm.training - step 20830 | epoch 0 | loss 3.5569 | lr 1.42e-03 | grad 0.2025 +2026-04-10 06:34:02 - INFO - parrotllm.training - step 20840 | epoch 0 | loss 3.4326 | lr 1.42e-03 | grad 0.2232 +2026-04-10 06:34:05 - INFO - parrotllm.training - step 20850 | epoch 0 | loss 3.5042 | lr 1.42e-03 | grad 0.1769 +2026-04-10 06:34:08 - INFO - parrotllm.training - step 20860 | epoch 0 | loss 3.4909 | lr 1.42e-03 | grad 0.2072 +2026-04-10 06:34:11 - INFO - parrotllm.training - step 20870 | epoch 0 | loss 3.5782 | lr 1.42e-03 | grad 0.1892 +2026-04-10 06:34:14 - INFO - parrotllm.training - step 20880 | epoch 0 | loss 3.4881 | lr 1.42e-03 | grad 0.1749 +2026-04-10 06:34:17 - INFO - parrotllm.training - step 20890 | epoch 0 | loss 3.4967 | lr 1.42e-03 | grad 0.1957 +2026-04-10 06:34:20 - INFO - parrotllm.training - step 20900 | epoch 0 | loss 3.4702 | lr 1.42e-03 | grad 0.2060 +2026-04-10 06:34:23 - INFO - parrotllm.training - step 20910 | epoch 0 | loss 3.4327 | lr 1.42e-03 | grad 0.1897 +2026-04-10 06:34:26 - INFO - parrotllm.training - step 20920 | epoch 0 | loss 3.4713 | lr 1.42e-03 | grad 0.1779 +2026-04-10 06:34:29 - INFO - parrotllm.training - step 20930 | epoch 0 | loss 3.5042 | lr 1.42e-03 | grad 0.1991 +2026-04-10 06:34:32 - INFO - parrotllm.training - step 20940 | epoch 0 | loss 3.4530 | lr 1.42e-03 | grad 0.2564 +2026-04-10 06:34:35 - INFO - parrotllm.training - step 20950 | epoch 0 | loss 3.4360 | lr 1.42e-03 | grad 0.1959 +2026-04-10 06:34:38 - INFO - parrotllm.training - step 20960 | epoch 0 | loss 3.4765 | lr 1.42e-03 | grad 0.2123 +2026-04-10 06:34:41 - INFO - parrotllm.training - step 20970 | epoch 0 | loss 3.4919 | lr 1.41e-03 | grad 0.2095 +2026-04-10 06:34:44 - INFO - parrotllm.training - step 20980 | epoch 0 | loss 3.5400 | lr 1.41e-03 | grad 0.2110 +2026-04-10 06:34:48 - INFO - parrotllm.training - step 20990 | epoch 0 | loss 3.5153 | lr 1.41e-03 | grad 0.2514 +2026-04-10 06:34:51 - INFO - parrotllm.training - step 21000 | epoch 0 | loss 3.5407 | lr 1.41e-03 | grad 0.2118 +2026-04-10 06:34:51 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:34:51 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:34:54 - INFO - parrotllm.training - Train: loss=3.5407, ppl=34.49 +2026-04-10 06:34:54 - INFO - parrotllm.training - Val: loss=3.4275, ppl=30.80 +2026-04-10 06:34:55 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4275_epoch_0000_step_0021000.pt +2026-04-10 06:34:56 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4279, min_delta=0.001000). +2026-04-10 06:34:56 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:34:59 - INFO - parrotllm.training - step 21010 | epoch 0 | loss 3.5236 | lr 1.41e-03 | grad 0.1958 +2026-04-10 06:35:02 - INFO - parrotllm.training - step 21020 | epoch 0 | loss 3.4366 | lr 1.41e-03 | grad 0.2166 +2026-04-10 06:35:05 - INFO - parrotllm.training - step 21030 | epoch 0 | loss 3.4745 | lr 1.41e-03 | grad 0.2102 +2026-04-10 06:35:08 - INFO - parrotllm.training - step 21040 | epoch 0 | loss 3.4754 | lr 1.41e-03 | grad 0.1835 +2026-04-10 06:35:11 - INFO - parrotllm.training - step 21050 | epoch 0 | loss 3.4813 | lr 1.41e-03 | grad 0.1834 +2026-04-10 06:35:14 - INFO - parrotllm.training - step 21060 | epoch 0 | loss 3.4471 | lr 1.41e-03 | grad 0.2064 +2026-04-10 06:35:17 - INFO - parrotllm.training - step 21070 | epoch 0 | loss 3.4693 | lr 1.41e-03 | grad 0.2421 +2026-04-10 06:35:20 - INFO - parrotllm.training - step 21080 | epoch 0 | loss 3.4394 | lr 1.41e-03 | grad 0.1907 +2026-04-10 06:35:23 - INFO - parrotllm.training - step 21090 | epoch 0 | loss 3.4420 | lr 1.41e-03 | grad 0.2039 +2026-04-10 06:35:26 - INFO - parrotllm.training - step 21100 | epoch 0 | loss 3.4977 | lr 1.41e-03 | grad 0.1930 +2026-04-10 06:35:29 - INFO - parrotllm.training - step 21110 | epoch 0 | loss 3.4851 | lr 1.41e-03 | grad 0.2318 +2026-04-10 06:35:32 - INFO - parrotllm.training - step 21120 | epoch 0 | loss 3.5330 | lr 1.41e-03 | grad 0.1920 +2026-04-10 06:35:35 - INFO - parrotllm.training - step 21130 | epoch 0 | loss 3.5871 | lr 1.41e-03 | grad 0.2161 +2026-04-10 06:35:38 - INFO - parrotllm.training - step 21140 | epoch 0 | loss 3.2881 | lr 1.41e-03 | grad 0.1869 +2026-04-10 06:35:41 - INFO - parrotllm.training - step 21150 | epoch 0 | loss 3.4958 | lr 1.41e-03 | grad 0.2097 +2026-04-10 06:35:45 - INFO - parrotllm.training - step 21160 | epoch 0 | loss 3.4159 | lr 1.41e-03 | grad 0.1981 +2026-04-10 06:35:48 - INFO - parrotllm.training - step 21170 | epoch 0 | loss 3.4454 | lr 1.41e-03 | grad 0.1858 +2026-04-10 06:35:51 - INFO - parrotllm.training - step 21180 | epoch 0 | loss 3.5003 | lr 1.41e-03 | grad 0.1952 +2026-04-10 06:35:54 - INFO - parrotllm.training - step 21190 | epoch 0 | loss 3.4962 | lr 1.41e-03 | grad 0.2399 +2026-04-10 06:35:57 - INFO - parrotllm.training - step 21200 | epoch 0 | loss 3.4826 | lr 1.41e-03 | grad 0.2309 +2026-04-10 06:36:00 - INFO - parrotllm.training - step 21210 | epoch 0 | loss 3.5396 | lr 1.41e-03 | grad 0.1932 +2026-04-10 06:36:03 - INFO - parrotllm.training - step 21220 | epoch 0 | loss 3.4825 | lr 1.41e-03 | grad 0.1983 +2026-04-10 06:36:06 - INFO - parrotllm.training - step 21230 | epoch 0 | loss 3.4282 | lr 1.41e-03 | grad 0.2349 +2026-04-10 06:36:09 - INFO - parrotllm.training - step 21240 | epoch 0 | loss 3.5329 | lr 1.41e-03 | grad 0.2250 +2026-04-10 06:36:12 - INFO - parrotllm.training - step 21250 | epoch 0 | loss 3.5550 | lr 1.41e-03 | grad 0.1887 +2026-04-10 06:36:15 - INFO - parrotllm.training - step 21260 | epoch 0 | loss 3.4954 | lr 1.41e-03 | grad 0.2068 +2026-04-10 06:36:18 - INFO - parrotllm.training - step 21270 | epoch 0 | loss 3.5873 | lr 1.41e-03 | grad 0.1912 +2026-04-10 06:36:21 - INFO - parrotllm.training - step 21280 | epoch 0 | loss 3.4545 | lr 1.41e-03 | grad 0.2030 +2026-04-10 06:36:24 - INFO - parrotllm.training - step 21290 | epoch 0 | loss 3.4891 | lr 1.41e-03 | grad 0.2241 +2026-04-10 06:36:27 - INFO - parrotllm.training - step 21300 | epoch 0 | loss 3.5561 | lr 1.41e-03 | grad 0.1888 +2026-04-10 06:36:30 - INFO - parrotllm.training - step 21310 | epoch 0 | loss 3.5235 | lr 1.41e-03 | grad 0.1970 +2026-04-10 06:36:33 - INFO - parrotllm.training - step 21320 | epoch 0 | loss 3.4432 | lr 1.41e-03 | grad 0.2150 +2026-04-10 06:36:36 - INFO - parrotllm.training - step 21330 | epoch 0 | loss 3.4619 | lr 1.41e-03 | grad 0.1950 +2026-04-10 06:36:39 - INFO - parrotllm.training - step 21340 | epoch 0 | loss 3.5097 | lr 1.41e-03 | grad 0.1918 +2026-04-10 06:36:43 - INFO - parrotllm.training - step 21350 | epoch 0 | loss 3.4456 | lr 1.41e-03 | grad 0.2091 +2026-04-10 06:36:46 - INFO - parrotllm.training - step 21360 | epoch 0 | loss 3.4845 | lr 1.41e-03 | grad 0.1966 +2026-04-10 06:36:49 - INFO - parrotllm.training - step 21370 | epoch 0 | loss 3.3631 | lr 1.41e-03 | grad 0.2000 +2026-04-10 06:36:52 - INFO - parrotllm.training - step 21380 | epoch 0 | loss 3.4842 | lr 1.41e-03 | grad 0.2596 +2026-04-10 06:36:55 - INFO - parrotllm.training - step 21390 | epoch 0 | loss 3.3767 | lr 1.41e-03 | grad 0.2169 +2026-04-10 06:36:58 - INFO - parrotllm.training - step 21400 | epoch 0 | loss 3.5268 | lr 1.41e-03 | grad 0.2021 +2026-04-10 06:37:01 - INFO - parrotllm.training - step 21410 | epoch 0 | loss 3.5029 | lr 1.41e-03 | grad 0.1896 +2026-04-10 06:37:04 - INFO - parrotllm.training - step 21420 | epoch 0 | loss 3.5398 | lr 1.41e-03 | grad 0.2002 +2026-04-10 06:37:07 - INFO - parrotllm.training - step 21430 | epoch 0 | loss 3.4006 | lr 1.41e-03 | grad 0.1933 +2026-04-10 06:37:10 - INFO - parrotllm.training - step 21440 | epoch 0 | loss 3.3956 | lr 1.41e-03 | grad 0.2333 +2026-04-10 06:37:13 - INFO - parrotllm.training - step 21450 | epoch 0 | loss 3.4579 | lr 1.41e-03 | grad 0.2206 +2026-04-10 06:37:16 - INFO - parrotllm.training - step 21460 | epoch 0 | loss 3.3983 | lr 1.41e-03 | grad 0.2387 +2026-04-10 06:37:19 - INFO - parrotllm.training - step 21470 | epoch 0 | loss 3.5113 | lr 1.41e-03 | grad 0.2202 +2026-04-10 06:37:22 - INFO - parrotllm.training - step 21480 | epoch 0 | loss 3.6001 | lr 1.41e-03 | grad 0.1918 +2026-04-10 06:37:25 - INFO - parrotllm.training - step 21490 | epoch 0 | loss 3.4105 | lr 1.41e-03 | grad 0.1961 +2026-04-10 06:37:28 - INFO - parrotllm.training - step 21500 | epoch 0 | loss 3.6126 | lr 1.41e-03 | grad 0.2182 +2026-04-10 06:37:28 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:37:28 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:37:31 - INFO - parrotllm.training - Train: loss=3.6126, ppl=37.06 +2026-04-10 06:37:31 - INFO - parrotllm.training - Val: loss=3.4241, ppl=30.70 +2026-04-10 06:37:31 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:37:32 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4241_epoch_0000_step_0021500.pt +2026-04-10 06:37:33 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:37:36 - INFO - parrotllm.training - step 21510 | epoch 0 | loss 3.4191 | lr 1.41e-03 | grad 0.1938 +2026-04-10 06:37:40 - INFO - parrotllm.training - step 21520 | epoch 0 | loss 3.3988 | lr 1.41e-03 | grad 0.2013 +2026-04-10 06:37:43 - INFO - parrotllm.training - step 21530 | epoch 0 | loss 3.4331 | lr 1.41e-03 | grad 0.1736 +2026-04-10 06:37:46 - INFO - parrotllm.training - step 21540 | epoch 0 | loss 3.5526 | lr 1.41e-03 | grad 0.2091 +2026-04-10 06:37:49 - INFO - parrotllm.training - step 21550 | epoch 0 | loss 3.5621 | lr 1.41e-03 | grad 0.1891 +2026-04-10 06:37:52 - INFO - parrotllm.training - step 21560 | epoch 0 | loss 3.3910 | lr 1.41e-03 | grad 0.1974 +2026-04-10 06:37:55 - INFO - parrotllm.training - step 21570 | epoch 0 | loss 3.5434 | lr 1.41e-03 | grad 0.2373 +2026-04-10 06:37:58 - INFO - parrotllm.training - step 21580 | epoch 0 | loss 3.4347 | lr 1.41e-03 | grad 0.1866 +2026-04-10 06:38:01 - INFO - parrotllm.training - step 21590 | epoch 0 | loss 3.4402 | lr 1.41e-03 | grad 0.2056 +2026-04-10 06:38:04 - INFO - parrotllm.training - step 21600 | epoch 0 | loss 3.4128 | lr 1.41e-03 | grad 0.2240 +2026-04-10 06:38:07 - INFO - parrotllm.training - step 21610 | epoch 0 | loss 3.5196 | lr 1.41e-03 | grad 0.2220 +2026-04-10 06:38:10 - INFO - parrotllm.training - step 21620 | epoch 0 | loss 3.5225 | lr 1.41e-03 | grad 0.2120 +2026-04-10 06:38:13 - INFO - parrotllm.training - step 21630 | epoch 0 | loss 3.4072 | lr 1.41e-03 | grad 0.2166 +2026-04-10 06:38:16 - INFO - parrotllm.training - step 21640 | epoch 0 | loss 3.3673 | lr 1.41e-03 | grad 0.2390 +2026-04-10 06:38:19 - INFO - parrotllm.training - step 21650 | epoch 0 | loss 3.5507 | lr 1.41e-03 | grad 0.2111 +2026-04-10 06:38:22 - INFO - parrotllm.training - step 21660 | epoch 0 | loss 3.4384 | lr 1.41e-03 | grad 0.2530 +2026-04-10 06:38:25 - INFO - parrotllm.training - step 21670 | epoch 0 | loss 3.4805 | lr 1.41e-03 | grad 0.2364 +2026-04-10 06:38:28 - INFO - parrotllm.training - step 21680 | epoch 0 | loss 3.4892 | lr 1.41e-03 | grad 0.2345 +2026-04-10 06:38:31 - INFO - parrotllm.training - step 21690 | epoch 0 | loss 3.5337 | lr 1.41e-03 | grad 0.2210 +2026-04-10 06:38:35 - INFO - parrotllm.training - step 21700 | epoch 0 | loss 3.4689 | lr 1.41e-03 | grad 0.2090 +2026-04-10 06:38:38 - INFO - parrotllm.training - step 21710 | epoch 0 | loss 3.5811 | lr 1.41e-03 | grad 0.2085 +2026-04-10 06:38:41 - INFO - parrotllm.training - step 21720 | epoch 0 | loss 3.5094 | lr 1.41e-03 | grad 0.2060 +2026-04-10 06:38:44 - INFO - parrotllm.training - step 21730 | epoch 0 | loss 3.4664 | lr 1.41e-03 | grad 0.2183 +2026-04-10 06:38:44 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 21738/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 06:38:47 - INFO - parrotllm.training - step 21740 | epoch 0 | loss 3.4495 | lr 1.41e-03 | grad 0.1949 +2026-04-10 06:38:50 - INFO - parrotllm.training - step 21750 | epoch 0 | loss 3.4122 | lr 1.41e-03 | grad 0.2045 +2026-04-10 06:38:53 - INFO - parrotllm.training - step 21760 | epoch 0 | loss 3.5184 | lr 1.41e-03 | grad 0.1952 +2026-04-10 06:38:56 - INFO - parrotllm.training - step 21770 | epoch 0 | loss 3.4092 | lr 1.41e-03 | grad 0.2006 +2026-04-10 06:38:59 - INFO - parrotllm.training - step 21780 | epoch 0 | loss 3.5511 | lr 1.41e-03 | grad 0.2048 +2026-04-10 06:39:02 - INFO - parrotllm.training - step 21790 | epoch 0 | loss 3.4587 | lr 1.41e-03 | grad 0.2081 +2026-04-10 06:39:05 - INFO - parrotllm.training - step 21800 | epoch 0 | loss 3.4810 | lr 1.41e-03 | grad 0.2283 +2026-04-10 06:39:08 - INFO - parrotllm.training - step 21810 | epoch 0 | loss 3.4383 | lr 1.41e-03 | grad 0.2042 +2026-04-10 06:39:11 - INFO - parrotllm.training - step 21820 | epoch 0 | loss 3.5328 | lr 1.41e-03 | grad 0.1920 +2026-04-10 06:39:15 - INFO - parrotllm.training - step 21830 | epoch 0 | loss 3.5574 | lr 1.41e-03 | grad 0.2093 +2026-04-10 06:39:18 - INFO - parrotllm.training - step 21840 | epoch 0 | loss 3.5544 | lr 1.41e-03 | grad 0.2075 +2026-04-10 06:39:21 - INFO - parrotllm.training - step 21850 | epoch 0 | loss 3.4851 | lr 1.41e-03 | grad 0.1886 +2026-04-10 06:39:24 - INFO - parrotllm.training - step 21860 | epoch 0 | loss 3.5165 | lr 1.41e-03 | grad 0.1989 +2026-04-10 06:39:27 - INFO - parrotllm.training - step 21870 | epoch 0 | loss 3.5284 | lr 1.41e-03 | grad 0.1927 +2026-04-10 06:39:30 - INFO - parrotllm.training - step 21880 | epoch 0 | loss 3.4149 | lr 1.41e-03 | grad 0.2072 +2026-04-10 06:39:33 - INFO - parrotllm.training - step 21890 | epoch 0 | loss 3.5309 | lr 1.41e-03 | grad 0.1919 +2026-04-10 06:39:36 - INFO - parrotllm.training - step 21900 | epoch 0 | loss 3.5662 | lr 1.41e-03 | grad 0.2220 +2026-04-10 06:39:39 - INFO - parrotllm.training - step 21910 | epoch 0 | loss 3.4380 | lr 1.41e-03 | grad 0.2073 +2026-04-10 06:39:42 - INFO - parrotllm.training - step 21920 | epoch 0 | loss 3.5634 | lr 1.41e-03 | grad 0.2237 +2026-04-10 06:39:45 - INFO - parrotllm.training - step 21930 | epoch 0 | loss 3.4640 | lr 1.41e-03 | grad 0.1916 +2026-04-10 06:39:48 - INFO - parrotllm.training - step 21940 | epoch 0 | loss 3.4099 | lr 1.41e-03 | grad 0.2486 +2026-04-10 06:39:51 - INFO - parrotllm.training - step 21950 | epoch 0 | loss 3.4413 | lr 1.41e-03 | grad 0.1938 +2026-04-10 06:39:54 - INFO - parrotllm.training - step 21960 | epoch 0 | loss 3.4452 | lr 1.41e-03 | grad 0.1896 +2026-04-10 06:39:57 - INFO - parrotllm.training - step 21970 | epoch 0 | loss 3.4881 | lr 1.41e-03 | grad 0.1899 +2026-04-10 06:40:00 - INFO - parrotllm.training - step 21980 | epoch 0 | loss 3.5478 | lr 1.41e-03 | grad 0.1977 +2026-04-10 06:40:03 - INFO - parrotllm.training - step 21990 | epoch 0 | loss 3.5389 | lr 1.41e-03 | grad 0.1873 +2026-04-10 06:40:06 - INFO - parrotllm.training - step 22000 | epoch 0 | loss 3.4923 | lr 1.41e-03 | grad 0.1923 +2026-04-10 06:40:06 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:40:06 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:40:09 - INFO - parrotllm.training - Train: loss=3.4923, ppl=32.86 +2026-04-10 06:40:09 - INFO - parrotllm.training - Val: loss=3.4208, ppl=30.59 +2026-04-10 06:40:09 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:40:10 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4208_epoch_0000_step_0022000.pt +2026-04-10 06:40:12 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:40:15 - INFO - parrotllm.training - step 22010 | epoch 0 | loss 3.4623 | lr 1.41e-03 | grad 0.2491 +2026-04-10 06:40:18 - INFO - parrotllm.training - step 22020 | epoch 0 | loss 3.5931 | lr 1.41e-03 | grad 0.2351 +2026-04-10 06:40:21 - INFO - parrotllm.training - step 22030 | epoch 0 | loss 3.4765 | lr 1.41e-03 | grad 0.2000 +2026-04-10 06:40:24 - INFO - parrotllm.training - step 22040 | epoch 0 | loss 3.5116 | lr 1.41e-03 | grad 0.2245 +2026-04-10 06:40:27 - INFO - parrotllm.training - step 22050 | epoch 0 | loss 3.5891 | lr 1.41e-03 | grad 0.2050 +2026-04-10 06:40:30 - INFO - parrotllm.training - step 22060 | epoch 0 | loss 3.4794 | lr 1.41e-03 | grad 0.2127 +2026-04-10 06:40:33 - INFO - parrotllm.training - step 22070 | epoch 0 | loss 3.5102 | lr 1.41e-03 | grad 0.2393 +2026-04-10 06:40:36 - INFO - parrotllm.training - step 22080 | epoch 0 | loss 3.6037 | lr 1.41e-03 | grad 0.1834 +2026-04-10 06:40:39 - INFO - parrotllm.training - step 22090 | epoch 0 | loss 3.5036 | lr 1.41e-03 | grad 0.1904 +2026-04-10 06:40:42 - INFO - parrotllm.training - step 22100 | epoch 0 | loss 3.4706 | lr 1.41e-03 | grad 0.1925 +2026-04-10 06:40:45 - INFO - parrotllm.training - step 22110 | epoch 0 | loss 3.6029 | lr 1.40e-03 | grad 0.1870 +2026-04-10 06:40:48 - INFO - parrotllm.training - step 22120 | epoch 0 | loss 3.5258 | lr 1.40e-03 | grad 0.2324 +2026-04-10 06:40:51 - INFO - parrotllm.training - step 22130 | epoch 0 | loss 3.5547 | lr 1.40e-03 | grad 0.1960 +2026-04-10 06:40:54 - INFO - parrotllm.training - step 22140 | epoch 0 | loss 3.3911 | lr 1.40e-03 | grad 0.2011 +2026-04-10 06:40:57 - INFO - parrotllm.training - step 22150 | epoch 0 | loss 3.5518 | lr 1.40e-03 | grad 0.2099 +2026-04-10 06:41:00 - INFO - parrotllm.training - step 22160 | epoch 0 | loss 3.3938 | lr 1.40e-03 | grad 0.2212 +2026-04-10 06:41:03 - INFO - parrotllm.training - step 22170 | epoch 0 | loss 3.3941 | lr 1.40e-03 | grad 0.2050 +2026-04-10 06:41:06 - INFO - parrotllm.training - step 22180 | epoch 0 | loss 3.5116 | lr 1.40e-03 | grad 0.2114 +2026-04-10 06:41:09 - INFO - parrotllm.training - step 22190 | epoch 0 | loss 3.3589 | lr 1.40e-03 | grad 0.2382 +2026-04-10 06:41:12 - INFO - parrotllm.training - step 22200 | epoch 0 | loss 3.4648 | lr 1.40e-03 | grad 0.2053 +2026-04-10 06:41:16 - INFO - parrotllm.training - step 22210 | epoch 0 | loss 3.4610 | lr 1.40e-03 | grad 0.1945 +2026-04-10 06:41:19 - INFO - parrotllm.training - step 22220 | epoch 0 | loss 3.4961 | lr 1.40e-03 | grad 0.2191 +2026-04-10 06:41:22 - INFO - parrotllm.training - step 22230 | epoch 0 | loss 3.5524 | lr 1.40e-03 | grad 0.1994 +2026-04-10 06:41:25 - INFO - parrotllm.training - step 22240 | epoch 0 | loss 3.3749 | lr 1.40e-03 | grad 0.2173 +2026-04-10 06:41:28 - INFO - parrotllm.training - step 22250 | epoch 0 | loss 3.4037 | lr 1.40e-03 | grad 0.2225 +2026-04-10 06:41:31 - INFO - parrotllm.training - step 22260 | epoch 0 | loss 3.4915 | lr 1.40e-03 | grad 0.2149 +2026-04-10 06:41:34 - INFO - parrotllm.training - step 22270 | epoch 0 | loss 3.5031 | lr 1.40e-03 | grad 0.2004 +2026-04-10 06:41:37 - INFO - parrotllm.training - step 22280 | epoch 0 | loss 3.3752 | lr 1.40e-03 | grad 0.1903 +2026-04-10 06:41:40 - INFO - parrotllm.training - step 22290 | epoch 0 | loss 3.6020 | lr 1.40e-03 | grad 0.2305 +2026-04-10 06:41:43 - INFO - parrotllm.training - step 22300 | epoch 0 | loss 3.4914 | lr 1.40e-03 | grad 0.2193 +2026-04-10 06:41:46 - INFO - parrotllm.training - step 22310 | epoch 0 | loss 3.4769 | lr 1.40e-03 | grad 0.1941 +2026-04-10 06:41:49 - INFO - parrotllm.training - step 22320 | epoch 0 | loss 3.4447 | lr 1.40e-03 | grad 0.2020 +2026-04-10 06:41:52 - INFO - parrotllm.training - step 22330 | epoch 0 | loss 3.4893 | lr 1.40e-03 | grad 0.2278 +2026-04-10 06:41:55 - INFO - parrotllm.training - step 22340 | epoch 0 | loss 3.5618 | lr 1.40e-03 | grad 0.2271 +2026-04-10 06:41:58 - INFO - parrotllm.training - step 22350 | epoch 0 | loss 3.4646 | lr 1.40e-03 | grad 0.2104 +2026-04-10 06:42:01 - INFO - parrotllm.training - step 22360 | epoch 0 | loss 3.4542 | lr 1.40e-03 | grad 0.2126 +2026-04-10 06:42:04 - INFO - parrotllm.training - step 22370 | epoch 0 | loss 3.3999 | lr 1.40e-03 | grad 0.2062 +2026-04-10 06:42:07 - INFO - parrotllm.training - step 22380 | epoch 0 | loss 3.5108 | lr 1.40e-03 | grad 0.2378 +2026-04-10 06:42:10 - INFO - parrotllm.training - step 22390 | epoch 0 | loss 3.5267 | lr 1.40e-03 | grad 0.2000 +2026-04-10 06:42:13 - INFO - parrotllm.training - step 22400 | epoch 0 | loss 3.4321 | lr 1.40e-03 | grad 0.2016 +2026-04-10 06:42:17 - INFO - parrotllm.training - step 22410 | epoch 0 | loss 3.5392 | lr 1.40e-03 | grad 0.2049 +2026-04-10 06:42:20 - INFO - parrotllm.training - step 22420 | epoch 0 | loss 3.4289 | lr 1.40e-03 | grad 0.1834 +2026-04-10 06:42:23 - INFO - parrotllm.training - step 22430 | epoch 0 | loss 3.4098 | lr 1.40e-03 | grad 0.2316 +2026-04-10 06:42:26 - INFO - parrotllm.training - step 22440 | epoch 0 | loss 3.4375 | lr 1.40e-03 | grad 0.2344 +2026-04-10 06:42:29 - INFO - parrotllm.training - step 22450 | epoch 0 | loss 3.4327 | lr 1.40e-03 | grad 0.2118 +2026-04-10 06:42:32 - INFO - parrotllm.training - step 22460 | epoch 0 | loss 3.5616 | lr 1.40e-03 | grad 0.1918 +2026-04-10 06:42:35 - INFO - parrotllm.training - step 22470 | epoch 0 | loss 3.4974 | lr 1.40e-03 | grad 0.1942 +2026-04-10 06:42:38 - INFO - parrotllm.training - step 22480 | epoch 0 | loss 3.4977 | lr 1.40e-03 | grad 0.2405 +2026-04-10 06:42:41 - INFO - parrotllm.training - step 22490 | epoch 0 | loss 3.3603 | lr 1.40e-03 | grad 0.2358 +2026-04-10 06:42:44 - INFO - parrotllm.training - step 22500 | epoch 0 | loss 3.5201 | lr 1.40e-03 | grad 0.2205 +2026-04-10 06:42:44 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:42:44 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:42:47 - INFO - parrotllm.training - Train: loss=3.5201, ppl=33.79 +2026-04-10 06:42:47 - INFO - parrotllm.training - Val: loss=3.4195, ppl=30.56 +2026-04-10 06:42:47 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:42:48 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4195_epoch_0000_step_0022500.pt +2026-04-10 06:42:49 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:42:50 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0022500.pt +2026-04-10 06:42:55 - INFO - parrotllm.training - step 22510 | epoch 0 | loss 3.5614 | lr 1.40e-03 | grad 0.2367 +2026-04-10 06:42:58 - INFO - parrotllm.training - step 22520 | epoch 0 | loss 3.4590 | lr 1.40e-03 | grad 0.1938 +2026-04-10 06:43:01 - INFO - parrotllm.training - step 22530 | epoch 0 | loss 3.4769 | lr 1.40e-03 | grad 0.2545 +2026-04-10 06:43:04 - INFO - parrotllm.training - step 22540 | epoch 0 | loss 3.4405 | lr 1.40e-03 | grad 0.2059 +2026-04-10 06:43:07 - INFO - parrotllm.training - step 22550 | epoch 0 | loss 3.4370 | lr 1.40e-03 | grad 0.2031 +2026-04-10 06:43:10 - INFO - parrotllm.training - step 22560 | epoch 0 | loss 3.4913 | lr 1.40e-03 | grad 0.2106 +2026-04-10 06:43:13 - INFO - parrotllm.training - step 22570 | epoch 0 | loss 3.5887 | lr 1.40e-03 | grad 0.1974 +2026-04-10 06:43:16 - INFO - parrotllm.training - step 22580 | epoch 0 | loss 3.5127 | lr 1.40e-03 | grad 0.2239 +2026-04-10 06:43:19 - INFO - parrotllm.training - step 22590 | epoch 0 | loss 3.4595 | lr 1.40e-03 | grad 0.2242 +2026-04-10 06:43:23 - INFO - parrotllm.training - step 22600 | epoch 0 | loss 3.5537 | lr 1.40e-03 | grad 0.1793 +2026-04-10 06:43:26 - INFO - parrotllm.training - step 22610 | epoch 0 | loss 3.5376 | lr 1.40e-03 | grad 0.2012 +2026-04-10 06:43:29 - INFO - parrotllm.training - step 22620 | epoch 0 | loss 3.4474 | lr 1.40e-03 | grad 0.2051 +2026-04-10 06:43:32 - INFO - parrotllm.training - step 22630 | epoch 0 | loss 3.4023 | lr 1.40e-03 | grad 0.2001 +2026-04-10 06:43:35 - INFO - parrotllm.training - step 22640 | epoch 0 | loss 3.5206 | lr 1.40e-03 | grad 0.2080 +2026-04-10 06:43:38 - INFO - parrotllm.training - step 22650 | epoch 0 | loss 3.5414 | lr 1.40e-03 | grad 0.2050 +2026-04-10 06:43:41 - INFO - parrotllm.training - step 22660 | epoch 0 | loss 3.4310 | lr 1.40e-03 | grad 0.1941 +2026-04-10 06:43:44 - INFO - parrotllm.training - step 22670 | epoch 0 | loss 3.4683 | lr 1.40e-03 | grad 0.1967 +2026-04-10 06:43:47 - INFO - parrotllm.training - step 22680 | epoch 0 | loss 3.4469 | lr 1.40e-03 | grad 0.1988 +2026-04-10 06:43:50 - INFO - parrotllm.training - step 22690 | epoch 0 | loss 3.5653 | lr 1.40e-03 | grad 0.2227 +2026-04-10 06:43:53 - INFO - parrotllm.training - step 22700 | epoch 0 | loss 3.4362 | lr 1.40e-03 | grad 0.2389 +2026-04-10 06:43:56 - INFO - parrotllm.training - step 22710 | epoch 0 | loss 3.5476 | lr 1.40e-03 | grad 0.1987 +2026-04-10 06:43:59 - INFO - parrotllm.training - step 22720 | epoch 0 | loss 3.4525 | lr 1.40e-03 | grad 0.2036 +2026-04-10 06:44:02 - INFO - parrotllm.training - step 22730 | epoch 0 | loss 3.4909 | lr 1.40e-03 | grad 0.2181 +2026-04-10 06:44:05 - INFO - parrotllm.training - step 22740 | epoch 0 | loss 3.4212 | lr 1.40e-03 | grad 0.2317 +2026-04-10 06:44:08 - INFO - parrotllm.training - step 22750 | epoch 0 | loss 3.3916 | lr 1.40e-03 | grad 0.2095 +2026-04-10 06:44:11 - INFO - parrotllm.training - step 22760 | epoch 0 | loss 3.4930 | lr 1.40e-03 | grad 0.2291 +2026-04-10 06:44:15 - INFO - parrotllm.training - step 22770 | epoch 0 | loss 3.3871 | lr 1.40e-03 | grad 0.1910 +2026-04-10 06:44:18 - INFO - parrotllm.training - step 22780 | epoch 0 | loss 3.5388 | lr 1.40e-03 | grad 0.2095 +2026-04-10 06:44:21 - INFO - parrotllm.training - step 22790 | epoch 0 | loss 3.5144 | lr 1.40e-03 | grad 0.2306 +2026-04-10 06:44:24 - INFO - parrotllm.training - step 22800 | epoch 0 | loss 3.3498 | lr 1.40e-03 | grad 0.2265 +2026-04-10 06:44:27 - INFO - parrotllm.training - step 22810 | epoch 0 | loss 3.4717 | lr 1.40e-03 | grad 0.1937 +2026-04-10 06:44:30 - INFO - parrotllm.training - step 22820 | epoch 0 | loss 3.4939 | lr 1.40e-03 | grad 0.1904 +2026-04-10 06:44:33 - INFO - parrotllm.training - step 22830 | epoch 0 | loss 3.4614 | lr 1.40e-03 | grad 0.2065 +2026-04-10 06:44:36 - INFO - parrotllm.training - step 22840 | epoch 0 | loss 3.4499 | lr 1.40e-03 | grad 0.2134 +2026-04-10 06:44:39 - INFO - parrotllm.training - step 22850 | epoch 0 | loss 3.4606 | lr 1.40e-03 | grad 0.2194 +2026-04-10 06:44:42 - INFO - parrotllm.training - step 22860 | epoch 0 | loss 3.4765 | lr 1.40e-03 | grad 0.1968 +2026-04-10 06:44:45 - INFO - parrotllm.training - step 22870 | epoch 0 | loss 3.5643 | lr 1.40e-03 | grad 0.2116 +2026-04-10 06:44:48 - INFO - parrotllm.training - step 22880 | epoch 0 | loss 3.5557 | lr 1.40e-03 | grad 0.1873 +2026-04-10 06:44:51 - INFO - parrotllm.training - step 22890 | epoch 0 | loss 3.4966 | lr 1.40e-03 | grad 0.2193 +2026-04-10 06:44:54 - INFO - parrotllm.training - step 22900 | epoch 0 | loss 3.4957 | lr 1.40e-03 | grad 0.1983 +2026-04-10 06:44:57 - INFO - parrotllm.training - step 22910 | epoch 0 | loss 3.5644 | lr 1.40e-03 | grad 0.1936 +2026-04-10 06:45:00 - INFO - parrotllm.training - step 22920 | epoch 0 | loss 3.5146 | lr 1.40e-03 | grad 0.2168 +2026-04-10 06:45:03 - INFO - parrotllm.training - step 22930 | epoch 0 | loss 3.5583 | lr 1.40e-03 | grad 0.1845 +2026-04-10 06:45:06 - INFO - parrotllm.training - step 22940 | epoch 0 | loss 3.5352 | lr 1.40e-03 | grad 0.2096 +2026-04-10 06:45:10 - INFO - parrotllm.training - step 22950 | epoch 0 | loss 3.5473 | lr 1.40e-03 | grad 0.2089 +2026-04-10 06:45:13 - INFO - parrotllm.training - step 22960 | epoch 0 | loss 3.4467 | lr 1.40e-03 | grad 0.2012 +2026-04-10 06:45:16 - INFO - parrotllm.training - step 22970 | epoch 0 | loss 3.4390 | lr 1.40e-03 | grad 0.2083 +2026-04-10 06:45:19 - INFO - parrotllm.training - step 22980 | epoch 0 | loss 3.3687 | lr 1.40e-03 | grad 0.2030 +2026-04-10 06:45:22 - INFO - parrotllm.training - step 22990 | epoch 0 | loss 3.5887 | lr 1.40e-03 | grad 0.2201 +2026-04-10 06:45:25 - INFO - parrotllm.training - step 23000 | epoch 0 | loss 3.5303 | lr 1.40e-03 | grad 0.2444 +2026-04-10 06:45:25 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:45:25 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:45:28 - INFO - parrotllm.training - Train: loss=3.5303, ppl=34.14 +2026-04-10 06:45:28 - INFO - parrotllm.training - Val: loss=3.4195, ppl=30.55 +2026-04-10 06:45:29 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4195_epoch_0000_step_0023000.pt +2026-04-10 06:45:30 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4195, min_delta=0.001000). +2026-04-10 06:45:30 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:45:33 - INFO - parrotllm.training - step 23010 | epoch 0 | loss 3.5207 | lr 1.40e-03 | grad 0.2146 +2026-04-10 06:45:36 - INFO - parrotllm.training - step 23020 | epoch 0 | loss 3.5016 | lr 1.40e-03 | grad 0.2161 +2026-04-10 06:45:39 - INFO - parrotllm.training - step 23030 | epoch 0 | loss 3.4641 | lr 1.40e-03 | grad 0.2636 +2026-04-10 06:45:42 - INFO - parrotllm.training - step 23040 | epoch 0 | loss 3.5191 | lr 1.40e-03 | grad 0.2142 +2026-04-10 06:45:45 - INFO - parrotllm.training - step 23050 | epoch 0 | loss 3.4378 | lr 1.40e-03 | grad 0.2019 +2026-04-10 06:45:48 - INFO - parrotllm.training - step 23060 | epoch 0 | loss 3.5142 | lr 1.40e-03 | grad 0.2104 +2026-04-10 06:45:51 - INFO - parrotllm.training - step 23070 | epoch 0 | loss 3.4221 | lr 1.40e-03 | grad 0.2308 +2026-04-10 06:45:54 - INFO - parrotllm.training - step 23080 | epoch 0 | loss 3.5707 | lr 1.40e-03 | grad 0.1946 +2026-04-10 06:45:57 - INFO - parrotllm.training - step 23090 | epoch 0 | loss 3.3981 | lr 1.40e-03 | grad 0.2100 +2026-04-10 06:46:00 - INFO - parrotllm.training - step 23100 | epoch 0 | loss 3.5131 | lr 1.40e-03 | grad 0.2009 +2026-04-10 06:46:04 - INFO - parrotllm.training - step 23110 | epoch 0 | loss 3.5195 | lr 1.40e-03 | grad 0.2003 +2026-04-10 06:46:07 - INFO - parrotllm.training - step 23120 | epoch 0 | loss 3.4994 | lr 1.40e-03 | grad 0.2575 +2026-04-10 06:46:10 - INFO - parrotllm.training - step 23130 | epoch 0 | loss 3.4820 | lr 1.40e-03 | grad 0.2238 +2026-04-10 06:46:13 - INFO - parrotllm.training - step 23140 | epoch 0 | loss 3.4941 | lr 1.40e-03 | grad 0.2053 +2026-04-10 06:46:16 - INFO - parrotllm.training - step 23150 | epoch 0 | loss 3.3967 | lr 1.40e-03 | grad 0.1718 +2026-04-10 06:46:19 - INFO - parrotllm.training - step 23160 | epoch 0 | loss 3.3911 | lr 1.40e-03 | grad 0.2073 +2026-04-10 06:46:22 - INFO - parrotllm.training - step 23170 | epoch 0 | loss 3.4307 | lr 1.40e-03 | grad 0.1898 +2026-04-10 06:46:25 - INFO - parrotllm.training - step 23180 | epoch 0 | loss 3.4580 | lr 1.40e-03 | grad 0.2184 +2026-04-10 06:46:28 - INFO - parrotllm.training - step 23190 | epoch 0 | loss 3.4934 | lr 1.40e-03 | grad 0.2171 +2026-04-10 06:46:31 - INFO - parrotllm.training - step 23200 | epoch 0 | loss 3.4685 | lr 1.39e-03 | grad 0.1982 +2026-04-10 06:46:34 - INFO - parrotllm.training - step 23210 | epoch 0 | loss 3.4652 | lr 1.39e-03 | grad 0.2203 +2026-04-10 06:46:37 - INFO - parrotllm.training - step 23220 | epoch 0 | loss 3.5035 | lr 1.39e-03 | grad 0.1858 +2026-04-10 06:46:40 - INFO - parrotllm.training - step 23230 | epoch 0 | loss 3.4051 | lr 1.39e-03 | grad 0.2146 +2026-04-10 06:46:43 - INFO - parrotllm.training - step 23240 | epoch 0 | loss 3.5353 | lr 1.39e-03 | grad 0.2419 +2026-04-10 06:46:46 - INFO - parrotllm.training - step 23250 | epoch 0 | loss 3.4757 | lr 1.39e-03 | grad 0.2137 +2026-04-10 06:46:49 - INFO - parrotllm.training - step 23260 | epoch 0 | loss 3.4574 | lr 1.39e-03 | grad 0.2037 +2026-04-10 06:46:52 - INFO - parrotllm.training - step 23270 | epoch 0 | loss 3.4422 | lr 1.39e-03 | grad 0.1973 +2026-04-10 06:46:55 - INFO - parrotllm.training - step 23280 | epoch 0 | loss 3.4818 | lr 1.39e-03 | grad 0.1989 +2026-04-10 06:46:59 - INFO - parrotllm.training - step 23290 | epoch 0 | loss 3.4223 | lr 1.39e-03 | grad 0.2283 +2026-04-10 06:47:02 - INFO - parrotllm.training - step 23300 | epoch 0 | loss 3.3659 | lr 1.39e-03 | grad 0.2043 +2026-04-10 06:47:05 - INFO - parrotllm.training - step 23310 | epoch 0 | loss 3.5121 | lr 1.39e-03 | grad 0.1963 +2026-04-10 06:47:08 - INFO - parrotllm.training - step 23320 | epoch 0 | loss 3.4730 | lr 1.39e-03 | grad 0.2052 +2026-04-10 06:47:11 - INFO - parrotllm.training - step 23330 | epoch 0 | loss 3.4460 | lr 1.39e-03 | grad 0.2134 +2026-04-10 06:47:14 - INFO - parrotllm.training - step 23340 | epoch 0 | loss 3.5241 | lr 1.39e-03 | grad 0.1924 +2026-04-10 06:47:17 - INFO - parrotllm.training - step 23350 | epoch 0 | loss 3.5683 | lr 1.39e-03 | grad 0.2149 +2026-04-10 06:47:20 - INFO - parrotllm.training - step 23360 | epoch 0 | loss 3.5070 | lr 1.39e-03 | grad 0.2280 +2026-04-10 06:47:23 - INFO - parrotllm.training - step 23370 | epoch 0 | loss 3.4939 | lr 1.39e-03 | grad 0.2128 +2026-04-10 06:47:26 - INFO - parrotllm.training - step 23380 | epoch 0 | loss 3.5190 | lr 1.39e-03 | grad 0.2177 +2026-04-10 06:47:29 - INFO - parrotllm.training - step 23390 | epoch 0 | loss 3.4622 | lr 1.39e-03 | grad 0.2159 +2026-04-10 06:47:32 - INFO - parrotllm.training - step 23400 | epoch 0 | loss 3.4991 | lr 1.39e-03 | grad 0.1895 +2026-04-10 06:47:35 - INFO - parrotllm.training - step 23410 | epoch 0 | loss 3.5203 | lr 1.39e-03 | grad 0.1862 +2026-04-10 06:47:38 - INFO - parrotllm.training - step 23420 | epoch 0 | loss 3.5916 | lr 1.39e-03 | grad 0.2107 +2026-04-10 06:47:41 - INFO - parrotllm.training - step 23430 | epoch 0 | loss 3.5055 | lr 1.39e-03 | grad 0.1957 +2026-04-10 06:47:44 - INFO - parrotllm.training - step 23440 | epoch 0 | loss 3.3575 | lr 1.39e-03 | grad 0.1991 +2026-04-10 06:47:47 - INFO - parrotllm.training - step 23450 | epoch 0 | loss 3.5461 | lr 1.39e-03 | grad 0.2180 +2026-04-10 06:47:50 - INFO - parrotllm.training - step 23460 | epoch 0 | loss 3.4936 | lr 1.39e-03 | grad 0.1899 +2026-04-10 06:47:53 - INFO - parrotllm.training - step 23470 | epoch 0 | loss 3.5376 | lr 1.39e-03 | grad 0.1806 +2026-04-10 06:47:57 - INFO - parrotllm.training - step 23480 | epoch 0 | loss 3.5087 | lr 1.39e-03 | grad 0.2285 +2026-04-10 06:48:00 - INFO - parrotllm.training - step 23490 | epoch 0 | loss 3.3376 | lr 1.39e-03 | grad 0.1965 +2026-04-10 06:48:03 - INFO - parrotllm.training - step 23500 | epoch 0 | loss 3.5944 | lr 1.39e-03 | grad 0.2493 +2026-04-10 06:48:03 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:48:03 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:48:06 - INFO - parrotllm.training - Train: loss=3.5944, ppl=36.39 +2026-04-10 06:48:06 - INFO - parrotllm.training - Val: loss=3.4189, ppl=30.54 +2026-04-10 06:48:07 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4189_epoch_0000_step_0023500.pt +2026-04-10 06:48:08 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.4195, min_delta=0.001000). +2026-04-10 06:48:08 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:48:11 - INFO - parrotllm.training - step 23510 | epoch 0 | loss 3.5462 | lr 1.39e-03 | grad 0.1981 +2026-04-10 06:48:14 - INFO - parrotllm.training - step 23520 | epoch 0 | loss 3.4672 | lr 1.39e-03 | grad 0.1957 +2026-04-10 06:48:17 - INFO - parrotllm.training - step 23530 | epoch 0 | loss 3.4603 | lr 1.39e-03 | grad 0.1978 +2026-04-10 06:48:20 - INFO - parrotllm.training - step 23540 | epoch 0 | loss 3.3584 | lr 1.39e-03 | grad 0.1996 +2026-04-10 06:48:23 - INFO - parrotllm.training - step 23550 | epoch 0 | loss 3.4746 | lr 1.39e-03 | grad 0.2095 +2026-04-10 06:48:26 - INFO - parrotllm.training - step 23560 | epoch 0 | loss 3.4605 | lr 1.39e-03 | grad 0.1965 +2026-04-10 06:48:29 - INFO - parrotllm.training - step 23570 | epoch 0 | loss 3.4240 | lr 1.39e-03 | grad 0.1926 +2026-04-10 06:48:32 - INFO - parrotllm.training - step 23580 | epoch 0 | loss 3.4578 | lr 1.39e-03 | grad 0.2192 +2026-04-10 06:48:35 - INFO - parrotllm.training - step 23590 | epoch 0 | loss 3.4356 | lr 1.39e-03 | grad 0.2008 +2026-04-10 06:48:38 - INFO - parrotllm.training - step 23600 | epoch 0 | loss 3.4513 | lr 1.39e-03 | grad 0.2098 +2026-04-10 06:48:41 - INFO - parrotllm.training - step 23610 | epoch 0 | loss 3.4962 | lr 1.39e-03 | grad 0.2184 +2026-04-10 06:48:44 - INFO - parrotllm.training - step 23620 | epoch 0 | loss 3.4384 | lr 1.39e-03 | grad 0.2289 +2026-04-10 06:48:47 - INFO - parrotllm.training - step 23630 | epoch 0 | loss 3.4565 | lr 1.39e-03 | grad 0.2234 +2026-04-10 06:48:50 - INFO - parrotllm.training - step 23640 | epoch 0 | loss 3.4965 | lr 1.39e-03 | grad 0.2255 +2026-04-10 06:48:54 - INFO - parrotllm.training - step 23650 | epoch 0 | loss 3.4808 | lr 1.39e-03 | grad 0.2299 +2026-04-10 06:48:57 - INFO - parrotllm.training - step 23660 | epoch 0 | loss 3.4866 | lr 1.39e-03 | grad 0.2029 +2026-04-10 06:49:00 - INFO - parrotllm.training - step 23670 | epoch 0 | loss 3.4488 | lr 1.39e-03 | grad 0.1849 +2026-04-10 06:49:03 - INFO - parrotllm.training - step 23680 | epoch 0 | loss 3.5274 | lr 1.39e-03 | grad 0.2318 +2026-04-10 06:49:06 - INFO - parrotllm.training - step 23690 | epoch 0 | loss 3.4650 | lr 1.39e-03 | grad 0.2049 +2026-04-10 06:49:09 - INFO - parrotllm.training - step 23700 | epoch 0 | loss 3.4441 | lr 1.39e-03 | grad 0.2326 +2026-04-10 06:49:12 - INFO - parrotllm.training - step 23710 | epoch 0 | loss 3.5623 | lr 1.39e-03 | grad 0.2150 +2026-04-10 06:49:15 - INFO - parrotllm.training - step 23720 | epoch 0 | loss 3.5031 | lr 1.39e-03 | grad 0.2012 +2026-04-10 06:49:18 - INFO - parrotllm.training - step 23730 | epoch 0 | loss 3.5002 | lr 1.39e-03 | grad 0.2037 +2026-04-10 06:49:21 - INFO - parrotllm.training - step 23740 | epoch 0 | loss 3.3814 | lr 1.39e-03 | grad 0.2167 +2026-04-10 06:49:24 - INFO - parrotllm.training - step 23750 | epoch 0 | loss 3.4718 | lr 1.39e-03 | grad 0.2035 +2026-04-10 06:49:27 - INFO - parrotllm.training - step 23760 | epoch 0 | loss 3.3837 | lr 1.39e-03 | grad 0.2391 +2026-04-10 06:49:30 - INFO - parrotllm.training - step 23770 | epoch 0 | loss 3.4228 | lr 1.39e-03 | grad 0.2073 +2026-04-10 06:49:33 - INFO - parrotllm.training - step 23780 | epoch 0 | loss 3.4281 | lr 1.39e-03 | grad 0.2069 +2026-04-10 06:49:33 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 23789/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 06:49:37 - INFO - parrotllm.training - step 23790 | epoch 0 | loss 3.4272 | lr 1.39e-03 | grad 0.2143 +2026-04-10 06:49:40 - INFO - parrotllm.training - step 23800 | epoch 0 | loss 3.4608 | lr 1.39e-03 | grad 0.2075 +2026-04-10 06:49:43 - INFO - parrotllm.training - step 23810 | epoch 0 | loss 3.4859 | lr 1.39e-03 | grad 0.1977 +2026-04-10 06:49:46 - INFO - parrotllm.training - step 23820 | epoch 0 | loss 3.4639 | lr 1.39e-03 | grad 0.2070 +2026-04-10 06:49:49 - INFO - parrotllm.training - step 23830 | epoch 0 | loss 3.5303 | lr 1.39e-03 | grad 0.2158 +2026-04-10 06:49:52 - INFO - parrotllm.training - step 23840 | epoch 0 | loss 3.3409 | lr 1.39e-03 | grad 0.1998 +2026-04-10 06:49:55 - INFO - parrotllm.training - step 23850 | epoch 0 | loss 3.5386 | lr 1.39e-03 | grad 0.1904 +2026-04-10 06:49:58 - INFO - parrotllm.training - step 23860 | epoch 0 | loss 3.4866 | lr 1.39e-03 | grad 0.2051 +2026-04-10 06:50:01 - INFO - parrotllm.training - step 23870 | epoch 0 | loss 3.4476 | lr 1.39e-03 | grad 0.2262 +2026-04-10 06:50:04 - INFO - parrotllm.training - step 23880 | epoch 0 | loss 3.3762 | lr 1.39e-03 | grad 0.2292 +2026-04-10 06:50:07 - INFO - parrotllm.training - step 23890 | epoch 0 | loss 3.4267 | lr 1.39e-03 | grad 0.2125 +2026-04-10 06:50:10 - INFO - parrotllm.training - step 23900 | epoch 0 | loss 3.5367 | lr 1.39e-03 | grad 0.2009 +2026-04-10 06:50:13 - INFO - parrotllm.training - step 23910 | epoch 0 | loss 3.5327 | lr 1.39e-03 | grad 0.1856 +2026-04-10 06:50:16 - INFO - parrotllm.training - step 23920 | epoch 0 | loss 3.4949 | lr 1.39e-03 | grad 0.1957 +2026-04-10 06:50:19 - INFO - parrotllm.training - step 23930 | epoch 0 | loss 3.4819 | lr 1.39e-03 | grad 0.2072 +2026-04-10 06:50:22 - INFO - parrotllm.training - step 23940 | epoch 0 | loss 3.5009 | lr 1.39e-03 | grad 0.2299 +2026-04-10 06:50:25 - INFO - parrotllm.training - step 23950 | epoch 0 | loss 3.4944 | lr 1.39e-03 | grad 0.2025 +2026-04-10 06:50:28 - INFO - parrotllm.training - step 23960 | epoch 0 | loss 3.5484 | lr 1.39e-03 | grad 0.1949 +2026-04-10 06:50:31 - INFO - parrotllm.training - step 23970 | epoch 0 | loss 3.4165 | lr 1.39e-03 | grad 0.2022 +2026-04-10 06:50:34 - INFO - parrotllm.training - step 23980 | epoch 0 | loss 3.3623 | lr 1.39e-03 | grad 0.2125 +2026-04-10 06:50:37 - INFO - parrotllm.training - step 23990 | epoch 0 | loss 3.4338 | lr 1.39e-03 | grad 0.2001 +2026-04-10 06:50:40 - INFO - parrotllm.training - step 24000 | epoch 0 | loss 3.4333 | lr 1.39e-03 | grad 0.2067 +2026-04-10 06:50:40 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:50:40 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:50:44 - INFO - parrotllm.training - Train: loss=3.4333, ppl=30.98 +2026-04-10 06:50:44 - INFO - parrotllm.training - Val: loss=3.4182, ppl=30.52 +2026-04-10 06:50:44 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:50:45 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4182_epoch_0000_step_0024000.pt +2026-04-10 06:50:46 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:50:49 - INFO - parrotllm.training - step 24010 | epoch 0 | loss 3.4438 | lr 1.39e-03 | grad 0.2042 +2026-04-10 06:50:52 - INFO - parrotllm.training - step 24020 | epoch 0 | loss 3.4264 | lr 1.39e-03 | grad 0.2173 +2026-04-10 06:50:55 - INFO - parrotllm.training - step 24030 | epoch 0 | loss 3.4921 | lr 1.39e-03 | grad 0.1976 +2026-04-10 06:50:58 - INFO - parrotllm.training - step 24040 | epoch 0 | loss 3.4287 | lr 1.39e-03 | grad 0.1915 +2026-04-10 06:51:01 - INFO - parrotllm.training - step 24050 | epoch 0 | loss 3.5131 | lr 1.39e-03 | grad 0.1950 +2026-04-10 06:51:04 - INFO - parrotllm.training - step 24060 | epoch 0 | loss 3.4196 | lr 1.39e-03 | grad 0.1981 +2026-04-10 06:51:07 - INFO - parrotllm.training - step 24070 | epoch 0 | loss 3.5031 | lr 1.39e-03 | grad 0.2344 +2026-04-10 06:51:10 - INFO - parrotllm.training - step 24080 | epoch 0 | loss 3.5678 | lr 1.39e-03 | grad 0.2034 +2026-04-10 06:51:13 - INFO - parrotllm.training - step 24090 | epoch 0 | loss 3.3928 | lr 1.39e-03 | grad 0.2429 +2026-04-10 06:51:16 - INFO - parrotllm.training - step 24100 | epoch 0 | loss 3.4453 | lr 1.39e-03 | grad 0.2028 +2026-04-10 06:51:19 - INFO - parrotllm.training - step 24110 | epoch 0 | loss 3.4138 | lr 1.39e-03 | grad 0.2002 +2026-04-10 06:51:22 - INFO - parrotllm.training - step 24120 | epoch 0 | loss 3.4846 | lr 1.39e-03 | grad 0.1778 +2026-04-10 06:51:25 - INFO - parrotllm.training - step 24130 | epoch 0 | loss 3.4496 | lr 1.39e-03 | grad 0.1944 +2026-04-10 06:51:28 - INFO - parrotllm.training - step 24140 | epoch 0 | loss 3.5131 | lr 1.39e-03 | grad 0.2079 +2026-04-10 06:51:31 - INFO - parrotllm.training - step 24150 | epoch 0 | loss 3.3556 | lr 1.39e-03 | grad 0.2046 +2026-04-10 06:51:34 - INFO - parrotllm.training - step 24160 | epoch 0 | loss 3.5499 | lr 1.39e-03 | grad 0.2080 +2026-04-10 06:51:37 - INFO - parrotllm.training - step 24170 | epoch 0 | loss 3.5278 | lr 1.39e-03 | grad 0.2147 +2026-04-10 06:51:41 - INFO - parrotllm.training - step 24180 | epoch 0 | loss 3.5328 | lr 1.39e-03 | grad 0.1968 +2026-04-10 06:51:44 - INFO - parrotllm.training - step 24190 | epoch 0 | loss 3.5094 | lr 1.39e-03 | grad 0.2162 +2026-04-10 06:51:47 - INFO - parrotllm.training - step 24200 | epoch 0 | loss 3.4373 | lr 1.39e-03 | grad 0.1959 +2026-04-10 06:51:50 - INFO - parrotllm.training - step 24210 | epoch 0 | loss 3.4541 | lr 1.39e-03 | grad 0.2110 +2026-04-10 06:51:53 - INFO - parrotllm.training - step 24220 | epoch 0 | loss 3.4660 | lr 1.39e-03 | grad 0.2156 +2026-04-10 06:51:56 - INFO - parrotllm.training - step 24230 | epoch 0 | loss 3.4048 | lr 1.39e-03 | grad 0.1854 +2026-04-10 06:51:59 - INFO - parrotllm.training - step 24240 | epoch 0 | loss 3.5356 | lr 1.38e-03 | grad 0.2247 +2026-04-10 06:52:02 - INFO - parrotllm.training - step 24250 | epoch 0 | loss 3.4354 | lr 1.38e-03 | grad 0.1959 +2026-04-10 06:52:05 - INFO - parrotllm.training - step 24260 | epoch 0 | loss 3.4626 | lr 1.38e-03 | grad 0.1983 +2026-04-10 06:52:08 - INFO - parrotllm.training - step 24270 | epoch 0 | loss 3.5182 | lr 1.38e-03 | grad 0.2392 +2026-04-10 06:52:11 - INFO - parrotllm.training - step 24280 | epoch 0 | loss 3.5257 | lr 1.38e-03 | grad 0.2404 +2026-04-10 06:52:14 - INFO - parrotllm.training - step 24290 | epoch 0 | loss 3.5407 | lr 1.38e-03 | grad 0.2075 +2026-04-10 06:52:17 - INFO - parrotllm.training - step 24300 | epoch 0 | loss 3.5244 | lr 1.38e-03 | grad 0.2764 +2026-04-10 06:52:20 - INFO - parrotllm.training - step 24310 | epoch 0 | loss 3.5189 | lr 1.38e-03 | grad 0.2268 +2026-04-10 06:52:23 - INFO - parrotllm.training - step 24320 | epoch 0 | loss 3.6196 | lr 1.38e-03 | grad 0.2278 +2026-04-10 06:52:26 - INFO - parrotllm.training - step 24330 | epoch 0 | loss 3.4420 | lr 1.38e-03 | grad 0.1900 +2026-04-10 06:52:29 - INFO - parrotllm.training - step 24340 | epoch 0 | loss 3.5387 | lr 1.38e-03 | grad 0.2075 +2026-04-10 06:52:32 - INFO - parrotllm.training - step 24350 | epoch 0 | loss 3.4244 | lr 1.38e-03 | grad 0.1982 +2026-04-10 06:52:35 - INFO - parrotllm.training - step 24360 | epoch 0 | loss 3.4412 | lr 1.38e-03 | grad 0.2432 +2026-04-10 06:52:38 - INFO - parrotllm.training - step 24370 | epoch 0 | loss 3.4860 | lr 1.38e-03 | grad 0.2228 +2026-04-10 06:52:42 - INFO - parrotllm.training - step 24380 | epoch 0 | loss 3.4799 | lr 1.38e-03 | grad 0.2374 +2026-04-10 06:52:45 - INFO - parrotllm.training - step 24390 | epoch 0 | loss 3.4477 | lr 1.38e-03 | grad 0.2099 +2026-04-10 06:52:48 - INFO - parrotllm.training - step 24400 | epoch 0 | loss 3.3696 | lr 1.38e-03 | grad 0.2165 +2026-04-10 06:52:51 - INFO - parrotllm.training - step 24410 | epoch 0 | loss 3.4116 | lr 1.38e-03 | grad 0.2142 +2026-04-10 06:52:54 - INFO - parrotllm.training - step 24420 | epoch 0 | loss 3.5259 | lr 1.38e-03 | grad 0.1914 +2026-04-10 06:52:57 - INFO - parrotllm.training - step 24430 | epoch 0 | loss 3.4264 | lr 1.38e-03 | grad 0.2112 +2026-04-10 06:53:00 - INFO - parrotllm.training - step 24440 | epoch 0 | loss 3.4467 | lr 1.38e-03 | grad 0.1976 +2026-04-10 06:53:03 - INFO - parrotllm.training - step 24450 | epoch 0 | loss 3.4548 | lr 1.38e-03 | grad 0.2070 +2026-04-10 06:53:06 - INFO - parrotllm.training - step 24460 | epoch 0 | loss 3.5107 | lr 1.38e-03 | grad 0.2169 +2026-04-10 06:53:09 - INFO - parrotllm.training - step 24470 | epoch 0 | loss 3.5209 | lr 1.38e-03 | grad 0.1998 +2026-04-10 06:53:12 - INFO - parrotllm.training - step 24480 | epoch 0 | loss 3.4332 | lr 1.38e-03 | grad 0.1926 +2026-04-10 06:53:15 - INFO - parrotllm.training - step 24490 | epoch 0 | loss 3.5172 | lr 1.38e-03 | grad 0.1957 +2026-04-10 06:53:18 - INFO - parrotllm.training - step 24500 | epoch 0 | loss 3.4247 | lr 1.38e-03 | grad 0.1956 +2026-04-10 06:53:18 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:53:18 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:53:21 - INFO - parrotllm.training - Train: loss=3.4247, ppl=30.71 +2026-04-10 06:53:21 - INFO - parrotllm.training - Val: loss=3.4169, ppl=30.47 +2026-04-10 06:53:21 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:53:22 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4169_epoch_0000_step_0024500.pt +2026-04-10 06:53:23 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:53:26 - INFO - parrotllm.training - step 24510 | epoch 0 | loss 3.4357 | lr 1.38e-03 | grad 0.2142 +2026-04-10 06:53:30 - INFO - parrotllm.training - step 24520 | epoch 0 | loss 3.4627 | lr 1.38e-03 | grad 0.2292 +2026-04-10 06:53:33 - INFO - parrotllm.training - step 24530 | epoch 0 | loss 3.5071 | lr 1.38e-03 | grad 0.1885 +2026-04-10 06:53:36 - INFO - parrotllm.training - step 24540 | epoch 0 | loss 3.4464 | lr 1.38e-03 | grad 0.1936 +2026-04-10 06:53:39 - INFO - parrotllm.training - step 24550 | epoch 0 | loss 3.3814 | lr 1.38e-03 | grad 0.2113 +2026-04-10 06:53:42 - INFO - parrotllm.training - step 24560 | epoch 0 | loss 3.5690 | lr 1.38e-03 | grad 0.2048 +2026-04-10 06:53:45 - INFO - parrotllm.training - step 24570 | epoch 0 | loss 3.4792 | lr 1.38e-03 | grad 0.2105 +2026-04-10 06:53:48 - INFO - parrotllm.training - step 24580 | epoch 0 | loss 3.4830 | lr 1.38e-03 | grad 0.2239 +2026-04-10 06:53:51 - INFO - parrotllm.training - step 24590 | epoch 0 | loss 3.3815 | lr 1.38e-03 | grad 0.1831 +2026-04-10 06:53:54 - INFO - parrotllm.training - step 24600 | epoch 0 | loss 3.4613 | lr 1.38e-03 | grad 0.1790 +2026-04-10 06:53:57 - INFO - parrotllm.training - step 24610 | epoch 0 | loss 3.5075 | lr 1.38e-03 | grad 0.2184 +2026-04-10 06:54:00 - INFO - parrotllm.training - step 24620 | epoch 0 | loss 3.4468 | lr 1.38e-03 | grad 0.2245 +2026-04-10 06:54:03 - INFO - parrotllm.training - step 24630 | epoch 0 | loss 3.4259 | lr 1.38e-03 | grad 0.2469 +2026-04-10 06:54:06 - INFO - parrotllm.training - step 24640 | epoch 0 | loss 3.4308 | lr 1.38e-03 | grad 0.2018 +2026-04-10 06:54:09 - INFO - parrotllm.training - step 24650 | epoch 0 | loss 3.4625 | lr 1.38e-03 | grad 0.2240 +2026-04-10 06:54:12 - INFO - parrotllm.training - step 24660 | epoch 0 | loss 3.4789 | lr 1.38e-03 | grad 0.1947 +2026-04-10 06:54:15 - INFO - parrotllm.training - step 24670 | epoch 0 | loss 3.4650 | lr 1.38e-03 | grad 0.2506 +2026-04-10 06:54:18 - INFO - parrotllm.training - step 24680 | epoch 0 | loss 3.5085 | lr 1.38e-03 | grad 0.2131 +2026-04-10 06:54:21 - INFO - parrotllm.training - step 24690 | epoch 0 | loss 3.4111 | lr 1.38e-03 | grad 0.1839 +2026-04-10 06:54:24 - INFO - parrotllm.training - step 24700 | epoch 0 | loss 3.5628 | lr 1.38e-03 | grad 0.2165 +2026-04-10 06:54:27 - INFO - parrotllm.training - step 24710 | epoch 0 | loss 3.5054 | lr 1.38e-03 | grad 0.1964 +2026-04-10 06:54:30 - INFO - parrotllm.training - step 24720 | epoch 0 | loss 3.5163 | lr 1.38e-03 | grad 0.2093 +2026-04-10 06:54:34 - INFO - parrotllm.training - step 24730 | epoch 0 | loss 3.4607 | lr 1.38e-03 | grad 0.1998 +2026-04-10 06:54:37 - INFO - parrotllm.training - step 24740 | epoch 0 | loss 3.3832 | lr 1.38e-03 | grad 0.2376 +2026-04-10 06:54:40 - INFO - parrotllm.training - step 24750 | epoch 0 | loss 3.4177 | lr 1.38e-03 | grad 0.1883 +2026-04-10 06:54:43 - INFO - parrotllm.training - step 24760 | epoch 0 | loss 3.5207 | lr 1.38e-03 | grad 0.2122 +2026-04-10 06:54:46 - INFO - parrotllm.training - step 24770 | epoch 0 | loss 3.4082 | lr 1.38e-03 | grad 0.2131 +2026-04-10 06:54:49 - INFO - parrotllm.training - step 24780 | epoch 0 | loss 3.4598 | lr 1.38e-03 | grad 0.2267 +2026-04-10 06:54:52 - INFO - parrotllm.training - step 24790 | epoch 0 | loss 3.5240 | lr 1.38e-03 | grad 0.2085 +2026-04-10 06:54:55 - INFO - parrotllm.training - step 24800 | epoch 0 | loss 3.4000 | lr 1.38e-03 | grad 0.1830 +2026-04-10 06:54:58 - INFO - parrotllm.training - step 24810 | epoch 0 | loss 3.3945 | lr 1.38e-03 | grad 0.2314 +2026-04-10 06:55:01 - INFO - parrotllm.training - step 24820 | epoch 0 | loss 3.4397 | lr 1.38e-03 | grad 0.2065 +2026-04-10 06:55:04 - INFO - parrotllm.training - step 24830 | epoch 0 | loss 3.4405 | lr 1.38e-03 | grad 0.2229 +2026-04-10 06:55:07 - INFO - parrotllm.training - step 24840 | epoch 0 | loss 3.4770 | lr 1.38e-03 | grad 0.2461 +2026-04-10 06:55:10 - INFO - parrotllm.training - step 24850 | epoch 0 | loss 3.3772 | lr 1.38e-03 | grad 0.2165 +2026-04-10 06:55:13 - INFO - parrotllm.training - step 24860 | epoch 0 | loss 3.4949 | lr 1.38e-03 | grad 0.2095 +2026-04-10 06:55:16 - INFO - parrotllm.training - step 24870 | epoch 0 | loss 3.4626 | lr 1.38e-03 | grad 0.2105 +2026-04-10 06:55:19 - INFO - parrotllm.training - step 24880 | epoch 0 | loss 3.3956 | lr 1.38e-03 | grad 0.2546 +2026-04-10 06:55:22 - INFO - parrotllm.training - step 24890 | epoch 0 | loss 3.4183 | lr 1.38e-03 | grad 0.2222 +2026-04-10 06:55:25 - INFO - parrotllm.training - step 24900 | epoch 0 | loss 3.5282 | lr 1.38e-03 | grad 0.1884 +2026-04-10 06:55:28 - INFO - parrotllm.training - step 24910 | epoch 0 | loss 3.5111 | lr 1.38e-03 | grad 0.2433 +2026-04-10 06:55:31 - INFO - parrotllm.training - step 24920 | epoch 0 | loss 3.4406 | lr 1.38e-03 | grad 0.2052 +2026-04-10 06:55:35 - INFO - parrotllm.training - step 24930 | epoch 0 | loss 3.3879 | lr 1.38e-03 | grad 0.2519 +2026-04-10 06:55:38 - INFO - parrotllm.training - step 24940 | epoch 0 | loss 3.4392 | lr 1.38e-03 | grad 0.2282 +2026-04-10 06:55:41 - INFO - parrotllm.training - step 24950 | epoch 0 | loss 3.5414 | lr 1.38e-03 | grad 0.2463 +2026-04-10 06:55:44 - INFO - parrotllm.training - step 24960 | epoch 0 | loss 3.4160 | lr 1.38e-03 | grad 0.1822 +2026-04-10 06:55:47 - INFO - parrotllm.training - step 24970 | epoch 0 | loss 3.4904 | lr 1.38e-03 | grad 0.2253 +2026-04-10 06:55:50 - INFO - parrotllm.training - step 24980 | epoch 0 | loss 3.3822 | lr 1.38e-03 | grad 0.1831 +2026-04-10 06:55:53 - INFO - parrotllm.training - step 24990 | epoch 0 | loss 3.5113 | lr 1.38e-03 | grad 0.2003 +2026-04-10 06:55:56 - INFO - parrotllm.training - step 25000 | epoch 0 | loss 3.4303 | lr 1.38e-03 | grad 0.2412 +2026-04-10 06:55:56 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:55:56 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:55:59 - INFO - parrotllm.training - Train: loss=3.4303, ppl=30.89 +2026-04-10 06:55:59 - INFO - parrotllm.training - Val: loss=3.4159, ppl=30.44 +2026-04-10 06:56:00 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4159_epoch_0000_step_0025000.pt +2026-04-10 06:56:01 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4169, min_delta=0.001000). +2026-04-10 06:56:01 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:56:02 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0025000.pt +2026-04-10 06:56:06 - INFO - parrotllm.training - step 25010 | epoch 0 | loss 3.5491 | lr 1.38e-03 | grad 0.2011 +2026-04-10 06:56:09 - INFO - parrotllm.training - step 25020 | epoch 0 | loss 3.4195 | lr 1.38e-03 | grad 0.2227 +2026-04-10 06:56:12 - INFO - parrotllm.training - step 25030 | epoch 0 | loss 3.5387 | lr 1.38e-03 | grad 0.1859 +2026-04-10 06:56:15 - INFO - parrotllm.training - step 25040 | epoch 0 | loss 3.5504 | lr 1.38e-03 | grad 0.2124 +2026-04-10 06:56:18 - INFO - parrotllm.training - step 25050 | epoch 0 | loss 3.4868 | lr 1.38e-03 | grad 0.2064 +2026-04-10 06:56:21 - INFO - parrotllm.training - step 25060 | epoch 0 | loss 3.3797 | lr 1.38e-03 | grad 0.1972 +2026-04-10 06:56:25 - INFO - parrotllm.training - step 25070 | epoch 0 | loss 3.4357 | lr 1.38e-03 | grad 0.2206 +2026-04-10 06:56:28 - INFO - parrotllm.training - step 25080 | epoch 0 | loss 3.6012 | lr 1.38e-03 | grad 0.2289 +2026-04-10 06:56:31 - INFO - parrotllm.training - step 25090 | epoch 0 | loss 3.5167 | lr 1.38e-03 | grad 0.2010 +2026-04-10 06:56:34 - INFO - parrotllm.training - step 25100 | epoch 0 | loss 3.4899 | lr 1.38e-03 | grad 0.2261 +2026-04-10 06:56:37 - INFO - parrotllm.training - step 25110 | epoch 0 | loss 3.5341 | lr 1.38e-03 | grad 0.2282 +2026-04-10 06:56:40 - INFO - parrotllm.training - step 25120 | epoch 0 | loss 3.5063 | lr 1.38e-03 | grad 0.2132 +2026-04-10 06:56:43 - INFO - parrotllm.training - step 25130 | epoch 0 | loss 3.5387 | lr 1.38e-03 | grad 0.2174 +2026-04-10 06:56:46 - INFO - parrotllm.training - step 25140 | epoch 0 | loss 3.5502 | lr 1.38e-03 | grad 0.2210 +2026-04-10 06:56:49 - INFO - parrotllm.training - step 25150 | epoch 0 | loss 3.5484 | lr 1.38e-03 | grad 0.2215 +2026-04-10 06:56:52 - INFO - parrotllm.training - step 25160 | epoch 0 | loss 3.4487 | lr 1.38e-03 | grad 0.2224 +2026-04-10 06:56:55 - INFO - parrotllm.training - step 25170 | epoch 0 | loss 3.5125 | lr 1.38e-03 | grad 0.2446 +2026-04-10 06:56:58 - INFO - parrotllm.training - step 25180 | epoch 0 | loss 3.3728 | lr 1.38e-03 | grad 0.2309 +2026-04-10 06:57:01 - INFO - parrotllm.training - step 25190 | epoch 0 | loss 3.5445 | lr 1.38e-03 | grad 0.1954 +2026-04-10 06:57:04 - INFO - parrotllm.training - step 25200 | epoch 0 | loss 3.4882 | lr 1.38e-03 | grad 0.1829 +2026-04-10 06:57:07 - INFO - parrotllm.training - step 25210 | epoch 0 | loss 3.3683 | lr 1.38e-03 | grad 0.1971 +2026-04-10 06:57:10 - INFO - parrotllm.training - step 25220 | epoch 0 | loss 3.4870 | lr 1.38e-03 | grad 0.2069 +2026-04-10 06:57:13 - INFO - parrotllm.training - step 25230 | epoch 0 | loss 3.3800 | lr 1.38e-03 | grad 0.2060 +2026-04-10 06:57:16 - INFO - parrotllm.training - step 25240 | epoch 0 | loss 3.5633 | lr 1.37e-03 | grad 0.1963 +2026-04-10 06:57:19 - INFO - parrotllm.training - step 25250 | epoch 0 | loss 3.4585 | lr 1.37e-03 | grad 0.2171 +2026-04-10 06:57:22 - INFO - parrotllm.training - step 25260 | epoch 0 | loss 3.5008 | lr 1.37e-03 | grad 0.1894 +2026-04-10 06:57:26 - INFO - parrotllm.training - step 25270 | epoch 0 | loss 3.4021 | lr 1.37e-03 | grad 0.1990 +2026-04-10 06:57:29 - INFO - parrotllm.training - step 25280 | epoch 0 | loss 3.4554 | lr 1.37e-03 | grad 0.2124 +2026-04-10 06:57:32 - INFO - parrotllm.training - step 25290 | epoch 0 | loss 3.4530 | lr 1.37e-03 | grad 0.1971 +2026-04-10 06:57:35 - INFO - parrotllm.training - step 25300 | epoch 0 | loss 3.5396 | lr 1.37e-03 | grad 0.2297 +2026-04-10 06:57:38 - INFO - parrotllm.training - step 25310 | epoch 0 | loss 3.4060 | lr 1.37e-03 | grad 0.2073 +2026-04-10 06:57:41 - INFO - parrotllm.training - step 25320 | epoch 0 | loss 3.4465 | lr 1.37e-03 | grad 0.2224 +2026-04-10 06:57:44 - INFO - parrotllm.training - step 25330 | epoch 0 | loss 3.4543 | lr 1.37e-03 | grad 0.2042 +2026-04-10 06:57:47 - INFO - parrotllm.training - step 25340 | epoch 0 | loss 3.5278 | lr 1.37e-03 | grad 0.1980 +2026-04-10 06:57:50 - INFO - parrotllm.training - step 25350 | epoch 0 | loss 3.4493 | lr 1.37e-03 | grad 0.2170 +2026-04-10 06:57:53 - INFO - parrotllm.training - step 25360 | epoch 0 | loss 3.4117 | lr 1.37e-03 | grad 0.2014 +2026-04-10 06:57:56 - INFO - parrotllm.training - step 25370 | epoch 0 | loss 3.4660 | lr 1.37e-03 | grad 0.2462 +2026-04-10 06:57:59 - INFO - parrotllm.training - step 25380 | epoch 0 | loss 3.3429 | lr 1.37e-03 | grad 0.2033 +2026-04-10 06:58:02 - INFO - parrotllm.training - step 25390 | epoch 0 | loss 3.4221 | lr 1.37e-03 | grad 0.2382 +2026-04-10 06:58:05 - INFO - parrotllm.training - step 25400 | epoch 0 | loss 3.5105 | lr 1.37e-03 | grad 0.2175 +2026-04-10 06:58:08 - INFO - parrotllm.training - step 25410 | epoch 0 | loss 3.4704 | lr 1.37e-03 | grad 0.2273 +2026-04-10 06:58:11 - INFO - parrotllm.training - step 25420 | epoch 0 | loss 3.4359 | lr 1.37e-03 | grad 0.2255 +2026-04-10 06:58:14 - INFO - parrotllm.training - step 25430 | epoch 0 | loss 3.4467 | lr 1.37e-03 | grad 0.1911 +2026-04-10 06:58:17 - INFO - parrotllm.training - step 25440 | epoch 0 | loss 3.4278 | lr 1.37e-03 | grad 0.2302 +2026-04-10 06:58:20 - INFO - parrotllm.training - step 25450 | epoch 0 | loss 3.4800 | lr 1.37e-03 | grad 0.1927 +2026-04-10 06:58:23 - INFO - parrotllm.training - step 25460 | epoch 0 | loss 3.5456 | lr 1.37e-03 | grad 0.2556 +2026-04-10 06:58:27 - INFO - parrotllm.training - step 25470 | epoch 0 | loss 3.4961 | lr 1.37e-03 | grad 0.2241 +2026-04-10 06:58:30 - INFO - parrotllm.training - step 25480 | epoch 0 | loss 3.4677 | lr 1.37e-03 | grad 0.2175 +2026-04-10 06:58:33 - INFO - parrotllm.training - step 25490 | epoch 0 | loss 3.4832 | lr 1.37e-03 | grad 0.2162 +2026-04-10 06:58:36 - INFO - parrotllm.training - step 25500 | epoch 0 | loss 3.4649 | lr 1.37e-03 | grad 0.1891 +2026-04-10 06:58:36 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 06:58:36 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:58:39 - INFO - parrotllm.training - Train: loss=3.4649, ppl=31.97 +2026-04-10 06:58:39 - INFO - parrotllm.training - Val: loss=3.4154, ppl=30.43 +2026-04-10 06:58:39 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 06:58:40 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4154_epoch_0000_step_0025500.pt +2026-04-10 06:58:41 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 06:58:44 - INFO - parrotllm.training - step 25510 | epoch 0 | loss 3.4838 | lr 1.37e-03 | grad 0.2381 +2026-04-10 06:58:47 - INFO - parrotllm.training - step 25520 | epoch 0 | loss 3.5126 | lr 1.37e-03 | grad 0.2158 +2026-04-10 06:58:50 - INFO - parrotllm.training - step 25530 | epoch 0 | loss 3.4253 | lr 1.37e-03 | grad 0.2072 +2026-04-10 06:58:53 - INFO - parrotllm.training - step 25540 | epoch 0 | loss 3.3904 | lr 1.37e-03 | grad 0.1824 +2026-04-10 06:58:56 - INFO - parrotllm.training - step 25550 | epoch 0 | loss 3.5515 | lr 1.37e-03 | grad 0.2430 +2026-04-10 06:58:59 - INFO - parrotllm.training - step 25560 | epoch 0 | loss 3.4897 | lr 1.37e-03 | grad 0.2133 +2026-04-10 06:59:02 - INFO - parrotllm.training - step 25570 | epoch 0 | loss 3.4544 | lr 1.37e-03 | grad 0.2072 +2026-04-10 06:59:05 - INFO - parrotllm.training - step 25580 | epoch 0 | loss 3.4451 | lr 1.37e-03 | grad 0.2085 +2026-04-10 06:59:08 - INFO - parrotllm.training - step 25590 | epoch 0 | loss 3.5516 | lr 1.37e-03 | grad 0.2009 +2026-04-10 06:59:11 - INFO - parrotllm.training - step 25600 | epoch 0 | loss 3.5406 | lr 1.37e-03 | grad 0.2103 +2026-04-10 06:59:14 - INFO - parrotllm.training - step 25610 | epoch 0 | loss 3.4754 | lr 1.37e-03 | grad 0.2187 +2026-04-10 06:59:17 - INFO - parrotllm.training - step 25620 | epoch 0 | loss 3.5321 | lr 1.37e-03 | grad 0.2128 +2026-04-10 06:59:20 - INFO - parrotllm.training - step 25630 | epoch 0 | loss 3.3672 | lr 1.37e-03 | grad 0.2192 +2026-04-10 06:59:23 - INFO - parrotllm.training - step 25640 | epoch 0 | loss 3.3620 | lr 1.37e-03 | grad 0.2365 +2026-04-10 06:59:26 - INFO - parrotllm.training - step 25650 | epoch 0 | loss 3.6083 | lr 1.37e-03 | grad 0.2059 +2026-04-10 06:59:30 - INFO - parrotllm.training - step 25660 | epoch 0 | loss 3.5701 | lr 1.37e-03 | grad 0.2058 +2026-04-10 06:59:33 - INFO - parrotllm.training - step 25670 | epoch 0 | loss 3.3791 | lr 1.37e-03 | grad 0.1883 +2026-04-10 06:59:36 - INFO - parrotllm.training - step 25680 | epoch 0 | loss 3.4776 | lr 1.37e-03 | grad 0.2248 +2026-04-10 06:59:39 - INFO - parrotllm.training - step 25690 | epoch 0 | loss 3.3991 | lr 1.37e-03 | grad 0.2036 +2026-04-10 06:59:42 - INFO - parrotllm.training - step 25700 | epoch 0 | loss 3.5352 | lr 1.37e-03 | grad 0.2268 +2026-04-10 06:59:45 - INFO - parrotllm.training - step 25710 | epoch 0 | loss 3.5069 | lr 1.37e-03 | grad 0.2313 +2026-04-10 06:59:48 - INFO - parrotllm.training - step 25720 | epoch 0 | loss 3.4730 | lr 1.37e-03 | grad 0.1891 +2026-04-10 06:59:51 - INFO - parrotllm.training - step 25730 | epoch 0 | loss 3.4582 | lr 1.37e-03 | grad 0.1833 +2026-04-10 06:59:54 - INFO - parrotllm.training - step 25740 | epoch 0 | loss 3.4837 | lr 1.37e-03 | grad 0.2154 +2026-04-10 06:59:57 - INFO - parrotllm.training - step 25750 | epoch 0 | loss 3.3908 | lr 1.37e-03 | grad 0.2156 +2026-04-10 07:00:00 - INFO - parrotllm.training - step 25760 | epoch 0 | loss 3.4449 | lr 1.37e-03 | grad 0.1939 +2026-04-10 07:00:03 - INFO - parrotllm.training - step 25770 | epoch 0 | loss 3.4289 | lr 1.37e-03 | grad 0.1965 +2026-04-10 07:00:06 - INFO - parrotllm.training - step 25780 | epoch 0 | loss 3.3424 | lr 1.37e-03 | grad 0.2508 +2026-04-10 07:00:09 - INFO - parrotllm.training - step 25790 | epoch 0 | loss 3.5153 | lr 1.37e-03 | grad 0.2082 +2026-04-10 07:00:12 - INFO - parrotllm.training - step 25800 | epoch 0 | loss 3.4864 | lr 1.37e-03 | grad 0.2018 +2026-04-10 07:00:15 - INFO - parrotllm.training - step 25810 | epoch 0 | loss 3.6204 | lr 1.37e-03 | grad 0.2010 +2026-04-10 07:00:18 - INFO - parrotllm.training - step 25820 | epoch 0 | loss 3.5153 | lr 1.37e-03 | grad 0.2371 +2026-04-10 07:00:21 - INFO - parrotllm.training - step 25830 | epoch 0 | loss 3.4751 | lr 1.37e-03 | grad 0.2642 +2026-04-10 07:00:24 - INFO - parrotllm.training - step 25840 | epoch 0 | loss 3.3974 | lr 1.37e-03 | grad 0.2232 +2026-04-10 07:00:28 - INFO - parrotllm.training - step 25850 | epoch 0 | loss 3.4427 | lr 1.37e-03 | grad 0.2179 +2026-04-10 07:00:31 - INFO - parrotllm.training - step 25860 | epoch 0 | loss 3.4614 | lr 1.37e-03 | grad 0.2225 +2026-04-10 07:00:34 - INFO - parrotllm.training - step 25870 | epoch 0 | loss 3.5250 | lr 1.37e-03 | grad 0.1974 +2026-04-10 07:00:37 - INFO - parrotllm.training - step 25880 | epoch 0 | loss 3.4741 | lr 1.37e-03 | grad 0.1970 +2026-04-10 07:00:40 - INFO - parrotllm.training - step 25890 | epoch 0 | loss 3.4218 | lr 1.37e-03 | grad 0.2193 +2026-04-10 07:00:43 - INFO - parrotllm.training - step 25900 | epoch 0 | loss 3.3698 | lr 1.37e-03 | grad 0.2034 +2026-04-10 07:00:46 - INFO - parrotllm.training - step 25910 | epoch 0 | loss 3.4546 | lr 1.37e-03 | grad 0.2028 +2026-04-10 07:00:49 - INFO - parrotllm.training - step 25920 | epoch 0 | loss 3.4776 | lr 1.37e-03 | grad 0.2018 +2026-04-10 07:00:52 - INFO - parrotllm.training - step 25930 | epoch 0 | loss 3.5951 | lr 1.37e-03 | grad 0.2131 +2026-04-10 07:00:55 - INFO - parrotllm.training - step 25940 | epoch 0 | loss 3.4556 | lr 1.37e-03 | grad 0.2261 +2026-04-10 07:00:58 - INFO - parrotllm.training - step 25950 | epoch 0 | loss 3.4380 | lr 1.37e-03 | grad 0.2073 +2026-04-10 07:01:01 - INFO - parrotllm.training - step 25960 | epoch 0 | loss 3.4442 | lr 1.37e-03 | grad 0.2295 +2026-04-10 07:01:04 - INFO - parrotllm.training - step 25970 | epoch 0 | loss 3.4467 | lr 1.37e-03 | grad 0.2076 +2026-04-10 07:01:07 - INFO - parrotllm.training - step 25980 | epoch 0 | loss 3.4887 | lr 1.37e-03 | grad 0.2240 +2026-04-10 07:01:10 - INFO - parrotllm.training - step 25990 | epoch 0 | loss 3.5117 | lr 1.37e-03 | grad 0.1963 +2026-04-10 07:01:13 - INFO - parrotllm.training - step 26000 | epoch 0 | loss 3.4447 | lr 1.37e-03 | grad 0.2142 +2026-04-10 07:01:13 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:01:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:01:16 - INFO - parrotllm.training - Train: loss=3.4447, ppl=31.34 +2026-04-10 07:01:16 - INFO - parrotllm.training - Val: loss=3.4130, ppl=30.36 +2026-04-10 07:01:16 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:01:17 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4130_epoch_0000_step_0026000.pt +2026-04-10 07:01:19 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:01:22 - INFO - parrotllm.training - step 26010 | epoch 0 | loss 3.5802 | lr 1.37e-03 | grad 0.2089 +2026-04-10 07:01:25 - INFO - parrotllm.training - step 26020 | epoch 0 | loss 3.4181 | lr 1.37e-03 | grad 0.2003 +2026-04-10 07:01:28 - INFO - parrotllm.training - step 26030 | epoch 0 | loss 3.3639 | lr 1.37e-03 | grad 0.2168 +2026-04-10 07:01:31 - INFO - parrotllm.training - step 26040 | epoch 0 | loss 3.3647 | lr 1.37e-03 | grad 0.2014 +2026-04-10 07:01:34 - INFO - parrotllm.training - step 26050 | epoch 0 | loss 3.6060 | lr 1.37e-03 | grad 0.2022 +2026-04-10 07:01:37 - INFO - parrotllm.training - step 26060 | epoch 0 | loss 3.4914 | lr 1.37e-03 | grad 0.2006 +2026-04-10 07:01:40 - INFO - parrotllm.training - step 26070 | epoch 0 | loss 3.4527 | lr 1.37e-03 | grad 0.1881 +2026-04-10 07:01:43 - INFO - parrotllm.training - step 26080 | epoch 0 | loss 3.4117 | lr 1.37e-03 | grad 0.2519 +2026-04-10 07:01:46 - INFO - parrotllm.training - step 26090 | epoch 0 | loss 3.4451 | lr 1.37e-03 | grad 0.2306 +2026-04-10 07:01:49 - INFO - parrotllm.training - step 26100 | epoch 0 | loss 3.5475 | lr 1.37e-03 | grad 0.1863 +2026-04-10 07:01:52 - INFO - parrotllm.training - step 26110 | epoch 0 | loss 3.5690 | lr 1.37e-03 | grad 0.2093 +2026-04-10 07:01:55 - INFO - parrotllm.training - step 26120 | epoch 0 | loss 3.5129 | lr 1.37e-03 | grad 0.2156 +2026-04-10 07:01:58 - INFO - parrotllm.training - step 26130 | epoch 0 | loss 3.5169 | lr 1.37e-03 | grad 0.2144 +2026-04-10 07:02:01 - INFO - parrotllm.training - step 26140 | epoch 0 | loss 3.4245 | lr 1.37e-03 | grad 0.2043 +2026-04-10 07:02:04 - INFO - parrotllm.training - step 26150 | epoch 0 | loss 3.3929 | lr 1.37e-03 | grad 0.1945 +2026-04-10 07:02:07 - INFO - parrotllm.training - step 26160 | epoch 0 | loss 3.5697 | lr 1.37e-03 | grad 0.2133 +2026-04-10 07:02:10 - INFO - parrotllm.training - step 26170 | epoch 0 | loss 3.4816 | lr 1.37e-03 | grad 0.2206 +2026-04-10 07:02:13 - INFO - parrotllm.training - step 26180 | epoch 0 | loss 3.4850 | lr 1.37e-03 | grad 0.2109 +2026-04-10 07:02:16 - INFO - parrotllm.training - step 26190 | epoch 0 | loss 3.4605 | lr 1.37e-03 | grad 0.2360 +2026-04-10 07:02:19 - INFO - parrotllm.training - step 26200 | epoch 0 | loss 3.4538 | lr 1.36e-03 | grad 0.1996 +2026-04-10 07:02:23 - INFO - parrotllm.training - step 26210 | epoch 0 | loss 3.4284 | lr 1.36e-03 | grad 0.2032 +2026-04-10 07:02:26 - INFO - parrotllm.training - step 26220 | epoch 0 | loss 3.5473 | lr 1.36e-03 | grad 0.2136 +2026-04-10 07:02:29 - INFO - parrotllm.training - step 26230 | epoch 0 | loss 3.4278 | lr 1.36e-03 | grad 0.2307 +2026-04-10 07:02:32 - INFO - parrotllm.training - step 26240 | epoch 0 | loss 3.4481 | lr 1.36e-03 | grad 0.2461 +2026-04-10 07:02:35 - INFO - parrotllm.training - step 26250 | epoch 0 | loss 3.4678 | lr 1.36e-03 | grad 0.1993 +2026-04-10 07:02:38 - INFO - parrotllm.training - step 26260 | epoch 0 | loss 3.5388 | lr 1.36e-03 | grad 0.1926 +2026-04-10 07:02:41 - INFO - parrotllm.training - step 26270 | epoch 0 | loss 3.5178 | lr 1.36e-03 | grad 0.2333 +2026-04-10 07:02:44 - INFO - parrotllm.training - step 26280 | epoch 0 | loss 3.5408 | lr 1.36e-03 | grad 0.2185 +2026-04-10 07:02:47 - INFO - parrotllm.training - step 26290 | epoch 0 | loss 3.4733 | lr 1.36e-03 | grad 0.2137 +2026-04-10 07:02:50 - INFO - parrotllm.training - step 26300 | epoch 0 | loss 3.4302 | lr 1.36e-03 | grad 0.2004 +2026-04-10 07:02:52 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 26316/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 07:02:53 - INFO - parrotllm.training - step 26310 | epoch 0 | loss 3.5520 | lr 1.36e-03 | grad 0.2111 +2026-04-10 07:02:56 - INFO - parrotllm.training - step 26320 | epoch 0 | loss 3.5169 | lr 1.36e-03 | grad 0.2207 +2026-04-10 07:02:59 - INFO - parrotllm.training - step 26330 | epoch 0 | loss 3.4102 | lr 1.36e-03 | grad 0.2146 +2026-04-10 07:03:03 - INFO - parrotllm.training - step 26340 | epoch 0 | loss 3.4107 | lr 1.36e-03 | grad 0.2001 +2026-04-10 07:03:06 - INFO - parrotllm.training - step 26350 | epoch 0 | loss 3.4512 | lr 1.36e-03 | grad 0.2070 +2026-04-10 07:03:09 - INFO - parrotllm.training - step 26360 | epoch 0 | loss 3.4942 | lr 1.36e-03 | grad 0.1972 +2026-04-10 07:03:12 - INFO - parrotllm.training - step 26370 | epoch 0 | loss 3.4957 | lr 1.36e-03 | grad 0.2034 +2026-04-10 07:03:15 - INFO - parrotllm.training - step 26380 | epoch 0 | loss 3.4888 | lr 1.36e-03 | grad 0.1879 +2026-04-10 07:03:18 - INFO - parrotllm.training - step 26390 | epoch 0 | loss 3.4007 | lr 1.36e-03 | grad 0.1942 +2026-04-10 07:03:21 - INFO - parrotllm.training - step 26400 | epoch 0 | loss 3.5615 | lr 1.36e-03 | grad 0.1909 +2026-04-10 07:03:24 - INFO - parrotllm.training - step 26410 | epoch 0 | loss 3.4791 | lr 1.36e-03 | grad 0.2302 +2026-04-10 07:03:27 - INFO - parrotllm.training - step 26420 | epoch 0 | loss 3.4366 | lr 1.36e-03 | grad 0.2270 +2026-04-10 07:03:30 - INFO - parrotllm.training - step 26430 | epoch 0 | loss 3.4386 | lr 1.36e-03 | grad 0.2097 +2026-04-10 07:03:33 - INFO - parrotllm.training - step 26440 | epoch 0 | loss 3.5984 | lr 1.36e-03 | grad 0.1906 +2026-04-10 07:03:36 - INFO - parrotllm.training - step 26450 | epoch 0 | loss 3.4814 | lr 1.36e-03 | grad 0.2160 +2026-04-10 07:03:39 - INFO - parrotllm.training - step 26460 | epoch 0 | loss 3.3809 | lr 1.36e-03 | grad 0.2038 +2026-04-10 07:03:42 - INFO - parrotllm.training - step 26470 | epoch 0 | loss 3.5324 | lr 1.36e-03 | grad 0.2185 +2026-04-10 07:03:45 - INFO - parrotllm.training - step 26480 | epoch 0 | loss 3.5132 | lr 1.36e-03 | grad 0.2700 +2026-04-10 07:03:48 - INFO - parrotllm.training - step 26490 | epoch 0 | loss 3.5163 | lr 1.36e-03 | grad 0.2267 +2026-04-10 07:03:51 - INFO - parrotllm.training - step 26500 | epoch 0 | loss 3.4564 | lr 1.36e-03 | grad 0.2086 +2026-04-10 07:03:51 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:03:51 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:03:54 - INFO - parrotllm.training - Train: loss=3.4564, ppl=31.70 +2026-04-10 07:03:54 - INFO - parrotllm.training - Val: loss=3.4133, ppl=30.36 +2026-04-10 07:03:55 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4133_epoch_0000_step_0026500.pt +2026-04-10 07:03:56 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4130, min_delta=0.001000). +2026-04-10 07:03:56 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:04:00 - INFO - parrotllm.training - step 26510 | epoch 0 | loss 3.4745 | lr 1.36e-03 | grad 0.2169 +2026-04-10 07:04:03 - INFO - parrotllm.training - step 26520 | epoch 0 | loss 3.5339 | lr 1.36e-03 | grad 0.1817 +2026-04-10 07:04:06 - INFO - parrotllm.training - step 26530 | epoch 0 | loss 3.4579 | lr 1.36e-03 | grad 0.2192 +2026-04-10 07:04:09 - INFO - parrotllm.training - step 26540 | epoch 0 | loss 3.4588 | lr 1.36e-03 | grad 0.2559 +2026-04-10 07:04:12 - INFO - parrotllm.training - step 26550 | epoch 0 | loss 3.4900 | lr 1.36e-03 | grad 0.2227 +2026-04-10 07:04:15 - INFO - parrotllm.training - step 26560 | epoch 0 | loss 3.4054 | lr 1.36e-03 | grad 0.2303 +2026-04-10 07:04:18 - INFO - parrotllm.training - step 26570 | epoch 0 | loss 3.4239 | lr 1.36e-03 | grad 0.2405 +2026-04-10 07:04:21 - INFO - parrotllm.training - step 26580 | epoch 0 | loss 3.4667 | lr 1.36e-03 | grad 0.2181 +2026-04-10 07:04:24 - INFO - parrotllm.training - step 26590 | epoch 0 | loss 3.4175 | lr 1.36e-03 | grad 0.2189 +2026-04-10 07:04:27 - INFO - parrotllm.training - step 26600 | epoch 0 | loss 3.4528 | lr 1.36e-03 | grad 0.2216 +2026-04-10 07:04:30 - INFO - parrotllm.training - step 26610 | epoch 0 | loss 3.4569 | lr 1.36e-03 | grad 0.2042 +2026-04-10 07:04:33 - INFO - parrotllm.training - step 26620 | epoch 0 | loss 3.4814 | lr 1.36e-03 | grad 0.2082 +2026-04-10 07:04:36 - INFO - parrotllm.training - step 26630 | epoch 0 | loss 3.3939 | lr 1.36e-03 | grad 0.2308 +2026-04-10 07:04:39 - INFO - parrotllm.training - step 26640 | epoch 0 | loss 3.4266 | lr 1.36e-03 | grad 0.2101 +2026-04-10 07:04:42 - INFO - parrotllm.training - step 26650 | epoch 0 | loss 3.5145 | lr 1.36e-03 | grad 0.1904 +2026-04-10 07:04:45 - INFO - parrotllm.training - step 26660 | epoch 0 | loss 3.4057 | lr 1.36e-03 | grad 0.2094 +2026-04-10 07:04:48 - INFO - parrotllm.training - step 26670 | epoch 0 | loss 3.4435 | lr 1.36e-03 | grad 0.2026 +2026-04-10 07:04:51 - INFO - parrotllm.training - step 26680 | epoch 0 | loss 3.4393 | lr 1.36e-03 | grad 0.2085 +2026-04-10 07:04:54 - INFO - parrotllm.training - step 26690 | epoch 0 | loss 3.4628 | lr 1.36e-03 | grad 0.1823 +2026-04-10 07:04:57 - INFO - parrotllm.training - step 26700 | epoch 0 | loss 3.4042 | lr 1.36e-03 | grad 0.2236 +2026-04-10 07:05:00 - INFO - parrotllm.training - step 26710 | epoch 0 | loss 3.5611 | lr 1.36e-03 | grad 0.2215 +2026-04-10 07:05:04 - INFO - parrotllm.training - step 26720 | epoch 0 | loss 3.4539 | lr 1.36e-03 | grad 0.1891 +2026-04-10 07:05:07 - INFO - parrotllm.training - step 26730 | epoch 0 | loss 3.5197 | lr 1.36e-03 | grad 0.2151 +2026-04-10 07:05:10 - INFO - parrotllm.training - step 26740 | epoch 0 | loss 3.4345 | lr 1.36e-03 | grad 0.2030 +2026-04-10 07:05:13 - INFO - parrotllm.training - step 26750 | epoch 0 | loss 3.4627 | lr 1.36e-03 | grad 0.2319 +2026-04-10 07:05:16 - INFO - parrotllm.training - step 26760 | epoch 0 | loss 3.4524 | lr 1.36e-03 | grad 0.2096 +2026-04-10 07:05:19 - INFO - parrotllm.training - step 26770 | epoch 0 | loss 3.5085 | lr 1.36e-03 | grad 0.2208 +2026-04-10 07:05:22 - INFO - parrotllm.training - step 26780 | epoch 0 | loss 3.5112 | lr 1.36e-03 | grad 0.1852 +2026-04-10 07:05:25 - INFO - parrotllm.training - step 26790 | epoch 0 | loss 3.3075 | lr 1.36e-03 | grad 0.2273 +2026-04-10 07:05:28 - INFO - parrotllm.training - step 26800 | epoch 0 | loss 3.4238 | lr 1.36e-03 | grad 0.2037 +2026-04-10 07:05:31 - INFO - parrotllm.training - step 26810 | epoch 0 | loss 3.4750 | lr 1.36e-03 | grad 0.1931 +2026-04-10 07:05:34 - INFO - parrotllm.training - step 26820 | epoch 0 | loss 3.4714 | lr 1.36e-03 | grad 0.1946 +2026-04-10 07:05:37 - INFO - parrotllm.training - step 26830 | epoch 0 | loss 3.4043 | lr 1.36e-03 | grad 0.2289 +2026-04-10 07:05:40 - INFO - parrotllm.training - step 26840 | epoch 0 | loss 3.4522 | lr 1.36e-03 | grad 0.2003 +2026-04-10 07:05:43 - INFO - parrotllm.training - step 26850 | epoch 0 | loss 3.4203 | lr 1.36e-03 | grad 0.2038 +2026-04-10 07:05:46 - INFO - parrotllm.training - step 26860 | epoch 0 | loss 3.4127 | lr 1.36e-03 | grad 0.2060 +2026-04-10 07:05:49 - INFO - parrotllm.training - step 26870 | epoch 0 | loss 3.4305 | lr 1.36e-03 | grad 0.2021 +2026-04-10 07:05:52 - INFO - parrotllm.training - step 26880 | epoch 0 | loss 3.4421 | lr 1.36e-03 | grad 0.2040 +2026-04-10 07:05:55 - INFO - parrotllm.training - step 26890 | epoch 0 | loss 3.5007 | lr 1.36e-03 | grad 0.2143 +2026-04-10 07:05:58 - INFO - parrotllm.training - step 26900 | epoch 0 | loss 3.5038 | lr 1.36e-03 | grad 0.1882 +2026-04-10 07:06:01 - INFO - parrotllm.training - step 26910 | epoch 0 | loss 3.4744 | lr 1.36e-03 | grad 0.2044 +2026-04-10 07:06:04 - INFO - parrotllm.training - step 26920 | epoch 0 | loss 3.4944 | lr 1.36e-03 | grad 0.2028 +2026-04-10 07:06:08 - INFO - parrotllm.training - step 26930 | epoch 0 | loss 3.4932 | lr 1.36e-03 | grad 0.1950 +2026-04-10 07:06:11 - INFO - parrotllm.training - step 26940 | epoch 0 | loss 3.4383 | lr 1.36e-03 | grad 0.2137 +2026-04-10 07:06:14 - INFO - parrotllm.training - step 26950 | epoch 0 | loss 3.4905 | lr 1.36e-03 | grad 0.2122 +2026-04-10 07:06:17 - INFO - parrotllm.training - step 26960 | epoch 0 | loss 3.4085 | lr 1.36e-03 | grad 0.2031 +2026-04-10 07:06:20 - INFO - parrotllm.training - step 26970 | epoch 0 | loss 3.4385 | lr 1.36e-03 | grad 0.1950 +2026-04-10 07:06:23 - INFO - parrotllm.training - step 26980 | epoch 0 | loss 3.5133 | lr 1.36e-03 | grad 0.2298 +2026-04-10 07:06:26 - INFO - parrotllm.training - step 26990 | epoch 0 | loss 3.5129 | lr 1.36e-03 | grad 0.2250 +2026-04-10 07:06:29 - INFO - parrotllm.training - step 27000 | epoch 0 | loss 3.5351 | lr 1.36e-03 | grad 0.2263 +2026-04-10 07:06:29 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:06:29 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:06:32 - INFO - parrotllm.training - Train: loss=3.5351, ppl=34.30 +2026-04-10 07:06:32 - INFO - parrotllm.training - Val: loss=3.4081, ppl=30.21 +2026-04-10 07:06:32 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:06:33 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4081_epoch_0000_step_0027000.pt +2026-04-10 07:06:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:06:37 - INFO - parrotllm.training - step 27010 | epoch 0 | loss 3.4378 | lr 1.36e-03 | grad 0.2067 +2026-04-10 07:06:40 - INFO - parrotllm.training - step 27020 | epoch 0 | loss 3.3732 | lr 1.36e-03 | grad 0.2019 +2026-04-10 07:06:43 - INFO - parrotllm.training - step 27030 | epoch 0 | loss 3.5398 | lr 1.36e-03 | grad 0.2015 +2026-04-10 07:06:46 - INFO - parrotllm.training - step 27040 | epoch 0 | loss 3.5704 | lr 1.36e-03 | grad 0.2478 +2026-04-10 07:06:49 - INFO - parrotllm.training - step 27050 | epoch 0 | loss 3.4660 | lr 1.36e-03 | grad 0.2186 +2026-04-10 07:06:52 - INFO - parrotllm.training - step 27060 | epoch 0 | loss 3.5525 | lr 1.36e-03 | grad 0.2176 +2026-04-10 07:06:55 - INFO - parrotllm.training - step 27070 | epoch 0 | loss 3.5508 | lr 1.36e-03 | grad 0.1965 +2026-04-10 07:06:58 - INFO - parrotllm.training - step 27080 | epoch 0 | loss 3.4761 | lr 1.36e-03 | grad 0.2345 +2026-04-10 07:07:01 - INFO - parrotllm.training - step 27090 | epoch 0 | loss 3.3618 | lr 1.36e-03 | grad 0.2180 +2026-04-10 07:07:04 - INFO - parrotllm.training - step 27100 | epoch 0 | loss 3.4566 | lr 1.36e-03 | grad 0.2407 +2026-04-10 07:07:08 - INFO - parrotllm.training - step 27110 | epoch 0 | loss 3.5671 | lr 1.36e-03 | grad 0.2017 +2026-04-10 07:07:11 - INFO - parrotllm.training - step 27120 | epoch 0 | loss 3.4024 | lr 1.36e-03 | grad 0.1873 +2026-04-10 07:07:14 - INFO - parrotllm.training - step 27130 | epoch 0 | loss 3.4876 | lr 1.35e-03 | grad 0.2123 +2026-04-10 07:07:17 - INFO - parrotllm.training - step 27140 | epoch 0 | loss 3.3858 | lr 1.35e-03 | grad 0.2038 +2026-04-10 07:07:20 - INFO - parrotllm.training - step 27150 | epoch 0 | loss 3.5511 | lr 1.35e-03 | grad 0.2137 +2026-04-10 07:07:23 - INFO - parrotllm.training - step 27160 | epoch 0 | loss 3.4510 | lr 1.35e-03 | grad 0.2013 +2026-04-10 07:07:26 - INFO - parrotllm.training - step 27170 | epoch 0 | loss 3.5038 | lr 1.35e-03 | grad 0.2004 +2026-04-10 07:07:29 - INFO - parrotllm.training - step 27180 | epoch 0 | loss 3.4658 | lr 1.35e-03 | grad 0.2194 +2026-04-10 07:07:32 - INFO - parrotllm.training - step 27190 | epoch 0 | loss 3.4083 | lr 1.35e-03 | grad 0.2076 +2026-04-10 07:07:35 - INFO - parrotllm.training - step 27200 | epoch 0 | loss 3.3992 | lr 1.35e-03 | grad 0.2283 +2026-04-10 07:07:38 - INFO - parrotllm.training - step 27210 | epoch 0 | loss 3.4880 | lr 1.35e-03 | grad 0.2097 +2026-04-10 07:07:41 - INFO - parrotllm.training - step 27220 | epoch 0 | loss 3.5387 | lr 1.35e-03 | grad 0.2181 +2026-04-10 07:07:44 - INFO - parrotllm.training - step 27230 | epoch 0 | loss 3.4563 | lr 1.35e-03 | grad 0.1966 +2026-04-10 07:07:47 - INFO - parrotllm.training - step 27240 | epoch 0 | loss 3.3552 | lr 1.35e-03 | grad 0.1974 +2026-04-10 07:07:50 - INFO - parrotllm.training - step 27250 | epoch 0 | loss 3.4832 | lr 1.35e-03 | grad 0.2023 +2026-04-10 07:07:53 - INFO - parrotllm.training - step 27260 | epoch 0 | loss 3.5580 | lr 1.35e-03 | grad 0.2301 +2026-04-10 07:07:56 - INFO - parrotllm.training - step 27270 | epoch 0 | loss 3.5045 | lr 1.35e-03 | grad 0.2121 +2026-04-10 07:07:59 - INFO - parrotllm.training - step 27280 | epoch 0 | loss 3.4205 | lr 1.35e-03 | grad 0.2183 +2026-04-10 07:08:02 - INFO - parrotllm.training - step 27290 | epoch 0 | loss 3.4208 | lr 1.35e-03 | grad 0.1990 +2026-04-10 07:08:06 - INFO - parrotllm.training - step 27300 | epoch 0 | loss 3.5456 | lr 1.35e-03 | grad 0.2499 +2026-04-10 07:08:09 - INFO - parrotllm.training - step 27310 | epoch 0 | loss 3.5619 | lr 1.35e-03 | grad 0.2013 +2026-04-10 07:08:12 - INFO - parrotllm.training - step 27320 | epoch 0 | loss 3.4141 | lr 1.35e-03 | grad 0.1973 +2026-04-10 07:08:15 - INFO - parrotllm.training - step 27330 | epoch 0 | loss 3.4532 | lr 1.35e-03 | grad 0.2132 +2026-04-10 07:08:18 - INFO - parrotllm.training - step 27340 | epoch 0 | loss 3.5087 | lr 1.35e-03 | grad 0.2078 +2026-04-10 07:08:21 - INFO - parrotllm.training - step 27350 | epoch 0 | loss 3.4626 | lr 1.35e-03 | grad 0.2307 +2026-04-10 07:08:24 - INFO - parrotllm.training - step 27360 | epoch 0 | loss 3.4714 | lr 1.35e-03 | grad 0.2282 +2026-04-10 07:08:27 - INFO - parrotllm.training - step 27370 | epoch 0 | loss 3.4694 | lr 1.35e-03 | grad 0.2139 +2026-04-10 07:08:30 - INFO - parrotllm.training - step 27380 | epoch 0 | loss 3.4551 | lr 1.35e-03 | grad 0.2178 +2026-04-10 07:08:33 - INFO - parrotllm.training - step 27390 | epoch 0 | loss 3.4310 | lr 1.35e-03 | grad 0.1928 +2026-04-10 07:08:36 - INFO - parrotllm.training - step 27400 | epoch 0 | loss 3.5507 | lr 1.35e-03 | grad 0.2394 +2026-04-10 07:08:39 - INFO - parrotllm.training - step 27410 | epoch 0 | loss 3.4609 | lr 1.35e-03 | grad 0.2178 +2026-04-10 07:08:42 - INFO - parrotllm.training - step 27420 | epoch 0 | loss 3.5786 | lr 1.35e-03 | grad 0.1964 +2026-04-10 07:08:45 - INFO - parrotllm.training - step 27430 | epoch 0 | loss 3.4818 | lr 1.35e-03 | grad 0.2062 +2026-04-10 07:08:48 - INFO - parrotllm.training - step 27440 | epoch 0 | loss 3.4195 | lr 1.35e-03 | grad 0.2037 +2026-04-10 07:08:51 - INFO - parrotllm.training - step 27450 | epoch 0 | loss 3.4585 | lr 1.35e-03 | grad 0.2608 +2026-04-10 07:08:54 - INFO - parrotllm.training - step 27460 | epoch 0 | loss 3.3962 | lr 1.35e-03 | grad 0.2042 +2026-04-10 07:08:57 - INFO - parrotllm.training - step 27470 | epoch 0 | loss 3.4447 | lr 1.35e-03 | grad 0.2023 +2026-04-10 07:09:00 - INFO - parrotllm.training - step 27480 | epoch 0 | loss 3.5539 | lr 1.35e-03 | grad 0.1998 +2026-04-10 07:09:04 - INFO - parrotllm.training - step 27490 | epoch 0 | loss 3.3283 | lr 1.35e-03 | grad 0.2286 +2026-04-10 07:09:07 - INFO - parrotllm.training - step 27500 | epoch 0 | loss 3.5230 | lr 1.35e-03 | grad 0.1998 +2026-04-10 07:09:07 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:09:07 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:09:10 - INFO - parrotllm.training - Train: loss=3.5230, ppl=33.89 +2026-04-10 07:09:10 - INFO - parrotllm.training - Val: loss=3.4082, ppl=30.21 +2026-04-10 07:09:11 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4082_epoch_0000_step_0027500.pt +2026-04-10 07:09:12 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4081, min_delta=0.001000). +2026-04-10 07:09:12 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:09:14 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0027500.pt +2026-04-10 07:09:18 - INFO - parrotllm.training - step 27510 | epoch 0 | loss 3.5009 | lr 1.35e-03 | grad 0.2193 +2026-04-10 07:09:21 - INFO - parrotllm.training - step 27520 | epoch 0 | loss 3.4118 | lr 1.35e-03 | grad 0.2301 +2026-04-10 07:09:24 - INFO - parrotllm.training - step 27530 | epoch 0 | loss 3.4739 | lr 1.35e-03 | grad 0.1926 +2026-04-10 07:09:27 - INFO - parrotllm.training - step 27540 | epoch 0 | loss 3.4992 | lr 1.35e-03 | grad 0.2301 +2026-04-10 07:09:30 - INFO - parrotllm.training - step 27550 | epoch 0 | loss 3.4580 | lr 1.35e-03 | grad 0.2088 +2026-04-10 07:09:33 - INFO - parrotllm.training - step 27560 | epoch 0 | loss 3.4088 | lr 1.35e-03 | grad 0.2089 +2026-04-10 07:09:36 - INFO - parrotllm.training - step 27570 | epoch 0 | loss 3.4251 | lr 1.35e-03 | grad 0.2054 +2026-04-10 07:09:39 - INFO - parrotllm.training - step 27580 | epoch 0 | loss 3.4174 | lr 1.35e-03 | grad 0.2307 +2026-04-10 07:09:42 - INFO - parrotllm.training - step 27590 | epoch 0 | loss 3.4879 | lr 1.35e-03 | grad 0.1994 +2026-04-10 07:09:45 - INFO - parrotllm.training - step 27600 | epoch 0 | loss 3.4488 | lr 1.35e-03 | grad 0.1992 +2026-04-10 07:09:48 - INFO - parrotllm.training - step 27610 | epoch 0 | loss 3.4180 | lr 1.35e-03 | grad 0.2124 +2026-04-10 07:09:51 - INFO - parrotllm.training - step 27620 | epoch 0 | loss 3.4329 | lr 1.35e-03 | grad 0.2610 +2026-04-10 07:09:54 - INFO - parrotllm.training - step 27630 | epoch 0 | loss 3.5003 | lr 1.35e-03 | grad 0.1962 +2026-04-10 07:09:57 - INFO - parrotllm.training - step 27640 | epoch 0 | loss 3.4828 | lr 1.35e-03 | grad 0.2066 +2026-04-10 07:10:01 - INFO - parrotllm.training - step 27650 | epoch 0 | loss 3.3648 | lr 1.35e-03 | grad 0.2327 +2026-04-10 07:10:04 - INFO - parrotllm.training - step 27660 | epoch 0 | loss 3.3183 | lr 1.35e-03 | grad 0.2134 +2026-04-10 07:10:07 - INFO - parrotllm.training - step 27670 | epoch 0 | loss 3.4233 | lr 1.35e-03 | grad 0.2282 +2026-04-10 07:10:10 - INFO - parrotllm.training - step 27680 | epoch 0 | loss 3.3548 | lr 1.35e-03 | grad 0.2351 +2026-04-10 07:10:13 - INFO - parrotllm.training - step 27690 | epoch 0 | loss 3.4303 | lr 1.35e-03 | grad 0.2210 +2026-04-10 07:10:16 - INFO - parrotllm.training - step 27700 | epoch 0 | loss 3.4871 | lr 1.35e-03 | grad 0.2152 +2026-04-10 07:10:19 - INFO - parrotllm.training - step 27710 | epoch 0 | loss 3.4932 | lr 1.35e-03 | grad 0.2311 +2026-04-10 07:10:22 - INFO - parrotllm.training - step 27720 | epoch 0 | loss 3.4565 | lr 1.35e-03 | grad 0.1886 +2026-04-10 07:10:25 - INFO - parrotllm.training - step 27730 | epoch 0 | loss 3.5192 | lr 1.35e-03 | grad 0.2196 +2026-04-10 07:10:28 - INFO - parrotllm.training - step 27740 | epoch 0 | loss 3.3239 | lr 1.35e-03 | grad 0.1925 +2026-04-10 07:10:31 - INFO - parrotllm.training - step 27750 | epoch 0 | loss 3.4574 | lr 1.35e-03 | grad 0.2158 +2026-04-10 07:10:34 - INFO - parrotllm.training - step 27760 | epoch 0 | loss 3.4342 | lr 1.35e-03 | grad 0.2011 +2026-04-10 07:10:37 - INFO - parrotllm.training - step 27770 | epoch 0 | loss 3.5242 | lr 1.35e-03 | grad 0.2066 +2026-04-10 07:10:40 - INFO - parrotllm.training - step 27780 | epoch 0 | loss 3.6646 | lr 1.35e-03 | grad 0.2109 +2026-04-10 07:10:43 - INFO - parrotllm.training - step 27790 | epoch 0 | loss 3.5139 | lr 1.35e-03 | grad 0.2085 +2026-04-10 07:10:46 - INFO - parrotllm.training - step 27800 | epoch 0 | loss 3.4629 | lr 1.35e-03 | grad 0.2202 +2026-04-10 07:10:49 - INFO - parrotllm.training - step 27810 | epoch 0 | loss 3.5703 | lr 1.35e-03 | grad 0.2084 +2026-04-10 07:10:52 - INFO - parrotllm.training - step 27820 | epoch 0 | loss 3.4170 | lr 1.35e-03 | grad 0.2053 +2026-04-10 07:10:55 - INFO - parrotllm.training - step 27830 | epoch 0 | loss 3.5578 | lr 1.35e-03 | grad 0.2114 +2026-04-10 07:10:58 - INFO - parrotllm.training - step 27840 | epoch 0 | loss 3.4786 | lr 1.35e-03 | grad 0.2145 +2026-04-10 07:11:02 - INFO - parrotllm.training - step 27850 | epoch 0 | loss 3.4396 | lr 1.35e-03 | grad 0.2319 +2026-04-10 07:11:05 - INFO - parrotllm.training - step 27860 | epoch 0 | loss 3.4396 | lr 1.35e-03 | grad 0.1793 +2026-04-10 07:11:08 - INFO - parrotllm.training - step 27870 | epoch 0 | loss 3.5157 | lr 1.35e-03 | grad 0.2278 +2026-04-10 07:11:11 - INFO - parrotllm.training - step 27880 | epoch 0 | loss 3.4779 | lr 1.35e-03 | grad 0.2108 +2026-04-10 07:11:14 - INFO - parrotllm.training - step 27890 | epoch 0 | loss 3.5289 | lr 1.35e-03 | grad 0.1852 +2026-04-10 07:11:17 - INFO - parrotllm.training - step 27900 | epoch 0 | loss 3.5830 | lr 1.35e-03 | grad 0.2492 +2026-04-10 07:11:20 - INFO - parrotllm.training - step 27910 | epoch 0 | loss 3.4966 | lr 1.35e-03 | grad 0.2195 +2026-04-10 07:11:23 - INFO - parrotllm.training - step 27920 | epoch 0 | loss 3.5784 | lr 1.35e-03 | grad 0.2342 +2026-04-10 07:11:26 - INFO - parrotllm.training - step 27930 | epoch 0 | loss 3.4579 | lr 1.35e-03 | grad 0.2006 +2026-04-10 07:11:29 - INFO - parrotllm.training - step 27940 | epoch 0 | loss 3.5555 | lr 1.35e-03 | grad 0.1991 +2026-04-10 07:11:32 - INFO - parrotllm.training - step 27950 | epoch 0 | loss 3.4769 | lr 1.35e-03 | grad 0.2245 +2026-04-10 07:11:35 - INFO - parrotllm.training - step 27960 | epoch 0 | loss 3.4539 | lr 1.35e-03 | grad 0.2140 +2026-04-10 07:11:38 - INFO - parrotllm.training - step 27970 | epoch 0 | loss 3.5702 | lr 1.35e-03 | grad 0.2044 +2026-04-10 07:11:41 - INFO - parrotllm.training - step 27980 | epoch 0 | loss 3.4715 | lr 1.35e-03 | grad 0.2056 +2026-04-10 07:11:44 - INFO - parrotllm.training - step 27990 | epoch 0 | loss 3.5367 | lr 1.35e-03 | grad 0.2083 +2026-04-10 07:11:47 - INFO - parrotllm.training - step 28000 | epoch 0 | loss 3.4434 | lr 1.35e-03 | grad 0.1929 +2026-04-10 07:11:47 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:11:47 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:11:50 - INFO - parrotllm.training - Train: loss=3.4434, ppl=31.29 +2026-04-10 07:11:50 - INFO - parrotllm.training - Val: loss=3.4055, ppl=30.13 +2026-04-10 07:11:50 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:11:51 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4055_epoch_0000_step_0028000.pt +2026-04-10 07:11:53 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:11:56 - INFO - parrotllm.training - step 28010 | epoch 0 | loss 3.4726 | lr 1.35e-03 | grad 0.2327 +2026-04-10 07:11:59 - INFO - parrotllm.training - step 28020 | epoch 0 | loss 3.4818 | lr 1.35e-03 | grad 0.2631 +2026-04-10 07:12:02 - INFO - parrotllm.training - step 28030 | epoch 0 | loss 3.5566 | lr 1.35e-03 | grad 0.2075 +2026-04-10 07:12:05 - INFO - parrotllm.training - step 28040 | epoch 0 | loss 3.5194 | lr 1.34e-03 | grad 0.2214 +2026-04-10 07:12:08 - INFO - parrotllm.training - step 28050 | epoch 0 | loss 3.4294 | lr 1.34e-03 | grad 0.1883 +2026-04-10 07:12:11 - INFO - parrotllm.training - step 28060 | epoch 0 | loss 3.4484 | lr 1.34e-03 | grad 0.2410 +2026-04-10 07:12:14 - INFO - parrotllm.training - step 28070 | epoch 0 | loss 3.3015 | lr 1.34e-03 | grad 0.2350 +2026-04-10 07:12:17 - INFO - parrotllm.training - step 28080 | epoch 0 | loss 3.6184 | lr 1.34e-03 | grad 0.2039 +2026-04-10 07:12:20 - INFO - parrotllm.training - step 28090 | epoch 0 | loss 3.4016 | lr 1.34e-03 | grad 0.2237 +2026-04-10 07:12:23 - INFO - parrotllm.training - step 28100 | epoch 0 | loss 3.5207 | lr 1.34e-03 | grad 0.2097 +2026-04-10 07:12:26 - INFO - parrotllm.training - step 28110 | epoch 0 | loss 3.4918 | lr 1.34e-03 | grad 0.2277 +2026-04-10 07:12:29 - INFO - parrotllm.training - step 28120 | epoch 0 | loss 3.4471 | lr 1.34e-03 | grad 0.2083 +2026-04-10 07:12:32 - INFO - parrotllm.training - step 28130 | epoch 0 | loss 3.4799 | lr 1.34e-03 | grad 0.2459 +2026-04-10 07:12:35 - INFO - parrotllm.training - step 28140 | epoch 0 | loss 3.4673 | lr 1.34e-03 | grad 0.2233 +2026-04-10 07:12:38 - INFO - parrotllm.training - step 28150 | epoch 0 | loss 3.3749 | lr 1.34e-03 | grad 0.2127 +2026-04-10 07:12:41 - INFO - parrotllm.training - step 28160 | epoch 0 | loss 3.5541 | lr 1.34e-03 | grad 0.1994 +2026-04-10 07:12:44 - INFO - parrotllm.training - step 28170 | epoch 0 | loss 3.4179 | lr 1.34e-03 | grad 0.2231 +2026-04-10 07:12:47 - INFO - parrotllm.training - step 28180 | epoch 0 | loss 3.4574 | lr 1.34e-03 | grad 0.2041 +2026-04-10 07:12:51 - INFO - parrotllm.training - step 28190 | epoch 0 | loss 3.4951 | lr 1.34e-03 | grad 0.2193 +2026-04-10 07:12:54 - INFO - parrotllm.training - step 28200 | epoch 0 | loss 3.3492 | lr 1.34e-03 | grad 0.2185 +2026-04-10 07:12:57 - INFO - parrotllm.training - step 28210 | epoch 0 | loss 3.5213 | lr 1.34e-03 | grad 0.2375 +2026-04-10 07:13:00 - INFO - parrotllm.training - step 28220 | epoch 0 | loss 3.4921 | lr 1.34e-03 | grad 0.2131 +2026-04-10 07:13:03 - INFO - parrotllm.training - step 28230 | epoch 0 | loss 3.4001 | lr 1.34e-03 | grad 0.2078 +2026-04-10 07:13:06 - INFO - parrotllm.training - step 28240 | epoch 0 | loss 3.4616 | lr 1.34e-03 | grad 0.2245 +2026-04-10 07:13:09 - INFO - parrotllm.training - step 28250 | epoch 0 | loss 3.4870 | lr 1.34e-03 | grad 0.1957 +2026-04-10 07:13:12 - INFO - parrotllm.training - step 28260 | epoch 0 | loss 3.4173 | lr 1.34e-03 | grad 0.1979 +2026-04-10 07:13:15 - INFO - parrotllm.training - step 28270 | epoch 0 | loss 3.4365 | lr 1.34e-03 | grad 0.2370 +2026-04-10 07:13:18 - INFO - parrotllm.training - step 28280 | epoch 0 | loss 3.5046 | lr 1.34e-03 | grad 0.1899 +2026-04-10 07:13:21 - INFO - parrotllm.training - step 28290 | epoch 0 | loss 3.4467 | lr 1.34e-03 | grad 0.1943 +2026-04-10 07:13:24 - INFO - parrotllm.training - step 28300 | epoch 0 | loss 3.5398 | lr 1.34e-03 | grad 0.1967 +2026-04-10 07:13:27 - INFO - parrotllm.training - step 28310 | epoch 0 | loss 3.3618 | lr 1.34e-03 | grad 0.2408 +2026-04-10 07:13:30 - INFO - parrotllm.training - step 28320 | epoch 0 | loss 3.4630 | lr 1.34e-03 | grad 0.1997 +2026-04-10 07:13:33 - INFO - parrotllm.training - step 28330 | epoch 0 | loss 3.5113 | lr 1.34e-03 | grad 0.2549 +2026-04-10 07:13:36 - INFO - parrotllm.training - step 28340 | epoch 0 | loss 3.3326 | lr 1.34e-03 | grad 0.1998 +2026-04-10 07:13:39 - INFO - parrotllm.training - step 28350 | epoch 0 | loss 3.4414 | lr 1.34e-03 | grad 0.2221 +2026-04-10 07:13:42 - INFO - parrotllm.training - step 28360 | epoch 0 | loss 3.4497 | lr 1.34e-03 | grad 0.2138 +2026-04-10 07:13:46 - INFO - parrotllm.training - step 28370 | epoch 0 | loss 3.5369 | lr 1.34e-03 | grad 0.2324 +2026-04-10 07:13:49 - INFO - parrotllm.training - step 28380 | epoch 0 | loss 3.4380 | lr 1.34e-03 | grad 0.2494 +2026-04-10 07:13:52 - INFO - parrotllm.training - step 28390 | epoch 0 | loss 3.4527 | lr 1.34e-03 | grad 0.1878 +2026-04-10 07:13:55 - INFO - parrotllm.training - step 28400 | epoch 0 | loss 3.5520 | lr 1.34e-03 | grad 0.2198 +2026-04-10 07:13:58 - INFO - parrotllm.training - step 28410 | epoch 0 | loss 3.4300 | lr 1.34e-03 | grad 0.2079 +2026-04-10 07:14:01 - INFO - parrotllm.training - step 28420 | epoch 0 | loss 3.5276 | lr 1.34e-03 | grad 0.2155 +2026-04-10 07:14:04 - INFO - parrotllm.training - step 28430 | epoch 0 | loss 3.5490 | lr 1.34e-03 | grad 0.2013 +2026-04-10 07:14:07 - INFO - parrotllm.training - step 28440 | epoch 0 | loss 3.4610 | lr 1.34e-03 | grad 0.2093 +2026-04-10 07:14:10 - INFO - parrotllm.training - step 28450 | epoch 0 | loss 3.4328 | lr 1.34e-03 | grad 0.2045 +2026-04-10 07:14:13 - INFO - parrotllm.training - step 28460 | epoch 0 | loss 3.5221 | lr 1.34e-03 | grad 0.1948 +2026-04-10 07:14:16 - INFO - parrotllm.training - step 28470 | epoch 0 | loss 3.4998 | lr 1.34e-03 | grad 0.2049 +2026-04-10 07:14:18 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 28485/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 07:14:20 - INFO - parrotllm.training - step 28480 | epoch 0 | loss 3.5038 | lr 1.34e-03 | grad 0.2087 +2026-04-10 07:14:23 - INFO - parrotllm.training - step 28490 | epoch 0 | loss 3.4020 | lr 1.34e-03 | grad 0.2296 +2026-04-10 07:14:26 - INFO - parrotllm.training - step 28500 | epoch 0 | loss 3.5186 | lr 1.34e-03 | grad 0.2051 +2026-04-10 07:14:26 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:14:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:14:29 - INFO - parrotllm.training - Train: loss=3.5186, ppl=33.74 +2026-04-10 07:14:29 - INFO - parrotllm.training - Val: loss=3.4071, ppl=30.18 +2026-04-10 07:14:30 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4071_epoch_0000_step_0028500.pt +2026-04-10 07:14:31 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4055, min_delta=0.001000). +2026-04-10 07:14:31 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:14:34 - INFO - parrotllm.training - step 28510 | epoch 0 | loss 3.4786 | lr 1.34e-03 | grad 0.2222 +2026-04-10 07:14:37 - INFO - parrotllm.training - step 28520 | epoch 0 | loss 3.3540 | lr 1.34e-03 | grad 0.2016 +2026-04-10 07:14:40 - INFO - parrotllm.training - step 28530 | epoch 0 | loss 3.4432 | lr 1.34e-03 | grad 0.2438 +2026-04-10 07:14:43 - INFO - parrotllm.training - step 28540 | epoch 0 | loss 3.4285 | lr 1.34e-03 | grad 0.2100 +2026-04-10 07:14:46 - INFO - parrotllm.training - step 28550 | epoch 0 | loss 3.5123 | lr 1.34e-03 | grad 0.2151 +2026-04-10 07:14:49 - INFO - parrotllm.training - step 28560 | epoch 0 | loss 3.5133 | lr 1.34e-03 | grad 0.2143 +2026-04-10 07:14:52 - INFO - parrotllm.training - step 28570 | epoch 0 | loss 3.4403 | lr 1.34e-03 | grad 0.2036 +2026-04-10 07:14:55 - INFO - parrotllm.training - step 28580 | epoch 0 | loss 3.3959 | lr 1.34e-03 | grad 0.1813 +2026-04-10 07:14:58 - INFO - parrotllm.training - step 28590 | epoch 0 | loss 3.4869 | lr 1.34e-03 | grad 0.2093 +2026-04-10 07:15:01 - INFO - parrotllm.training - step 28600 | epoch 0 | loss 3.4531 | lr 1.34e-03 | grad 0.1953 +2026-04-10 07:15:04 - INFO - parrotllm.training - step 28610 | epoch 0 | loss 3.5015 | lr 1.34e-03 | grad 0.2183 +2026-04-10 07:15:07 - INFO - parrotllm.training - step 28620 | epoch 0 | loss 3.4506 | lr 1.34e-03 | grad 0.1959 +2026-04-10 07:15:10 - INFO - parrotllm.training - step 28630 | epoch 0 | loss 3.3863 | lr 1.34e-03 | grad 0.2212 +2026-04-10 07:15:13 - INFO - parrotllm.training - step 28640 | epoch 0 | loss 3.4326 | lr 1.34e-03 | grad 0.2072 +2026-04-10 07:15:17 - INFO - parrotllm.training - step 28650 | epoch 0 | loss 3.4521 | lr 1.34e-03 | grad 0.1953 +2026-04-10 07:15:20 - INFO - parrotllm.training - step 28660 | epoch 0 | loss 3.4817 | lr 1.34e-03 | grad 0.1947 +2026-04-10 07:15:23 - INFO - parrotllm.training - step 28670 | epoch 0 | loss 3.4913 | lr 1.34e-03 | grad 0.2109 +2026-04-10 07:15:26 - INFO - parrotllm.training - step 28680 | epoch 0 | loss 3.4759 | lr 1.34e-03 | grad 0.1992 +2026-04-10 07:15:29 - INFO - parrotllm.training - step 28690 | epoch 0 | loss 3.4322 | lr 1.34e-03 | grad 0.2216 +2026-04-10 07:15:32 - INFO - parrotllm.training - step 28700 | epoch 0 | loss 3.4479 | lr 1.34e-03 | grad 0.2195 +2026-04-10 07:15:35 - INFO - parrotllm.training - step 28710 | epoch 0 | loss 3.4763 | lr 1.34e-03 | grad 0.1896 +2026-04-10 07:15:38 - INFO - parrotllm.training - step 28720 | epoch 0 | loss 3.4044 | lr 1.34e-03 | grad 0.2009 +2026-04-10 07:15:41 - INFO - parrotllm.training - step 28730 | epoch 0 | loss 3.4566 | lr 1.34e-03 | grad 0.2037 +2026-04-10 07:15:44 - INFO - parrotllm.training - step 28740 | epoch 0 | loss 3.4527 | lr 1.34e-03 | grad 0.2084 +2026-04-10 07:15:47 - INFO - parrotllm.training - step 28750 | epoch 0 | loss 3.4803 | lr 1.34e-03 | grad 0.2627 +2026-04-10 07:15:50 - INFO - parrotllm.training - step 28760 | epoch 0 | loss 3.5364 | lr 1.34e-03 | grad 0.1933 +2026-04-10 07:15:53 - INFO - parrotllm.training - step 28770 | epoch 0 | loss 3.4590 | lr 1.34e-03 | grad 0.2541 +2026-04-10 07:15:56 - INFO - parrotllm.training - step 28780 | epoch 0 | loss 3.4732 | lr 1.34e-03 | grad 0.2222 +2026-04-10 07:15:59 - INFO - parrotllm.training - step 28790 | epoch 0 | loss 3.4715 | lr 1.34e-03 | grad 0.2409 +2026-04-10 07:16:02 - INFO - parrotllm.training - step 28800 | epoch 0 | loss 3.5164 | lr 1.34e-03 | grad 0.2532 +2026-04-10 07:16:05 - INFO - parrotllm.training - step 28810 | epoch 0 | loss 3.4770 | lr 1.34e-03 | grad 0.2071 +2026-04-10 07:16:08 - INFO - parrotllm.training - step 28820 | epoch 0 | loss 3.3855 | lr 1.34e-03 | grad 0.2391 +2026-04-10 07:16:12 - INFO - parrotllm.training - step 28830 | epoch 0 | loss 3.4959 | lr 1.34e-03 | grad 0.2153 +2026-04-10 07:16:15 - INFO - parrotllm.training - step 28840 | epoch 0 | loss 3.4223 | lr 1.34e-03 | grad 0.2394 +2026-04-10 07:16:18 - INFO - parrotllm.training - step 28850 | epoch 0 | loss 3.5150 | lr 1.34e-03 | grad 0.2239 +2026-04-10 07:16:21 - INFO - parrotllm.training - step 28860 | epoch 0 | loss 3.3898 | lr 1.34e-03 | grad 0.2118 +2026-04-10 07:16:24 - INFO - parrotllm.training - step 28870 | epoch 0 | loss 3.5615 | lr 1.34e-03 | grad 0.2042 +2026-04-10 07:16:27 - INFO - parrotllm.training - step 28880 | epoch 0 | loss 3.4727 | lr 1.34e-03 | grad 0.1967 +2026-04-10 07:16:30 - INFO - parrotllm.training - step 28890 | epoch 0 | loss 3.3680 | lr 1.34e-03 | grad 0.2091 +2026-04-10 07:16:33 - INFO - parrotllm.training - step 28900 | epoch 0 | loss 3.5040 | lr 1.34e-03 | grad 0.2088 +2026-04-10 07:16:36 - INFO - parrotllm.training - step 28910 | epoch 0 | loss 3.3765 | lr 1.34e-03 | grad 0.2045 +2026-04-10 07:16:39 - INFO - parrotllm.training - step 28920 | epoch 0 | loss 3.3894 | lr 1.33e-03 | grad 0.2055 +2026-04-10 07:16:42 - INFO - parrotllm.training - step 28930 | epoch 0 | loss 3.5323 | lr 1.33e-03 | grad 0.2008 +2026-04-10 07:16:45 - INFO - parrotllm.training - step 28940 | epoch 0 | loss 3.4247 | lr 1.33e-03 | grad 0.2127 +2026-04-10 07:16:48 - INFO - parrotllm.training - step 28950 | epoch 0 | loss 3.4214 | lr 1.33e-03 | grad 0.2254 +2026-04-10 07:16:51 - INFO - parrotllm.training - step 28960 | epoch 0 | loss 3.5481 | lr 1.33e-03 | grad 0.2760 +2026-04-10 07:16:54 - INFO - parrotllm.training - step 28970 | epoch 0 | loss 3.3798 | lr 1.33e-03 | grad 0.2262 +2026-04-10 07:16:57 - INFO - parrotllm.training - step 28980 | epoch 0 | loss 3.4635 | lr 1.33e-03 | grad 0.2342 +2026-04-10 07:17:00 - INFO - parrotllm.training - step 28990 | epoch 0 | loss 3.4156 | lr 1.33e-03 | grad 0.2002 +2026-04-10 07:17:03 - INFO - parrotllm.training - step 29000 | epoch 0 | loss 3.5688 | lr 1.33e-03 | grad 0.2230 +2026-04-10 07:17:03 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:17:03 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:17:07 - INFO - parrotllm.training - Train: loss=3.5688, ppl=35.47 +2026-04-10 07:17:07 - INFO - parrotllm.training - Val: loss=3.4041, ppl=30.09 +2026-04-10 07:17:07 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:17:08 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4041_epoch_0000_step_0029000.pt +2026-04-10 07:17:09 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:17:12 - INFO - parrotllm.training - step 29010 | epoch 0 | loss 3.4149 | lr 1.33e-03 | grad 0.1974 +2026-04-10 07:17:15 - INFO - parrotllm.training - step 29020 | epoch 0 | loss 3.3725 | lr 1.33e-03 | grad 0.2140 +2026-04-10 07:17:18 - INFO - parrotllm.training - step 29030 | epoch 0 | loss 3.4055 | lr 1.33e-03 | grad 0.1976 +2026-04-10 07:17:21 - INFO - parrotllm.training - step 29040 | epoch 0 | loss 3.4743 | lr 1.33e-03 | grad 0.2094 +2026-04-10 07:17:24 - INFO - parrotllm.training - step 29050 | epoch 0 | loss 3.6020 | lr 1.33e-03 | grad 0.1994 +2026-04-10 07:17:27 - INFO - parrotllm.training - step 29060 | epoch 0 | loss 3.4313 | lr 1.33e-03 | grad 0.2239 +2026-04-10 07:17:30 - INFO - parrotllm.training - step 29070 | epoch 0 | loss 3.4926 | lr 1.33e-03 | grad 0.2494 +2026-04-10 07:17:33 - INFO - parrotllm.training - step 29080 | epoch 0 | loss 3.4803 | lr 1.33e-03 | grad 0.1927 +2026-04-10 07:17:36 - INFO - parrotllm.training - step 29090 | epoch 0 | loss 3.4290 | lr 1.33e-03 | grad 0.1871 +2026-04-10 07:17:39 - INFO - parrotllm.training - step 29100 | epoch 0 | loss 3.4219 | lr 1.33e-03 | grad 0.2168 +2026-04-10 07:17:42 - INFO - parrotllm.training - step 29110 | epoch 0 | loss 3.3146 | lr 1.33e-03 | grad 0.2367 +2026-04-10 07:17:46 - INFO - parrotllm.training - step 29120 | epoch 0 | loss 3.4702 | lr 1.33e-03 | grad 0.2273 +2026-04-10 07:17:49 - INFO - parrotllm.training - step 29130 | epoch 0 | loss 3.4842 | lr 1.33e-03 | grad 0.2187 +2026-04-10 07:17:52 - INFO - parrotllm.training - step 29140 | epoch 0 | loss 3.4690 | lr 1.33e-03 | grad 0.1963 +2026-04-10 07:17:55 - INFO - parrotllm.training - step 29150 | epoch 0 | loss 3.5017 | lr 1.33e-03 | grad 0.2051 +2026-04-10 07:17:58 - INFO - parrotllm.training - step 29160 | epoch 0 | loss 3.4691 | lr 1.33e-03 | grad 0.2441 +2026-04-10 07:18:01 - INFO - parrotllm.training - step 29170 | epoch 0 | loss 3.4326 | lr 1.33e-03 | grad 0.2156 +2026-04-10 07:18:04 - INFO - parrotllm.training - step 29180 | epoch 0 | loss 3.4326 | lr 1.33e-03 | grad 0.2341 +2026-04-10 07:18:07 - INFO - parrotllm.training - step 29190 | epoch 0 | loss 3.5119 | lr 1.33e-03 | grad 0.2295 +2026-04-10 07:18:10 - INFO - parrotllm.training - step 29200 | epoch 0 | loss 3.4704 | lr 1.33e-03 | grad 0.1889 +2026-04-10 07:18:13 - INFO - parrotllm.training - step 29210 | epoch 0 | loss 3.4106 | lr 1.33e-03 | grad 0.2008 +2026-04-10 07:18:16 - INFO - parrotllm.training - step 29220 | epoch 0 | loss 3.3525 | lr 1.33e-03 | grad 0.2152 +2026-04-10 07:18:19 - INFO - parrotllm.training - step 29230 | epoch 0 | loss 3.4826 | lr 1.33e-03 | grad 0.2113 +2026-04-10 07:18:22 - INFO - parrotllm.training - step 29240 | epoch 0 | loss 3.3797 | lr 1.33e-03 | grad 0.1893 +2026-04-10 07:18:25 - INFO - parrotllm.training - step 29250 | epoch 0 | loss 3.4525 | lr 1.33e-03 | grad 0.1876 +2026-04-10 07:18:28 - INFO - parrotllm.training - step 29260 | epoch 0 | loss 3.3761 | lr 1.33e-03 | grad 0.1974 +2026-04-10 07:18:31 - INFO - parrotllm.training - step 29270 | epoch 0 | loss 3.3959 | lr 1.33e-03 | grad 0.2201 +2026-04-10 07:18:34 - INFO - parrotllm.training - step 29280 | epoch 0 | loss 3.4876 | lr 1.33e-03 | grad 0.1989 +2026-04-10 07:18:37 - INFO - parrotllm.training - step 29290 | epoch 0 | loss 3.4955 | lr 1.33e-03 | grad 0.2316 +2026-04-10 07:18:41 - INFO - parrotllm.training - step 29300 | epoch 0 | loss 3.4063 | lr 1.33e-03 | grad 0.2168 +2026-04-10 07:18:44 - INFO - parrotllm.training - step 29310 | epoch 0 | loss 3.4741 | lr 1.33e-03 | grad 0.2426 +2026-04-10 07:18:47 - INFO - parrotllm.training - step 29320 | epoch 0 | loss 3.5164 | lr 1.33e-03 | grad 0.2118 +2026-04-10 07:18:50 - INFO - parrotllm.training - step 29330 | epoch 0 | loss 3.4446 | lr 1.33e-03 | grad 0.2019 +2026-04-10 07:18:53 - INFO - parrotllm.training - step 29340 | epoch 0 | loss 3.4148 | lr 1.33e-03 | grad 0.2005 +2026-04-10 07:18:56 - INFO - parrotllm.training - step 29350 | epoch 0 | loss 3.4211 | lr 1.33e-03 | grad 0.2500 +2026-04-10 07:18:59 - INFO - parrotllm.training - step 29360 | epoch 0 | loss 3.4913 | lr 1.33e-03 | grad 0.2051 +2026-04-10 07:19:02 - INFO - parrotllm.training - step 29370 | epoch 0 | loss 3.5286 | lr 1.33e-03 | grad 0.2078 +2026-04-10 07:19:05 - INFO - parrotllm.training - step 29380 | epoch 0 | loss 3.5246 | lr 1.33e-03 | grad 0.2214 +2026-04-10 07:19:08 - INFO - parrotllm.training - step 29390 | epoch 0 | loss 3.5656 | lr 1.33e-03 | grad 0.1936 +2026-04-10 07:19:11 - INFO - parrotllm.training - step 29400 | epoch 0 | loss 3.4127 | lr 1.33e-03 | grad 0.1935 +2026-04-10 07:19:14 - INFO - parrotllm.training - step 29410 | epoch 0 | loss 3.4821 | lr 1.33e-03 | grad 0.2242 +2026-04-10 07:19:17 - INFO - parrotllm.training - step 29420 | epoch 0 | loss 3.4112 | lr 1.33e-03 | grad 0.2070 +2026-04-10 07:19:20 - INFO - parrotllm.training - step 29430 | epoch 0 | loss 3.5120 | lr 1.33e-03 | grad 0.2257 +2026-04-10 07:19:23 - INFO - parrotllm.training - step 29440 | epoch 0 | loss 3.5187 | lr 1.33e-03 | grad 0.2205 +2026-04-10 07:19:26 - INFO - parrotllm.training - step 29450 | epoch 0 | loss 3.4576 | lr 1.33e-03 | grad 0.2158 +2026-04-10 07:19:29 - INFO - parrotllm.training - step 29460 | epoch 0 | loss 3.4690 | lr 1.33e-03 | grad 0.1847 +2026-04-10 07:19:32 - INFO - parrotllm.training - step 29470 | epoch 0 | loss 3.4960 | lr 1.33e-03 | grad 0.2074 +2026-04-10 07:19:35 - INFO - parrotllm.training - step 29480 | epoch 0 | loss 3.3911 | lr 1.33e-03 | grad 0.2196 +2026-04-10 07:19:39 - INFO - parrotllm.training - step 29490 | epoch 0 | loss 3.4206 | lr 1.33e-03 | grad 0.2224 +2026-04-10 07:19:42 - INFO - parrotllm.training - step 29500 | epoch 0 | loss 3.5171 | lr 1.33e-03 | grad 0.1917 +2026-04-10 07:19:42 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:19:42 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:19:45 - INFO - parrotllm.training - Train: loss=3.5171, ppl=33.69 +2026-04-10 07:19:45 - INFO - parrotllm.training - Val: loss=3.4042, ppl=30.09 +2026-04-10 07:19:46 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4042_epoch_0000_step_0029500.pt +2026-04-10 07:19:47 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.4041, min_delta=0.001000). +2026-04-10 07:19:47 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:19:50 - INFO - parrotllm.training - step 29510 | epoch 0 | loss 3.4796 | lr 1.33e-03 | grad 0.2223 +2026-04-10 07:19:53 - INFO - parrotllm.training - step 29520 | epoch 0 | loss 3.4904 | lr 1.33e-03 | grad 0.2280 +2026-04-10 07:19:56 - INFO - parrotllm.training - step 29530 | epoch 0 | loss 3.4468 | lr 1.33e-03 | grad 0.2425 +2026-04-10 07:19:59 - INFO - parrotllm.training - step 29540 | epoch 0 | loss 3.4952 | lr 1.33e-03 | grad 0.2115 +2026-04-10 07:20:02 - INFO - parrotllm.training - step 29550 | epoch 0 | loss 3.4891 | lr 1.33e-03 | grad 0.2113 +2026-04-10 07:20:05 - INFO - parrotllm.training - step 29560 | epoch 0 | loss 3.3526 | lr 1.33e-03 | grad 0.1942 +2026-04-10 07:20:08 - INFO - parrotllm.training - step 29570 | epoch 0 | loss 3.4872 | lr 1.33e-03 | grad 0.1974 +2026-04-10 07:20:11 - INFO - parrotllm.training - step 29580 | epoch 0 | loss 3.5880 | lr 1.33e-03 | grad 0.2106 +2026-04-10 07:20:14 - INFO - parrotllm.training - step 29590 | epoch 0 | loss 3.4652 | lr 1.33e-03 | grad 0.2270 +2026-04-10 07:20:17 - INFO - parrotllm.training - step 29600 | epoch 0 | loss 3.5082 | lr 1.33e-03 | grad 0.2085 +2026-04-10 07:20:20 - INFO - parrotllm.training - step 29610 | epoch 0 | loss 3.4773 | lr 1.33e-03 | grad 0.2154 +2026-04-10 07:20:23 - INFO - parrotllm.training - step 29620 | epoch 0 | loss 3.3793 | lr 1.33e-03 | grad 0.2275 +2026-04-10 07:20:26 - INFO - parrotllm.training - step 29630 | epoch 0 | loss 3.4636 | lr 1.33e-03 | grad 0.2525 +2026-04-10 07:20:30 - INFO - parrotllm.training - step 29640 | epoch 0 | loss 3.3944 | lr 1.33e-03 | grad 0.2215 +2026-04-10 07:20:33 - INFO - parrotllm.training - step 29650 | epoch 0 | loss 3.4133 | lr 1.33e-03 | grad 0.2435 +2026-04-10 07:20:36 - INFO - parrotllm.training - step 29660 | epoch 0 | loss 3.3936 | lr 1.33e-03 | grad 0.2241 +2026-04-10 07:20:39 - INFO - parrotllm.training - step 29670 | epoch 0 | loss 3.5082 | lr 1.33e-03 | grad 0.2250 +2026-04-10 07:20:42 - INFO - parrotllm.training - step 29680 | epoch 0 | loss 3.4231 | lr 1.33e-03 | grad 0.2409 +2026-04-10 07:20:45 - INFO - parrotllm.training - step 29690 | epoch 0 | loss 3.4856 | lr 1.33e-03 | grad 0.2072 +2026-04-10 07:20:48 - INFO - parrotllm.training - step 29700 | epoch 0 | loss 3.4976 | lr 1.33e-03 | grad 0.1966 +2026-04-10 07:20:51 - INFO - parrotllm.training - step 29710 | epoch 0 | loss 3.5434 | lr 1.33e-03 | grad 0.2093 +2026-04-10 07:20:54 - INFO - parrotllm.training - step 29720 | epoch 0 | loss 3.5343 | lr 1.33e-03 | grad 0.1960 +2026-04-10 07:20:57 - INFO - parrotllm.training - step 29730 | epoch 0 | loss 3.3855 | lr 1.33e-03 | grad 0.1953 +2026-04-10 07:21:00 - INFO - parrotllm.training - step 29740 | epoch 0 | loss 3.4170 | lr 1.33e-03 | grad 0.2134 +2026-04-10 07:21:03 - INFO - parrotllm.training - step 29750 | epoch 0 | loss 3.5305 | lr 1.33e-03 | grad 0.2145 +2026-04-10 07:21:06 - INFO - parrotllm.training - step 29760 | epoch 0 | loss 3.6100 | lr 1.33e-03 | grad 0.2166 +2026-04-10 07:21:09 - INFO - parrotllm.training - step 29770 | epoch 0 | loss 3.5652 | lr 1.33e-03 | grad 0.2005 +2026-04-10 07:21:12 - INFO - parrotllm.training - step 29780 | epoch 0 | loss 3.4245 | lr 1.32e-03 | grad 0.2002 +2026-04-10 07:21:15 - INFO - parrotllm.training - step 29790 | epoch 0 | loss 3.5171 | lr 1.32e-03 | grad 0.2011 +2026-04-10 07:21:18 - INFO - parrotllm.training - step 29800 | epoch 0 | loss 3.4828 | lr 1.32e-03 | grad 0.2320 +2026-04-10 07:21:22 - INFO - parrotllm.training - step 29810 | epoch 0 | loss 3.5129 | lr 1.32e-03 | grad 0.1929 +2026-04-10 07:21:25 - INFO - parrotllm.training - step 29820 | epoch 0 | loss 3.3763 | lr 1.32e-03 | grad 0.2396 +2026-04-10 07:21:28 - INFO - parrotllm.training - step 29830 | epoch 0 | loss 3.4090 | lr 1.32e-03 | grad 0.2162 +2026-04-10 07:21:31 - INFO - parrotllm.training - step 29840 | epoch 0 | loss 3.4650 | lr 1.32e-03 | grad 0.2067 +2026-04-10 07:21:34 - INFO - parrotllm.training - step 29850 | epoch 0 | loss 3.3436 | lr 1.32e-03 | grad 0.1963 +2026-04-10 07:21:37 - INFO - parrotllm.training - step 29860 | epoch 0 | loss 3.4491 | lr 1.32e-03 | grad 0.2187 +2026-04-10 07:21:40 - INFO - parrotllm.training - step 29870 | epoch 0 | loss 3.4815 | lr 1.32e-03 | grad 0.2362 +2026-04-10 07:21:43 - INFO - parrotllm.training - step 29880 | epoch 0 | loss 3.5336 | lr 1.32e-03 | grad 0.2315 +2026-04-10 07:21:46 - INFO - parrotllm.training - step 29890 | epoch 0 | loss 3.4709 | lr 1.32e-03 | grad 0.2182 +2026-04-10 07:21:49 - INFO - parrotllm.training - step 29900 | epoch 0 | loss 3.4681 | lr 1.32e-03 | grad 0.2108 +2026-04-10 07:21:52 - INFO - parrotllm.training - step 29910 | epoch 0 | loss 3.3937 | lr 1.32e-03 | grad 0.2014 +2026-04-10 07:21:55 - INFO - parrotllm.training - step 29920 | epoch 0 | loss 3.5364 | lr 1.32e-03 | grad 0.1965 +2026-04-10 07:21:58 - INFO - parrotllm.training - step 29930 | epoch 0 | loss 3.5457 | lr 1.32e-03 | grad 0.1960 +2026-04-10 07:22:01 - INFO - parrotllm.training - step 29940 | epoch 0 | loss 3.4670 | lr 1.32e-03 | grad 0.2208 +2026-04-10 07:22:04 - INFO - parrotllm.training - step 29950 | epoch 0 | loss 3.4143 | lr 1.32e-03 | grad 0.2272 +2026-04-10 07:22:07 - INFO - parrotllm.training - step 29960 | epoch 0 | loss 3.4162 | lr 1.32e-03 | grad 0.1986 +2026-04-10 07:22:10 - INFO - parrotllm.training - step 29970 | epoch 0 | loss 3.4548 | lr 1.32e-03 | grad 0.2201 +2026-04-10 07:22:13 - INFO - parrotllm.training - step 29980 | epoch 0 | loss 3.4200 | lr 1.32e-03 | grad 0.1910 +2026-04-10 07:22:17 - INFO - parrotllm.training - step 29990 | epoch 0 | loss 3.4482 | lr 1.32e-03 | grad 0.2112 +2026-04-10 07:22:20 - INFO - parrotllm.training - step 30000 | epoch 0 | loss 3.4868 | lr 1.32e-03 | grad 0.2227 +2026-04-10 07:22:20 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:22:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:22:23 - INFO - parrotllm.training - Train: loss=3.4868, ppl=32.68 +2026-04-10 07:22:23 - INFO - parrotllm.training - Val: loss=3.4030, ppl=30.05 +2026-04-10 07:22:23 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:22:24 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4030_epoch_0000_step_0030000.pt +2026-04-10 07:22:25 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:22:26 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0030000.pt +2026-04-10 07:22:30 - INFO - parrotllm.training - step 30010 | epoch 0 | loss 3.4197 | lr 1.32e-03 | grad 0.2190 +2026-04-10 07:22:33 - INFO - parrotllm.training - step 30020 | epoch 0 | loss 3.5988 | lr 1.32e-03 | grad 0.2209 +2026-04-10 07:22:36 - INFO - parrotllm.training - step 30030 | epoch 0 | loss 3.4116 | lr 1.32e-03 | grad 0.2241 +2026-04-10 07:22:39 - INFO - parrotllm.training - step 30040 | epoch 0 | loss 3.5113 | lr 1.32e-03 | grad 0.2131 +2026-04-10 07:22:42 - INFO - parrotllm.training - step 30050 | epoch 0 | loss 3.4597 | lr 1.32e-03 | grad 0.2186 +2026-04-10 07:22:45 - INFO - parrotllm.training - step 30060 | epoch 0 | loss 3.2655 | lr 1.32e-03 | grad 0.2065 +2026-04-10 07:22:49 - INFO - parrotllm.training - step 30070 | epoch 0 | loss 3.4442 | lr 1.32e-03 | grad 0.2279 +2026-04-10 07:22:52 - INFO - parrotllm.training - step 30080 | epoch 0 | loss 3.5120 | lr 1.32e-03 | grad 0.2017 +2026-04-10 07:22:55 - INFO - parrotllm.training - step 30090 | epoch 0 | loss 3.4734 | lr 1.32e-03 | grad 0.2003 +2026-04-10 07:22:58 - INFO - parrotllm.training - step 30100 | epoch 0 | loss 3.4151 | lr 1.32e-03 | grad 0.2265 +2026-04-10 07:23:01 - INFO - parrotllm.training - step 30110 | epoch 0 | loss 3.4189 | lr 1.32e-03 | grad 0.2405 +2026-04-10 07:23:04 - INFO - parrotllm.training - step 30120 | epoch 0 | loss 3.4614 | lr 1.32e-03 | grad 0.2091 +2026-04-10 07:23:07 - INFO - parrotllm.training - step 30130 | epoch 0 | loss 3.4604 | lr 1.32e-03 | grad 0.2067 +2026-04-10 07:23:10 - INFO - parrotllm.training - step 30140 | epoch 0 | loss 3.5016 | lr 1.32e-03 | grad 0.2318 +2026-04-10 07:23:13 - INFO - parrotllm.training - step 30150 | epoch 0 | loss 3.5315 | lr 1.32e-03 | grad 0.2284 +2026-04-10 07:23:16 - INFO - parrotllm.training - step 30160 | epoch 0 | loss 3.4414 | lr 1.32e-03 | grad 0.2016 +2026-04-10 07:23:19 - INFO - parrotllm.training - step 30170 | epoch 0 | loss 3.5288 | lr 1.32e-03 | grad 0.2012 +2026-04-10 07:23:22 - INFO - parrotllm.training - step 30180 | epoch 0 | loss 3.3733 | lr 1.32e-03 | grad 0.2070 +2026-04-10 07:23:25 - INFO - parrotllm.training - step 30190 | epoch 0 | loss 3.4361 | lr 1.32e-03 | grad 0.2105 +2026-04-10 07:23:28 - INFO - parrotllm.training - step 30200 | epoch 0 | loss 3.4706 | lr 1.32e-03 | grad 0.2704 +2026-04-10 07:23:31 - INFO - parrotllm.training - step 30210 | epoch 0 | loss 3.4107 | lr 1.32e-03 | grad 0.2531 +2026-04-10 07:23:34 - INFO - parrotllm.training - step 30220 | epoch 0 | loss 3.3984 | lr 1.32e-03 | grad 0.1949 +2026-04-10 07:23:37 - INFO - parrotllm.training - step 30230 | epoch 0 | loss 3.3245 | lr 1.32e-03 | grad 0.2060 +2026-04-10 07:23:40 - INFO - parrotllm.training - step 30240 | epoch 0 | loss 3.4776 | lr 1.32e-03 | grad 0.1963 +2026-04-10 07:23:44 - INFO - parrotllm.training - step 30250 | epoch 0 | loss 3.5039 | lr 1.32e-03 | grad 0.2166 +2026-04-10 07:23:47 - INFO - parrotllm.training - step 30260 | epoch 0 | loss 3.4663 | lr 1.32e-03 | grad 0.2243 +2026-04-10 07:23:50 - INFO - parrotllm.training - step 30270 | epoch 0 | loss 3.5139 | lr 1.32e-03 | grad 0.1914 +2026-04-10 07:23:53 - INFO - parrotllm.training - step 30280 | epoch 0 | loss 3.4677 | lr 1.32e-03 | grad 0.2536 +2026-04-10 07:23:56 - INFO - parrotllm.training - step 30290 | epoch 0 | loss 3.4549 | lr 1.32e-03 | grad 0.2089 +2026-04-10 07:23:59 - INFO - parrotllm.training - step 30300 | epoch 0 | loss 3.5811 | lr 1.32e-03 | grad 0.2067 +2026-04-10 07:24:02 - INFO - parrotllm.training - step 30310 | epoch 0 | loss 3.5293 | lr 1.32e-03 | grad 0.1978 +2026-04-10 07:24:05 - INFO - parrotllm.training - step 30320 | epoch 0 | loss 3.3941 | lr 1.32e-03 | grad 0.2554 +2026-04-10 07:24:08 - INFO - parrotllm.training - step 30330 | epoch 0 | loss 3.4563 | lr 1.32e-03 | grad 0.2089 +2026-04-10 07:24:11 - INFO - parrotllm.training - step 30340 | epoch 0 | loss 3.3494 | lr 1.32e-03 | grad 0.2177 +2026-04-10 07:24:14 - INFO - parrotllm.training - step 30350 | epoch 0 | loss 3.4410 | lr 1.32e-03 | grad 0.2169 +2026-04-10 07:24:17 - INFO - parrotllm.training - step 30360 | epoch 0 | loss 3.4240 | lr 1.32e-03 | grad 0.2231 +2026-04-10 07:24:20 - INFO - parrotllm.training - step 30370 | epoch 0 | loss 3.4265 | lr 1.32e-03 | grad 0.2186 +2026-04-10 07:24:23 - INFO - parrotllm.training - step 30380 | epoch 0 | loss 3.5482 | lr 1.32e-03 | grad 0.1979 +2026-04-10 07:24:26 - INFO - parrotllm.training - step 30390 | epoch 0 | loss 3.5247 | lr 1.32e-03 | grad 0.2394 +2026-04-10 07:24:29 - INFO - parrotllm.training - step 30400 | epoch 0 | loss 3.4706 | lr 1.32e-03 | grad 0.2166 +2026-04-10 07:24:32 - INFO - parrotllm.training - step 30410 | epoch 0 | loss 3.5381 | lr 1.32e-03 | grad 0.2174 +2026-04-10 07:24:36 - INFO - parrotllm.training - step 30420 | epoch 0 | loss 3.5345 | lr 1.32e-03 | grad 0.2130 +2026-04-10 07:24:39 - INFO - parrotllm.training - step 30430 | epoch 0 | loss 3.4433 | lr 1.32e-03 | grad 0.2096 +2026-04-10 07:24:42 - INFO - parrotllm.training - step 30440 | epoch 0 | loss 3.4169 | lr 1.32e-03 | grad 0.2248 +2026-04-10 07:24:45 - INFO - parrotllm.training - step 30450 | epoch 0 | loss 3.4800 | lr 1.32e-03 | grad 0.2133 +2026-04-10 07:24:48 - INFO - parrotllm.training - step 30460 | epoch 0 | loss 3.4591 | lr 1.32e-03 | grad 0.2412 +2026-04-10 07:24:51 - INFO - parrotllm.training - step 30470 | epoch 0 | loss 3.4839 | lr 1.32e-03 | grad 0.2267 +2026-04-10 07:24:54 - INFO - parrotllm.training - step 30480 | epoch 0 | loss 3.3662 | lr 1.32e-03 | grad 0.1831 +2026-04-10 07:24:57 - INFO - parrotllm.training - step 30490 | epoch 0 | loss 3.5338 | lr 1.32e-03 | grad 0.2045 +2026-04-10 07:25:00 - INFO - parrotllm.training - step 30500 | epoch 0 | loss 3.4504 | lr 1.32e-03 | grad 0.2443 +2026-04-10 07:25:00 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:25:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:25:03 - INFO - parrotllm.training - Train: loss=3.4504, ppl=31.51 +2026-04-10 07:25:03 - INFO - parrotllm.training - Val: loss=3.4020, ppl=30.02 +2026-04-10 07:25:03 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:25:04 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4020_epoch_0000_step_0030500.pt +2026-04-10 07:25:05 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:25:08 - INFO - parrotllm.training - step 30510 | epoch 0 | loss 3.4983 | lr 1.32e-03 | grad 0.2114 +2026-04-10 07:25:11 - INFO - parrotllm.training - step 30520 | epoch 0 | loss 3.5292 | lr 1.32e-03 | grad 0.2167 +2026-04-10 07:25:14 - INFO - parrotllm.training - step 30530 | epoch 0 | loss 3.3897 | lr 1.32e-03 | grad 0.2257 +2026-04-10 07:25:17 - INFO - parrotllm.training - step 30540 | epoch 0 | loss 3.5702 | lr 1.32e-03 | grad 0.2203 +2026-04-10 07:25:20 - INFO - parrotllm.training - step 30550 | epoch 0 | loss 3.6009 | lr 1.32e-03 | grad 0.2222 +2026-04-10 07:25:24 - INFO - parrotllm.training - step 30560 | epoch 0 | loss 3.4374 | lr 1.32e-03 | grad 0.2000 +2026-04-10 07:25:27 - INFO - parrotllm.training - step 30570 | epoch 0 | loss 3.4803 | lr 1.32e-03 | grad 0.1914 +2026-04-10 07:25:30 - INFO - parrotllm.training - step 30580 | epoch 0 | loss 3.3909 | lr 1.32e-03 | grad 0.2562 +2026-04-10 07:25:33 - INFO - parrotllm.training - step 30590 | epoch 0 | loss 3.3952 | lr 1.32e-03 | grad 0.2144 +2026-04-10 07:25:36 - INFO - parrotllm.training - step 30600 | epoch 0 | loss 3.5066 | lr 1.32e-03 | grad 0.2104 +2026-04-10 07:25:39 - INFO - parrotllm.training - step 30610 | epoch 0 | loss 3.5149 | lr 1.31e-03 | grad 0.2250 +2026-04-10 07:25:42 - INFO - parrotllm.training - step 30620 | epoch 0 | loss 3.4011 | lr 1.31e-03 | grad 0.2252 +2026-04-10 07:25:45 - INFO - parrotllm.training - step 30630 | epoch 0 | loss 3.4555 | lr 1.31e-03 | grad 0.1998 +2026-04-10 07:25:48 - INFO - parrotllm.training - step 30640 | epoch 0 | loss 3.4772 | lr 1.31e-03 | grad 0.2096 +2026-04-10 07:25:51 - INFO - parrotllm.training - step 30650 | epoch 0 | loss 3.3576 | lr 1.31e-03 | grad 0.2255 +2026-04-10 07:25:54 - INFO - parrotllm.training - step 30660 | epoch 0 | loss 3.4905 | lr 1.31e-03 | grad 0.2035 +2026-04-10 07:25:57 - INFO - parrotllm.training - step 30670 | epoch 0 | loss 3.4991 | lr 1.31e-03 | grad 0.2078 +2026-04-10 07:26:00 - INFO - parrotllm.training - step 30680 | epoch 0 | loss 3.4470 | lr 1.31e-03 | grad 0.2067 +2026-04-10 07:26:03 - INFO - parrotllm.training - step 30690 | epoch 0 | loss 3.4704 | lr 1.31e-03 | grad 0.2459 +2026-04-10 07:26:06 - INFO - parrotllm.training - step 30700 | epoch 0 | loss 3.4815 | lr 1.31e-03 | grad 0.2032 +2026-04-10 07:26:09 - INFO - parrotllm.training - step 30710 | epoch 0 | loss 3.4113 | lr 1.31e-03 | grad 0.1858 +2026-04-10 07:26:12 - INFO - parrotllm.training - step 30720 | epoch 0 | loss 3.3697 | lr 1.31e-03 | grad 0.2221 +2026-04-10 07:26:16 - INFO - parrotllm.training - step 30730 | epoch 0 | loss 3.5191 | lr 1.31e-03 | grad 0.2079 +2026-04-10 07:26:19 - INFO - parrotllm.training - step 30740 | epoch 0 | loss 3.4504 | lr 1.31e-03 | grad 0.2129 +2026-04-10 07:26:22 - INFO - parrotllm.training - step 30750 | epoch 0 | loss 3.5276 | lr 1.31e-03 | grad 0.2350 +2026-04-10 07:26:25 - INFO - parrotllm.training - step 30760 | epoch 0 | loss 3.4948 | lr 1.31e-03 | grad 0.1872 +2026-04-10 07:26:28 - INFO - parrotllm.training - step 30770 | epoch 0 | loss 3.4335 | lr 1.31e-03 | grad 0.2062 +2026-04-10 07:26:31 - INFO - parrotllm.training - step 30780 | epoch 0 | loss 3.4293 | lr 1.31e-03 | grad 0.2191 +2026-04-10 07:26:34 - INFO - parrotllm.training - step 30790 | epoch 0 | loss 3.5241 | lr 1.31e-03 | grad 0.2340 +2026-04-10 07:26:37 - INFO - parrotllm.training - step 30800 | epoch 0 | loss 3.4162 | lr 1.31e-03 | grad 0.2538 +2026-04-10 07:26:40 - INFO - parrotllm.training - step 30810 | epoch 0 | loss 3.4688 | lr 1.31e-03 | grad 0.2248 +2026-04-10 07:26:43 - INFO - parrotllm.training - step 30820 | epoch 0 | loss 3.4688 | lr 1.31e-03 | grad 0.2038 +2026-04-10 07:26:46 - INFO - parrotllm.training - step 30830 | epoch 0 | loss 3.4028 | lr 1.31e-03 | grad 0.2154 +2026-04-10 07:26:49 - INFO - parrotllm.training - step 30840 | epoch 0 | loss 3.4863 | lr 1.31e-03 | grad 0.2235 +2026-04-10 07:26:52 - INFO - parrotllm.training - step 30850 | epoch 0 | loss 3.4303 | lr 1.31e-03 | grad 0.2186 +2026-04-10 07:26:55 - INFO - parrotllm.training - step 30860 | epoch 0 | loss 3.4778 | lr 1.31e-03 | grad 0.2298 +2026-04-10 07:26:58 - INFO - parrotllm.training - step 30870 | epoch 0 | loss 3.4462 | lr 1.31e-03 | grad 0.1974 +2026-04-10 07:27:01 - INFO - parrotllm.training - step 30880 | epoch 0 | loss 3.3880 | lr 1.31e-03 | grad 0.1896 +2026-04-10 07:27:05 - INFO - parrotllm.training - step 30890 | epoch 0 | loss 3.4855 | lr 1.31e-03 | grad 0.2005 +2026-04-10 07:27:08 - INFO - parrotllm.training - step 30900 | epoch 0 | loss 3.4898 | lr 1.31e-03 | grad 0.2105 +2026-04-10 07:27:11 - INFO - parrotllm.training - step 30910 | epoch 0 | loss 3.5634 | lr 1.31e-03 | grad 0.2091 +2026-04-10 07:27:14 - INFO - parrotllm.training - step 30920 | epoch 0 | loss 3.4763 | lr 1.31e-03 | grad 0.2421 +2026-04-10 07:27:17 - INFO - parrotllm.training - step 30930 | epoch 0 | loss 3.4263 | lr 1.31e-03 | grad 0.2373 +2026-04-10 07:27:20 - INFO - parrotllm.training - step 30940 | epoch 0 | loss 3.4413 | lr 1.31e-03 | grad 0.2105 +2026-04-10 07:27:23 - INFO - parrotllm.training - step 30950 | epoch 0 | loss 3.4256 | lr 1.31e-03 | grad 0.2529 +2026-04-10 07:27:26 - INFO - parrotllm.training - step 30960 | epoch 0 | loss 3.3999 | lr 1.31e-03 | grad 0.2321 +2026-04-10 07:27:29 - INFO - parrotllm.training - step 30970 | epoch 0 | loss 3.4617 | lr 1.31e-03 | grad 0.2143 +2026-04-10 07:27:32 - INFO - parrotllm.training - step 30980 | epoch 0 | loss 3.5273 | lr 1.31e-03 | grad 0.2150 +2026-04-10 07:27:34 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 30997/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 07:27:35 - INFO - parrotllm.training - step 30990 | epoch 0 | loss 3.3943 | lr 1.31e-03 | grad 0.2206 +2026-04-10 07:27:38 - INFO - parrotllm.training - step 31000 | epoch 0 | loss 3.5173 | lr 1.31e-03 | grad 0.2053 +2026-04-10 07:27:38 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:27:38 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:27:42 - INFO - parrotllm.training - Train: loss=3.5173, ppl=33.69 +2026-04-10 07:27:42 - INFO - parrotllm.training - Val: loss=3.4003, ppl=29.97 +2026-04-10 07:27:42 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:27:42 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p4003_epoch_0000_step_0031000.pt +2026-04-10 07:27:44 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:27:47 - INFO - parrotllm.training - step 31010 | epoch 0 | loss 3.4460 | lr 1.31e-03 | grad 0.1895 +2026-04-10 07:27:50 - INFO - parrotllm.training - step 31020 | epoch 0 | loss 3.4758 | lr 1.31e-03 | grad 0.2083 +2026-04-10 07:27:53 - INFO - parrotllm.training - step 31030 | epoch 0 | loss 3.3989 | lr 1.31e-03 | grad 0.2229 +2026-04-10 07:27:56 - INFO - parrotllm.training - step 31040 | epoch 0 | loss 3.4160 | lr 1.31e-03 | grad 0.2276 +2026-04-10 07:27:59 - INFO - parrotllm.training - step 31050 | epoch 0 | loss 3.5495 | lr 1.31e-03 | grad 0.2181 +2026-04-10 07:28:02 - INFO - parrotllm.training - step 31060 | epoch 0 | loss 3.4584 | lr 1.31e-03 | grad 0.2462 +2026-04-10 07:28:05 - INFO - parrotllm.training - step 31070 | epoch 0 | loss 3.4739 | lr 1.31e-03 | grad 0.2257 +2026-04-10 07:28:08 - INFO - parrotllm.training - step 31080 | epoch 0 | loss 3.4614 | lr 1.31e-03 | grad 0.2070 +2026-04-10 07:28:11 - INFO - parrotllm.training - step 31090 | epoch 0 | loss 3.5172 | lr 1.31e-03 | grad 0.2233 +2026-04-10 07:28:14 - INFO - parrotllm.training - step 31100 | epoch 0 | loss 3.3283 | lr 1.31e-03 | grad 0.2280 +2026-04-10 07:28:17 - INFO - parrotllm.training - step 31110 | epoch 0 | loss 3.4858 | lr 1.31e-03 | grad 0.2583 +2026-04-10 07:28:20 - INFO - parrotllm.training - step 31120 | epoch 0 | loss 3.4159 | lr 1.31e-03 | grad 0.1917 +2026-04-10 07:28:24 - INFO - parrotllm.training - step 31130 | epoch 0 | loss 3.4350 | lr 1.31e-03 | grad 0.1958 +2026-04-10 07:28:27 - INFO - parrotllm.training - step 31140 | epoch 0 | loss 3.5779 | lr 1.31e-03 | grad 0.1989 +2026-04-10 07:28:30 - INFO - parrotllm.training - step 31150 | epoch 0 | loss 3.3859 | lr 1.31e-03 | grad 0.2189 +2026-04-10 07:28:33 - INFO - parrotllm.training - step 31160 | epoch 0 | loss 3.4840 | lr 1.31e-03 | grad 0.2337 +2026-04-10 07:28:36 - INFO - parrotllm.training - step 31170 | epoch 0 | loss 3.4750 | lr 1.31e-03 | grad 0.2064 +2026-04-10 07:28:39 - INFO - parrotllm.training - step 31180 | epoch 0 | loss 3.3817 | lr 1.31e-03 | grad 0.2240 +2026-04-10 07:28:42 - INFO - parrotllm.training - step 31190 | epoch 0 | loss 3.5652 | lr 1.31e-03 | grad 0.2084 +2026-04-10 07:28:45 - INFO - parrotllm.training - step 31200 | epoch 0 | loss 3.4426 | lr 1.31e-03 | grad 0.2080 +2026-04-10 07:28:48 - INFO - parrotllm.training - step 31210 | epoch 0 | loss 3.4965 | lr 1.31e-03 | grad 0.2044 +2026-04-10 07:28:51 - INFO - parrotllm.training - step 31220 | epoch 0 | loss 3.5050 | lr 1.31e-03 | grad 0.2252 +2026-04-10 07:28:54 - INFO - parrotllm.training - step 31230 | epoch 0 | loss 3.4071 | lr 1.31e-03 | grad 0.2377 +2026-04-10 07:28:57 - INFO - parrotllm.training - step 31240 | epoch 0 | loss 3.4884 | lr 1.31e-03 | grad 0.2021 +2026-04-10 07:29:00 - INFO - parrotllm.training - step 31250 | epoch 0 | loss 3.4437 | lr 1.31e-03 | grad 0.2491 +2026-04-10 07:29:03 - INFO - parrotllm.training - step 31260 | epoch 0 | loss 3.5115 | lr 1.31e-03 | grad 0.2100 +2026-04-10 07:29:06 - INFO - parrotllm.training - step 31270 | epoch 0 | loss 3.4160 | lr 1.31e-03 | grad 0.2518 +2026-04-10 07:29:09 - INFO - parrotllm.training - step 31280 | epoch 0 | loss 3.4459 | lr 1.31e-03 | grad 0.2116 +2026-04-10 07:29:12 - INFO - parrotllm.training - step 31290 | epoch 0 | loss 3.5110 | lr 1.31e-03 | grad 0.2302 +2026-04-10 07:29:15 - INFO - parrotllm.training - step 31300 | epoch 0 | loss 3.4198 | lr 1.31e-03 | grad 0.2176 +2026-04-10 07:29:18 - INFO - parrotllm.training - step 31310 | epoch 0 | loss 3.4398 | lr 1.31e-03 | grad 0.2002 +2026-04-10 07:29:22 - INFO - parrotllm.training - step 31320 | epoch 0 | loss 3.5271 | lr 1.31e-03 | grad 0.1886 +2026-04-10 07:29:25 - INFO - parrotllm.training - step 31330 | epoch 0 | loss 3.4780 | lr 1.31e-03 | grad 0.2280 +2026-04-10 07:29:28 - INFO - parrotllm.training - step 31340 | epoch 0 | loss 3.5540 | lr 1.31e-03 | grad 0.2187 +2026-04-10 07:29:31 - INFO - parrotllm.training - step 31350 | epoch 0 | loss 3.4647 | lr 1.31e-03 | grad 0.2181 +2026-04-10 07:29:34 - INFO - parrotllm.training - step 31360 | epoch 0 | loss 3.4658 | lr 1.31e-03 | grad 0.2208 +2026-04-10 07:29:37 - INFO - parrotllm.training - step 31370 | epoch 0 | loss 3.5284 | lr 1.31e-03 | grad 0.2104 +2026-04-10 07:29:40 - INFO - parrotllm.training - step 31380 | epoch 0 | loss 3.4019 | lr 1.31e-03 | grad 0.2123 +2026-04-10 07:29:43 - INFO - parrotllm.training - step 31390 | epoch 0 | loss 3.4835 | lr 1.31e-03 | grad 0.2271 +2026-04-10 07:29:46 - INFO - parrotllm.training - step 31400 | epoch 0 | loss 3.5408 | lr 1.31e-03 | grad 0.2024 +2026-04-10 07:29:49 - INFO - parrotllm.training - step 31410 | epoch 0 | loss 3.4368 | lr 1.31e-03 | grad 0.2000 +2026-04-10 07:29:52 - INFO - parrotllm.training - step 31420 | epoch 0 | loss 3.4965 | lr 1.31e-03 | grad 0.1983 +2026-04-10 07:29:55 - INFO - parrotllm.training - step 31430 | epoch 0 | loss 3.4379 | lr 1.30e-03 | grad 0.2082 +2026-04-10 07:29:58 - INFO - parrotllm.training - step 31440 | epoch 0 | loss 3.4255 | lr 1.30e-03 | grad 0.2171 +2026-04-10 07:30:01 - INFO - parrotllm.training - step 31450 | epoch 0 | loss 3.5168 | lr 1.30e-03 | grad 0.2106 +2026-04-10 07:30:04 - INFO - parrotllm.training - step 31460 | epoch 0 | loss 3.5119 | lr 1.30e-03 | grad 0.2017 +2026-04-10 07:30:07 - INFO - parrotllm.training - step 31470 | epoch 0 | loss 3.5001 | lr 1.30e-03 | grad 0.2064 +2026-04-10 07:30:10 - INFO - parrotllm.training - step 31480 | epoch 0 | loss 3.4457 | lr 1.30e-03 | grad 0.2349 +2026-04-10 07:30:13 - INFO - parrotllm.training - step 31490 | epoch 0 | loss 3.5392 | lr 1.30e-03 | grad 0.2265 +2026-04-10 07:30:16 - INFO - parrotllm.training - step 31500 | epoch 0 | loss 3.3699 | lr 1.30e-03 | grad 0.2203 +2026-04-10 07:30:16 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:30:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:30:20 - INFO - parrotllm.training - Train: loss=3.3699, ppl=29.08 +2026-04-10 07:30:20 - INFO - parrotllm.training - Val: loss=3.3975, ppl=29.89 +2026-04-10 07:30:20 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:30:20 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3975_epoch_0000_step_0031500.pt +2026-04-10 07:30:22 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:30:25 - INFO - parrotllm.training - step 31510 | epoch 0 | loss 3.4842 | lr 1.30e-03 | grad 0.2153 +2026-04-10 07:30:28 - INFO - parrotllm.training - step 31520 | epoch 0 | loss 3.4557 | lr 1.30e-03 | grad 0.2164 +2026-04-10 07:30:31 - INFO - parrotllm.training - step 31530 | epoch 0 | loss 3.4548 | lr 1.30e-03 | grad 0.2168 +2026-04-10 07:30:34 - INFO - parrotllm.training - step 31540 | epoch 0 | loss 3.4455 | lr 1.30e-03 | grad 0.2169 +2026-04-10 07:30:37 - INFO - parrotllm.training - step 31550 | epoch 0 | loss 3.3672 | lr 1.30e-03 | grad 0.2170 +2026-04-10 07:30:40 - INFO - parrotllm.training - step 31560 | epoch 0 | loss 3.4651 | lr 1.30e-03 | grad 0.2133 +2026-04-10 07:30:43 - INFO - parrotllm.training - step 31570 | epoch 0 | loss 3.4903 | lr 1.30e-03 | grad 0.1840 +2026-04-10 07:30:46 - INFO - parrotllm.training - step 31580 | epoch 0 | loss 3.5410 | lr 1.30e-03 | grad 0.2086 +2026-04-10 07:30:49 - INFO - parrotllm.training - step 31590 | epoch 0 | loss 3.4755 | lr 1.30e-03 | grad 0.2013 +2026-04-10 07:30:52 - INFO - parrotllm.training - step 31600 | epoch 0 | loss 3.4629 | lr 1.30e-03 | grad 0.2118 +2026-04-10 07:30:55 - INFO - parrotllm.training - step 31610 | epoch 0 | loss 3.4524 | lr 1.30e-03 | grad 0.2444 +2026-04-10 07:30:58 - INFO - parrotllm.training - step 31620 | epoch 0 | loss 3.4988 | lr 1.30e-03 | grad 0.2002 +2026-04-10 07:31:01 - INFO - parrotllm.training - step 31630 | epoch 0 | loss 3.4919 | lr 1.30e-03 | grad 0.2002 +2026-04-10 07:31:04 - INFO - parrotllm.training - step 31640 | epoch 0 | loss 3.3729 | lr 1.30e-03 | grad 0.2049 +2026-04-10 07:31:07 - INFO - parrotllm.training - step 31650 | epoch 0 | loss 3.4139 | lr 1.30e-03 | grad 0.2425 +2026-04-10 07:31:11 - INFO - parrotllm.training - step 31660 | epoch 0 | loss 3.4009 | lr 1.30e-03 | grad 0.2416 +2026-04-10 07:31:14 - INFO - parrotllm.training - step 31670 | epoch 0 | loss 3.4112 | lr 1.30e-03 | grad 0.2103 +2026-04-10 07:31:17 - INFO - parrotllm.training - step 31680 | epoch 0 | loss 3.5506 | lr 1.30e-03 | grad 0.2113 +2026-04-10 07:31:20 - INFO - parrotllm.training - step 31690 | epoch 0 | loss 3.5551 | lr 1.30e-03 | grad 0.2536 +2026-04-10 07:31:23 - INFO - parrotllm.training - step 31700 | epoch 0 | loss 3.4557 | lr 1.30e-03 | grad 0.1996 +2026-04-10 07:31:26 - INFO - parrotllm.training - step 31710 | epoch 0 | loss 3.4044 | lr 1.30e-03 | grad 0.2052 +2026-04-10 07:31:29 - INFO - parrotllm.training - step 31720 | epoch 0 | loss 3.4023 | lr 1.30e-03 | grad 0.2025 +2026-04-10 07:31:32 - INFO - parrotllm.training - step 31730 | epoch 0 | loss 3.4802 | lr 1.30e-03 | grad 0.2300 +2026-04-10 07:31:35 - INFO - parrotllm.training - step 31740 | epoch 0 | loss 3.4456 | lr 1.30e-03 | grad 0.1942 +2026-04-10 07:31:38 - INFO - parrotllm.training - step 31750 | epoch 0 | loss 3.5499 | lr 1.30e-03 | grad 0.2093 +2026-04-10 07:31:41 - INFO - parrotllm.training - step 31760 | epoch 0 | loss 3.6110 | lr 1.30e-03 | grad 0.2066 +2026-04-10 07:31:44 - INFO - parrotllm.training - step 31770 | epoch 0 | loss 3.4499 | lr 1.30e-03 | grad 0.2173 +2026-04-10 07:31:47 - INFO - parrotllm.training - step 31780 | epoch 0 | loss 3.4119 | lr 1.30e-03 | grad 0.2061 +2026-04-10 07:31:50 - INFO - parrotllm.training - step 31790 | epoch 0 | loss 3.4776 | lr 1.30e-03 | grad 0.2103 +2026-04-10 07:31:53 - INFO - parrotllm.training - step 31800 | epoch 0 | loss 3.3777 | lr 1.30e-03 | grad 0.2357 +2026-04-10 07:31:56 - INFO - parrotllm.training - step 31810 | epoch 0 | loss 3.5826 | lr 1.30e-03 | grad 0.2084 +2026-04-10 07:31:59 - INFO - parrotllm.training - step 31820 | epoch 0 | loss 3.4778 | lr 1.30e-03 | grad 0.2277 +2026-04-10 07:32:02 - INFO - parrotllm.training - step 31830 | epoch 0 | loss 3.4361 | lr 1.30e-03 | grad 0.2195 +2026-04-10 07:32:05 - INFO - parrotllm.training - step 31840 | epoch 0 | loss 3.4299 | lr 1.30e-03 | grad 0.2055 +2026-04-10 07:32:09 - INFO - parrotllm.training - step 31850 | epoch 0 | loss 3.4460 | lr 1.30e-03 | grad 0.2171 +2026-04-10 07:32:12 - INFO - parrotllm.training - step 31860 | epoch 0 | loss 3.4889 | lr 1.30e-03 | grad 0.2118 +2026-04-10 07:32:15 - INFO - parrotllm.training - step 31870 | epoch 0 | loss 3.5864 | lr 1.30e-03 | grad 0.2091 +2026-04-10 07:32:18 - INFO - parrotllm.training - step 31880 | epoch 0 | loss 3.4541 | lr 1.30e-03 | grad 0.1929 +2026-04-10 07:32:21 - INFO - parrotllm.training - step 31890 | epoch 0 | loss 3.4414 | lr 1.30e-03 | grad 0.2182 +2026-04-10 07:32:24 - INFO - parrotllm.training - step 31900 | epoch 0 | loss 3.5087 | lr 1.30e-03 | grad 0.2099 +2026-04-10 07:32:27 - INFO - parrotllm.training - step 31910 | epoch 0 | loss 3.4062 | lr 1.30e-03 | grad 0.2224 +2026-04-10 07:32:30 - INFO - parrotllm.training - step 31920 | epoch 0 | loss 3.4890 | lr 1.30e-03 | grad 0.1925 +2026-04-10 07:32:33 - INFO - parrotllm.training - step 31930 | epoch 0 | loss 3.4226 | lr 1.30e-03 | grad 0.1978 +2026-04-10 07:32:36 - INFO - parrotllm.training - step 31940 | epoch 0 | loss 3.3419 | lr 1.30e-03 | grad 0.2241 +2026-04-10 07:32:39 - INFO - parrotllm.training - step 31950 | epoch 0 | loss 3.5412 | lr 1.30e-03 | grad 0.2032 +2026-04-10 07:32:42 - INFO - parrotllm.training - step 31960 | epoch 0 | loss 3.4582 | lr 1.30e-03 | grad 0.1896 +2026-04-10 07:32:45 - INFO - parrotllm.training - step 31970 | epoch 0 | loss 3.3692 | lr 1.30e-03 | grad 0.2056 +2026-04-10 07:32:48 - INFO - parrotllm.training - step 31980 | epoch 0 | loss 3.4619 | lr 1.30e-03 | grad 0.2619 +2026-04-10 07:32:51 - INFO - parrotllm.training - step 31990 | epoch 0 | loss 3.5118 | lr 1.30e-03 | grad 0.2217 +2026-04-10 07:32:54 - INFO - parrotllm.training - step 32000 | epoch 0 | loss 3.5830 | lr 1.30e-03 | grad 0.2264 +2026-04-10 07:32:54 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:32:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:32:57 - INFO - parrotllm.training - Train: loss=3.5830, ppl=35.98 +2026-04-10 07:32:57 - INFO - parrotllm.training - Val: loss=3.3990, ppl=29.93 +2026-04-10 07:32:58 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3990_epoch_0000_step_0032000.pt +2026-04-10 07:32:59 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3975, min_delta=0.001000). +2026-04-10 07:33:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:33:03 - INFO - parrotllm.training - step 32010 | epoch 0 | loss 3.3676 | lr 1.30e-03 | grad 0.2335 +2026-04-10 07:33:06 - INFO - parrotllm.training - step 32020 | epoch 0 | loss 3.5241 | lr 1.30e-03 | grad 0.2109 +2026-04-10 07:33:09 - INFO - parrotllm.training - step 32030 | epoch 0 | loss 3.4468 | lr 1.30e-03 | grad 0.2247 +2026-04-10 07:33:12 - INFO - parrotllm.training - step 32040 | epoch 0 | loss 3.4686 | lr 1.30e-03 | grad 0.2087 +2026-04-10 07:33:15 - INFO - parrotllm.training - step 32050 | epoch 0 | loss 3.5675 | lr 1.30e-03 | grad 0.1919 +2026-04-10 07:33:18 - INFO - parrotllm.training - step 32060 | epoch 0 | loss 3.4177 | lr 1.30e-03 | grad 0.2058 +2026-04-10 07:33:21 - INFO - parrotllm.training - step 32070 | epoch 0 | loss 3.5240 | lr 1.30e-03 | grad 0.2087 +2026-04-10 07:33:24 - INFO - parrotllm.training - step 32080 | epoch 0 | loss 3.5876 | lr 1.30e-03 | grad 0.2350 +2026-04-10 07:33:27 - INFO - parrotllm.training - step 32090 | epoch 0 | loss 3.5176 | lr 1.30e-03 | grad 0.1931 +2026-04-10 07:33:30 - INFO - parrotllm.training - step 32100 | epoch 0 | loss 3.4226 | lr 1.30e-03 | grad 0.2369 +2026-04-10 07:33:33 - INFO - parrotllm.training - step 32110 | epoch 0 | loss 3.3620 | lr 1.30e-03 | grad 0.2523 +2026-04-10 07:33:36 - INFO - parrotllm.training - step 32120 | epoch 0 | loss 3.3929 | lr 1.30e-03 | grad 0.2199 +2026-04-10 07:33:39 - INFO - parrotllm.training - step 32130 | epoch 0 | loss 3.5896 | lr 1.30e-03 | grad 0.2271 +2026-04-10 07:33:42 - INFO - parrotllm.training - step 32140 | epoch 0 | loss 3.4385 | lr 1.30e-03 | grad 0.2171 +2026-04-10 07:33:46 - INFO - parrotllm.training - step 32150 | epoch 0 | loss 3.4121 | lr 1.30e-03 | grad 0.2012 +2026-04-10 07:33:49 - INFO - parrotllm.training - step 32160 | epoch 0 | loss 3.5252 | lr 1.30e-03 | grad 0.2210 +2026-04-10 07:33:52 - INFO - parrotllm.training - step 32170 | epoch 0 | loss 3.4442 | lr 1.30e-03 | grad 0.1937 +2026-04-10 07:33:55 - INFO - parrotllm.training - step 32180 | epoch 0 | loss 3.4430 | lr 1.30e-03 | grad 0.2021 +2026-04-10 07:33:58 - INFO - parrotllm.training - step 32190 | epoch 0 | loss 3.4735 | lr 1.30e-03 | grad 0.2689 +2026-04-10 07:34:01 - INFO - parrotllm.training - step 32200 | epoch 0 | loss 3.3223 | lr 1.30e-03 | grad 0.2112 +2026-04-10 07:34:04 - INFO - parrotllm.training - step 32210 | epoch 0 | loss 3.5015 | lr 1.30e-03 | grad 0.2035 +2026-04-10 07:34:07 - INFO - parrotllm.training - step 32220 | epoch 0 | loss 3.3761 | lr 1.30e-03 | grad 0.2148 +2026-04-10 07:34:10 - INFO - parrotllm.training - step 32230 | epoch 0 | loss 3.4335 | lr 1.29e-03 | grad 0.2011 +2026-04-10 07:34:13 - INFO - parrotllm.training - step 32240 | epoch 0 | loss 3.5568 | lr 1.29e-03 | grad 0.2179 +2026-04-10 07:34:16 - INFO - parrotllm.training - step 32250 | epoch 0 | loss 3.4209 | lr 1.29e-03 | grad 0.1981 +2026-04-10 07:34:19 - INFO - parrotllm.training - step 32260 | epoch 0 | loss 3.4509 | lr 1.29e-03 | grad 0.2232 +2026-04-10 07:34:22 - INFO - parrotllm.training - step 32270 | epoch 0 | loss 3.4577 | lr 1.29e-03 | grad 0.2329 +2026-04-10 07:34:25 - INFO - parrotllm.training - step 32280 | epoch 0 | loss 3.5810 | lr 1.29e-03 | grad 0.2112 +2026-04-10 07:34:28 - INFO - parrotllm.training - step 32290 | epoch 0 | loss 3.3915 | lr 1.29e-03 | grad 0.2372 +2026-04-10 07:34:31 - INFO - parrotllm.training - step 32300 | epoch 0 | loss 3.4425 | lr 1.29e-03 | grad 0.1878 +2026-04-10 07:34:34 - INFO - parrotllm.training - step 32310 | epoch 0 | loss 3.5033 | lr 1.29e-03 | grad 0.1985 +2026-04-10 07:34:37 - INFO - parrotllm.training - step 32320 | epoch 0 | loss 3.5133 | lr 1.29e-03 | grad 0.2230 +2026-04-10 07:34:40 - INFO - parrotllm.training - step 32330 | epoch 0 | loss 3.5170 | lr 1.29e-03 | grad 0.2201 +2026-04-10 07:34:44 - INFO - parrotllm.training - step 32340 | epoch 0 | loss 3.3693 | lr 1.29e-03 | grad 0.2106 +2026-04-10 07:34:47 - INFO - parrotllm.training - step 32350 | epoch 0 | loss 3.4509 | lr 1.29e-03 | grad 0.2274 +2026-04-10 07:34:50 - INFO - parrotllm.training - step 32360 | epoch 0 | loss 3.3259 | lr 1.29e-03 | grad 0.2163 +2026-04-10 07:34:53 - INFO - parrotllm.training - step 32370 | epoch 0 | loss 3.4816 | lr 1.29e-03 | grad 0.2159 +2026-04-10 07:34:56 - INFO - parrotllm.training - step 32380 | epoch 0 | loss 3.3960 | lr 1.29e-03 | grad 0.2215 +2026-04-10 07:34:59 - INFO - parrotllm.training - step 32390 | epoch 0 | loss 3.4256 | lr 1.29e-03 | grad 0.2288 +2026-04-10 07:35:02 - INFO - parrotllm.training - step 32400 | epoch 0 | loss 3.3251 | lr 1.29e-03 | grad 0.1963 +2026-04-10 07:35:05 - INFO - parrotllm.training - step 32410 | epoch 0 | loss 3.5163 | lr 1.29e-03 | grad 0.2137 +2026-04-10 07:35:08 - INFO - parrotllm.training - step 32420 | epoch 0 | loss 3.4573 | lr 1.29e-03 | grad 0.2330 +2026-04-10 07:35:11 - INFO - parrotllm.training - step 32430 | epoch 0 | loss 3.4184 | lr 1.29e-03 | grad 0.2097 +2026-04-10 07:35:14 - INFO - parrotllm.training - step 32440 | epoch 0 | loss 3.3864 | lr 1.29e-03 | grad 0.2303 +2026-04-10 07:35:17 - INFO - parrotllm.training - step 32450 | epoch 0 | loss 3.4922 | lr 1.29e-03 | grad 0.2249 +2026-04-10 07:35:20 - INFO - parrotllm.training - step 32460 | epoch 0 | loss 3.4912 | lr 1.29e-03 | grad 0.2385 +2026-04-10 07:35:23 - INFO - parrotllm.training - step 32470 | epoch 0 | loss 3.4623 | lr 1.29e-03 | grad 0.2303 +2026-04-10 07:35:26 - INFO - parrotllm.training - step 32480 | epoch 0 | loss 3.4811 | lr 1.29e-03 | grad 0.2040 +2026-04-10 07:35:29 - INFO - parrotllm.training - step 32490 | epoch 0 | loss 3.4302 | lr 1.29e-03 | grad 0.2015 +2026-04-10 07:35:32 - INFO - parrotllm.training - step 32500 | epoch 0 | loss 3.4186 | lr 1.29e-03 | grad 0.2244 +2026-04-10 07:35:32 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:35:32 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:35:35 - INFO - parrotllm.training - Train: loss=3.4186, ppl=30.53 +2026-04-10 07:35:35 - INFO - parrotllm.training - Val: loss=3.3981, ppl=29.91 +2026-04-10 07:35:36 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3981_epoch_0000_step_0032500.pt +2026-04-10 07:35:38 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3975, min_delta=0.001000). +2026-04-10 07:35:38 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:35:39 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0032500.pt +2026-04-10 07:35:43 - INFO - parrotllm.training - step 32510 | epoch 0 | loss 3.4562 | lr 1.29e-03 | grad 0.2173 +2026-04-10 07:35:46 - INFO - parrotllm.training - step 32520 | epoch 0 | loss 3.4855 | lr 1.29e-03 | grad 0.2163 +2026-04-10 07:35:49 - INFO - parrotllm.training - step 32530 | epoch 0 | loss 3.4439 | lr 1.29e-03 | grad 0.2268 +2026-04-10 07:35:52 - INFO - parrotllm.training - step 32540 | epoch 0 | loss 3.5980 | lr 1.29e-03 | grad 0.2249 +2026-04-10 07:35:55 - INFO - parrotllm.training - step 32550 | epoch 0 | loss 3.4686 | lr 1.29e-03 | grad 0.2322 +2026-04-10 07:35:58 - INFO - parrotllm.training - step 32560 | epoch 0 | loss 3.4659 | lr 1.29e-03 | grad 0.2035 +2026-04-10 07:36:01 - INFO - parrotllm.training - step 32570 | epoch 0 | loss 3.4785 | lr 1.29e-03 | grad 0.2218 +2026-04-10 07:36:04 - INFO - parrotllm.training - step 32580 | epoch 0 | loss 3.5472 | lr 1.29e-03 | grad 0.1967 +2026-04-10 07:36:07 - INFO - parrotllm.training - step 32590 | epoch 0 | loss 3.4933 | lr 1.29e-03 | grad 0.2103 +2026-04-10 07:36:11 - INFO - parrotllm.training - step 32600 | epoch 0 | loss 3.5271 | lr 1.29e-03 | grad 0.2209 +2026-04-10 07:36:14 - INFO - parrotllm.training - step 32610 | epoch 0 | loss 3.4799 | lr 1.29e-03 | grad 0.2057 +2026-04-10 07:36:17 - INFO - parrotllm.training - step 32620 | epoch 0 | loss 3.4142 | lr 1.29e-03 | grad 0.1986 +2026-04-10 07:36:20 - INFO - parrotllm.training - step 32630 | epoch 0 | loss 3.4406 | lr 1.29e-03 | grad 0.2111 +2026-04-10 07:36:23 - INFO - parrotllm.training - step 32640 | epoch 0 | loss 3.5452 | lr 1.29e-03 | grad 0.2301 +2026-04-10 07:36:26 - INFO - parrotllm.training - step 32650 | epoch 0 | loss 3.4408 | lr 1.29e-03 | grad 0.2260 +2026-04-10 07:36:29 - INFO - parrotllm.training - step 32660 | epoch 0 | loss 3.5071 | lr 1.29e-03 | grad 0.1958 +2026-04-10 07:36:32 - INFO - parrotllm.training - step 32670 | epoch 0 | loss 3.4295 | lr 1.29e-03 | grad 0.2293 +2026-04-10 07:36:35 - INFO - parrotllm.training - step 32680 | epoch 0 | loss 3.4675 | lr 1.29e-03 | grad 0.2200 +2026-04-10 07:36:38 - INFO - parrotllm.training - step 32690 | epoch 0 | loss 3.4244 | lr 1.29e-03 | grad 0.2020 +2026-04-10 07:36:41 - INFO - parrotllm.training - step 32700 | epoch 0 | loss 3.3708 | lr 1.29e-03 | grad 0.2562 +2026-04-10 07:36:44 - INFO - parrotllm.training - step 32710 | epoch 0 | loss 3.4432 | lr 1.29e-03 | grad 0.2105 +2026-04-10 07:36:47 - INFO - parrotllm.training - step 32720 | epoch 0 | loss 3.3933 | lr 1.29e-03 | grad 0.1954 +2026-04-10 07:36:50 - INFO - parrotllm.training - step 32730 | epoch 0 | loss 3.5238 | lr 1.29e-03 | grad 0.2351 +2026-04-10 07:36:53 - INFO - parrotllm.training - step 32740 | epoch 0 | loss 3.5202 | lr 1.29e-03 | grad 0.2063 +2026-04-10 07:36:56 - INFO - parrotllm.training - step 32750 | epoch 0 | loss 3.3927 | lr 1.29e-03 | grad 0.2311 +2026-04-10 07:36:59 - INFO - parrotllm.training - step 32760 | epoch 0 | loss 3.4227 | lr 1.29e-03 | grad 0.2101 +2026-04-10 07:37:02 - INFO - parrotllm.training - step 32770 | epoch 0 | loss 3.4983 | lr 1.29e-03 | grad 0.2071 +2026-04-10 07:37:05 - INFO - parrotllm.training - step 32780 | epoch 0 | loss 3.4242 | lr 1.29e-03 | grad 0.2151 +2026-04-10 07:37:09 - INFO - parrotllm.training - step 32790 | epoch 0 | loss 3.4166 | lr 1.29e-03 | grad 0.2274 +2026-04-10 07:37:12 - INFO - parrotllm.training - step 32800 | epoch 0 | loss 3.4414 | lr 1.29e-03 | grad 0.2051 +2026-04-10 07:37:15 - INFO - parrotllm.training - step 32810 | epoch 0 | loss 3.4660 | lr 1.29e-03 | grad 0.2041 +2026-04-10 07:37:18 - INFO - parrotllm.training - step 32820 | epoch 0 | loss 3.4371 | lr 1.29e-03 | grad 0.2114 +2026-04-10 07:37:21 - INFO - parrotllm.training - step 32830 | epoch 0 | loss 3.5301 | lr 1.29e-03 | grad 0.2781 +2026-04-10 07:37:24 - INFO - parrotllm.training - step 32840 | epoch 0 | loss 3.4396 | lr 1.29e-03 | grad 0.2182 +2026-04-10 07:37:27 - INFO - parrotllm.training - step 32850 | epoch 0 | loss 3.3589 | lr 1.29e-03 | grad 0.2170 +2026-04-10 07:37:30 - INFO - parrotllm.training - step 32860 | epoch 0 | loss 3.4461 | lr 1.29e-03 | grad 0.2148 +2026-04-10 07:37:33 - INFO - parrotllm.training - step 32870 | epoch 0 | loss 3.6045 | lr 1.29e-03 | grad 0.2237 +2026-04-10 07:37:36 - INFO - parrotllm.training - step 32880 | epoch 0 | loss 3.3748 | lr 1.29e-03 | grad 0.2011 +2026-04-10 07:37:39 - INFO - parrotllm.training - step 32890 | epoch 0 | loss 3.4057 | lr 1.29e-03 | grad 0.2190 +2026-04-10 07:37:42 - INFO - parrotllm.training - step 32900 | epoch 0 | loss 3.4296 | lr 1.29e-03 | grad 0.2309 +2026-04-10 07:37:45 - INFO - parrotllm.training - step 32910 | epoch 0 | loss 3.4070 | lr 1.29e-03 | grad 0.2087 +2026-04-10 07:37:48 - INFO - parrotllm.training - step 32920 | epoch 0 | loss 3.5157 | lr 1.29e-03 | grad 0.2129 +2026-04-10 07:37:51 - INFO - parrotllm.training - step 32930 | epoch 0 | loss 3.5579 | lr 1.29e-03 | grad 0.1990 +2026-04-10 07:37:54 - INFO - parrotllm.training - step 32940 | epoch 0 | loss 3.4322 | lr 1.29e-03 | grad 0.2362 +2026-04-10 07:37:57 - INFO - parrotllm.training - step 32950 | epoch 0 | loss 3.4277 | lr 1.29e-03 | grad 0.2005 +2026-04-10 07:38:00 - INFO - parrotllm.training - step 32960 | epoch 0 | loss 3.4405 | lr 1.29e-03 | grad 0.2350 +2026-04-10 07:38:04 - INFO - parrotllm.training - step 32970 | epoch 0 | loss 3.5441 | lr 1.29e-03 | grad 0.2170 +2026-04-10 07:38:07 - INFO - parrotllm.training - step 32980 | epoch 0 | loss 3.5394 | lr 1.29e-03 | grad 0.2058 +2026-04-10 07:38:10 - INFO - parrotllm.training - step 32990 | epoch 0 | loss 3.3733 | lr 1.29e-03 | grad 0.2154 +2026-04-10 07:38:13 - INFO - parrotllm.training - step 33000 | epoch 0 | loss 3.3957 | lr 1.29e-03 | grad 0.2084 +2026-04-10 07:38:13 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:38:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:38:16 - INFO - parrotllm.training - Train: loss=3.3957, ppl=29.84 +2026-04-10 07:38:16 - INFO - parrotllm.training - Val: loss=3.3977, ppl=29.90 +2026-04-10 07:38:17 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3977_epoch_0000_step_0033000.pt +2026-04-10 07:38:18 - INFO - parrotllm.training - No validation improvement for 3/15 evaluation(s) (best=3.3975, min_delta=0.001000). +2026-04-10 07:38:18 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:38:21 - INFO - parrotllm.training - step 33010 | epoch 0 | loss 3.4934 | lr 1.28e-03 | grad 0.2621 +2026-04-10 07:38:24 - INFO - parrotllm.training - step 33020 | epoch 0 | loss 3.4322 | lr 1.28e-03 | grad 0.2122 +2026-04-10 07:38:27 - INFO - parrotllm.training - step 33030 | epoch 0 | loss 3.5131 | lr 1.28e-03 | grad 0.2142 +2026-04-10 07:38:30 - INFO - parrotllm.training - step 33040 | epoch 0 | loss 3.4547 | lr 1.28e-03 | grad 0.2203 +2026-04-10 07:38:33 - INFO - parrotllm.training - step 33050 | epoch 0 | loss 3.4887 | lr 1.28e-03 | grad 0.2039 +2026-04-10 07:38:36 - INFO - parrotllm.training - step 33060 | epoch 0 | loss 3.3980 | lr 1.28e-03 | grad 0.2136 +2026-04-10 07:38:39 - INFO - parrotllm.training - step 33070 | epoch 0 | loss 3.4904 | lr 1.28e-03 | grad 0.2025 +2026-04-10 07:38:42 - INFO - parrotllm.training - step 33080 | epoch 0 | loss 3.4586 | lr 1.28e-03 | grad 0.2428 +2026-04-10 07:38:45 - INFO - parrotllm.training - step 33090 | epoch 0 | loss 3.4505 | lr 1.28e-03 | grad 0.2165 +2026-04-10 07:38:48 - INFO - parrotllm.training - step 33100 | epoch 0 | loss 3.3173 | lr 1.28e-03 | grad 0.1936 +2026-04-10 07:38:52 - INFO - parrotllm.training - step 33110 | epoch 0 | loss 3.4780 | lr 1.28e-03 | grad 0.2066 +2026-04-10 07:38:55 - INFO - parrotllm.training - step 33120 | epoch 0 | loss 3.4311 | lr 1.28e-03 | grad 0.2265 +2026-04-10 07:38:58 - INFO - parrotllm.training - step 33130 | epoch 0 | loss 3.4328 | lr 1.28e-03 | grad 0.1948 +2026-04-10 07:39:01 - INFO - parrotllm.training - step 33140 | epoch 0 | loss 3.4863 | lr 1.28e-03 | grad 0.2263 +2026-04-10 07:39:04 - INFO - parrotllm.training - step 33150 | epoch 0 | loss 3.3652 | lr 1.28e-03 | grad 0.2189 +2026-04-10 07:39:07 - INFO - parrotllm.training - step 33160 | epoch 0 | loss 3.3734 | lr 1.28e-03 | grad 0.2260 +2026-04-10 07:39:10 - INFO - parrotllm.training - step 33170 | epoch 0 | loss 3.4772 | lr 1.28e-03 | grad 0.1977 +2026-04-10 07:39:13 - INFO - parrotllm.training - step 33180 | epoch 0 | loss 3.3873 | lr 1.28e-03 | grad 0.2270 +2026-04-10 07:39:16 - INFO - parrotllm.training - step 33190 | epoch 0 | loss 3.3925 | lr 1.28e-03 | grad 0.2288 +2026-04-10 07:39:19 - INFO - parrotllm.training - step 33200 | epoch 0 | loss 3.5672 | lr 1.28e-03 | grad 0.2267 +2026-04-10 07:39:22 - INFO - parrotllm.training - step 33210 | epoch 0 | loss 3.5362 | lr 1.28e-03 | grad 0.2424 +2026-04-10 07:39:24 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 33227/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 07:39:25 - INFO - parrotllm.training - step 33220 | epoch 0 | loss 3.3889 | lr 1.28e-03 | grad 0.2359 +2026-04-10 07:39:28 - INFO - parrotllm.training - step 33230 | epoch 0 | loss 3.4789 | lr 1.28e-03 | grad 0.2298 +2026-04-10 07:39:32 - INFO - parrotllm.training - step 33240 | epoch 0 | loss 3.3686 | lr 1.28e-03 | grad 0.2309 +2026-04-10 07:39:35 - INFO - parrotllm.training - step 33250 | epoch 0 | loss 3.5111 | lr 1.28e-03 | grad 0.1986 +2026-04-10 07:39:38 - INFO - parrotllm.training - step 33260 | epoch 0 | loss 3.4090 | lr 1.28e-03 | grad 0.2111 +2026-04-10 07:39:41 - INFO - parrotllm.training - step 33270 | epoch 0 | loss 3.5020 | lr 1.28e-03 | grad 0.2002 +2026-04-10 07:39:44 - INFO - parrotllm.training - step 33280 | epoch 0 | loss 3.5330 | lr 1.28e-03 | grad 0.2152 +2026-04-10 07:39:47 - INFO - parrotllm.training - step 33290 | epoch 0 | loss 3.3705 | lr 1.28e-03 | grad 0.2328 +2026-04-10 07:39:50 - INFO - parrotllm.training - step 33300 | epoch 0 | loss 3.3819 | lr 1.28e-03 | grad 0.2007 +2026-04-10 07:39:53 - INFO - parrotllm.training - step 33310 | epoch 0 | loss 3.3834 | lr 1.28e-03 | grad 0.2086 +2026-04-10 07:39:56 - INFO - parrotllm.training - step 33320 | epoch 0 | loss 3.5070 | lr 1.28e-03 | grad 0.2210 +2026-04-10 07:39:59 - INFO - parrotllm.training - step 33330 | epoch 0 | loss 3.4308 | lr 1.28e-03 | grad 0.2218 +2026-04-10 07:40:02 - INFO - parrotllm.training - step 33340 | epoch 0 | loss 3.4264 | lr 1.28e-03 | grad 0.2012 +2026-04-10 07:40:05 - INFO - parrotllm.training - step 33350 | epoch 0 | loss 3.4803 | lr 1.28e-03 | grad 0.2211 +2026-04-10 07:40:08 - INFO - parrotllm.training - step 33360 | epoch 0 | loss 3.4737 | lr 1.28e-03 | grad 0.2195 +2026-04-10 07:40:11 - INFO - parrotllm.training - step 33370 | epoch 0 | loss 3.4532 | lr 1.28e-03 | grad 0.2078 +2026-04-10 07:40:14 - INFO - parrotllm.training - step 33380 | epoch 0 | loss 3.4372 | lr 1.28e-03 | grad 0.2193 +2026-04-10 07:40:17 - INFO - parrotllm.training - step 33390 | epoch 0 | loss 3.5451 | lr 1.28e-03 | grad 0.2171 +2026-04-10 07:40:20 - INFO - parrotllm.training - step 33400 | epoch 0 | loss 3.4524 | lr 1.28e-03 | grad 0.2028 +2026-04-10 07:40:23 - INFO - parrotllm.training - step 33410 | epoch 0 | loss 3.5248 | lr 1.28e-03 | grad 0.2304 +2026-04-10 07:40:26 - INFO - parrotllm.training - step 33420 | epoch 0 | loss 3.3875 | lr 1.28e-03 | grad 0.1894 +2026-04-10 07:40:30 - INFO - parrotllm.training - step 33430 | epoch 0 | loss 3.4916 | lr 1.28e-03 | grad 0.2107 +2026-04-10 07:40:33 - INFO - parrotllm.training - step 33440 | epoch 0 | loss 3.4167 | lr 1.28e-03 | grad 0.1945 +2026-04-10 07:40:36 - INFO - parrotllm.training - step 33450 | epoch 0 | loss 3.5259 | lr 1.28e-03 | grad 0.2216 +2026-04-10 07:40:39 - INFO - parrotllm.training - step 33460 | epoch 0 | loss 3.3896 | lr 1.28e-03 | grad 0.2054 +2026-04-10 07:40:42 - INFO - parrotllm.training - step 33470 | epoch 0 | loss 3.4769 | lr 1.28e-03 | grad 0.2227 +2026-04-10 07:40:45 - INFO - parrotllm.training - step 33480 | epoch 0 | loss 3.4866 | lr 1.28e-03 | grad 0.2455 +2026-04-10 07:40:48 - INFO - parrotllm.training - step 33490 | epoch 0 | loss 3.4084 | lr 1.28e-03 | grad 0.2386 +2026-04-10 07:40:51 - INFO - parrotllm.training - step 33500 | epoch 0 | loss 3.4362 | lr 1.28e-03 | grad 0.2392 +2026-04-10 07:40:51 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:40:51 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:40:54 - INFO - parrotllm.training - Train: loss=3.4362, ppl=31.07 +2026-04-10 07:40:54 - INFO - parrotllm.training - Val: loss=3.3939, ppl=29.78 +2026-04-10 07:40:54 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:40:55 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3939_epoch_0000_step_0033500.pt +2026-04-10 07:40:56 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:40:59 - INFO - parrotllm.training - step 33510 | epoch 0 | loss 3.4999 | lr 1.28e-03 | grad 0.2053 +2026-04-10 07:41:03 - INFO - parrotllm.training - step 33520 | epoch 0 | loss 3.5015 | lr 1.28e-03 | grad 0.2060 +2026-04-10 07:41:06 - INFO - parrotllm.training - step 33530 | epoch 0 | loss 3.5730 | lr 1.28e-03 | grad 0.1958 +2026-04-10 07:41:09 - INFO - parrotllm.training - step 33540 | epoch 0 | loss 3.5236 | lr 1.28e-03 | grad 0.2130 +2026-04-10 07:41:12 - INFO - parrotllm.training - step 33550 | epoch 0 | loss 3.3543 | lr 1.28e-03 | grad 0.2282 +2026-04-10 07:41:15 - INFO - parrotllm.training - step 33560 | epoch 0 | loss 3.4625 | lr 1.28e-03 | grad 0.2086 +2026-04-10 07:41:18 - INFO - parrotllm.training - step 33570 | epoch 0 | loss 3.4308 | lr 1.28e-03 | grad 0.2048 +2026-04-10 07:41:21 - INFO - parrotllm.training - step 33580 | epoch 0 | loss 3.3554 | lr 1.28e-03 | grad 0.2099 +2026-04-10 07:41:24 - INFO - parrotllm.training - step 33590 | epoch 0 | loss 3.4119 | lr 1.28e-03 | grad 0.2019 +2026-04-10 07:41:27 - INFO - parrotllm.training - step 33600 | epoch 0 | loss 3.4488 | lr 1.28e-03 | grad 0.1959 +2026-04-10 07:41:30 - INFO - parrotllm.training - step 33610 | epoch 0 | loss 3.4350 | lr 1.28e-03 | grad 0.2102 +2026-04-10 07:41:33 - INFO - parrotllm.training - step 33620 | epoch 0 | loss 3.4884 | lr 1.28e-03 | grad 0.2172 +2026-04-10 07:41:36 - INFO - parrotllm.training - step 33630 | epoch 0 | loss 3.3658 | lr 1.28e-03 | grad 0.2173 +2026-04-10 07:41:39 - INFO - parrotllm.training - step 33640 | epoch 0 | loss 3.3875 | lr 1.28e-03 | grad 0.2424 +2026-04-10 07:41:42 - INFO - parrotllm.training - step 33650 | epoch 0 | loss 3.3556 | lr 1.28e-03 | grad 0.2233 +2026-04-10 07:41:45 - INFO - parrotllm.training - step 33660 | epoch 0 | loss 3.4682 | lr 1.28e-03 | grad 0.2369 +2026-04-10 07:41:48 - INFO - parrotllm.training - step 33670 | epoch 0 | loss 3.3702 | lr 1.28e-03 | grad 0.2310 +2026-04-10 07:41:51 - INFO - parrotllm.training - step 33680 | epoch 0 | loss 3.3736 | lr 1.28e-03 | grad 0.2175 +2026-04-10 07:41:54 - INFO - parrotllm.training - step 33690 | epoch 0 | loss 3.4362 | lr 1.28e-03 | grad 0.2180 +2026-04-10 07:41:57 - INFO - parrotllm.training - step 33700 | epoch 0 | loss 3.4245 | lr 1.28e-03 | grad 0.2377 +2026-04-10 07:42:00 - INFO - parrotllm.training - step 33710 | epoch 0 | loss 3.5204 | lr 1.28e-03 | grad 0.1977 +2026-04-10 07:42:03 - INFO - parrotllm.training - step 33720 | epoch 0 | loss 3.4092 | lr 1.28e-03 | grad 0.2146 +2026-04-10 07:42:06 - INFO - parrotllm.training - step 33730 | epoch 0 | loss 3.3661 | lr 1.28e-03 | grad 0.2405 +2026-04-10 07:42:10 - INFO - parrotllm.training - step 33740 | epoch 0 | loss 3.5175 | lr 1.28e-03 | grad 0.1932 +2026-04-10 07:42:13 - INFO - parrotllm.training - step 33750 | epoch 0 | loss 3.3851 | lr 1.28e-03 | grad 0.2198 +2026-04-10 07:42:16 - INFO - parrotllm.training - step 33760 | epoch 0 | loss 3.4045 | lr 1.28e-03 | grad 0.2169 +2026-04-10 07:42:19 - INFO - parrotllm.training - step 33770 | epoch 0 | loss 3.4325 | lr 1.28e-03 | grad 0.2254 +2026-04-10 07:42:22 - INFO - parrotllm.training - step 33780 | epoch 0 | loss 3.5061 | lr 1.27e-03 | grad 0.2233 +2026-04-10 07:42:25 - INFO - parrotllm.training - step 33790 | epoch 0 | loss 3.4077 | lr 1.27e-03 | grad 0.2027 +2026-04-10 07:42:28 - INFO - parrotllm.training - step 33800 | epoch 0 | loss 3.4722 | lr 1.27e-03 | grad 0.2561 +2026-04-10 07:42:31 - INFO - parrotllm.training - step 33810 | epoch 0 | loss 3.3937 | lr 1.27e-03 | grad 0.2038 +2026-04-10 07:42:34 - INFO - parrotllm.training - step 33820 | epoch 0 | loss 3.4247 | lr 1.27e-03 | grad 0.2127 +2026-04-10 07:42:37 - INFO - parrotllm.training - step 33830 | epoch 0 | loss 3.4322 | lr 1.27e-03 | grad 0.2539 +2026-04-10 07:42:40 - INFO - parrotllm.training - step 33840 | epoch 0 | loss 3.4285 | lr 1.27e-03 | grad 0.2242 +2026-04-10 07:42:43 - INFO - parrotllm.training - step 33850 | epoch 0 | loss 3.4538 | lr 1.27e-03 | grad 0.2264 +2026-04-10 07:42:46 - INFO - parrotllm.training - step 33860 | epoch 0 | loss 3.5201 | lr 1.27e-03 | grad 0.2239 +2026-04-10 07:42:49 - INFO - parrotllm.training - step 33870 | epoch 0 | loss 3.4433 | lr 1.27e-03 | grad 0.2001 +2026-04-10 07:42:52 - INFO - parrotllm.training - step 33880 | epoch 0 | loss 3.4973 | lr 1.27e-03 | grad 0.2070 +2026-04-10 07:42:55 - INFO - parrotllm.training - step 33890 | epoch 0 | loss 3.6085 | lr 1.27e-03 | grad 0.2199 +2026-04-10 07:42:58 - INFO - parrotllm.training - step 33900 | epoch 0 | loss 3.4653 | lr 1.27e-03 | grad 0.2070 +2026-04-10 07:43:01 - INFO - parrotllm.training - step 33910 | epoch 0 | loss 3.5089 | lr 1.27e-03 | grad 0.2415 +2026-04-10 07:43:04 - INFO - parrotllm.training - step 33920 | epoch 0 | loss 3.4572 | lr 1.27e-03 | grad 0.2176 +2026-04-10 07:43:08 - INFO - parrotllm.training - step 33930 | epoch 0 | loss 3.4515 | lr 1.27e-03 | grad 0.2179 +2026-04-10 07:43:11 - INFO - parrotllm.training - step 33940 | epoch 0 | loss 3.4725 | lr 1.27e-03 | grad 0.2134 +2026-04-10 07:43:14 - INFO - parrotllm.training - step 33950 | epoch 0 | loss 3.4677 | lr 1.27e-03 | grad 0.2183 +2026-04-10 07:43:17 - INFO - parrotllm.training - step 33960 | epoch 0 | loss 3.4882 | lr 1.27e-03 | grad 0.2079 +2026-04-10 07:43:20 - INFO - parrotllm.training - step 33970 | epoch 0 | loss 3.4170 | lr 1.27e-03 | grad 0.2216 +2026-04-10 07:43:23 - INFO - parrotllm.training - step 33980 | epoch 0 | loss 3.4067 | lr 1.27e-03 | grad 0.2487 +2026-04-10 07:43:26 - INFO - parrotllm.training - step 33990 | epoch 0 | loss 3.4997 | lr 1.27e-03 | grad 0.2636 +2026-04-10 07:43:29 - INFO - parrotllm.training - step 34000 | epoch 0 | loss 3.4523 | lr 1.27e-03 | grad 0.2106 +2026-04-10 07:43:29 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:43:29 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:43:32 - INFO - parrotllm.training - Train: loss=3.4523, ppl=31.57 +2026-04-10 07:43:32 - INFO - parrotllm.training - Val: loss=3.3968, ppl=29.87 +2026-04-10 07:43:33 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3968_epoch_0000_step_0034000.pt +2026-04-10 07:43:34 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3939, min_delta=0.001000). +2026-04-10 07:43:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:43:37 - INFO - parrotllm.training - step 34010 | epoch 0 | loss 3.4683 | lr 1.27e-03 | grad 0.2290 +2026-04-10 07:43:40 - INFO - parrotllm.training - step 34020 | epoch 0 | loss 3.5100 | lr 1.27e-03 | grad 0.2024 +2026-04-10 07:43:43 - INFO - parrotllm.training - step 34030 | epoch 0 | loss 3.4788 | lr 1.27e-03 | grad 0.2038 +2026-04-10 07:43:46 - INFO - parrotllm.training - step 34040 | epoch 0 | loss 3.3803 | lr 1.27e-03 | grad 0.2574 +2026-04-10 07:43:49 - INFO - parrotllm.training - step 34050 | epoch 0 | loss 3.4005 | lr 1.27e-03 | grad 0.2191 +2026-04-10 07:43:52 - INFO - parrotllm.training - step 34060 | epoch 0 | loss 3.3499 | lr 1.27e-03 | grad 0.2275 +2026-04-10 07:43:55 - INFO - parrotllm.training - step 34070 | epoch 0 | loss 3.4107 | lr 1.27e-03 | grad 0.2420 +2026-04-10 07:43:58 - INFO - parrotllm.training - step 34080 | epoch 0 | loss 3.3517 | lr 1.27e-03 | grad 0.2185 +2026-04-10 07:44:02 - INFO - parrotllm.training - step 34090 | epoch 0 | loss 3.3972 | lr 1.27e-03 | grad 0.2126 +2026-04-10 07:44:05 - INFO - parrotllm.training - step 34100 | epoch 0 | loss 3.4623 | lr 1.27e-03 | grad 0.2133 +2026-04-10 07:44:08 - INFO - parrotllm.training - step 34110 | epoch 0 | loss 3.5571 | lr 1.27e-03 | grad 0.2063 +2026-04-10 07:44:11 - INFO - parrotllm.training - step 34120 | epoch 0 | loss 3.3521 | lr 1.27e-03 | grad 0.1939 +2026-04-10 07:44:14 - INFO - parrotllm.training - step 34130 | epoch 0 | loss 3.4380 | lr 1.27e-03 | grad 0.2274 +2026-04-10 07:44:17 - INFO - parrotllm.training - step 34140 | epoch 0 | loss 3.5310 | lr 1.27e-03 | grad 0.2222 +2026-04-10 07:44:20 - INFO - parrotllm.training - step 34150 | epoch 0 | loss 3.4509 | lr 1.27e-03 | grad 0.2207 +2026-04-10 07:44:23 - INFO - parrotllm.training - step 34160 | epoch 0 | loss 3.5719 | lr 1.27e-03 | grad 0.2000 +2026-04-10 07:44:26 - INFO - parrotllm.training - step 34170 | epoch 0 | loss 3.4819 | lr 1.27e-03 | grad 0.2402 +2026-04-10 07:44:29 - INFO - parrotllm.training - step 34180 | epoch 0 | loss 3.5062 | lr 1.27e-03 | grad 0.2319 +2026-04-10 07:44:32 - INFO - parrotllm.training - step 34190 | epoch 0 | loss 3.4722 | lr 1.27e-03 | grad 0.2280 +2026-04-10 07:44:35 - INFO - parrotllm.training - step 34200 | epoch 0 | loss 3.4052 | lr 1.27e-03 | grad 0.2102 +2026-04-10 07:44:38 - INFO - parrotllm.training - step 34210 | epoch 0 | loss 3.5425 | lr 1.27e-03 | grad 0.2086 +2026-04-10 07:44:41 - INFO - parrotllm.training - step 34220 | epoch 0 | loss 3.4012 | lr 1.27e-03 | grad 0.1950 +2026-04-10 07:44:44 - INFO - parrotllm.training - step 34230 | epoch 0 | loss 3.4222 | lr 1.27e-03 | grad 0.2167 +2026-04-10 07:44:47 - INFO - parrotllm.training - step 34240 | epoch 0 | loss 3.4134 | lr 1.27e-03 | grad 0.2418 +2026-04-10 07:44:50 - INFO - parrotllm.training - step 34250 | epoch 0 | loss 3.4499 | lr 1.27e-03 | grad 0.2353 +2026-04-10 07:44:53 - INFO - parrotllm.training - step 34260 | epoch 0 | loss 3.5234 | lr 1.27e-03 | grad 0.2079 +2026-04-10 07:44:56 - INFO - parrotllm.training - step 34270 | epoch 0 | loss 3.4519 | lr 1.27e-03 | grad 0.2076 +2026-04-10 07:45:00 - INFO - parrotllm.training - step 34280 | epoch 0 | loss 3.4884 | lr 1.27e-03 | grad 0.2341 +2026-04-10 07:45:03 - INFO - parrotllm.training - step 34290 | epoch 0 | loss 3.4764 | lr 1.27e-03 | grad 0.2125 +2026-04-10 07:45:06 - INFO - parrotllm.training - step 34300 | epoch 0 | loss 3.4371 | lr 1.27e-03 | grad 0.2441 +2026-04-10 07:45:09 - INFO - parrotllm.training - step 34310 | epoch 0 | loss 3.4767 | lr 1.27e-03 | grad 0.2317 +2026-04-10 07:45:12 - INFO - parrotllm.training - step 34320 | epoch 0 | loss 3.4117 | lr 1.27e-03 | grad 0.2314 +2026-04-10 07:45:15 - INFO - parrotllm.training - step 34330 | epoch 0 | loss 3.4420 | lr 1.27e-03 | grad 0.2285 +2026-04-10 07:45:18 - INFO - parrotllm.training - step 34340 | epoch 0 | loss 3.3791 | lr 1.27e-03 | grad 0.2152 +2026-04-10 07:45:21 - INFO - parrotllm.training - step 34350 | epoch 0 | loss 3.3894 | lr 1.27e-03 | grad 0.2097 +2026-04-10 07:45:24 - INFO - parrotllm.training - step 34360 | epoch 0 | loss 3.4699 | lr 1.27e-03 | grad 0.2428 +2026-04-10 07:45:27 - INFO - parrotllm.training - step 34370 | epoch 0 | loss 3.5200 | lr 1.27e-03 | grad 0.2357 +2026-04-10 07:45:30 - INFO - parrotllm.training - step 34380 | epoch 0 | loss 3.4120 | lr 1.27e-03 | grad 0.2176 +2026-04-10 07:45:33 - INFO - parrotllm.training - step 34390 | epoch 0 | loss 3.5636 | lr 1.27e-03 | grad 0.2246 +2026-04-10 07:45:36 - INFO - parrotllm.training - step 34400 | epoch 0 | loss 3.3585 | lr 1.27e-03 | grad 0.2159 +2026-04-10 07:45:39 - INFO - parrotllm.training - step 34410 | epoch 0 | loss 3.4243 | lr 1.27e-03 | grad 0.1860 +2026-04-10 07:45:42 - INFO - parrotllm.training - step 34420 | epoch 0 | loss 3.4958 | lr 1.27e-03 | grad 0.2012 +2026-04-10 07:45:45 - INFO - parrotllm.training - step 34430 | epoch 0 | loss 3.4806 | lr 1.27e-03 | grad 0.2162 +2026-04-10 07:45:48 - INFO - parrotllm.training - step 34440 | epoch 0 | loss 3.4665 | lr 1.27e-03 | grad 0.2118 +2026-04-10 07:45:51 - INFO - parrotllm.training - step 34450 | epoch 0 | loss 3.5324 | lr 1.27e-03 | grad 0.2582 +2026-04-10 07:45:55 - INFO - parrotllm.training - step 34460 | epoch 0 | loss 3.4731 | lr 1.27e-03 | grad 0.2018 +2026-04-10 07:45:58 - INFO - parrotllm.training - step 34470 | epoch 0 | loss 3.4054 | lr 1.27e-03 | grad 0.2487 +2026-04-10 07:46:01 - INFO - parrotllm.training - step 34480 | epoch 0 | loss 3.5623 | lr 1.27e-03 | grad 0.2070 +2026-04-10 07:46:04 - INFO - parrotllm.training - step 34490 | epoch 0 | loss 3.4079 | lr 1.27e-03 | grad 0.2030 +2026-04-10 07:46:07 - INFO - parrotllm.training - step 34500 | epoch 0 | loss 3.5151 | lr 1.27e-03 | grad 0.2563 +2026-04-10 07:46:07 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:46:07 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:46:10 - INFO - parrotllm.training - Train: loss=3.5151, ppl=33.62 +2026-04-10 07:46:10 - INFO - parrotllm.training - Val: loss=3.3930, ppl=29.75 +2026-04-10 07:46:11 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3930_epoch_0000_step_0034500.pt +2026-04-10 07:46:12 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3939, min_delta=0.001000). +2026-04-10 07:46:12 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:46:15 - INFO - parrotllm.training - step 34510 | epoch 0 | loss 3.4088 | lr 1.27e-03 | grad 0.1947 +2026-04-10 07:46:18 - INFO - parrotllm.training - step 34520 | epoch 0 | loss 3.4504 | lr 1.27e-03 | grad 0.2318 +2026-04-10 07:46:21 - INFO - parrotllm.training - step 34530 | epoch 0 | loss 3.3854 | lr 1.27e-03 | grad 0.2172 +2026-04-10 07:46:24 - INFO - parrotllm.training - step 34540 | epoch 0 | loss 3.5124 | lr 1.26e-03 | grad 0.2166 +2026-04-10 07:46:27 - INFO - parrotllm.training - step 34550 | epoch 0 | loss 3.4238 | lr 1.26e-03 | grad 0.2083 +2026-04-10 07:46:30 - INFO - parrotllm.training - step 34560 | epoch 0 | loss 3.5140 | lr 1.26e-03 | grad 0.2289 +2026-04-10 07:46:33 - INFO - parrotllm.training - step 34570 | epoch 0 | loss 3.4208 | lr 1.26e-03 | grad 0.2188 +2026-04-10 07:46:36 - INFO - parrotllm.training - step 34580 | epoch 0 | loss 3.4556 | lr 1.26e-03 | grad 0.2207 +2026-04-10 07:46:40 - INFO - parrotllm.training - step 34590 | epoch 0 | loss 3.4788 | lr 1.26e-03 | grad 0.1952 +2026-04-10 07:46:43 - INFO - parrotllm.training - step 34600 | epoch 0 | loss 3.5446 | lr 1.26e-03 | grad 0.2183 +2026-04-10 07:46:46 - INFO - parrotllm.training - step 34610 | epoch 0 | loss 3.4478 | lr 1.26e-03 | grad 0.2108 +2026-04-10 07:46:49 - INFO - parrotllm.training - step 34620 | epoch 0 | loss 3.4329 | lr 1.26e-03 | grad 0.2113 +2026-04-10 07:46:52 - INFO - parrotllm.training - step 34630 | epoch 0 | loss 3.5096 | lr 1.26e-03 | grad 0.2148 +2026-04-10 07:46:55 - INFO - parrotllm.training - step 34640 | epoch 0 | loss 3.5056 | lr 1.26e-03 | grad 0.1989 +2026-04-10 07:46:58 - INFO - parrotllm.training - step 34650 | epoch 0 | loss 3.4496 | lr 1.26e-03 | grad 0.2183 +2026-04-10 07:47:01 - INFO - parrotllm.training - step 34660 | epoch 0 | loss 3.3824 | lr 1.26e-03 | grad 0.1952 +2026-04-10 07:47:04 - INFO - parrotllm.training - step 34670 | epoch 0 | loss 3.5785 | lr 1.26e-03 | grad 0.2134 +2026-04-10 07:47:07 - INFO - parrotllm.training - step 34680 | epoch 0 | loss 3.4795 | lr 1.26e-03 | grad 0.2558 +2026-04-10 07:47:10 - INFO - parrotllm.training - step 34690 | epoch 0 | loss 3.3601 | lr 1.26e-03 | grad 0.2041 +2026-04-10 07:47:13 - INFO - parrotllm.training - step 34700 | epoch 0 | loss 3.4802 | lr 1.26e-03 | grad 0.1888 +2026-04-10 07:47:16 - INFO - parrotllm.training - step 34710 | epoch 0 | loss 3.4897 | lr 1.26e-03 | grad 0.2249 +2026-04-10 07:47:19 - INFO - parrotllm.training - step 34720 | epoch 0 | loss 3.3840 | lr 1.26e-03 | grad 0.2080 +2026-04-10 07:47:22 - INFO - parrotllm.training - step 34730 | epoch 0 | loss 3.4725 | lr 1.26e-03 | grad 0.2264 +2026-04-10 07:47:25 - INFO - parrotllm.training - step 34740 | epoch 0 | loss 3.5248 | lr 1.26e-03 | grad 0.2284 +2026-04-10 07:47:28 - INFO - parrotllm.training - step 34750 | epoch 0 | loss 3.5287 | lr 1.26e-03 | grad 0.2154 +2026-04-10 07:47:31 - INFO - parrotllm.training - step 34760 | epoch 0 | loss 3.3845 | lr 1.26e-03 | grad 0.1922 +2026-04-10 07:47:34 - INFO - parrotllm.training - step 34770 | epoch 0 | loss 3.3879 | lr 1.26e-03 | grad 0.2673 +2026-04-10 07:47:37 - INFO - parrotllm.training - step 34780 | epoch 0 | loss 3.3299 | lr 1.26e-03 | grad 0.2518 +2026-04-10 07:47:40 - INFO - parrotllm.training - step 34790 | epoch 0 | loss 3.5067 | lr 1.26e-03 | grad 0.2235 +2026-04-10 07:47:44 - INFO - parrotllm.training - step 34800 | epoch 0 | loss 3.5621 | lr 1.26e-03 | grad 0.2105 +2026-04-10 07:47:47 - INFO - parrotllm.training - step 34810 | epoch 0 | loss 3.4128 | lr 1.26e-03 | grad 0.2108 +2026-04-10 07:47:50 - INFO - parrotllm.training - step 34820 | epoch 0 | loss 3.5179 | lr 1.26e-03 | grad 0.2260 +2026-04-10 07:47:53 - INFO - parrotllm.training - step 34830 | epoch 0 | loss 3.4541 | lr 1.26e-03 | grad 0.1948 +2026-04-10 07:47:56 - INFO - parrotllm.training - step 34840 | epoch 0 | loss 3.5022 | lr 1.26e-03 | grad 0.2374 +2026-04-10 07:47:59 - INFO - parrotllm.training - step 34850 | epoch 0 | loss 3.4646 | lr 1.26e-03 | grad 0.2167 +2026-04-10 07:48:02 - INFO - parrotllm.training - step 34860 | epoch 0 | loss 3.5247 | lr 1.26e-03 | grad 0.2072 +2026-04-10 07:48:05 - INFO - parrotllm.training - step 34870 | epoch 0 | loss 3.3841 | lr 1.26e-03 | grad 0.1976 +2026-04-10 07:48:08 - INFO - parrotllm.training - step 34880 | epoch 0 | loss 3.4392 | lr 1.26e-03 | grad 0.2287 +2026-04-10 07:48:11 - INFO - parrotllm.training - step 34890 | epoch 0 | loss 3.4691 | lr 1.26e-03 | grad 0.2421 +2026-04-10 07:48:14 - INFO - parrotllm.training - step 34900 | epoch 0 | loss 3.2605 | lr 1.26e-03 | grad 0.2277 +2026-04-10 07:48:17 - INFO - parrotllm.training - step 34910 | epoch 0 | loss 3.5172 | lr 1.26e-03 | grad 0.1936 +2026-04-10 07:48:20 - INFO - parrotllm.training - step 34920 | epoch 0 | loss 3.5023 | lr 1.26e-03 | grad 0.2063 +2026-04-10 07:48:23 - INFO - parrotllm.training - step 34930 | epoch 0 | loss 3.5004 | lr 1.26e-03 | grad 0.2106 +2026-04-10 07:48:26 - INFO - parrotllm.training - step 34940 | epoch 0 | loss 3.5246 | lr 1.26e-03 | grad 0.2355 +2026-04-10 07:48:29 - INFO - parrotllm.training - step 34950 | epoch 0 | loss 3.4222 | lr 1.26e-03 | grad 0.2553 +2026-04-10 07:48:32 - INFO - parrotllm.training - step 34960 | epoch 0 | loss 3.4356 | lr 1.26e-03 | grad 0.2166 +2026-04-10 07:48:35 - INFO - parrotllm.training - step 34970 | epoch 0 | loss 3.3745 | lr 1.26e-03 | grad 0.2142 +2026-04-10 07:48:38 - INFO - parrotllm.training - step 34980 | epoch 0 | loss 3.5157 | lr 1.26e-03 | grad 0.1979 +2026-04-10 07:48:42 - INFO - parrotllm.training - step 34990 | epoch 0 | loss 3.4130 | lr 1.26e-03 | grad 0.1884 +2026-04-10 07:48:45 - INFO - parrotllm.training - step 35000 | epoch 0 | loss 3.4042 | lr 1.26e-03 | grad 0.2590 +2026-04-10 07:48:45 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:48:45 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:48:48 - INFO - parrotllm.training - Train: loss=3.4042, ppl=30.09 +2026-04-10 07:48:48 - INFO - parrotllm.training - Val: loss=3.3933, ppl=29.76 +2026-04-10 07:48:49 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3933_epoch_0000_step_0035000.pt +2026-04-10 07:48:50 - INFO - parrotllm.training - No validation improvement for 3/15 evaluation(s) (best=3.3939, min_delta=0.001000). +2026-04-10 07:48:50 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:48:51 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0035000.pt +2026-04-10 07:48:56 - INFO - parrotllm.training - step 35010 | epoch 0 | loss 3.4259 | lr 1.26e-03 | grad 0.2206 +2026-04-10 07:48:59 - INFO - parrotllm.training - step 35020 | epoch 0 | loss 3.4595 | lr 1.26e-03 | grad 0.2323 +2026-04-10 07:49:02 - INFO - parrotllm.training - step 35030 | epoch 0 | loss 3.4690 | lr 1.26e-03 | grad 0.2276 +2026-04-10 07:49:05 - INFO - parrotllm.training - step 35040 | epoch 0 | loss 3.4366 | lr 1.26e-03 | grad 0.2386 +2026-04-10 07:49:08 - INFO - parrotllm.training - step 35050 | epoch 0 | loss 3.3987 | lr 1.26e-03 | grad 0.2188 +2026-04-10 07:49:11 - INFO - parrotllm.training - step 35060 | epoch 0 | loss 3.5023 | lr 1.26e-03 | grad 0.2361 +2026-04-10 07:49:14 - INFO - parrotllm.training - step 35070 | epoch 0 | loss 3.4969 | lr 1.26e-03 | grad 0.2093 +2026-04-10 07:49:17 - INFO - parrotllm.training - step 35080 | epoch 0 | loss 3.4299 | lr 1.26e-03 | grad 0.2108 +2026-04-10 07:49:20 - INFO - parrotllm.training - step 35090 | epoch 0 | loss 3.5880 | lr 1.26e-03 | grad 0.2351 +2026-04-10 07:49:23 - INFO - parrotllm.training - step 35100 | epoch 0 | loss 3.5169 | lr 1.26e-03 | grad 0.2258 +2026-04-10 07:49:26 - INFO - parrotllm.training - step 35110 | epoch 0 | loss 3.5003 | lr 1.26e-03 | grad 0.2321 +2026-04-10 07:49:29 - INFO - parrotllm.training - step 35120 | epoch 0 | loss 3.5188 | lr 1.26e-03 | grad 0.2078 +2026-04-10 07:49:32 - INFO - parrotllm.training - step 35130 | epoch 0 | loss 3.4589 | lr 1.26e-03 | grad 0.2081 +2026-04-10 07:49:35 - INFO - parrotllm.training - step 35140 | epoch 0 | loss 3.5164 | lr 1.26e-03 | grad 0.2388 +2026-04-10 07:49:38 - INFO - parrotllm.training - step 35150 | epoch 0 | loss 3.4829 | lr 1.26e-03 | grad 0.2286 +2026-04-10 07:49:42 - INFO - parrotllm.training - step 35160 | epoch 0 | loss 3.4350 | lr 1.26e-03 | grad 0.2647 +2026-04-10 07:49:45 - INFO - parrotllm.training - step 35170 | epoch 0 | loss 3.5126 | lr 1.26e-03 | grad 0.2305 +2026-04-10 07:49:48 - INFO - parrotllm.training - step 35180 | epoch 0 | loss 3.4401 | lr 1.26e-03 | grad 0.2301 +2026-04-10 07:49:51 - INFO - parrotllm.training - step 35190 | epoch 0 | loss 3.3262 | lr 1.26e-03 | grad 0.2531 +2026-04-10 07:49:54 - INFO - parrotllm.training - step 35200 | epoch 0 | loss 3.4150 | lr 1.26e-03 | grad 0.1972 +2026-04-10 07:49:57 - INFO - parrotllm.training - step 35210 | epoch 0 | loss 3.4196 | lr 1.26e-03 | grad 0.2423 +2026-04-10 07:50:00 - INFO - parrotllm.training - step 35220 | epoch 0 | loss 3.3558 | lr 1.26e-03 | grad 0.2506 +2026-04-10 07:50:03 - INFO - parrotllm.training - step 35230 | epoch 0 | loss 3.4203 | lr 1.26e-03 | grad 0.2016 +2026-04-10 07:50:03 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 35244/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 07:50:06 - INFO - parrotllm.training - step 35240 | epoch 0 | loss 3.3816 | lr 1.26e-03 | grad 0.2175 +2026-04-10 07:50:09 - INFO - parrotllm.training - step 35250 | epoch 0 | loss 3.4911 | lr 1.26e-03 | grad 0.2149 +2026-04-10 07:50:12 - INFO - parrotllm.training - step 35260 | epoch 0 | loss 3.4363 | lr 1.26e-03 | grad 0.2051 +2026-04-10 07:50:15 - INFO - parrotllm.training - step 35270 | epoch 0 | loss 3.5038 | lr 1.26e-03 | grad 0.2045 +2026-04-10 07:50:18 - INFO - parrotllm.training - step 35280 | epoch 0 | loss 3.4340 | lr 1.25e-03 | grad 0.2363 +2026-04-10 07:50:22 - INFO - parrotllm.training - step 35290 | epoch 0 | loss 3.3571 | lr 1.25e-03 | grad 0.2381 +2026-04-10 07:50:25 - INFO - parrotllm.training - step 35300 | epoch 0 | loss 3.5340 | lr 1.25e-03 | grad 0.2196 +2026-04-10 07:50:28 - INFO - parrotllm.training - step 35310 | epoch 0 | loss 3.5176 | lr 1.25e-03 | grad 0.2051 +2026-04-10 07:50:31 - INFO - parrotllm.training - step 35320 | epoch 0 | loss 3.4256 | lr 1.25e-03 | grad 0.2273 +2026-04-10 07:50:34 - INFO - parrotllm.training - step 35330 | epoch 0 | loss 3.4099 | lr 1.25e-03 | grad 0.2177 +2026-04-10 07:50:37 - INFO - parrotllm.training - step 35340 | epoch 0 | loss 3.4159 | lr 1.25e-03 | grad 0.2253 +2026-04-10 07:50:40 - INFO - parrotllm.training - step 35350 | epoch 0 | loss 3.5445 | lr 1.25e-03 | grad 0.2328 +2026-04-10 07:50:43 - INFO - parrotllm.training - step 35360 | epoch 0 | loss 3.4298 | lr 1.25e-03 | grad 0.2187 +2026-04-10 07:50:46 - INFO - parrotllm.training - step 35370 | epoch 0 | loss 3.3797 | lr 1.25e-03 | grad 0.2017 +2026-04-10 07:50:49 - INFO - parrotllm.training - step 35380 | epoch 0 | loss 3.4006 | lr 1.25e-03 | grad 0.2215 +2026-04-10 07:50:52 - INFO - parrotllm.training - step 35390 | epoch 0 | loss 3.3392 | lr 1.25e-03 | grad 0.2697 +2026-04-10 07:50:55 - INFO - parrotllm.training - step 35400 | epoch 0 | loss 3.5670 | lr 1.25e-03 | grad 0.2373 +2026-04-10 07:50:58 - INFO - parrotllm.training - step 35410 | epoch 0 | loss 3.4303 | lr 1.25e-03 | grad 0.2078 +2026-04-10 07:51:01 - INFO - parrotllm.training - step 35420 | epoch 0 | loss 3.4437 | lr 1.25e-03 | grad 0.2200 +2026-04-10 07:51:04 - INFO - parrotllm.training - step 35430 | epoch 0 | loss 3.5159 | lr 1.25e-03 | grad 0.2123 +2026-04-10 07:51:07 - INFO - parrotllm.training - step 35440 | epoch 0 | loss 3.4347 | lr 1.25e-03 | grad 0.2027 +2026-04-10 07:51:10 - INFO - parrotllm.training - step 35450 | epoch 0 | loss 3.4081 | lr 1.25e-03 | grad 0.1930 +2026-04-10 07:51:13 - INFO - parrotllm.training - step 35460 | epoch 0 | loss 3.5387 | lr 1.25e-03 | grad 0.2137 +2026-04-10 07:51:17 - INFO - parrotllm.training - step 35470 | epoch 0 | loss 3.5114 | lr 1.25e-03 | grad 0.2126 +2026-04-10 07:51:20 - INFO - parrotllm.training - step 35480 | epoch 0 | loss 3.5041 | lr 1.25e-03 | grad 0.2123 +2026-04-10 07:51:23 - INFO - parrotllm.training - step 35490 | epoch 0 | loss 3.4920 | lr 1.25e-03 | grad 0.1985 +2026-04-10 07:51:26 - INFO - parrotllm.training - step 35500 | epoch 0 | loss 3.3588 | lr 1.25e-03 | grad 0.2232 +2026-04-10 07:51:26 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:51:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:51:29 - INFO - parrotllm.training - Train: loss=3.3588, ppl=28.75 +2026-04-10 07:51:29 - INFO - parrotllm.training - Val: loss=3.3904, ppl=29.68 +2026-04-10 07:51:29 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:51:30 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3904_epoch_0000_step_0035500.pt +2026-04-10 07:51:31 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:51:34 - INFO - parrotllm.training - step 35510 | epoch 0 | loss 3.4586 | lr 1.25e-03 | grad 0.2287 +2026-04-10 07:51:37 - INFO - parrotllm.training - step 35520 | epoch 0 | loss 3.4724 | lr 1.25e-03 | grad 0.2748 +2026-04-10 07:51:40 - INFO - parrotllm.training - step 35530 | epoch 0 | loss 3.3704 | lr 1.25e-03 | grad 0.2687 +2026-04-10 07:51:43 - INFO - parrotllm.training - step 35540 | epoch 0 | loss 3.4644 | lr 1.25e-03 | grad 0.2176 +2026-04-10 07:51:46 - INFO - parrotllm.training - step 35550 | epoch 0 | loss 3.5428 | lr 1.25e-03 | grad 0.2044 +2026-04-10 07:51:49 - INFO - parrotllm.training - step 35560 | epoch 0 | loss 3.5543 | lr 1.25e-03 | grad 0.2232 +2026-04-10 07:51:52 - INFO - parrotllm.training - step 35570 | epoch 0 | loss 3.4478 | lr 1.25e-03 | grad 0.2234 +2026-04-10 07:51:55 - INFO - parrotllm.training - step 35580 | epoch 0 | loss 3.4213 | lr 1.25e-03 | grad 0.2560 +2026-04-10 07:51:59 - INFO - parrotllm.training - step 35590 | epoch 0 | loss 3.4593 | lr 1.25e-03 | grad 0.2427 +2026-04-10 07:52:02 - INFO - parrotllm.training - step 35600 | epoch 0 | loss 3.5094 | lr 1.25e-03 | grad 0.2057 +2026-04-10 07:52:05 - INFO - parrotllm.training - step 35610 | epoch 0 | loss 3.4552 | lr 1.25e-03 | grad 0.2754 +2026-04-10 07:52:08 - INFO - parrotllm.training - step 35620 | epoch 0 | loss 3.4088 | lr 1.25e-03 | grad 0.2455 +2026-04-10 07:52:11 - INFO - parrotllm.training - step 35630 | epoch 0 | loss 3.4215 | lr 1.25e-03 | grad 0.2091 +2026-04-10 07:52:14 - INFO - parrotllm.training - step 35640 | epoch 0 | loss 3.4986 | lr 1.25e-03 | grad 0.2135 +2026-04-10 07:52:17 - INFO - parrotllm.training - step 35650 | epoch 0 | loss 3.4951 | lr 1.25e-03 | grad 0.2128 +2026-04-10 07:52:20 - INFO - parrotllm.training - step 35660 | epoch 0 | loss 3.5286 | lr 1.25e-03 | grad 0.2371 +2026-04-10 07:52:23 - INFO - parrotllm.training - step 35670 | epoch 0 | loss 3.3879 | lr 1.25e-03 | grad 0.2577 +2026-04-10 07:52:26 - INFO - parrotllm.training - step 35680 | epoch 0 | loss 3.5075 | lr 1.25e-03 | grad 0.2467 +2026-04-10 07:52:29 - INFO - parrotllm.training - step 35690 | epoch 0 | loss 3.4425 | lr 1.25e-03 | grad 0.2137 +2026-04-10 07:52:32 - INFO - parrotllm.training - step 35700 | epoch 0 | loss 3.3419 | lr 1.25e-03 | grad 0.2202 +2026-04-10 07:52:35 - INFO - parrotllm.training - step 35710 | epoch 0 | loss 3.4844 | lr 1.25e-03 | grad 0.2173 +2026-04-10 07:52:38 - INFO - parrotllm.training - step 35720 | epoch 0 | loss 3.5225 | lr 1.25e-03 | grad 0.2342 +2026-04-10 07:52:41 - INFO - parrotllm.training - step 35730 | epoch 0 | loss 3.3974 | lr 1.25e-03 | grad 0.2192 +2026-04-10 07:52:44 - INFO - parrotllm.training - step 35740 | epoch 0 | loss 3.4795 | lr 1.25e-03 | grad 0.2400 +2026-04-10 07:52:47 - INFO - parrotllm.training - step 35750 | epoch 0 | loss 3.5107 | lr 1.25e-03 | grad 0.2183 +2026-04-10 07:52:50 - INFO - parrotllm.training - step 35760 | epoch 0 | loss 3.4805 | lr 1.25e-03 | grad 0.2446 +2026-04-10 07:52:53 - INFO - parrotllm.training - step 35770 | epoch 0 | loss 3.4078 | lr 1.25e-03 | grad 0.2508 +2026-04-10 07:52:57 - INFO - parrotllm.training - step 35780 | epoch 0 | loss 3.4295 | lr 1.25e-03 | grad 0.1995 +2026-04-10 07:53:00 - INFO - parrotllm.training - step 35790 | epoch 0 | loss 3.3978 | lr 1.25e-03 | grad 0.2150 +2026-04-10 07:53:03 - INFO - parrotllm.training - step 35800 | epoch 0 | loss 3.4049 | lr 1.25e-03 | grad 0.2752 +2026-04-10 07:53:06 - INFO - parrotllm.training - step 35810 | epoch 0 | loss 3.5268 | lr 1.25e-03 | grad 0.1926 +2026-04-10 07:53:09 - INFO - parrotllm.training - step 35820 | epoch 0 | loss 3.4827 | lr 1.25e-03 | grad 0.2029 +2026-04-10 07:53:12 - INFO - parrotllm.training - step 35830 | epoch 0 | loss 3.4485 | lr 1.25e-03 | grad 0.2377 +2026-04-10 07:53:15 - INFO - parrotllm.training - step 35840 | epoch 0 | loss 3.4115 | lr 1.25e-03 | grad 0.2126 +2026-04-10 07:53:18 - INFO - parrotllm.training - step 35850 | epoch 0 | loss 3.4371 | lr 1.25e-03 | grad 0.2220 +2026-04-10 07:53:21 - INFO - parrotllm.training - step 35860 | epoch 0 | loss 3.5183 | lr 1.25e-03 | grad 0.1880 +2026-04-10 07:53:24 - INFO - parrotllm.training - step 35870 | epoch 0 | loss 3.5447 | lr 1.25e-03 | grad 0.2058 +2026-04-10 07:53:27 - INFO - parrotllm.training - step 35880 | epoch 0 | loss 3.4845 | lr 1.25e-03 | grad 0.2080 +2026-04-10 07:53:30 - INFO - parrotllm.training - step 35890 | epoch 0 | loss 3.4766 | lr 1.25e-03 | grad 0.2285 +2026-04-10 07:53:33 - INFO - parrotllm.training - step 35900 | epoch 0 | loss 3.5349 | lr 1.25e-03 | grad 0.2097 +2026-04-10 07:53:36 - INFO - parrotllm.training - step 35910 | epoch 0 | loss 3.3708 | lr 1.25e-03 | grad 0.2193 +2026-04-10 07:53:39 - INFO - parrotllm.training - step 35920 | epoch 0 | loss 3.4185 | lr 1.25e-03 | grad 0.2105 +2026-04-10 07:53:42 - INFO - parrotllm.training - step 35930 | epoch 0 | loss 3.3806 | lr 1.25e-03 | grad 0.2330 +2026-04-10 07:53:45 - INFO - parrotllm.training - step 35940 | epoch 0 | loss 3.3944 | lr 1.25e-03 | grad 0.2148 +2026-04-10 07:53:48 - INFO - parrotllm.training - step 35950 | epoch 0 | loss 3.4371 | lr 1.25e-03 | grad 0.2398 +2026-04-10 07:53:51 - INFO - parrotllm.training - step 35960 | epoch 0 | loss 3.4393 | lr 1.25e-03 | grad 0.2027 +2026-04-10 07:53:55 - INFO - parrotllm.training - step 35970 | epoch 0 | loss 3.3679 | lr 1.25e-03 | grad 0.2009 +2026-04-10 07:53:58 - INFO - parrotllm.training - step 35980 | epoch 0 | loss 3.5074 | lr 1.25e-03 | grad 0.2446 +2026-04-10 07:54:01 - INFO - parrotllm.training - step 35990 | epoch 0 | loss 3.4708 | lr 1.25e-03 | grad 0.2186 +2026-04-10 07:54:04 - INFO - parrotllm.training - step 36000 | epoch 0 | loss 3.4250 | lr 1.25e-03 | grad 0.1858 +2026-04-10 07:54:04 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:54:04 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:54:07 - INFO - parrotllm.training - Train: loss=3.4250, ppl=30.72 +2026-04-10 07:54:07 - INFO - parrotllm.training - Val: loss=3.3934, ppl=29.77 +2026-04-10 07:54:08 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3934_epoch_0000_step_0036000.pt +2026-04-10 07:54:09 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3904, min_delta=0.001000). +2026-04-10 07:54:09 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:54:12 - INFO - parrotllm.training - step 36010 | epoch 0 | loss 3.4734 | lr 1.25e-03 | grad 0.2228 +2026-04-10 07:54:15 - INFO - parrotllm.training - step 36020 | epoch 0 | loss 3.4437 | lr 1.24e-03 | grad 0.2115 +2026-04-10 07:54:18 - INFO - parrotllm.training - step 36030 | epoch 0 | loss 3.4205 | lr 1.24e-03 | grad 0.2602 +2026-04-10 07:54:21 - INFO - parrotllm.training - step 36040 | epoch 0 | loss 3.4727 | lr 1.24e-03 | grad 0.2525 +2026-04-10 07:54:24 - INFO - parrotllm.training - step 36050 | epoch 0 | loss 3.4022 | lr 1.24e-03 | grad 0.2032 +2026-04-10 07:54:27 - INFO - parrotllm.training - step 36060 | epoch 0 | loss 3.5656 | lr 1.24e-03 | grad 0.2176 +2026-04-10 07:54:30 - INFO - parrotllm.training - step 36070 | epoch 0 | loss 3.4482 | lr 1.24e-03 | grad 0.1935 +2026-04-10 07:54:33 - INFO - parrotllm.training - step 36080 | epoch 0 | loss 3.4612 | lr 1.24e-03 | grad 0.2242 +2026-04-10 07:54:36 - INFO - parrotllm.training - step 36090 | epoch 0 | loss 3.4969 | lr 1.24e-03 | grad 0.3709 +2026-04-10 07:54:39 - INFO - parrotllm.training - step 36100 | epoch 0 | loss 3.4713 | lr 1.24e-03 | grad 0.2130 +2026-04-10 07:54:43 - INFO - parrotllm.training - step 36110 | epoch 0 | loss 3.5152 | lr 1.24e-03 | grad 0.2284 +2026-04-10 07:54:46 - INFO - parrotllm.training - step 36120 | epoch 0 | loss 3.5250 | lr 1.24e-03 | grad 0.2035 +2026-04-10 07:54:49 - INFO - parrotllm.training - step 36130 | epoch 0 | loss 3.4894 | lr 1.24e-03 | grad 0.1981 +2026-04-10 07:54:52 - INFO - parrotllm.training - step 36140 | epoch 0 | loss 3.4294 | lr 1.24e-03 | grad 0.2303 +2026-04-10 07:54:55 - INFO - parrotllm.training - step 36150 | epoch 0 | loss 3.4456 | lr 1.24e-03 | grad 0.2221 +2026-04-10 07:54:58 - INFO - parrotllm.training - step 36160 | epoch 0 | loss 3.3738 | lr 1.24e-03 | grad 0.2001 +2026-04-10 07:55:01 - INFO - parrotllm.training - step 36170 | epoch 0 | loss 3.4115 | lr 1.24e-03 | grad 0.2580 +2026-04-10 07:55:04 - INFO - parrotllm.training - step 36180 | epoch 0 | loss 3.5199 | lr 1.24e-03 | grad 0.2252 +2026-04-10 07:55:07 - INFO - parrotllm.training - step 36190 | epoch 0 | loss 3.4564 | lr 1.24e-03 | grad 0.2101 +2026-04-10 07:55:10 - INFO - parrotllm.training - step 36200 | epoch 0 | loss 3.5634 | lr 1.24e-03 | grad 0.2224 +2026-04-10 07:55:13 - INFO - parrotllm.training - step 36210 | epoch 0 | loss 3.4203 | lr 1.24e-03 | grad 0.2173 +2026-04-10 07:55:16 - INFO - parrotllm.training - step 36220 | epoch 0 | loss 3.4991 | lr 1.24e-03 | grad 0.2046 +2026-04-10 07:55:19 - INFO - parrotllm.training - step 36230 | epoch 0 | loss 3.4463 | lr 1.24e-03 | grad 0.2323 +2026-04-10 07:55:22 - INFO - parrotllm.training - step 36240 | epoch 0 | loss 3.3570 | lr 1.24e-03 | grad 0.2408 +2026-04-10 07:55:25 - INFO - parrotllm.training - step 36250 | epoch 0 | loss 3.4978 | lr 1.24e-03 | grad 0.2003 +2026-04-10 07:55:28 - INFO - parrotllm.training - step 36260 | epoch 0 | loss 3.5155 | lr 1.24e-03 | grad 0.2131 +2026-04-10 07:55:31 - INFO - parrotllm.training - step 36270 | epoch 0 | loss 3.5117 | lr 1.24e-03 | grad 0.2244 +2026-04-10 07:55:34 - INFO - parrotllm.training - step 36280 | epoch 0 | loss 3.4881 | lr 1.24e-03 | grad 0.2052 +2026-04-10 07:55:37 - INFO - parrotllm.training - step 36290 | epoch 0 | loss 3.4716 | lr 1.24e-03 | grad 0.2086 +2026-04-10 07:55:40 - INFO - parrotllm.training - step 36300 | epoch 0 | loss 3.3983 | lr 1.24e-03 | grad 0.2088 +2026-04-10 07:55:44 - INFO - parrotllm.training - step 36310 | epoch 0 | loss 3.3950 | lr 1.24e-03 | grad 0.2178 +2026-04-10 07:55:47 - INFO - parrotllm.training - step 36320 | epoch 0 | loss 3.3073 | lr 1.24e-03 | grad 0.2162 +2026-04-10 07:55:50 - INFO - parrotllm.training - step 36330 | epoch 0 | loss 3.4402 | lr 1.24e-03 | grad 0.2230 +2026-04-10 07:55:53 - INFO - parrotllm.training - step 36340 | epoch 0 | loss 3.3807 | lr 1.24e-03 | grad 0.2262 +2026-04-10 07:55:56 - INFO - parrotllm.training - step 36350 | epoch 0 | loss 3.4500 | lr 1.24e-03 | grad 0.2242 +2026-04-10 07:55:59 - INFO - parrotllm.training - step 36360 | epoch 0 | loss 3.5010 | lr 1.24e-03 | grad 0.2373 +2026-04-10 07:56:02 - INFO - parrotllm.training - step 36370 | epoch 0 | loss 3.4157 | lr 1.24e-03 | grad 0.2490 +2026-04-10 07:56:05 - INFO - parrotllm.training - step 36380 | epoch 0 | loss 3.4378 | lr 1.24e-03 | grad 0.2156 +2026-04-10 07:56:08 - INFO - parrotllm.training - step 36390 | epoch 0 | loss 3.4456 | lr 1.24e-03 | grad 0.1980 +2026-04-10 07:56:11 - INFO - parrotllm.training - step 36400 | epoch 0 | loss 3.4493 | lr 1.24e-03 | grad 0.2174 +2026-04-10 07:56:14 - INFO - parrotllm.training - step 36410 | epoch 0 | loss 3.5763 | lr 1.24e-03 | grad 0.2081 +2026-04-10 07:56:17 - INFO - parrotllm.training - step 36420 | epoch 0 | loss 3.3999 | lr 1.24e-03 | grad 0.2299 +2026-04-10 07:56:20 - INFO - parrotllm.training - step 36430 | epoch 0 | loss 3.4392 | lr 1.24e-03 | grad 0.2332 +2026-04-10 07:56:23 - INFO - parrotllm.training - step 36440 | epoch 0 | loss 3.4815 | lr 1.24e-03 | grad 0.2028 +2026-04-10 07:56:26 - INFO - parrotllm.training - step 36450 | epoch 0 | loss 3.4691 | lr 1.24e-03 | grad 0.2053 +2026-04-10 07:56:29 - INFO - parrotllm.training - step 36460 | epoch 0 | loss 3.4345 | lr 1.24e-03 | grad 0.2258 +2026-04-10 07:56:32 - INFO - parrotllm.training - step 36470 | epoch 0 | loss 3.4856 | lr 1.24e-03 | grad 0.2008 +2026-04-10 07:56:35 - INFO - parrotllm.training - step 36480 | epoch 0 | loss 3.4576 | lr 1.24e-03 | grad 0.2060 +2026-04-10 07:56:38 - INFO - parrotllm.training - step 36490 | epoch 0 | loss 3.4434 | lr 1.24e-03 | grad 0.2092 +2026-04-10 07:56:41 - INFO - parrotllm.training - step 36500 | epoch 0 | loss 3.5011 | lr 1.24e-03 | grad 0.2337 +2026-04-10 07:56:41 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:56:41 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:56:45 - INFO - parrotllm.training - Train: loss=3.5011, ppl=33.15 +2026-04-10 07:56:45 - INFO - parrotllm.training - Val: loss=3.3889, ppl=29.63 +2026-04-10 07:56:45 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 07:56:45 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3889_epoch_0000_step_0036500.pt +2026-04-10 07:56:47 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:56:50 - INFO - parrotllm.training - step 36510 | epoch 0 | loss 3.5019 | lr 1.24e-03 | grad 0.2214 +2026-04-10 07:56:53 - INFO - parrotllm.training - step 36520 | epoch 0 | loss 3.6230 | lr 1.24e-03 | grad 0.2170 +2026-04-10 07:56:56 - INFO - parrotllm.training - step 36530 | epoch 0 | loss 3.3814 | lr 1.24e-03 | grad 0.2350 +2026-04-10 07:56:59 - INFO - parrotllm.training - step 36540 | epoch 0 | loss 3.3716 | lr 1.24e-03 | grad 0.2040 +2026-04-10 07:57:02 - INFO - parrotllm.training - step 36550 | epoch 0 | loss 3.4046 | lr 1.24e-03 | grad 0.2377 +2026-04-10 07:57:05 - INFO - parrotllm.training - step 36560 | epoch 0 | loss 3.4431 | lr 1.24e-03 | grad 0.2012 +2026-04-10 07:57:08 - INFO - parrotllm.training - step 36570 | epoch 0 | loss 3.5020 | lr 1.24e-03 | grad 0.2294 +2026-04-10 07:57:12 - INFO - parrotllm.training - step 36580 | epoch 0 | loss 3.4453 | lr 1.24e-03 | grad 0.2265 +2026-04-10 07:57:15 - INFO - parrotllm.training - step 36590 | epoch 0 | loss 3.3647 | lr 1.24e-03 | grad 0.2367 +2026-04-10 07:57:18 - INFO - parrotllm.training - step 36600 | epoch 0 | loss 3.4131 | lr 1.24e-03 | grad 0.2131 +2026-04-10 07:57:21 - INFO - parrotllm.training - step 36610 | epoch 0 | loss 3.4960 | lr 1.24e-03 | grad 0.2220 +2026-04-10 07:57:24 - INFO - parrotllm.training - step 36620 | epoch 0 | loss 3.4921 | lr 1.24e-03 | grad 0.1883 +2026-04-10 07:57:27 - INFO - parrotllm.training - step 36630 | epoch 0 | loss 3.4462 | lr 1.24e-03 | grad 0.1874 +2026-04-10 07:57:30 - INFO - parrotllm.training - step 36640 | epoch 0 | loss 3.4603 | lr 1.24e-03 | grad 0.2376 +2026-04-10 07:57:33 - INFO - parrotllm.training - step 36650 | epoch 0 | loss 3.4551 | lr 1.24e-03 | grad 0.2662 +2026-04-10 07:57:36 - INFO - parrotllm.training - step 36660 | epoch 0 | loss 3.3850 | lr 1.24e-03 | grad 0.2083 +2026-04-10 07:57:39 - INFO - parrotllm.training - step 36670 | epoch 0 | loss 3.4202 | lr 1.24e-03 | grad 0.2260 +2026-04-10 07:57:42 - INFO - parrotllm.training - step 36680 | epoch 0 | loss 3.4350 | lr 1.24e-03 | grad 0.2112 +2026-04-10 07:57:45 - INFO - parrotllm.training - step 36690 | epoch 0 | loss 3.3494 | lr 1.24e-03 | grad 0.2223 +2026-04-10 07:57:48 - INFO - parrotllm.training - step 36700 | epoch 0 | loss 3.4648 | lr 1.24e-03 | grad 0.2197 +2026-04-10 07:57:51 - INFO - parrotllm.training - step 36710 | epoch 0 | loss 3.4199 | lr 1.24e-03 | grad 0.2064 +2026-04-10 07:57:54 - INFO - parrotllm.training - step 36720 | epoch 0 | loss 3.3912 | lr 1.24e-03 | grad 0.1988 +2026-04-10 07:57:57 - INFO - parrotllm.training - step 36730 | epoch 0 | loss 3.3949 | lr 1.24e-03 | grad 0.2096 +2026-04-10 07:58:00 - INFO - parrotllm.training - step 36740 | epoch 0 | loss 3.4434 | lr 1.23e-03 | grad 0.2100 +2026-04-10 07:58:03 - INFO - parrotllm.training - step 36750 | epoch 0 | loss 3.2942 | lr 1.23e-03 | grad 0.2024 +2026-04-10 07:58:06 - INFO - parrotllm.training - step 36760 | epoch 0 | loss 3.4873 | lr 1.23e-03 | grad 0.2104 +2026-04-10 07:58:10 - INFO - parrotllm.training - step 36770 | epoch 0 | loss 3.4823 | lr 1.23e-03 | grad 0.1994 +2026-04-10 07:58:13 - INFO - parrotllm.training - step 36780 | epoch 0 | loss 3.4328 | lr 1.23e-03 | grad 0.2260 +2026-04-10 07:58:16 - INFO - parrotllm.training - step 36790 | epoch 0 | loss 3.4817 | lr 1.23e-03 | grad 0.2243 +2026-04-10 07:58:19 - INFO - parrotllm.training - step 36800 | epoch 0 | loss 3.3627 | lr 1.23e-03 | grad 0.2309 +2026-04-10 07:58:22 - INFO - parrotllm.training - step 36810 | epoch 0 | loss 3.4771 | lr 1.23e-03 | grad 0.2188 +2026-04-10 07:58:25 - INFO - parrotllm.training - step 36820 | epoch 0 | loss 3.4287 | lr 1.23e-03 | grad 0.2528 +2026-04-10 07:58:28 - INFO - parrotllm.training - step 36830 | epoch 0 | loss 3.5395 | lr 1.23e-03 | grad 0.2137 +2026-04-10 07:58:31 - INFO - parrotllm.training - step 36840 | epoch 0 | loss 3.4224 | lr 1.23e-03 | grad 0.2347 +2026-04-10 07:58:34 - INFO - parrotllm.training - step 36850 | epoch 0 | loss 3.4544 | lr 1.23e-03 | grad 0.2097 +2026-04-10 07:58:37 - INFO - parrotllm.training - step 36860 | epoch 0 | loss 3.4039 | lr 1.23e-03 | grad 0.2415 +2026-04-10 07:58:40 - INFO - parrotllm.training - step 36870 | epoch 0 | loss 3.3964 | lr 1.23e-03 | grad 0.2318 +2026-04-10 07:58:43 - INFO - parrotllm.training - step 36880 | epoch 0 | loss 3.4591 | lr 1.23e-03 | grad 0.2127 +2026-04-10 07:58:46 - INFO - parrotllm.training - step 36890 | epoch 0 | loss 3.4886 | lr 1.23e-03 | grad 0.2279 +2026-04-10 07:58:49 - INFO - parrotllm.training - step 36900 | epoch 0 | loss 3.4667 | lr 1.23e-03 | grad 0.2267 +2026-04-10 07:58:52 - INFO - parrotllm.training - step 36910 | epoch 0 | loss 3.3521 | lr 1.23e-03 | grad 0.2134 +2026-04-10 07:58:55 - INFO - parrotllm.training - step 36920 | epoch 0 | loss 3.3947 | lr 1.23e-03 | grad 0.2229 +2026-04-10 07:58:58 - INFO - parrotllm.training - step 36930 | epoch 0 | loss 3.5139 | lr 1.23e-03 | grad 0.2179 +2026-04-10 07:59:01 - INFO - parrotllm.training - step 36940 | epoch 0 | loss 3.3216 | lr 1.23e-03 | grad 0.2401 +2026-04-10 07:59:04 - INFO - parrotllm.training - step 36950 | epoch 0 | loss 3.5064 | lr 1.23e-03 | grad 0.2192 +2026-04-10 07:59:08 - INFO - parrotllm.training - step 36960 | epoch 0 | loss 3.4780 | lr 1.23e-03 | grad 0.2518 +2026-04-10 07:59:11 - INFO - parrotllm.training - step 36970 | epoch 0 | loss 3.4267 | lr 1.23e-03 | grad 0.2561 +2026-04-10 07:59:14 - INFO - parrotllm.training - step 36980 | epoch 0 | loss 3.5132 | lr 1.23e-03 | grad 0.2010 +2026-04-10 07:59:17 - INFO - parrotllm.training - step 36990 | epoch 0 | loss 3.5165 | lr 1.23e-03 | grad 0.2130 +2026-04-10 07:59:20 - INFO - parrotllm.training - step 37000 | epoch 0 | loss 3.4083 | lr 1.23e-03 | grad 0.1951 +2026-04-10 07:59:20 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 07:59:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:59:23 - INFO - parrotllm.training - Train: loss=3.4083, ppl=30.21 +2026-04-10 07:59:23 - INFO - parrotllm.training - Val: loss=3.3887, ppl=29.63 +2026-04-10 07:59:24 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3887_epoch_0000_step_0037000.pt +2026-04-10 07:59:25 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3889, min_delta=0.001000). +2026-04-10 07:59:25 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 07:59:28 - INFO - parrotllm.training - step 37010 | epoch 0 | loss 3.5348 | lr 1.23e-03 | grad 0.2528 +2026-04-10 07:59:31 - INFO - parrotllm.training - step 37020 | epoch 0 | loss 3.3558 | lr 1.23e-03 | grad 0.2320 +2026-04-10 07:59:34 - INFO - parrotllm.training - step 37030 | epoch 0 | loss 3.4616 | lr 1.23e-03 | grad 0.2061 +2026-04-10 07:59:37 - INFO - parrotllm.training - step 37040 | epoch 0 | loss 3.3936 | lr 1.23e-03 | grad 0.2816 +2026-04-10 07:59:40 - INFO - parrotllm.training - step 37050 | epoch 0 | loss 3.4399 | lr 1.23e-03 | grad 0.2409 +2026-04-10 07:59:43 - INFO - parrotllm.training - step 37060 | epoch 0 | loss 3.3852 | lr 1.23e-03 | grad 0.2476 +2026-04-10 07:59:46 - INFO - parrotllm.training - step 37070 | epoch 0 | loss 3.4514 | lr 1.23e-03 | grad 0.2062 +2026-04-10 07:59:50 - INFO - parrotllm.training - step 37080 | epoch 0 | loss 3.4457 | lr 1.23e-03 | grad 0.2349 +2026-04-10 07:59:53 - INFO - parrotllm.training - step 37090 | epoch 0 | loss 3.4739 | lr 1.23e-03 | grad 0.2140 +2026-04-10 07:59:56 - INFO - parrotllm.training - step 37100 | epoch 0 | loss 3.6276 | lr 1.23e-03 | grad 0.2163 +2026-04-10 07:59:59 - INFO - parrotllm.training - step 37110 | epoch 0 | loss 3.4859 | lr 1.23e-03 | grad 0.1932 +2026-04-10 08:00:02 - INFO - parrotllm.training - step 37120 | epoch 0 | loss 3.4282 | lr 1.23e-03 | grad 0.2498 +2026-04-10 08:00:05 - INFO - parrotllm.training - step 37130 | epoch 0 | loss 3.5554 | lr 1.23e-03 | grad 0.2320 +2026-04-10 08:00:08 - INFO - parrotllm.training - step 37140 | epoch 0 | loss 3.4864 | lr 1.23e-03 | grad 0.2381 +2026-04-10 08:00:11 - INFO - parrotllm.training - step 37150 | epoch 0 | loss 3.4139 | lr 1.23e-03 | grad 0.2352 +2026-04-10 08:00:14 - INFO - parrotllm.training - step 37160 | epoch 0 | loss 3.3575 | lr 1.23e-03 | grad 0.2197 +2026-04-10 08:00:17 - INFO - parrotllm.training - step 37170 | epoch 0 | loss 3.4995 | lr 1.23e-03 | grad 0.2533 +2026-04-10 08:00:20 - INFO - parrotllm.training - step 37180 | epoch 0 | loss 3.5762 | lr 1.23e-03 | grad 0.2246 +2026-04-10 08:00:23 - INFO - parrotllm.training - step 37190 | epoch 0 | loss 3.4489 | lr 1.23e-03 | grad 0.2011 +2026-04-10 08:00:26 - INFO - parrotllm.training - step 37200 | epoch 0 | loss 3.4831 | lr 1.23e-03 | grad 0.2056 +2026-04-10 08:00:29 - INFO - parrotllm.training - step 37210 | epoch 0 | loss 3.3861 | lr 1.23e-03 | grad 0.2131 +2026-04-10 08:00:32 - INFO - parrotllm.training - step 37220 | epoch 0 | loss 3.4140 | lr 1.23e-03 | grad 0.2037 +2026-04-10 08:00:35 - INFO - parrotllm.training - step 37230 | epoch 0 | loss 3.4256 | lr 1.23e-03 | grad 0.2022 +2026-04-10 08:00:38 - INFO - parrotllm.training - step 37240 | epoch 0 | loss 3.5773 | lr 1.23e-03 | grad 0.2142 +2026-04-10 08:00:41 - INFO - parrotllm.training - step 37250 | epoch 0 | loss 3.5068 | lr 1.23e-03 | grad 0.1997 +2026-04-10 08:00:44 - INFO - parrotllm.training - step 37260 | epoch 0 | loss 3.5106 | lr 1.23e-03 | grad 0.2909 +2026-04-10 08:00:48 - INFO - parrotllm.training - step 37270 | epoch 0 | loss 3.4135 | lr 1.23e-03 | grad 0.2394 +2026-04-10 08:00:51 - INFO - parrotllm.training - step 37280 | epoch 0 | loss 3.4054 | lr 1.23e-03 | grad 0.2257 +2026-04-10 08:00:51 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 37295/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 08:00:54 - INFO - parrotllm.training - step 37290 | epoch 0 | loss 3.4474 | lr 1.23e-03 | grad 0.2439 +2026-04-10 08:00:57 - INFO - parrotllm.training - step 37300 | epoch 0 | loss 3.3332 | lr 1.23e-03 | grad 0.2370 +2026-04-10 08:01:00 - INFO - parrotllm.training - step 37310 | epoch 0 | loss 3.4325 | lr 1.23e-03 | grad 0.2295 +2026-04-10 08:01:03 - INFO - parrotllm.training - step 37320 | epoch 0 | loss 3.4431 | lr 1.23e-03 | grad 0.2283 +2026-04-10 08:01:06 - INFO - parrotllm.training - step 37330 | epoch 0 | loss 3.4687 | lr 1.23e-03 | grad 0.2231 +2026-04-10 08:01:09 - INFO - parrotllm.training - step 37340 | epoch 0 | loss 3.4981 | lr 1.23e-03 | grad 0.2133 +2026-04-10 08:01:12 - INFO - parrotllm.training - step 37350 | epoch 0 | loss 3.4188 | lr 1.23e-03 | grad 0.2318 +2026-04-10 08:01:15 - INFO - parrotllm.training - step 37360 | epoch 0 | loss 3.4196 | lr 1.23e-03 | grad 0.2158 +2026-04-10 08:01:18 - INFO - parrotllm.training - step 37370 | epoch 0 | loss 3.4351 | lr 1.23e-03 | grad 0.2186 +2026-04-10 08:01:21 - INFO - parrotllm.training - step 37380 | epoch 0 | loss 3.4830 | lr 1.23e-03 | grad 0.2007 +2026-04-10 08:01:24 - INFO - parrotllm.training - step 37390 | epoch 0 | loss 3.3985 | lr 1.23e-03 | grad 0.2348 +2026-04-10 08:01:28 - INFO - parrotllm.training - step 37400 | epoch 0 | loss 3.5676 | lr 1.23e-03 | grad 0.2141 +2026-04-10 08:01:31 - INFO - parrotllm.training - step 37410 | epoch 0 | loss 3.4002 | lr 1.23e-03 | grad 0.2520 +2026-04-10 08:01:34 - INFO - parrotllm.training - step 37420 | epoch 0 | loss 3.3122 | lr 1.23e-03 | grad 0.2299 +2026-04-10 08:01:37 - INFO - parrotllm.training - step 37430 | epoch 0 | loss 3.3840 | lr 1.23e-03 | grad 0.2143 +2026-04-10 08:01:40 - INFO - parrotllm.training - step 37440 | epoch 0 | loss 3.3930 | lr 1.23e-03 | grad 0.2145 +2026-04-10 08:01:43 - INFO - parrotllm.training - step 37450 | epoch 0 | loss 3.3650 | lr 1.22e-03 | grad 0.2442 +2026-04-10 08:01:46 - INFO - parrotllm.training - step 37460 | epoch 0 | loss 3.4740 | lr 1.22e-03 | grad 0.1975 +2026-04-10 08:01:49 - INFO - parrotllm.training - step 37470 | epoch 0 | loss 3.4507 | lr 1.22e-03 | grad 0.2160 +2026-04-10 08:01:52 - INFO - parrotllm.training - step 37480 | epoch 0 | loss 3.5174 | lr 1.22e-03 | grad 0.1931 +2026-04-10 08:01:55 - INFO - parrotllm.training - step 37490 | epoch 0 | loss 3.4922 | lr 1.22e-03 | grad 0.1992 +2026-04-10 08:01:58 - INFO - parrotllm.training - step 37500 | epoch 0 | loss 3.4478 | lr 1.22e-03 | grad 0.2120 +2026-04-10 08:01:58 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:01:58 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:02:01 - INFO - parrotllm.training - Train: loss=3.4478, ppl=31.43 +2026-04-10 08:02:01 - INFO - parrotllm.training - Val: loss=3.3865, ppl=29.56 +2026-04-10 08:02:01 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 08:02:02 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3865_epoch_0000_step_0037500.pt +2026-04-10 08:02:04 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:02:05 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0037500.pt +2026-04-10 08:02:09 - INFO - parrotllm.training - step 37510 | epoch 0 | loss 3.5071 | lr 1.22e-03 | grad 0.2368 +2026-04-10 08:02:12 - INFO - parrotllm.training - step 37520 | epoch 0 | loss 3.4827 | lr 1.22e-03 | grad 0.2402 +2026-04-10 08:02:15 - INFO - parrotllm.training - step 37530 | epoch 0 | loss 3.4674 | lr 1.22e-03 | grad 0.2126 +2026-04-10 08:02:18 - INFO - parrotllm.training - step 37540 | epoch 0 | loss 3.3376 | lr 1.22e-03 | grad 0.2153 +2026-04-10 08:02:21 - INFO - parrotllm.training - step 37550 | epoch 0 | loss 3.3902 | lr 1.22e-03 | grad 0.2383 +2026-04-10 08:02:24 - INFO - parrotllm.training - step 37560 | epoch 0 | loss 3.3937 | lr 1.22e-03 | grad 0.2630 +2026-04-10 08:02:27 - INFO - parrotllm.training - step 37570 | epoch 0 | loss 3.4961 | lr 1.22e-03 | grad 0.2134 +2026-04-10 08:02:30 - INFO - parrotllm.training - step 37580 | epoch 0 | loss 3.5216 | lr 1.22e-03 | grad 0.2162 +2026-04-10 08:02:33 - INFO - parrotllm.training - step 37590 | epoch 0 | loss 3.3629 | lr 1.22e-03 | grad 0.2218 +2026-04-10 08:02:37 - INFO - parrotllm.training - step 37600 | epoch 0 | loss 3.3981 | lr 1.22e-03 | grad 0.2160 +2026-04-10 08:02:40 - INFO - parrotllm.training - step 37610 | epoch 0 | loss 3.4267 | lr 1.22e-03 | grad 0.2389 +2026-04-10 08:02:43 - INFO - parrotllm.training - step 37620 | epoch 0 | loss 3.4610 | lr 1.22e-03 | grad 0.2163 +2026-04-10 08:02:46 - INFO - parrotllm.training - step 37630 | epoch 0 | loss 3.4854 | lr 1.22e-03 | grad 0.2038 +2026-04-10 08:02:49 - INFO - parrotllm.training - step 37640 | epoch 0 | loss 3.4251 | lr 1.22e-03 | grad 0.2243 +2026-04-10 08:02:52 - INFO - parrotllm.training - step 37650 | epoch 0 | loss 3.4523 | lr 1.22e-03 | grad 0.2226 +2026-04-10 08:02:55 - INFO - parrotllm.training - step 37660 | epoch 0 | loss 3.4658 | lr 1.22e-03 | grad 0.2147 +2026-04-10 08:02:58 - INFO - parrotllm.training - step 37670 | epoch 0 | loss 3.5314 | lr 1.22e-03 | grad 0.2057 +2026-04-10 08:03:01 - INFO - parrotllm.training - step 37680 | epoch 0 | loss 3.4114 | lr 1.22e-03 | grad 0.2112 +2026-04-10 08:03:04 - INFO - parrotllm.training - step 37690 | epoch 0 | loss 3.4391 | lr 1.22e-03 | grad 0.2842 +2026-04-10 08:03:07 - INFO - parrotllm.training - step 37700 | epoch 0 | loss 3.4927 | lr 1.22e-03 | grad 0.2278 +2026-04-10 08:03:10 - INFO - parrotllm.training - step 37710 | epoch 0 | loss 3.3866 | lr 1.22e-03 | grad 0.2275 +2026-04-10 08:03:13 - INFO - parrotllm.training - step 37720 | epoch 0 | loss 3.4332 | lr 1.22e-03 | grad 0.2193 +2026-04-10 08:03:16 - INFO - parrotllm.training - step 37730 | epoch 0 | loss 3.5315 | lr 1.22e-03 | grad 0.2079 +2026-04-10 08:03:19 - INFO - parrotllm.training - step 37740 | epoch 0 | loss 3.4215 | lr 1.22e-03 | grad 0.2149 +2026-04-10 08:03:22 - INFO - parrotllm.training - step 37750 | epoch 0 | loss 3.4562 | lr 1.22e-03 | grad 0.2120 +2026-04-10 08:03:25 - INFO - parrotllm.training - step 37760 | epoch 0 | loss 3.5534 | lr 1.22e-03 | grad 0.2046 +2026-04-10 08:03:28 - INFO - parrotllm.training - step 37770 | epoch 0 | loss 3.5051 | lr 1.22e-03 | grad 0.2616 +2026-04-10 08:03:31 - INFO - parrotllm.training - step 37780 | epoch 0 | loss 3.4148 | lr 1.22e-03 | grad 0.2168 +2026-04-10 08:03:34 - INFO - parrotllm.training - step 37790 | epoch 0 | loss 3.4886 | lr 1.22e-03 | grad 0.2021 +2026-04-10 08:03:38 - INFO - parrotllm.training - step 37800 | epoch 0 | loss 3.3834 | lr 1.22e-03 | grad 0.2191 +2026-04-10 08:03:41 - INFO - parrotllm.training - step 37810 | epoch 0 | loss 3.4495 | lr 1.22e-03 | grad 0.2128 +2026-04-10 08:03:44 - INFO - parrotllm.training - step 37820 | epoch 0 | loss 3.4345 | lr 1.22e-03 | grad 0.2290 +2026-04-10 08:03:47 - INFO - parrotllm.training - step 37830 | epoch 0 | loss 3.4152 | lr 1.22e-03 | grad 0.1953 +2026-04-10 08:03:50 - INFO - parrotllm.training - step 37840 | epoch 0 | loss 3.4391 | lr 1.22e-03 | grad 0.2171 +2026-04-10 08:03:53 - INFO - parrotllm.training - step 37850 | epoch 0 | loss 3.5642 | lr 1.22e-03 | grad 0.2442 +2026-04-10 08:03:56 - INFO - parrotllm.training - step 37860 | epoch 0 | loss 3.3757 | lr 1.22e-03 | grad 0.2054 +2026-04-10 08:03:59 - INFO - parrotllm.training - step 37870 | epoch 0 | loss 3.4639 | lr 1.22e-03 | grad 0.2085 +2026-04-10 08:04:02 - INFO - parrotllm.training - step 37880 | epoch 0 | loss 3.4500 | lr 1.22e-03 | grad 0.1948 +2026-04-10 08:04:05 - INFO - parrotllm.training - step 37890 | epoch 0 | loss 3.4342 | lr 1.22e-03 | grad 0.2255 +2026-04-10 08:04:08 - INFO - parrotllm.training - step 37900 | epoch 0 | loss 3.4880 | lr 1.22e-03 | grad 0.2208 +2026-04-10 08:04:11 - INFO - parrotllm.training - step 37910 | epoch 0 | loss 3.3963 | lr 1.22e-03 | grad 0.2370 +2026-04-10 08:04:14 - INFO - parrotllm.training - step 37920 | epoch 0 | loss 3.4861 | lr 1.22e-03 | grad 0.2234 +2026-04-10 08:04:17 - INFO - parrotllm.training - step 37930 | epoch 0 | loss 3.4286 | lr 1.22e-03 | grad 0.1954 +2026-04-10 08:04:20 - INFO - parrotllm.training - step 37940 | epoch 0 | loss 3.4134 | lr 1.22e-03 | grad 0.2572 +2026-04-10 08:04:23 - INFO - parrotllm.training - step 37950 | epoch 0 | loss 3.3919 | lr 1.22e-03 | grad 0.2087 +2026-04-10 08:04:26 - INFO - parrotllm.training - step 37960 | epoch 0 | loss 3.3780 | lr 1.22e-03 | grad 0.2394 +2026-04-10 08:04:29 - INFO - parrotllm.training - step 37970 | epoch 0 | loss 3.4660 | lr 1.22e-03 | grad 0.2214 +2026-04-10 08:04:32 - INFO - parrotllm.training - step 37980 | epoch 0 | loss 3.4484 | lr 1.22e-03 | grad 0.2231 +2026-04-10 08:04:36 - INFO - parrotllm.training - step 37990 | epoch 0 | loss 3.5067 | lr 1.22e-03 | grad 0.2100 +2026-04-10 08:04:39 - INFO - parrotllm.training - step 38000 | epoch 0 | loss 3.4963 | lr 1.22e-03 | grad 0.2249 +2026-04-10 08:04:39 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:04:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:04:42 - INFO - parrotllm.training - Train: loss=3.4963, ppl=32.99 +2026-04-10 08:04:42 - INFO - parrotllm.training - Val: loss=3.3856, ppl=29.54 +2026-04-10 08:04:43 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3856_epoch_0000_step_0038000.pt +2026-04-10 08:04:44 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3865, min_delta=0.001000). +2026-04-10 08:04:44 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:04:47 - INFO - parrotllm.training - step 38010 | epoch 0 | loss 3.3801 | lr 1.22e-03 | grad 0.2278 +2026-04-10 08:04:50 - INFO - parrotllm.training - step 38020 | epoch 0 | loss 3.4813 | lr 1.22e-03 | grad 0.2322 +2026-04-10 08:04:53 - INFO - parrotllm.training - step 38030 | epoch 0 | loss 3.3789 | lr 1.22e-03 | grad 0.2534 +2026-04-10 08:04:56 - INFO - parrotllm.training - step 38040 | epoch 0 | loss 3.4875 | lr 1.22e-03 | grad 0.2161 +2026-04-10 08:04:59 - INFO - parrotllm.training - step 38050 | epoch 0 | loss 3.3702 | lr 1.22e-03 | grad 0.2289 +2026-04-10 08:05:02 - INFO - parrotllm.training - step 38060 | epoch 0 | loss 3.4625 | lr 1.22e-03 | grad 0.2329 +2026-04-10 08:05:05 - INFO - parrotllm.training - step 38070 | epoch 0 | loss 3.4367 | lr 1.22e-03 | grad 0.2354 +2026-04-10 08:05:08 - INFO - parrotllm.training - step 38080 | epoch 0 | loss 3.5397 | lr 1.22e-03 | grad 0.2181 +2026-04-10 08:05:11 - INFO - parrotllm.training - step 38090 | epoch 0 | loss 3.3462 | lr 1.22e-03 | grad 0.2492 +2026-04-10 08:05:14 - INFO - parrotllm.training - step 38100 | epoch 0 | loss 3.4644 | lr 1.22e-03 | grad 0.2851 +2026-04-10 08:05:18 - INFO - parrotllm.training - step 38110 | epoch 0 | loss 3.4546 | lr 1.22e-03 | grad 0.2431 +2026-04-10 08:05:21 - INFO - parrotllm.training - step 38120 | epoch 0 | loss 3.4012 | lr 1.22e-03 | grad 0.2298 +2026-04-10 08:05:24 - INFO - parrotllm.training - step 38130 | epoch 0 | loss 3.4773 | lr 1.22e-03 | grad 0.2228 +2026-04-10 08:05:27 - INFO - parrotllm.training - step 38140 | epoch 0 | loss 3.4641 | lr 1.22e-03 | grad 0.2147 +2026-04-10 08:05:30 - INFO - parrotllm.training - step 38150 | epoch 0 | loss 3.4465 | lr 1.21e-03 | grad 0.2097 +2026-04-10 08:05:33 - INFO - parrotllm.training - step 38160 | epoch 0 | loss 3.4983 | lr 1.21e-03 | grad 0.2220 +2026-04-10 08:05:36 - INFO - parrotllm.training - step 38170 | epoch 0 | loss 3.4142 | lr 1.21e-03 | grad 0.2941 +2026-04-10 08:05:39 - INFO - parrotllm.training - step 38180 | epoch 0 | loss 3.3401 | lr 1.21e-03 | grad 0.2164 +2026-04-10 08:05:42 - INFO - parrotllm.training - step 38190 | epoch 0 | loss 3.4864 | lr 1.21e-03 | grad 0.2135 +2026-04-10 08:05:45 - INFO - parrotllm.training - step 38200 | epoch 0 | loss 3.4966 | lr 1.21e-03 | grad 0.1991 +2026-04-10 08:05:48 - INFO - parrotllm.training - step 38210 | epoch 0 | loss 3.3781 | lr 1.21e-03 | grad 0.2260 +2026-04-10 08:05:51 - INFO - parrotllm.training - step 38220 | epoch 0 | loss 3.4147 | lr 1.21e-03 | grad 0.2285 +2026-04-10 08:05:54 - INFO - parrotllm.training - step 38230 | epoch 0 | loss 3.4171 | lr 1.21e-03 | grad 0.2100 +2026-04-10 08:05:57 - INFO - parrotllm.training - step 38240 | epoch 0 | loss 3.4063 | lr 1.21e-03 | grad 0.2031 +2026-04-10 08:06:00 - INFO - parrotllm.training - step 38250 | epoch 0 | loss 3.5478 | lr 1.21e-03 | grad 0.1996 +2026-04-10 08:06:03 - INFO - parrotllm.training - step 38260 | epoch 0 | loss 3.5040 | lr 1.21e-03 | grad 0.2498 +2026-04-10 08:06:06 - INFO - parrotllm.training - step 38270 | epoch 0 | loss 3.4516 | lr 1.21e-03 | grad 0.2233 +2026-04-10 08:06:09 - INFO - parrotllm.training - step 38280 | epoch 0 | loss 3.5397 | lr 1.21e-03 | grad 0.2476 +2026-04-10 08:06:12 - INFO - parrotllm.training - step 38290 | epoch 0 | loss 3.4770 | lr 1.21e-03 | grad 0.2463 +2026-04-10 08:06:15 - INFO - parrotllm.training - step 38300 | epoch 0 | loss 3.5149 | lr 1.21e-03 | grad 0.2389 +2026-04-10 08:06:19 - INFO - parrotllm.training - step 38310 | epoch 0 | loss 3.4415 | lr 1.21e-03 | grad 0.2293 +2026-04-10 08:06:22 - INFO - parrotllm.training - step 38320 | epoch 0 | loss 3.4359 | lr 1.21e-03 | grad 0.2178 +2026-04-10 08:06:25 - INFO - parrotllm.training - step 38330 | epoch 0 | loss 3.3893 | lr 1.21e-03 | grad 0.2277 +2026-04-10 08:06:28 - INFO - parrotllm.training - step 38340 | epoch 0 | loss 3.5292 | lr 1.21e-03 | grad 0.2670 +2026-04-10 08:06:31 - INFO - parrotllm.training - step 38350 | epoch 0 | loss 3.5199 | lr 1.21e-03 | grad 0.2201 +2026-04-10 08:06:34 - INFO - parrotllm.training - step 38360 | epoch 0 | loss 3.5467 | lr 1.21e-03 | grad 0.2112 +2026-04-10 08:06:37 - INFO - parrotllm.training - step 38370 | epoch 0 | loss 3.4387 | lr 1.21e-03 | grad 0.2294 +2026-04-10 08:06:40 - INFO - parrotllm.training - step 38380 | epoch 0 | loss 3.5007 | lr 1.21e-03 | grad 0.2266 +2026-04-10 08:06:43 - INFO - parrotllm.training - step 38390 | epoch 0 | loss 3.4530 | lr 1.21e-03 | grad 0.2274 +2026-04-10 08:06:46 - INFO - parrotllm.training - step 38400 | epoch 0 | loss 3.3996 | lr 1.21e-03 | grad 0.2449 +2026-04-10 08:06:49 - INFO - parrotllm.training - step 38410 | epoch 0 | loss 3.3178 | lr 1.21e-03 | grad 0.2207 +2026-04-10 08:06:52 - INFO - parrotllm.training - step 38420 | epoch 0 | loss 3.4046 | lr 1.21e-03 | grad 0.2105 +2026-04-10 08:06:55 - INFO - parrotllm.training - step 38430 | epoch 0 | loss 3.4241 | lr 1.21e-03 | grad 0.2251 +2026-04-10 08:06:58 - INFO - parrotllm.training - step 38440 | epoch 0 | loss 3.4783 | lr 1.21e-03 | grad 0.2202 +2026-04-10 08:07:01 - INFO - parrotllm.training - step 38450 | epoch 0 | loss 3.4630 | lr 1.21e-03 | grad 0.2155 +2026-04-10 08:07:04 - INFO - parrotllm.training - step 38460 | epoch 0 | loss 3.4377 | lr 1.21e-03 | grad 0.2366 +2026-04-10 08:07:07 - INFO - parrotllm.training - step 38470 | epoch 0 | loss 3.5062 | lr 1.21e-03 | grad 0.2310 +2026-04-10 08:07:10 - INFO - parrotllm.training - step 38480 | epoch 0 | loss 3.4909 | lr 1.21e-03 | grad 0.2180 +2026-04-10 08:07:13 - INFO - parrotllm.training - step 38490 | epoch 0 | loss 3.4580 | lr 1.21e-03 | grad 0.2345 +2026-04-10 08:07:17 - INFO - parrotllm.training - step 38500 | epoch 0 | loss 3.4290 | lr 1.21e-03 | grad 0.2285 +2026-04-10 08:07:17 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:07:17 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:07:20 - INFO - parrotllm.training - Train: loss=3.4290, ppl=30.85 +2026-04-10 08:07:20 - INFO - parrotllm.training - Val: loss=3.3835, ppl=29.47 +2026-04-10 08:07:20 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 08:07:21 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3835_epoch_0000_step_0038500.pt +2026-04-10 08:07:22 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:07:25 - INFO - parrotllm.training - step 38510 | epoch 0 | loss 3.4134 | lr 1.21e-03 | grad 0.2139 +2026-04-10 08:07:28 - INFO - parrotllm.training - step 38520 | epoch 0 | loss 3.4125 | lr 1.21e-03 | grad 0.2310 +2026-04-10 08:07:31 - INFO - parrotllm.training - step 38530 | epoch 0 | loss 3.4335 | lr 1.21e-03 | grad 0.2127 +2026-04-10 08:07:34 - INFO - parrotllm.training - step 38540 | epoch 0 | loss 3.4809 | lr 1.21e-03 | grad 0.2258 +2026-04-10 08:07:37 - INFO - parrotllm.training - step 38550 | epoch 0 | loss 3.4461 | lr 1.21e-03 | grad 0.2031 +2026-04-10 08:07:40 - INFO - parrotllm.training - step 38560 | epoch 0 | loss 3.3506 | lr 1.21e-03 | grad 0.2030 +2026-04-10 08:07:43 - INFO - parrotllm.training - step 38570 | epoch 0 | loss 3.4933 | lr 1.21e-03 | grad 0.2371 +2026-04-10 08:07:46 - INFO - parrotllm.training - step 38580 | epoch 0 | loss 3.4407 | lr 1.21e-03 | grad 0.2229 +2026-04-10 08:07:49 - INFO - parrotllm.training - step 38590 | epoch 0 | loss 3.3629 | lr 1.21e-03 | grad 0.2167 +2026-04-10 08:07:52 - INFO - parrotllm.training - step 38600 | epoch 0 | loss 3.4139 | lr 1.21e-03 | grad 0.2209 +2026-04-10 08:07:55 - INFO - parrotllm.training - step 38610 | epoch 0 | loss 3.3199 | lr 1.21e-03 | grad 0.2318 +2026-04-10 08:07:58 - INFO - parrotllm.training - step 38620 | epoch 0 | loss 3.3977 | lr 1.21e-03 | grad 0.1857 +2026-04-10 08:08:01 - INFO - parrotllm.training - step 38630 | epoch 0 | loss 3.4806 | lr 1.21e-03 | grad 0.2172 +2026-04-10 08:08:04 - INFO - parrotllm.training - step 38640 | epoch 0 | loss 3.3470 | lr 1.21e-03 | grad 0.2500 +2026-04-10 08:08:08 - INFO - parrotllm.training - step 38650 | epoch 0 | loss 3.4315 | lr 1.21e-03 | grad 0.2267 +2026-04-10 08:08:11 - INFO - parrotllm.training - step 38660 | epoch 0 | loss 3.3892 | lr 1.21e-03 | grad 0.2173 +2026-04-10 08:08:14 - INFO - parrotllm.training - step 38670 | epoch 0 | loss 3.4373 | lr 1.21e-03 | grad 0.2386 +2026-04-10 08:08:17 - INFO - parrotllm.training - step 38680 | epoch 0 | loss 3.4254 | lr 1.21e-03 | grad 0.2093 +2026-04-10 08:08:20 - INFO - parrotllm.training - step 38690 | epoch 0 | loss 3.4159 | lr 1.21e-03 | grad 0.2100 +2026-04-10 08:08:23 - INFO - parrotllm.training - step 38700 | epoch 0 | loss 3.4347 | lr 1.21e-03 | grad 0.2356 +2026-04-10 08:08:26 - INFO - parrotllm.training - step 38710 | epoch 0 | loss 3.4621 | lr 1.21e-03 | grad 0.2068 +2026-04-10 08:08:29 - INFO - parrotllm.training - step 38720 | epoch 0 | loss 3.3853 | lr 1.21e-03 | grad 0.2046 +2026-04-10 08:08:32 - INFO - parrotllm.training - step 38730 | epoch 0 | loss 3.3864 | lr 1.21e-03 | grad 0.2374 +2026-04-10 08:08:35 - INFO - parrotllm.training - step 38740 | epoch 0 | loss 3.3657 | lr 1.21e-03 | grad 0.2233 +2026-04-10 08:08:38 - INFO - parrotllm.training - step 38750 | epoch 0 | loss 3.4340 | lr 1.21e-03 | grad 0.2311 +2026-04-10 08:08:41 - INFO - parrotllm.training - step 38760 | epoch 0 | loss 3.4315 | lr 1.21e-03 | grad 0.2121 +2026-04-10 08:08:44 - INFO - parrotllm.training - step 38770 | epoch 0 | loss 3.4552 | lr 1.21e-03 | grad 0.2098 +2026-04-10 08:08:47 - INFO - parrotllm.training - step 38780 | epoch 0 | loss 3.5307 | lr 1.21e-03 | grad 0.2330 +2026-04-10 08:08:50 - INFO - parrotllm.training - step 38790 | epoch 0 | loss 3.4148 | lr 1.21e-03 | grad 0.2178 +2026-04-10 08:08:53 - INFO - parrotllm.training - step 38800 | epoch 0 | loss 3.4250 | lr 1.21e-03 | grad 0.2535 +2026-04-10 08:08:56 - INFO - parrotllm.training - step 38810 | epoch 0 | loss 3.4253 | lr 1.21e-03 | grad 0.2179 +2026-04-10 08:08:59 - INFO - parrotllm.training - step 38820 | epoch 0 | loss 3.4948 | lr 1.21e-03 | grad 0.2157 +2026-04-10 08:09:02 - INFO - parrotllm.training - step 38830 | epoch 0 | loss 3.4212 | lr 1.21e-03 | grad 0.1953 +2026-04-10 08:09:05 - INFO - parrotllm.training - step 38840 | epoch 0 | loss 3.4198 | lr 1.20e-03 | grad 0.2009 +2026-04-10 08:09:09 - INFO - parrotllm.training - step 38850 | epoch 0 | loss 3.4097 | lr 1.20e-03 | grad 0.2096 +2026-04-10 08:09:12 - INFO - parrotllm.training - step 38860 | epoch 0 | loss 3.4068 | lr 1.20e-03 | grad 0.2267 +2026-04-10 08:09:15 - INFO - parrotllm.training - step 38870 | epoch 0 | loss 3.2925 | lr 1.20e-03 | grad 0.2178 +2026-04-10 08:09:18 - INFO - parrotllm.training - step 38880 | epoch 0 | loss 3.5304 | lr 1.20e-03 | grad 0.2624 +2026-04-10 08:09:21 - INFO - parrotllm.training - step 38890 | epoch 0 | loss 3.4000 | lr 1.20e-03 | grad 0.2090 +2026-04-10 08:09:24 - INFO - parrotllm.training - step 38900 | epoch 0 | loss 3.5265 | lr 1.20e-03 | grad 0.2242 +2026-04-10 08:09:27 - INFO - parrotllm.training - step 38910 | epoch 0 | loss 3.4521 | lr 1.20e-03 | grad 0.2074 +2026-04-10 08:09:30 - INFO - parrotllm.training - step 38920 | epoch 0 | loss 3.3308 | lr 1.20e-03 | grad 0.2105 +2026-04-10 08:09:33 - INFO - parrotllm.training - step 38930 | epoch 0 | loss 3.4100 | lr 1.20e-03 | grad 0.2541 +2026-04-10 08:09:36 - INFO - parrotllm.training - step 38940 | epoch 0 | loss 3.3856 | lr 1.20e-03 | grad 0.2192 +2026-04-10 08:09:39 - INFO - parrotllm.training - step 38950 | epoch 0 | loss 3.4077 | lr 1.20e-03 | grad 0.2098 +2026-04-10 08:09:42 - INFO - parrotllm.training - step 38960 | epoch 0 | loss 3.3894 | lr 1.20e-03 | grad 0.2693 +2026-04-10 08:09:45 - INFO - parrotllm.training - step 38970 | epoch 0 | loss 3.3128 | lr 1.20e-03 | grad 0.2103 +2026-04-10 08:09:48 - INFO - parrotllm.training - step 38980 | epoch 0 | loss 3.4834 | lr 1.20e-03 | grad 0.2522 +2026-04-10 08:09:51 - INFO - parrotllm.training - step 38990 | epoch 0 | loss 3.4928 | lr 1.20e-03 | grad 0.2017 +2026-04-10 08:09:54 - INFO - parrotllm.training - step 39000 | epoch 0 | loss 3.4011 | lr 1.20e-03 | grad 0.2230 +2026-04-10 08:09:54 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:09:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:09:57 - INFO - parrotllm.training - Train: loss=3.4011, ppl=30.00 +2026-04-10 08:09:57 - INFO - parrotllm.training - Val: loss=3.3855, ppl=29.53 +2026-04-10 08:09:58 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3855_epoch_0000_step_0039000.pt +2026-04-10 08:10:00 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3835, min_delta=0.001000). +2026-04-10 08:10:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:10:03 - INFO - parrotllm.training - step 39010 | epoch 0 | loss 3.3789 | lr 1.20e-03 | grad 0.2078 +2026-04-10 08:10:06 - INFO - parrotllm.training - step 39020 | epoch 0 | loss 3.4450 | lr 1.20e-03 | grad 0.2008 +2026-04-10 08:10:09 - INFO - parrotllm.training - step 39030 | epoch 0 | loss 3.4666 | lr 1.20e-03 | grad 0.2268 +2026-04-10 08:10:12 - INFO - parrotllm.training - step 39040 | epoch 0 | loss 3.3966 | lr 1.20e-03 | grad 0.2102 +2026-04-10 08:10:15 - INFO - parrotllm.training - step 39050 | epoch 0 | loss 3.4147 | lr 1.20e-03 | grad 0.2466 +2026-04-10 08:10:18 - INFO - parrotllm.training - step 39060 | epoch 0 | loss 3.4168 | lr 1.20e-03 | grad 0.1952 +2026-04-10 08:10:21 - INFO - parrotllm.training - step 39070 | epoch 0 | loss 3.4904 | lr 1.20e-03 | grad 0.2533 +2026-04-10 08:10:24 - INFO - parrotllm.training - step 39080 | epoch 0 | loss 3.4792 | lr 1.20e-03 | grad 0.2484 +2026-04-10 08:10:27 - INFO - parrotllm.training - step 39090 | epoch 0 | loss 3.5370 | lr 1.20e-03 | grad 0.2194 +2026-04-10 08:10:30 - INFO - parrotllm.training - step 39100 | epoch 0 | loss 3.4927 | lr 1.20e-03 | grad 0.2313 +2026-04-10 08:10:33 - INFO - parrotllm.training - step 39110 | epoch 0 | loss 3.4930 | lr 1.20e-03 | grad 0.2226 +2026-04-10 08:10:36 - INFO - parrotllm.training - step 39120 | epoch 0 | loss 3.4287 | lr 1.20e-03 | grad 0.2091 +2026-04-10 08:10:39 - INFO - parrotllm.training - step 39130 | epoch 0 | loss 3.3947 | lr 1.20e-03 | grad 0.2106 +2026-04-10 08:10:42 - INFO - parrotllm.training - step 39140 | epoch 0 | loss 3.5247 | lr 1.20e-03 | grad 0.2146 +2026-04-10 08:10:45 - INFO - parrotllm.training - step 39150 | epoch 0 | loss 3.4088 | lr 1.20e-03 | grad 0.2521 +2026-04-10 08:10:48 - INFO - parrotllm.training - step 39160 | epoch 0 | loss 3.4565 | lr 1.20e-03 | grad 0.2192 +2026-04-10 08:10:51 - INFO - parrotllm.training - step 39170 | epoch 0 | loss 3.4346 | lr 1.20e-03 | grad 0.2199 +2026-04-10 08:10:55 - INFO - parrotllm.training - step 39180 | epoch 0 | loss 3.4757 | lr 1.20e-03 | grad 0.2453 +2026-04-10 08:10:58 - INFO - parrotllm.training - step 39190 | epoch 0 | loss 3.3739 | lr 1.20e-03 | grad 0.2213 +2026-04-10 08:11:01 - INFO - parrotllm.training - step 39200 | epoch 0 | loss 3.4019 | lr 1.20e-03 | grad 0.2239 +2026-04-10 08:11:04 - INFO - parrotllm.training - step 39210 | epoch 0 | loss 3.4430 | lr 1.20e-03 | grad 0.2157 +2026-04-10 08:11:07 - INFO - parrotllm.training - step 39220 | epoch 0 | loss 3.4921 | lr 1.20e-03 | grad 0.2557 +2026-04-10 08:11:10 - INFO - parrotllm.training - step 39230 | epoch 0 | loss 3.5245 | lr 1.20e-03 | grad 0.2281 +2026-04-10 08:11:13 - INFO - parrotllm.training - step 39240 | epoch 0 | loss 3.4541 | lr 1.20e-03 | grad 0.2533 +2026-04-10 08:11:16 - INFO - parrotllm.training - step 39250 | epoch 0 | loss 3.3997 | lr 1.20e-03 | grad 0.2184 +2026-04-10 08:11:19 - INFO - parrotllm.training - step 39260 | epoch 0 | loss 3.4643 | lr 1.20e-03 | grad 0.2306 +2026-04-10 08:11:22 - INFO - parrotllm.training - step 39270 | epoch 0 | loss 3.5362 | lr 1.20e-03 | grad 0.2146 +2026-04-10 08:11:25 - INFO - parrotllm.training - step 39280 | epoch 0 | loss 3.4487 | lr 1.20e-03 | grad 0.2209 +2026-04-10 08:11:28 - INFO - parrotllm.training - step 39290 | epoch 0 | loss 3.4558 | lr 1.20e-03 | grad 0.2456 +2026-04-10 08:11:31 - INFO - parrotllm.training - step 39300 | epoch 0 | loss 3.5266 | lr 1.20e-03 | grad 0.2081 +2026-04-10 08:11:34 - INFO - parrotllm.training - step 39310 | epoch 0 | loss 3.4426 | lr 1.20e-03 | grad 0.2087 +2026-04-10 08:11:37 - INFO - parrotllm.training - step 39320 | epoch 0 | loss 3.4352 | lr 1.20e-03 | grad 0.2177 +2026-04-10 08:11:40 - INFO - parrotllm.training - step 39330 | epoch 0 | loss 3.4839 | lr 1.20e-03 | grad 0.2277 +2026-04-10 08:11:43 - INFO - parrotllm.training - step 39340 | epoch 0 | loss 3.4892 | lr 1.20e-03 | grad 0.2062 +2026-04-10 08:11:46 - INFO - parrotllm.training - step 39350 | epoch 0 | loss 3.5090 | lr 1.20e-03 | grad 0.2219 +2026-04-10 08:11:49 - INFO - parrotllm.training - step 39360 | epoch 0 | loss 3.4443 | lr 1.20e-03 | grad 0.2373 +2026-04-10 08:11:52 - INFO - parrotllm.training - step 39370 | epoch 0 | loss 3.4173 | lr 1.20e-03 | grad 0.2375 +2026-04-10 08:11:55 - INFO - parrotllm.training - step 39380 | epoch 0 | loss 3.3580 | lr 1.20e-03 | grad 0.2020 +2026-04-10 08:11:59 - INFO - parrotllm.training - step 39390 | epoch 0 | loss 3.3810 | lr 1.20e-03 | grad 0.2136 +2026-04-10 08:12:02 - INFO - parrotllm.training - step 39400 | epoch 0 | loss 3.3896 | lr 1.20e-03 | grad 0.2089 +2026-04-10 08:12:05 - INFO - parrotllm.training - step 39410 | epoch 0 | loss 3.3615 | lr 1.20e-03 | grad 0.2257 +2026-04-10 08:12:08 - INFO - parrotllm.training - step 39420 | epoch 0 | loss 3.4442 | lr 1.20e-03 | grad 0.2366 +2026-04-10 08:12:11 - INFO - parrotllm.training - step 39430 | epoch 0 | loss 3.4417 | lr 1.20e-03 | grad 0.2275 +2026-04-10 08:12:14 - INFO - parrotllm.training - step 39440 | epoch 0 | loss 3.4494 | lr 1.20e-03 | grad 0.2050 +2026-04-10 08:12:17 - INFO - parrotllm.training - step 39450 | epoch 0 | loss 3.4024 | lr 1.20e-03 | grad 0.2281 +2026-04-10 08:12:20 - INFO - parrotllm.training - step 39460 | epoch 0 | loss 3.4477 | lr 1.20e-03 | grad 0.2353 +2026-04-10 08:12:23 - INFO - parrotllm.training - step 39470 | epoch 0 | loss 3.6049 | lr 1.20e-03 | grad 0.2204 +2026-04-10 08:12:26 - INFO - parrotllm.training - step 39480 | epoch 0 | loss 3.4511 | lr 1.20e-03 | grad 0.2176 +2026-04-10 08:12:29 - INFO - parrotllm.training - step 39490 | epoch 0 | loss 3.4331 | lr 1.20e-03 | grad 0.2126 +2026-04-10 08:12:32 - INFO - parrotllm.training - step 39500 | epoch 0 | loss 3.4598 | lr 1.20e-03 | grad 0.2071 +2026-04-10 08:12:32 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:12:32 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:12:35 - INFO - parrotllm.training - Train: loss=3.4598, ppl=31.81 +2026-04-10 08:12:35 - INFO - parrotllm.training - Val: loss=3.3836, ppl=29.48 +2026-04-10 08:12:36 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3836_epoch_0000_step_0039500.pt +2026-04-10 08:12:37 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3835, min_delta=0.001000). +2026-04-10 08:12:37 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:12:40 - INFO - parrotllm.training - step 39510 | epoch 0 | loss 3.4597 | lr 1.20e-03 | grad 0.2158 +2026-04-10 08:12:43 - INFO - parrotllm.training - step 39520 | epoch 0 | loss 3.4336 | lr 1.19e-03 | grad 0.2350 +2026-04-10 08:12:47 - INFO - parrotllm.training - step 39530 | epoch 0 | loss 3.4371 | lr 1.19e-03 | grad 0.2186 +2026-04-10 08:12:50 - INFO - parrotllm.training - step 39540 | epoch 0 | loss 3.4288 | lr 1.19e-03 | grad 0.2102 +2026-04-10 08:12:53 - INFO - parrotllm.training - step 39550 | epoch 0 | loss 3.5094 | lr 1.19e-03 | grad 0.2224 +2026-04-10 08:12:56 - INFO - parrotllm.training - step 39560 | epoch 0 | loss 3.3583 | lr 1.19e-03 | grad 0.2505 +2026-04-10 08:12:59 - INFO - parrotllm.training - step 39570 | epoch 0 | loss 3.2766 | lr 1.19e-03 | grad 0.2130 +2026-04-10 08:13:02 - INFO - parrotllm.training - step 39580 | epoch 0 | loss 3.3181 | lr 1.19e-03 | grad 0.2117 +2026-04-10 08:13:05 - INFO - parrotllm.training - step 39590 | epoch 0 | loss 3.5956 | lr 1.19e-03 | grad 0.2276 +2026-04-10 08:13:05 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 39607/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 08:13:08 - INFO - parrotllm.training - step 39600 | epoch 0 | loss 3.3991 | lr 1.19e-03 | grad 0.2291 +2026-04-10 08:13:11 - INFO - parrotllm.training - step 39610 | epoch 0 | loss 3.4381 | lr 1.19e-03 | grad 0.2254 +2026-04-10 08:13:14 - INFO - parrotllm.training - step 39620 | epoch 0 | loss 3.4645 | lr 1.19e-03 | grad 0.2148 +2026-04-10 08:13:17 - INFO - parrotllm.training - step 39630 | epoch 0 | loss 3.5039 | lr 1.19e-03 | grad 0.2290 +2026-04-10 08:13:20 - INFO - parrotllm.training - step 39640 | epoch 0 | loss 3.5349 | lr 1.19e-03 | grad 0.1989 +2026-04-10 08:13:23 - INFO - parrotllm.training - step 39650 | epoch 0 | loss 3.4384 | lr 1.19e-03 | grad 0.2230 +2026-04-10 08:13:26 - INFO - parrotllm.training - step 39660 | epoch 0 | loss 3.4321 | lr 1.19e-03 | grad 0.2017 +2026-04-10 08:13:30 - INFO - parrotllm.training - step 39670 | epoch 0 | loss 3.5015 | lr 1.19e-03 | grad 0.2132 +2026-04-10 08:13:33 - INFO - parrotllm.training - step 39680 | epoch 0 | loss 3.5259 | lr 1.19e-03 | grad 0.2258 +2026-04-10 08:13:36 - INFO - parrotllm.training - step 39690 | epoch 0 | loss 3.3933 | lr 1.19e-03 | grad 0.2238 +2026-04-10 08:13:39 - INFO - parrotllm.training - step 39700 | epoch 0 | loss 3.4642 | lr 1.19e-03 | grad 0.2636 +2026-04-10 08:13:42 - INFO - parrotllm.training - step 39710 | epoch 0 | loss 3.4243 | lr 1.19e-03 | grad 0.2067 +2026-04-10 08:13:45 - INFO - parrotllm.training - step 39720 | epoch 0 | loss 3.4125 | lr 1.19e-03 | grad 0.2174 +2026-04-10 08:13:48 - INFO - parrotllm.training - step 39730 | epoch 0 | loss 3.4559 | lr 1.19e-03 | grad 0.2357 +2026-04-10 08:13:51 - INFO - parrotllm.training - step 39740 | epoch 0 | loss 3.5229 | lr 1.19e-03 | grad 0.2234 +2026-04-10 08:13:54 - INFO - parrotllm.training - step 39750 | epoch 0 | loss 3.4043 | lr 1.19e-03 | grad 0.2329 +2026-04-10 08:13:57 - INFO - parrotllm.training - step 39760 | epoch 0 | loss 3.4996 | lr 1.19e-03 | grad 0.2463 +2026-04-10 08:14:00 - INFO - parrotllm.training - step 39770 | epoch 0 | loss 3.4203 | lr 1.19e-03 | grad 0.2182 +2026-04-10 08:14:03 - INFO - parrotllm.training - step 39780 | epoch 0 | loss 3.5198 | lr 1.19e-03 | grad 0.2256 +2026-04-10 08:14:06 - INFO - parrotllm.training - step 39790 | epoch 0 | loss 3.4389 | lr 1.19e-03 | grad 0.2083 +2026-04-10 08:14:09 - INFO - parrotllm.training - step 39800 | epoch 0 | loss 3.4495 | lr 1.19e-03 | grad 0.2353 +2026-04-10 08:14:12 - INFO - parrotllm.training - step 39810 | epoch 0 | loss 3.4982 | lr 1.19e-03 | grad 0.1961 +2026-04-10 08:14:15 - INFO - parrotllm.training - step 39820 | epoch 0 | loss 3.4548 | lr 1.19e-03 | grad 0.2104 +2026-04-10 08:14:18 - INFO - parrotllm.training - step 39830 | epoch 0 | loss 3.3969 | lr 1.19e-03 | grad 0.2373 +2026-04-10 08:14:21 - INFO - parrotllm.training - step 39840 | epoch 0 | loss 3.3784 | lr 1.19e-03 | grad 0.1977 +2026-04-10 08:14:24 - INFO - parrotllm.training - step 39850 | epoch 0 | loss 3.5239 | lr 1.19e-03 | grad 0.2060 +2026-04-10 08:14:27 - INFO - parrotllm.training - step 39860 | epoch 0 | loss 3.3825 | lr 1.19e-03 | grad 0.2231 +2026-04-10 08:14:30 - INFO - parrotllm.training - step 39870 | epoch 0 | loss 3.4198 | lr 1.19e-03 | grad 0.2163 +2026-04-10 08:14:34 - INFO - parrotllm.training - step 39880 | epoch 0 | loss 3.4499 | lr 1.19e-03 | grad 0.2377 +2026-04-10 08:14:37 - INFO - parrotllm.training - step 39890 | epoch 0 | loss 3.3998 | lr 1.19e-03 | grad 0.2363 +2026-04-10 08:14:40 - INFO - parrotllm.training - step 39900 | epoch 0 | loss 3.4183 | lr 1.19e-03 | grad 0.2121 +2026-04-10 08:14:43 - INFO - parrotllm.training - step 39910 | epoch 0 | loss 3.4046 | lr 1.19e-03 | grad 0.2211 +2026-04-10 08:14:46 - INFO - parrotllm.training - step 39920 | epoch 0 | loss 3.4167 | lr 1.19e-03 | grad 0.2240 +2026-04-10 08:14:49 - INFO - parrotllm.training - step 39930 | epoch 0 | loss 3.4018 | lr 1.19e-03 | grad 0.2399 +2026-04-10 08:14:52 - INFO - parrotllm.training - step 39940 | epoch 0 | loss 3.4814 | lr 1.19e-03 | grad 0.2317 +2026-04-10 08:14:55 - INFO - parrotllm.training - step 39950 | epoch 0 | loss 3.4392 | lr 1.19e-03 | grad 0.2143 +2026-04-10 08:14:58 - INFO - parrotllm.training - step 39960 | epoch 0 | loss 3.5167 | lr 1.19e-03 | grad 0.2406 +2026-04-10 08:15:01 - INFO - parrotllm.training - step 39970 | epoch 0 | loss 3.4535 | lr 1.19e-03 | grad 0.2362 +2026-04-10 08:15:04 - INFO - parrotllm.training - step 39980 | epoch 0 | loss 3.4282 | lr 1.19e-03 | grad 0.2241 +2026-04-10 08:15:07 - INFO - parrotllm.training - step 39990 | epoch 0 | loss 3.4144 | lr 1.19e-03 | grad 0.2158 +2026-04-10 08:15:10 - INFO - parrotllm.training - step 40000 | epoch 0 | loss 3.4842 | lr 1.19e-03 | grad 0.2134 +2026-04-10 08:15:10 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:15:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:15:13 - INFO - parrotllm.training - Train: loss=3.4842, ppl=32.60 +2026-04-10 08:15:13 - INFO - parrotllm.training - Val: loss=3.3810, ppl=29.40 +2026-04-10 08:15:13 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 08:15:14 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3810_epoch_0000_step_0040000.pt +2026-04-10 08:15:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:15:17 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0040000.pt +2026-04-10 08:15:21 - INFO - parrotllm.training - step 40010 | epoch 0 | loss 3.4471 | lr 1.19e-03 | grad 0.2366 +2026-04-10 08:15:24 - INFO - parrotllm.training - step 40020 | epoch 0 | loss 3.4943 | lr 1.19e-03 | grad 0.2694 +2026-04-10 08:15:28 - INFO - parrotllm.training - step 40030 | epoch 0 | loss 3.3613 | lr 1.19e-03 | grad 0.2217 +2026-04-10 08:15:31 - INFO - parrotllm.training - step 40040 | epoch 0 | loss 3.4202 | lr 1.19e-03 | grad 0.2057 +2026-04-10 08:15:34 - INFO - parrotllm.training - step 40050 | epoch 0 | loss 3.3541 | lr 1.19e-03 | grad 0.1964 +2026-04-10 08:15:37 - INFO - parrotllm.training - step 40060 | epoch 0 | loss 3.4243 | lr 1.19e-03 | grad 0.2009 +2026-04-10 08:15:40 - INFO - parrotllm.training - step 40070 | epoch 0 | loss 3.3749 | lr 1.19e-03 | grad 0.2198 +2026-04-10 08:15:43 - INFO - parrotllm.training - step 40080 | epoch 0 | loss 3.4388 | lr 1.19e-03 | grad 0.2076 +2026-04-10 08:15:46 - INFO - parrotllm.training - step 40090 | epoch 0 | loss 3.4168 | lr 1.19e-03 | grad 0.2116 +2026-04-10 08:15:49 - INFO - parrotllm.training - step 40100 | epoch 0 | loss 3.4570 | lr 1.19e-03 | grad 0.2140 +2026-04-10 08:15:52 - INFO - parrotllm.training - step 40110 | epoch 0 | loss 3.5185 | lr 1.19e-03 | grad 0.2388 +2026-04-10 08:15:55 - INFO - parrotllm.training - step 40120 | epoch 0 | loss 3.4460 | lr 1.19e-03 | grad 0.2135 +2026-04-10 08:15:58 - INFO - parrotllm.training - step 40130 | epoch 0 | loss 3.4310 | lr 1.19e-03 | grad 0.2427 +2026-04-10 08:16:01 - INFO - parrotllm.training - step 40140 | epoch 0 | loss 3.3304 | lr 1.19e-03 | grad 0.2445 +2026-04-10 08:16:04 - INFO - parrotllm.training - step 40150 | epoch 0 | loss 3.5323 | lr 1.19e-03 | grad 0.2269 +2026-04-10 08:16:07 - INFO - parrotllm.training - step 40160 | epoch 0 | loss 3.3800 | lr 1.19e-03 | grad 0.2042 +2026-04-10 08:16:10 - INFO - parrotllm.training - step 40170 | epoch 0 | loss 3.4518 | lr 1.19e-03 | grad 0.2496 +2026-04-10 08:16:13 - INFO - parrotllm.training - step 40180 | epoch 0 | loss 3.4491 | lr 1.19e-03 | grad 0.2257 +2026-04-10 08:16:16 - INFO - parrotllm.training - step 40190 | epoch 0 | loss 3.3558 | lr 1.19e-03 | grad 0.2079 +2026-04-10 08:16:19 - INFO - parrotllm.training - step 40200 | epoch 0 | loss 3.4036 | lr 1.18e-03 | grad 0.2096 +2026-04-10 08:16:22 - INFO - parrotllm.training - step 40210 | epoch 0 | loss 3.4165 | lr 1.18e-03 | grad 0.2272 +2026-04-10 08:16:26 - INFO - parrotllm.training - step 40220 | epoch 0 | loss 3.4199 | lr 1.18e-03 | grad 0.2099 +2026-04-10 08:16:29 - INFO - parrotllm.training - step 40230 | epoch 0 | loss 3.4678 | lr 1.18e-03 | grad 0.2482 +2026-04-10 08:16:32 - INFO - parrotllm.training - step 40240 | epoch 0 | loss 3.4261 | lr 1.18e-03 | grad 0.2520 +2026-04-10 08:16:35 - INFO - parrotllm.training - step 40250 | epoch 0 | loss 3.5827 | lr 1.18e-03 | grad 0.2374 +2026-04-10 08:16:38 - INFO - parrotllm.training - step 40260 | epoch 0 | loss 3.4327 | lr 1.18e-03 | grad 0.2342 +2026-04-10 08:16:41 - INFO - parrotllm.training - step 40270 | epoch 0 | loss 3.5413 | lr 1.18e-03 | grad 0.2412 +2026-04-10 08:16:44 - INFO - parrotllm.training - step 40280 | epoch 0 | loss 3.5525 | lr 1.18e-03 | grad 0.2360 +2026-04-10 08:16:47 - INFO - parrotllm.training - step 40290 | epoch 0 | loss 3.3787 | lr 1.18e-03 | grad 0.2092 +2026-04-10 08:16:50 - INFO - parrotllm.training - step 40300 | epoch 0 | loss 3.4391 | lr 1.18e-03 | grad 0.2412 +2026-04-10 08:16:53 - INFO - parrotllm.training - step 40310 | epoch 0 | loss 3.3829 | lr 1.18e-03 | grad 0.2242 +2026-04-10 08:16:56 - INFO - parrotllm.training - step 40320 | epoch 0 | loss 3.4664 | lr 1.18e-03 | grad 0.2182 +2026-04-10 08:16:59 - INFO - parrotllm.training - step 40330 | epoch 0 | loss 3.4505 | lr 1.18e-03 | grad 0.2073 +2026-04-10 08:17:02 - INFO - parrotllm.training - step 40340 | epoch 0 | loss 3.3633 | lr 1.18e-03 | grad 0.2290 +2026-04-10 08:17:05 - INFO - parrotllm.training - step 40350 | epoch 0 | loss 3.3761 | lr 1.18e-03 | grad 0.2375 +2026-04-10 08:17:08 - INFO - parrotllm.training - step 40360 | epoch 0 | loss 3.4624 | lr 1.18e-03 | grad 0.2172 +2026-04-10 08:17:11 - INFO - parrotllm.training - step 40370 | epoch 0 | loss 3.4535 | lr 1.18e-03 | grad 0.2355 +2026-04-10 08:17:14 - INFO - parrotllm.training - step 40380 | epoch 0 | loss 3.3586 | lr 1.18e-03 | grad 0.2436 +2026-04-10 08:17:17 - INFO - parrotllm.training - step 40390 | epoch 0 | loss 3.3507 | lr 1.18e-03 | grad 0.1995 +2026-04-10 08:17:21 - INFO - parrotllm.training - step 40400 | epoch 0 | loss 3.4581 | lr 1.18e-03 | grad 0.2518 +2026-04-10 08:17:24 - INFO - parrotllm.training - step 40410 | epoch 0 | loss 3.3690 | lr 1.18e-03 | grad 0.2637 +2026-04-10 08:17:27 - INFO - parrotllm.training - step 40420 | epoch 0 | loss 3.4457 | lr 1.18e-03 | grad 0.2133 +2026-04-10 08:17:30 - INFO - parrotllm.training - step 40430 | epoch 0 | loss 3.4657 | lr 1.18e-03 | grad 0.2353 +2026-04-10 08:17:33 - INFO - parrotllm.training - step 40440 | epoch 0 | loss 3.3464 | lr 1.18e-03 | grad 0.2142 +2026-04-10 08:17:36 - INFO - parrotllm.training - step 40450 | epoch 0 | loss 3.4702 | lr 1.18e-03 | grad 0.2120 +2026-04-10 08:17:39 - INFO - parrotllm.training - step 40460 | epoch 0 | loss 3.5204 | lr 1.18e-03 | grad 0.2095 +2026-04-10 08:17:42 - INFO - parrotllm.training - step 40470 | epoch 0 | loss 3.4196 | lr 1.18e-03 | grad 0.2585 +2026-04-10 08:17:45 - INFO - parrotllm.training - step 40480 | epoch 0 | loss 3.5043 | lr 1.18e-03 | grad 0.2094 +2026-04-10 08:17:48 - INFO - parrotllm.training - step 40490 | epoch 0 | loss 3.5146 | lr 1.18e-03 | grad 0.2302 +2026-04-10 08:17:51 - INFO - parrotllm.training - step 40500 | epoch 0 | loss 3.5051 | lr 1.18e-03 | grad 0.2513 +2026-04-10 08:17:51 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:17:51 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:17:54 - INFO - parrotllm.training - Train: loss=3.5051, ppl=33.29 +2026-04-10 08:17:54 - INFO - parrotllm.training - Val: loss=3.3833, ppl=29.47 +2026-04-10 08:17:55 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3833_epoch_0000_step_0040500.pt +2026-04-10 08:17:57 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3810, min_delta=0.001000). +2026-04-10 08:17:57 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:18:00 - INFO - parrotllm.training - step 40510 | epoch 0 | loss 3.3960 | lr 1.18e-03 | grad 0.2226 +2026-04-10 08:18:03 - INFO - parrotllm.training - step 40520 | epoch 0 | loss 3.4900 | lr 1.18e-03 | grad 0.2156 +2026-04-10 08:18:06 - INFO - parrotllm.training - step 40530 | epoch 0 | loss 3.4323 | lr 1.18e-03 | grad 0.2273 +2026-04-10 08:18:09 - INFO - parrotllm.training - step 40540 | epoch 0 | loss 3.4926 | lr 1.18e-03 | grad 0.2202 +2026-04-10 08:18:12 - INFO - parrotllm.training - step 40550 | epoch 0 | loss 3.3974 | lr 1.18e-03 | grad 0.2093 +2026-04-10 08:18:15 - INFO - parrotllm.training - step 40560 | epoch 0 | loss 3.3555 | lr 1.18e-03 | grad 0.2163 +2026-04-10 08:18:18 - INFO - parrotllm.training - step 40570 | epoch 0 | loss 3.4707 | lr 1.18e-03 | grad 0.2108 +2026-04-10 08:18:21 - INFO - parrotllm.training - step 40580 | epoch 0 | loss 3.4782 | lr 1.18e-03 | grad 0.2156 +2026-04-10 08:18:24 - INFO - parrotllm.training - step 40590 | epoch 0 | loss 3.4886 | lr 1.18e-03 | grad 0.2670 +2026-04-10 08:18:27 - INFO - parrotllm.training - step 40600 | epoch 0 | loss 3.3341 | lr 1.18e-03 | grad 0.2327 +2026-04-10 08:18:30 - INFO - parrotllm.training - step 40610 | epoch 0 | loss 3.3503 | lr 1.18e-03 | grad 0.2145 +2026-04-10 08:18:33 - INFO - parrotllm.training - step 40620 | epoch 0 | loss 3.4319 | lr 1.18e-03 | grad 0.2255 +2026-04-10 08:18:36 - INFO - parrotllm.training - step 40630 | epoch 0 | loss 3.4154 | lr 1.18e-03 | grad 0.2326 +2026-04-10 08:18:39 - INFO - parrotllm.training - step 40640 | epoch 0 | loss 3.4271 | lr 1.18e-03 | grad 0.2244 +2026-04-10 08:18:42 - INFO - parrotllm.training - step 40650 | epoch 0 | loss 3.3862 | lr 1.18e-03 | grad 0.2270 +2026-04-10 08:18:45 - INFO - parrotllm.training - step 40660 | epoch 0 | loss 3.4305 | lr 1.18e-03 | grad 0.2067 +2026-04-10 08:18:49 - INFO - parrotllm.training - step 40670 | epoch 0 | loss 3.4615 | lr 1.18e-03 | grad 0.2175 +2026-04-10 08:18:52 - INFO - parrotllm.training - step 40680 | epoch 0 | loss 3.5397 | lr 1.18e-03 | grad 0.2817 +2026-04-10 08:18:55 - INFO - parrotllm.training - step 40690 | epoch 0 | loss 3.3787 | lr 1.18e-03 | grad 0.2152 +2026-04-10 08:18:58 - INFO - parrotllm.training - step 40700 | epoch 0 | loss 3.4852 | lr 1.18e-03 | grad 0.2315 +2026-04-10 08:19:01 - INFO - parrotllm.training - step 40710 | epoch 0 | loss 3.3762 | lr 1.18e-03 | grad 0.2102 +2026-04-10 08:19:04 - INFO - parrotllm.training - step 40720 | epoch 0 | loss 3.4472 | lr 1.18e-03 | grad 0.3017 +2026-04-10 08:19:07 - INFO - parrotllm.training - step 40730 | epoch 0 | loss 3.5170 | lr 1.18e-03 | grad 0.2223 +2026-04-10 08:19:10 - INFO - parrotllm.training - step 40740 | epoch 0 | loss 3.3951 | lr 1.18e-03 | grad 0.2225 +2026-04-10 08:19:13 - INFO - parrotllm.training - step 40750 | epoch 0 | loss 3.4710 | lr 1.18e-03 | grad 0.2216 +2026-04-10 08:19:16 - INFO - parrotllm.training - step 40760 | epoch 0 | loss 3.4189 | lr 1.18e-03 | grad 0.2275 +2026-04-10 08:19:19 - INFO - parrotllm.training - step 40770 | epoch 0 | loss 3.4577 | lr 1.18e-03 | grad 0.2069 +2026-04-10 08:19:22 - INFO - parrotllm.training - step 40780 | epoch 0 | loss 3.3522 | lr 1.18e-03 | grad 0.2411 +2026-04-10 08:19:25 - INFO - parrotllm.training - step 40790 | epoch 0 | loss 3.3541 | lr 1.18e-03 | grad 0.2105 +2026-04-10 08:19:28 - INFO - parrotllm.training - step 40800 | epoch 0 | loss 3.4699 | lr 1.18e-03 | grad 0.2175 +2026-04-10 08:19:31 - INFO - parrotllm.training - step 40810 | epoch 0 | loss 3.3999 | lr 1.18e-03 | grad 0.2022 +2026-04-10 08:19:34 - INFO - parrotllm.training - step 40820 | epoch 0 | loss 3.3666 | lr 1.18e-03 | grad 0.2555 +2026-04-10 08:19:37 - INFO - parrotllm.training - step 40830 | epoch 0 | loss 3.4758 | lr 1.18e-03 | grad 0.2001 +2026-04-10 08:19:40 - INFO - parrotllm.training - step 40840 | epoch 0 | loss 3.4313 | lr 1.18e-03 | grad 0.2327 +2026-04-10 08:19:43 - INFO - parrotllm.training - step 40850 | epoch 0 | loss 3.4178 | lr 1.18e-03 | grad 0.2182 +2026-04-10 08:19:47 - INFO - parrotllm.training - step 40860 | epoch 0 | loss 3.3880 | lr 1.18e-03 | grad 0.2334 +2026-04-10 08:19:50 - INFO - parrotllm.training - step 40870 | epoch 0 | loss 3.3776 | lr 1.17e-03 | grad 0.2258 +2026-04-10 08:19:53 - INFO - parrotllm.training - step 40880 | epoch 0 | loss 3.4566 | lr 1.17e-03 | grad 0.2306 +2026-04-10 08:19:56 - INFO - parrotllm.training - step 40890 | epoch 0 | loss 3.5177 | lr 1.17e-03 | grad 0.2058 +2026-04-10 08:19:59 - INFO - parrotllm.training - step 40900 | epoch 0 | loss 3.5045 | lr 1.17e-03 | grad 0.2299 +2026-04-10 08:20:02 - INFO - parrotllm.training - step 40910 | epoch 0 | loss 3.4773 | lr 1.17e-03 | grad 0.2576 +2026-04-10 08:20:05 - INFO - parrotllm.training - step 40920 | epoch 0 | loss 3.3703 | lr 1.17e-03 | grad 0.2208 +2026-04-10 08:20:08 - INFO - parrotllm.training - step 40930 | epoch 0 | loss 3.4975 | lr 1.17e-03 | grad 0.2070 +2026-04-10 08:20:11 - INFO - parrotllm.training - step 40940 | epoch 0 | loss 3.4746 | lr 1.17e-03 | grad 0.2144 +2026-04-10 08:20:14 - INFO - parrotllm.training - step 40950 | epoch 0 | loss 3.4302 | lr 1.17e-03 | grad 0.2397 +2026-04-10 08:20:17 - INFO - parrotllm.training - step 40960 | epoch 0 | loss 3.4384 | lr 1.17e-03 | grad 0.2448 +2026-04-10 08:20:20 - INFO - parrotllm.training - step 40970 | epoch 0 | loss 3.4100 | lr 1.17e-03 | grad 0.2623 +2026-04-10 08:20:23 - INFO - parrotllm.training - step 40980 | epoch 0 | loss 3.5054 | lr 1.17e-03 | grad 0.2316 +2026-04-10 08:20:26 - INFO - parrotllm.training - step 40990 | epoch 0 | loss 3.3659 | lr 1.17e-03 | grad 0.2020 +2026-04-10 08:20:29 - INFO - parrotllm.training - step 41000 | epoch 0 | loss 3.4572 | lr 1.17e-03 | grad 0.2619 +2026-04-10 08:20:29 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:20:29 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:20:32 - INFO - parrotllm.training - Train: loss=3.4572, ppl=31.73 +2026-04-10 08:20:32 - INFO - parrotllm.training - Val: loss=3.3814, ppl=29.41 +2026-04-10 08:20:33 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3814_epoch_0000_step_0041000.pt +2026-04-10 08:20:35 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3810, min_delta=0.001000). +2026-04-10 08:20:35 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:20:38 - INFO - parrotllm.training - step 41010 | epoch 0 | loss 3.4384 | lr 1.17e-03 | grad 0.2148 +2026-04-10 08:20:41 - INFO - parrotllm.training - step 41020 | epoch 0 | loss 3.5363 | lr 1.17e-03 | grad 0.2334 +2026-04-10 08:20:44 - INFO - parrotllm.training - step 41030 | epoch 0 | loss 3.4883 | lr 1.17e-03 | grad 0.2422 +2026-04-10 08:20:47 - INFO - parrotllm.training - step 41040 | epoch 0 | loss 3.5179 | lr 1.17e-03 | grad 0.2318 +2026-04-10 08:20:50 - INFO - parrotllm.training - step 41050 | epoch 0 | loss 3.3691 | lr 1.17e-03 | grad 0.2298 +2026-04-10 08:20:53 - INFO - parrotllm.training - step 41060 | epoch 0 | loss 3.4971 | lr 1.17e-03 | grad 0.2237 +2026-04-10 08:20:56 - INFO - parrotllm.training - step 41070 | epoch 0 | loss 3.5103 | lr 1.17e-03 | grad 0.2251 +2026-04-10 08:20:59 - INFO - parrotllm.training - step 41080 | epoch 0 | loss 3.4082 | lr 1.17e-03 | grad 0.2637 +2026-04-10 08:21:02 - INFO - parrotllm.training - step 41090 | epoch 0 | loss 3.3592 | lr 1.17e-03 | grad 0.2310 +2026-04-10 08:21:05 - INFO - parrotllm.training - step 41100 | epoch 0 | loss 3.5107 | lr 1.17e-03 | grad 0.2339 +2026-04-10 08:21:08 - INFO - parrotllm.training - step 41110 | epoch 0 | loss 3.4883 | lr 1.17e-03 | grad 0.2507 +2026-04-10 08:21:11 - INFO - parrotllm.training - step 41120 | epoch 0 | loss 3.4486 | lr 1.17e-03 | grad 0.2284 +2026-04-10 08:21:14 - INFO - parrotllm.training - step 41130 | epoch 0 | loss 3.3401 | lr 1.17e-03 | grad 0.2257 +2026-04-10 08:21:17 - INFO - parrotllm.training - step 41140 | epoch 0 | loss 3.6652 | lr 1.17e-03 | grad 0.2133 +2026-04-10 08:21:20 - INFO - parrotllm.training - step 41150 | epoch 0 | loss 3.4789 | lr 1.17e-03 | grad 0.2085 +2026-04-10 08:21:24 - INFO - parrotllm.training - step 41160 | epoch 0 | loss 3.5291 | lr 1.17e-03 | grad 0.2253 +2026-04-10 08:21:27 - INFO - parrotllm.training - step 41170 | epoch 0 | loss 3.4381 | lr 1.17e-03 | grad 0.2303 +2026-04-10 08:21:30 - INFO - parrotllm.training - step 41180 | epoch 0 | loss 3.5943 | lr 1.17e-03 | grad 0.2518 +2026-04-10 08:21:33 - INFO - parrotllm.training - step 41190 | epoch 0 | loss 3.4402 | lr 1.17e-03 | grad 0.2207 +2026-04-10 08:21:36 - INFO - parrotllm.training - step 41200 | epoch 0 | loss 3.5296 | lr 1.17e-03 | grad 0.2202 +2026-04-10 08:21:39 - INFO - parrotllm.training - step 41210 | epoch 0 | loss 3.4187 | lr 1.17e-03 | grad 0.2315 +2026-04-10 08:21:42 - INFO - parrotllm.training - step 41220 | epoch 0 | loss 3.4289 | lr 1.17e-03 | grad 0.2056 +2026-04-10 08:21:45 - INFO - parrotllm.training - step 41230 | epoch 0 | loss 3.5649 | lr 1.17e-03 | grad 0.2105 +2026-04-10 08:21:48 - INFO - parrotllm.training - step 41240 | epoch 0 | loss 3.4607 | lr 1.17e-03 | grad 0.2540 +2026-04-10 08:21:51 - INFO - parrotllm.training - step 41250 | epoch 0 | loss 3.4642 | lr 1.17e-03 | grad 0.2396 +2026-04-10 08:21:54 - INFO - parrotllm.training - step 41260 | epoch 0 | loss 3.4486 | lr 1.17e-03 | grad 0.2221 +2026-04-10 08:21:57 - INFO - parrotllm.training - step 41270 | epoch 0 | loss 3.3914 | lr 1.17e-03 | grad 0.1904 +2026-04-10 08:22:00 - INFO - parrotllm.training - step 41280 | epoch 0 | loss 3.3717 | lr 1.17e-03 | grad 0.2250 +2026-04-10 08:22:03 - INFO - parrotllm.training - step 41290 | epoch 0 | loss 3.3743 | lr 1.17e-03 | grad 0.2028 +2026-04-10 08:22:06 - INFO - parrotllm.training - step 41300 | epoch 0 | loss 3.4628 | lr 1.17e-03 | grad 0.2395 +2026-04-10 08:22:09 - INFO - parrotllm.training - step 41310 | epoch 0 | loss 3.3697 | lr 1.17e-03 | grad 0.2238 +2026-04-10 08:22:12 - INFO - parrotllm.training - step 41320 | epoch 0 | loss 3.4830 | lr 1.17e-03 | grad 0.2114 +2026-04-10 08:22:15 - INFO - parrotllm.training - step 41330 | epoch 0 | loss 3.4107 | lr 1.17e-03 | grad 0.2092 +2026-04-10 08:22:18 - INFO - parrotllm.training - step 41340 | epoch 0 | loss 3.4595 | lr 1.17e-03 | grad 0.2590 +2026-04-10 08:22:22 - INFO - parrotllm.training - step 41350 | epoch 0 | loss 3.4566 | lr 1.17e-03 | grad 0.2437 +2026-04-10 08:22:25 - INFO - parrotllm.training - step 41360 | epoch 0 | loss 3.4250 | lr 1.17e-03 | grad 0.2309 +2026-04-10 08:22:28 - INFO - parrotllm.training - step 41370 | epoch 0 | loss 3.4872 | lr 1.17e-03 | grad 0.2127 +2026-04-10 08:22:31 - INFO - parrotllm.training - step 41380 | epoch 0 | loss 3.4967 | lr 1.17e-03 | grad 0.2494 +2026-04-10 08:22:34 - INFO - parrotllm.training - step 41390 | epoch 0 | loss 3.5244 | lr 1.17e-03 | grad 0.2433 +2026-04-10 08:22:37 - INFO - parrotllm.training - step 41400 | epoch 0 | loss 3.4118 | lr 1.17e-03 | grad 0.2063 +2026-04-10 08:22:40 - INFO - parrotllm.training - step 41410 | epoch 0 | loss 3.4192 | lr 1.17e-03 | grad 0.2198 +2026-04-10 08:22:43 - INFO - parrotllm.training - step 41420 | epoch 0 | loss 3.5108 | lr 1.17e-03 | grad 0.2066 +2026-04-10 08:22:46 - INFO - parrotllm.training - step 41430 | epoch 0 | loss 3.5150 | lr 1.17e-03 | grad 0.2340 +2026-04-10 08:22:49 - INFO - parrotllm.training - step 41440 | epoch 0 | loss 3.4030 | lr 1.17e-03 | grad 0.2002 +2026-04-10 08:22:52 - INFO - parrotllm.training - step 41450 | epoch 0 | loss 3.3728 | lr 1.17e-03 | grad 0.2248 +2026-04-10 08:22:55 - INFO - parrotllm.training - step 41460 | epoch 0 | loss 3.3465 | lr 1.17e-03 | grad 0.2321 +2026-04-10 08:22:58 - INFO - parrotllm.training - step 41470 | epoch 0 | loss 3.3248 | lr 1.17e-03 | grad 0.2465 +2026-04-10 08:23:01 - INFO - parrotllm.training - step 41480 | epoch 0 | loss 3.3379 | lr 1.17e-03 | grad 0.2282 +2026-04-10 08:23:04 - INFO - parrotllm.training - step 41490 | epoch 0 | loss 3.4832 | lr 1.17e-03 | grad 0.2397 +2026-04-10 08:23:07 - INFO - parrotllm.training - step 41500 | epoch 0 | loss 3.4186 | lr 1.17e-03 | grad 0.2201 +2026-04-10 08:23:07 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:23:07 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:23:10 - INFO - parrotllm.training - Train: loss=3.4186, ppl=30.53 +2026-04-10 08:23:10 - INFO - parrotllm.training - Val: loss=3.3808, ppl=29.39 +2026-04-10 08:23:11 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3808_epoch_0000_step_0041500.pt +2026-04-10 08:23:13 - INFO - parrotllm.training - No validation improvement for 3/15 evaluation(s) (best=3.3810, min_delta=0.001000). +2026-04-10 08:23:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:23:16 - INFO - parrotllm.training - step 41510 | epoch 0 | loss 3.4166 | lr 1.17e-03 | grad 0.2356 +2026-04-10 08:23:19 - INFO - parrotllm.training - step 41520 | epoch 0 | loss 3.4153 | lr 1.17e-03 | grad 0.2228 +2026-04-10 08:23:22 - INFO - parrotllm.training - step 41530 | epoch 0 | loss 3.5450 | lr 1.16e-03 | grad 0.2078 +2026-04-10 08:23:25 - INFO - parrotllm.training - step 41540 | epoch 0 | loss 3.4406 | lr 1.16e-03 | grad 0.2115 +2026-04-10 08:23:28 - INFO - parrotllm.training - step 41550 | epoch 0 | loss 3.4086 | lr 1.16e-03 | grad 0.2272 +2026-04-10 08:23:31 - INFO - parrotllm.training - step 41560 | epoch 0 | loss 3.5025 | lr 1.16e-03 | grad 0.2314 +2026-04-10 08:23:34 - INFO - parrotllm.training - step 41570 | epoch 0 | loss 3.4570 | lr 1.16e-03 | grad 0.2340 +2026-04-10 08:23:37 - INFO - parrotllm.training - step 41580 | epoch 0 | loss 3.4277 | lr 1.16e-03 | grad 0.2354 +2026-04-10 08:23:40 - INFO - parrotllm.training - step 41590 | epoch 0 | loss 3.4440 | lr 1.16e-03 | grad 0.2518 +2026-04-10 08:23:43 - INFO - parrotllm.training - step 41600 | epoch 0 | loss 3.3791 | lr 1.16e-03 | grad 0.2407 +2026-04-10 08:23:46 - INFO - parrotllm.training - step 41610 | epoch 0 | loss 3.4334 | lr 1.16e-03 | grad 0.2465 +2026-04-10 08:23:49 - INFO - parrotllm.training - step 41620 | epoch 0 | loss 3.3970 | lr 1.16e-03 | grad 0.2473 +2026-04-10 08:23:52 - INFO - parrotllm.training - step 41630 | epoch 0 | loss 3.4474 | lr 1.16e-03 | grad 0.2316 +2026-04-10 08:23:56 - INFO - parrotllm.training - step 41640 | epoch 0 | loss 3.4379 | lr 1.16e-03 | grad 0.2603 +2026-04-10 08:23:59 - INFO - parrotllm.training - step 41650 | epoch 0 | loss 3.3080 | lr 1.16e-03 | grad 0.2189 +2026-04-10 08:24:02 - INFO - parrotllm.training - step 41660 | epoch 0 | loss 3.3540 | lr 1.16e-03 | grad 0.2048 +2026-04-10 08:24:05 - INFO - parrotllm.training - step 41670 | epoch 0 | loss 3.4582 | lr 1.16e-03 | grad 0.2196 +2026-04-10 08:24:08 - INFO - parrotllm.training - step 41680 | epoch 0 | loss 3.3931 | lr 1.16e-03 | grad 0.2101 +2026-04-10 08:24:11 - INFO - parrotllm.training - step 41690 | epoch 0 | loss 3.5511 | lr 1.16e-03 | grad 0.2283 +2026-04-10 08:24:14 - INFO - parrotllm.training - step 41700 | epoch 0 | loss 3.4030 | lr 1.16e-03 | grad 0.2176 +2026-04-10 08:24:17 - INFO - parrotllm.training - step 41710 | epoch 0 | loss 3.5357 | lr 1.16e-03 | grad 0.2191 +2026-04-10 08:24:20 - INFO - parrotllm.training - step 41720 | epoch 0 | loss 3.5120 | lr 1.16e-03 | grad 0.2190 +2026-04-10 08:24:23 - INFO - parrotllm.training - step 41730 | epoch 0 | loss 3.4902 | lr 1.16e-03 | grad 0.2548 +2026-04-10 08:24:26 - INFO - parrotllm.training - step 41740 | epoch 0 | loss 3.4099 | lr 1.16e-03 | grad 0.2139 +2026-04-10 08:24:29 - INFO - parrotllm.training - step 41750 | epoch 0 | loss 3.3971 | lr 1.16e-03 | grad 0.2201 +2026-04-10 08:24:32 - INFO - parrotllm.training - step 41760 | epoch 0 | loss 3.4434 | lr 1.16e-03 | grad 0.2598 +2026-04-10 08:24:35 - INFO - parrotllm.training - step 41770 | epoch 0 | loss 3.4772 | lr 1.16e-03 | grad 0.2265 +2026-04-10 08:24:38 - INFO - parrotllm.training - step 41780 | epoch 0 | loss 3.3687 | lr 1.16e-03 | grad 0.2082 +2026-04-10 08:24:41 - INFO - parrotllm.training - step 41790 | epoch 0 | loss 3.3748 | lr 1.16e-03 | grad 0.2492 +2026-04-10 08:24:44 - INFO - parrotllm.training - step 41800 | epoch 0 | loss 3.4478 | lr 1.16e-03 | grad 0.2230 +2026-04-10 08:24:47 - INFO - parrotllm.training - step 41810 | epoch 0 | loss 3.4479 | lr 1.16e-03 | grad 0.2420 +2026-04-10 08:24:50 - INFO - parrotllm.training - step 41820 | epoch 0 | loss 3.4337 | lr 1.16e-03 | grad 0.2120 +2026-04-10 08:24:54 - INFO - parrotllm.training - step 41830 | epoch 0 | loss 3.4516 | lr 1.16e-03 | grad 0.1941 +2026-04-10 08:24:57 - INFO - parrotllm.training - step 41840 | epoch 0 | loss 3.3468 | lr 1.16e-03 | grad 0.2438 +2026-04-10 08:25:00 - INFO - parrotllm.training - step 41850 | epoch 0 | loss 3.3741 | lr 1.16e-03 | grad 0.2265 +2026-04-10 08:25:03 - INFO - parrotllm.training - step 41860 | epoch 0 | loss 3.3673 | lr 1.16e-03 | grad 0.2243 +2026-04-10 08:25:06 - INFO - parrotllm.training - step 41870 | epoch 0 | loss 3.4413 | lr 1.16e-03 | grad 0.2206 +2026-04-10 08:25:09 - INFO - parrotllm.training - step 41880 | epoch 0 | loss 3.5047 | lr 1.16e-03 | grad 0.2325 +2026-04-10 08:25:12 - INFO - parrotllm.training - step 41890 | epoch 0 | loss 3.3897 | lr 1.16e-03 | grad 0.2480 +2026-04-10 08:25:15 - INFO - parrotllm.training - step 41900 | epoch 0 | loss 3.3948 | lr 1.16e-03 | grad 0.2162 +2026-04-10 08:25:18 - INFO - parrotllm.training - step 41910 | epoch 0 | loss 3.3429 | lr 1.16e-03 | grad 0.2173 +2026-04-10 08:25:21 - INFO - parrotllm.training - step 41920 | epoch 0 | loss 3.5336 | lr 1.16e-03 | grad 0.2290 +2026-04-10 08:25:24 - INFO - parrotllm.training - step 41930 | epoch 0 | loss 3.5005 | lr 1.16e-03 | grad 0.2227 +2026-04-10 08:25:27 - INFO - parrotllm.training - step 41940 | epoch 0 | loss 3.3998 | lr 1.16e-03 | grad 0.2151 +2026-04-10 08:25:30 - INFO - parrotllm.training - step 41950 | epoch 0 | loss 3.4739 | lr 1.16e-03 | grad 0.2320 +2026-04-10 08:25:33 - INFO - parrotllm.training - step 41960 | epoch 0 | loss 3.4171 | lr 1.16e-03 | grad 0.2305 +2026-04-10 08:25:36 - INFO - parrotllm.training - step 41970 | epoch 0 | loss 3.5549 | lr 1.16e-03 | grad 0.2285 +2026-04-10 08:25:37 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 41990/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 08:25:40 - INFO - parrotllm.training - step 41980 | epoch 0 | loss 3.4038 | lr 1.16e-03 | grad 0.2172 +2026-04-10 08:25:43 - INFO - parrotllm.training - step 41990 | epoch 0 | loss 3.4197 | lr 1.16e-03 | grad 0.2574 +2026-04-10 08:25:46 - INFO - parrotllm.training - step 42000 | epoch 0 | loss 3.3945 | lr 1.16e-03 | grad 0.2437 +2026-04-10 08:25:46 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:25:46 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:25:49 - INFO - parrotllm.training - Train: loss=3.3945, ppl=29.80 +2026-04-10 08:25:49 - INFO - parrotllm.training - Val: loss=3.3774, ppl=29.29 +2026-04-10 08:25:49 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 08:25:50 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3774_epoch_0000_step_0042000.pt +2026-04-10 08:25:51 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:25:54 - INFO - parrotllm.training - step 42010 | epoch 0 | loss 3.4800 | lr 1.16e-03 | grad 0.2422 +2026-04-10 08:25:57 - INFO - parrotllm.training - step 42020 | epoch 0 | loss 3.4488 | lr 1.16e-03 | grad 0.2328 +2026-04-10 08:26:00 - INFO - parrotllm.training - step 42030 | epoch 0 | loss 3.4290 | lr 1.16e-03 | grad 0.2388 +2026-04-10 08:26:03 - INFO - parrotllm.training - step 42040 | epoch 0 | loss 3.3104 | lr 1.16e-03 | grad 0.2380 +2026-04-10 08:26:06 - INFO - parrotllm.training - step 42050 | epoch 0 | loss 3.3547 | lr 1.16e-03 | grad 0.2098 +2026-04-10 08:26:10 - INFO - parrotllm.training - step 42060 | epoch 0 | loss 3.4938 | lr 1.16e-03 | grad 0.2198 +2026-04-10 08:26:13 - INFO - parrotllm.training - step 42070 | epoch 0 | loss 3.4295 | lr 1.16e-03 | grad 0.2006 +2026-04-10 08:26:16 - INFO - parrotllm.training - step 42080 | epoch 0 | loss 3.4199 | lr 1.16e-03 | grad 0.2231 +2026-04-10 08:26:19 - INFO - parrotllm.training - step 42090 | epoch 0 | loss 3.4298 | lr 1.16e-03 | grad 0.2302 +2026-04-10 08:26:22 - INFO - parrotllm.training - step 42100 | epoch 0 | loss 3.4117 | lr 1.16e-03 | grad 0.2572 +2026-04-10 08:26:25 - INFO - parrotllm.training - step 42110 | epoch 0 | loss 3.3829 | lr 1.16e-03 | grad 0.2204 +2026-04-10 08:26:28 - INFO - parrotllm.training - step 42120 | epoch 0 | loss 3.4383 | lr 1.16e-03 | grad 0.2487 +2026-04-10 08:26:31 - INFO - parrotllm.training - step 42130 | epoch 0 | loss 3.4181 | lr 1.16e-03 | grad 0.2265 +2026-04-10 08:26:34 - INFO - parrotllm.training - step 42140 | epoch 0 | loss 3.4497 | lr 1.16e-03 | grad 0.2194 +2026-04-10 08:26:37 - INFO - parrotllm.training - step 42150 | epoch 0 | loss 3.4303 | lr 1.16e-03 | grad 0.2218 +2026-04-10 08:26:40 - INFO - parrotllm.training - step 42160 | epoch 0 | loss 3.5076 | lr 1.16e-03 | grad 0.2141 +2026-04-10 08:26:43 - INFO - parrotllm.training - step 42170 | epoch 0 | loss 3.5174 | lr 1.16e-03 | grad 0.2243 +2026-04-10 08:26:46 - INFO - parrotllm.training - step 42180 | epoch 0 | loss 3.4697 | lr 1.16e-03 | grad 0.2229 +2026-04-10 08:26:49 - INFO - parrotllm.training - step 42190 | epoch 0 | loss 3.3996 | lr 1.15e-03 | grad 0.2250 +2026-04-10 08:26:52 - INFO - parrotllm.training - step 42200 | epoch 0 | loss 3.4738 | lr 1.15e-03 | grad 0.2189 +2026-04-10 08:26:55 - INFO - parrotllm.training - step 42210 | epoch 0 | loss 3.4544 | lr 1.15e-03 | grad 0.2145 +2026-04-10 08:26:58 - INFO - parrotllm.training - step 42220 | epoch 0 | loss 3.4042 | lr 1.15e-03 | grad 0.2175 +2026-04-10 08:27:01 - INFO - parrotllm.training - step 42230 | epoch 0 | loss 3.3649 | lr 1.15e-03 | grad 0.2265 +2026-04-10 08:27:04 - INFO - parrotllm.training - step 42240 | epoch 0 | loss 3.4570 | lr 1.15e-03 | grad 0.2136 +2026-04-10 08:27:07 - INFO - parrotllm.training - step 42250 | epoch 0 | loss 3.3592 | lr 1.15e-03 | grad 0.2280 +2026-04-10 08:27:11 - INFO - parrotllm.training - step 42260 | epoch 0 | loss 3.3839 | lr 1.15e-03 | grad 0.2297 +2026-04-10 08:27:14 - INFO - parrotllm.training - step 42270 | epoch 0 | loss 3.4120 | lr 1.15e-03 | grad 0.2204 +2026-04-10 08:27:17 - INFO - parrotllm.training - step 42280 | epoch 0 | loss 3.4723 | lr 1.15e-03 | grad 0.2304 +2026-04-10 08:27:20 - INFO - parrotllm.training - step 42290 | epoch 0 | loss 3.4622 | lr 1.15e-03 | grad 0.2035 +2026-04-10 08:27:23 - INFO - parrotllm.training - step 42300 | epoch 0 | loss 3.4949 | lr 1.15e-03 | grad 0.2597 +2026-04-10 08:27:26 - INFO - parrotllm.training - step 42310 | epoch 0 | loss 3.4725 | lr 1.15e-03 | grad 0.2087 +2026-04-10 08:27:29 - INFO - parrotllm.training - step 42320 | epoch 0 | loss 3.4802 | lr 1.15e-03 | grad 0.2187 +2026-04-10 08:27:32 - INFO - parrotllm.training - step 42330 | epoch 0 | loss 3.4495 | lr 1.15e-03 | grad 0.2356 +2026-04-10 08:27:35 - INFO - parrotllm.training - step 42340 | epoch 0 | loss 3.3694 | lr 1.15e-03 | grad 0.2339 +2026-04-10 08:27:38 - INFO - parrotllm.training - step 42350 | epoch 0 | loss 3.4001 | lr 1.15e-03 | grad 0.2260 +2026-04-10 08:27:41 - INFO - parrotllm.training - step 42360 | epoch 0 | loss 3.3771 | lr 1.15e-03 | grad 0.2824 +2026-04-10 08:27:44 - INFO - parrotllm.training - step 42370 | epoch 0 | loss 3.4602 | lr 1.15e-03 | grad 0.2212 +2026-04-10 08:27:47 - INFO - parrotllm.training - step 42380 | epoch 0 | loss 3.5008 | lr 1.15e-03 | grad 0.2201 +2026-04-10 08:27:50 - INFO - parrotllm.training - step 42390 | epoch 0 | loss 3.5699 | lr 1.15e-03 | grad 0.2301 +2026-04-10 08:27:53 - INFO - parrotllm.training - step 42400 | epoch 0 | loss 3.4498 | lr 1.15e-03 | grad 0.2078 +2026-04-10 08:27:56 - INFO - parrotllm.training - step 42410 | epoch 0 | loss 3.3575 | lr 1.15e-03 | grad 0.2044 +2026-04-10 08:27:59 - INFO - parrotllm.training - step 42420 | epoch 0 | loss 3.4760 | lr 1.15e-03 | grad 0.2260 +2026-04-10 08:28:02 - INFO - parrotllm.training - step 42430 | epoch 0 | loss 3.2757 | lr 1.15e-03 | grad 0.2263 +2026-04-10 08:28:05 - INFO - parrotllm.training - step 42440 | epoch 0 | loss 3.3844 | lr 1.15e-03 | grad 0.2675 +2026-04-10 08:28:08 - INFO - parrotllm.training - step 42450 | epoch 0 | loss 3.4859 | lr 1.15e-03 | grad 0.2032 +2026-04-10 08:28:12 - INFO - parrotllm.training - step 42460 | epoch 0 | loss 3.3730 | lr 1.15e-03 | grad 0.2600 +2026-04-10 08:28:15 - INFO - parrotllm.training - step 42470 | epoch 0 | loss 3.4210 | lr 1.15e-03 | grad 0.2431 +2026-04-10 08:28:18 - INFO - parrotllm.training - step 42480 | epoch 0 | loss 3.4479 | lr 1.15e-03 | grad 0.2311 +2026-04-10 08:28:21 - INFO - parrotllm.training - step 42490 | epoch 0 | loss 3.4772 | lr 1.15e-03 | grad 0.2555 +2026-04-10 08:28:24 - INFO - parrotllm.training - step 42500 | epoch 0 | loss 3.4812 | lr 1.15e-03 | grad 0.2183 +2026-04-10 08:28:24 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:28:24 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:28:27 - INFO - parrotllm.training - Train: loss=3.4812, ppl=32.50 +2026-04-10 08:28:27 - INFO - parrotllm.training - Val: loss=3.3804, ppl=29.38 +2026-04-10 08:28:28 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3804_epoch_0000_step_0042500.pt +2026-04-10 08:28:29 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3774, min_delta=0.001000). +2026-04-10 08:28:29 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:28:30 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0042500.pt +2026-04-10 08:28:35 - INFO - parrotllm.training - step 42510 | epoch 0 | loss 3.3877 | lr 1.15e-03 | grad 0.2380 +2026-04-10 08:28:38 - INFO - parrotllm.training - step 42520 | epoch 0 | loss 3.5632 | lr 1.15e-03 | grad 0.2157 +2026-04-10 08:28:41 - INFO - parrotllm.training - step 42530 | epoch 0 | loss 3.3990 | lr 1.15e-03 | grad 0.2312 +2026-04-10 08:28:44 - INFO - parrotllm.training - step 42540 | epoch 0 | loss 3.3777 | lr 1.15e-03 | grad 0.2172 +2026-04-10 08:28:47 - INFO - parrotllm.training - step 42550 | epoch 0 | loss 3.4252 | lr 1.15e-03 | grad 0.2135 +2026-04-10 08:28:50 - INFO - parrotllm.training - step 42560 | epoch 0 | loss 3.4612 | lr 1.15e-03 | grad 0.2287 +2026-04-10 08:28:53 - INFO - parrotllm.training - step 42570 | epoch 0 | loss 3.4644 | lr 1.15e-03 | grad 0.2296 +2026-04-10 08:28:56 - INFO - parrotllm.training - step 42580 | epoch 0 | loss 3.4061 | lr 1.15e-03 | grad 0.2401 +2026-04-10 08:28:59 - INFO - parrotllm.training - step 42590 | epoch 0 | loss 3.3472 | lr 1.15e-03 | grad 0.2096 +2026-04-10 08:29:02 - INFO - parrotllm.training - step 42600 | epoch 0 | loss 3.3761 | lr 1.15e-03 | grad 0.2256 +2026-04-10 08:29:05 - INFO - parrotllm.training - step 42610 | epoch 0 | loss 3.5842 | lr 1.15e-03 | grad 0.2489 +2026-04-10 08:29:08 - INFO - parrotllm.training - step 42620 | epoch 0 | loss 3.5145 | lr 1.15e-03 | grad 0.2523 +2026-04-10 08:29:11 - INFO - parrotllm.training - step 42630 | epoch 0 | loss 3.4998 | lr 1.15e-03 | grad 0.2153 +2026-04-10 08:29:14 - INFO - parrotllm.training - step 42640 | epoch 0 | loss 3.2799 | lr 1.15e-03 | grad 0.2286 +2026-04-10 08:29:17 - INFO - parrotllm.training - step 42650 | epoch 0 | loss 3.4303 | lr 1.15e-03 | grad 0.2271 +2026-04-10 08:29:20 - INFO - parrotllm.training - step 42660 | epoch 0 | loss 3.4648 | lr 1.15e-03 | grad 0.2084 +2026-04-10 08:29:23 - INFO - parrotllm.training - step 42670 | epoch 0 | loss 3.4986 | lr 1.15e-03 | grad 0.2536 +2026-04-10 08:29:26 - INFO - parrotllm.training - step 42680 | epoch 0 | loss 3.3984 | lr 1.15e-03 | grad 0.2322 +2026-04-10 08:29:30 - INFO - parrotllm.training - step 42690 | epoch 0 | loss 3.4572 | lr 1.15e-03 | grad 0.2128 +2026-04-10 08:29:33 - INFO - parrotllm.training - step 42700 | epoch 0 | loss 3.3976 | lr 1.15e-03 | grad 0.2105 +2026-04-10 08:29:36 - INFO - parrotllm.training - step 42710 | epoch 0 | loss 3.3739 | lr 1.15e-03 | grad 0.2188 +2026-04-10 08:29:39 - INFO - parrotllm.training - step 42720 | epoch 0 | loss 3.3799 | lr 1.15e-03 | grad 0.2094 +2026-04-10 08:29:42 - INFO - parrotllm.training - step 42730 | epoch 0 | loss 3.4147 | lr 1.15e-03 | grad 0.2215 +2026-04-10 08:29:45 - INFO - parrotllm.training - step 42740 | epoch 0 | loss 3.2410 | lr 1.15e-03 | grad 0.2191 +2026-04-10 08:29:48 - INFO - parrotllm.training - step 42750 | epoch 0 | loss 3.3992 | lr 1.15e-03 | grad 0.2298 +2026-04-10 08:29:51 - INFO - parrotllm.training - step 42760 | epoch 0 | loss 3.6132 | lr 1.15e-03 | grad 0.2369 +2026-04-10 08:29:54 - INFO - parrotllm.training - step 42770 | epoch 0 | loss 3.2944 | lr 1.15e-03 | grad 0.2147 +2026-04-10 08:29:57 - INFO - parrotllm.training - step 42780 | epoch 0 | loss 3.3397 | lr 1.15e-03 | grad 0.2424 +2026-04-10 08:30:00 - INFO - parrotllm.training - step 42790 | epoch 0 | loss 3.4722 | lr 1.15e-03 | grad 0.2542 +2026-04-10 08:30:03 - INFO - parrotllm.training - step 42800 | epoch 0 | loss 3.3811 | lr 1.15e-03 | grad 0.2492 +2026-04-10 08:30:06 - INFO - parrotllm.training - step 42810 | epoch 0 | loss 3.4711 | lr 1.15e-03 | grad 0.2202 +2026-04-10 08:30:09 - INFO - parrotllm.training - step 42820 | epoch 0 | loss 3.3732 | lr 1.15e-03 | grad 0.2085 +2026-04-10 08:30:12 - INFO - parrotllm.training - step 42830 | epoch 0 | loss 3.4028 | lr 1.14e-03 | grad 0.2346 +2026-04-10 08:30:15 - INFO - parrotllm.training - step 42840 | epoch 0 | loss 3.4305 | lr 1.14e-03 | grad 0.2145 +2026-04-10 08:30:18 - INFO - parrotllm.training - step 42850 | epoch 0 | loss 3.3697 | lr 1.14e-03 | grad 0.2632 +2026-04-10 08:30:21 - INFO - parrotllm.training - step 42860 | epoch 0 | loss 3.4767 | lr 1.14e-03 | grad 0.2251 +2026-04-10 08:30:24 - INFO - parrotllm.training - step 42870 | epoch 0 | loss 3.3892 | lr 1.14e-03 | grad 0.2244 +2026-04-10 08:30:28 - INFO - parrotllm.training - step 42880 | epoch 0 | loss 3.4498 | lr 1.14e-03 | grad 0.1945 +2026-04-10 08:30:31 - INFO - parrotllm.training - step 42890 | epoch 0 | loss 3.4147 | lr 1.14e-03 | grad 0.2326 +2026-04-10 08:30:34 - INFO - parrotllm.training - step 42900 | epoch 0 | loss 3.5001 | lr 1.14e-03 | grad 0.2132 +2026-04-10 08:30:37 - INFO - parrotllm.training - step 42910 | epoch 0 | loss 3.5045 | lr 1.14e-03 | grad 0.2429 +2026-04-10 08:30:40 - INFO - parrotllm.training - step 42920 | epoch 0 | loss 3.4037 | lr 1.14e-03 | grad 0.2098 +2026-04-10 08:30:43 - INFO - parrotllm.training - step 42930 | epoch 0 | loss 3.4166 | lr 1.14e-03 | grad 0.2178 +2026-04-10 08:30:46 - INFO - parrotllm.training - step 42940 | epoch 0 | loss 3.4460 | lr 1.14e-03 | grad 0.2631 +2026-04-10 08:30:49 - INFO - parrotllm.training - step 42950 | epoch 0 | loss 3.4529 | lr 1.14e-03 | grad 0.2254 +2026-04-10 08:30:52 - INFO - parrotllm.training - step 42960 | epoch 0 | loss 3.4046 | lr 1.14e-03 | grad 0.2590 +2026-04-10 08:30:55 - INFO - parrotllm.training - step 42970 | epoch 0 | loss 3.5148 | lr 1.14e-03 | grad 0.2116 +2026-04-10 08:30:58 - INFO - parrotllm.training - step 42980 | epoch 0 | loss 3.4074 | lr 1.14e-03 | grad 0.2176 +2026-04-10 08:31:01 - INFO - parrotllm.training - step 42990 | epoch 0 | loss 3.3806 | lr 1.14e-03 | grad 0.2183 +2026-04-10 08:31:04 - INFO - parrotllm.training - step 43000 | epoch 0 | loss 3.4875 | lr 1.14e-03 | grad 0.2148 +2026-04-10 08:31:04 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:31:04 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:31:07 - INFO - parrotllm.training - Train: loss=3.4875, ppl=32.70 +2026-04-10 08:31:07 - INFO - parrotllm.training - Val: loss=3.3784, ppl=29.32 +2026-04-10 08:31:08 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3784_epoch_0000_step_0043000.pt +2026-04-10 08:31:10 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3774, min_delta=0.001000). +2026-04-10 08:31:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:31:13 - INFO - parrotllm.training - step 43010 | epoch 0 | loss 3.5331 | lr 1.14e-03 | grad 0.2255 +2026-04-10 08:31:16 - INFO - parrotllm.training - step 43020 | epoch 0 | loss 3.4830 | lr 1.14e-03 | grad 0.2214 +2026-04-10 08:31:19 - INFO - parrotllm.training - step 43030 | epoch 0 | loss 3.4144 | lr 1.14e-03 | grad 0.2531 +2026-04-10 08:31:22 - INFO - parrotllm.training - step 43040 | epoch 0 | loss 3.3739 | lr 1.14e-03 | grad 0.2406 +2026-04-10 08:31:25 - INFO - parrotllm.training - step 43050 | epoch 0 | loss 3.4469 | lr 1.14e-03 | grad 0.2362 +2026-04-10 08:31:28 - INFO - parrotllm.training - step 43060 | epoch 0 | loss 3.4147 | lr 1.14e-03 | grad 0.2146 +2026-04-10 08:31:31 - INFO - parrotllm.training - step 43070 | epoch 0 | loss 3.5746 | lr 1.14e-03 | grad 0.2465 +2026-04-10 08:31:34 - INFO - parrotllm.training - step 43080 | epoch 0 | loss 3.5375 | lr 1.14e-03 | grad 0.2053 +2026-04-10 08:31:37 - INFO - parrotllm.training - step 43090 | epoch 0 | loss 3.5563 | lr 1.14e-03 | grad 0.2156 +2026-04-10 08:31:40 - INFO - parrotllm.training - step 43100 | epoch 0 | loss 3.4896 | lr 1.14e-03 | grad 0.2339 +2026-04-10 08:31:43 - INFO - parrotllm.training - step 43110 | epoch 0 | loss 3.4684 | lr 1.14e-03 | grad 0.2102 +2026-04-10 08:31:46 - INFO - parrotllm.training - step 43120 | epoch 0 | loss 3.3838 | lr 1.14e-03 | grad 0.2061 +2026-04-10 08:31:49 - INFO - parrotllm.training - step 43130 | epoch 0 | loss 3.4557 | lr 1.14e-03 | grad 0.2413 +2026-04-10 08:31:52 - INFO - parrotllm.training - step 43140 | epoch 0 | loss 3.4174 | lr 1.14e-03 | grad 0.2373 +2026-04-10 08:31:55 - INFO - parrotllm.training - step 43150 | epoch 0 | loss 3.3148 | lr 1.14e-03 | grad 0.2267 +2026-04-10 08:31:58 - INFO - parrotllm.training - step 43160 | epoch 0 | loss 3.3925 | lr 1.14e-03 | grad 0.2444 +2026-04-10 08:32:01 - INFO - parrotllm.training - step 43170 | epoch 0 | loss 3.4032 | lr 1.14e-03 | grad 0.2295 +2026-04-10 08:32:04 - INFO - parrotllm.training - step 43180 | epoch 0 | loss 3.4040 | lr 1.14e-03 | grad 0.2598 +2026-04-10 08:32:08 - INFO - parrotllm.training - step 43190 | epoch 0 | loss 3.3405 | lr 1.14e-03 | grad 0.2418 +2026-04-10 08:32:11 - INFO - parrotllm.training - step 43200 | epoch 0 | loss 3.4126 | lr 1.14e-03 | grad 0.2196 +2026-04-10 08:32:14 - INFO - parrotllm.training - step 43210 | epoch 0 | loss 3.5266 | lr 1.14e-03 | grad 0.2419 +2026-04-10 08:32:17 - INFO - parrotllm.training - step 43220 | epoch 0 | loss 3.3685 | lr 1.14e-03 | grad 0.2247 +2026-04-10 08:32:20 - INFO - parrotllm.training - step 43230 | epoch 0 | loss 3.4222 | lr 1.14e-03 | grad 0.2438 +2026-04-10 08:32:23 - INFO - parrotllm.training - step 43240 | epoch 0 | loss 3.3981 | lr 1.14e-03 | grad 0.2341 +2026-04-10 08:32:26 - INFO - parrotllm.training - step 43250 | epoch 0 | loss 3.4906 | lr 1.14e-03 | grad 0.2522 +2026-04-10 08:32:29 - INFO - parrotllm.training - step 43260 | epoch 0 | loss 3.3802 | lr 1.14e-03 | grad 0.2039 +2026-04-10 08:32:32 - INFO - parrotllm.training - step 43270 | epoch 0 | loss 3.3995 | lr 1.14e-03 | grad 0.2233 +2026-04-10 08:32:35 - INFO - parrotllm.training - step 43280 | epoch 0 | loss 3.4286 | lr 1.14e-03 | grad 0.2448 +2026-04-10 08:32:38 - INFO - parrotllm.training - step 43290 | epoch 0 | loss 3.4614 | lr 1.14e-03 | grad 0.2154 +2026-04-10 08:32:41 - INFO - parrotllm.training - step 43300 | epoch 0 | loss 3.4035 | lr 1.14e-03 | grad 0.2540 +2026-04-10 08:32:44 - INFO - parrotllm.training - step 43310 | epoch 0 | loss 3.4289 | lr 1.14e-03 | grad 0.2242 +2026-04-10 08:32:47 - INFO - parrotllm.training - step 43320 | epoch 0 | loss 3.4761 | lr 1.14e-03 | grad 0.2477 +2026-04-10 08:32:50 - INFO - parrotllm.training - step 43330 | epoch 0 | loss 3.4148 | lr 1.14e-03 | grad 0.2257 +2026-04-10 08:32:53 - INFO - parrotllm.training - step 43340 | epoch 0 | loss 3.5069 | lr 1.14e-03 | grad 0.2423 +2026-04-10 08:32:56 - INFO - parrotllm.training - step 43350 | epoch 0 | loss 3.4195 | lr 1.14e-03 | grad 0.2214 +2026-04-10 08:32:59 - INFO - parrotllm.training - step 43360 | epoch 0 | loss 3.3620 | lr 1.14e-03 | grad 0.2040 +2026-04-10 08:33:02 - INFO - parrotllm.training - step 43370 | epoch 0 | loss 3.5036 | lr 1.14e-03 | grad 0.2162 +2026-04-10 08:33:05 - INFO - parrotllm.training - step 43380 | epoch 0 | loss 3.4995 | lr 1.14e-03 | grad 0.2055 +2026-04-10 08:33:09 - INFO - parrotllm.training - step 43390 | epoch 0 | loss 3.5232 | lr 1.14e-03 | grad 0.2280 +2026-04-10 08:33:12 - INFO - parrotllm.training - step 43400 | epoch 0 | loss 3.4606 | lr 1.14e-03 | grad 0.2526 +2026-04-10 08:33:15 - INFO - parrotllm.training - step 43410 | epoch 0 | loss 3.3678 | lr 1.14e-03 | grad 0.2155 +2026-04-10 08:33:18 - INFO - parrotllm.training - step 43420 | epoch 0 | loss 3.4585 | lr 1.14e-03 | grad 0.1948 +2026-04-10 08:33:21 - INFO - parrotllm.training - step 43430 | epoch 0 | loss 3.4183 | lr 1.14e-03 | grad 0.2384 +2026-04-10 08:33:24 - INFO - parrotllm.training - step 43440 | epoch 0 | loss 3.4767 | lr 1.14e-03 | grad 0.2287 +2026-04-10 08:33:27 - INFO - parrotllm.training - step 43450 | epoch 0 | loss 3.5138 | lr 1.14e-03 | grad 0.2274 +2026-04-10 08:33:30 - INFO - parrotllm.training - step 43460 | epoch 0 | loss 3.4043 | lr 1.14e-03 | grad 0.2388 +2026-04-10 08:33:33 - INFO - parrotllm.training - step 43470 | epoch 0 | loss 3.4386 | lr 1.14e-03 | grad 0.2768 +2026-04-10 08:33:36 - INFO - parrotllm.training - step 43480 | epoch 0 | loss 3.5725 | lr 1.13e-03 | grad 0.2298 +2026-04-10 08:33:39 - INFO - parrotllm.training - step 43490 | epoch 0 | loss 3.3946 | lr 1.13e-03 | grad 0.2270 +2026-04-10 08:33:42 - INFO - parrotllm.training - step 43500 | epoch 0 | loss 3.5464 | lr 1.13e-03 | grad 0.2346 +2026-04-10 08:33:42 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:33:42 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:33:45 - INFO - parrotllm.training - Train: loss=3.5464, ppl=34.69 +2026-04-10 08:33:45 - INFO - parrotllm.training - Val: loss=3.3764, ppl=29.27 +2026-04-10 08:33:46 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3764_epoch_0000_step_0043500.pt +2026-04-10 08:33:48 - INFO - parrotllm.training - No validation improvement for 3/15 evaluation(s) (best=3.3774, min_delta=0.001000). +2026-04-10 08:33:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:33:51 - INFO - parrotllm.training - step 43510 | epoch 0 | loss 3.5368 | lr 1.13e-03 | grad 0.2140 +2026-04-10 08:33:54 - INFO - parrotllm.training - step 43520 | epoch 0 | loss 3.3906 | lr 1.13e-03 | grad 0.2101 +2026-04-10 08:33:57 - INFO - parrotllm.training - step 43530 | epoch 0 | loss 3.4257 | lr 1.13e-03 | grad 0.2087 +2026-04-10 08:34:00 - INFO - parrotllm.training - step 43540 | epoch 0 | loss 3.4921 | lr 1.13e-03 | grad 0.2160 +2026-04-10 08:34:03 - INFO - parrotllm.training - step 43550 | epoch 0 | loss 3.4226 | lr 1.13e-03 | grad 0.2279 +2026-04-10 08:34:06 - INFO - parrotllm.training - step 43560 | epoch 0 | loss 3.3519 | lr 1.13e-03 | grad 0.2235 +2026-04-10 08:34:09 - INFO - parrotllm.training - step 43570 | epoch 0 | loss 3.4096 | lr 1.13e-03 | grad 0.2422 +2026-04-10 08:34:12 - INFO - parrotllm.training - step 43580 | epoch 0 | loss 3.4803 | lr 1.13e-03 | grad 0.2232 +2026-04-10 08:34:15 - INFO - parrotllm.training - step 43590 | epoch 0 | loss 3.3678 | lr 1.13e-03 | grad 0.2113 +2026-04-10 08:34:18 - INFO - parrotllm.training - step 43600 | epoch 0 | loss 3.3978 | lr 1.13e-03 | grad 0.2602 +2026-04-10 08:34:21 - INFO - parrotllm.training - step 43610 | epoch 0 | loss 3.4193 | lr 1.13e-03 | grad 0.2447 +2026-04-10 08:34:24 - INFO - parrotllm.training - step 43620 | epoch 0 | loss 3.4302 | lr 1.13e-03 | grad 0.2473 +2026-04-10 08:34:27 - INFO - parrotllm.training - step 43630 | epoch 0 | loss 3.4031 | lr 1.13e-03 | grad 0.2428 +2026-04-10 08:34:30 - INFO - parrotllm.training - step 43640 | epoch 0 | loss 3.6052 | lr 1.13e-03 | grad 0.2204 +2026-04-10 08:34:33 - INFO - parrotllm.training - step 43650 | epoch 0 | loss 3.4682 | lr 1.13e-03 | grad 0.2278 +2026-04-10 08:34:36 - INFO - parrotllm.training - step 43660 | epoch 0 | loss 3.4615 | lr 1.13e-03 | grad 0.2148 +2026-04-10 08:34:39 - INFO - parrotllm.training - step 43670 | epoch 0 | loss 3.5271 | lr 1.13e-03 | grad 0.2463 +2026-04-10 08:34:42 - INFO - parrotllm.training - step 43680 | epoch 0 | loss 3.3936 | lr 1.13e-03 | grad 0.2257 +2026-04-10 08:34:45 - INFO - parrotllm.training - step 43690 | epoch 0 | loss 3.4053 | lr 1.13e-03 | grad 0.2060 +2026-04-10 08:34:49 - INFO - parrotllm.training - step 43700 | epoch 0 | loss 3.5124 | lr 1.13e-03 | grad 0.2074 +2026-04-10 08:34:52 - INFO - parrotllm.training - step 43710 | epoch 0 | loss 3.3595 | lr 1.13e-03 | grad 0.2634 +2026-04-10 08:34:55 - INFO - parrotllm.training - step 43720 | epoch 0 | loss 3.5691 | lr 1.13e-03 | grad 0.2177 +2026-04-10 08:34:58 - INFO - parrotllm.training - step 43730 | epoch 0 | loss 3.4184 | lr 1.13e-03 | grad 0.2529 +2026-04-10 08:35:01 - INFO - parrotllm.training - step 43740 | epoch 0 | loss 3.3484 | lr 1.13e-03 | grad 0.2331 +2026-04-10 08:35:04 - INFO - parrotllm.training - step 43750 | epoch 0 | loss 3.4898 | lr 1.13e-03 | grad 0.2475 +2026-04-10 08:35:07 - INFO - parrotllm.training - step 43760 | epoch 0 | loss 3.4868 | lr 1.13e-03 | grad 0.2184 +2026-04-10 08:35:10 - INFO - parrotllm.training - step 43770 | epoch 0 | loss 3.5327 | lr 1.13e-03 | grad 0.2421 +2026-04-10 08:35:13 - INFO - parrotllm.training - step 43780 | epoch 0 | loss 3.4416 | lr 1.13e-03 | grad 0.2201 +2026-04-10 08:35:16 - INFO - parrotllm.training - step 43790 | epoch 0 | loss 3.3104 | lr 1.13e-03 | grad 0.2003 +2026-04-10 08:35:19 - INFO - parrotllm.training - step 43800 | epoch 0 | loss 3.4557 | lr 1.13e-03 | grad 0.2555 +2026-04-10 08:35:22 - INFO - parrotllm.training - step 43810 | epoch 0 | loss 3.4664 | lr 1.13e-03 | grad 0.2049 +2026-04-10 08:35:25 - INFO - parrotllm.training - step 43820 | epoch 0 | loss 3.4726 | lr 1.13e-03 | grad 0.2314 +2026-04-10 08:35:28 - INFO - parrotllm.training - step 43830 | epoch 0 | loss 3.4625 | lr 1.13e-03 | grad 0.2265 +2026-04-10 08:35:31 - INFO - parrotllm.training - step 43840 | epoch 0 | loss 3.3785 | lr 1.13e-03 | grad 0.2233 +2026-04-10 08:35:34 - INFO - parrotllm.training - step 43850 | epoch 0 | loss 3.3628 | lr 1.13e-03 | grad 0.2201 +2026-04-10 08:35:37 - INFO - parrotllm.training - step 43860 | epoch 0 | loss 3.3505 | lr 1.13e-03 | grad 0.2496 +2026-04-10 08:35:40 - INFO - parrotllm.training - step 43870 | epoch 0 | loss 3.4372 | lr 1.13e-03 | grad 0.2232 +2026-04-10 08:35:43 - INFO - parrotllm.training - step 43880 | epoch 0 | loss 3.3270 | lr 1.13e-03 | grad 0.2356 +2026-04-10 08:35:46 - INFO - parrotllm.training - step 43890 | epoch 0 | loss 3.4309 | lr 1.13e-03 | grad 0.2230 +2026-04-10 08:35:49 - INFO - parrotllm.training - step 43900 | epoch 0 | loss 3.3998 | lr 1.13e-03 | grad 0.2379 +2026-04-10 08:35:52 - INFO - parrotllm.training - step 43910 | epoch 0 | loss 3.4849 | lr 1.13e-03 | grad 0.2374 +2026-04-10 08:35:56 - INFO - parrotllm.training - step 43920 | epoch 0 | loss 3.5179 | lr 1.13e-03 | grad 0.2075 +2026-04-10 08:35:59 - INFO - parrotllm.training - step 43930 | epoch 0 | loss 3.4126 | lr 1.13e-03 | grad 0.2305 +2026-04-10 08:36:02 - INFO - parrotllm.training - step 43940 | epoch 0 | loss 3.4134 | lr 1.13e-03 | grad 0.2310 +2026-04-10 08:36:05 - INFO - parrotllm.training - step 43950 | epoch 0 | loss 3.4807 | lr 1.13e-03 | grad 0.2237 +2026-04-10 08:36:08 - INFO - parrotllm.training - step 43960 | epoch 0 | loss 3.3770 | lr 1.13e-03 | grad 0.2163 +2026-04-10 08:36:11 - INFO - parrotllm.training - step 43970 | epoch 0 | loss 3.3549 | lr 1.13e-03 | grad 0.2299 +2026-04-10 08:36:14 - INFO - parrotllm.training - step 43980 | epoch 0 | loss 3.4015 | lr 1.13e-03 | grad 0.2794 +2026-04-10 08:36:17 - INFO - parrotllm.training - step 43990 | epoch 0 | loss 3.4398 | lr 1.13e-03 | grad 0.2274 +2026-04-10 08:36:20 - INFO - parrotllm.training - step 44000 | epoch 0 | loss 3.3954 | lr 1.13e-03 | grad 0.2585 +2026-04-10 08:36:20 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:36:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:36:23 - INFO - parrotllm.training - Train: loss=3.3954, ppl=29.83 +2026-04-10 08:36:23 - INFO - parrotllm.training - Val: loss=3.3760, ppl=29.25 +2026-04-10 08:36:23 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 08:36:24 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3760_epoch_0000_step_0044000.pt +2026-04-10 08:36:25 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:36:28 - INFO - parrotllm.training - step 44010 | epoch 0 | loss 3.3677 | lr 1.13e-03 | grad 0.2405 +2026-04-10 08:36:31 - INFO - parrotllm.training - step 44020 | epoch 0 | loss 3.4642 | lr 1.13e-03 | grad 0.2538 +2026-04-10 08:36:34 - INFO - parrotllm.training - step 44030 | epoch 0 | loss 3.5018 | lr 1.13e-03 | grad 0.2413 +2026-04-10 08:36:37 - INFO - parrotllm.training - step 44040 | epoch 0 | loss 3.3896 | lr 1.13e-03 | grad 0.2179 +2026-04-10 08:36:40 - INFO - parrotllm.training - step 44050 | epoch 0 | loss 3.4207 | lr 1.13e-03 | grad 0.2330 +2026-04-10 08:36:43 - INFO - parrotllm.training - step 44060 | epoch 0 | loss 3.3841 | lr 1.13e-03 | grad 0.2359 +2026-04-10 08:36:47 - INFO - parrotllm.training - step 44070 | epoch 0 | loss 3.3403 | lr 1.13e-03 | grad 0.2075 +2026-04-10 08:36:50 - INFO - parrotllm.training - step 44080 | epoch 0 | loss 3.4265 | lr 1.13e-03 | grad 0.2412 +2026-04-10 08:36:53 - INFO - parrotllm.training - step 44090 | epoch 0 | loss 3.4108 | lr 1.13e-03 | grad 0.2306 +2026-04-10 08:36:56 - INFO - parrotllm.training - step 44100 | epoch 0 | loss 3.3746 | lr 1.13e-03 | grad 0.2115 +2026-04-10 08:36:59 - INFO - parrotllm.training - step 44110 | epoch 0 | loss 3.3757 | lr 1.12e-03 | grad 0.2068 +2026-04-10 08:37:02 - INFO - parrotllm.training - step 44120 | epoch 0 | loss 3.5185 | lr 1.12e-03 | grad 0.2306 +2026-04-10 08:37:04 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 44143/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 08:37:05 - INFO - parrotllm.training - step 44130 | epoch 0 | loss 3.2821 | lr 1.12e-03 | grad 0.2134 +2026-04-10 08:37:08 - INFO - parrotllm.training - step 44140 | epoch 0 | loss 3.3812 | lr 1.12e-03 | grad 0.2190 +2026-04-10 08:37:11 - INFO - parrotllm.training - step 44150 | epoch 0 | loss 3.4323 | lr 1.12e-03 | grad 0.2209 +2026-04-10 08:37:14 - INFO - parrotllm.training - step 44160 | epoch 0 | loss 3.5054 | lr 1.12e-03 | grad 0.2208 +2026-04-10 08:37:17 - INFO - parrotllm.training - step 44170 | epoch 0 | loss 3.4099 | lr 1.12e-03 | grad 0.2077 +2026-04-10 08:37:20 - INFO - parrotllm.training - step 44180 | epoch 0 | loss 3.4543 | lr 1.12e-03 | grad 0.2180 +2026-04-10 08:37:23 - INFO - parrotllm.training - step 44190 | epoch 0 | loss 3.3114 | lr 1.12e-03 | grad 0.2567 +2026-04-10 08:37:26 - INFO - parrotllm.training - step 44200 | epoch 0 | loss 3.5499 | lr 1.12e-03 | grad 0.2257 +2026-04-10 08:37:29 - INFO - parrotllm.training - step 44210 | epoch 0 | loss 3.4684 | lr 1.12e-03 | grad 0.2214 +2026-04-10 08:37:33 - INFO - parrotllm.training - step 44220 | epoch 0 | loss 3.4421 | lr 1.12e-03 | grad 0.2173 +2026-04-10 08:37:36 - INFO - parrotllm.training - step 44230 | epoch 0 | loss 3.3838 | lr 1.12e-03 | grad 0.2397 +2026-04-10 08:37:39 - INFO - parrotllm.training - step 44240 | epoch 0 | loss 3.5209 | lr 1.12e-03 | grad 0.2039 +2026-04-10 08:37:42 - INFO - parrotllm.training - step 44250 | epoch 0 | loss 3.3221 | lr 1.12e-03 | grad 0.2343 +2026-04-10 08:37:45 - INFO - parrotllm.training - step 44260 | epoch 0 | loss 3.4822 | lr 1.12e-03 | grad 0.2487 +2026-04-10 08:37:48 - INFO - parrotllm.training - step 44270 | epoch 0 | loss 3.4758 | lr 1.12e-03 | grad 0.2585 +2026-04-10 08:37:51 - INFO - parrotllm.training - step 44280 | epoch 0 | loss 3.4375 | lr 1.12e-03 | grad 0.2519 +2026-04-10 08:37:54 - INFO - parrotllm.training - step 44290 | epoch 0 | loss 3.4231 | lr 1.12e-03 | grad 0.2363 +2026-04-10 08:37:57 - INFO - parrotllm.training - step 44300 | epoch 0 | loss 3.2727 | lr 1.12e-03 | grad 0.2184 +2026-04-10 08:38:00 - INFO - parrotllm.training - step 44310 | epoch 0 | loss 3.4274 | lr 1.12e-03 | grad 0.2207 +2026-04-10 08:38:03 - INFO - parrotllm.training - step 44320 | epoch 0 | loss 3.5189 | lr 1.12e-03 | grad 0.2181 +2026-04-10 08:38:06 - INFO - parrotllm.training - step 44330 | epoch 0 | loss 3.3868 | lr 1.12e-03 | grad 0.2145 +2026-04-10 08:38:09 - INFO - parrotllm.training - step 44340 | epoch 0 | loss 3.4112 | lr 1.12e-03 | grad 0.2125 +2026-04-10 08:38:12 - INFO - parrotllm.training - step 44350 | epoch 0 | loss 3.4456 | lr 1.12e-03 | grad 0.2184 +2026-04-10 08:38:15 - INFO - parrotllm.training - step 44360 | epoch 0 | loss 3.4883 | lr 1.12e-03 | grad 0.2479 +2026-04-10 08:38:18 - INFO - parrotllm.training - step 44370 | epoch 0 | loss 3.5052 | lr 1.12e-03 | grad 0.2060 +2026-04-10 08:38:21 - INFO - parrotllm.training - step 44380 | epoch 0 | loss 3.5284 | lr 1.12e-03 | grad 0.2361 +2026-04-10 08:38:24 - INFO - parrotllm.training - step 44390 | epoch 0 | loss 3.4223 | lr 1.12e-03 | grad 0.2441 +2026-04-10 08:38:27 - INFO - parrotllm.training - step 44400 | epoch 0 | loss 3.4744 | lr 1.12e-03 | grad 0.2337 +2026-04-10 08:38:30 - INFO - parrotllm.training - step 44410 | epoch 0 | loss 3.4092 | lr 1.12e-03 | grad 0.2111 +2026-04-10 08:38:34 - INFO - parrotllm.training - step 44420 | epoch 0 | loss 3.3305 | lr 1.12e-03 | grad 0.2094 +2026-04-10 08:38:37 - INFO - parrotllm.training - step 44430 | epoch 0 | loss 3.4727 | lr 1.12e-03 | grad 0.2564 +2026-04-10 08:38:40 - INFO - parrotllm.training - step 44440 | epoch 0 | loss 3.4389 | lr 1.12e-03 | grad 0.2237 +2026-04-10 08:38:43 - INFO - parrotllm.training - step 44450 | epoch 0 | loss 3.3198 | lr 1.12e-03 | grad 0.2837 +2026-04-10 08:38:46 - INFO - parrotllm.training - step 44460 | epoch 0 | loss 3.3618 | lr 1.12e-03 | grad 0.2221 +2026-04-10 08:38:49 - INFO - parrotllm.training - step 44470 | epoch 0 | loss 3.3905 | lr 1.12e-03 | grad 0.2066 +2026-04-10 08:38:52 - INFO - parrotllm.training - step 44480 | epoch 0 | loss 3.3653 | lr 1.12e-03 | grad 0.2326 +2026-04-10 08:38:55 - INFO - parrotllm.training - step 44490 | epoch 0 | loss 3.4479 | lr 1.12e-03 | grad 0.2626 +2026-04-10 08:38:58 - INFO - parrotllm.training - step 44500 | epoch 0 | loss 3.4081 | lr 1.12e-03 | grad 0.2400 +2026-04-10 08:38:58 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:38:58 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:39:01 - INFO - parrotllm.training - Train: loss=3.4081, ppl=30.21 +2026-04-10 08:39:01 - INFO - parrotllm.training - Val: loss=3.3725, ppl=29.15 +2026-04-10 08:39:01 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 08:39:02 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3725_epoch_0000_step_0044500.pt +2026-04-10 08:39:03 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:39:06 - INFO - parrotllm.training - step 44510 | epoch 0 | loss 3.4781 | lr 1.12e-03 | grad 0.2362 +2026-04-10 08:39:09 - INFO - parrotllm.training - step 44520 | epoch 0 | loss 3.5545 | lr 1.12e-03 | grad 0.2335 +2026-04-10 08:39:12 - INFO - parrotllm.training - step 44530 | epoch 0 | loss 3.3505 | lr 1.12e-03 | grad 0.2314 +2026-04-10 08:39:15 - INFO - parrotllm.training - step 44540 | epoch 0 | loss 3.5290 | lr 1.12e-03 | grad 0.2320 +2026-04-10 08:39:18 - INFO - parrotllm.training - step 44550 | epoch 0 | loss 3.3764 | lr 1.12e-03 | grad 0.2109 +2026-04-10 08:39:21 - INFO - parrotllm.training - step 44560 | epoch 0 | loss 3.4103 | lr 1.12e-03 | grad 0.2226 +2026-04-10 08:39:25 - INFO - parrotllm.training - step 44570 | epoch 0 | loss 3.4738 | lr 1.12e-03 | grad 0.2063 +2026-04-10 08:39:28 - INFO - parrotllm.training - step 44580 | epoch 0 | loss 3.3957 | lr 1.12e-03 | grad 0.2199 +2026-04-10 08:39:31 - INFO - parrotllm.training - step 44590 | epoch 0 | loss 3.4752 | lr 1.12e-03 | grad 0.2441 +2026-04-10 08:39:34 - INFO - parrotllm.training - step 44600 | epoch 0 | loss 3.5007 | lr 1.12e-03 | grad 0.2562 +2026-04-10 08:39:37 - INFO - parrotllm.training - step 44610 | epoch 0 | loss 3.4798 | lr 1.12e-03 | grad 0.2059 +2026-04-10 08:39:40 - INFO - parrotllm.training - step 44620 | epoch 0 | loss 3.3818 | lr 1.12e-03 | grad 0.2423 +2026-04-10 08:39:43 - INFO - parrotllm.training - step 44630 | epoch 0 | loss 3.4247 | lr 1.12e-03 | grad 0.2583 +2026-04-10 08:39:46 - INFO - parrotllm.training - step 44640 | epoch 0 | loss 3.4902 | lr 1.12e-03 | grad 0.2223 +2026-04-10 08:39:49 - INFO - parrotllm.training - step 44650 | epoch 0 | loss 3.4434 | lr 1.12e-03 | grad 0.2338 +2026-04-10 08:39:52 - INFO - parrotllm.training - step 44660 | epoch 0 | loss 3.5516 | lr 1.12e-03 | grad 0.2463 +2026-04-10 08:39:55 - INFO - parrotllm.training - step 44670 | epoch 0 | loss 3.4436 | lr 1.12e-03 | grad 0.2387 +2026-04-10 08:39:58 - INFO - parrotllm.training - step 44680 | epoch 0 | loss 3.4362 | lr 1.12e-03 | grad 0.2206 +2026-04-10 08:40:01 - INFO - parrotllm.training - step 44690 | epoch 0 | loss 3.4263 | lr 1.12e-03 | grad 0.2074 +2026-04-10 08:40:04 - INFO - parrotllm.training - step 44700 | epoch 0 | loss 3.3862 | lr 1.12e-03 | grad 0.2136 +2026-04-10 08:40:07 - INFO - parrotllm.training - step 44710 | epoch 0 | loss 3.4351 | lr 1.12e-03 | grad 0.2226 +2026-04-10 08:40:10 - INFO - parrotllm.training - step 44720 | epoch 0 | loss 3.3758 | lr 1.12e-03 | grad 0.2387 +2026-04-10 08:40:13 - INFO - parrotllm.training - step 44730 | epoch 0 | loss 3.3700 | lr 1.12e-03 | grad 0.2434 +2026-04-10 08:40:16 - INFO - parrotllm.training - step 44740 | epoch 0 | loss 3.4874 | lr 1.12e-03 | grad 0.2167 +2026-04-10 08:40:19 - INFO - parrotllm.training - step 44750 | epoch 0 | loss 3.4352 | lr 1.11e-03 | grad 0.2312 +2026-04-10 08:40:22 - INFO - parrotllm.training - step 44760 | epoch 0 | loss 3.5128 | lr 1.11e-03 | grad 0.2292 +2026-04-10 08:40:26 - INFO - parrotllm.training - step 44770 | epoch 0 | loss 3.4781 | lr 1.11e-03 | grad 0.2346 +2026-04-10 08:40:29 - INFO - parrotllm.training - step 44780 | epoch 0 | loss 3.4180 | lr 1.11e-03 | grad 0.2238 +2026-04-10 08:40:32 - INFO - parrotllm.training - step 44790 | epoch 0 | loss 3.4289 | lr 1.11e-03 | grad 0.2256 +2026-04-10 08:40:35 - INFO - parrotllm.training - step 44800 | epoch 0 | loss 3.3770 | lr 1.11e-03 | grad 0.2099 +2026-04-10 08:40:38 - INFO - parrotllm.training - step 44810 | epoch 0 | loss 3.5537 | lr 1.11e-03 | grad 0.1975 +2026-04-10 08:40:41 - INFO - parrotllm.training - step 44820 | epoch 0 | loss 3.5600 | lr 1.11e-03 | grad 0.2247 +2026-04-10 08:40:44 - INFO - parrotllm.training - step 44830 | epoch 0 | loss 3.5046 | lr 1.11e-03 | grad 0.2060 +2026-04-10 08:40:47 - INFO - parrotllm.training - step 44840 | epoch 0 | loss 3.3342 | lr 1.11e-03 | grad 0.2408 +2026-04-10 08:40:50 - INFO - parrotllm.training - step 44850 | epoch 0 | loss 3.4738 | lr 1.11e-03 | grad 0.2319 +2026-04-10 08:40:53 - INFO - parrotllm.training - step 44860 | epoch 0 | loss 3.4661 | lr 1.11e-03 | grad 0.2272 +2026-04-10 08:40:56 - INFO - parrotllm.training - step 44870 | epoch 0 | loss 3.4690 | lr 1.11e-03 | grad 0.2064 +2026-04-10 08:40:59 - INFO - parrotllm.training - step 44880 | epoch 0 | loss 3.4370 | lr 1.11e-03 | grad 0.2263 +2026-04-10 08:41:02 - INFO - parrotllm.training - step 44890 | epoch 0 | loss 3.4314 | lr 1.11e-03 | grad 0.2394 +2026-04-10 08:41:05 - INFO - parrotllm.training - step 44900 | epoch 0 | loss 3.5642 | lr 1.11e-03 | grad 0.2412 +2026-04-10 08:41:08 - INFO - parrotllm.training - step 44910 | epoch 0 | loss 3.4764 | lr 1.11e-03 | grad 0.2158 +2026-04-10 08:41:11 - INFO - parrotllm.training - step 44920 | epoch 0 | loss 3.4377 | lr 1.11e-03 | grad 0.2170 +2026-04-10 08:41:14 - INFO - parrotllm.training - step 44930 | epoch 0 | loss 3.3023 | lr 1.11e-03 | grad 0.2147 +2026-04-10 08:41:17 - INFO - parrotllm.training - step 44940 | epoch 0 | loss 3.2987 | lr 1.11e-03 | grad 0.2202 +2026-04-10 08:41:20 - INFO - parrotllm.training - step 44950 | epoch 0 | loss 3.3749 | lr 1.11e-03 | grad 0.2392 +2026-04-10 08:41:23 - INFO - parrotllm.training - step 44960 | epoch 0 | loss 3.5097 | lr 1.11e-03 | grad 0.2217 +2026-04-10 08:41:27 - INFO - parrotllm.training - step 44970 | epoch 0 | loss 3.3376 | lr 1.11e-03 | grad 0.2515 +2026-04-10 08:41:30 - INFO - parrotllm.training - step 44980 | epoch 0 | loss 3.4851 | lr 1.11e-03 | grad 0.2476 +2026-04-10 08:41:33 - INFO - parrotllm.training - step 44990 | epoch 0 | loss 3.3905 | lr 1.11e-03 | grad 0.2217 +2026-04-10 08:41:36 - INFO - parrotllm.training - step 45000 | epoch 0 | loss 3.4091 | lr 1.11e-03 | grad 0.2412 +2026-04-10 08:41:36 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:41:36 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:41:39 - INFO - parrotllm.training - Train: loss=3.4091, ppl=30.24 +2026-04-10 08:41:39 - INFO - parrotllm.training - Val: loss=3.3701, ppl=29.08 +2026-04-10 08:41:39 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 08:41:40 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3701_epoch_0000_step_0045000.pt +2026-04-10 08:41:41 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:41:42 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0045000.pt +2026-04-10 08:41:46 - INFO - parrotllm.training - step 45010 | epoch 0 | loss 3.4080 | lr 1.11e-03 | grad 0.2258 +2026-04-10 08:41:50 - INFO - parrotllm.training - step 45020 | epoch 0 | loss 3.3939 | lr 1.11e-03 | grad 0.2295 +2026-04-10 08:41:53 - INFO - parrotllm.training - step 45030 | epoch 0 | loss 3.4940 | lr 1.11e-03 | grad 0.2123 +2026-04-10 08:41:56 - INFO - parrotllm.training - step 45040 | epoch 0 | loss 3.5470 | lr 1.11e-03 | grad 0.2214 +2026-04-10 08:41:59 - INFO - parrotllm.training - step 45050 | epoch 0 | loss 3.4712 | lr 1.11e-03 | grad 0.2285 +2026-04-10 08:42:02 - INFO - parrotllm.training - step 45060 | epoch 0 | loss 3.4199 | lr 1.11e-03 | grad 0.2202 +2026-04-10 08:42:05 - INFO - parrotllm.training - step 45070 | epoch 0 | loss 3.3716 | lr 1.11e-03 | grad 0.2246 +2026-04-10 08:42:08 - INFO - parrotllm.training - step 45080 | epoch 0 | loss 3.4589 | lr 1.11e-03 | grad 0.2052 +2026-04-10 08:42:11 - INFO - parrotllm.training - step 45090 | epoch 0 | loss 3.4803 | lr 1.11e-03 | grad 0.2122 +2026-04-10 08:42:14 - INFO - parrotllm.training - step 45100 | epoch 0 | loss 3.3744 | lr 1.11e-03 | grad 0.2270 +2026-04-10 08:42:17 - INFO - parrotllm.training - step 45110 | epoch 0 | loss 3.4608 | lr 1.11e-03 | grad 0.2225 +2026-04-10 08:42:20 - INFO - parrotllm.training - step 45120 | epoch 0 | loss 3.4305 | lr 1.11e-03 | grad 0.2368 +2026-04-10 08:42:23 - INFO - parrotllm.training - step 45130 | epoch 0 | loss 3.4347 | lr 1.11e-03 | grad 0.2290 +2026-04-10 08:42:26 - INFO - parrotllm.training - step 45140 | epoch 0 | loss 3.3911 | lr 1.11e-03 | grad 0.2510 +2026-04-10 08:42:29 - INFO - parrotllm.training - step 45150 | epoch 0 | loss 3.5051 | lr 1.11e-03 | grad 0.2111 +2026-04-10 08:42:32 - INFO - parrotllm.training - step 45160 | epoch 0 | loss 3.3699 | lr 1.11e-03 | grad 0.2231 +2026-04-10 08:42:35 - INFO - parrotllm.training - step 45170 | epoch 0 | loss 3.3999 | lr 1.11e-03 | grad 0.2544 +2026-04-10 08:42:38 - INFO - parrotllm.training - step 45180 | epoch 0 | loss 3.4097 | lr 1.11e-03 | grad 0.2132 +2026-04-10 08:42:41 - INFO - parrotllm.training - step 45190 | epoch 0 | loss 3.4451 | lr 1.11e-03 | grad 0.2694 +2026-04-10 08:42:45 - INFO - parrotllm.training - step 45200 | epoch 0 | loss 3.3875 | lr 1.11e-03 | grad 0.2306 +2026-04-10 08:42:48 - INFO - parrotllm.training - step 45210 | epoch 0 | loss 3.5264 | lr 1.11e-03 | grad 0.2255 +2026-04-10 08:42:51 - INFO - parrotllm.training - step 45220 | epoch 0 | loss 3.3314 | lr 1.11e-03 | grad 0.2471 +2026-04-10 08:42:54 - INFO - parrotllm.training - step 45230 | epoch 0 | loss 3.4370 | lr 1.11e-03 | grad 0.2540 +2026-04-10 08:42:57 - INFO - parrotllm.training - step 45240 | epoch 0 | loss 3.2603 | lr 1.11e-03 | grad 0.2489 +2026-04-10 08:43:00 - INFO - parrotllm.training - step 45250 | epoch 0 | loss 3.4055 | lr 1.11e-03 | grad 0.2127 +2026-04-10 08:43:03 - INFO - parrotllm.training - step 45260 | epoch 0 | loss 3.4010 | lr 1.11e-03 | grad 0.1952 +2026-04-10 08:43:06 - INFO - parrotllm.training - step 45270 | epoch 0 | loss 3.4671 | lr 1.11e-03 | grad 0.2006 +2026-04-10 08:43:09 - INFO - parrotllm.training - step 45280 | epoch 0 | loss 3.3938 | lr 1.11e-03 | grad 0.2092 +2026-04-10 08:43:12 - INFO - parrotllm.training - step 45290 | epoch 0 | loss 3.3626 | lr 1.11e-03 | grad 0.2325 +2026-04-10 08:43:15 - INFO - parrotllm.training - step 45300 | epoch 0 | loss 3.4710 | lr 1.11e-03 | grad 0.2325 +2026-04-10 08:43:18 - INFO - parrotllm.training - step 45310 | epoch 0 | loss 3.3721 | lr 1.11e-03 | grad 0.2387 +2026-04-10 08:43:21 - INFO - parrotllm.training - step 45320 | epoch 0 | loss 3.4095 | lr 1.11e-03 | grad 0.2007 +2026-04-10 08:43:24 - INFO - parrotllm.training - step 45330 | epoch 0 | loss 3.4866 | lr 1.11e-03 | grad 0.2270 +2026-04-10 08:43:27 - INFO - parrotllm.training - step 45340 | epoch 0 | loss 3.4163 | lr 1.11e-03 | grad 0.2617 +2026-04-10 08:43:30 - INFO - parrotllm.training - step 45350 | epoch 0 | loss 3.3996 | lr 1.11e-03 | grad 0.2336 +2026-04-10 08:43:33 - INFO - parrotllm.training - step 45360 | epoch 0 | loss 3.4265 | lr 1.11e-03 | grad 0.2139 +2026-04-10 08:43:36 - INFO - parrotllm.training - step 45370 | epoch 0 | loss 3.4440 | lr 1.10e-03 | grad 0.2172 +2026-04-10 08:43:39 - INFO - parrotllm.training - step 45380 | epoch 0 | loss 3.5111 | lr 1.10e-03 | grad 0.2433 +2026-04-10 08:43:42 - INFO - parrotllm.training - step 45390 | epoch 0 | loss 3.3939 | lr 1.10e-03 | grad 0.2213 +2026-04-10 08:43:46 - INFO - parrotllm.training - step 45400 | epoch 0 | loss 3.4110 | lr 1.10e-03 | grad 0.2358 +2026-04-10 08:43:49 - INFO - parrotllm.training - step 45410 | epoch 0 | loss 3.4026 | lr 1.10e-03 | grad 0.2436 +2026-04-10 08:43:52 - INFO - parrotllm.training - step 45420 | epoch 0 | loss 3.5780 | lr 1.10e-03 | grad 0.2167 +2026-04-10 08:43:55 - INFO - parrotllm.training - step 45430 | epoch 0 | loss 3.5065 | lr 1.10e-03 | grad 0.2470 +2026-04-10 08:43:58 - INFO - parrotllm.training - step 45440 | epoch 0 | loss 3.4220 | lr 1.10e-03 | grad 0.2194 +2026-04-10 08:44:01 - INFO - parrotllm.training - step 45450 | epoch 0 | loss 3.4035 | lr 1.10e-03 | grad 0.2090 +2026-04-10 08:44:04 - INFO - parrotllm.training - step 45460 | epoch 0 | loss 3.5207 | lr 1.10e-03 | grad 0.3070 +2026-04-10 08:44:07 - INFO - parrotllm.training - step 45470 | epoch 0 | loss 3.4408 | lr 1.10e-03 | grad 0.2486 +2026-04-10 08:44:10 - INFO - parrotllm.training - step 45480 | epoch 0 | loss 3.3934 | lr 1.10e-03 | grad 0.2350 +2026-04-10 08:44:13 - INFO - parrotllm.training - step 45490 | epoch 0 | loss 3.5230 | lr 1.10e-03 | grad 0.2230 +2026-04-10 08:44:16 - INFO - parrotllm.training - step 45500 | epoch 0 | loss 3.4006 | lr 1.10e-03 | grad 0.2423 +2026-04-10 08:44:16 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:44:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:44:19 - INFO - parrotllm.training - Train: loss=3.4006, ppl=29.98 +2026-04-10 08:44:19 - INFO - parrotllm.training - Val: loss=3.3725, ppl=29.15 +2026-04-10 08:44:20 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3725_epoch_0000_step_0045500.pt +2026-04-10 08:44:22 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3701, min_delta=0.001000). +2026-04-10 08:44:22 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:44:25 - INFO - parrotllm.training - step 45510 | epoch 0 | loss 3.4991 | lr 1.10e-03 | grad 0.2152 +2026-04-10 08:44:28 - INFO - parrotllm.training - step 45520 | epoch 0 | loss 3.3104 | lr 1.10e-03 | grad 0.2127 +2026-04-10 08:44:31 - INFO - parrotllm.training - step 45530 | epoch 0 | loss 3.4794 | lr 1.10e-03 | grad 0.2203 +2026-04-10 08:44:34 - INFO - parrotllm.training - step 45540 | epoch 0 | loss 3.3764 | lr 1.10e-03 | grad 0.2111 +2026-04-10 08:44:37 - INFO - parrotllm.training - step 45550 | epoch 0 | loss 3.3016 | lr 1.10e-03 | grad 0.2330 +2026-04-10 08:44:40 - INFO - parrotllm.training - step 45560 | epoch 0 | loss 3.4670 | lr 1.10e-03 | grad 0.2309 +2026-04-10 08:44:43 - INFO - parrotllm.training - step 45570 | epoch 0 | loss 3.5492 | lr 1.10e-03 | grad 0.2331 +2026-04-10 08:44:46 - INFO - parrotllm.training - step 45580 | epoch 0 | loss 3.3522 | lr 1.10e-03 | grad 0.2233 +2026-04-10 08:44:49 - INFO - parrotllm.training - step 45590 | epoch 0 | loss 3.4222 | lr 1.10e-03 | grad 0.2612 +2026-04-10 08:44:52 - INFO - parrotllm.training - step 45600 | epoch 0 | loss 3.2985 | lr 1.10e-03 | grad 0.2051 +2026-04-10 08:44:55 - INFO - parrotllm.training - step 45610 | epoch 0 | loss 3.3760 | lr 1.10e-03 | grad 0.2233 +2026-04-10 08:44:58 - INFO - parrotllm.training - step 45620 | epoch 0 | loss 3.4222 | lr 1.10e-03 | grad 0.2343 +2026-04-10 08:45:01 - INFO - parrotllm.training - step 45630 | epoch 0 | loss 3.4217 | lr 1.10e-03 | grad 0.2649 +2026-04-10 08:45:04 - INFO - parrotllm.training - step 45640 | epoch 0 | loss 3.4325 | lr 1.10e-03 | grad 0.2293 +2026-04-10 08:45:07 - INFO - parrotllm.training - step 45650 | epoch 0 | loss 3.3877 | lr 1.10e-03 | grad 0.2243 +2026-04-10 08:45:10 - INFO - parrotllm.training - step 45660 | epoch 0 | loss 3.4845 | lr 1.10e-03 | grad 0.2539 +2026-04-10 08:45:13 - INFO - parrotllm.training - step 45670 | epoch 0 | loss 3.5641 | lr 1.10e-03 | grad 0.2674 +2026-04-10 08:45:17 - INFO - parrotllm.training - step 45680 | epoch 0 | loss 3.4375 | lr 1.10e-03 | grad 0.2186 +2026-04-10 08:45:20 - INFO - parrotllm.training - step 45690 | epoch 0 | loss 3.4363 | lr 1.10e-03 | grad 0.2334 +2026-04-10 08:45:23 - INFO - parrotllm.training - step 45700 | epoch 0 | loss 3.4299 | lr 1.10e-03 | grad 0.2282 +2026-04-10 08:45:26 - INFO - parrotllm.training - step 45710 | epoch 0 | loss 3.6217 | lr 1.10e-03 | grad 0.2199 +2026-04-10 08:45:29 - INFO - parrotllm.training - step 45720 | epoch 0 | loss 3.3925 | lr 1.10e-03 | grad 0.2187 +2026-04-10 08:45:32 - INFO - parrotllm.training - step 45730 | epoch 0 | loss 3.4444 | lr 1.10e-03 | grad 0.2280 +2026-04-10 08:45:35 - INFO - parrotllm.training - step 45740 | epoch 0 | loss 3.5602 | lr 1.10e-03 | grad 0.2102 +2026-04-10 08:45:38 - INFO - parrotllm.training - step 45750 | epoch 0 | loss 3.4037 | lr 1.10e-03 | grad 0.2146 +2026-04-10 08:45:41 - INFO - parrotllm.training - step 45760 | epoch 0 | loss 3.4153 | lr 1.10e-03 | grad 0.2251 +2026-04-10 08:45:44 - INFO - parrotllm.training - step 45770 | epoch 0 | loss 3.4465 | lr 1.10e-03 | grad 0.2375 +2026-04-10 08:45:47 - INFO - parrotllm.training - step 45780 | epoch 0 | loss 3.4064 | lr 1.10e-03 | grad 0.2302 +2026-04-10 08:45:50 - INFO - parrotllm.training - step 45790 | epoch 0 | loss 3.4053 | lr 1.10e-03 | grad 0.2241 +2026-04-10 08:45:53 - INFO - parrotllm.training - step 45800 | epoch 0 | loss 3.4563 | lr 1.10e-03 | grad 0.2261 +2026-04-10 08:45:56 - INFO - parrotllm.training - step 45810 | epoch 0 | loss 3.4064 | lr 1.10e-03 | grad 0.2345 +2026-04-10 08:45:59 - INFO - parrotllm.training - step 45820 | epoch 0 | loss 3.4384 | lr 1.10e-03 | grad 0.2441 +2026-04-10 08:46:02 - INFO - parrotllm.training - step 45830 | epoch 0 | loss 3.3545 | lr 1.10e-03 | grad 0.2373 +2026-04-10 08:46:05 - INFO - parrotllm.training - step 45840 | epoch 0 | loss 3.4483 | lr 1.10e-03 | grad 0.2412 +2026-04-10 08:46:08 - INFO - parrotllm.training - step 45850 | epoch 0 | loss 3.4114 | lr 1.10e-03 | grad 0.2195 +2026-04-10 08:46:11 - INFO - parrotllm.training - step 45860 | epoch 0 | loss 3.4172 | lr 1.10e-03 | grad 0.2199 +2026-04-10 08:46:14 - INFO - parrotllm.training - step 45870 | epoch 0 | loss 3.4307 | lr 1.10e-03 | grad 0.2101 +2026-04-10 08:46:17 - INFO - parrotllm.training - step 45880 | epoch 0 | loss 3.3730 | lr 1.10e-03 | grad 0.2880 +2026-04-10 08:46:20 - INFO - parrotllm.training - step 45890 | epoch 0 | loss 3.3970 | lr 1.10e-03 | grad 0.2318 +2026-04-10 08:46:24 - INFO - parrotllm.training - step 45900 | epoch 0 | loss 3.4270 | lr 1.10e-03 | grad 0.2555 +2026-04-10 08:46:27 - INFO - parrotllm.training - step 45910 | epoch 0 | loss 3.3906 | lr 1.10e-03 | grad 0.2186 +2026-04-10 08:46:30 - INFO - parrotllm.training - step 45920 | epoch 0 | loss 3.4923 | lr 1.10e-03 | grad 0.2681 +2026-04-10 08:46:33 - INFO - parrotllm.training - step 45930 | epoch 0 | loss 3.4609 | lr 1.10e-03 | grad 0.2309 +2026-04-10 08:46:36 - INFO - parrotllm.training - step 45940 | epoch 0 | loss 3.3927 | lr 1.10e-03 | grad 0.2078 +2026-04-10 08:46:39 - INFO - parrotllm.training - step 45950 | epoch 0 | loss 3.4791 | lr 1.10e-03 | grad 0.2196 +2026-04-10 08:46:42 - INFO - parrotllm.training - step 45960 | epoch 0 | loss 3.4078 | lr 1.10e-03 | grad 0.2208 +2026-04-10 08:46:45 - INFO - parrotllm.training - step 45970 | epoch 0 | loss 3.4279 | lr 1.10e-03 | grad 0.2251 +2026-04-10 08:46:48 - INFO - parrotllm.training - step 45980 | epoch 0 | loss 3.3927 | lr 1.10e-03 | grad 0.2418 +2026-04-10 08:46:51 - INFO - parrotllm.training - step 45990 | epoch 0 | loss 3.3232 | lr 1.10e-03 | grad 0.2322 +2026-04-10 08:46:54 - INFO - parrotllm.training - step 46000 | epoch 0 | loss 3.3973 | lr 1.09e-03 | grad 0.2186 +2026-04-10 08:46:54 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:46:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:46:57 - INFO - parrotllm.training - Train: loss=3.3973, ppl=29.88 +2026-04-10 08:46:57 - INFO - parrotllm.training - Val: loss=3.3719, ppl=29.13 +2026-04-10 08:46:58 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3719_epoch_0000_step_0046000.pt +2026-04-10 08:46:59 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3701, min_delta=0.001000). +2026-04-10 08:46:59 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:47:02 - INFO - parrotllm.training - step 46010 | epoch 0 | loss 3.4111 | lr 1.09e-03 | grad 0.2144 +2026-04-10 08:47:05 - INFO - parrotllm.training - step 46020 | epoch 0 | loss 3.3836 | lr 1.09e-03 | grad 0.2328 +2026-04-10 08:47:09 - INFO - parrotllm.training - step 46030 | epoch 0 | loss 3.4590 | lr 1.09e-03 | grad 0.2350 +2026-04-10 08:47:12 - INFO - parrotllm.training - step 46040 | epoch 0 | loss 3.4600 | lr 1.09e-03 | grad 0.2223 +2026-04-10 08:47:15 - INFO - parrotllm.training - step 46050 | epoch 0 | loss 3.4178 | lr 1.09e-03 | grad 0.2223 +2026-04-10 08:47:18 - INFO - parrotllm.training - step 46060 | epoch 0 | loss 3.4953 | lr 1.09e-03 | grad 0.2325 +2026-04-10 08:47:21 - INFO - parrotllm.training - step 46070 | epoch 0 | loss 3.3960 | lr 1.09e-03 | grad 0.2485 +2026-04-10 08:47:24 - INFO - parrotllm.training - step 46080 | epoch 0 | loss 3.3059 | lr 1.09e-03 | grad 0.2297 +2026-04-10 08:47:27 - INFO - parrotllm.training - step 46090 | epoch 0 | loss 3.4577 | lr 1.09e-03 | grad 0.2184 +2026-04-10 08:47:30 - INFO - parrotllm.training - step 46100 | epoch 0 | loss 3.4269 | lr 1.09e-03 | grad 0.2496 +2026-04-10 08:47:33 - INFO - parrotllm.training - step 46110 | epoch 0 | loss 3.3459 | lr 1.09e-03 | grad 0.2376 +2026-04-10 08:47:36 - INFO - parrotllm.training - step 46120 | epoch 0 | loss 3.4302 | lr 1.09e-03 | grad 0.2093 +2026-04-10 08:47:39 - INFO - parrotllm.training - step 46130 | epoch 0 | loss 3.4869 | lr 1.09e-03 | grad 0.2407 +2026-04-10 08:47:42 - INFO - parrotllm.training - step 46140 | epoch 0 | loss 3.3455 | lr 1.09e-03 | grad 0.2215 +2026-04-10 08:47:45 - INFO - parrotllm.training - step 46150 | epoch 0 | loss 3.3985 | lr 1.09e-03 | grad 0.2084 +2026-04-10 08:47:48 - INFO - parrotllm.training - step 46160 | epoch 0 | loss 3.4710 | lr 1.09e-03 | grad 0.2036 +2026-04-10 08:47:51 - INFO - parrotllm.training - step 46170 | epoch 0 | loss 3.5099 | lr 1.09e-03 | grad 0.2300 +2026-04-10 08:47:54 - INFO - parrotllm.training - step 46180 | epoch 0 | loss 3.4038 | lr 1.09e-03 | grad 0.2467 +2026-04-10 08:47:57 - INFO - parrotllm.training - step 46190 | epoch 0 | loss 3.4228 | lr 1.09e-03 | grad 0.2393 +2026-04-10 08:48:00 - INFO - parrotllm.training - step 46200 | epoch 0 | loss 3.4591 | lr 1.09e-03 | grad 0.2634 +2026-04-10 08:48:03 - INFO - parrotllm.training - step 46210 | epoch 0 | loss 3.5436 | lr 1.09e-03 | grad 0.2168 +2026-04-10 08:48:06 - INFO - parrotllm.training - step 46220 | epoch 0 | loss 3.4498 | lr 1.09e-03 | grad 0.2325 +2026-04-10 08:48:10 - INFO - parrotllm.training - step 46230 | epoch 0 | loss 3.3897 | lr 1.09e-03 | grad 0.2173 +2026-04-10 08:48:13 - INFO - parrotllm.training - step 46240 | epoch 0 | loss 3.3763 | lr 1.09e-03 | grad 0.2213 +2026-04-10 08:48:16 - INFO - parrotllm.training - step 46250 | epoch 0 | loss 3.4936 | lr 1.09e-03 | grad 0.2400 +2026-04-10 08:48:19 - INFO - parrotllm.training - step 46260 | epoch 0 | loss 3.4138 | lr 1.09e-03 | grad 0.2281 +2026-04-10 08:48:22 - INFO - parrotllm.training - step 46270 | epoch 0 | loss 3.5093 | lr 1.09e-03 | grad 0.2125 +2026-04-10 08:48:25 - INFO - parrotllm.training - step 46280 | epoch 0 | loss 3.5052 | lr 1.09e-03 | grad 0.2178 +2026-04-10 08:48:28 - INFO - parrotllm.training - step 46290 | epoch 0 | loss 3.4000 | lr 1.09e-03 | grad 0.2232 +2026-04-10 08:48:31 - INFO - parrotllm.training - step 46300 | epoch 0 | loss 3.4200 | lr 1.09e-03 | grad 0.2227 +2026-04-10 08:48:34 - INFO - parrotllm.training - step 46310 | epoch 0 | loss 3.4477 | lr 1.09e-03 | grad 0.2035 +2026-04-10 08:48:37 - INFO - parrotllm.training - step 46320 | epoch 0 | loss 3.4840 | lr 1.09e-03 | grad 0.2162 +2026-04-10 08:48:40 - INFO - parrotllm.training - step 46330 | epoch 0 | loss 3.4549 | lr 1.09e-03 | grad 0.2094 +2026-04-10 08:48:43 - INFO - parrotllm.training - step 46340 | epoch 0 | loss 3.4295 | lr 1.09e-03 | grad 0.2321 +2026-04-10 08:48:46 - INFO - parrotllm.training - step 46350 | epoch 0 | loss 3.5115 | lr 1.09e-03 | grad 0.2612 +2026-04-10 08:48:49 - INFO - parrotllm.training - step 46360 | epoch 0 | loss 3.4294 | lr 1.09e-03 | grad 0.2002 +2026-04-10 08:48:52 - INFO - parrotllm.training - step 46370 | epoch 0 | loss 3.3479 | lr 1.09e-03 | grad 0.2102 +2026-04-10 08:48:55 - INFO - parrotllm.training - step 46380 | epoch 0 | loss 3.4619 | lr 1.09e-03 | grad 0.2430 +2026-04-10 08:48:58 - INFO - parrotllm.training - step 46390 | epoch 0 | loss 3.3099 | lr 1.09e-03 | grad 0.2135 +2026-04-10 08:49:01 - INFO - parrotllm.training - step 46400 | epoch 0 | loss 3.4391 | lr 1.09e-03 | grad 0.2451 +2026-04-10 08:49:04 - INFO - parrotllm.training - step 46410 | epoch 0 | loss 3.4212 | lr 1.09e-03 | grad 0.2499 +2026-04-10 08:49:08 - INFO - parrotllm.training - step 46420 | epoch 0 | loss 3.4721 | lr 1.09e-03 | grad 0.2219 +2026-04-10 08:49:11 - INFO - parrotllm.training - step 46430 | epoch 0 | loss 3.4015 | lr 1.09e-03 | grad 0.2051 +2026-04-10 08:49:14 - INFO - parrotllm.training - step 46440 | epoch 0 | loss 3.3914 | lr 1.09e-03 | grad 0.2084 +2026-04-10 08:49:17 - INFO - parrotllm.training - step 46450 | epoch 0 | loss 3.4548 | lr 1.09e-03 | grad 0.2500 +2026-04-10 08:49:20 - INFO - parrotllm.training - step 46460 | epoch 0 | loss 3.4354 | lr 1.09e-03 | grad 0.2202 +2026-04-10 08:49:23 - INFO - parrotllm.training - step 46470 | epoch 0 | loss 3.4239 | lr 1.09e-03 | grad 0.2256 +2026-04-10 08:49:26 - INFO - parrotllm.training - step 46480 | epoch 0 | loss 3.2970 | lr 1.09e-03 | grad 0.2331 +2026-04-10 08:49:29 - INFO - parrotllm.training - step 46490 | epoch 0 | loss 3.5058 | lr 1.09e-03 | grad 0.2231 +2026-04-10 08:49:32 - INFO - parrotllm.training - step 46500 | epoch 0 | loss 3.4816 | lr 1.09e-03 | grad 0.2183 +2026-04-10 08:49:32 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:49:32 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:49:35 - INFO - parrotllm.training - Train: loss=3.4816, ppl=32.51 +2026-04-10 08:49:35 - INFO - parrotllm.training - Val: loss=3.3702, ppl=29.08 +2026-04-10 08:49:36 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3702_epoch_0000_step_0046500.pt +2026-04-10 08:49:37 - INFO - parrotllm.training - No validation improvement for 3/15 evaluation(s) (best=3.3701, min_delta=0.001000). +2026-04-10 08:49:37 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:49:40 - INFO - parrotllm.training - step 46510 | epoch 0 | loss 3.4350 | lr 1.09e-03 | grad 0.2232 +2026-04-10 08:49:43 - INFO - parrotllm.training - step 46520 | epoch 0 | loss 3.3941 | lr 1.09e-03 | grad 0.2313 +2026-04-10 08:49:46 - INFO - parrotllm.training - step 46530 | epoch 0 | loss 3.3837 | lr 1.09e-03 | grad 0.2121 +2026-04-10 08:49:50 - INFO - parrotllm.training - step 46540 | epoch 0 | loss 3.3884 | lr 1.09e-03 | grad 0.2279 +2026-04-10 08:49:53 - INFO - parrotllm.training - step 46550 | epoch 0 | loss 3.4957 | lr 1.09e-03 | grad 0.2270 +2026-04-10 08:49:56 - INFO - parrotllm.training - step 46560 | epoch 0 | loss 3.4234 | lr 1.09e-03 | grad 0.2183 +2026-04-10 08:49:56 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 46580/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 08:49:59 - INFO - parrotllm.training - step 46570 | epoch 0 | loss 3.5222 | lr 1.09e-03 | grad 0.2281 +2026-04-10 08:50:02 - INFO - parrotllm.training - step 46580 | epoch 0 | loss 3.4749 | lr 1.09e-03 | grad 0.2655 +2026-04-10 08:50:05 - INFO - parrotllm.training - step 46590 | epoch 0 | loss 3.4134 | lr 1.09e-03 | grad 0.2229 +2026-04-10 08:50:08 - INFO - parrotllm.training - step 46600 | epoch 0 | loss 3.4041 | lr 1.09e-03 | grad 0.2336 +2026-04-10 08:50:11 - INFO - parrotllm.training - step 46610 | epoch 0 | loss 3.4153 | lr 1.08e-03 | grad 0.2251 +2026-04-10 08:50:14 - INFO - parrotllm.training - step 46620 | epoch 0 | loss 3.4505 | lr 1.08e-03 | grad 0.2582 +2026-04-10 08:50:17 - INFO - parrotllm.training - step 46630 | epoch 0 | loss 3.3619 | lr 1.08e-03 | grad 0.2283 +2026-04-10 08:50:20 - INFO - parrotllm.training - step 46640 | epoch 0 | loss 3.4001 | lr 1.08e-03 | grad 0.2208 +2026-04-10 08:50:23 - INFO - parrotllm.training - step 46650 | epoch 0 | loss 3.5386 | lr 1.08e-03 | grad 0.2197 +2026-04-10 08:50:26 - INFO - parrotllm.training - step 46660 | epoch 0 | loss 3.3910 | lr 1.08e-03 | grad 0.2337 +2026-04-10 08:50:30 - INFO - parrotllm.training - step 46670 | epoch 0 | loss 3.3809 | lr 1.08e-03 | grad 0.2396 +2026-04-10 08:50:33 - INFO - parrotllm.training - step 46680 | epoch 0 | loss 3.3879 | lr 1.08e-03 | grad 0.2150 +2026-04-10 08:50:36 - INFO - parrotllm.training - step 46690 | epoch 0 | loss 3.5512 | lr 1.08e-03 | grad 0.2262 +2026-04-10 08:50:39 - INFO - parrotllm.training - step 46700 | epoch 0 | loss 3.4312 | lr 1.08e-03 | grad 0.2231 +2026-04-10 08:50:42 - INFO - parrotllm.training - step 46710 | epoch 0 | loss 3.4096 | lr 1.08e-03 | grad 0.2401 +2026-04-10 08:50:45 - INFO - parrotllm.training - step 46720 | epoch 0 | loss 3.3841 | lr 1.08e-03 | grad 0.2580 +2026-04-10 08:50:48 - INFO - parrotllm.training - step 46730 | epoch 0 | loss 3.3925 | lr 1.08e-03 | grad 0.2361 +2026-04-10 08:50:51 - INFO - parrotllm.training - step 46740 | epoch 0 | loss 3.3792 | lr 1.08e-03 | grad 0.2105 +2026-04-10 08:50:54 - INFO - parrotllm.training - step 46750 | epoch 0 | loss 3.4666 | lr 1.08e-03 | grad 0.2095 +2026-04-10 08:50:57 - INFO - parrotllm.training - step 46760 | epoch 0 | loss 3.4354 | lr 1.08e-03 | grad 0.2362 +2026-04-10 08:51:00 - INFO - parrotllm.training - step 46770 | epoch 0 | loss 3.3924 | lr 1.08e-03 | grad 0.2353 +2026-04-10 08:51:03 - INFO - parrotllm.training - step 46780 | epoch 0 | loss 3.4909 | lr 1.08e-03 | grad 0.2344 +2026-04-10 08:51:06 - INFO - parrotllm.training - step 46790 | epoch 0 | loss 3.4149 | lr 1.08e-03 | grad 0.2579 +2026-04-10 08:51:09 - INFO - parrotllm.training - step 46800 | epoch 0 | loss 3.3657 | lr 1.08e-03 | grad 0.2141 +2026-04-10 08:51:12 - INFO - parrotllm.training - step 46810 | epoch 0 | loss 3.5202 | lr 1.08e-03 | grad 0.2286 +2026-04-10 08:51:15 - INFO - parrotllm.training - step 46820 | epoch 0 | loss 3.3965 | lr 1.08e-03 | grad 0.2620 +2026-04-10 08:51:18 - INFO - parrotllm.training - step 46830 | epoch 0 | loss 3.3817 | lr 1.08e-03 | grad 0.2216 +2026-04-10 08:51:21 - INFO - parrotllm.training - step 46840 | epoch 0 | loss 3.4335 | lr 1.08e-03 | grad 0.2168 +2026-04-10 08:51:24 - INFO - parrotllm.training - step 46850 | epoch 0 | loss 3.4830 | lr 1.08e-03 | grad 0.2218 +2026-04-10 08:51:28 - INFO - parrotllm.training - step 46860 | epoch 0 | loss 3.4176 | lr 1.08e-03 | grad 0.2280 +2026-04-10 08:51:31 - INFO - parrotllm.training - step 46870 | epoch 0 | loss 3.4565 | lr 1.08e-03 | grad 0.2186 +2026-04-10 08:51:34 - INFO - parrotllm.training - step 46880 | epoch 0 | loss 3.3559 | lr 1.08e-03 | grad 0.2422 +2026-04-10 08:51:37 - INFO - parrotllm.training - step 46890 | epoch 0 | loss 3.3288 | lr 1.08e-03 | grad 0.2341 +2026-04-10 08:51:40 - INFO - parrotllm.training - step 46900 | epoch 0 | loss 3.3391 | lr 1.08e-03 | grad 0.2092 +2026-04-10 08:51:43 - INFO - parrotllm.training - step 46910 | epoch 0 | loss 3.3980 | lr 1.08e-03 | grad 0.2511 +2026-04-10 08:51:46 - INFO - parrotllm.training - step 46920 | epoch 0 | loss 3.3737 | lr 1.08e-03 | grad 0.2533 +2026-04-10 08:51:49 - INFO - parrotllm.training - step 46930 | epoch 0 | loss 3.3618 | lr 1.08e-03 | grad 0.2407 +2026-04-10 08:51:52 - INFO - parrotllm.training - step 46940 | epoch 0 | loss 3.4363 | lr 1.08e-03 | grad 0.2546 +2026-04-10 08:51:55 - INFO - parrotllm.training - step 46950 | epoch 0 | loss 3.4472 | lr 1.08e-03 | grad 0.2302 +2026-04-10 08:51:58 - INFO - parrotllm.training - step 46960 | epoch 0 | loss 3.3067 | lr 1.08e-03 | grad 0.2670 +2026-04-10 08:52:01 - INFO - parrotllm.training - step 46970 | epoch 0 | loss 3.3415 | lr 1.08e-03 | grad 0.2113 +2026-04-10 08:52:04 - INFO - parrotllm.training - step 46980 | epoch 0 | loss 3.3773 | lr 1.08e-03 | grad 0.2101 +2026-04-10 08:52:07 - INFO - parrotllm.training - step 46990 | epoch 0 | loss 3.4303 | lr 1.08e-03 | grad 0.2264 +2026-04-10 08:52:10 - INFO - parrotllm.training - step 47000 | epoch 0 | loss 3.4727 | lr 1.08e-03 | grad 0.2318 +2026-04-10 08:52:10 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:52:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:52:13 - INFO - parrotllm.training - Train: loss=3.4727, ppl=32.22 +2026-04-10 08:52:13 - INFO - parrotllm.training - Val: loss=3.3700, ppl=29.08 +2026-04-10 08:52:14 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3700_epoch_0000_step_0047000.pt +2026-04-10 08:52:16 - INFO - parrotllm.training - No validation improvement for 4/15 evaluation(s) (best=3.3701, min_delta=0.001000). +2026-04-10 08:52:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:52:19 - INFO - parrotllm.training - step 47010 | epoch 0 | loss 3.4407 | lr 1.08e-03 | grad 0.2184 +2026-04-10 08:52:22 - INFO - parrotllm.training - step 47020 | epoch 0 | loss 3.4071 | lr 1.08e-03 | grad 0.2551 +2026-04-10 08:52:25 - INFO - parrotllm.training - step 47030 | epoch 0 | loss 3.5521 | lr 1.08e-03 | grad 0.2195 +2026-04-10 08:52:28 - INFO - parrotllm.training - step 47040 | epoch 0 | loss 3.3854 | lr 1.08e-03 | grad 0.2315 +2026-04-10 08:52:31 - INFO - parrotllm.training - step 47050 | epoch 0 | loss 3.4544 | lr 1.08e-03 | grad 0.2196 +2026-04-10 08:52:34 - INFO - parrotllm.training - step 47060 | epoch 0 | loss 3.4098 | lr 1.08e-03 | grad 0.2362 +2026-04-10 08:52:37 - INFO - parrotllm.training - step 47070 | epoch 0 | loss 3.4980 | lr 1.08e-03 | grad 0.2066 +2026-04-10 08:52:40 - INFO - parrotllm.training - step 47080 | epoch 0 | loss 3.4128 | lr 1.08e-03 | grad 0.2123 +2026-04-10 08:52:43 - INFO - parrotllm.training - step 47090 | epoch 0 | loss 3.5134 | lr 1.08e-03 | grad 0.2240 +2026-04-10 08:52:46 - INFO - parrotllm.training - step 47100 | epoch 0 | loss 3.4875 | lr 1.08e-03 | grad 0.2425 +2026-04-10 08:52:49 - INFO - parrotllm.training - step 47110 | epoch 0 | loss 3.2768 | lr 1.08e-03 | grad 0.2495 +2026-04-10 08:52:52 - INFO - parrotllm.training - step 47120 | epoch 0 | loss 3.5219 | lr 1.08e-03 | grad 0.2359 +2026-04-10 08:52:55 - INFO - parrotllm.training - step 47130 | epoch 0 | loss 3.4802 | lr 1.08e-03 | grad 0.2276 +2026-04-10 08:52:58 - INFO - parrotllm.training - step 47140 | epoch 0 | loss 3.3460 | lr 1.08e-03 | grad 0.2203 +2026-04-10 08:53:01 - INFO - parrotllm.training - step 47150 | epoch 0 | loss 3.3971 | lr 1.08e-03 | grad 0.2320 +2026-04-10 08:53:05 - INFO - parrotllm.training - step 47160 | epoch 0 | loss 3.4660 | lr 1.08e-03 | grad 0.2581 +2026-04-10 08:53:08 - INFO - parrotllm.training - step 47170 | epoch 0 | loss 3.4227 | lr 1.08e-03 | grad 0.2372 +2026-04-10 08:53:11 - INFO - parrotllm.training - step 47180 | epoch 0 | loss 3.4355 | lr 1.08e-03 | grad 0.2372 +2026-04-10 08:53:14 - INFO - parrotllm.training - step 47190 | epoch 0 | loss 3.3468 | lr 1.08e-03 | grad 0.2165 +2026-04-10 08:53:17 - INFO - parrotllm.training - step 47200 | epoch 0 | loss 3.3974 | lr 1.08e-03 | grad 0.2189 +2026-04-10 08:53:20 - INFO - parrotllm.training - step 47210 | epoch 0 | loss 3.4656 | lr 1.08e-03 | grad 0.2432 +2026-04-10 08:53:23 - INFO - parrotllm.training - step 47220 | epoch 0 | loss 3.3915 | lr 1.08e-03 | grad 0.2209 +2026-04-10 08:53:26 - INFO - parrotllm.training - step 47230 | epoch 0 | loss 3.3805 | lr 1.07e-03 | grad 0.2404 +2026-04-10 08:53:29 - INFO - parrotllm.training - step 47240 | epoch 0 | loss 3.5017 | lr 1.07e-03 | grad 0.2400 +2026-04-10 08:53:32 - INFO - parrotllm.training - step 47250 | epoch 0 | loss 3.4229 | lr 1.07e-03 | grad 0.2506 +2026-04-10 08:53:35 - INFO - parrotllm.training - step 47260 | epoch 0 | loss 3.4514 | lr 1.07e-03 | grad 0.2334 +2026-04-10 08:53:38 - INFO - parrotllm.training - step 47270 | epoch 0 | loss 3.4856 | lr 1.07e-03 | grad 0.2037 +2026-04-10 08:53:41 - INFO - parrotllm.training - step 47280 | epoch 0 | loss 3.3693 | lr 1.07e-03 | grad 0.2304 +2026-04-10 08:53:44 - INFO - parrotllm.training - step 47290 | epoch 0 | loss 3.3325 | lr 1.07e-03 | grad 0.2348 +2026-04-10 08:53:47 - INFO - parrotllm.training - step 47300 | epoch 0 | loss 3.4769 | lr 1.07e-03 | grad 0.2652 +2026-04-10 08:53:50 - INFO - parrotllm.training - step 47310 | epoch 0 | loss 3.3253 | lr 1.07e-03 | grad 0.2443 +2026-04-10 08:53:53 - INFO - parrotllm.training - step 47320 | epoch 0 | loss 3.4193 | lr 1.07e-03 | grad 0.2164 +2026-04-10 08:53:56 - INFO - parrotllm.training - step 47330 | epoch 0 | loss 3.2691 | lr 1.07e-03 | grad 0.2204 +2026-04-10 08:53:59 - INFO - parrotllm.training - step 47340 | epoch 0 | loss 3.4888 | lr 1.07e-03 | grad 0.2434 +2026-04-10 08:54:02 - INFO - parrotllm.training - step 47350 | epoch 0 | loss 3.3518 | lr 1.07e-03 | grad 0.2119 +2026-04-10 08:54:06 - INFO - parrotllm.training - step 47360 | epoch 0 | loss 3.3286 | lr 1.07e-03 | grad 0.2370 +2026-04-10 08:54:09 - INFO - parrotllm.training - step 47370 | epoch 0 | loss 3.4177 | lr 1.07e-03 | grad 0.2652 +2026-04-10 08:54:12 - INFO - parrotllm.training - step 47380 | epoch 0 | loss 3.3455 | lr 1.07e-03 | grad 0.2663 +2026-04-10 08:54:15 - INFO - parrotllm.training - step 47390 | epoch 0 | loss 3.4253 | lr 1.07e-03 | grad 0.2324 +2026-04-10 08:54:18 - INFO - parrotllm.training - step 47400 | epoch 0 | loss 3.4503 | lr 1.07e-03 | grad 0.2341 +2026-04-10 08:54:21 - INFO - parrotllm.training - step 47410 | epoch 0 | loss 3.3587 | lr 1.07e-03 | grad 0.2199 +2026-04-10 08:54:24 - INFO - parrotllm.training - step 47420 | epoch 0 | loss 3.3966 | lr 1.07e-03 | grad 0.2176 +2026-04-10 08:54:27 - INFO - parrotllm.training - step 47430 | epoch 0 | loss 3.4543 | lr 1.07e-03 | grad 0.2560 +2026-04-10 08:54:30 - INFO - parrotllm.training - step 47440 | epoch 0 | loss 3.4402 | lr 1.07e-03 | grad 0.2301 +2026-04-10 08:54:33 - INFO - parrotllm.training - step 47450 | epoch 0 | loss 3.3077 | lr 1.07e-03 | grad 0.2236 +2026-04-10 08:54:36 - INFO - parrotllm.training - step 47460 | epoch 0 | loss 3.3132 | lr 1.07e-03 | grad 0.2460 +2026-04-10 08:54:39 - INFO - parrotllm.training - step 47470 | epoch 0 | loss 3.2813 | lr 1.07e-03 | grad 0.2190 +2026-04-10 08:54:42 - INFO - parrotllm.training - step 47480 | epoch 0 | loss 3.4535 | lr 1.07e-03 | grad 0.2076 +2026-04-10 08:54:45 - INFO - parrotllm.training - step 47490 | epoch 0 | loss 3.4283 | lr 1.07e-03 | grad 0.2113 +2026-04-10 08:54:48 - INFO - parrotllm.training - step 47500 | epoch 0 | loss 3.4319 | lr 1.07e-03 | grad 0.2636 +2026-04-10 08:54:48 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:54:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:54:51 - INFO - parrotllm.training - Train: loss=3.4319, ppl=30.94 +2026-04-10 08:54:51 - INFO - parrotllm.training - Val: loss=3.3680, ppl=29.02 +2026-04-10 08:54:51 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 08:54:52 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3680_epoch_0000_step_0047500.pt +2026-04-10 08:54:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:54:55 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0047500.pt +2026-04-10 08:55:00 - INFO - parrotllm.training - step 47510 | epoch 0 | loss 3.3047 | lr 1.07e-03 | grad 0.2129 +2026-04-10 08:55:03 - INFO - parrotllm.training - step 47520 | epoch 0 | loss 3.3866 | lr 1.07e-03 | grad 0.2168 +2026-04-10 08:55:06 - INFO - parrotllm.training - step 47530 | epoch 0 | loss 3.4568 | lr 1.07e-03 | grad 0.2437 +2026-04-10 08:55:09 - INFO - parrotllm.training - step 47540 | epoch 0 | loss 3.3636 | lr 1.07e-03 | grad 0.2153 +2026-04-10 08:55:12 - INFO - parrotllm.training - step 47550 | epoch 0 | loss 3.4296 | lr 1.07e-03 | grad 0.2144 +2026-04-10 08:55:15 - INFO - parrotllm.training - step 47560 | epoch 0 | loss 3.4027 | lr 1.07e-03 | grad 0.2399 +2026-04-10 08:55:18 - INFO - parrotllm.training - step 47570 | epoch 0 | loss 3.3826 | lr 1.07e-03 | grad 0.2131 +2026-04-10 08:55:21 - INFO - parrotllm.training - step 47580 | epoch 0 | loss 3.4686 | lr 1.07e-03 | grad 0.2363 +2026-04-10 08:55:25 - INFO - parrotllm.training - step 47590 | epoch 0 | loss 3.3140 | lr 1.07e-03 | grad 0.2201 +2026-04-10 08:55:28 - INFO - parrotllm.training - step 47600 | epoch 0 | loss 3.3562 | lr 1.07e-03 | grad 0.2641 +2026-04-10 08:55:31 - INFO - parrotllm.training - step 47610 | epoch 0 | loss 3.3579 | lr 1.07e-03 | grad 0.2335 +2026-04-10 08:55:34 - INFO - parrotllm.training - step 47620 | epoch 0 | loss 3.4269 | lr 1.07e-03 | grad 0.2716 +2026-04-10 08:55:37 - INFO - parrotllm.training - step 47630 | epoch 0 | loss 3.4343 | lr 1.07e-03 | grad 0.2047 +2026-04-10 08:55:40 - INFO - parrotllm.training - step 47640 | epoch 0 | loss 3.4155 | lr 1.07e-03 | grad 0.2286 +2026-04-10 08:55:43 - INFO - parrotllm.training - step 47650 | epoch 0 | loss 3.4314 | lr 1.07e-03 | grad 0.2181 +2026-04-10 08:55:46 - INFO - parrotllm.training - step 47660 | epoch 0 | loss 3.3665 | lr 1.07e-03 | grad 0.2458 +2026-04-10 08:55:49 - INFO - parrotllm.training - step 47670 | epoch 0 | loss 3.4133 | lr 1.07e-03 | grad 0.2380 +2026-04-10 08:55:52 - INFO - parrotllm.training - step 47680 | epoch 0 | loss 3.3926 | lr 1.07e-03 | grad 0.2126 +2026-04-10 08:55:55 - INFO - parrotllm.training - step 47690 | epoch 0 | loss 3.4162 | lr 1.07e-03 | grad 0.2238 +2026-04-10 08:55:58 - INFO - parrotllm.training - step 47700 | epoch 0 | loss 3.4286 | lr 1.07e-03 | grad 0.2545 +2026-04-10 08:56:01 - INFO - parrotllm.training - step 47710 | epoch 0 | loss 3.4427 | lr 1.07e-03 | grad 0.2477 +2026-04-10 08:56:04 - INFO - parrotllm.training - step 47720 | epoch 0 | loss 3.5289 | lr 1.07e-03 | grad 0.2447 +2026-04-10 08:56:07 - INFO - parrotllm.training - step 47730 | epoch 0 | loss 3.3853 | lr 1.07e-03 | grad 0.2599 +2026-04-10 08:56:10 - INFO - parrotllm.training - step 47740 | epoch 0 | loss 3.4946 | lr 1.07e-03 | grad 0.2194 +2026-04-10 08:56:13 - INFO - parrotllm.training - step 47750 | epoch 0 | loss 3.3444 | lr 1.07e-03 | grad 0.2406 +2026-04-10 08:56:16 - INFO - parrotllm.training - step 47760 | epoch 0 | loss 3.4979 | lr 1.07e-03 | grad 0.2110 +2026-04-10 08:56:19 - INFO - parrotllm.training - step 47770 | epoch 0 | loss 3.4019 | lr 1.07e-03 | grad 0.2596 +2026-04-10 08:56:22 - INFO - parrotllm.training - step 47780 | epoch 0 | loss 3.4574 | lr 1.07e-03 | grad 0.2229 +2026-04-10 08:56:26 - INFO - parrotllm.training - step 47790 | epoch 0 | loss 3.4880 | lr 1.07e-03 | grad 0.2088 +2026-04-10 08:56:29 - INFO - parrotllm.training - step 47800 | epoch 0 | loss 3.3282 | lr 1.07e-03 | grad 0.2742 +2026-04-10 08:56:32 - INFO - parrotllm.training - step 47810 | epoch 0 | loss 3.3450 | lr 1.07e-03 | grad 0.2601 +2026-04-10 08:56:35 - INFO - parrotllm.training - step 47820 | epoch 0 | loss 3.4582 | lr 1.07e-03 | grad 0.2589 +2026-04-10 08:56:38 - INFO - parrotllm.training - step 47830 | epoch 0 | loss 3.3658 | lr 1.07e-03 | grad 0.2057 +2026-04-10 08:56:41 - INFO - parrotllm.training - step 47840 | epoch 0 | loss 3.4930 | lr 1.06e-03 | grad 0.2496 +2026-04-10 08:56:44 - INFO - parrotllm.training - step 47850 | epoch 0 | loss 3.2993 | lr 1.06e-03 | grad 0.2175 +2026-04-10 08:56:47 - INFO - parrotllm.training - step 47860 | epoch 0 | loss 3.4836 | lr 1.06e-03 | grad 0.2198 +2026-04-10 08:56:50 - INFO - parrotllm.training - step 47870 | epoch 0 | loss 3.4272 | lr 1.06e-03 | grad 0.2487 +2026-04-10 08:56:53 - INFO - parrotllm.training - step 47880 | epoch 0 | loss 3.3753 | lr 1.06e-03 | grad 0.2542 +2026-04-10 08:56:56 - INFO - parrotllm.training - step 47890 | epoch 0 | loss 3.4318 | lr 1.06e-03 | grad 0.2061 +2026-04-10 08:56:59 - INFO - parrotllm.training - step 47900 | epoch 0 | loss 3.4119 | lr 1.06e-03 | grad 0.2346 +2026-04-10 08:57:02 - INFO - parrotllm.training - step 47910 | epoch 0 | loss 3.4217 | lr 1.06e-03 | grad 0.2451 +2026-04-10 08:57:05 - INFO - parrotllm.training - step 47920 | epoch 0 | loss 3.3933 | lr 1.06e-03 | grad 0.2098 +2026-04-10 08:57:08 - INFO - parrotllm.training - step 47930 | epoch 0 | loss 3.2617 | lr 1.06e-03 | grad 0.2635 +2026-04-10 08:57:11 - INFO - parrotllm.training - step 47940 | epoch 0 | loss 3.4428 | lr 1.06e-03 | grad 0.2291 +2026-04-10 08:57:14 - INFO - parrotllm.training - step 47950 | epoch 0 | loss 3.4728 | lr 1.06e-03 | grad 0.2254 +2026-04-10 08:57:17 - INFO - parrotllm.training - step 47960 | epoch 0 | loss 3.4123 | lr 1.06e-03 | grad 0.2196 +2026-04-10 08:57:20 - INFO - parrotllm.training - step 47970 | epoch 0 | loss 3.4577 | lr 1.06e-03 | grad 0.2429 +2026-04-10 08:57:23 - INFO - parrotllm.training - step 47980 | epoch 0 | loss 3.4813 | lr 1.06e-03 | grad 0.2880 +2026-04-10 08:57:26 - INFO - parrotllm.training - step 47990 | epoch 0 | loss 3.4022 | lr 1.06e-03 | grad 0.2132 +2026-04-10 08:57:30 - INFO - parrotllm.training - step 48000 | epoch 0 | loss 3.3915 | lr 1.06e-03 | grad 0.2309 +2026-04-10 08:57:30 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 08:57:30 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:57:33 - INFO - parrotllm.training - Train: loss=3.3915, ppl=29.71 +2026-04-10 08:57:33 - INFO - parrotllm.training - Val: loss=3.3661, ppl=28.97 +2026-04-10 08:57:33 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 08:57:34 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3661_epoch_0000_step_0048000.pt +2026-04-10 08:57:35 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 08:57:38 - INFO - parrotllm.training - step 48010 | epoch 0 | loss 3.4262 | lr 1.06e-03 | grad 0.2457 +2026-04-10 08:57:41 - INFO - parrotllm.training - step 48020 | epoch 0 | loss 3.4015 | lr 1.06e-03 | grad 0.2181 +2026-04-10 08:57:44 - INFO - parrotllm.training - step 48030 | epoch 0 | loss 3.3988 | lr 1.06e-03 | grad 0.2303 +2026-04-10 08:57:47 - INFO - parrotllm.training - step 48040 | epoch 0 | loss 3.4298 | lr 1.06e-03 | grad 0.2265 +2026-04-10 08:57:50 - INFO - parrotllm.training - step 48050 | epoch 0 | loss 3.3565 | lr 1.06e-03 | grad 0.2274 +2026-04-10 08:57:53 - INFO - parrotllm.training - step 48060 | epoch 0 | loss 3.4146 | lr 1.06e-03 | grad 0.2126 +2026-04-10 08:57:57 - INFO - parrotllm.training - step 48070 | epoch 0 | loss 3.4505 | lr 1.06e-03 | grad 0.2173 +2026-04-10 08:58:00 - INFO - parrotllm.training - step 48080 | epoch 0 | loss 3.4946 | lr 1.06e-03 | grad 0.2219 +2026-04-10 08:58:03 - INFO - parrotllm.training - step 48090 | epoch 0 | loss 3.3859 | lr 1.06e-03 | grad 0.2539 +2026-04-10 08:58:06 - INFO - parrotllm.training - step 48100 | epoch 0 | loss 3.4780 | lr 1.06e-03 | grad 0.2409 +2026-04-10 08:58:09 - INFO - parrotllm.training - step 48110 | epoch 0 | loss 3.3786 | lr 1.06e-03 | grad 0.2323 +2026-04-10 08:58:12 - INFO - parrotllm.training - step 48120 | epoch 0 | loss 3.4287 | lr 1.06e-03 | grad 0.2459 +2026-04-10 08:58:15 - INFO - parrotllm.training - step 48130 | epoch 0 | loss 3.4265 | lr 1.06e-03 | grad 0.2313 +2026-04-10 08:58:18 - INFO - parrotllm.training - step 48140 | epoch 0 | loss 3.2779 | lr 1.06e-03 | grad 0.2238 +2026-04-10 08:58:21 - INFO - parrotllm.training - step 48150 | epoch 0 | loss 3.4122 | lr 1.06e-03 | grad 0.2221 +2026-04-10 08:58:24 - INFO - parrotllm.training - step 48160 | epoch 0 | loss 3.3951 | lr 1.06e-03 | grad 0.2445 +2026-04-10 08:58:27 - INFO - parrotllm.training - step 48170 | epoch 0 | loss 3.3491 | lr 1.06e-03 | grad 0.2846 +2026-04-10 08:58:30 - INFO - parrotllm.training - step 48180 | epoch 0 | loss 3.4052 | lr 1.06e-03 | grad 0.2252 +2026-04-10 08:58:33 - INFO - parrotllm.training - step 48190 | epoch 0 | loss 3.3488 | lr 1.06e-03 | grad 0.2377 +2026-04-10 08:58:36 - INFO - parrotllm.training - step 48200 | epoch 0 | loss 3.3243 | lr 1.06e-03 | grad 0.2310 +2026-04-10 08:58:39 - INFO - parrotllm.training - step 48210 | epoch 0 | loss 3.3689 | lr 1.06e-03 | grad 0.2135 +2026-04-10 08:58:42 - INFO - parrotllm.training - step 48220 | epoch 0 | loss 3.4052 | lr 1.06e-03 | grad 0.2304 +2026-04-10 08:58:45 - INFO - parrotllm.training - step 48230 | epoch 0 | loss 3.5113 | lr 1.06e-03 | grad 0.2239 +2026-04-10 08:58:48 - INFO - parrotllm.training - step 48240 | epoch 0 | loss 3.4691 | lr 1.06e-03 | grad 0.2460 +2026-04-10 08:58:51 - INFO - parrotllm.training - step 48250 | epoch 0 | loss 3.4101 | lr 1.06e-03 | grad 0.2809 +2026-04-10 08:58:54 - INFO - parrotllm.training - step 48260 | epoch 0 | loss 3.4739 | lr 1.06e-03 | grad 0.2286 +2026-04-10 08:58:58 - INFO - parrotllm.training - step 48270 | epoch 0 | loss 3.3900 | lr 1.06e-03 | grad 0.2557 +2026-04-10 08:59:01 - INFO - parrotllm.training - step 48280 | epoch 0 | loss 3.5007 | lr 1.06e-03 | grad 0.2101 +2026-04-10 08:59:04 - INFO - parrotllm.training - step 48290 | epoch 0 | loss 3.3368 | lr 1.06e-03 | grad 0.2632 +2026-04-10 08:59:07 - INFO - parrotllm.training - step 48300 | epoch 0 | loss 3.3803 | lr 1.06e-03 | grad 0.2375 +2026-04-10 08:59:10 - INFO - parrotllm.training - step 48310 | epoch 0 | loss 3.5396 | lr 1.06e-03 | grad 0.2124 +2026-04-10 08:59:13 - INFO - parrotllm.training - step 48320 | epoch 0 | loss 3.4757 | lr 1.06e-03 | grad 0.2410 +2026-04-10 08:59:16 - INFO - parrotllm.training - step 48330 | epoch 0 | loss 3.4037 | lr 1.06e-03 | grad 0.2286 +2026-04-10 08:59:19 - INFO - parrotllm.training - step 48340 | epoch 0 | loss 3.4494 | lr 1.06e-03 | grad 0.2292 +2026-04-10 08:59:22 - INFO - parrotllm.training - step 48350 | epoch 0 | loss 3.4892 | lr 1.06e-03 | grad 0.2238 +2026-04-10 08:59:25 - INFO - parrotllm.training - step 48360 | epoch 0 | loss 3.4121 | lr 1.06e-03 | grad 0.2378 +2026-04-10 08:59:28 - INFO - parrotllm.training - step 48370 | epoch 0 | loss 3.4009 | lr 1.06e-03 | grad 0.2407 +2026-04-10 08:59:31 - INFO - parrotllm.training - step 48380 | epoch 0 | loss 3.3948 | lr 1.06e-03 | grad 0.2124 +2026-04-10 08:59:34 - INFO - parrotllm.training - step 48390 | epoch 0 | loss 3.4918 | lr 1.06e-03 | grad 0.2264 +2026-04-10 08:59:37 - INFO - parrotllm.training - step 48400 | epoch 0 | loss 3.4038 | lr 1.06e-03 | grad 0.2267 +2026-04-10 08:59:40 - INFO - parrotllm.training - step 48410 | epoch 0 | loss 3.4655 | lr 1.06e-03 | grad 0.2227 +2026-04-10 08:59:43 - INFO - parrotllm.training - step 48420 | epoch 0 | loss 3.3745 | lr 1.06e-03 | grad 0.2723 +2026-04-10 08:59:46 - INFO - parrotllm.training - step 48430 | epoch 0 | loss 3.3827 | lr 1.06e-03 | grad 0.2213 +2026-04-10 08:59:49 - INFO - parrotllm.training - step 48440 | epoch 0 | loss 3.4396 | lr 1.06e-03 | grad 0.2330 +2026-04-10 08:59:52 - INFO - parrotllm.training - step 48450 | epoch 0 | loss 3.5643 | lr 1.05e-03 | grad 0.2059 +2026-04-10 08:59:55 - INFO - parrotllm.training - step 48460 | epoch 0 | loss 3.4277 | lr 1.05e-03 | grad 0.2248 +2026-04-10 08:59:58 - INFO - parrotllm.training - step 48470 | epoch 0 | loss 3.2690 | lr 1.05e-03 | grad 0.2564 +2026-04-10 09:00:01 - INFO - parrotllm.training - step 48480 | epoch 0 | loss 3.3390 | lr 1.05e-03 | grad 0.2232 +2026-04-10 09:00:05 - INFO - parrotllm.training - step 48490 | epoch 0 | loss 3.4945 | lr 1.05e-03 | grad 0.2453 +2026-04-10 09:00:08 - INFO - parrotllm.training - step 48500 | epoch 0 | loss 3.3119 | lr 1.05e-03 | grad 0.2498 +2026-04-10 09:00:08 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:00:08 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:00:11 - INFO - parrotllm.training - Train: loss=3.3119, ppl=27.44 +2026-04-10 09:00:11 - INFO - parrotllm.training - Val: loss=3.3654, ppl=28.95 +2026-04-10 09:00:12 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3654_epoch_0000_step_0048500.pt +2026-04-10 09:00:13 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3661, min_delta=0.001000). +2026-04-10 09:00:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:00:16 - INFO - parrotllm.training - step 48510 | epoch 0 | loss 3.3518 | lr 1.05e-03 | grad 0.2368 +2026-04-10 09:00:19 - INFO - parrotllm.training - step 48520 | epoch 0 | loss 3.4023 | lr 1.05e-03 | grad 0.2113 +2026-04-10 09:00:22 - INFO - parrotllm.training - step 48530 | epoch 0 | loss 3.4950 | lr 1.05e-03 | grad 0.2240 +2026-04-10 09:00:25 - INFO - parrotllm.training - step 48540 | epoch 0 | loss 3.4492 | lr 1.05e-03 | grad 0.2149 +2026-04-10 09:00:28 - INFO - parrotllm.training - step 48550 | epoch 0 | loss 3.4463 | lr 1.05e-03 | grad 0.2372 +2026-04-10 09:00:31 - INFO - parrotllm.training - step 48560 | epoch 0 | loss 3.4195 | lr 1.05e-03 | grad 0.2147 +2026-04-10 09:00:35 - INFO - parrotllm.training - step 48570 | epoch 0 | loss 3.3760 | lr 1.05e-03 | grad 0.2420 +2026-04-10 09:00:38 - INFO - parrotllm.training - step 48580 | epoch 0 | loss 3.3883 | lr 1.05e-03 | grad 0.2228 +2026-04-10 09:00:41 - INFO - parrotllm.training - step 48590 | epoch 0 | loss 3.4093 | lr 1.05e-03 | grad 0.2248 +2026-04-10 09:00:44 - INFO - parrotllm.training - step 48600 | epoch 0 | loss 3.3573 | lr 1.05e-03 | grad 0.2527 +2026-04-10 09:00:47 - INFO - parrotllm.training - step 48610 | epoch 0 | loss 3.5406 | lr 1.05e-03 | grad 0.2295 +2026-04-10 09:00:50 - INFO - parrotllm.training - step 48620 | epoch 0 | loss 3.3723 | lr 1.05e-03 | grad 0.2240 +2026-04-10 09:00:53 - INFO - parrotllm.training - step 48630 | epoch 0 | loss 3.5682 | lr 1.05e-03 | grad 0.2332 +2026-04-10 09:00:56 - INFO - parrotllm.training - step 48640 | epoch 0 | loss 3.3857 | lr 1.05e-03 | grad 0.2194 +2026-04-10 09:00:59 - INFO - parrotllm.training - step 48650 | epoch 0 | loss 3.4178 | lr 1.05e-03 | grad 0.2238 +2026-04-10 09:01:02 - INFO - parrotllm.training - step 48660 | epoch 0 | loss 3.3560 | lr 1.05e-03 | grad 0.2240 +2026-04-10 09:01:05 - INFO - parrotllm.training - step 48670 | epoch 0 | loss 3.3574 | lr 1.05e-03 | grad 0.2350 +2026-04-10 09:01:08 - INFO - parrotllm.training - step 48680 | epoch 0 | loss 3.4450 | lr 1.05e-03 | grad 0.2453 +2026-04-10 09:01:11 - INFO - parrotllm.training - step 48690 | epoch 0 | loss 3.4845 | lr 1.05e-03 | grad 0.2510 +2026-04-10 09:01:14 - INFO - parrotllm.training - step 48700 | epoch 0 | loss 3.5056 | lr 1.05e-03 | grad 0.2244 +2026-04-10 09:01:17 - INFO - parrotllm.training - step 48710 | epoch 0 | loss 3.6048 | lr 1.05e-03 | grad 0.2519 +2026-04-10 09:01:20 - INFO - parrotllm.training - step 48720 | epoch 0 | loss 3.3517 | lr 1.05e-03 | grad 0.2384 +2026-04-10 09:01:23 - INFO - parrotllm.training - step 48730 | epoch 0 | loss 3.4454 | lr 1.05e-03 | grad 0.2507 +2026-04-10 09:01:26 - INFO - parrotllm.training - step 48740 | epoch 0 | loss 3.5864 | lr 1.05e-03 | grad 0.2198 +2026-04-10 09:01:29 - INFO - parrotllm.training - step 48750 | epoch 0 | loss 3.4342 | lr 1.05e-03 | grad 0.2224 +2026-04-10 09:01:32 - INFO - parrotllm.training - step 48760 | epoch 0 | loss 3.5685 | lr 1.05e-03 | grad 0.2203 +2026-04-10 09:01:36 - INFO - parrotllm.training - step 48770 | epoch 0 | loss 3.5049 | lr 1.05e-03 | grad 0.2186 +2026-04-10 09:01:39 - INFO - parrotllm.training - step 48780 | epoch 0 | loss 3.3739 | lr 1.05e-03 | grad 0.2398 +2026-04-10 09:01:42 - INFO - parrotllm.training - step 48790 | epoch 0 | loss 3.3564 | lr 1.05e-03 | grad 0.2866 +2026-04-10 09:01:45 - INFO - parrotllm.training - step 48800 | epoch 0 | loss 3.5122 | lr 1.05e-03 | grad 0.2396 +2026-04-10 09:01:48 - INFO - parrotllm.training - step 48810 | epoch 0 | loss 3.3545 | lr 1.05e-03 | grad 0.2398 +2026-04-10 09:01:51 - INFO - parrotllm.training - step 48820 | epoch 0 | loss 3.3949 | lr 1.05e-03 | grad 0.2224 +2026-04-10 09:01:54 - INFO - parrotllm.training - step 48830 | epoch 0 | loss 3.4255 | lr 1.05e-03 | grad 0.2345 +2026-04-10 09:01:57 - INFO - parrotllm.training - step 48840 | epoch 0 | loss 3.4360 | lr 1.05e-03 | grad 0.2281 +2026-04-10 09:02:00 - INFO - parrotllm.training - step 48850 | epoch 0 | loss 3.4428 | lr 1.05e-03 | grad 0.2362 +2026-04-10 09:02:03 - INFO - parrotllm.training - step 48860 | epoch 0 | loss 3.4329 | lr 1.05e-03 | grad 0.2100 +2026-04-10 09:02:06 - INFO - parrotllm.training - step 48870 | epoch 0 | loss 3.4203 | lr 1.05e-03 | grad 0.2104 +2026-04-10 09:02:09 - INFO - parrotllm.training - step 48880 | epoch 0 | loss 3.4562 | lr 1.05e-03 | grad 0.2360 +2026-04-10 09:02:12 - INFO - parrotllm.training - step 48890 | epoch 0 | loss 3.4602 | lr 1.05e-03 | grad 0.2155 +2026-04-10 09:02:15 - INFO - parrotllm.training - step 48900 | epoch 0 | loss 3.4205 | lr 1.05e-03 | grad 0.2360 +2026-04-10 09:02:18 - INFO - parrotllm.training - step 48910 | epoch 0 | loss 3.3508 | lr 1.05e-03 | grad 0.2272 +2026-04-10 09:02:21 - INFO - parrotllm.training - step 48920 | epoch 0 | loss 3.4029 | lr 1.05e-03 | grad 0.2390 +2026-04-10 09:02:24 - INFO - parrotllm.training - step 48930 | epoch 0 | loss 3.3958 | lr 1.05e-03 | grad 0.2241 +2026-04-10 09:02:27 - INFO - parrotllm.training - step 48940 | epoch 0 | loss 3.4617 | lr 1.05e-03 | grad 0.2071 +2026-04-10 09:02:30 - INFO - parrotllm.training - step 48950 | epoch 0 | loss 3.5091 | lr 1.05e-03 | grad 0.2535 +2026-04-10 09:02:34 - INFO - parrotllm.training - step 48960 | epoch 0 | loss 3.4966 | lr 1.05e-03 | grad 0.2325 +2026-04-10 09:02:37 - INFO - parrotllm.training - step 48970 | epoch 0 | loss 3.4183 | lr 1.05e-03 | grad 0.2211 +2026-04-10 09:02:40 - INFO - parrotllm.training - step 48980 | epoch 0 | loss 3.4218 | lr 1.05e-03 | grad 0.2539 +2026-04-10 09:02:43 - INFO - parrotllm.training - step 48990 | epoch 0 | loss 3.3312 | lr 1.05e-03 | grad 0.2292 +2026-04-10 09:02:46 - INFO - parrotllm.training - step 49000 | epoch 0 | loss 3.4590 | lr 1.05e-03 | grad 0.2191 +2026-04-10 09:02:46 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:02:46 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:02:49 - INFO - parrotllm.training - Train: loss=3.4590, ppl=31.79 +2026-04-10 09:02:49 - INFO - parrotllm.training - Val: loss=3.3639, ppl=28.90 +2026-04-10 09:02:49 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 09:02:50 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3639_epoch_0000_step_0049000.pt +2026-04-10 09:02:51 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:02:54 - INFO - parrotllm.training - step 49010 | epoch 0 | loss 3.3926 | lr 1.05e-03 | grad 0.2070 +2026-04-10 09:02:58 - INFO - parrotllm.training - step 49020 | epoch 0 | loss 3.3902 | lr 1.05e-03 | grad 0.2205 +2026-04-10 09:03:01 - INFO - parrotllm.training - step 49030 | epoch 0 | loss 3.3027 | lr 1.05e-03 | grad 0.2315 +2026-04-10 09:03:04 - INFO - parrotllm.training - step 49040 | epoch 0 | loss 3.4482 | lr 1.05e-03 | grad 0.2252 +2026-04-10 09:03:07 - INFO - parrotllm.training - step 49050 | epoch 0 | loss 3.4752 | lr 1.04e-03 | grad 0.2430 +2026-04-10 09:03:10 - INFO - parrotllm.training - step 49060 | epoch 0 | loss 3.4307 | lr 1.04e-03 | grad 0.2412 +2026-04-10 09:03:13 - INFO - parrotllm.training - step 49070 | epoch 0 | loss 3.4372 | lr 1.04e-03 | grad 0.2266 +2026-04-10 09:03:16 - INFO - parrotllm.training - step 49080 | epoch 0 | loss 3.3989 | lr 1.04e-03 | grad 0.2229 +2026-04-10 09:03:19 - INFO - parrotllm.training - step 49090 | epoch 0 | loss 3.4879 | lr 1.04e-03 | grad 0.2382 +2026-04-10 09:03:22 - INFO - parrotllm.training - step 49100 | epoch 0 | loss 3.4481 | lr 1.04e-03 | grad 0.2123 +2026-04-10 09:03:25 - INFO - parrotllm.training - step 49110 | epoch 0 | loss 3.4567 | lr 1.04e-03 | grad 0.2340 +2026-04-10 09:03:28 - INFO - parrotllm.training - step 49120 | epoch 0 | loss 3.3561 | lr 1.04e-03 | grad 0.2278 +2026-04-10 09:03:31 - INFO - parrotllm.training - step 49130 | epoch 0 | loss 3.4531 | lr 1.04e-03 | grad 0.2514 +2026-04-10 09:03:34 - INFO - parrotllm.training - step 49140 | epoch 0 | loss 3.4047 | lr 1.04e-03 | grad 0.2158 +2026-04-10 09:03:37 - INFO - parrotllm.training - step 49150 | epoch 0 | loss 3.4427 | lr 1.04e-03 | grad 0.2248 +2026-04-10 09:03:40 - INFO - parrotllm.training - step 49160 | epoch 0 | loss 3.3752 | lr 1.04e-03 | grad 0.2582 +2026-04-10 09:03:43 - INFO - parrotllm.training - step 49170 | epoch 0 | loss 3.3845 | lr 1.04e-03 | grad 0.2217 +2026-04-10 09:03:46 - INFO - parrotllm.training - step 49180 | epoch 0 | loss 3.4417 | lr 1.04e-03 | grad 0.2526 +2026-04-10 09:03:49 - INFO - parrotllm.training - step 49190 | epoch 0 | loss 3.5469 | lr 1.04e-03 | grad 0.2584 +2026-04-10 09:03:52 - INFO - parrotllm.training - step 49200 | epoch 0 | loss 3.4741 | lr 1.04e-03 | grad 0.2596 +2026-04-10 09:03:53 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 49221/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 09:03:56 - INFO - parrotllm.training - step 49210 | epoch 0 | loss 3.4633 | lr 1.04e-03 | grad 0.2387 +2026-04-10 09:03:59 - INFO - parrotllm.training - step 49220 | epoch 0 | loss 3.4729 | lr 1.04e-03 | grad 0.2551 +2026-04-10 09:04:02 - INFO - parrotllm.training - step 49230 | epoch 0 | loss 3.4918 | lr 1.04e-03 | grad 0.2220 +2026-04-10 09:04:05 - INFO - parrotllm.training - step 49240 | epoch 0 | loss 3.4963 | lr 1.04e-03 | grad 0.2775 +2026-04-10 09:04:08 - INFO - parrotllm.training - step 49250 | epoch 0 | loss 3.3755 | lr 1.04e-03 | grad 0.2112 +2026-04-10 09:04:11 - INFO - parrotllm.training - step 49260 | epoch 0 | loss 3.3679 | lr 1.04e-03 | grad 0.2452 +2026-04-10 09:04:14 - INFO - parrotllm.training - step 49270 | epoch 0 | loss 3.4186 | lr 1.04e-03 | grad 0.3162 +2026-04-10 09:04:17 - INFO - parrotllm.training - step 49280 | epoch 0 | loss 3.4570 | lr 1.04e-03 | grad 0.2183 +2026-04-10 09:04:20 - INFO - parrotllm.training - step 49290 | epoch 0 | loss 3.4351 | lr 1.04e-03 | grad 0.2288 +2026-04-10 09:04:23 - INFO - parrotllm.training - step 49300 | epoch 0 | loss 3.4964 | lr 1.04e-03 | grad 0.2461 +2026-04-10 09:04:26 - INFO - parrotllm.training - step 49310 | epoch 0 | loss 3.5436 | lr 1.04e-03 | grad 0.2396 +2026-04-10 09:04:29 - INFO - parrotllm.training - step 49320 | epoch 0 | loss 3.2909 | lr 1.04e-03 | grad 0.2479 +2026-04-10 09:04:32 - INFO - parrotllm.training - step 49330 | epoch 0 | loss 3.4774 | lr 1.04e-03 | grad 0.2412 +2026-04-10 09:04:35 - INFO - parrotllm.training - step 49340 | epoch 0 | loss 3.3926 | lr 1.04e-03 | grad 0.2334 +2026-04-10 09:04:38 - INFO - parrotllm.training - step 49350 | epoch 0 | loss 3.3803 | lr 1.04e-03 | grad 0.2233 +2026-04-10 09:04:42 - INFO - parrotllm.training - step 49360 | epoch 0 | loss 3.4587 | lr 1.04e-03 | grad 0.2378 +2026-04-10 09:04:45 - INFO - parrotllm.training - step 49370 | epoch 0 | loss 3.4957 | lr 1.04e-03 | grad 0.2256 +2026-04-10 09:04:48 - INFO - parrotllm.training - step 49380 | epoch 0 | loss 3.4109 | lr 1.04e-03 | grad 0.2707 +2026-04-10 09:04:51 - INFO - parrotllm.training - step 49390 | epoch 0 | loss 3.4191 | lr 1.04e-03 | grad 0.2364 +2026-04-10 09:04:54 - INFO - parrotllm.training - step 49400 | epoch 0 | loss 3.4699 | lr 1.04e-03 | grad 0.2300 +2026-04-10 09:04:57 - INFO - parrotllm.training - step 49410 | epoch 0 | loss 3.4731 | lr 1.04e-03 | grad 0.2170 +2026-04-10 09:05:00 - INFO - parrotllm.training - step 49420 | epoch 0 | loss 3.2867 | lr 1.04e-03 | grad 0.2470 +2026-04-10 09:05:03 - INFO - parrotllm.training - step 49430 | epoch 0 | loss 3.4194 | lr 1.04e-03 | grad 0.2582 +2026-04-10 09:05:06 - INFO - parrotllm.training - step 49440 | epoch 0 | loss 3.4964 | lr 1.04e-03 | grad 0.2258 +2026-04-10 09:05:09 - INFO - parrotllm.training - step 49450 | epoch 0 | loss 3.4258 | lr 1.04e-03 | grad 0.2119 +2026-04-10 09:05:12 - INFO - parrotllm.training - step 49460 | epoch 0 | loss 3.4428 | lr 1.04e-03 | grad 0.2432 +2026-04-10 09:05:15 - INFO - parrotllm.training - step 49470 | epoch 0 | loss 3.4584 | lr 1.04e-03 | grad 0.2296 +2026-04-10 09:05:18 - INFO - parrotllm.training - step 49480 | epoch 0 | loss 3.4052 | lr 1.04e-03 | grad 0.2398 +2026-04-10 09:05:21 - INFO - parrotllm.training - step 49490 | epoch 0 | loss 3.4085 | lr 1.04e-03 | grad 0.2373 +2026-04-10 09:05:24 - INFO - parrotllm.training - step 49500 | epoch 0 | loss 3.4772 | lr 1.04e-03 | grad 0.2373 +2026-04-10 09:05:24 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:05:24 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:05:27 - INFO - parrotllm.training - Train: loss=3.4772, ppl=32.37 +2026-04-10 09:05:27 - INFO - parrotllm.training - Val: loss=3.3631, ppl=28.88 +2026-04-10 09:05:28 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3631_epoch_0000_step_0049500.pt +2026-04-10 09:05:30 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3639, min_delta=0.001000). +2026-04-10 09:05:30 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:05:33 - INFO - parrotllm.training - step 49510 | epoch 0 | loss 3.5026 | lr 1.04e-03 | grad 0.2010 +2026-04-10 09:05:36 - INFO - parrotllm.training - step 49520 | epoch 0 | loss 3.4474 | lr 1.04e-03 | grad 0.2289 +2026-04-10 09:05:39 - INFO - parrotllm.training - step 49530 | epoch 0 | loss 3.4087 | lr 1.04e-03 | grad 0.2363 +2026-04-10 09:05:42 - INFO - parrotllm.training - step 49540 | epoch 0 | loss 3.3418 | lr 1.04e-03 | grad 0.2373 +2026-04-10 09:05:45 - INFO - parrotllm.training - step 49550 | epoch 0 | loss 3.4345 | lr 1.04e-03 | grad 0.2401 +2026-04-10 09:05:48 - INFO - parrotllm.training - step 49560 | epoch 0 | loss 3.6038 | lr 1.04e-03 | grad 0.2345 +2026-04-10 09:05:51 - INFO - parrotllm.training - step 49570 | epoch 0 | loss 3.4648 | lr 1.04e-03 | grad 0.2495 +2026-04-10 09:05:54 - INFO - parrotllm.training - step 49580 | epoch 0 | loss 3.2834 | lr 1.04e-03 | grad 0.2318 +2026-04-10 09:05:57 - INFO - parrotllm.training - step 49590 | epoch 0 | loss 3.3772 | lr 1.04e-03 | grad 0.2283 +2026-04-10 09:06:00 - INFO - parrotllm.training - step 49600 | epoch 0 | loss 3.4484 | lr 1.04e-03 | grad 0.2251 +2026-04-10 09:06:03 - INFO - parrotllm.training - step 49610 | epoch 0 | loss 3.3781 | lr 1.04e-03 | grad 0.2305 +2026-04-10 09:06:06 - INFO - parrotllm.training - step 49620 | epoch 0 | loss 3.4901 | lr 1.04e-03 | grad 0.2460 +2026-04-10 09:06:09 - INFO - parrotllm.training - step 49630 | epoch 0 | loss 3.4790 | lr 1.04e-03 | grad 0.2497 +2026-04-10 09:06:12 - INFO - parrotllm.training - step 49640 | epoch 0 | loss 3.3934 | lr 1.04e-03 | grad 0.2053 +2026-04-10 09:06:15 - INFO - parrotllm.training - step 49650 | epoch 0 | loss 3.4725 | lr 1.03e-03 | grad 0.2433 +2026-04-10 09:06:18 - INFO - parrotllm.training - step 49660 | epoch 0 | loss 3.3211 | lr 1.03e-03 | grad 0.2344 +2026-04-10 09:06:21 - INFO - parrotllm.training - step 49670 | epoch 0 | loss 3.3632 | lr 1.03e-03 | grad 0.2053 +2026-04-10 09:06:25 - INFO - parrotllm.training - step 49680 | epoch 0 | loss 3.3705 | lr 1.03e-03 | grad 0.2378 +2026-04-10 09:06:28 - INFO - parrotllm.training - step 49690 | epoch 0 | loss 3.4176 | lr 1.03e-03 | grad 0.2319 +2026-04-10 09:06:31 - INFO - parrotllm.training - step 49700 | epoch 0 | loss 3.3571 | lr 1.03e-03 | grad 0.2434 +2026-04-10 09:06:34 - INFO - parrotllm.training - step 49710 | epoch 0 | loss 3.4748 | lr 1.03e-03 | grad 0.2216 +2026-04-10 09:06:37 - INFO - parrotllm.training - step 49720 | epoch 0 | loss 3.2972 | lr 1.03e-03 | grad 0.2374 +2026-04-10 09:06:40 - INFO - parrotllm.training - step 49730 | epoch 0 | loss 3.3477 | lr 1.03e-03 | grad 0.2250 +2026-04-10 09:06:43 - INFO - parrotllm.training - step 49740 | epoch 0 | loss 3.4109 | lr 1.03e-03 | grad 0.2536 +2026-04-10 09:06:46 - INFO - parrotllm.training - step 49750 | epoch 0 | loss 3.5014 | lr 1.03e-03 | grad 0.2293 +2026-04-10 09:06:49 - INFO - parrotllm.training - step 49760 | epoch 0 | loss 3.3672 | lr 1.03e-03 | grad 0.2362 +2026-04-10 09:06:52 - INFO - parrotllm.training - step 49770 | epoch 0 | loss 3.4013 | lr 1.03e-03 | grad 0.2499 +2026-04-10 09:06:55 - INFO - parrotllm.training - step 49780 | epoch 0 | loss 3.3702 | lr 1.03e-03 | grad 0.2339 +2026-04-10 09:06:58 - INFO - parrotllm.training - step 49790 | epoch 0 | loss 3.5723 | lr 1.03e-03 | grad 0.2251 +2026-04-10 09:07:01 - INFO - parrotllm.training - step 49800 | epoch 0 | loss 3.4500 | lr 1.03e-03 | grad 0.2716 +2026-04-10 09:07:04 - INFO - parrotllm.training - step 49810 | epoch 0 | loss 3.3232 | lr 1.03e-03 | grad 0.2380 +2026-04-10 09:07:07 - INFO - parrotllm.training - step 49820 | epoch 0 | loss 3.3704 | lr 1.03e-03 | grad 0.2198 +2026-04-10 09:07:10 - INFO - parrotllm.training - step 49830 | epoch 0 | loss 3.4219 | lr 1.03e-03 | grad 0.2338 +2026-04-10 09:07:13 - INFO - parrotllm.training - step 49840 | epoch 0 | loss 3.4301 | lr 1.03e-03 | grad 0.2293 +2026-04-10 09:07:16 - INFO - parrotllm.training - step 49850 | epoch 0 | loss 3.4574 | lr 1.03e-03 | grad 0.2071 +2026-04-10 09:07:19 - INFO - parrotllm.training - step 49860 | epoch 0 | loss 3.4315 | lr 1.03e-03 | grad 0.2200 +2026-04-10 09:07:22 - INFO - parrotllm.training - step 49870 | epoch 0 | loss 3.3100 | lr 1.03e-03 | grad 0.2577 +2026-04-10 09:07:26 - INFO - parrotllm.training - step 49880 | epoch 0 | loss 3.4519 | lr 1.03e-03 | grad 0.2807 +2026-04-10 09:07:29 - INFO - parrotllm.training - step 49890 | epoch 0 | loss 3.4899 | lr 1.03e-03 | grad 0.2360 +2026-04-10 09:07:32 - INFO - parrotllm.training - step 49900 | epoch 0 | loss 3.4199 | lr 1.03e-03 | grad 0.2222 +2026-04-10 09:07:35 - INFO - parrotllm.training - step 49910 | epoch 0 | loss 3.4178 | lr 1.03e-03 | grad 0.2247 +2026-04-10 09:07:38 - INFO - parrotllm.training - step 49920 | epoch 0 | loss 3.3570 | lr 1.03e-03 | grad 0.2273 +2026-04-10 09:07:41 - INFO - parrotllm.training - step 49930 | epoch 0 | loss 3.4276 | lr 1.03e-03 | grad 0.2483 +2026-04-10 09:07:44 - INFO - parrotllm.training - step 49940 | epoch 0 | loss 3.3912 | lr 1.03e-03 | grad 0.2173 +2026-04-10 09:07:47 - INFO - parrotllm.training - step 49950 | epoch 0 | loss 3.3659 | lr 1.03e-03 | grad 0.2239 +2026-04-10 09:07:50 - INFO - parrotllm.training - step 49960 | epoch 0 | loss 3.4386 | lr 1.03e-03 | grad 0.2500 +2026-04-10 09:07:53 - INFO - parrotllm.training - step 49970 | epoch 0 | loss 3.4389 | lr 1.03e-03 | grad 0.2351 +2026-04-10 09:07:56 - INFO - parrotllm.training - step 49980 | epoch 0 | loss 3.3916 | lr 1.03e-03 | grad 0.2525 +2026-04-10 09:07:59 - INFO - parrotllm.training - step 49990 | epoch 0 | loss 3.2905 | lr 1.03e-03 | grad 0.2245 +2026-04-10 09:08:02 - INFO - parrotllm.training - step 50000 | epoch 0 | loss 3.3881 | lr 1.03e-03 | grad 0.2202 +2026-04-10 09:08:02 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:08:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:08:05 - INFO - parrotllm.training - Train: loss=3.3881, ppl=29.61 +2026-04-10 09:08:05 - INFO - parrotllm.training - Val: loss=3.3630, ppl=28.88 +2026-04-10 09:08:06 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3630_epoch_0000_step_0050000.pt +2026-04-10 09:08:08 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3639, min_delta=0.001000). +2026-04-10 09:08:08 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:08:09 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0050000.pt +2026-04-10 09:08:13 - INFO - parrotllm.training - step 50010 | epoch 0 | loss 3.3678 | lr 1.03e-03 | grad 0.2341 +2026-04-10 09:08:16 - INFO - parrotllm.training - step 50020 | epoch 0 | loss 3.3614 | lr 1.03e-03 | grad 0.2351 +2026-04-10 09:08:19 - INFO - parrotllm.training - step 50030 | epoch 0 | loss 3.5286 | lr 1.03e-03 | grad 0.2660 +2026-04-10 09:08:22 - INFO - parrotllm.training - step 50040 | epoch 0 | loss 3.3575 | lr 1.03e-03 | grad 0.2330 +2026-04-10 09:08:25 - INFO - parrotllm.training - step 50050 | epoch 0 | loss 3.3615 | lr 1.03e-03 | grad 0.2116 +2026-04-10 09:08:29 - INFO - parrotllm.training - step 50060 | epoch 0 | loss 3.4943 | lr 1.03e-03 | grad 0.2471 +2026-04-10 09:08:32 - INFO - parrotllm.training - step 50070 | epoch 0 | loss 3.4114 | lr 1.03e-03 | grad 0.2303 +2026-04-10 09:08:35 - INFO - parrotllm.training - step 50080 | epoch 0 | loss 3.3172 | lr 1.03e-03 | grad 0.2410 +2026-04-10 09:08:38 - INFO - parrotllm.training - step 50090 | epoch 0 | loss 3.4483 | lr 1.03e-03 | grad 0.2534 +2026-04-10 09:08:41 - INFO - parrotllm.training - step 50100 | epoch 0 | loss 3.4650 | lr 1.03e-03 | grad 0.2388 +2026-04-10 09:08:44 - INFO - parrotllm.training - step 50110 | epoch 0 | loss 3.3630 | lr 1.03e-03 | grad 0.2391 +2026-04-10 09:08:47 - INFO - parrotllm.training - step 50120 | epoch 0 | loss 3.4524 | lr 1.03e-03 | grad 0.2424 +2026-04-10 09:08:50 - INFO - parrotllm.training - step 50130 | epoch 0 | loss 3.4584 | lr 1.03e-03 | grad 0.2247 +2026-04-10 09:08:53 - INFO - parrotllm.training - step 50140 | epoch 0 | loss 3.3560 | lr 1.03e-03 | grad 0.2231 +2026-04-10 09:08:56 - INFO - parrotllm.training - step 50150 | epoch 0 | loss 3.3975 | lr 1.03e-03 | grad 0.2508 +2026-04-10 09:08:59 - INFO - parrotllm.training - step 50160 | epoch 0 | loss 3.3499 | lr 1.03e-03 | grad 0.2399 +2026-04-10 09:09:02 - INFO - parrotllm.training - step 50170 | epoch 0 | loss 3.3571 | lr 1.03e-03 | grad 0.2076 +2026-04-10 09:09:05 - INFO - parrotllm.training - step 50180 | epoch 0 | loss 3.3848 | lr 1.03e-03 | grad 0.2206 +2026-04-10 09:09:08 - INFO - parrotllm.training - step 50190 | epoch 0 | loss 3.4350 | lr 1.03e-03 | grad 0.2386 +2026-04-10 09:09:11 - INFO - parrotllm.training - step 50200 | epoch 0 | loss 3.3999 | lr 1.03e-03 | grad 0.2354 +2026-04-10 09:09:14 - INFO - parrotllm.training - step 50210 | epoch 0 | loss 3.4376 | lr 1.03e-03 | grad 0.2354 +2026-04-10 09:09:17 - INFO - parrotllm.training - step 50220 | epoch 0 | loss 3.3949 | lr 1.03e-03 | grad 0.2392 +2026-04-10 09:09:20 - INFO - parrotllm.training - step 50230 | epoch 0 | loss 3.4617 | lr 1.03e-03 | grad 0.2416 +2026-04-10 09:09:23 - INFO - parrotllm.training - step 50240 | epoch 0 | loss 3.3364 | lr 1.03e-03 | grad 0.2278 +2026-04-10 09:09:26 - INFO - parrotllm.training - step 50250 | epoch 0 | loss 3.3274 | lr 1.02e-03 | grad 0.2388 +2026-04-10 09:09:29 - INFO - parrotllm.training - step 50260 | epoch 0 | loss 3.4290 | lr 1.02e-03 | grad 0.2547 +2026-04-10 09:09:33 - INFO - parrotllm.training - step 50270 | epoch 0 | loss 3.4412 | lr 1.02e-03 | grad 0.2473 +2026-04-10 09:09:36 - INFO - parrotllm.training - step 50280 | epoch 0 | loss 3.3702 | lr 1.02e-03 | grad 0.2404 +2026-04-10 09:09:39 - INFO - parrotllm.training - step 50290 | epoch 0 | loss 3.4672 | lr 1.02e-03 | grad 0.2320 +2026-04-10 09:09:42 - INFO - parrotllm.training - step 50300 | epoch 0 | loss 3.3576 | lr 1.02e-03 | grad 0.2143 +2026-04-10 09:09:45 - INFO - parrotllm.training - step 50310 | epoch 0 | loss 3.4326 | lr 1.02e-03 | grad 0.2422 +2026-04-10 09:09:48 - INFO - parrotllm.training - step 50320 | epoch 0 | loss 3.4412 | lr 1.02e-03 | grad 0.2835 +2026-04-10 09:09:51 - INFO - parrotllm.training - step 50330 | epoch 0 | loss 3.3911 | lr 1.02e-03 | grad 0.2461 +2026-04-10 09:09:54 - INFO - parrotllm.training - step 50340 | epoch 0 | loss 3.4967 | lr 1.02e-03 | grad 0.2302 +2026-04-10 09:09:57 - INFO - parrotllm.training - step 50350 | epoch 0 | loss 3.3773 | lr 1.02e-03 | grad 0.2371 +2026-04-10 09:10:00 - INFO - parrotllm.training - step 50360 | epoch 0 | loss 3.3322 | lr 1.02e-03 | grad 0.2261 +2026-04-10 09:10:03 - INFO - parrotllm.training - step 50370 | epoch 0 | loss 3.4154 | lr 1.02e-03 | grad 0.2409 +2026-04-10 09:10:06 - INFO - parrotllm.training - step 50380 | epoch 0 | loss 3.4659 | lr 1.02e-03 | grad 0.2378 +2026-04-10 09:10:09 - INFO - parrotllm.training - step 50390 | epoch 0 | loss 3.3407 | lr 1.02e-03 | grad 0.2562 +2026-04-10 09:10:12 - INFO - parrotllm.training - step 50400 | epoch 0 | loss 3.3551 | lr 1.02e-03 | grad 0.2236 +2026-04-10 09:10:15 - INFO - parrotllm.training - step 50410 | epoch 0 | loss 3.3849 | lr 1.02e-03 | grad 0.2098 +2026-04-10 09:10:18 - INFO - parrotllm.training - step 50420 | epoch 0 | loss 3.3882 | lr 1.02e-03 | grad 0.2583 +2026-04-10 09:10:21 - INFO - parrotllm.training - step 50430 | epoch 0 | loss 3.3498 | lr 1.02e-03 | grad 0.2142 +2026-04-10 09:10:24 - INFO - parrotllm.training - step 50440 | epoch 0 | loss 3.4943 | lr 1.02e-03 | grad 0.2093 +2026-04-10 09:10:27 - INFO - parrotllm.training - step 50450 | epoch 0 | loss 3.4896 | lr 1.02e-03 | grad 0.2460 +2026-04-10 09:10:30 - INFO - parrotllm.training - step 50460 | epoch 0 | loss 3.3562 | lr 1.02e-03 | grad 0.3454 +2026-04-10 09:10:33 - INFO - parrotllm.training - step 50470 | epoch 0 | loss 3.4316 | lr 1.02e-03 | grad 0.2708 +2026-04-10 09:10:37 - INFO - parrotllm.training - step 50480 | epoch 0 | loss 3.4001 | lr 1.02e-03 | grad 0.2673 +2026-04-10 09:10:40 - INFO - parrotllm.training - step 50490 | epoch 0 | loss 3.3470 | lr 1.02e-03 | grad 0.2276 +2026-04-10 09:10:43 - INFO - parrotllm.training - step 50500 | epoch 0 | loss 3.5043 | lr 1.02e-03 | grad 0.2323 +2026-04-10 09:10:43 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:10:43 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:10:46 - INFO - parrotllm.training - Train: loss=3.5043, ppl=33.26 +2026-04-10 09:10:46 - INFO - parrotllm.training - Val: loss=3.3585, ppl=28.75 +2026-04-10 09:10:46 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 09:10:47 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3585_epoch_0000_step_0050500.pt +2026-04-10 09:10:49 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:10:52 - INFO - parrotllm.training - step 50510 | epoch 0 | loss 3.5693 | lr 1.02e-03 | grad 0.2263 +2026-04-10 09:10:55 - INFO - parrotllm.training - step 50520 | epoch 0 | loss 3.3918 | lr 1.02e-03 | grad 0.2419 +2026-04-10 09:10:58 - INFO - parrotllm.training - step 50530 | epoch 0 | loss 3.3578 | lr 1.02e-03 | grad 0.2277 +2026-04-10 09:11:01 - INFO - parrotllm.training - step 50540 | epoch 0 | loss 3.4398 | lr 1.02e-03 | grad 0.2406 +2026-04-10 09:11:04 - INFO - parrotllm.training - step 50550 | epoch 0 | loss 3.4158 | lr 1.02e-03 | grad 0.2302 +2026-04-10 09:11:07 - INFO - parrotllm.training - step 50560 | epoch 0 | loss 3.4219 | lr 1.02e-03 | grad 0.2803 +2026-04-10 09:11:10 - INFO - parrotllm.training - step 50570 | epoch 0 | loss 3.3986 | lr 1.02e-03 | grad 0.2349 +2026-04-10 09:11:13 - INFO - parrotllm.training - step 50580 | epoch 0 | loss 3.4028 | lr 1.02e-03 | grad 0.2335 +2026-04-10 09:11:16 - INFO - parrotllm.training - step 50590 | epoch 0 | loss 3.3676 | lr 1.02e-03 | grad 0.2414 +2026-04-10 09:11:19 - INFO - parrotllm.training - step 50600 | epoch 0 | loss 3.4797 | lr 1.02e-03 | grad 0.2123 +2026-04-10 09:11:22 - INFO - parrotllm.training - step 50610 | epoch 0 | loss 3.4011 | lr 1.02e-03 | grad 0.2186 +2026-04-10 09:11:25 - INFO - parrotllm.training - step 50620 | epoch 0 | loss 3.3613 | lr 1.02e-03 | grad 0.2270 +2026-04-10 09:11:28 - INFO - parrotllm.training - step 50630 | epoch 0 | loss 3.3551 | lr 1.02e-03 | grad 0.2235 +2026-04-10 09:11:31 - INFO - parrotllm.training - step 50640 | epoch 0 | loss 3.4066 | lr 1.02e-03 | grad 0.2221 +2026-04-10 09:11:34 - INFO - parrotllm.training - step 50650 | epoch 0 | loss 3.3552 | lr 1.02e-03 | grad 0.2306 +2026-04-10 09:11:37 - INFO - parrotllm.training - step 50660 | epoch 0 | loss 3.4416 | lr 1.02e-03 | grad 0.2235 +2026-04-10 09:11:40 - INFO - parrotllm.training - step 50670 | epoch 0 | loss 3.4110 | lr 1.02e-03 | grad 0.2336 +2026-04-10 09:11:43 - INFO - parrotllm.training - step 50680 | epoch 0 | loss 3.4474 | lr 1.02e-03 | grad 0.2259 +2026-04-10 09:11:47 - INFO - parrotllm.training - step 50690 | epoch 0 | loss 3.4464 | lr 1.02e-03 | grad 0.2185 +2026-04-10 09:11:50 - INFO - parrotllm.training - step 50700 | epoch 0 | loss 3.3822 | lr 1.02e-03 | grad 0.2841 +2026-04-10 09:11:53 - INFO - parrotllm.training - step 50710 | epoch 0 | loss 3.5150 | lr 1.02e-03 | grad 0.2596 +2026-04-10 09:11:56 - INFO - parrotllm.training - step 50720 | epoch 0 | loss 3.4278 | lr 1.02e-03 | grad 0.2487 +2026-04-10 09:11:59 - INFO - parrotllm.training - step 50730 | epoch 0 | loss 3.3555 | lr 1.02e-03 | grad 0.2222 +2026-04-10 09:12:02 - INFO - parrotllm.training - step 50740 | epoch 0 | loss 3.3803 | lr 1.02e-03 | grad 0.2245 +2026-04-10 09:12:05 - INFO - parrotllm.training - step 50750 | epoch 0 | loss 3.4717 | lr 1.02e-03 | grad 0.2220 +2026-04-10 09:12:08 - INFO - parrotllm.training - step 50760 | epoch 0 | loss 3.4431 | lr 1.02e-03 | grad 0.2501 +2026-04-10 09:12:11 - INFO - parrotllm.training - step 50770 | epoch 0 | loss 3.3637 | lr 1.02e-03 | grad 0.2341 +2026-04-10 09:12:14 - INFO - parrotllm.training - step 50780 | epoch 0 | loss 3.4146 | lr 1.02e-03 | grad 0.2286 +2026-04-10 09:12:17 - INFO - parrotllm.training - step 50790 | epoch 0 | loss 3.3858 | lr 1.02e-03 | grad 0.2395 +2026-04-10 09:12:20 - INFO - parrotllm.training - step 50800 | epoch 0 | loss 3.5064 | lr 1.02e-03 | grad 0.2169 +2026-04-10 09:12:23 - INFO - parrotllm.training - step 50810 | epoch 0 | loss 3.4472 | lr 1.02e-03 | grad 0.2540 +2026-04-10 09:12:26 - INFO - parrotllm.training - step 50820 | epoch 0 | loss 3.4044 | lr 1.02e-03 | grad 0.2235 +2026-04-10 09:12:29 - INFO - parrotllm.training - step 50830 | epoch 0 | loss 3.4409 | lr 1.02e-03 | grad 0.2746 +2026-04-10 09:12:32 - INFO - parrotllm.training - step 50840 | epoch 0 | loss 3.3123 | lr 1.01e-03 | grad 0.2436 +2026-04-10 09:12:35 - INFO - parrotllm.training - step 50850 | epoch 0 | loss 3.3983 | lr 1.01e-03 | grad 0.2334 +2026-04-10 09:12:38 - INFO - parrotllm.training - step 50860 | epoch 0 | loss 3.4920 | lr 1.01e-03 | grad 0.2253 +2026-04-10 09:12:41 - INFO - parrotllm.training - step 50870 | epoch 0 | loss 3.5016 | lr 1.01e-03 | grad 0.2250 +2026-04-10 09:12:44 - INFO - parrotllm.training - step 50880 | epoch 0 | loss 3.4054 | lr 1.01e-03 | grad 0.2398 +2026-04-10 09:12:47 - INFO - parrotllm.training - step 50890 | epoch 0 | loss 3.4786 | lr 1.01e-03 | grad 0.2341 +2026-04-10 09:12:51 - INFO - parrotllm.training - step 50900 | epoch 0 | loss 3.3759 | lr 1.01e-03 | grad 0.2683 +2026-04-10 09:12:54 - INFO - parrotllm.training - step 50910 | epoch 0 | loss 3.5431 | lr 1.01e-03 | grad 0.2581 +2026-04-10 09:12:57 - INFO - parrotllm.training - step 50920 | epoch 0 | loss 3.3956 | lr 1.01e-03 | grad 0.2403 +2026-04-10 09:13:00 - INFO - parrotllm.training - step 50930 | epoch 0 | loss 3.5024 | lr 1.01e-03 | grad 0.2331 +2026-04-10 09:13:03 - INFO - parrotllm.training - step 50940 | epoch 0 | loss 3.4464 | lr 1.01e-03 | grad 0.2216 +2026-04-10 09:13:06 - INFO - parrotllm.training - step 50950 | epoch 0 | loss 3.3416 | lr 1.01e-03 | grad 0.2427 +2026-04-10 09:13:09 - INFO - parrotllm.training - step 50960 | epoch 0 | loss 3.4410 | lr 1.01e-03 | grad 0.2237 +2026-04-10 09:13:12 - INFO - parrotllm.training - step 50970 | epoch 0 | loss 3.3171 | lr 1.01e-03 | grad 0.2731 +2026-04-10 09:13:15 - INFO - parrotllm.training - step 50980 | epoch 0 | loss 3.4837 | lr 1.01e-03 | grad 0.2095 +2026-04-10 09:13:18 - INFO - parrotllm.training - step 50990 | epoch 0 | loss 3.2937 | lr 1.01e-03 | grad 0.2419 +2026-04-10 09:13:21 - INFO - parrotllm.training - step 51000 | epoch 0 | loss 3.4150 | lr 1.01e-03 | grad 0.2267 +2026-04-10 09:13:21 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:13:21 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:13:24 - INFO - parrotllm.training - Train: loss=3.4150, ppl=30.42 +2026-04-10 09:13:24 - INFO - parrotllm.training - Val: loss=3.3599, ppl=28.79 +2026-04-10 09:13:25 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3599_epoch_0000_step_0051000.pt +2026-04-10 09:13:26 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3585, min_delta=0.001000). +2026-04-10 09:13:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:13:30 - INFO - parrotllm.training - step 51010 | epoch 0 | loss 3.3643 | lr 1.01e-03 | grad 0.2119 +2026-04-10 09:13:33 - INFO - parrotllm.training - step 51020 | epoch 0 | loss 3.4473 | lr 1.01e-03 | grad 0.2245 +2026-04-10 09:13:36 - INFO - parrotllm.training - step 51030 | epoch 0 | loss 3.4480 | lr 1.01e-03 | grad 0.2404 +2026-04-10 09:13:39 - INFO - parrotllm.training - step 51040 | epoch 0 | loss 3.4595 | lr 1.01e-03 | grad 0.2883 +2026-04-10 09:13:42 - INFO - parrotllm.training - step 51050 | epoch 0 | loss 3.3980 | lr 1.01e-03 | grad 0.2351 +2026-04-10 09:13:45 - INFO - parrotllm.training - step 51060 | epoch 0 | loss 3.3606 | lr 1.01e-03 | grad 0.2389 +2026-04-10 09:13:48 - INFO - parrotllm.training - step 51070 | epoch 0 | loss 3.4844 | lr 1.01e-03 | grad 0.2740 +2026-04-10 09:13:51 - INFO - parrotllm.training - step 51080 | epoch 0 | loss 3.3725 | lr 1.01e-03 | grad 0.2398 +2026-04-10 09:13:54 - INFO - parrotllm.training - step 51090 | epoch 0 | loss 3.4514 | lr 1.01e-03 | grad 0.2472 +2026-04-10 09:13:57 - INFO - parrotllm.training - step 51100 | epoch 0 | loss 3.3276 | lr 1.01e-03 | grad 0.2399 +2026-04-10 09:14:00 - INFO - parrotllm.training - step 51110 | epoch 0 | loss 3.4289 | lr 1.01e-03 | grad 0.2408 +2026-04-10 09:14:03 - INFO - parrotllm.training - step 51120 | epoch 0 | loss 3.5206 | lr 1.01e-03 | grad 0.2625 +2026-04-10 09:14:06 - INFO - parrotllm.training - step 51130 | epoch 0 | loss 3.4268 | lr 1.01e-03 | grad 0.2463 +2026-04-10 09:14:09 - INFO - parrotllm.training - step 51140 | epoch 0 | loss 3.4574 | lr 1.01e-03 | grad 0.2158 +2026-04-10 09:14:12 - INFO - parrotllm.training - step 51150 | epoch 0 | loss 3.5318 | lr 1.01e-03 | grad 0.2179 +2026-04-10 09:14:15 - INFO - parrotllm.training - step 51160 | epoch 0 | loss 3.5176 | lr 1.01e-03 | grad 0.2449 +2026-04-10 09:14:18 - INFO - parrotllm.training - step 51170 | epoch 0 | loss 3.4638 | lr 1.01e-03 | grad 0.2624 +2026-04-10 09:14:21 - INFO - parrotllm.training - step 51180 | epoch 0 | loss 3.4764 | lr 1.01e-03 | grad 0.2598 +2026-04-10 09:14:24 - INFO - parrotllm.training - step 51190 | epoch 0 | loss 3.4202 | lr 1.01e-03 | grad 0.2258 +2026-04-10 09:14:27 - INFO - parrotllm.training - step 51200 | epoch 0 | loss 3.4084 | lr 1.01e-03 | grad 0.2251 +2026-04-10 09:14:30 - INFO - parrotllm.training - step 51210 | epoch 0 | loss 3.3967 | lr 1.01e-03 | grad 0.2427 +2026-04-10 09:14:34 - INFO - parrotllm.training - step 51220 | epoch 0 | loss 3.4552 | lr 1.01e-03 | grad 0.2177 +2026-04-10 09:14:37 - INFO - parrotllm.training - step 51230 | epoch 0 | loss 3.4295 | lr 1.01e-03 | grad 0.2226 +2026-04-10 09:14:40 - INFO - parrotllm.training - step 51240 | epoch 0 | loss 3.4975 | lr 1.01e-03 | grad 0.2179 +2026-04-10 09:14:43 - INFO - parrotllm.training - step 51250 | epoch 0 | loss 3.3613 | lr 1.01e-03 | grad 0.2649 +2026-04-10 09:14:46 - INFO - parrotllm.training - step 51260 | epoch 0 | loss 3.4399 | lr 1.01e-03 | grad 0.2278 +2026-04-10 09:14:49 - INFO - parrotllm.training - step 51270 | epoch 0 | loss 3.3119 | lr 1.01e-03 | grad 0.2229 +2026-04-10 09:14:52 - INFO - parrotllm.training - step 51280 | epoch 0 | loss 3.4623 | lr 1.01e-03 | grad 0.2057 +2026-04-10 09:14:55 - INFO - parrotllm.training - step 51290 | epoch 0 | loss 3.4749 | lr 1.01e-03 | grad 0.2281 +2026-04-10 09:14:58 - INFO - parrotllm.training - step 51300 | epoch 0 | loss 3.4074 | lr 1.01e-03 | grad 0.2234 +2026-04-10 09:15:01 - INFO - parrotllm.training - step 51310 | epoch 0 | loss 3.3641 | lr 1.01e-03 | grad 0.2057 +2026-04-10 09:15:04 - INFO - parrotllm.training - step 51320 | epoch 0 | loss 3.3072 | lr 1.01e-03 | grad 0.2579 +2026-04-10 09:15:07 - INFO - parrotllm.training - step 51330 | epoch 0 | loss 3.4048 | lr 1.01e-03 | grad 0.2379 +2026-04-10 09:15:10 - INFO - parrotllm.training - step 51340 | epoch 0 | loss 3.4190 | lr 1.01e-03 | grad 0.2336 +2026-04-10 09:15:13 - INFO - parrotllm.training - step 51350 | epoch 0 | loss 3.3277 | lr 1.01e-03 | grad 0.2474 +2026-04-10 09:15:16 - INFO - parrotllm.training - step 51360 | epoch 0 | loss 3.3907 | lr 1.01e-03 | grad 0.2310 +2026-04-10 09:15:19 - INFO - parrotllm.training - step 51370 | epoch 0 | loss 3.5116 | lr 1.01e-03 | grad 0.2362 +2026-04-10 09:15:22 - INFO - parrotllm.training - step 51380 | epoch 0 | loss 3.3491 | lr 1.01e-03 | grad 0.2410 +2026-04-10 09:15:25 - INFO - parrotllm.training - step 51390 | epoch 0 | loss 3.3407 | lr 1.01e-03 | grad 0.2495 +2026-04-10 09:15:28 - INFO - parrotllm.training - step 51400 | epoch 0 | loss 3.3932 | lr 1.01e-03 | grad 0.2641 +2026-04-10 09:15:31 - INFO - parrotllm.training - step 51410 | epoch 0 | loss 3.3869 | lr 1.01e-03 | grad 0.2479 +2026-04-10 09:15:35 - INFO - parrotllm.training - step 51420 | epoch 0 | loss 3.4810 | lr 1.01e-03 | grad 0.2203 +2026-04-10 09:15:38 - INFO - parrotllm.training - step 51430 | epoch 0 | loss 3.4075 | lr 1.00e-03 | grad 0.2520 +2026-04-10 09:15:41 - INFO - parrotllm.training - step 51440 | epoch 0 | loss 3.5463 | lr 1.00e-03 | grad 0.2418 +2026-04-10 09:15:44 - INFO - parrotllm.training - step 51450 | epoch 0 | loss 3.3992 | lr 1.00e-03 | grad 0.2266 +2026-04-10 09:15:47 - INFO - parrotllm.training - step 51460 | epoch 0 | loss 3.4319 | lr 1.00e-03 | grad 0.2548 +2026-04-10 09:15:50 - INFO - parrotllm.training - step 51470 | epoch 0 | loss 3.4647 | lr 1.00e-03 | grad 0.2248 +2026-04-10 09:15:53 - INFO - parrotllm.training - step 51480 | epoch 0 | loss 3.3554 | lr 1.00e-03 | grad 0.2202 +2026-04-10 09:15:56 - INFO - parrotllm.training - step 51490 | epoch 0 | loss 3.3798 | lr 1.00e-03 | grad 0.2239 +2026-04-10 09:15:59 - INFO - parrotllm.training - step 51500 | epoch 0 | loss 3.4343 | lr 1.00e-03 | grad 0.2186 +2026-04-10 09:15:59 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:15:59 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:16:02 - INFO - parrotllm.training - Train: loss=3.4343, ppl=31.01 +2026-04-10 09:16:02 - INFO - parrotllm.training - Val: loss=3.3575, ppl=28.72 +2026-04-10 09:16:02 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 09:16:03 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3575_epoch_0000_step_0051500.pt +2026-04-10 09:16:05 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:16:08 - INFO - parrotllm.training - step 51510 | epoch 0 | loss 3.3800 | lr 1.00e-03 | grad 0.2194 +2026-04-10 09:16:11 - INFO - parrotllm.training - step 51520 | epoch 0 | loss 3.4944 | lr 1.00e-03 | grad 0.2488 +2026-04-10 09:16:14 - INFO - parrotllm.training - step 51530 | epoch 0 | loss 3.3721 | lr 1.00e-03 | grad 0.2356 +2026-04-10 09:16:17 - INFO - parrotllm.training - step 51540 | epoch 0 | loss 3.3746 | lr 1.00e-03 | grad 0.2305 +2026-04-10 09:16:20 - INFO - parrotllm.training - step 51550 | epoch 0 | loss 3.3809 | lr 1.00e-03 | grad 0.2457 +2026-04-10 09:16:23 - INFO - parrotllm.training - step 51560 | epoch 0 | loss 3.4438 | lr 1.00e-03 | grad 0.2403 +2026-04-10 09:16:26 - INFO - parrotllm.training - step 51570 | epoch 0 | loss 3.4113 | lr 1.00e-03 | grad 0.2294 +2026-04-10 09:16:29 - INFO - parrotllm.training - step 51580 | epoch 0 | loss 3.3822 | lr 1.00e-03 | grad 0.2250 +2026-04-10 09:16:32 - INFO - parrotllm.training - step 51590 | epoch 0 | loss 3.4374 | lr 1.00e-03 | grad 0.2309 +2026-04-10 09:16:34 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 51616/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 09:16:36 - INFO - parrotllm.training - step 51600 | epoch 0 | loss 3.3672 | lr 1.00e-03 | grad 0.2175 +2026-04-10 09:16:39 - INFO - parrotllm.training - step 51610 | epoch 0 | loss 3.4719 | lr 1.00e-03 | grad 0.2340 +2026-04-10 09:16:42 - INFO - parrotllm.training - step 51620 | epoch 0 | loss 3.3586 | lr 1.00e-03 | grad 0.2278 +2026-04-10 09:16:45 - INFO - parrotllm.training - step 51630 | epoch 0 | loss 3.3614 | lr 1.00e-03 | grad 0.2355 +2026-04-10 09:16:48 - INFO - parrotllm.training - step 51640 | epoch 0 | loss 3.4433 | lr 1.00e-03 | grad 0.2392 +2026-04-10 09:16:51 - INFO - parrotllm.training - step 51650 | epoch 0 | loss 3.3143 | lr 1.00e-03 | grad 0.2341 +2026-04-10 09:16:54 - INFO - parrotllm.training - step 51660 | epoch 0 | loss 3.4047 | lr 1.00e-03 | grad 0.2329 +2026-04-10 09:16:57 - INFO - parrotllm.training - step 51670 | epoch 0 | loss 3.3703 | lr 1.00e-03 | grad 0.2382 +2026-04-10 09:17:00 - INFO - parrotllm.training - step 51680 | epoch 0 | loss 3.3446 | lr 1.00e-03 | grad 0.2192 +2026-04-10 09:17:03 - INFO - parrotllm.training - step 51690 | epoch 0 | loss 3.4594 | lr 1.00e-03 | grad 0.2177 +2026-04-10 09:17:06 - INFO - parrotllm.training - step 51700 | epoch 0 | loss 3.4082 | lr 1.00e-03 | grad 0.2444 +2026-04-10 09:17:09 - INFO - parrotllm.training - step 51710 | epoch 0 | loss 3.3329 | lr 1.00e-03 | grad 0.2377 +2026-04-10 09:17:12 - INFO - parrotllm.training - step 51720 | epoch 0 | loss 3.4513 | lr 1.00e-03 | grad 0.2607 +2026-04-10 09:17:15 - INFO - parrotllm.training - step 51730 | epoch 0 | loss 3.3986 | lr 1.00e-03 | grad 0.2180 +2026-04-10 09:17:18 - INFO - parrotllm.training - step 51740 | epoch 0 | loss 3.4648 | lr 1.00e-03 | grad 0.2210 +2026-04-10 09:17:21 - INFO - parrotllm.training - step 51750 | epoch 0 | loss 3.3381 | lr 9.99e-04 | grad 0.2447 +2026-04-10 09:17:25 - INFO - parrotllm.training - step 51760 | epoch 0 | loss 3.2471 | lr 9.99e-04 | grad 0.2421 +2026-04-10 09:17:28 - INFO - parrotllm.training - step 51770 | epoch 0 | loss 3.4234 | lr 9.99e-04 | grad 0.2258 +2026-04-10 09:17:31 - INFO - parrotllm.training - step 51780 | epoch 0 | loss 3.4419 | lr 9.99e-04 | grad 0.2347 +2026-04-10 09:17:34 - INFO - parrotllm.training - step 51790 | epoch 0 | loss 3.4403 | lr 9.99e-04 | grad 0.2322 +2026-04-10 09:17:37 - INFO - parrotllm.training - step 51800 | epoch 0 | loss 3.3769 | lr 9.99e-04 | grad 0.2189 +2026-04-10 09:17:40 - INFO - parrotllm.training - step 51810 | epoch 0 | loss 3.4299 | lr 9.98e-04 | grad 0.2425 +2026-04-10 09:17:43 - INFO - parrotllm.training - step 51820 | epoch 0 | loss 3.3816 | lr 9.98e-04 | grad 0.2210 +2026-04-10 09:17:46 - INFO - parrotllm.training - step 51830 | epoch 0 | loss 3.3701 | lr 9.98e-04 | grad 0.2224 +2026-04-10 09:17:49 - INFO - parrotllm.training - step 51840 | epoch 0 | loss 3.3366 | lr 9.98e-04 | grad 0.2307 +2026-04-10 09:17:52 - INFO - parrotllm.training - step 51850 | epoch 0 | loss 3.3132 | lr 9.98e-04 | grad 0.2248 +2026-04-10 09:17:55 - INFO - parrotllm.training - step 51860 | epoch 0 | loss 3.3940 | lr 9.98e-04 | grad 0.2209 +2026-04-10 09:17:58 - INFO - parrotllm.training - step 51870 | epoch 0 | loss 3.4424 | lr 9.97e-04 | grad 0.2355 +2026-04-10 09:18:01 - INFO - parrotllm.training - step 51880 | epoch 0 | loss 3.3511 | lr 9.97e-04 | grad 0.2340 +2026-04-10 09:18:04 - INFO - parrotllm.training - step 51890 | epoch 0 | loss 3.3107 | lr 9.97e-04 | grad 0.2257 +2026-04-10 09:18:07 - INFO - parrotllm.training - step 51900 | epoch 0 | loss 3.3920 | lr 9.97e-04 | grad 0.2360 +2026-04-10 09:18:10 - INFO - parrotllm.training - step 51910 | epoch 0 | loss 3.4080 | lr 9.97e-04 | grad 0.2485 +2026-04-10 09:18:13 - INFO - parrotllm.training - step 51920 | epoch 0 | loss 3.3057 | lr 9.97e-04 | grad 0.2235 +2026-04-10 09:18:17 - INFO - parrotllm.training - step 51930 | epoch 0 | loss 3.4490 | lr 9.96e-04 | grad 0.2488 +2026-04-10 09:18:20 - INFO - parrotllm.training - step 51940 | epoch 0 | loss 3.4277 | lr 9.96e-04 | grad 0.2422 +2026-04-10 09:18:23 - INFO - parrotllm.training - step 51950 | epoch 0 | loss 3.4670 | lr 9.96e-04 | grad 0.2224 +2026-04-10 09:18:26 - INFO - parrotllm.training - step 51960 | epoch 0 | loss 3.4198 | lr 9.96e-04 | grad 0.2647 +2026-04-10 09:18:29 - INFO - parrotllm.training - step 51970 | epoch 0 | loss 3.4274 | lr 9.96e-04 | grad 0.2759 +2026-04-10 09:18:32 - INFO - parrotllm.training - step 51980 | epoch 0 | loss 3.5226 | lr 9.96e-04 | grad 0.2258 +2026-04-10 09:18:35 - INFO - parrotllm.training - step 51990 | epoch 0 | loss 3.4457 | lr 9.95e-04 | grad 0.2305 +2026-04-10 09:18:38 - INFO - parrotllm.training - step 52000 | epoch 0 | loss 3.3203 | lr 9.95e-04 | grad 0.2472 +2026-04-10 09:18:38 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:18:38 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:18:41 - INFO - parrotllm.training - Train: loss=3.3203, ppl=27.67 +2026-04-10 09:18:41 - INFO - parrotllm.training - Val: loss=3.3578, ppl=28.73 +2026-04-10 09:18:42 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3578_epoch_0000_step_0052000.pt +2026-04-10 09:18:44 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3575, min_delta=0.001000). +2026-04-10 09:18:44 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:18:47 - INFO - parrotllm.training - step 52010 | epoch 0 | loss 3.4155 | lr 9.95e-04 | grad 0.2153 +2026-04-10 09:18:50 - INFO - parrotllm.training - step 52020 | epoch 0 | loss 3.3557 | lr 9.95e-04 | grad 0.2194 +2026-04-10 09:18:53 - INFO - parrotllm.training - step 52030 | epoch 0 | loss 3.3863 | lr 9.95e-04 | grad 0.2605 +2026-04-10 09:18:56 - INFO - parrotllm.training - step 52040 | epoch 0 | loss 3.3431 | lr 9.95e-04 | grad 0.2507 +2026-04-10 09:18:59 - INFO - parrotllm.training - step 52050 | epoch 0 | loss 3.5011 | lr 9.94e-04 | grad 0.2359 +2026-04-10 09:19:02 - INFO - parrotllm.training - step 52060 | epoch 0 | loss 3.3655 | lr 9.94e-04 | grad 0.2133 +2026-04-10 09:19:05 - INFO - parrotllm.training - step 52070 | epoch 0 | loss 3.3955 | lr 9.94e-04 | grad 0.2317 +2026-04-10 09:19:08 - INFO - parrotllm.training - step 52080 | epoch 0 | loss 3.4849 | lr 9.94e-04 | grad 0.2262 +2026-04-10 09:19:11 - INFO - parrotllm.training - step 52090 | epoch 0 | loss 3.4077 | lr 9.94e-04 | grad 0.2519 +2026-04-10 09:19:14 - INFO - parrotllm.training - step 52100 | epoch 0 | loss 3.3906 | lr 9.94e-04 | grad 0.2408 +2026-04-10 09:19:17 - INFO - parrotllm.training - step 52110 | epoch 0 | loss 3.3456 | lr 9.93e-04 | grad 0.2160 +2026-04-10 09:19:20 - INFO - parrotllm.training - step 52120 | epoch 0 | loss 3.3787 | lr 9.93e-04 | grad 0.2282 +2026-04-10 09:19:23 - INFO - parrotllm.training - step 52130 | epoch 0 | loss 3.3984 | lr 9.93e-04 | grad 0.2328 +2026-04-10 09:19:26 - INFO - parrotllm.training - step 52140 | epoch 0 | loss 3.3898 | lr 9.93e-04 | grad 0.2688 +2026-04-10 09:19:29 - INFO - parrotllm.training - step 52150 | epoch 0 | loss 3.5104 | lr 9.93e-04 | grad 0.2170 +2026-04-10 09:19:32 - INFO - parrotllm.training - step 52160 | epoch 0 | loss 3.3473 | lr 9.93e-04 | grad 0.2406 +2026-04-10 09:19:35 - INFO - parrotllm.training - step 52170 | epoch 0 | loss 3.3285 | lr 9.92e-04 | grad 0.2398 +2026-04-10 09:19:39 - INFO - parrotllm.training - step 52180 | epoch 0 | loss 3.3789 | lr 9.92e-04 | grad 0.2281 +2026-04-10 09:19:42 - INFO - parrotllm.training - step 52190 | epoch 0 | loss 3.4580 | lr 9.92e-04 | grad 0.2486 +2026-04-10 09:19:45 - INFO - parrotllm.training - step 52200 | epoch 0 | loss 3.3573 | lr 9.92e-04 | grad 0.2296 +2026-04-10 09:19:48 - INFO - parrotllm.training - step 52210 | epoch 0 | loss 3.3885 | lr 9.92e-04 | grad 0.2215 +2026-04-10 09:19:51 - INFO - parrotllm.training - step 52220 | epoch 0 | loss 3.3878 | lr 9.91e-04 | grad 0.2808 +2026-04-10 09:19:54 - INFO - parrotllm.training - step 52230 | epoch 0 | loss 3.4257 | lr 9.91e-04 | grad 0.2193 +2026-04-10 09:19:57 - INFO - parrotllm.training - step 52240 | epoch 0 | loss 3.3972 | lr 9.91e-04 | grad 0.2240 +2026-04-10 09:20:00 - INFO - parrotllm.training - step 52250 | epoch 0 | loss 3.4323 | lr 9.91e-04 | grad 0.2150 +2026-04-10 09:20:03 - INFO - parrotllm.training - step 52260 | epoch 0 | loss 3.4375 | lr 9.91e-04 | grad 0.2247 +2026-04-10 09:20:06 - INFO - parrotllm.training - step 52270 | epoch 0 | loss 3.2697 | lr 9.91e-04 | grad 0.2308 +2026-04-10 09:20:09 - INFO - parrotllm.training - step 52280 | epoch 0 | loss 3.4748 | lr 9.90e-04 | grad 0.2345 +2026-04-10 09:20:12 - INFO - parrotllm.training - step 52290 | epoch 0 | loss 3.3352 | lr 9.90e-04 | grad 0.2410 +2026-04-10 09:20:15 - INFO - parrotllm.training - step 52300 | epoch 0 | loss 3.4209 | lr 9.90e-04 | grad 0.2357 +2026-04-10 09:20:18 - INFO - parrotllm.training - step 52310 | epoch 0 | loss 3.5586 | lr 9.90e-04 | grad 0.2269 +2026-04-10 09:20:21 - INFO - parrotllm.training - step 52320 | epoch 0 | loss 3.5185 | lr 9.90e-04 | grad 0.2437 +2026-04-10 09:20:24 - INFO - parrotllm.training - step 52330 | epoch 0 | loss 3.3707 | lr 9.90e-04 | grad 0.2411 +2026-04-10 09:20:27 - INFO - parrotllm.training - step 52340 | epoch 0 | loss 3.4275 | lr 9.89e-04 | grad 0.2262 +2026-04-10 09:20:30 - INFO - parrotllm.training - step 52350 | epoch 0 | loss 3.4029 | lr 9.89e-04 | grad 0.2298 +2026-04-10 09:20:33 - INFO - parrotllm.training - step 52360 | epoch 0 | loss 3.2946 | lr 9.89e-04 | grad 0.2186 +2026-04-10 09:20:36 - INFO - parrotllm.training - step 52370 | epoch 0 | loss 3.3301 | lr 9.89e-04 | grad 0.2446 +2026-04-10 09:20:40 - INFO - parrotllm.training - step 52380 | epoch 0 | loss 3.4582 | lr 9.89e-04 | grad 0.2224 +2026-04-10 09:20:43 - INFO - parrotllm.training - step 52390 | epoch 0 | loss 3.4553 | lr 9.89e-04 | grad 0.2862 +2026-04-10 09:20:46 - INFO - parrotllm.training - step 52400 | epoch 0 | loss 3.4235 | lr 9.88e-04 | grad 0.2596 +2026-04-10 09:20:49 - INFO - parrotllm.training - step 52410 | epoch 0 | loss 3.3724 | lr 9.88e-04 | grad 0.2504 +2026-04-10 09:20:52 - INFO - parrotllm.training - step 52420 | epoch 0 | loss 3.5896 | lr 9.88e-04 | grad 0.2201 +2026-04-10 09:20:55 - INFO - parrotllm.training - step 52430 | epoch 0 | loss 3.5191 | lr 9.88e-04 | grad 0.2282 +2026-04-10 09:20:58 - INFO - parrotllm.training - step 52440 | epoch 0 | loss 3.4482 | lr 9.88e-04 | grad 0.2433 +2026-04-10 09:21:01 - INFO - parrotllm.training - step 52450 | epoch 0 | loss 3.4327 | lr 9.88e-04 | grad 0.2471 +2026-04-10 09:21:04 - INFO - parrotllm.training - step 52460 | epoch 0 | loss 3.4705 | lr 9.87e-04 | grad 0.2204 +2026-04-10 09:21:07 - INFO - parrotllm.training - step 52470 | epoch 0 | loss 3.4340 | lr 9.87e-04 | grad 0.2151 +2026-04-10 09:21:10 - INFO - parrotllm.training - step 52480 | epoch 0 | loss 3.4491 | lr 9.87e-04 | grad 0.2030 +2026-04-10 09:21:13 - INFO - parrotllm.training - step 52490 | epoch 0 | loss 3.4281 | lr 9.87e-04 | grad 0.2350 +2026-04-10 09:21:16 - INFO - parrotllm.training - step 52500 | epoch 0 | loss 3.3880 | lr 9.87e-04 | grad 0.2358 +2026-04-10 09:21:16 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:21:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:21:19 - INFO - parrotllm.training - Train: loss=3.3880, ppl=29.61 +2026-04-10 09:21:19 - INFO - parrotllm.training - Val: loss=3.3558, ppl=28.67 +2026-04-10 09:21:19 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 09:21:20 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3558_epoch_0000_step_0052500.pt +2026-04-10 09:21:22 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:21:23 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0052500.pt +2026-04-10 09:21:27 - INFO - parrotllm.training - step 52510 | epoch 0 | loss 3.4107 | lr 9.87e-04 | grad 0.2493 +2026-04-10 09:21:30 - INFO - parrotllm.training - step 52520 | epoch 0 | loss 3.4251 | lr 9.86e-04 | grad 0.2902 +2026-04-10 09:21:34 - INFO - parrotllm.training - step 52530 | epoch 0 | loss 3.4273 | lr 9.86e-04 | grad 0.2523 +2026-04-10 09:21:37 - INFO - parrotllm.training - step 52540 | epoch 0 | loss 3.4317 | lr 9.86e-04 | grad 0.2483 +2026-04-10 09:21:40 - INFO - parrotllm.training - step 52550 | epoch 0 | loss 3.5193 | lr 9.86e-04 | grad 0.2349 +2026-04-10 09:21:43 - INFO - parrotllm.training - step 52560 | epoch 0 | loss 3.4545 | lr 9.86e-04 | grad 0.2254 +2026-04-10 09:21:46 - INFO - parrotllm.training - step 52570 | epoch 0 | loss 3.3219 | lr 9.86e-04 | grad 0.2285 +2026-04-10 09:21:49 - INFO - parrotllm.training - step 52580 | epoch 0 | loss 3.4889 | lr 9.85e-04 | grad 0.2071 +2026-04-10 09:21:52 - INFO - parrotllm.training - step 52590 | epoch 0 | loss 3.4571 | lr 9.85e-04 | grad 0.2391 +2026-04-10 09:21:55 - INFO - parrotllm.training - step 52600 | epoch 0 | loss 3.4029 | lr 9.85e-04 | grad 0.2473 +2026-04-10 09:21:58 - INFO - parrotllm.training - step 52610 | epoch 0 | loss 3.3156 | lr 9.85e-04 | grad 0.2445 +2026-04-10 09:22:01 - INFO - parrotllm.training - step 52620 | epoch 0 | loss 3.4817 | lr 9.85e-04 | grad 0.2357 +2026-04-10 09:22:04 - INFO - parrotllm.training - step 52630 | epoch 0 | loss 3.3866 | lr 9.84e-04 | grad 0.2656 +2026-04-10 09:22:07 - INFO - parrotllm.training - step 52640 | epoch 0 | loss 3.3847 | lr 9.84e-04 | grad 0.2488 +2026-04-10 09:22:10 - INFO - parrotllm.training - step 52650 | epoch 0 | loss 3.3266 | lr 9.84e-04 | grad 0.2162 +2026-04-10 09:22:13 - INFO - parrotllm.training - step 52660 | epoch 0 | loss 3.3948 | lr 9.84e-04 | grad 0.2489 +2026-04-10 09:22:16 - INFO - parrotllm.training - step 52670 | epoch 0 | loss 3.4545 | lr 9.84e-04 | grad 0.2261 +2026-04-10 09:22:19 - INFO - parrotllm.training - step 52680 | epoch 0 | loss 3.3964 | lr 9.84e-04 | grad 0.2254 +2026-04-10 09:22:22 - INFO - parrotllm.training - step 52690 | epoch 0 | loss 3.5240 | lr 9.83e-04 | grad 0.2259 +2026-04-10 09:22:25 - INFO - parrotllm.training - step 52700 | epoch 0 | loss 3.4577 | lr 9.83e-04 | grad 0.2574 +2026-04-10 09:22:28 - INFO - parrotllm.training - step 52710 | epoch 0 | loss 3.4343 | lr 9.83e-04 | grad 0.2505 +2026-04-10 09:22:31 - INFO - parrotllm.training - step 52720 | epoch 0 | loss 3.4181 | lr 9.83e-04 | grad 0.2576 +2026-04-10 09:22:35 - INFO - parrotllm.training - step 52730 | epoch 0 | loss 3.4186 | lr 9.83e-04 | grad 0.2167 +2026-04-10 09:22:38 - INFO - parrotllm.training - step 52740 | epoch 0 | loss 3.4481 | lr 9.83e-04 | grad 0.2476 +2026-04-10 09:22:41 - INFO - parrotllm.training - step 52750 | epoch 0 | loss 3.4445 | lr 9.82e-04 | grad 0.2928 +2026-04-10 09:22:44 - INFO - parrotllm.training - step 52760 | epoch 0 | loss 3.3974 | lr 9.82e-04 | grad 0.2342 +2026-04-10 09:22:47 - INFO - parrotllm.training - step 52770 | epoch 0 | loss 3.3526 | lr 9.82e-04 | grad 0.2499 +2026-04-10 09:22:50 - INFO - parrotllm.training - step 52780 | epoch 0 | loss 3.4240 | lr 9.82e-04 | grad 0.2247 +2026-04-10 09:22:53 - INFO - parrotllm.training - step 52790 | epoch 0 | loss 3.3627 | lr 9.82e-04 | grad 0.2478 +2026-04-10 09:22:56 - INFO - parrotllm.training - step 52800 | epoch 0 | loss 3.4538 | lr 9.82e-04 | grad 0.2490 +2026-04-10 09:22:59 - INFO - parrotllm.training - step 52810 | epoch 0 | loss 3.3619 | lr 9.81e-04 | grad 0.2421 +2026-04-10 09:23:02 - INFO - parrotllm.training - step 52820 | epoch 0 | loss 3.3904 | lr 9.81e-04 | grad 0.2655 +2026-04-10 09:23:05 - INFO - parrotllm.training - step 52830 | epoch 0 | loss 3.3572 | lr 9.81e-04 | grad 0.2532 +2026-04-10 09:23:08 - INFO - parrotllm.training - step 52840 | epoch 0 | loss 3.2877 | lr 9.81e-04 | grad 0.2345 +2026-04-10 09:23:11 - INFO - parrotllm.training - step 52850 | epoch 0 | loss 3.4547 | lr 9.81e-04 | grad 0.2384 +2026-04-10 09:23:14 - INFO - parrotllm.training - step 52860 | epoch 0 | loss 3.4696 | lr 9.81e-04 | grad 0.2339 +2026-04-10 09:23:17 - INFO - parrotllm.training - step 52870 | epoch 0 | loss 3.5099 | lr 9.80e-04 | grad 0.2180 +2026-04-10 09:23:20 - INFO - parrotllm.training - step 52880 | epoch 0 | loss 3.4583 | lr 9.80e-04 | grad 0.2318 +2026-04-10 09:23:23 - INFO - parrotllm.training - step 52890 | epoch 0 | loss 3.3000 | lr 9.80e-04 | grad 0.2313 +2026-04-10 09:23:26 - INFO - parrotllm.training - step 52900 | epoch 0 | loss 3.2961 | lr 9.80e-04 | grad 0.2358 +2026-04-10 09:23:29 - INFO - parrotllm.training - step 52910 | epoch 0 | loss 3.5470 | lr 9.80e-04 | grad 0.2346 +2026-04-10 09:23:33 - INFO - parrotllm.training - step 52920 | epoch 0 | loss 3.3715 | lr 9.80e-04 | grad 0.2254 +2026-04-10 09:23:36 - INFO - parrotllm.training - step 52930 | epoch 0 | loss 3.4151 | lr 9.79e-04 | grad 0.2396 +2026-04-10 09:23:39 - INFO - parrotllm.training - step 52940 | epoch 0 | loss 3.4003 | lr 9.79e-04 | grad 0.2292 +2026-04-10 09:23:42 - INFO - parrotllm.training - step 52950 | epoch 0 | loss 3.4399 | lr 9.79e-04 | grad 0.2681 +2026-04-10 09:23:45 - INFO - parrotllm.training - step 52960 | epoch 0 | loss 3.4038 | lr 9.79e-04 | grad 0.2359 +2026-04-10 09:23:48 - INFO - parrotllm.training - step 52970 | epoch 0 | loss 3.3431 | lr 9.79e-04 | grad 0.2380 +2026-04-10 09:23:51 - INFO - parrotllm.training - step 52980 | epoch 0 | loss 3.3354 | lr 9.79e-04 | grad 0.2421 +2026-04-10 09:23:54 - INFO - parrotllm.training - step 52990 | epoch 0 | loss 3.4563 | lr 9.78e-04 | grad 0.2425 +2026-04-10 09:23:57 - INFO - parrotllm.training - step 53000 | epoch 0 | loss 3.4124 | lr 9.78e-04 | grad 0.2482 +2026-04-10 09:23:57 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:23:57 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:24:00 - INFO - parrotllm.training - Train: loss=3.4124, ppl=30.34 +2026-04-10 09:24:00 - INFO - parrotllm.training - Val: loss=3.3560, ppl=28.67 +2026-04-10 09:24:01 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3560_epoch_0000_step_0053000.pt +2026-04-10 09:24:03 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3558, min_delta=0.001000). +2026-04-10 09:24:03 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:24:06 - INFO - parrotllm.training - step 53010 | epoch 0 | loss 3.4332 | lr 9.78e-04 | grad 0.2601 +2026-04-10 09:24:09 - INFO - parrotllm.training - step 53020 | epoch 0 | loss 3.3955 | lr 9.78e-04 | grad 0.2340 +2026-04-10 09:24:12 - INFO - parrotllm.training - step 53030 | epoch 0 | loss 3.3954 | lr 9.78e-04 | grad 0.2191 +2026-04-10 09:24:15 - INFO - parrotllm.training - step 53040 | epoch 0 | loss 3.3744 | lr 9.77e-04 | grad 0.2405 +2026-04-10 09:24:18 - INFO - parrotllm.training - step 53050 | epoch 0 | loss 3.3465 | lr 9.77e-04 | grad 0.2395 +2026-04-10 09:24:21 - INFO - parrotllm.training - step 53060 | epoch 0 | loss 3.4545 | lr 9.77e-04 | grad 0.2285 +2026-04-10 09:24:24 - INFO - parrotllm.training - step 53070 | epoch 0 | loss 3.3924 | lr 9.77e-04 | grad 0.2353 +2026-04-10 09:24:27 - INFO - parrotllm.training - step 53080 | epoch 0 | loss 3.4716 | lr 9.77e-04 | grad 0.2474 +2026-04-10 09:24:30 - INFO - parrotllm.training - step 53090 | epoch 0 | loss 3.4474 | lr 9.77e-04 | grad 0.2408 +2026-04-10 09:24:33 - INFO - parrotllm.training - step 53100 | epoch 0 | loss 3.4660 | lr 9.76e-04 | grad 0.2142 +2026-04-10 09:24:36 - INFO - parrotllm.training - step 53110 | epoch 0 | loss 3.3978 | lr 9.76e-04 | grad 0.2347 +2026-04-10 09:24:39 - INFO - parrotllm.training - step 53120 | epoch 0 | loss 3.4234 | lr 9.76e-04 | grad 0.2346 +2026-04-10 09:24:42 - INFO - parrotllm.training - step 53130 | epoch 0 | loss 3.4308 | lr 9.76e-04 | grad 0.2378 +2026-04-10 09:24:45 - INFO - parrotllm.training - step 53140 | epoch 0 | loss 3.4243 | lr 9.76e-04 | grad 0.2221 +2026-04-10 09:24:49 - INFO - parrotllm.training - step 53150 | epoch 0 | loss 3.4823 | lr 9.76e-04 | grad 0.2455 +2026-04-10 09:24:52 - INFO - parrotllm.training - step 53160 | epoch 0 | loss 3.3566 | lr 9.75e-04 | grad 0.2502 +2026-04-10 09:24:55 - INFO - parrotllm.training - step 53170 | epoch 0 | loss 3.4765 | lr 9.75e-04 | grad 0.2227 +2026-04-10 09:24:58 - INFO - parrotllm.training - step 53180 | epoch 0 | loss 3.3918 | lr 9.75e-04 | grad 0.2237 +2026-04-10 09:25:01 - INFO - parrotllm.training - step 53190 | epoch 0 | loss 3.3136 | lr 9.75e-04 | grad 0.2388 +2026-04-10 09:25:04 - INFO - parrotllm.training - step 53200 | epoch 0 | loss 3.3336 | lr 9.75e-04 | grad 0.2471 +2026-04-10 09:25:07 - INFO - parrotllm.training - step 53210 | epoch 0 | loss 3.5207 | lr 9.75e-04 | grad 0.2327 +2026-04-10 09:25:10 - INFO - parrotllm.training - step 53220 | epoch 0 | loss 3.3516 | lr 9.74e-04 | grad 0.2437 +2026-04-10 09:25:13 - INFO - parrotllm.training - step 53230 | epoch 0 | loss 3.4577 | lr 9.74e-04 | grad 0.2271 +2026-04-10 09:25:16 - INFO - parrotllm.training - step 53240 | epoch 0 | loss 3.4521 | lr 9.74e-04 | grad 0.2218 +2026-04-10 09:25:19 - INFO - parrotllm.training - step 53250 | epoch 0 | loss 3.4157 | lr 9.74e-04 | grad 0.2384 +2026-04-10 09:25:22 - INFO - parrotllm.training - step 53260 | epoch 0 | loss 3.5054 | lr 9.74e-04 | grad 0.2258 +2026-04-10 09:25:25 - INFO - parrotllm.training - step 53270 | epoch 0 | loss 3.3648 | lr 9.74e-04 | grad 0.2788 +2026-04-10 09:25:28 - INFO - parrotllm.training - step 53280 | epoch 0 | loss 3.4052 | lr 9.73e-04 | grad 0.2465 +2026-04-10 09:25:31 - INFO - parrotllm.training - step 53290 | epoch 0 | loss 3.4198 | lr 9.73e-04 | grad 0.2311 +2026-04-10 09:25:34 - INFO - parrotllm.training - step 53300 | epoch 0 | loss 3.4047 | lr 9.73e-04 | grad 0.2355 +2026-04-10 09:25:37 - INFO - parrotllm.training - step 53310 | epoch 0 | loss 3.3596 | lr 9.73e-04 | grad 0.2359 +2026-04-10 09:25:40 - INFO - parrotllm.training - step 53320 | epoch 0 | loss 3.5062 | lr 9.73e-04 | grad 0.2807 +2026-04-10 09:25:44 - INFO - parrotllm.training - step 53330 | epoch 0 | loss 3.4939 | lr 9.73e-04 | grad 0.2387 +2026-04-10 09:25:47 - INFO - parrotllm.training - step 53340 | epoch 0 | loss 3.2710 | lr 9.72e-04 | grad 0.2348 +2026-04-10 09:25:50 - INFO - parrotllm.training - step 53350 | epoch 0 | loss 3.3824 | lr 9.72e-04 | grad 0.2122 +2026-04-10 09:25:53 - INFO - parrotllm.training - step 53360 | epoch 0 | loss 3.4285 | lr 9.72e-04 | grad 0.2474 +2026-04-10 09:25:56 - INFO - parrotllm.training - step 53370 | epoch 0 | loss 3.3948 | lr 9.72e-04 | grad 0.2119 +2026-04-10 09:25:59 - INFO - parrotllm.training - step 53380 | epoch 0 | loss 3.4116 | lr 9.72e-04 | grad 0.2252 +2026-04-10 09:26:02 - INFO - parrotllm.training - step 53390 | epoch 0 | loss 3.5555 | lr 9.71e-04 | grad 0.2528 +2026-04-10 09:26:05 - INFO - parrotllm.training - step 53400 | epoch 0 | loss 3.4504 | lr 9.71e-04 | grad 0.2408 +2026-04-10 09:26:08 - INFO - parrotllm.training - step 53410 | epoch 0 | loss 3.3980 | lr 9.71e-04 | grad 0.2813 +2026-04-10 09:26:11 - INFO - parrotllm.training - step 53420 | epoch 0 | loss 3.4710 | lr 9.71e-04 | grad 0.2735 +2026-04-10 09:26:14 - INFO - parrotllm.training - step 53430 | epoch 0 | loss 3.3814 | lr 9.71e-04 | grad 0.2209 +2026-04-10 09:26:17 - INFO - parrotllm.training - step 53440 | epoch 0 | loss 3.3615 | lr 9.71e-04 | grad 0.2092 +2026-04-10 09:26:20 - INFO - parrotllm.training - step 53450 | epoch 0 | loss 3.3520 | lr 9.70e-04 | grad 0.2274 +2026-04-10 09:26:23 - INFO - parrotllm.training - step 53460 | epoch 0 | loss 3.3094 | lr 9.70e-04 | grad 0.2234 +2026-04-10 09:26:26 - INFO - parrotllm.training - step 53470 | epoch 0 | loss 3.4571 | lr 9.70e-04 | grad 0.2375 +2026-04-10 09:26:29 - INFO - parrotllm.training - step 53480 | epoch 0 | loss 3.3967 | lr 9.70e-04 | grad 0.2335 +2026-04-10 09:26:32 - INFO - parrotllm.training - step 53490 | epoch 0 | loss 3.3821 | lr 9.70e-04 | grad 0.2235 +2026-04-10 09:26:35 - INFO - parrotllm.training - step 53500 | epoch 0 | loss 3.4786 | lr 9.70e-04 | grad 0.2203 +2026-04-10 09:26:35 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:26:35 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:26:38 - INFO - parrotllm.training - Train: loss=3.4786, ppl=32.41 +2026-04-10 09:26:38 - INFO - parrotllm.training - Val: loss=3.3546, ppl=28.64 +2026-04-10 09:26:38 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 09:26:39 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3546_epoch_0000_step_0053500.pt +2026-04-10 09:26:41 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:26:44 - INFO - parrotllm.training - step 53510 | epoch 0 | loss 3.4227 | lr 9.69e-04 | grad 0.2383 +2026-04-10 09:26:47 - INFO - parrotllm.training - step 53520 | epoch 0 | loss 3.3219 | lr 9.69e-04 | grad 0.2512 +2026-04-10 09:26:50 - INFO - parrotllm.training - step 53530 | epoch 0 | loss 3.4861 | lr 9.69e-04 | grad 0.2332 +2026-04-10 09:26:53 - INFO - parrotllm.training - step 53540 | epoch 0 | loss 3.4274 | lr 9.69e-04 | grad 0.2357 +2026-04-10 09:26:56 - INFO - parrotllm.training - step 53550 | epoch 0 | loss 3.3125 | lr 9.69e-04 | grad 0.2171 +2026-04-10 09:26:59 - INFO - parrotllm.training - step 53560 | epoch 0 | loss 3.3437 | lr 9.69e-04 | grad 0.2179 +2026-04-10 09:27:02 - INFO - parrotllm.training - step 53570 | epoch 0 | loss 3.2677 | lr 9.68e-04 | grad 0.2271 +2026-04-10 09:27:05 - INFO - parrotllm.training - step 53580 | epoch 0 | loss 3.4119 | lr 9.68e-04 | grad 0.2348 +2026-04-10 09:27:08 - INFO - parrotllm.training - step 53590 | epoch 0 | loss 3.4720 | lr 9.68e-04 | grad 0.2591 +2026-04-10 09:27:11 - INFO - parrotllm.training - step 53600 | epoch 0 | loss 3.4231 | lr 9.68e-04 | grad 0.2744 +2026-04-10 09:27:15 - INFO - parrotllm.training - step 53610 | epoch 0 | loss 3.4704 | lr 9.68e-04 | grad 0.3063 +2026-04-10 09:27:18 - INFO - parrotllm.training - step 53620 | epoch 0 | loss 3.4319 | lr 9.68e-04 | grad 0.2532 +2026-04-10 09:27:21 - INFO - parrotllm.training - step 53630 | epoch 0 | loss 3.3566 | lr 9.67e-04 | grad 0.2263 +2026-04-10 09:27:24 - INFO - parrotllm.training - step 53640 | epoch 0 | loss 3.4844 | lr 9.67e-04 | grad 0.2208 +2026-04-10 09:27:27 - INFO - parrotllm.training - step 53650 | epoch 0 | loss 3.4074 | lr 9.67e-04 | grad 0.2262 +2026-04-10 09:27:30 - INFO - parrotllm.training - step 53660 | epoch 0 | loss 3.4080 | lr 9.67e-04 | grad 0.2204 +2026-04-10 09:27:33 - INFO - parrotllm.training - step 53670 | epoch 0 | loss 3.3556 | lr 9.67e-04 | grad 0.2466 +2026-04-10 09:27:36 - INFO - parrotllm.training - step 53680 | epoch 0 | loss 3.3990 | lr 9.66e-04 | grad 0.2361 +2026-04-10 09:27:39 - INFO - parrotllm.training - step 53690 | epoch 0 | loss 3.4316 | lr 9.66e-04 | grad 0.2549 +2026-04-10 09:27:42 - INFO - parrotllm.training - step 53700 | epoch 0 | loss 3.2668 | lr 9.66e-04 | grad 0.2611 +2026-04-10 09:27:45 - INFO - parrotllm.training - step 53710 | epoch 0 | loss 3.4090 | lr 9.66e-04 | grad 0.2315 +2026-04-10 09:27:47 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 53738/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 09:27:49 - INFO - parrotllm.training - step 53720 | epoch 0 | loss 3.4282 | lr 9.66e-04 | grad 0.2248 +2026-04-10 09:27:52 - INFO - parrotllm.training - step 53730 | epoch 0 | loss 3.3562 | lr 9.66e-04 | grad 0.2486 +2026-04-10 09:27:55 - INFO - parrotllm.training - step 53740 | epoch 0 | loss 3.4225 | lr 9.65e-04 | grad 0.2060 +2026-04-10 09:27:58 - INFO - parrotllm.training - step 53750 | epoch 0 | loss 3.4192 | lr 9.65e-04 | grad 0.2281 +2026-04-10 09:28:01 - INFO - parrotllm.training - step 53760 | epoch 0 | loss 3.3979 | lr 9.65e-04 | grad 0.2166 +2026-04-10 09:28:04 - INFO - parrotllm.training - step 53770 | epoch 0 | loss 3.4413 | lr 9.65e-04 | grad 0.2544 +2026-04-10 09:28:07 - INFO - parrotllm.training - step 53780 | epoch 0 | loss 3.5708 | lr 9.65e-04 | grad 0.2419 +2026-04-10 09:28:10 - INFO - parrotllm.training - step 53790 | epoch 0 | loss 3.3905 | lr 9.65e-04 | grad 0.2715 +2026-04-10 09:28:13 - INFO - parrotllm.training - step 53800 | epoch 0 | loss 3.3989 | lr 9.64e-04 | grad 0.2707 +2026-04-10 09:28:16 - INFO - parrotllm.training - step 53810 | epoch 0 | loss 3.4468 | lr 9.64e-04 | grad 0.2490 +2026-04-10 09:28:19 - INFO - parrotllm.training - step 53820 | epoch 0 | loss 3.3747 | lr 9.64e-04 | grad 0.2250 +2026-04-10 09:28:22 - INFO - parrotllm.training - step 53830 | epoch 0 | loss 3.4990 | lr 9.64e-04 | grad 0.2120 +2026-04-10 09:28:25 - INFO - parrotllm.training - step 53840 | epoch 0 | loss 3.3864 | lr 9.64e-04 | grad 0.2456 +2026-04-10 09:28:28 - INFO - parrotllm.training - step 53850 | epoch 0 | loss 3.3659 | lr 9.64e-04 | grad 0.2522 +2026-04-10 09:28:31 - INFO - parrotllm.training - step 53860 | epoch 0 | loss 3.4637 | lr 9.63e-04 | grad 0.2515 +2026-04-10 09:28:34 - INFO - parrotllm.training - step 53870 | epoch 0 | loss 3.2948 | lr 9.63e-04 | grad 0.2269 +2026-04-10 09:28:37 - INFO - parrotllm.training - step 53880 | epoch 0 | loss 3.4012 | lr 9.63e-04 | grad 0.2421 +2026-04-10 09:28:40 - INFO - parrotllm.training - step 53890 | epoch 0 | loss 3.3679 | lr 9.63e-04 | grad 0.2372 +2026-04-10 09:28:43 - INFO - parrotllm.training - step 53900 | epoch 0 | loss 3.5721 | lr 9.63e-04 | grad 0.2364 +2026-04-10 09:28:47 - INFO - parrotllm.training - step 53910 | epoch 0 | loss 3.4973 | lr 9.63e-04 | grad 0.2513 +2026-04-10 09:28:50 - INFO - parrotllm.training - step 53920 | epoch 0 | loss 3.3911 | lr 9.62e-04 | grad 0.2201 +2026-04-10 09:28:53 - INFO - parrotllm.training - step 53930 | epoch 0 | loss 3.3397 | lr 9.62e-04 | grad 0.2391 +2026-04-10 09:28:56 - INFO - parrotllm.training - step 53940 | epoch 0 | loss 3.3074 | lr 9.62e-04 | grad 0.2416 +2026-04-10 09:28:59 - INFO - parrotllm.training - step 53950 | epoch 0 | loss 3.3539 | lr 9.62e-04 | grad 0.2379 +2026-04-10 09:29:02 - INFO - parrotllm.training - step 53960 | epoch 0 | loss 3.4806 | lr 9.62e-04 | grad 0.2448 +2026-04-10 09:29:05 - INFO - parrotllm.training - step 53970 | epoch 0 | loss 3.3970 | lr 9.62e-04 | grad 0.2430 +2026-04-10 09:29:08 - INFO - parrotllm.training - step 53980 | epoch 0 | loss 3.4306 | lr 9.61e-04 | grad 0.2501 +2026-04-10 09:29:11 - INFO - parrotllm.training - step 53990 | epoch 0 | loss 3.4934 | lr 9.61e-04 | grad 0.2261 +2026-04-10 09:29:14 - INFO - parrotllm.training - step 54000 | epoch 0 | loss 3.3631 | lr 9.61e-04 | grad 0.2162 +2026-04-10 09:29:14 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:29:14 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:29:17 - INFO - parrotllm.training - Train: loss=3.3631, ppl=28.88 +2026-04-10 09:29:17 - INFO - parrotllm.training - Val: loss=3.3536, ppl=28.61 +2026-04-10 09:29:18 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3536_epoch_0000_step_0054000.pt +2026-04-10 09:29:20 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3546, min_delta=0.001000). +2026-04-10 09:29:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:29:23 - INFO - parrotllm.training - step 54010 | epoch 0 | loss 3.4223 | lr 9.61e-04 | grad 0.2547 +2026-04-10 09:29:26 - INFO - parrotllm.training - step 54020 | epoch 0 | loss 3.2482 | lr 9.61e-04 | grad 0.2336 +2026-04-10 09:29:29 - INFO - parrotllm.training - step 54030 | epoch 0 | loss 3.2405 | lr 9.60e-04 | grad 0.2238 +2026-04-10 09:29:32 - INFO - parrotllm.training - step 54040 | epoch 0 | loss 3.4262 | lr 9.60e-04 | grad 0.2550 +2026-04-10 09:29:35 - INFO - parrotllm.training - step 54050 | epoch 0 | loss 3.3503 | lr 9.60e-04 | grad 0.2247 +2026-04-10 09:29:38 - INFO - parrotllm.training - step 54060 | epoch 0 | loss 3.3932 | lr 9.60e-04 | grad 0.2282 +2026-04-10 09:29:41 - INFO - parrotllm.training - step 54070 | epoch 0 | loss 3.4654 | lr 9.60e-04 | grad 0.2551 +2026-04-10 09:29:44 - INFO - parrotllm.training - step 54080 | epoch 0 | loss 3.2921 | lr 9.60e-04 | grad 0.2323 +2026-04-10 09:29:47 - INFO - parrotllm.training - step 54090 | epoch 0 | loss 3.3523 | lr 9.59e-04 | grad 0.2360 +2026-04-10 09:29:50 - INFO - parrotllm.training - step 54100 | epoch 0 | loss 3.4828 | lr 9.59e-04 | grad 0.2215 +2026-04-10 09:29:53 - INFO - parrotllm.training - step 54110 | epoch 0 | loss 3.4483 | lr 9.59e-04 | grad 0.2526 +2026-04-10 09:29:56 - INFO - parrotllm.training - step 54120 | epoch 0 | loss 3.5722 | lr 9.59e-04 | grad 0.2890 +2026-04-10 09:29:59 - INFO - parrotllm.training - step 54130 | epoch 0 | loss 3.3562 | lr 9.59e-04 | grad 0.2250 +2026-04-10 09:30:02 - INFO - parrotllm.training - step 54140 | epoch 0 | loss 3.4072 | lr 9.59e-04 | grad 0.2441 +2026-04-10 09:30:05 - INFO - parrotllm.training - step 54150 | epoch 0 | loss 3.4169 | lr 9.58e-04 | grad 0.2477 +2026-04-10 09:30:08 - INFO - parrotllm.training - step 54160 | epoch 0 | loss 3.4240 | lr 9.58e-04 | grad 0.2155 +2026-04-10 09:30:11 - INFO - parrotllm.training - step 54170 | epoch 0 | loss 3.4643 | lr 9.58e-04 | grad 0.2792 +2026-04-10 09:30:15 - INFO - parrotllm.training - step 54180 | epoch 0 | loss 3.3247 | lr 9.58e-04 | grad 0.2309 +2026-04-10 09:30:18 - INFO - parrotllm.training - step 54190 | epoch 0 | loss 3.4414 | lr 9.58e-04 | grad 0.2632 +2026-04-10 09:30:21 - INFO - parrotllm.training - step 54200 | epoch 0 | loss 3.4178 | lr 9.58e-04 | grad 0.2602 +2026-04-10 09:30:24 - INFO - parrotllm.training - step 54210 | epoch 0 | loss 3.3635 | lr 9.57e-04 | grad 0.2525 +2026-04-10 09:30:27 - INFO - parrotllm.training - step 54220 | epoch 0 | loss 3.3872 | lr 9.57e-04 | grad 0.2640 +2026-04-10 09:30:30 - INFO - parrotllm.training - step 54230 | epoch 0 | loss 3.3614 | lr 9.57e-04 | grad 0.2359 +2026-04-10 09:30:33 - INFO - parrotllm.training - step 54240 | epoch 0 | loss 3.5120 | lr 9.57e-04 | grad 0.2486 +2026-04-10 09:30:36 - INFO - parrotllm.training - step 54250 | epoch 0 | loss 3.4498 | lr 9.57e-04 | grad 0.2365 +2026-04-10 09:30:39 - INFO - parrotllm.training - step 54260 | epoch 0 | loss 3.4304 | lr 9.57e-04 | grad 0.2365 +2026-04-10 09:30:42 - INFO - parrotllm.training - step 54270 | epoch 0 | loss 3.4515 | lr 9.56e-04 | grad 0.2269 +2026-04-10 09:30:45 - INFO - parrotllm.training - step 54280 | epoch 0 | loss 3.3295 | lr 9.56e-04 | grad 0.2420 +2026-04-10 09:30:48 - INFO - parrotllm.training - step 54290 | epoch 0 | loss 3.5062 | lr 9.56e-04 | grad 0.2429 +2026-04-10 09:30:51 - INFO - parrotllm.training - step 54300 | epoch 0 | loss 3.4043 | lr 9.56e-04 | grad 0.2432 +2026-04-10 09:30:54 - INFO - parrotllm.training - step 54310 | epoch 0 | loss 3.4654 | lr 9.56e-04 | grad 0.2353 +2026-04-10 09:30:57 - INFO - parrotllm.training - step 54320 | epoch 0 | loss 3.4395 | lr 9.55e-04 | grad 0.2534 +2026-04-10 09:31:00 - INFO - parrotllm.training - step 54330 | epoch 0 | loss 3.6138 | lr 9.55e-04 | grad 0.2289 +2026-04-10 09:31:03 - INFO - parrotllm.training - step 54340 | epoch 0 | loss 3.4403 | lr 9.55e-04 | grad 0.2763 +2026-04-10 09:31:06 - INFO - parrotllm.training - step 54350 | epoch 0 | loss 3.4523 | lr 9.55e-04 | grad 0.2459 +2026-04-10 09:31:09 - INFO - parrotllm.training - step 54360 | epoch 0 | loss 3.4323 | lr 9.55e-04 | grad 0.2975 +2026-04-10 09:31:12 - INFO - parrotllm.training - step 54370 | epoch 0 | loss 3.3808 | lr 9.55e-04 | grad 0.2507 +2026-04-10 09:31:16 - INFO - parrotllm.training - step 54380 | epoch 0 | loss 3.4442 | lr 9.54e-04 | grad 0.2554 +2026-04-10 09:31:19 - INFO - parrotllm.training - step 54390 | epoch 0 | loss 3.4487 | lr 9.54e-04 | grad 0.2641 +2026-04-10 09:31:22 - INFO - parrotllm.training - step 54400 | epoch 0 | loss 3.3825 | lr 9.54e-04 | grad 0.2193 +2026-04-10 09:31:25 - INFO - parrotllm.training - step 54410 | epoch 0 | loss 3.4241 | lr 9.54e-04 | grad 0.2393 +2026-04-10 09:31:28 - INFO - parrotllm.training - step 54420 | epoch 0 | loss 3.4481 | lr 9.54e-04 | grad 0.2287 +2026-04-10 09:31:31 - INFO - parrotllm.training - step 54430 | epoch 0 | loss 3.3925 | lr 9.54e-04 | grad 0.2446 +2026-04-10 09:31:34 - INFO - parrotllm.training - step 54440 | epoch 0 | loss 3.4411 | lr 9.53e-04 | grad 0.2272 +2026-04-10 09:31:37 - INFO - parrotllm.training - step 54450 | epoch 0 | loss 3.3254 | lr 9.53e-04 | grad 0.2184 +2026-04-10 09:31:40 - INFO - parrotllm.training - step 54460 | epoch 0 | loss 3.3770 | lr 9.53e-04 | grad 0.2317 +2026-04-10 09:31:43 - INFO - parrotllm.training - step 54470 | epoch 0 | loss 3.4856 | lr 9.53e-04 | grad 0.2595 +2026-04-10 09:31:46 - INFO - parrotllm.training - step 54480 | epoch 0 | loss 3.4441 | lr 9.53e-04 | grad 0.2285 +2026-04-10 09:31:49 - INFO - parrotllm.training - step 54490 | epoch 0 | loss 3.4208 | lr 9.53e-04 | grad 0.2278 +2026-04-10 09:31:52 - INFO - parrotllm.training - step 54500 | epoch 0 | loss 3.4656 | lr 9.52e-04 | grad 0.2358 +2026-04-10 09:31:52 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:31:52 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:31:55 - INFO - parrotllm.training - Train: loss=3.4656, ppl=32.00 +2026-04-10 09:31:55 - INFO - parrotllm.training - Val: loss=3.3525, ppl=28.57 +2026-04-10 09:31:55 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 09:31:56 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3525_epoch_0000_step_0054500.pt +2026-04-10 09:31:58 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:32:01 - INFO - parrotllm.training - step 54510 | epoch 0 | loss 3.4202 | lr 9.52e-04 | grad 0.2342 +2026-04-10 09:32:04 - INFO - parrotllm.training - step 54520 | epoch 0 | loss 3.4246 | lr 9.52e-04 | grad 0.2417 +2026-04-10 09:32:07 - INFO - parrotllm.training - step 54530 | epoch 0 | loss 3.4605 | lr 9.52e-04 | grad 0.2260 +2026-04-10 09:32:10 - INFO - parrotllm.training - step 54540 | epoch 0 | loss 3.4331 | lr 9.52e-04 | grad 0.2375 +2026-04-10 09:32:13 - INFO - parrotllm.training - step 54550 | epoch 0 | loss 3.3600 | lr 9.51e-04 | grad 0.2344 +2026-04-10 09:32:16 - INFO - parrotllm.training - step 54560 | epoch 0 | loss 3.3832 | lr 9.51e-04 | grad 0.2716 +2026-04-10 09:32:19 - INFO - parrotllm.training - step 54570 | epoch 0 | loss 3.3951 | lr 9.51e-04 | grad 0.2194 +2026-04-10 09:32:22 - INFO - parrotllm.training - step 54580 | epoch 0 | loss 3.5377 | lr 9.51e-04 | grad 0.2299 +2026-04-10 09:32:25 - INFO - parrotllm.training - step 54590 | epoch 0 | loss 3.4532 | lr 9.51e-04 | grad 0.2459 +2026-04-10 09:32:28 - INFO - parrotllm.training - step 54600 | epoch 0 | loss 3.4864 | lr 9.51e-04 | grad 0.2405 +2026-04-10 09:32:31 - INFO - parrotllm.training - step 54610 | epoch 0 | loss 3.4615 | lr 9.50e-04 | grad 0.2397 +2026-04-10 09:32:34 - INFO - parrotllm.training - step 54620 | epoch 0 | loss 3.3392 | lr 9.50e-04 | grad 0.2195 +2026-04-10 09:32:37 - INFO - parrotllm.training - step 54630 | epoch 0 | loss 3.3927 | lr 9.50e-04 | grad 0.2154 +2026-04-10 09:32:41 - INFO - parrotllm.training - step 54640 | epoch 0 | loss 3.4554 | lr 9.50e-04 | grad 0.2333 +2026-04-10 09:32:44 - INFO - parrotllm.training - step 54650 | epoch 0 | loss 3.3757 | lr 9.50e-04 | grad 0.2307 +2026-04-10 09:32:47 - INFO - parrotllm.training - step 54660 | epoch 0 | loss 3.2914 | lr 9.50e-04 | grad 0.2537 +2026-04-10 09:32:50 - INFO - parrotllm.training - step 54670 | epoch 0 | loss 3.3727 | lr 9.49e-04 | grad 0.2295 +2026-04-10 09:32:53 - INFO - parrotllm.training - step 54680 | epoch 0 | loss 3.4481 | lr 9.49e-04 | grad 0.2468 +2026-04-10 09:32:56 - INFO - parrotllm.training - step 54690 | epoch 0 | loss 3.4068 | lr 9.49e-04 | grad 0.2189 +2026-04-10 09:32:59 - INFO - parrotllm.training - step 54700 | epoch 0 | loss 3.3851 | lr 9.49e-04 | grad 0.2284 +2026-04-10 09:33:02 - INFO - parrotllm.training - step 54710 | epoch 0 | loss 3.4237 | lr 9.49e-04 | grad 0.2343 +2026-04-10 09:33:05 - INFO - parrotllm.training - step 54720 | epoch 0 | loss 3.3534 | lr 9.49e-04 | grad 0.2474 +2026-04-10 09:33:08 - INFO - parrotllm.training - step 54730 | epoch 0 | loss 3.4387 | lr 9.48e-04 | grad 0.2300 +2026-04-10 09:33:11 - INFO - parrotllm.training - step 54740 | epoch 0 | loss 3.4484 | lr 9.48e-04 | grad 0.2597 +2026-04-10 09:33:14 - INFO - parrotllm.training - step 54750 | epoch 0 | loss 3.4266 | lr 9.48e-04 | grad 0.2297 +2026-04-10 09:33:17 - INFO - parrotllm.training - step 54760 | epoch 0 | loss 3.4313 | lr 9.48e-04 | grad 0.2249 +2026-04-10 09:33:20 - INFO - parrotllm.training - step 54770 | epoch 0 | loss 3.4084 | lr 9.48e-04 | grad 0.2239 +2026-04-10 09:33:23 - INFO - parrotllm.training - step 54780 | epoch 0 | loss 3.4052 | lr 9.48e-04 | grad 0.2389 +2026-04-10 09:33:26 - INFO - parrotllm.training - step 54790 | epoch 0 | loss 3.4962 | lr 9.47e-04 | grad 0.2455 +2026-04-10 09:33:29 - INFO - parrotllm.training - step 54800 | epoch 0 | loss 3.3764 | lr 9.47e-04 | grad 0.2502 +2026-04-10 09:33:32 - INFO - parrotllm.training - step 54810 | epoch 0 | loss 3.3964 | lr 9.47e-04 | grad 0.2492 +2026-04-10 09:33:36 - INFO - parrotllm.training - step 54820 | epoch 0 | loss 3.3520 | lr 9.47e-04 | grad 0.2262 +2026-04-10 09:33:39 - INFO - parrotllm.training - step 54830 | epoch 0 | loss 3.3961 | lr 9.47e-04 | grad 0.2464 +2026-04-10 09:33:42 - INFO - parrotllm.training - step 54840 | epoch 0 | loss 3.3654 | lr 9.46e-04 | grad 0.2349 +2026-04-10 09:33:45 - INFO - parrotllm.training - step 54850 | epoch 0 | loss 3.3589 | lr 9.46e-04 | grad 0.2790 +2026-04-10 09:33:48 - INFO - parrotllm.training - step 54860 | epoch 0 | loss 3.5509 | lr 9.46e-04 | grad 0.2426 +2026-04-10 09:33:51 - INFO - parrotllm.training - step 54870 | epoch 0 | loss 3.4954 | lr 9.46e-04 | grad 0.2462 +2026-04-10 09:33:54 - INFO - parrotllm.training - step 54880 | epoch 0 | loss 3.3104 | lr 9.46e-04 | grad 0.2430 +2026-04-10 09:33:57 - INFO - parrotllm.training - step 54890 | epoch 0 | loss 3.4366 | lr 9.46e-04 | grad 0.2324 +2026-04-10 09:34:00 - INFO - parrotllm.training - step 54900 | epoch 0 | loss 3.3186 | lr 9.45e-04 | grad 0.2346 +2026-04-10 09:34:03 - INFO - parrotllm.training - step 54910 | epoch 0 | loss 3.4688 | lr 9.45e-04 | grad 0.2430 +2026-04-10 09:34:06 - INFO - parrotllm.training - step 54920 | epoch 0 | loss 3.4463 | lr 9.45e-04 | grad 0.2370 +2026-04-10 09:34:09 - INFO - parrotllm.training - step 54930 | epoch 0 | loss 3.4261 | lr 9.45e-04 | grad 0.2436 +2026-04-10 09:34:12 - INFO - parrotllm.training - step 54940 | epoch 0 | loss 3.3982 | lr 9.45e-04 | grad 0.2623 +2026-04-10 09:34:15 - INFO - parrotllm.training - step 54950 | epoch 0 | loss 3.4337 | lr 9.45e-04 | grad 0.2447 +2026-04-10 09:34:18 - INFO - parrotllm.training - step 54960 | epoch 0 | loss 3.4186 | lr 9.44e-04 | grad 0.2686 +2026-04-10 09:34:21 - INFO - parrotllm.training - step 54970 | epoch 0 | loss 3.4659 | lr 9.44e-04 | grad 0.2452 +2026-04-10 09:34:24 - INFO - parrotllm.training - step 54980 | epoch 0 | loss 3.4146 | lr 9.44e-04 | grad 0.2548 +2026-04-10 09:34:27 - INFO - parrotllm.training - step 54990 | epoch 0 | loss 3.4304 | lr 9.44e-04 | grad 0.2751 +2026-04-10 09:34:30 - INFO - parrotllm.training - step 55000 | epoch 0 | loss 3.4389 | lr 9.44e-04 | grad 0.2280 +2026-04-10 09:34:30 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:34:30 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:34:34 - INFO - parrotllm.training - Train: loss=3.4389, ppl=31.15 +2026-04-10 09:34:34 - INFO - parrotllm.training - Val: loss=3.3526, ppl=28.58 +2026-04-10 09:34:35 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3526_epoch_0000_step_0055000.pt +2026-04-10 09:34:36 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3525, min_delta=0.001000). +2026-04-10 09:34:36 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:34:37 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0055000.pt +2026-04-10 09:34:42 - INFO - parrotllm.training - step 55010 | epoch 0 | loss 3.3694 | lr 9.44e-04 | grad 0.2500 +2026-04-10 09:34:45 - INFO - parrotllm.training - step 55020 | epoch 0 | loss 3.3738 | lr 9.43e-04 | grad 0.2457 +2026-04-10 09:34:48 - INFO - parrotllm.training - step 55030 | epoch 0 | loss 3.3696 | lr 9.43e-04 | grad 0.2756 +2026-04-10 09:34:51 - INFO - parrotllm.training - step 55040 | epoch 0 | loss 3.4084 | lr 9.43e-04 | grad 0.2344 +2026-04-10 09:34:54 - INFO - parrotllm.training - step 55050 | epoch 0 | loss 3.3696 | lr 9.43e-04 | grad 0.2280 +2026-04-10 09:34:57 - INFO - parrotllm.training - step 55060 | epoch 0 | loss 3.4614 | lr 9.43e-04 | grad 0.2657 +2026-04-10 09:35:00 - INFO - parrotllm.training - step 55070 | epoch 0 | loss 3.3714 | lr 9.43e-04 | grad 0.2535 +2026-04-10 09:35:03 - INFO - parrotllm.training - step 55080 | epoch 0 | loss 3.4091 | lr 9.42e-04 | grad 0.2451 +2026-04-10 09:35:06 - INFO - parrotllm.training - step 55090 | epoch 0 | loss 3.4179 | lr 9.42e-04 | grad 0.2190 +2026-04-10 09:35:09 - INFO - parrotllm.training - step 55100 | epoch 0 | loss 3.3650 | lr 9.42e-04 | grad 0.2366 +2026-04-10 09:35:13 - INFO - parrotllm.training - step 55110 | epoch 0 | loss 3.4338 | lr 9.42e-04 | grad 0.2247 +2026-04-10 09:35:16 - INFO - parrotllm.training - step 55120 | epoch 0 | loss 3.3356 | lr 9.42e-04 | grad 0.2148 +2026-04-10 09:35:19 - INFO - parrotllm.training - step 55130 | epoch 0 | loss 3.4974 | lr 9.41e-04 | grad 0.2182 +2026-04-10 09:35:22 - INFO - parrotllm.training - step 55140 | epoch 0 | loss 3.3911 | lr 9.41e-04 | grad 0.2214 +2026-04-10 09:35:25 - INFO - parrotllm.training - step 55150 | epoch 0 | loss 3.2702 | lr 9.41e-04 | grad 0.2280 +2026-04-10 09:35:28 - INFO - parrotllm.training - step 55160 | epoch 0 | loss 3.3393 | lr 9.41e-04 | grad 0.2215 +2026-04-10 09:35:31 - INFO - parrotllm.training - step 55170 | epoch 0 | loss 3.3533 | lr 9.41e-04 | grad 0.2274 +2026-04-10 09:35:34 - INFO - parrotllm.training - step 55180 | epoch 0 | loss 3.4777 | lr 9.41e-04 | grad 0.2386 +2026-04-10 09:35:37 - INFO - parrotllm.training - step 55190 | epoch 0 | loss 3.4398 | lr 9.40e-04 | grad 0.2306 +2026-04-10 09:35:40 - INFO - parrotllm.training - step 55200 | epoch 0 | loss 3.4666 | lr 9.40e-04 | grad 0.2431 +2026-04-10 09:35:43 - INFO - parrotllm.training - step 55210 | epoch 0 | loss 3.4315 | lr 9.40e-04 | grad 0.2221 +2026-04-10 09:35:46 - INFO - parrotllm.training - step 55220 | epoch 0 | loss 3.3717 | lr 9.40e-04 | grad 0.2126 +2026-04-10 09:35:49 - INFO - parrotllm.training - step 55230 | epoch 0 | loss 3.4222 | lr 9.40e-04 | grad 0.2220 +2026-04-10 09:35:52 - INFO - parrotllm.training - step 55240 | epoch 0 | loss 3.3188 | lr 9.40e-04 | grad 0.2517 +2026-04-10 09:35:55 - INFO - parrotllm.training - step 55250 | epoch 0 | loss 3.4596 | lr 9.39e-04 | grad 0.2375 +2026-04-10 09:35:58 - INFO - parrotllm.training - step 55260 | epoch 0 | loss 3.3302 | lr 9.39e-04 | grad 0.2717 +2026-04-10 09:36:01 - INFO - parrotllm.training - step 55270 | epoch 0 | loss 3.4991 | lr 9.39e-04 | grad 0.2380 +2026-04-10 09:36:04 - INFO - parrotllm.training - step 55280 | epoch 0 | loss 3.4120 | lr 9.39e-04 | grad 0.2313 +2026-04-10 09:36:07 - INFO - parrotllm.training - step 55290 | epoch 0 | loss 3.3099 | lr 9.39e-04 | grad 0.2443 +2026-04-10 09:36:11 - INFO - parrotllm.training - step 55300 | epoch 0 | loss 3.4166 | lr 9.39e-04 | grad 0.2360 +2026-04-10 09:36:14 - INFO - parrotllm.training - step 55310 | epoch 0 | loss 3.3915 | lr 9.38e-04 | grad 0.2232 +2026-04-10 09:36:17 - INFO - parrotllm.training - step 55320 | epoch 0 | loss 3.4244 | lr 9.38e-04 | grad 0.2143 +2026-04-10 09:36:20 - INFO - parrotllm.training - step 55330 | epoch 0 | loss 3.4349 | lr 9.38e-04 | grad 0.2173 +2026-04-10 09:36:23 - INFO - parrotllm.training - step 55340 | epoch 0 | loss 3.4852 | lr 9.38e-04 | grad 0.2686 +2026-04-10 09:36:26 - INFO - parrotllm.training - step 55350 | epoch 0 | loss 3.3512 | lr 9.38e-04 | grad 0.2449 +2026-04-10 09:36:29 - INFO - parrotllm.training - step 55360 | epoch 0 | loss 3.3965 | lr 9.37e-04 | grad 0.2512 +2026-04-10 09:36:32 - INFO - parrotllm.training - step 55370 | epoch 0 | loss 3.3716 | lr 9.37e-04 | grad 0.2430 +2026-04-10 09:36:35 - INFO - parrotllm.training - step 55380 | epoch 0 | loss 3.5068 | lr 9.37e-04 | grad 0.2582 +2026-04-10 09:36:38 - INFO - parrotllm.training - step 55390 | epoch 0 | loss 3.4177 | lr 9.37e-04 | grad 0.2344 +2026-04-10 09:36:41 - INFO - parrotllm.training - step 55400 | epoch 0 | loss 3.4608 | lr 9.37e-04 | grad 0.2546 +2026-04-10 09:36:44 - INFO - parrotllm.training - step 55410 | epoch 0 | loss 3.4192 | lr 9.37e-04 | grad 0.2542 +2026-04-10 09:36:47 - INFO - parrotllm.training - step 55420 | epoch 0 | loss 3.4522 | lr 9.36e-04 | grad 0.2211 +2026-04-10 09:36:50 - INFO - parrotllm.training - step 55430 | epoch 0 | loss 3.3910 | lr 9.36e-04 | grad 0.2978 +2026-04-10 09:36:53 - INFO - parrotllm.training - step 55440 | epoch 0 | loss 3.4723 | lr 9.36e-04 | grad 0.2539 +2026-04-10 09:36:56 - INFO - parrotllm.training - step 55450 | epoch 0 | loss 3.4028 | lr 9.36e-04 | grad 0.2215 +2026-04-10 09:36:59 - INFO - parrotllm.training - step 55460 | epoch 0 | loss 3.3581 | lr 9.36e-04 | grad 0.2414 +2026-04-10 09:37:02 - INFO - parrotllm.training - step 55470 | epoch 0 | loss 3.3707 | lr 9.36e-04 | grad 0.2765 +2026-04-10 09:37:05 - INFO - parrotllm.training - step 55480 | epoch 0 | loss 3.3087 | lr 9.35e-04 | grad 0.2289 +2026-04-10 09:37:09 - INFO - parrotllm.training - step 55490 | epoch 0 | loss 3.4460 | lr 9.35e-04 | grad 0.2352 +2026-04-10 09:37:12 - INFO - parrotllm.training - step 55500 | epoch 0 | loss 3.4965 | lr 9.35e-04 | grad 0.2678 +2026-04-10 09:37:12 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:37:12 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:37:15 - INFO - parrotllm.training - Train: loss=3.4965, ppl=33.00 +2026-04-10 09:37:15 - INFO - parrotllm.training - Val: loss=3.3519, ppl=28.56 +2026-04-10 09:37:16 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3519_epoch_0000_step_0055500.pt +2026-04-10 09:37:17 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3525, min_delta=0.001000). +2026-04-10 09:37:17 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:37:20 - INFO - parrotllm.training - step 55510 | epoch 0 | loss 3.4566 | lr 9.35e-04 | grad 0.3116 +2026-04-10 09:37:23 - INFO - parrotllm.training - step 55520 | epoch 0 | loss 3.3934 | lr 9.35e-04 | grad 0.2395 +2026-04-10 09:37:26 - INFO - parrotllm.training - step 55530 | epoch 0 | loss 3.2987 | lr 9.35e-04 | grad 0.2628 +2026-04-10 09:37:29 - INFO - parrotllm.training - step 55540 | epoch 0 | loss 3.3559 | lr 9.34e-04 | grad 0.2515 +2026-04-10 09:37:32 - INFO - parrotllm.training - step 55550 | epoch 0 | loss 3.3817 | lr 9.34e-04 | grad 0.2202 +2026-04-10 09:37:35 - INFO - parrotllm.training - step 55560 | epoch 0 | loss 3.4498 | lr 9.34e-04 | grad 0.2316 +2026-04-10 09:37:38 - INFO - parrotllm.training - step 55570 | epoch 0 | loss 3.4489 | lr 9.34e-04 | grad 0.2975 +2026-04-10 09:37:42 - INFO - parrotllm.training - step 55580 | epoch 0 | loss 3.3286 | lr 9.34e-04 | grad 0.2108 +2026-04-10 09:37:45 - INFO - parrotllm.training - step 55590 | epoch 0 | loss 3.3568 | lr 9.33e-04 | grad 0.2432 +2026-04-10 09:37:48 - INFO - parrotllm.training - step 55600 | epoch 0 | loss 3.3013 | lr 9.33e-04 | grad 0.2662 +2026-04-10 09:37:51 - INFO - parrotllm.training - step 55610 | epoch 0 | loss 3.3737 | lr 9.33e-04 | grad 0.2694 +2026-04-10 09:37:54 - INFO - parrotllm.training - step 55620 | epoch 0 | loss 3.3212 | lr 9.33e-04 | grad 0.2268 +2026-04-10 09:37:57 - INFO - parrotllm.training - step 55630 | epoch 0 | loss 3.3852 | lr 9.33e-04 | grad 0.2317 +2026-04-10 09:38:00 - INFO - parrotllm.training - step 55640 | epoch 0 | loss 3.3053 | lr 9.33e-04 | grad 0.2570 +2026-04-10 09:38:03 - INFO - parrotllm.training - step 55650 | epoch 0 | loss 3.3743 | lr 9.32e-04 | grad 0.2149 +2026-04-10 09:38:06 - INFO - parrotllm.training - step 55660 | epoch 0 | loss 3.4602 | lr 9.32e-04 | grad 0.2310 +2026-04-10 09:38:09 - INFO - parrotllm.training - step 55670 | epoch 0 | loss 3.3505 | lr 9.32e-04 | grad 0.2541 +2026-04-10 09:38:12 - INFO - parrotllm.training - step 55680 | epoch 0 | loss 3.4235 | lr 9.32e-04 | grad 0.2305 +2026-04-10 09:38:15 - INFO - parrotllm.training - step 55690 | epoch 0 | loss 3.4049 | lr 9.32e-04 | grad 0.2491 +2026-04-10 09:38:18 - INFO - parrotllm.training - step 55700 | epoch 0 | loss 3.4696 | lr 9.32e-04 | grad 0.2488 +2026-04-10 09:38:21 - INFO - parrotllm.training - step 55710 | epoch 0 | loss 3.3977 | lr 9.31e-04 | grad 0.2419 +2026-04-10 09:38:24 - INFO - parrotllm.training - step 55720 | epoch 0 | loss 3.4256 | lr 9.31e-04 | grad 0.2442 +2026-04-10 09:38:27 - INFO - parrotllm.training - step 55730 | epoch 0 | loss 3.4492 | lr 9.31e-04 | grad 0.2735 +2026-04-10 09:38:30 - INFO - parrotllm.training - step 55740 | epoch 0 | loss 3.4273 | lr 9.31e-04 | grad 0.2393 +2026-04-10 09:38:33 - INFO - parrotllm.training - step 55750 | epoch 0 | loss 3.5380 | lr 9.31e-04 | grad 0.2657 +2026-04-10 09:38:36 - INFO - parrotllm.training - step 55760 | epoch 0 | loss 3.4343 | lr 9.31e-04 | grad 0.2379 +2026-04-10 09:38:39 - INFO - parrotllm.training - step 55770 | epoch 0 | loss 3.4572 | lr 9.30e-04 | grad 0.2369 +2026-04-10 09:38:43 - INFO - parrotllm.training - step 55780 | epoch 0 | loss 3.4668 | lr 9.30e-04 | grad 0.2219 +2026-04-10 09:38:46 - INFO - parrotllm.training - step 55790 | epoch 0 | loss 3.4642 | lr 9.30e-04 | grad 0.2240 +2026-04-10 09:38:49 - INFO - parrotllm.training - step 55800 | epoch 0 | loss 3.4390 | lr 9.30e-04 | grad 0.2342 +2026-04-10 09:38:52 - INFO - parrotllm.training - step 55810 | epoch 0 | loss 3.4389 | lr 9.30e-04 | grad 0.2609 +2026-04-10 09:38:55 - INFO - parrotllm.training - step 55820 | epoch 0 | loss 3.3851 | lr 9.29e-04 | grad 0.2524 +2026-04-10 09:38:58 - INFO - parrotllm.training - step 55830 | epoch 0 | loss 3.3591 | lr 9.29e-04 | grad 0.2300 +2026-04-10 09:39:01 - INFO - parrotllm.training - step 55840 | epoch 0 | loss 3.3538 | lr 9.29e-04 | grad 0.2316 +2026-04-10 09:39:04 - INFO - parrotllm.training - step 55850 | epoch 0 | loss 3.4083 | lr 9.29e-04 | grad 0.2467 +2026-04-10 09:39:07 - INFO - parrotllm.training - step 55860 | epoch 0 | loss 3.5083 | lr 9.29e-04 | grad 0.2346 +2026-04-10 09:39:10 - INFO - parrotllm.training - step 55870 | epoch 0 | loss 3.4224 | lr 9.29e-04 | grad 0.2709 +2026-04-10 09:39:13 - INFO - parrotllm.training - step 55880 | epoch 0 | loss 3.4320 | lr 9.28e-04 | grad 0.2428 +2026-04-10 09:39:16 - INFO - parrotllm.training - step 55890 | epoch 0 | loss 3.3239 | lr 9.28e-04 | grad 0.2234 +2026-04-10 09:39:19 - INFO - parrotllm.training - step 55900 | epoch 0 | loss 3.3709 | lr 9.28e-04 | grad 0.2330 +2026-04-10 09:39:22 - INFO - parrotllm.training - step 55910 | epoch 0 | loss 3.4041 | lr 9.28e-04 | grad 0.2613 +2026-04-10 09:39:25 - INFO - parrotllm.training - step 55920 | epoch 0 | loss 3.4410 | lr 9.28e-04 | grad 0.2856 +2026-04-10 09:39:28 - INFO - parrotllm.training - step 55930 | epoch 0 | loss 3.3627 | lr 9.28e-04 | grad 0.2330 +2026-04-10 09:39:31 - INFO - parrotllm.training - step 55940 | epoch 0 | loss 3.3887 | lr 9.27e-04 | grad 0.2472 +2026-04-10 09:39:34 - INFO - parrotllm.training - step 55950 | epoch 0 | loss 3.4309 | lr 9.27e-04 | grad 0.2366 +2026-04-10 09:39:37 - INFO - parrotllm.training - step 55960 | epoch 0 | loss 3.3344 | lr 9.27e-04 | grad 0.2459 +2026-04-10 09:39:41 - INFO - parrotllm.training - step 55970 | epoch 0 | loss 3.3948 | lr 9.27e-04 | grad 0.2727 +2026-04-10 09:39:44 - INFO - parrotllm.training - step 55980 | epoch 0 | loss 3.3848 | lr 9.27e-04 | grad 0.2370 +2026-04-10 09:39:47 - INFO - parrotllm.training - step 55990 | epoch 0 | loss 3.5064 | lr 9.27e-04 | grad 0.2462 +2026-04-10 09:39:50 - INFO - parrotllm.training - step 56000 | epoch 0 | loss 3.4684 | lr 9.26e-04 | grad 0.2453 +2026-04-10 09:39:50 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:39:50 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:39:53 - INFO - parrotllm.training - Train: loss=3.4684, ppl=32.08 +2026-04-10 09:39:53 - INFO - parrotllm.training - Val: loss=3.3493, ppl=28.48 +2026-04-10 09:39:53 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 09:39:54 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3493_epoch_0000_step_0056000.pt +2026-04-10 09:39:55 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:39:58 - INFO - parrotllm.training - step 56010 | epoch 0 | loss 3.4747 | lr 9.26e-04 | grad 0.2262 +2026-04-10 09:40:01 - INFO - parrotllm.training - step 56020 | epoch 0 | loss 3.4045 | lr 9.26e-04 | grad 0.2285 +2026-04-10 09:40:04 - INFO - parrotllm.training - step 56030 | epoch 0 | loss 3.3257 | lr 9.26e-04 | grad 0.2631 +2026-04-10 09:40:07 - INFO - parrotllm.training - step 56040 | epoch 0 | loss 3.3885 | lr 9.26e-04 | grad 0.2349 +2026-04-10 09:40:10 - INFO - parrotllm.training - step 56050 | epoch 0 | loss 3.4873 | lr 9.25e-04 | grad 0.2366 +2026-04-10 09:40:13 - INFO - parrotllm.training - step 56060 | epoch 0 | loss 3.4754 | lr 9.25e-04 | grad 0.2436 +2026-04-10 09:40:16 - INFO - parrotllm.training - step 56070 | epoch 0 | loss 3.4911 | lr 9.25e-04 | grad 0.2615 +2026-04-10 09:40:20 - INFO - parrotllm.training - step 56080 | epoch 0 | loss 3.3895 | lr 9.25e-04 | grad 0.2354 +2026-04-10 09:40:23 - INFO - parrotllm.training - step 56090 | epoch 0 | loss 3.4103 | lr 9.25e-04 | grad 0.2178 +2026-04-10 09:40:26 - INFO - parrotllm.training - step 56100 | epoch 0 | loss 3.3935 | lr 9.25e-04 | grad 0.2336 +2026-04-10 09:40:29 - INFO - parrotllm.training - step 56110 | epoch 0 | loss 3.4218 | lr 9.24e-04 | grad 0.2481 +2026-04-10 09:40:32 - INFO - parrotllm.training - step 56120 | epoch 0 | loss 3.4549 | lr 9.24e-04 | grad 0.2614 +2026-04-10 09:40:35 - INFO - parrotllm.training - step 56130 | epoch 0 | loss 3.4563 | lr 9.24e-04 | grad 0.2564 +2026-04-10 09:40:38 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 56161/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 09:40:38 - INFO - parrotllm.training - step 56140 | epoch 0 | loss 3.3646 | lr 9.24e-04 | grad 0.2441 +2026-04-10 09:40:41 - INFO - parrotllm.training - step 56150 | epoch 0 | loss 3.4091 | lr 9.24e-04 | grad 0.2415 +2026-04-10 09:40:44 - INFO - parrotllm.training - step 56160 | epoch 0 | loss 3.4779 | lr 9.24e-04 | grad 0.2707 +2026-04-10 09:40:47 - INFO - parrotllm.training - step 56170 | epoch 0 | loss 3.4147 | lr 9.23e-04 | grad 0.2571 +2026-04-10 09:40:50 - INFO - parrotllm.training - step 56180 | epoch 0 | loss 3.4460 | lr 9.23e-04 | grad 0.2363 +2026-04-10 09:40:53 - INFO - parrotllm.training - step 56190 | epoch 0 | loss 3.4098 | lr 9.23e-04 | grad 0.2409 +2026-04-10 09:40:56 - INFO - parrotllm.training - step 56200 | epoch 0 | loss 3.3930 | lr 9.23e-04 | grad 0.2401 +2026-04-10 09:40:59 - INFO - parrotllm.training - step 56210 | epoch 0 | loss 3.4608 | lr 9.23e-04 | grad 0.2730 +2026-04-10 09:41:02 - INFO - parrotllm.training - step 56220 | epoch 0 | loss 3.4991 | lr 9.23e-04 | grad 0.2337 +2026-04-10 09:41:06 - INFO - parrotllm.training - step 56230 | epoch 0 | loss 3.4714 | lr 9.22e-04 | grad 0.2404 +2026-04-10 09:41:09 - INFO - parrotllm.training - step 56240 | epoch 0 | loss 3.4454 | lr 9.22e-04 | grad 0.2505 +2026-04-10 09:41:12 - INFO - parrotllm.training - step 56250 | epoch 0 | loss 3.3827 | lr 9.22e-04 | grad 0.2361 +2026-04-10 09:41:15 - INFO - parrotllm.training - step 56260 | epoch 0 | loss 3.5003 | lr 9.22e-04 | grad 0.2271 +2026-04-10 09:41:18 - INFO - parrotllm.training - step 56270 | epoch 0 | loss 3.4122 | lr 9.22e-04 | grad 0.2383 +2026-04-10 09:41:21 - INFO - parrotllm.training - step 56280 | epoch 0 | loss 3.4637 | lr 9.21e-04 | grad 0.2524 +2026-04-10 09:41:24 - INFO - parrotllm.training - step 56290 | epoch 0 | loss 3.5393 | lr 9.21e-04 | grad 0.2425 +2026-04-10 09:41:27 - INFO - parrotllm.training - step 56300 | epoch 0 | loss 3.3733 | lr 9.21e-04 | grad 0.2414 +2026-04-10 09:41:30 - INFO - parrotllm.training - step 56310 | epoch 0 | loss 3.3588 | lr 9.21e-04 | grad 0.2452 +2026-04-10 09:41:33 - INFO - parrotllm.training - step 56320 | epoch 0 | loss 3.3658 | lr 9.21e-04 | grad 0.2485 +2026-04-10 09:41:36 - INFO - parrotllm.training - step 56330 | epoch 0 | loss 3.3077 | lr 9.21e-04 | grad 0.2563 +2026-04-10 09:41:39 - INFO - parrotllm.training - step 56340 | epoch 0 | loss 3.4739 | lr 9.20e-04 | grad 0.2324 +2026-04-10 09:41:42 - INFO - parrotllm.training - step 56350 | epoch 0 | loss 3.4189 | lr 9.20e-04 | grad 0.2381 +2026-04-10 09:41:45 - INFO - parrotllm.training - step 56360 | epoch 0 | loss 3.4672 | lr 9.20e-04 | grad 0.2497 +2026-04-10 09:41:48 - INFO - parrotllm.training - step 56370 | epoch 0 | loss 3.3909 | lr 9.20e-04 | grad 0.2835 +2026-04-10 09:41:51 - INFO - parrotllm.training - step 56380 | epoch 0 | loss 3.3822 | lr 9.20e-04 | grad 0.2494 +2026-04-10 09:41:54 - INFO - parrotllm.training - step 56390 | epoch 0 | loss 3.3508 | lr 9.20e-04 | grad 0.2420 +2026-04-10 09:41:57 - INFO - parrotllm.training - step 56400 | epoch 0 | loss 3.4264 | lr 9.19e-04 | grad 0.2548 +2026-04-10 09:42:00 - INFO - parrotllm.training - step 56410 | epoch 0 | loss 3.3648 | lr 9.19e-04 | grad 0.2304 +2026-04-10 09:42:03 - INFO - parrotllm.training - step 56420 | epoch 0 | loss 3.4106 | lr 9.19e-04 | grad 0.2610 +2026-04-10 09:42:07 - INFO - parrotllm.training - step 56430 | epoch 0 | loss 3.3948 | lr 9.19e-04 | grad 0.2801 +2026-04-10 09:42:10 - INFO - parrotllm.training - step 56440 | epoch 0 | loss 3.4613 | lr 9.19e-04 | grad 0.2374 +2026-04-10 09:42:13 - INFO - parrotllm.training - step 56450 | epoch 0 | loss 3.4323 | lr 9.19e-04 | grad 0.2677 +2026-04-10 09:42:16 - INFO - parrotllm.training - step 56460 | epoch 0 | loss 3.3819 | lr 9.18e-04 | grad 0.2627 +2026-04-10 09:42:19 - INFO - parrotllm.training - step 56470 | epoch 0 | loss 3.4304 | lr 9.18e-04 | grad 0.2677 +2026-04-10 09:42:22 - INFO - parrotllm.training - step 56480 | epoch 0 | loss 3.3165 | lr 9.18e-04 | grad 0.2224 +2026-04-10 09:42:25 - INFO - parrotllm.training - step 56490 | epoch 0 | loss 3.4559 | lr 9.18e-04 | grad 0.2312 +2026-04-10 09:42:28 - INFO - parrotllm.training - step 56500 | epoch 0 | loss 3.4657 | lr 9.18e-04 | grad 0.2434 +2026-04-10 09:42:28 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:42:28 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:42:31 - INFO - parrotllm.training - Train: loss=3.4657, ppl=32.00 +2026-04-10 09:42:31 - INFO - parrotllm.training - Val: loss=3.3501, ppl=28.50 +2026-04-10 09:42:32 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3501_epoch_0000_step_0056500.pt +2026-04-10 09:42:33 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3493, min_delta=0.001000). +2026-04-10 09:42:33 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:42:36 - INFO - parrotllm.training - step 56510 | epoch 0 | loss 3.5102 | lr 9.17e-04 | grad 0.2676 +2026-04-10 09:42:39 - INFO - parrotllm.training - step 56520 | epoch 0 | loss 3.3898 | lr 9.17e-04 | grad 0.2517 +2026-04-10 09:42:42 - INFO - parrotllm.training - step 56530 | epoch 0 | loss 3.2867 | lr 9.17e-04 | grad 0.2519 +2026-04-10 09:42:46 - INFO - parrotllm.training - step 56540 | epoch 0 | loss 3.3829 | lr 9.17e-04 | grad 0.2515 +2026-04-10 09:42:49 - INFO - parrotllm.training - step 56550 | epoch 0 | loss 3.4427 | lr 9.17e-04 | grad 0.2413 +2026-04-10 09:42:52 - INFO - parrotllm.training - step 56560 | epoch 0 | loss 3.4316 | lr 9.17e-04 | grad 0.2239 +2026-04-10 09:42:55 - INFO - parrotllm.training - step 56570 | epoch 0 | loss 3.4100 | lr 9.16e-04 | grad 0.2228 +2026-04-10 09:42:58 - INFO - parrotllm.training - step 56580 | epoch 0 | loss 3.4140 | lr 9.16e-04 | grad 0.2504 +2026-04-10 09:43:01 - INFO - parrotllm.training - step 56590 | epoch 0 | loss 3.4830 | lr 9.16e-04 | grad 0.2713 +2026-04-10 09:43:04 - INFO - parrotllm.training - step 56600 | epoch 0 | loss 3.5411 | lr 9.16e-04 | grad 0.2314 +2026-04-10 09:43:07 - INFO - parrotllm.training - step 56610 | epoch 0 | loss 3.4293 | lr 9.16e-04 | grad 0.2278 +2026-04-10 09:43:10 - INFO - parrotllm.training - step 56620 | epoch 0 | loss 3.5111 | lr 9.16e-04 | grad 0.2315 +2026-04-10 09:43:13 - INFO - parrotllm.training - step 56630 | epoch 0 | loss 3.4644 | lr 9.15e-04 | grad 0.2282 +2026-04-10 09:43:16 - INFO - parrotllm.training - step 56640 | epoch 0 | loss 3.3156 | lr 9.15e-04 | grad 0.2285 +2026-04-10 09:43:19 - INFO - parrotllm.training - step 56650 | epoch 0 | loss 3.4652 | lr 9.15e-04 | grad 0.2481 +2026-04-10 09:43:22 - INFO - parrotllm.training - step 56660 | epoch 0 | loss 3.3882 | lr 9.15e-04 | grad 0.2877 +2026-04-10 09:43:25 - INFO - parrotllm.training - step 56670 | epoch 0 | loss 3.2216 | lr 9.15e-04 | grad 0.2432 +2026-04-10 09:43:28 - INFO - parrotllm.training - step 56680 | epoch 0 | loss 3.4391 | lr 9.15e-04 | grad 0.2503 +2026-04-10 09:43:31 - INFO - parrotllm.training - step 56690 | epoch 0 | loss 3.4043 | lr 9.14e-04 | grad 0.2372 +2026-04-10 09:43:34 - INFO - parrotllm.training - step 56700 | epoch 0 | loss 3.4576 | lr 9.14e-04 | grad 0.2642 +2026-04-10 09:43:37 - INFO - parrotllm.training - step 56710 | epoch 0 | loss 3.4832 | lr 9.14e-04 | grad 0.2336 +2026-04-10 09:43:40 - INFO - parrotllm.training - step 56720 | epoch 0 | loss 3.2958 | lr 9.14e-04 | grad 0.2302 +2026-04-10 09:43:43 - INFO - parrotllm.training - step 56730 | epoch 0 | loss 3.4568 | lr 9.14e-04 | grad 0.2338 +2026-04-10 09:43:46 - INFO - parrotllm.training - step 56740 | epoch 0 | loss 3.3730 | lr 9.13e-04 | grad 0.2390 +2026-04-10 09:43:49 - INFO - parrotllm.training - step 56750 | epoch 0 | loss 3.4645 | lr 9.13e-04 | grad 0.2486 +2026-04-10 09:43:53 - INFO - parrotllm.training - step 56760 | epoch 0 | loss 3.3295 | lr 9.13e-04 | grad 0.2509 +2026-04-10 09:43:56 - INFO - parrotllm.training - step 56770 | epoch 0 | loss 3.3785 | lr 9.13e-04 | grad 0.2237 +2026-04-10 09:43:59 - INFO - parrotllm.training - step 56780 | epoch 0 | loss 3.5598 | lr 9.13e-04 | grad 0.2896 +2026-04-10 09:44:02 - INFO - parrotllm.training - step 56790 | epoch 0 | loss 3.4500 | lr 9.13e-04 | grad 0.2358 +2026-04-10 09:44:05 - INFO - parrotllm.training - step 56800 | epoch 0 | loss 3.2710 | lr 9.12e-04 | grad 0.2249 +2026-04-10 09:44:08 - INFO - parrotllm.training - step 56810 | epoch 0 | loss 3.3282 | lr 9.12e-04 | grad 0.2173 +2026-04-10 09:44:11 - INFO - parrotllm.training - step 56820 | epoch 0 | loss 3.4312 | lr 9.12e-04 | grad 0.2347 +2026-04-10 09:44:14 - INFO - parrotllm.training - step 56830 | epoch 0 | loss 3.5114 | lr 9.12e-04 | grad 0.2664 +2026-04-10 09:44:17 - INFO - parrotllm.training - step 56840 | epoch 0 | loss 3.3775 | lr 9.12e-04 | grad 0.2312 +2026-04-10 09:44:20 - INFO - parrotllm.training - step 56850 | epoch 0 | loss 3.4086 | lr 9.12e-04 | grad 0.2451 +2026-04-10 09:44:23 - INFO - parrotllm.training - step 56860 | epoch 0 | loss 3.5136 | lr 9.11e-04 | grad 0.2225 +2026-04-10 09:44:26 - INFO - parrotllm.training - step 56870 | epoch 0 | loss 3.2802 | lr 9.11e-04 | grad 0.2359 +2026-04-10 09:44:29 - INFO - parrotllm.training - step 56880 | epoch 0 | loss 3.3347 | lr 9.11e-04 | grad 0.2559 +2026-04-10 09:44:32 - INFO - parrotllm.training - step 56890 | epoch 0 | loss 3.4370 | lr 9.11e-04 | grad 0.2603 +2026-04-10 09:44:35 - INFO - parrotllm.training - step 56900 | epoch 0 | loss 3.3526 | lr 9.11e-04 | grad 0.2765 +2026-04-10 09:44:38 - INFO - parrotllm.training - step 56910 | epoch 0 | loss 3.3948 | lr 9.11e-04 | grad 0.2445 +2026-04-10 09:44:41 - INFO - parrotllm.training - step 56920 | epoch 0 | loss 3.4045 | lr 9.10e-04 | grad 0.2366 +2026-04-10 09:44:44 - INFO - parrotllm.training - step 56930 | epoch 0 | loss 3.4048 | lr 9.10e-04 | grad 0.2527 +2026-04-10 09:44:47 - INFO - parrotllm.training - step 56940 | epoch 0 | loss 3.4057 | lr 9.10e-04 | grad 0.2223 +2026-04-10 09:44:50 - INFO - parrotllm.training - step 56950 | epoch 0 | loss 3.3630 | lr 9.10e-04 | grad 0.2344 +2026-04-10 09:44:53 - INFO - parrotllm.training - step 56960 | epoch 0 | loss 3.4481 | lr 9.10e-04 | grad 0.2851 +2026-04-10 09:44:56 - INFO - parrotllm.training - step 56970 | epoch 0 | loss 3.5053 | lr 9.09e-04 | grad 0.2305 +2026-04-10 09:45:00 - INFO - parrotllm.training - step 56980 | epoch 0 | loss 3.3283 | lr 9.09e-04 | grad 0.2731 +2026-04-10 09:45:03 - INFO - parrotllm.training - step 56990 | epoch 0 | loss 3.3976 | lr 9.09e-04 | grad 0.2242 +2026-04-10 09:45:06 - INFO - parrotllm.training - step 57000 | epoch 0 | loss 3.3940 | lr 9.09e-04 | grad 0.2401 +2026-04-10 09:45:06 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:45:06 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:45:09 - INFO - parrotllm.training - Train: loss=3.3940, ppl=29.79 +2026-04-10 09:45:09 - INFO - parrotllm.training - Val: loss=3.3473, ppl=28.43 +2026-04-10 09:45:09 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 09:45:10 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3473_epoch_0000_step_0057000.pt +2026-04-10 09:45:11 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:45:15 - INFO - parrotllm.training - step 57010 | epoch 0 | loss 3.4675 | lr 9.09e-04 | grad 0.2701 +2026-04-10 09:45:18 - INFO - parrotllm.training - step 57020 | epoch 0 | loss 3.2865 | lr 9.09e-04 | grad 0.2434 +2026-04-10 09:45:21 - INFO - parrotllm.training - step 57030 | epoch 0 | loss 3.3878 | lr 9.08e-04 | grad 0.2568 +2026-04-10 09:45:24 - INFO - parrotllm.training - step 57040 | epoch 0 | loss 3.4160 | lr 9.08e-04 | grad 0.2632 +2026-04-10 09:45:27 - INFO - parrotllm.training - step 57050 | epoch 0 | loss 3.3838 | lr 9.08e-04 | grad 0.2291 +2026-04-10 09:45:30 - INFO - parrotllm.training - step 57060 | epoch 0 | loss 3.3710 | lr 9.08e-04 | grad 0.2646 +2026-04-10 09:45:33 - INFO - parrotllm.training - step 57070 | epoch 0 | loss 3.3312 | lr 9.08e-04 | grad 0.2493 +2026-04-10 09:45:36 - INFO - parrotllm.training - step 57080 | epoch 0 | loss 3.4388 | lr 9.08e-04 | grad 0.2380 +2026-04-10 09:45:39 - INFO - parrotllm.training - step 57090 | epoch 0 | loss 3.3225 | lr 9.07e-04 | grad 0.2262 +2026-04-10 09:45:42 - INFO - parrotllm.training - step 57100 | epoch 0 | loss 3.4388 | lr 9.07e-04 | grad 0.2328 +2026-04-10 09:45:45 - INFO - parrotllm.training - step 57110 | epoch 0 | loss 3.4576 | lr 9.07e-04 | grad 0.2397 +2026-04-10 09:45:48 - INFO - parrotllm.training - step 57120 | epoch 0 | loss 3.4195 | lr 9.07e-04 | grad 0.2257 +2026-04-10 09:45:51 - INFO - parrotllm.training - step 57130 | epoch 0 | loss 3.3602 | lr 9.07e-04 | grad 0.2639 +2026-04-10 09:45:54 - INFO - parrotllm.training - step 57140 | epoch 0 | loss 3.4564 | lr 9.06e-04 | grad 0.2393 +2026-04-10 09:45:57 - INFO - parrotllm.training - step 57150 | epoch 0 | loss 3.4778 | lr 9.06e-04 | grad 0.2430 +2026-04-10 09:46:00 - INFO - parrotllm.training - step 57160 | epoch 0 | loss 3.4066 | lr 9.06e-04 | grad 0.2305 +2026-04-10 09:46:03 - INFO - parrotllm.training - step 57170 | epoch 0 | loss 3.3782 | lr 9.06e-04 | grad 0.2518 +2026-04-10 09:46:06 - INFO - parrotllm.training - step 57180 | epoch 0 | loss 3.3319 | lr 9.06e-04 | grad 0.2529 +2026-04-10 09:46:09 - INFO - parrotllm.training - step 57190 | epoch 0 | loss 3.5049 | lr 9.06e-04 | grad 0.2524 +2026-04-10 09:46:12 - INFO - parrotllm.training - step 57200 | epoch 0 | loss 3.4291 | lr 9.05e-04 | grad 0.2610 +2026-04-10 09:46:15 - INFO - parrotllm.training - step 57210 | epoch 0 | loss 3.4706 | lr 9.05e-04 | grad 0.2296 +2026-04-10 09:46:18 - INFO - parrotllm.training - step 57220 | epoch 0 | loss 3.4651 | lr 9.05e-04 | grad 0.2668 +2026-04-10 09:46:22 - INFO - parrotllm.training - step 57230 | epoch 0 | loss 3.2978 | lr 9.05e-04 | grad 0.2367 +2026-04-10 09:46:25 - INFO - parrotllm.training - step 57240 | epoch 0 | loss 3.3495 | lr 9.05e-04 | grad 0.2515 +2026-04-10 09:46:28 - INFO - parrotllm.training - step 57250 | epoch 0 | loss 3.3885 | lr 9.05e-04 | grad 0.2329 +2026-04-10 09:46:31 - INFO - parrotllm.training - step 57260 | epoch 0 | loss 3.4343 | lr 9.04e-04 | grad 0.2413 +2026-04-10 09:46:34 - INFO - parrotllm.training - step 57270 | epoch 0 | loss 3.3967 | lr 9.04e-04 | grad 0.2229 +2026-04-10 09:46:37 - INFO - parrotllm.training - step 57280 | epoch 0 | loss 3.4163 | lr 9.04e-04 | grad 0.2326 +2026-04-10 09:46:40 - INFO - parrotllm.training - step 57290 | epoch 0 | loss 3.3942 | lr 9.04e-04 | grad 0.2625 +2026-04-10 09:46:43 - INFO - parrotllm.training - step 57300 | epoch 0 | loss 3.3697 | lr 9.04e-04 | grad 0.2304 +2026-04-10 09:46:46 - INFO - parrotllm.training - step 57310 | epoch 0 | loss 3.4295 | lr 9.04e-04 | grad 0.2243 +2026-04-10 09:46:49 - INFO - parrotllm.training - step 57320 | epoch 0 | loss 3.3151 | lr 9.03e-04 | grad 0.2789 +2026-04-10 09:46:52 - INFO - parrotllm.training - step 57330 | epoch 0 | loss 3.4853 | lr 9.03e-04 | grad 0.2386 +2026-04-10 09:46:55 - INFO - parrotllm.training - step 57340 | epoch 0 | loss 3.3283 | lr 9.03e-04 | grad 0.2407 +2026-04-10 09:46:58 - INFO - parrotllm.training - step 57350 | epoch 0 | loss 3.4603 | lr 9.03e-04 | grad 0.2418 +2026-04-10 09:47:01 - INFO - parrotllm.training - step 57360 | epoch 0 | loss 3.3471 | lr 9.03e-04 | grad 0.2374 +2026-04-10 09:47:04 - INFO - parrotllm.training - step 57370 | epoch 0 | loss 3.4164 | lr 9.02e-04 | grad 0.2415 +2026-04-10 09:47:07 - INFO - parrotllm.training - step 57380 | epoch 0 | loss 3.3894 | lr 9.02e-04 | grad 0.2342 +2026-04-10 09:47:10 - INFO - parrotllm.training - step 57390 | epoch 0 | loss 3.4707 | lr 9.02e-04 | grad 0.2347 +2026-04-10 09:47:13 - INFO - parrotllm.training - step 57400 | epoch 0 | loss 3.3807 | lr 9.02e-04 | grad 0.2534 +2026-04-10 09:47:16 - INFO - parrotllm.training - step 57410 | epoch 0 | loss 3.3735 | lr 9.02e-04 | grad 0.2414 +2026-04-10 09:47:19 - INFO - parrotllm.training - step 57420 | epoch 0 | loss 3.3598 | lr 9.02e-04 | grad 0.2301 +2026-04-10 09:47:22 - INFO - parrotllm.training - step 57430 | epoch 0 | loss 3.3406 | lr 9.01e-04 | grad 0.2423 +2026-04-10 09:47:25 - INFO - parrotllm.training - step 57440 | epoch 0 | loss 3.3434 | lr 9.01e-04 | grad 0.2351 +2026-04-10 09:47:29 - INFO - parrotllm.training - step 57450 | epoch 0 | loss 3.3955 | lr 9.01e-04 | grad 0.2416 +2026-04-10 09:47:32 - INFO - parrotllm.training - step 57460 | epoch 0 | loss 3.4839 | lr 9.01e-04 | grad 0.2446 +2026-04-10 09:47:35 - INFO - parrotllm.training - step 57470 | epoch 0 | loss 3.4145 | lr 9.01e-04 | grad 0.2356 +2026-04-10 09:47:38 - INFO - parrotllm.training - step 57480 | epoch 0 | loss 3.3789 | lr 9.01e-04 | grad 0.2514 +2026-04-10 09:47:41 - INFO - parrotllm.training - step 57490 | epoch 0 | loss 3.4749 | lr 9.00e-04 | grad 0.2733 +2026-04-10 09:47:44 - INFO - parrotllm.training - step 57500 | epoch 0 | loss 3.4180 | lr 9.00e-04 | grad 0.2665 +2026-04-10 09:47:44 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:47:44 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:47:47 - INFO - parrotllm.training - Train: loss=3.4180, ppl=30.51 +2026-04-10 09:47:47 - INFO - parrotllm.training - Val: loss=3.3445, ppl=28.35 +2026-04-10 09:47:47 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 09:47:48 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3445_epoch_0000_step_0057500.pt +2026-04-10 09:47:49 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:47:50 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0057500.pt +2026-04-10 09:47:55 - INFO - parrotllm.training - step 57510 | epoch 0 | loss 3.4297 | lr 9.00e-04 | grad 0.2281 +2026-04-10 09:47:58 - INFO - parrotllm.training - step 57520 | epoch 0 | loss 3.3819 | lr 9.00e-04 | grad 0.2431 +2026-04-10 09:48:01 - INFO - parrotllm.training - step 57530 | epoch 0 | loss 3.4215 | lr 9.00e-04 | grad 0.2314 +2026-04-10 09:48:04 - INFO - parrotllm.training - step 57540 | epoch 0 | loss 3.4217 | lr 9.00e-04 | grad 0.2298 +2026-04-10 09:48:07 - INFO - parrotllm.training - step 57550 | epoch 0 | loss 3.3310 | lr 8.99e-04 | grad 0.2341 +2026-04-10 09:48:10 - INFO - parrotllm.training - step 57560 | epoch 0 | loss 3.4106 | lr 8.99e-04 | grad 0.2451 +2026-04-10 09:48:13 - INFO - parrotllm.training - step 57570 | epoch 0 | loss 3.3777 | lr 8.99e-04 | grad 0.2129 +2026-04-10 09:48:16 - INFO - parrotllm.training - step 57580 | epoch 0 | loss 3.3011 | lr 8.99e-04 | grad 0.2318 +2026-04-10 09:48:19 - INFO - parrotllm.training - step 57590 | epoch 0 | loss 3.4830 | lr 8.99e-04 | grad 0.2857 +2026-04-10 09:48:22 - INFO - parrotllm.training - step 57600 | epoch 0 | loss 3.4267 | lr 8.98e-04 | grad 0.2355 +2026-04-10 09:48:25 - INFO - parrotllm.training - step 57610 | epoch 0 | loss 3.4098 | lr 8.98e-04 | grad 0.2495 +2026-04-10 09:48:28 - INFO - parrotllm.training - step 57620 | epoch 0 | loss 3.4626 | lr 8.98e-04 | grad 0.2546 +2026-04-10 09:48:31 - INFO - parrotllm.training - step 57630 | epoch 0 | loss 3.4584 | lr 8.98e-04 | grad 0.2474 +2026-04-10 09:48:34 - INFO - parrotllm.training - step 57640 | epoch 0 | loss 3.4130 | lr 8.98e-04 | grad 0.2316 +2026-04-10 09:48:37 - INFO - parrotllm.training - step 57650 | epoch 0 | loss 3.3715 | lr 8.98e-04 | grad 0.2272 +2026-04-10 09:48:41 - INFO - parrotllm.training - step 57660 | epoch 0 | loss 3.4579 | lr 8.97e-04 | grad 0.2313 +2026-04-10 09:48:44 - INFO - parrotllm.training - step 57670 | epoch 0 | loss 3.4308 | lr 8.97e-04 | grad 0.2631 +2026-04-10 09:48:47 - INFO - parrotllm.training - step 57680 | epoch 0 | loss 3.5035 | lr 8.97e-04 | grad 0.2645 +2026-04-10 09:48:50 - INFO - parrotllm.training - step 57690 | epoch 0 | loss 3.3413 | lr 8.97e-04 | grad 0.2550 +2026-04-10 09:48:53 - INFO - parrotllm.training - step 57700 | epoch 0 | loss 3.4533 | lr 8.97e-04 | grad 0.2145 +2026-04-10 09:48:56 - INFO - parrotllm.training - step 57710 | epoch 0 | loss 3.4980 | lr 8.97e-04 | grad 0.2454 +2026-04-10 09:48:59 - INFO - parrotllm.training - step 57720 | epoch 0 | loss 3.4149 | lr 8.96e-04 | grad 0.2316 +2026-04-10 09:49:02 - INFO - parrotllm.training - step 57730 | epoch 0 | loss 3.3593 | lr 8.96e-04 | grad 0.2394 +2026-04-10 09:49:05 - INFO - parrotllm.training - step 57740 | epoch 0 | loss 3.4365 | lr 8.96e-04 | grad 0.2303 +2026-04-10 09:49:08 - INFO - parrotllm.training - step 57750 | epoch 0 | loss 3.4554 | lr 8.96e-04 | grad 0.2364 +2026-04-10 09:49:11 - INFO - parrotllm.training - step 57760 | epoch 0 | loss 3.3794 | lr 8.96e-04 | grad 0.2359 +2026-04-10 09:49:14 - INFO - parrotllm.training - step 57770 | epoch 0 | loss 3.5746 | lr 8.95e-04 | grad 0.2655 +2026-04-10 09:49:17 - INFO - parrotllm.training - step 57780 | epoch 0 | loss 3.3689 | lr 8.95e-04 | grad 0.2268 +2026-04-10 09:49:20 - INFO - parrotllm.training - step 57790 | epoch 0 | loss 3.3775 | lr 8.95e-04 | grad 0.2395 +2026-04-10 09:49:23 - INFO - parrotllm.training - step 57800 | epoch 0 | loss 3.3496 | lr 8.95e-04 | grad 0.2788 +2026-04-10 09:49:26 - INFO - parrotllm.training - step 57810 | epoch 0 | loss 3.4529 | lr 8.95e-04 | grad 0.2355 +2026-04-10 09:49:29 - INFO - parrotllm.training - step 57820 | epoch 0 | loss 3.4675 | lr 8.95e-04 | grad 0.2645 +2026-04-10 09:49:32 - INFO - parrotllm.training - step 57830 | epoch 0 | loss 3.4186 | lr 8.94e-04 | grad 0.2569 +2026-04-10 09:49:35 - INFO - parrotllm.training - step 57840 | epoch 0 | loss 3.3892 | lr 8.94e-04 | grad 0.2504 +2026-04-10 09:49:38 - INFO - parrotllm.training - step 57850 | epoch 0 | loss 3.3878 | lr 8.94e-04 | grad 0.2451 +2026-04-10 09:49:41 - INFO - parrotllm.training - step 57860 | epoch 0 | loss 3.3816 | lr 8.94e-04 | grad 0.2265 +2026-04-10 09:49:45 - INFO - parrotllm.training - step 57870 | epoch 0 | loss 3.4913 | lr 8.94e-04 | grad 0.2466 +2026-04-10 09:49:48 - INFO - parrotllm.training - step 57880 | epoch 0 | loss 3.3454 | lr 8.94e-04 | grad 0.2352 +2026-04-10 09:49:51 - INFO - parrotllm.training - step 57890 | epoch 0 | loss 3.3923 | lr 8.93e-04 | grad 0.2321 +2026-04-10 09:49:54 - INFO - parrotllm.training - step 57900 | epoch 0 | loss 3.3924 | lr 8.93e-04 | grad 0.2522 +2026-04-10 09:49:57 - INFO - parrotllm.training - step 57910 | epoch 0 | loss 3.4326 | lr 8.93e-04 | grad 0.2342 +2026-04-10 09:50:00 - INFO - parrotllm.training - step 57920 | epoch 0 | loss 3.3804 | lr 8.93e-04 | grad 0.2116 +2026-04-10 09:50:03 - INFO - parrotllm.training - step 57930 | epoch 0 | loss 3.4428 | lr 8.93e-04 | grad 0.2404 +2026-04-10 09:50:06 - INFO - parrotllm.training - step 57940 | epoch 0 | loss 3.4544 | lr 8.93e-04 | grad 0.3068 +2026-04-10 09:50:09 - INFO - parrotllm.training - step 57950 | epoch 0 | loss 3.4670 | lr 8.92e-04 | grad 0.2390 +2026-04-10 09:50:12 - INFO - parrotllm.training - step 57960 | epoch 0 | loss 3.3659 | lr 8.92e-04 | grad 0.2356 +2026-04-10 09:50:15 - INFO - parrotllm.training - step 57970 | epoch 0 | loss 3.4594 | lr 8.92e-04 | grad 0.2306 +2026-04-10 09:50:18 - INFO - parrotllm.training - step 57980 | epoch 0 | loss 3.4057 | lr 8.92e-04 | grad 0.2387 +2026-04-10 09:50:21 - INFO - parrotllm.training - step 57990 | epoch 0 | loss 3.4770 | lr 8.92e-04 | grad 0.2720 +2026-04-10 09:50:24 - INFO - parrotllm.training - step 58000 | epoch 0 | loss 3.3656 | lr 8.91e-04 | grad 0.2394 +2026-04-10 09:50:24 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:50:24 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:50:27 - INFO - parrotllm.training - Train: loss=3.3656, ppl=28.95 +2026-04-10 09:50:27 - INFO - parrotllm.training - Val: loss=3.3456, ppl=28.38 +2026-04-10 09:50:28 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3456_epoch_0000_step_0058000.pt +2026-04-10 09:50:30 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3445, min_delta=0.001000). +2026-04-10 09:50:30 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:50:33 - INFO - parrotllm.training - step 58010 | epoch 0 | loss 3.3965 | lr 8.91e-04 | grad 0.2735 +2026-04-10 09:50:36 - INFO - parrotllm.training - step 58020 | epoch 0 | loss 3.4258 | lr 8.91e-04 | grad 0.2269 +2026-04-10 09:50:39 - INFO - parrotllm.training - step 58030 | epoch 0 | loss 3.3591 | lr 8.91e-04 | grad 0.2192 +2026-04-10 09:50:42 - INFO - parrotllm.training - step 58040 | epoch 0 | loss 3.4524 | lr 8.91e-04 | grad 0.2324 +2026-04-10 09:50:45 - INFO - parrotllm.training - step 58050 | epoch 0 | loss 3.3518 | lr 8.91e-04 | grad 0.2485 +2026-04-10 09:50:48 - INFO - parrotllm.training - step 58060 | epoch 0 | loss 3.4394 | lr 8.90e-04 | grad 0.2585 +2026-04-10 09:50:51 - INFO - parrotllm.training - step 58070 | epoch 0 | loss 3.4859 | lr 8.90e-04 | grad 0.2490 +2026-04-10 09:50:54 - INFO - parrotllm.training - step 58080 | epoch 0 | loss 3.4236 | lr 8.90e-04 | grad 0.2727 +2026-04-10 09:50:57 - INFO - parrotllm.training - step 58090 | epoch 0 | loss 3.4972 | lr 8.90e-04 | grad 0.2483 +2026-04-10 09:51:00 - INFO - parrotllm.training - step 58100 | epoch 0 | loss 3.4009 | lr 8.90e-04 | grad 0.2813 +2026-04-10 09:51:03 - INFO - parrotllm.training - step 58110 | epoch 0 | loss 3.3990 | lr 8.90e-04 | grad 0.2379 +2026-04-10 09:51:06 - INFO - parrotllm.training - step 58120 | epoch 0 | loss 3.4805 | lr 8.89e-04 | grad 0.2354 +2026-04-10 09:51:09 - INFO - parrotllm.training - step 58130 | epoch 0 | loss 3.3668 | lr 8.89e-04 | grad 0.2630 +2026-04-10 09:51:12 - INFO - parrotllm.training - step 58140 | epoch 0 | loss 3.4894 | lr 8.89e-04 | grad 0.2281 +2026-04-10 09:51:15 - INFO - parrotllm.training - step 58150 | epoch 0 | loss 3.4289 | lr 8.89e-04 | grad 0.2543 +2026-04-10 09:51:18 - INFO - parrotllm.training - step 58160 | epoch 0 | loss 3.3244 | lr 8.89e-04 | grad 0.2435 +2026-04-10 09:51:22 - INFO - parrotllm.training - step 58170 | epoch 0 | loss 3.3836 | lr 8.88e-04 | grad 0.2654 +2026-04-10 09:51:25 - INFO - parrotllm.training - step 58180 | epoch 0 | loss 3.3508 | lr 8.88e-04 | grad 0.2477 +2026-04-10 09:51:28 - INFO - parrotllm.training - step 58190 | epoch 0 | loss 3.4351 | lr 8.88e-04 | grad 0.2549 +2026-04-10 09:51:31 - INFO - parrotllm.training - step 58200 | epoch 0 | loss 3.4128 | lr 8.88e-04 | grad 0.2384 +2026-04-10 09:51:34 - INFO - parrotllm.training - step 58210 | epoch 0 | loss 3.3074 | lr 8.88e-04 | grad 0.2687 +2026-04-10 09:51:37 - INFO - parrotllm.training - step 58220 | epoch 0 | loss 3.4177 | lr 8.88e-04 | grad 0.2570 +2026-04-10 09:51:40 - INFO - parrotllm.training - step 58230 | epoch 0 | loss 3.3989 | lr 8.87e-04 | grad 0.2266 +2026-04-10 09:51:43 - INFO - parrotllm.training - step 58240 | epoch 0 | loss 3.3384 | lr 8.87e-04 | grad 0.2578 +2026-04-10 09:51:46 - INFO - parrotllm.training - step 58250 | epoch 0 | loss 3.3956 | lr 8.87e-04 | grad 0.2577 +2026-04-10 09:51:49 - INFO - parrotllm.training - step 58260 | epoch 0 | loss 3.4231 | lr 8.87e-04 | grad 0.2311 +2026-04-10 09:51:52 - INFO - parrotllm.training - step 58270 | epoch 0 | loss 3.4446 | lr 8.87e-04 | grad 0.2625 +2026-04-10 09:51:55 - INFO - parrotllm.training - step 58280 | epoch 0 | loss 3.4678 | lr 8.87e-04 | grad 0.2358 +2026-04-10 09:51:58 - INFO - parrotllm.training - step 58290 | epoch 0 | loss 3.3502 | lr 8.86e-04 | grad 0.2571 +2026-04-10 09:52:01 - INFO - parrotllm.training - step 58300 | epoch 0 | loss 3.3366 | lr 8.86e-04 | grad 0.2597 +2026-04-10 09:52:04 - INFO - parrotllm.training - step 58310 | epoch 0 | loss 3.3892 | lr 8.86e-04 | grad 0.2292 +2026-04-10 09:52:07 - INFO - parrotllm.training - step 58320 | epoch 0 | loss 3.4948 | lr 8.86e-04 | grad 0.2674 +2026-04-10 09:52:10 - INFO - parrotllm.training - step 58330 | epoch 0 | loss 3.3387 | lr 8.86e-04 | grad 0.2293 +2026-04-10 09:52:13 - INFO - parrotllm.training - step 58340 | epoch 0 | loss 3.3914 | lr 8.86e-04 | grad 0.2403 +2026-04-10 09:52:16 - INFO - parrotllm.training - step 58350 | epoch 0 | loss 3.3366 | lr 8.85e-04 | grad 0.2360 +2026-04-10 09:52:19 - INFO - parrotllm.training - step 58360 | epoch 0 | loss 3.4334 | lr 8.85e-04 | grad 0.2599 +2026-04-10 09:52:22 - INFO - parrotllm.training - step 58370 | epoch 0 | loss 3.4034 | lr 8.85e-04 | grad 0.2275 +2026-04-10 09:52:26 - INFO - parrotllm.training - step 58380 | epoch 0 | loss 3.4531 | lr 8.85e-04 | grad 0.2613 +2026-04-10 09:52:29 - INFO - parrotllm.training - step 58390 | epoch 0 | loss 3.3884 | lr 8.85e-04 | grad 0.2396 +2026-04-10 09:52:32 - INFO - parrotllm.training - step 58400 | epoch 0 | loss 3.3982 | lr 8.84e-04 | grad 0.2297 +2026-04-10 09:52:35 - INFO - parrotllm.training - step 58410 | epoch 0 | loss 3.4067 | lr 8.84e-04 | grad 0.2331 +2026-04-10 09:52:38 - INFO - parrotllm.training - step 58420 | epoch 0 | loss 3.4862 | lr 8.84e-04 | grad 0.2439 +2026-04-10 09:52:41 - INFO - parrotllm.training - step 58430 | epoch 0 | loss 3.4410 | lr 8.84e-04 | grad 0.2671 +2026-04-10 09:52:44 - INFO - parrotllm.training - step 58440 | epoch 0 | loss 3.3419 | lr 8.84e-04 | grad 0.2612 +2026-04-10 09:52:47 - INFO - parrotllm.training - step 58450 | epoch 0 | loss 3.4097 | lr 8.84e-04 | grad 0.2408 +2026-04-10 09:52:50 - INFO - parrotllm.training - step 58460 | epoch 0 | loss 3.4155 | lr 8.83e-04 | grad 0.2419 +2026-04-10 09:52:53 - INFO - parrotllm.training - step 58470 | epoch 0 | loss 3.4222 | lr 8.83e-04 | grad 0.2298 +2026-04-10 09:52:56 - INFO - parrotllm.training - step 58480 | epoch 0 | loss 3.4544 | lr 8.83e-04 | grad 0.2663 +2026-04-10 09:52:59 - INFO - parrotllm.training - step 58490 | epoch 0 | loss 3.4093 | lr 8.83e-04 | grad 0.2267 +2026-04-10 09:53:02 - INFO - parrotllm.training - step 58500 | epoch 0 | loss 3.2779 | lr 8.83e-04 | grad 0.2533 +2026-04-10 09:53:02 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:53:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:53:05 - INFO - parrotllm.training - Train: loss=3.2779, ppl=26.52 +2026-04-10 09:53:05 - INFO - parrotllm.training - Val: loss=3.3425, ppl=28.29 +2026-04-10 09:53:05 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 09:53:06 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3425_epoch_0000_step_0058500.pt +2026-04-10 09:53:08 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:53:11 - INFO - parrotllm.training - step 58510 | epoch 0 | loss 3.3611 | lr 8.83e-04 | grad 0.2436 +2026-04-10 09:53:14 - INFO - parrotllm.training - step 58520 | epoch 0 | loss 3.4338 | lr 8.82e-04 | grad 0.2454 +2026-04-10 09:53:17 - INFO - parrotllm.training - step 58530 | epoch 0 | loss 3.3976 | lr 8.82e-04 | grad 0.2488 +2026-04-10 09:53:18 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 58555/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 09:53:20 - INFO - parrotllm.training - step 58540 | epoch 0 | loss 3.3813 | lr 8.82e-04 | grad 0.2412 +2026-04-10 09:53:23 - INFO - parrotllm.training - step 58550 | epoch 0 | loss 3.4264 | lr 8.82e-04 | grad 0.2240 +2026-04-10 09:53:26 - INFO - parrotllm.training - step 58560 | epoch 0 | loss 3.3525 | lr 8.82e-04 | grad 0.2555 +2026-04-10 09:53:30 - INFO - parrotllm.training - step 58570 | epoch 0 | loss 3.3615 | lr 8.81e-04 | grad 0.2512 +2026-04-10 09:53:33 - INFO - parrotllm.training - step 58580 | epoch 0 | loss 3.3524 | lr 8.81e-04 | grad 0.2390 +2026-04-10 09:53:36 - INFO - parrotllm.training - step 58590 | epoch 0 | loss 3.4359 | lr 8.81e-04 | grad 0.2297 +2026-04-10 09:53:39 - INFO - parrotllm.training - step 58600 | epoch 0 | loss 3.3802 | lr 8.81e-04 | grad 0.2466 +2026-04-10 09:53:42 - INFO - parrotllm.training - step 58610 | epoch 0 | loss 3.2883 | lr 8.81e-04 | grad 0.2497 +2026-04-10 09:53:45 - INFO - parrotllm.training - step 58620 | epoch 0 | loss 3.4196 | lr 8.81e-04 | grad 0.2235 +2026-04-10 09:53:48 - INFO - parrotllm.training - step 58630 | epoch 0 | loss 3.3699 | lr 8.80e-04 | grad 0.2576 +2026-04-10 09:53:51 - INFO - parrotllm.training - step 58640 | epoch 0 | loss 3.4372 | lr 8.80e-04 | grad 0.2612 +2026-04-10 09:53:54 - INFO - parrotllm.training - step 58650 | epoch 0 | loss 3.4043 | lr 8.80e-04 | grad 0.2508 +2026-04-10 09:53:57 - INFO - parrotllm.training - step 58660 | epoch 0 | loss 3.3583 | lr 8.80e-04 | grad 0.2098 +2026-04-10 09:54:00 - INFO - parrotllm.training - step 58670 | epoch 0 | loss 3.4457 | lr 8.80e-04 | grad 0.2360 +2026-04-10 09:54:03 - INFO - parrotllm.training - step 58680 | epoch 0 | loss 3.4046 | lr 8.80e-04 | grad 0.2571 +2026-04-10 09:54:06 - INFO - parrotllm.training - step 58690 | epoch 0 | loss 3.5208 | lr 8.79e-04 | grad 0.2760 +2026-04-10 09:54:09 - INFO - parrotllm.training - step 58700 | epoch 0 | loss 3.3766 | lr 8.79e-04 | grad 0.2750 +2026-04-10 09:54:12 - INFO - parrotllm.training - step 58710 | epoch 0 | loss 3.4747 | lr 8.79e-04 | grad 0.2397 +2026-04-10 09:54:15 - INFO - parrotllm.training - step 58720 | epoch 0 | loss 3.2158 | lr 8.79e-04 | grad 0.2335 +2026-04-10 09:54:18 - INFO - parrotllm.training - step 58730 | epoch 0 | loss 3.5034 | lr 8.79e-04 | grad 0.2559 +2026-04-10 09:54:21 - INFO - parrotllm.training - step 58740 | epoch 0 | loss 3.3804 | lr 8.78e-04 | grad 0.2306 +2026-04-10 09:54:24 - INFO - parrotllm.training - step 58750 | epoch 0 | loss 3.3877 | lr 8.78e-04 | grad 0.2768 +2026-04-10 09:54:27 - INFO - parrotllm.training - step 58760 | epoch 0 | loss 3.3669 | lr 8.78e-04 | grad 0.2410 +2026-04-10 09:54:30 - INFO - parrotllm.training - step 58770 | epoch 0 | loss 3.3018 | lr 8.78e-04 | grad 0.2870 +2026-04-10 09:54:34 - INFO - parrotllm.training - step 58780 | epoch 0 | loss 3.4471 | lr 8.78e-04 | grad 0.2347 +2026-04-10 09:54:37 - INFO - parrotllm.training - step 58790 | epoch 0 | loss 3.3722 | lr 8.78e-04 | grad 0.2347 +2026-04-10 09:54:40 - INFO - parrotllm.training - step 58800 | epoch 0 | loss 3.4894 | lr 8.77e-04 | grad 0.2604 +2026-04-10 09:54:43 - INFO - parrotllm.training - step 58810 | epoch 0 | loss 3.2952 | lr 8.77e-04 | grad 0.2709 +2026-04-10 09:54:46 - INFO - parrotllm.training - step 58820 | epoch 0 | loss 3.4683 | lr 8.77e-04 | grad 0.2740 +2026-04-10 09:54:49 - INFO - parrotllm.training - step 58830 | epoch 0 | loss 3.5081 | lr 8.77e-04 | grad 0.2786 +2026-04-10 09:54:52 - INFO - parrotllm.training - step 58840 | epoch 0 | loss 3.2837 | lr 8.77e-04 | grad 0.2748 +2026-04-10 09:54:55 - INFO - parrotllm.training - step 58850 | epoch 0 | loss 3.4459 | lr 8.77e-04 | grad 0.2901 +2026-04-10 09:54:58 - INFO - parrotllm.training - step 58860 | epoch 0 | loss 3.3513 | lr 8.76e-04 | grad 0.2220 +2026-04-10 09:55:01 - INFO - parrotllm.training - step 58870 | epoch 0 | loss 3.3995 | lr 8.76e-04 | grad 0.2692 +2026-04-10 09:55:04 - INFO - parrotllm.training - step 58880 | epoch 0 | loss 3.4181 | lr 8.76e-04 | grad 0.2296 +2026-04-10 09:55:07 - INFO - parrotllm.training - step 58890 | epoch 0 | loss 3.3861 | lr 8.76e-04 | grad 0.2372 +2026-04-10 09:55:10 - INFO - parrotllm.training - step 58900 | epoch 0 | loss 3.2931 | lr 8.76e-04 | grad 0.2347 +2026-04-10 09:55:13 - INFO - parrotllm.training - step 58910 | epoch 0 | loss 3.4354 | lr 8.76e-04 | grad 0.2462 +2026-04-10 09:55:16 - INFO - parrotllm.training - step 58920 | epoch 0 | loss 3.3579 | lr 8.75e-04 | grad 0.2686 +2026-04-10 09:55:19 - INFO - parrotllm.training - step 58930 | epoch 0 | loss 3.4548 | lr 8.75e-04 | grad 0.2317 +2026-04-10 09:55:22 - INFO - parrotllm.training - step 58940 | epoch 0 | loss 3.2536 | lr 8.75e-04 | grad 0.2335 +2026-04-10 09:55:25 - INFO - parrotllm.training - step 58950 | epoch 0 | loss 3.5367 | lr 8.75e-04 | grad 0.2745 +2026-04-10 09:55:28 - INFO - parrotllm.training - step 58960 | epoch 0 | loss 3.3593 | lr 8.75e-04 | grad 0.2254 +2026-04-10 09:55:31 - INFO - parrotllm.training - step 58970 | epoch 0 | loss 3.4183 | lr 8.74e-04 | grad 0.2576 +2026-04-10 09:55:35 - INFO - parrotllm.training - step 58980 | epoch 0 | loss 3.3126 | lr 8.74e-04 | grad 0.2653 +2026-04-10 09:55:38 - INFO - parrotllm.training - step 58990 | epoch 0 | loss 3.4397 | lr 8.74e-04 | grad 0.2444 +2026-04-10 09:55:41 - INFO - parrotllm.training - step 59000 | epoch 0 | loss 3.3801 | lr 8.74e-04 | grad 0.2361 +2026-04-10 09:55:41 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:55:41 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:55:44 - INFO - parrotllm.training - Train: loss=3.3801, ppl=29.38 +2026-04-10 09:55:44 - INFO - parrotllm.training - Val: loss=3.3443, ppl=28.34 +2026-04-10 09:55:45 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3443_epoch_0000_step_0059000.pt +2026-04-10 09:55:46 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3425, min_delta=0.001000). +2026-04-10 09:55:46 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:55:49 - INFO - parrotllm.training - step 59010 | epoch 0 | loss 3.4703 | lr 8.74e-04 | grad 0.2505 +2026-04-10 09:55:52 - INFO - parrotllm.training - step 59020 | epoch 0 | loss 3.4385 | lr 8.74e-04 | grad 0.2484 +2026-04-10 09:55:55 - INFO - parrotllm.training - step 59030 | epoch 0 | loss 3.3604 | lr 8.73e-04 | grad 0.2644 +2026-04-10 09:55:59 - INFO - parrotllm.training - step 59040 | epoch 0 | loss 3.4347 | lr 8.73e-04 | grad 0.2255 +2026-04-10 09:56:02 - INFO - parrotllm.training - step 59050 | epoch 0 | loss 3.4311 | lr 8.73e-04 | grad 0.2384 +2026-04-10 09:56:05 - INFO - parrotllm.training - step 59060 | epoch 0 | loss 3.4230 | lr 8.73e-04 | grad 0.2278 +2026-04-10 09:56:08 - INFO - parrotllm.training - step 59070 | epoch 0 | loss 3.4872 | lr 8.73e-04 | grad 0.2510 +2026-04-10 09:56:11 - INFO - parrotllm.training - step 59080 | epoch 0 | loss 3.3245 | lr 8.73e-04 | grad 0.2391 +2026-04-10 09:56:14 - INFO - parrotllm.training - step 59090 | epoch 0 | loss 3.3241 | lr 8.72e-04 | grad 0.2443 +2026-04-10 09:56:17 - INFO - parrotllm.training - step 59100 | epoch 0 | loss 3.4300 | lr 8.72e-04 | grad 0.2740 +2026-04-10 09:56:20 - INFO - parrotllm.training - step 59110 | epoch 0 | loss 3.3048 | lr 8.72e-04 | grad 0.2407 +2026-04-10 09:56:23 - INFO - parrotllm.training - step 59120 | epoch 0 | loss 3.3927 | lr 8.72e-04 | grad 0.2455 +2026-04-10 09:56:26 - INFO - parrotllm.training - step 59130 | epoch 0 | loss 3.4351 | lr 8.72e-04 | grad 0.2728 +2026-04-10 09:56:29 - INFO - parrotllm.training - step 59140 | epoch 0 | loss 3.4479 | lr 8.71e-04 | grad 0.2378 +2026-04-10 09:56:32 - INFO - parrotllm.training - step 59150 | epoch 0 | loss 3.5074 | lr 8.71e-04 | grad 0.2422 +2026-04-10 09:56:35 - INFO - parrotllm.training - step 59160 | epoch 0 | loss 3.3842 | lr 8.71e-04 | grad 0.2301 +2026-04-10 09:56:38 - INFO - parrotllm.training - step 59170 | epoch 0 | loss 3.3792 | lr 8.71e-04 | grad 0.2414 +2026-04-10 09:56:41 - INFO - parrotllm.training - step 59180 | epoch 0 | loss 3.3762 | lr 8.71e-04 | grad 0.2256 +2026-04-10 09:56:44 - INFO - parrotllm.training - step 59190 | epoch 0 | loss 3.4164 | lr 8.71e-04 | grad 0.2184 +2026-04-10 09:56:47 - INFO - parrotllm.training - step 59200 | epoch 0 | loss 3.4170 | lr 8.70e-04 | grad 0.2465 +2026-04-10 09:56:50 - INFO - parrotllm.training - step 59210 | epoch 0 | loss 3.4264 | lr 8.70e-04 | grad 0.2249 +2026-04-10 09:56:53 - INFO - parrotllm.training - step 59220 | epoch 0 | loss 3.4464 | lr 8.70e-04 | grad 0.2469 +2026-04-10 09:56:56 - INFO - parrotllm.training - step 59230 | epoch 0 | loss 3.4088 | lr 8.70e-04 | grad 0.2706 +2026-04-10 09:56:59 - INFO - parrotllm.training - step 59240 | epoch 0 | loss 3.4199 | lr 8.70e-04 | grad 0.2416 +2026-04-10 09:57:03 - INFO - parrotllm.training - step 59250 | epoch 0 | loss 3.3328 | lr 8.70e-04 | grad 0.2319 +2026-04-10 09:57:06 - INFO - parrotllm.training - step 59260 | epoch 0 | loss 3.3996 | lr 8.69e-04 | grad 0.2340 +2026-04-10 09:57:09 - INFO - parrotllm.training - step 59270 | epoch 0 | loss 3.3773 | lr 8.69e-04 | grad 0.2582 +2026-04-10 09:57:12 - INFO - parrotllm.training - step 59280 | epoch 0 | loss 3.4634 | lr 8.69e-04 | grad 0.2449 +2026-04-10 09:57:15 - INFO - parrotllm.training - step 59290 | epoch 0 | loss 3.3347 | lr 8.69e-04 | grad 0.2307 +2026-04-10 09:57:18 - INFO - parrotllm.training - step 59300 | epoch 0 | loss 3.3994 | lr 8.69e-04 | grad 0.2235 +2026-04-10 09:57:21 - INFO - parrotllm.training - step 59310 | epoch 0 | loss 3.4720 | lr 8.68e-04 | grad 0.2274 +2026-04-10 09:57:24 - INFO - parrotllm.training - step 59320 | epoch 0 | loss 3.3711 | lr 8.68e-04 | grad 0.2732 +2026-04-10 09:57:27 - INFO - parrotllm.training - step 59330 | epoch 0 | loss 3.4365 | lr 8.68e-04 | grad 0.2873 +2026-04-10 09:57:30 - INFO - parrotllm.training - step 59340 | epoch 0 | loss 3.3876 | lr 8.68e-04 | grad 0.2334 +2026-04-10 09:57:33 - INFO - parrotllm.training - step 59350 | epoch 0 | loss 3.4018 | lr 8.68e-04 | grad 0.2516 +2026-04-10 09:57:36 - INFO - parrotllm.training - step 59360 | epoch 0 | loss 3.4380 | lr 8.68e-04 | grad 0.2302 +2026-04-10 09:57:39 - INFO - parrotllm.training - step 59370 | epoch 0 | loss 3.3781 | lr 8.67e-04 | grad 0.2601 +2026-04-10 09:57:42 - INFO - parrotllm.training - step 59380 | epoch 0 | loss 3.3557 | lr 8.67e-04 | grad 0.2471 +2026-04-10 09:57:45 - INFO - parrotllm.training - step 59390 | epoch 0 | loss 3.3846 | lr 8.67e-04 | grad 0.2420 +2026-04-10 09:57:48 - INFO - parrotllm.training - step 59400 | epoch 0 | loss 3.3855 | lr 8.67e-04 | grad 0.2426 +2026-04-10 09:57:51 - INFO - parrotllm.training - step 59410 | epoch 0 | loss 3.3753 | lr 8.67e-04 | grad 0.2321 +2026-04-10 09:57:54 - INFO - parrotllm.training - step 59420 | epoch 0 | loss 3.4106 | lr 8.67e-04 | grad 0.2231 +2026-04-10 09:57:57 - INFO - parrotllm.training - step 59430 | epoch 0 | loss 3.3559 | lr 8.66e-04 | grad 0.2521 +2026-04-10 09:58:00 - INFO - parrotllm.training - step 59440 | epoch 0 | loss 3.3109 | lr 8.66e-04 | grad 0.2679 +2026-04-10 09:58:04 - INFO - parrotllm.training - step 59450 | epoch 0 | loss 3.4848 | lr 8.66e-04 | grad 0.2539 +2026-04-10 09:58:07 - INFO - parrotllm.training - step 59460 | epoch 0 | loss 3.3249 | lr 8.66e-04 | grad 0.2535 +2026-04-10 09:58:10 - INFO - parrotllm.training - step 59470 | epoch 0 | loss 3.4169 | lr 8.66e-04 | grad 0.2269 +2026-04-10 09:58:13 - INFO - parrotllm.training - step 59480 | epoch 0 | loss 3.4273 | lr 8.66e-04 | grad 0.2482 +2026-04-10 09:58:16 - INFO - parrotllm.training - step 59490 | epoch 0 | loss 3.4058 | lr 8.65e-04 | grad 0.2652 +2026-04-10 09:58:19 - INFO - parrotllm.training - step 59500 | epoch 0 | loss 3.4372 | lr 8.65e-04 | grad 0.2294 +2026-04-10 09:58:19 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 09:58:19 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:58:22 - INFO - parrotllm.training - Train: loss=3.4372, ppl=31.10 +2026-04-10 09:58:22 - INFO - parrotllm.training - Val: loss=3.3429, ppl=28.30 +2026-04-10 09:58:23 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3429_epoch_0000_step_0059500.pt +2026-04-10 09:58:24 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3425, min_delta=0.001000). +2026-04-10 09:58:24 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 09:58:27 - INFO - parrotllm.training - step 59510 | epoch 0 | loss 3.3882 | lr 8.65e-04 | grad 0.2605 +2026-04-10 09:58:31 - INFO - parrotllm.training - step 59520 | epoch 0 | loss 3.3052 | lr 8.65e-04 | grad 0.2571 +2026-04-10 09:58:34 - INFO - parrotllm.training - step 59530 | epoch 0 | loss 3.3593 | lr 8.65e-04 | grad 0.2386 +2026-04-10 09:58:37 - INFO - parrotllm.training - step 59540 | epoch 0 | loss 3.3928 | lr 8.64e-04 | grad 0.2728 +2026-04-10 09:58:40 - INFO - parrotllm.training - step 59550 | epoch 0 | loss 3.4972 | lr 8.64e-04 | grad 0.2527 +2026-04-10 09:58:43 - INFO - parrotllm.training - step 59560 | epoch 0 | loss 3.3931 | lr 8.64e-04 | grad 0.2545 +2026-04-10 09:58:46 - INFO - parrotllm.training - step 59570 | epoch 0 | loss 3.2943 | lr 8.64e-04 | grad 0.2643 +2026-04-10 09:58:49 - INFO - parrotllm.training - step 59580 | epoch 0 | loss 3.4664 | lr 8.64e-04 | grad 0.2546 +2026-04-10 09:58:52 - INFO - parrotllm.training - step 59590 | epoch 0 | loss 3.4726 | lr 8.64e-04 | grad 0.2432 +2026-04-10 09:58:55 - INFO - parrotllm.training - step 59600 | epoch 0 | loss 3.2890 | lr 8.63e-04 | grad 0.2499 +2026-04-10 09:58:58 - INFO - parrotllm.training - step 59610 | epoch 0 | loss 3.2674 | lr 8.63e-04 | grad 0.2360 +2026-04-10 09:59:01 - INFO - parrotllm.training - step 59620 | epoch 0 | loss 3.4244 | lr 8.63e-04 | grad 0.2507 +2026-04-10 09:59:04 - INFO - parrotllm.training - step 59630 | epoch 0 | loss 3.4268 | lr 8.63e-04 | grad 0.2404 +2026-04-10 09:59:07 - INFO - parrotllm.training - step 59640 | epoch 0 | loss 3.5385 | lr 8.63e-04 | grad 0.3082 +2026-04-10 09:59:10 - INFO - parrotllm.training - step 59650 | epoch 0 | loss 3.4615 | lr 8.63e-04 | grad 0.2493 +2026-04-10 09:59:13 - INFO - parrotllm.training - step 59660 | epoch 0 | loss 3.4677 | lr 8.62e-04 | grad 0.2205 +2026-04-10 09:59:16 - INFO - parrotllm.training - step 59670 | epoch 0 | loss 3.4315 | lr 8.62e-04 | grad 0.2922 +2026-04-10 09:59:19 - INFO - parrotllm.training - step 59680 | epoch 0 | loss 3.3089 | lr 8.62e-04 | grad 0.2353 +2026-04-10 09:59:22 - INFO - parrotllm.training - step 59690 | epoch 0 | loss 3.3775 | lr 8.62e-04 | grad 0.2411 +2026-04-10 09:59:25 - INFO - parrotllm.training - step 59700 | epoch 0 | loss 3.4357 | lr 8.62e-04 | grad 0.2536 +2026-04-10 09:59:28 - INFO - parrotllm.training - step 59710 | epoch 0 | loss 3.3716 | lr 8.61e-04 | grad 0.2267 +2026-04-10 09:59:31 - INFO - parrotllm.training - step 59720 | epoch 0 | loss 3.4229 | lr 8.61e-04 | grad 0.2531 +2026-04-10 09:59:35 - INFO - parrotllm.training - step 59730 | epoch 0 | loss 3.4615 | lr 8.61e-04 | grad 0.2530 +2026-04-10 09:59:38 - INFO - parrotllm.training - step 59740 | epoch 0 | loss 3.4137 | lr 8.61e-04 | grad 0.2459 +2026-04-10 09:59:41 - INFO - parrotllm.training - step 59750 | epoch 0 | loss 3.4370 | lr 8.61e-04 | grad 0.2465 +2026-04-10 09:59:44 - INFO - parrotllm.training - step 59760 | epoch 0 | loss 3.3616 | lr 8.61e-04 | grad 0.2525 +2026-04-10 09:59:47 - INFO - parrotllm.training - step 59770 | epoch 0 | loss 3.3699 | lr 8.60e-04 | grad 0.2704 +2026-04-10 09:59:50 - INFO - parrotllm.training - step 59780 | epoch 0 | loss 3.3239 | lr 8.60e-04 | grad 0.2273 +2026-04-10 09:59:53 - INFO - parrotllm.training - step 59790 | epoch 0 | loss 3.3837 | lr 8.60e-04 | grad 0.2603 +2026-04-10 09:59:56 - INFO - parrotllm.training - step 59800 | epoch 0 | loss 3.4301 | lr 8.60e-04 | grad 0.2341 +2026-04-10 09:59:59 - INFO - parrotllm.training - step 59810 | epoch 0 | loss 3.3916 | lr 8.60e-04 | grad 0.2417 +2026-04-10 10:00:02 - INFO - parrotllm.training - step 59820 | epoch 0 | loss 3.3930 | lr 8.60e-04 | grad 0.2330 +2026-04-10 10:00:05 - INFO - parrotllm.training - step 59830 | epoch 0 | loss 3.4643 | lr 8.59e-04 | grad 0.2418 +2026-04-10 10:00:08 - INFO - parrotllm.training - step 59840 | epoch 0 | loss 3.3952 | lr 8.59e-04 | grad 0.2314 +2026-04-10 10:00:11 - INFO - parrotllm.training - step 59850 | epoch 0 | loss 3.3389 | lr 8.59e-04 | grad 0.2521 +2026-04-10 10:00:14 - INFO - parrotllm.training - step 59860 | epoch 0 | loss 3.3808 | lr 8.59e-04 | grad 0.2295 +2026-04-10 10:00:17 - INFO - parrotllm.training - step 59870 | epoch 0 | loss 3.4281 | lr 8.59e-04 | grad 0.2509 +2026-04-10 10:00:20 - INFO - parrotllm.training - step 59880 | epoch 0 | loss 3.3129 | lr 8.58e-04 | grad 0.2568 +2026-04-10 10:00:23 - INFO - parrotllm.training - step 59890 | epoch 0 | loss 3.4014 | lr 8.58e-04 | grad 0.2733 +2026-04-10 10:00:26 - INFO - parrotllm.training - step 59900 | epoch 0 | loss 3.5286 | lr 8.58e-04 | grad 0.2215 +2026-04-10 10:00:29 - INFO - parrotllm.training - step 59910 | epoch 0 | loss 3.2706 | lr 8.58e-04 | grad 0.2687 +2026-04-10 10:00:33 - INFO - parrotllm.training - step 59920 | epoch 0 | loss 3.4249 | lr 8.58e-04 | grad 0.2619 +2026-04-10 10:00:36 - INFO - parrotllm.training - step 59930 | epoch 0 | loss 3.4016 | lr 8.58e-04 | grad 0.2259 +2026-04-10 10:00:39 - INFO - parrotllm.training - step 59940 | epoch 0 | loss 3.3168 | lr 8.57e-04 | grad 0.2524 +2026-04-10 10:00:42 - INFO - parrotllm.training - step 59950 | epoch 0 | loss 3.4205 | lr 8.57e-04 | grad 0.2348 +2026-04-10 10:00:45 - INFO - parrotllm.training - step 59960 | epoch 0 | loss 3.3414 | lr 8.57e-04 | grad 0.2335 +2026-04-10 10:00:48 - INFO - parrotllm.training - step 59970 | epoch 0 | loss 3.4880 | lr 8.57e-04 | grad 0.2305 +2026-04-10 10:00:51 - INFO - parrotllm.training - step 59980 | epoch 0 | loss 3.2702 | lr 8.57e-04 | grad 0.3031 +2026-04-10 10:00:54 - INFO - parrotllm.training - step 59990 | epoch 0 | loss 3.4907 | lr 8.57e-04 | grad 0.2339 +2026-04-10 10:00:57 - INFO - parrotllm.training - step 60000 | epoch 0 | loss 3.3919 | lr 8.56e-04 | grad 0.2376 +2026-04-10 10:00:57 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:00:57 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:01:00 - INFO - parrotllm.training - Train: loss=3.3919, ppl=29.72 +2026-04-10 10:01:00 - INFO - parrotllm.training - Val: loss=3.3414, ppl=28.26 +2026-04-10 10:01:00 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:01:01 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3414_epoch_0000_step_0060000.pt +2026-04-10 10:01:03 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:01:04 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0060000.pt +2026-04-10 10:01:08 - INFO - parrotllm.training - step 60010 | epoch 0 | loss 3.3493 | lr 8.56e-04 | grad 0.2690 +2026-04-10 10:01:12 - INFO - parrotllm.training - step 60020 | epoch 0 | loss 3.3701 | lr 8.56e-04 | grad 0.2516 +2026-04-10 10:01:15 - INFO - parrotllm.training - step 60030 | epoch 0 | loss 3.4039 | lr 8.56e-04 | grad 0.2275 +2026-04-10 10:01:18 - INFO - parrotllm.training - step 60040 | epoch 0 | loss 3.3188 | lr 8.56e-04 | grad 0.2450 +2026-04-10 10:01:21 - INFO - parrotllm.training - step 60050 | epoch 0 | loss 3.4027 | lr 8.56e-04 | grad 0.2428 +2026-04-10 10:01:24 - INFO - parrotllm.training - step 60060 | epoch 0 | loss 3.4720 | lr 8.55e-04 | grad 0.2548 +2026-04-10 10:01:27 - INFO - parrotllm.training - step 60070 | epoch 0 | loss 3.3684 | lr 8.55e-04 | grad 0.2374 +2026-04-10 10:01:30 - INFO - parrotllm.training - step 60080 | epoch 0 | loss 3.3669 | lr 8.55e-04 | grad 0.2710 +2026-04-10 10:01:33 - INFO - parrotllm.training - step 60090 | epoch 0 | loss 3.3790 | lr 8.55e-04 | grad 0.2465 +2026-04-10 10:01:36 - INFO - parrotllm.training - step 60100 | epoch 0 | loss 3.4346 | lr 8.55e-04 | grad 0.2702 +2026-04-10 10:01:39 - INFO - parrotllm.training - step 60110 | epoch 0 | loss 3.4426 | lr 8.54e-04 | grad 0.2497 +2026-04-10 10:01:42 - INFO - parrotllm.training - step 60120 | epoch 0 | loss 3.3150 | lr 8.54e-04 | grad 0.2624 +2026-04-10 10:01:45 - INFO - parrotllm.training - step 60130 | epoch 0 | loss 3.2995 | lr 8.54e-04 | grad 0.2995 +2026-04-10 10:01:48 - INFO - parrotllm.training - step 60140 | epoch 0 | loss 3.3609 | lr 8.54e-04 | grad 0.2519 +2026-04-10 10:01:51 - INFO - parrotllm.training - step 60150 | epoch 0 | loss 3.4681 | lr 8.54e-04 | grad 0.2410 +2026-04-10 10:01:54 - INFO - parrotllm.training - step 60160 | epoch 0 | loss 3.3662 | lr 8.54e-04 | grad 0.2296 +2026-04-10 10:01:57 - INFO - parrotllm.training - step 60170 | epoch 0 | loss 3.4236 | lr 8.53e-04 | grad 0.2501 +2026-04-10 10:02:00 - INFO - parrotllm.training - step 60180 | epoch 0 | loss 3.3764 | lr 8.53e-04 | grad 0.2620 +2026-04-10 10:02:03 - INFO - parrotllm.training - step 60190 | epoch 0 | loss 3.4567 | lr 8.53e-04 | grad 0.2429 +2026-04-10 10:02:07 - INFO - parrotllm.training - step 60200 | epoch 0 | loss 3.3649 | lr 8.53e-04 | grad 0.2384 +2026-04-10 10:02:10 - INFO - parrotllm.training - step 60210 | epoch 0 | loss 3.3515 | lr 8.53e-04 | grad 0.2572 +2026-04-10 10:02:13 - INFO - parrotllm.training - step 60220 | epoch 0 | loss 3.3034 | lr 8.53e-04 | grad 0.2326 +2026-04-10 10:02:16 - INFO - parrotllm.training - step 60230 | epoch 0 | loss 3.3916 | lr 8.52e-04 | grad 0.2474 +2026-04-10 10:02:19 - INFO - parrotllm.training - step 60240 | epoch 0 | loss 3.4124 | lr 8.52e-04 | grad 0.2652 +2026-04-10 10:02:22 - INFO - parrotllm.training - step 60250 | epoch 0 | loss 3.3540 | lr 8.52e-04 | grad 0.2268 +2026-04-10 10:02:25 - INFO - parrotllm.training - step 60260 | epoch 0 | loss 3.4293 | lr 8.52e-04 | grad 0.2490 +2026-04-10 10:02:28 - INFO - parrotllm.training - step 60270 | epoch 0 | loss 3.4651 | lr 8.52e-04 | grad 0.2575 +2026-04-10 10:02:31 - INFO - parrotllm.training - step 60280 | epoch 0 | loss 3.4446 | lr 8.51e-04 | grad 0.2276 +2026-04-10 10:02:34 - INFO - parrotllm.training - step 60290 | epoch 0 | loss 3.4600 | lr 8.51e-04 | grad 0.2312 +2026-04-10 10:02:37 - INFO - parrotllm.training - step 60300 | epoch 0 | loss 3.4964 | lr 8.51e-04 | grad 0.2351 +2026-04-10 10:02:40 - INFO - parrotllm.training - step 60310 | epoch 0 | loss 3.3937 | lr 8.51e-04 | grad 0.2306 +2026-04-10 10:02:43 - INFO - parrotllm.training - step 60320 | epoch 0 | loss 3.3391 | lr 8.51e-04 | grad 0.2577 +2026-04-10 10:02:46 - INFO - parrotllm.training - step 60330 | epoch 0 | loss 3.3499 | lr 8.51e-04 | grad 0.2366 +2026-04-10 10:02:49 - INFO - parrotllm.training - step 60340 | epoch 0 | loss 3.4014 | lr 8.50e-04 | grad 0.2432 +2026-04-10 10:02:52 - INFO - parrotllm.training - step 60350 | epoch 0 | loss 3.4054 | lr 8.50e-04 | grad 0.2316 +2026-04-10 10:02:55 - INFO - parrotllm.training - step 60360 | epoch 0 | loss 3.4193 | lr 8.50e-04 | grad 0.2510 +2026-04-10 10:02:58 - INFO - parrotllm.training - step 60370 | epoch 0 | loss 3.3832 | lr 8.50e-04 | grad 0.2403 +2026-04-10 10:03:01 - INFO - parrotllm.training - step 60380 | epoch 0 | loss 3.3535 | lr 8.50e-04 | grad 0.2435 +2026-04-10 10:03:04 - INFO - parrotllm.training - step 60390 | epoch 0 | loss 3.3650 | lr 8.50e-04 | grad 0.2624 +2026-04-10 10:03:07 - INFO - parrotllm.training - step 60400 | epoch 0 | loss 3.4902 | lr 8.49e-04 | grad 0.2589 +2026-04-10 10:03:11 - INFO - parrotllm.training - step 60410 | epoch 0 | loss 3.2901 | lr 8.49e-04 | grad 0.2593 +2026-04-10 10:03:14 - INFO - parrotllm.training - step 60420 | epoch 0 | loss 3.3959 | lr 8.49e-04 | grad 0.2582 +2026-04-10 10:03:17 - INFO - parrotllm.training - step 60430 | epoch 0 | loss 3.3082 | lr 8.49e-04 | grad 0.2461 +2026-04-10 10:03:20 - INFO - parrotllm.training - step 60440 | epoch 0 | loss 3.3669 | lr 8.49e-04 | grad 0.2285 +2026-04-10 10:03:23 - INFO - parrotllm.training - step 60450 | epoch 0 | loss 3.4028 | lr 8.48e-04 | grad 0.2318 +2026-04-10 10:03:26 - INFO - parrotllm.training - step 60460 | epoch 0 | loss 3.4082 | lr 8.48e-04 | grad 0.2754 +2026-04-10 10:03:29 - INFO - parrotllm.training - step 60470 | epoch 0 | loss 3.4456 | lr 8.48e-04 | grad 0.2377 +2026-04-10 10:03:32 - INFO - parrotllm.training - step 60480 | epoch 0 | loss 3.3272 | lr 8.48e-04 | grad 0.2512 +2026-04-10 10:03:35 - INFO - parrotllm.training - step 60490 | epoch 0 | loss 3.4258 | lr 8.48e-04 | grad 0.2394 +2026-04-10 10:03:38 - INFO - parrotllm.training - step 60500 | epoch 0 | loss 3.4979 | lr 8.48e-04 | grad 0.2368 +2026-04-10 10:03:38 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:03:38 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:03:41 - INFO - parrotllm.training - Train: loss=3.4979, ppl=33.04 +2026-04-10 10:03:41 - INFO - parrotllm.training - Val: loss=3.3403, ppl=28.23 +2026-04-10 10:03:41 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:03:42 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3403_epoch_0000_step_0060500.pt +2026-04-10 10:03:44 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:03:47 - INFO - parrotllm.training - step 60510 | epoch 0 | loss 3.4732 | lr 8.47e-04 | grad 0.2421 +2026-04-10 10:03:50 - INFO - parrotllm.training - step 60520 | epoch 0 | loss 3.3686 | lr 8.47e-04 | grad 0.2382 +2026-04-10 10:03:53 - INFO - parrotllm.training - step 60530 | epoch 0 | loss 3.4576 | lr 8.47e-04 | grad 0.2322 +2026-04-10 10:03:56 - INFO - parrotllm.training - step 60540 | epoch 0 | loss 3.4324 | lr 8.47e-04 | grad 0.2500 +2026-04-10 10:03:59 - INFO - parrotllm.training - step 60550 | epoch 0 | loss 3.4755 | lr 8.47e-04 | grad 0.2275 +2026-04-10 10:04:02 - INFO - parrotllm.training - step 60560 | epoch 0 | loss 3.4902 | lr 8.47e-04 | grad 0.2632 +2026-04-10 10:04:05 - INFO - parrotllm.training - step 60570 | epoch 0 | loss 3.3496 | lr 8.46e-04 | grad 0.2394 +2026-04-10 10:04:08 - INFO - parrotllm.training - step 60580 | epoch 0 | loss 3.2779 | lr 8.46e-04 | grad 0.2536 +2026-04-10 10:04:11 - INFO - parrotllm.training - step 60590 | epoch 0 | loss 3.4545 | lr 8.46e-04 | grad 0.2265 +2026-04-10 10:04:14 - INFO - parrotllm.training - step 60600 | epoch 0 | loss 3.4300 | lr 8.46e-04 | grad 0.2914 +2026-04-10 10:04:17 - INFO - parrotllm.training - step 60610 | epoch 0 | loss 3.4194 | lr 8.46e-04 | grad 0.2585 +2026-04-10 10:04:20 - INFO - parrotllm.training - step 60620 | epoch 0 | loss 3.4029 | lr 8.45e-04 | grad 0.2558 +2026-04-10 10:04:23 - INFO - parrotllm.training - step 60630 | epoch 0 | loss 3.4362 | lr 8.45e-04 | grad 0.2364 +2026-04-10 10:04:26 - INFO - parrotllm.training - step 60640 | epoch 0 | loss 3.3378 | lr 8.45e-04 | grad 0.2367 +2026-04-10 10:04:29 - INFO - parrotllm.training - step 60650 | epoch 0 | loss 3.4351 | lr 8.45e-04 | grad 0.2447 +2026-04-10 10:04:32 - INFO - parrotllm.training - step 60660 | epoch 0 | loss 3.4595 | lr 8.45e-04 | grad 0.2367 +2026-04-10 10:04:35 - INFO - parrotllm.training - step 60670 | epoch 0 | loss 3.4877 | lr 8.45e-04 | grad 0.2772 +2026-04-10 10:04:38 - INFO - parrotllm.training - step 60680 | epoch 0 | loss 3.3592 | lr 8.44e-04 | grad 0.2379 +2026-04-10 10:04:39 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 60707/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 10:04:42 - INFO - parrotllm.training - step 60690 | epoch 0 | loss 3.3387 | lr 8.44e-04 | grad 0.2348 +2026-04-10 10:04:45 - INFO - parrotllm.training - step 60700 | epoch 0 | loss 3.3780 | lr 8.44e-04 | grad 0.2530 +2026-04-10 10:04:48 - INFO - parrotllm.training - step 60710 | epoch 0 | loss 3.4248 | lr 8.44e-04 | grad 0.2524 +2026-04-10 10:04:51 - INFO - parrotllm.training - step 60720 | epoch 0 | loss 3.4755 | lr 8.44e-04 | grad 0.2400 +2026-04-10 10:04:54 - INFO - parrotllm.training - step 60730 | epoch 0 | loss 3.3521 | lr 8.44e-04 | grad 0.2601 +2026-04-10 10:04:57 - INFO - parrotllm.training - step 60740 | epoch 0 | loss 3.4652 | lr 8.43e-04 | grad 0.2510 +2026-04-10 10:05:00 - INFO - parrotllm.training - step 60750 | epoch 0 | loss 3.5221 | lr 8.43e-04 | grad 0.2544 +2026-04-10 10:05:03 - INFO - parrotllm.training - step 60760 | epoch 0 | loss 3.5118 | lr 8.43e-04 | grad 0.2500 +2026-04-10 10:05:06 - INFO - parrotllm.training - step 60770 | epoch 0 | loss 3.4376 | lr 8.43e-04 | grad 0.2502 +2026-04-10 10:05:09 - INFO - parrotllm.training - step 60780 | epoch 0 | loss 3.3181 | lr 8.43e-04 | grad 0.2414 +2026-04-10 10:05:12 - INFO - parrotllm.training - step 60790 | epoch 0 | loss 3.5071 | lr 8.42e-04 | grad 0.2461 +2026-04-10 10:05:15 - INFO - parrotllm.training - step 60800 | epoch 0 | loss 3.4096 | lr 8.42e-04 | grad 0.2410 +2026-04-10 10:05:18 - INFO - parrotllm.training - step 60810 | epoch 0 | loss 3.3806 | lr 8.42e-04 | grad 0.2283 +2026-04-10 10:05:21 - INFO - parrotllm.training - step 60820 | epoch 0 | loss 3.3228 | lr 8.42e-04 | grad 0.2348 +2026-04-10 10:05:25 - INFO - parrotllm.training - step 60830 | epoch 0 | loss 3.4404 | lr 8.42e-04 | grad 0.2319 +2026-04-10 10:05:28 - INFO - parrotllm.training - step 60840 | epoch 0 | loss 3.4279 | lr 8.42e-04 | grad 0.2847 +2026-04-10 10:05:31 - INFO - parrotllm.training - step 60850 | epoch 0 | loss 3.4409 | lr 8.41e-04 | grad 0.2531 +2026-04-10 10:05:34 - INFO - parrotllm.training - step 60860 | epoch 0 | loss 3.3255 | lr 8.41e-04 | grad 0.2499 +2026-04-10 10:05:37 - INFO - parrotllm.training - step 60870 | epoch 0 | loss 3.3901 | lr 8.41e-04 | grad 0.2523 +2026-04-10 10:05:40 - INFO - parrotllm.training - step 60880 | epoch 0 | loss 3.3996 | lr 8.41e-04 | grad 0.2679 +2026-04-10 10:05:43 - INFO - parrotllm.training - step 60890 | epoch 0 | loss 3.4791 | lr 8.41e-04 | grad 0.2391 +2026-04-10 10:05:46 - INFO - parrotllm.training - step 60900 | epoch 0 | loss 3.3962 | lr 8.41e-04 | grad 0.2414 +2026-04-10 10:05:49 - INFO - parrotllm.training - step 60910 | epoch 0 | loss 3.4739 | lr 8.40e-04 | grad 0.2356 +2026-04-10 10:05:52 - INFO - parrotllm.training - step 60920 | epoch 0 | loss 3.4519 | lr 8.40e-04 | grad 0.2468 +2026-04-10 10:05:55 - INFO - parrotllm.training - step 60930 | epoch 0 | loss 3.4632 | lr 8.40e-04 | grad 0.2540 +2026-04-10 10:05:58 - INFO - parrotllm.training - step 60940 | epoch 0 | loss 3.4268 | lr 8.40e-04 | grad 0.2554 +2026-04-10 10:06:01 - INFO - parrotllm.training - step 60950 | epoch 0 | loss 3.3995 | lr 8.40e-04 | grad 0.2675 +2026-04-10 10:06:04 - INFO - parrotllm.training - step 60960 | epoch 0 | loss 3.3488 | lr 8.40e-04 | grad 0.2380 +2026-04-10 10:06:07 - INFO - parrotllm.training - step 60970 | epoch 0 | loss 3.4032 | lr 8.39e-04 | grad 0.2577 +2026-04-10 10:06:10 - INFO - parrotllm.training - step 60980 | epoch 0 | loss 3.3971 | lr 8.39e-04 | grad 0.2626 +2026-04-10 10:06:13 - INFO - parrotllm.training - step 60990 | epoch 0 | loss 3.4724 | lr 8.39e-04 | grad 0.2503 +2026-04-10 10:06:16 - INFO - parrotllm.training - step 61000 | epoch 0 | loss 3.4500 | lr 8.39e-04 | grad 0.2502 +2026-04-10 10:06:16 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:06:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:06:19 - INFO - parrotllm.training - Train: loss=3.4500, ppl=31.50 +2026-04-10 10:06:19 - INFO - parrotllm.training - Val: loss=3.3356, ppl=28.10 +2026-04-10 10:06:19 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:06:20 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3356_epoch_0000_step_0061000.pt +2026-04-10 10:06:22 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:06:25 - INFO - parrotllm.training - step 61010 | epoch 0 | loss 3.4136 | lr 8.39e-04 | grad 0.2520 +2026-04-10 10:06:28 - INFO - parrotllm.training - step 61020 | epoch 0 | loss 3.4083 | lr 8.38e-04 | grad 0.2421 +2026-04-10 10:06:31 - INFO - parrotllm.training - step 61030 | epoch 0 | loss 3.3878 | lr 8.38e-04 | grad 0.2703 +2026-04-10 10:06:34 - INFO - parrotllm.training - step 61040 | epoch 0 | loss 3.4789 | lr 8.38e-04 | grad 0.2458 +2026-04-10 10:06:38 - INFO - parrotllm.training - step 61050 | epoch 0 | loss 3.3507 | lr 8.38e-04 | grad 0.2510 +2026-04-10 10:06:41 - INFO - parrotllm.training - step 61060 | epoch 0 | loss 3.4641 | lr 8.38e-04 | grad 0.2319 +2026-04-10 10:06:44 - INFO - parrotllm.training - step 61070 | epoch 0 | loss 3.3887 | lr 8.38e-04 | grad 0.2346 +2026-04-10 10:06:47 - INFO - parrotllm.training - step 61080 | epoch 0 | loss 3.4451 | lr 8.37e-04 | grad 0.2426 +2026-04-10 10:06:50 - INFO - parrotllm.training - step 61090 | epoch 0 | loss 3.3116 | lr 8.37e-04 | grad 0.2291 +2026-04-10 10:06:53 - INFO - parrotllm.training - step 61100 | epoch 0 | loss 3.4191 | lr 8.37e-04 | grad 0.2549 +2026-04-10 10:06:56 - INFO - parrotllm.training - step 61110 | epoch 0 | loss 3.2415 | lr 8.37e-04 | grad 0.2861 +2026-04-10 10:06:59 - INFO - parrotllm.training - step 61120 | epoch 0 | loss 3.4104 | lr 8.37e-04 | grad 0.2471 +2026-04-10 10:07:02 - INFO - parrotllm.training - step 61130 | epoch 0 | loss 3.4630 | lr 8.37e-04 | grad 0.2286 +2026-04-10 10:07:05 - INFO - parrotllm.training - step 61140 | epoch 0 | loss 3.3990 | lr 8.36e-04 | grad 0.2660 +2026-04-10 10:07:08 - INFO - parrotllm.training - step 61150 | epoch 0 | loss 3.3344 | lr 8.36e-04 | grad 0.2418 +2026-04-10 10:07:11 - INFO - parrotllm.training - step 61160 | epoch 0 | loss 3.3702 | lr 8.36e-04 | grad 0.2659 +2026-04-10 10:07:14 - INFO - parrotllm.training - step 61170 | epoch 0 | loss 3.5133 | lr 8.36e-04 | grad 0.2802 +2026-04-10 10:07:17 - INFO - parrotllm.training - step 61180 | epoch 0 | loss 3.4101 | lr 8.36e-04 | grad 0.2972 +2026-04-10 10:07:20 - INFO - parrotllm.training - step 61190 | epoch 0 | loss 3.4432 | lr 8.35e-04 | grad 0.2479 +2026-04-10 10:07:23 - INFO - parrotllm.training - step 61200 | epoch 0 | loss 3.4530 | lr 8.35e-04 | grad 0.2486 +2026-04-10 10:07:26 - INFO - parrotllm.training - step 61210 | epoch 0 | loss 3.3421 | lr 8.35e-04 | grad 0.2402 +2026-04-10 10:07:29 - INFO - parrotllm.training - step 61220 | epoch 0 | loss 3.3775 | lr 8.35e-04 | grad 0.2455 +2026-04-10 10:07:33 - INFO - parrotllm.training - step 61230 | epoch 0 | loss 3.4339 | lr 8.35e-04 | grad 0.2580 +2026-04-10 10:07:36 - INFO - parrotllm.training - step 61240 | epoch 0 | loss 3.3800 | lr 8.35e-04 | grad 0.2343 +2026-04-10 10:07:39 - INFO - parrotllm.training - step 61250 | epoch 0 | loss 3.4250 | lr 8.34e-04 | grad 0.2443 +2026-04-10 10:07:42 - INFO - parrotllm.training - step 61260 | epoch 0 | loss 3.4821 | lr 8.34e-04 | grad 0.2187 +2026-04-10 10:07:45 - INFO - parrotllm.training - step 61270 | epoch 0 | loss 3.3066 | lr 8.34e-04 | grad 0.2606 +2026-04-10 10:07:48 - INFO - parrotllm.training - step 61280 | epoch 0 | loss 3.4160 | lr 8.34e-04 | grad 0.2440 +2026-04-10 10:07:51 - INFO - parrotllm.training - step 61290 | epoch 0 | loss 3.4440 | lr 8.34e-04 | grad 0.2323 +2026-04-10 10:07:54 - INFO - parrotllm.training - step 61300 | epoch 0 | loss 3.3626 | lr 8.34e-04 | grad 0.2292 +2026-04-10 10:07:57 - INFO - parrotllm.training - step 61310 | epoch 0 | loss 3.4385 | lr 8.33e-04 | grad 0.2665 +2026-04-10 10:08:00 - INFO - parrotllm.training - step 61320 | epoch 0 | loss 3.4491 | lr 8.33e-04 | grad 0.2266 +2026-04-10 10:08:03 - INFO - parrotllm.training - step 61330 | epoch 0 | loss 3.4219 | lr 8.33e-04 | grad 0.2419 +2026-04-10 10:08:06 - INFO - parrotllm.training - step 61340 | epoch 0 | loss 3.4332 | lr 8.33e-04 | grad 0.2282 +2026-04-10 10:08:09 - INFO - parrotllm.training - step 61350 | epoch 0 | loss 3.4516 | lr 8.33e-04 | grad 0.2284 +2026-04-10 10:08:12 - INFO - parrotllm.training - step 61360 | epoch 0 | loss 3.3255 | lr 8.32e-04 | grad 0.2293 +2026-04-10 10:08:15 - INFO - parrotllm.training - step 61370 | epoch 0 | loss 3.3173 | lr 8.32e-04 | grad 0.2277 +2026-04-10 10:08:18 - INFO - parrotllm.training - step 61380 | epoch 0 | loss 3.4335 | lr 8.32e-04 | grad 0.2415 +2026-04-10 10:08:21 - INFO - parrotllm.training - step 61390 | epoch 0 | loss 3.4256 | lr 8.32e-04 | grad 0.2434 +2026-04-10 10:08:24 - INFO - parrotllm.training - step 61400 | epoch 0 | loss 3.3479 | lr 8.32e-04 | grad 0.2301 +2026-04-10 10:08:27 - INFO - parrotllm.training - step 61410 | epoch 0 | loss 3.4241 | lr 8.32e-04 | grad 0.2814 +2026-04-10 10:08:30 - INFO - parrotllm.training - step 61420 | epoch 0 | loss 3.4297 | lr 8.31e-04 | grad 0.2330 +2026-04-10 10:08:34 - INFO - parrotllm.training - step 61430 | epoch 0 | loss 3.4128 | lr 8.31e-04 | grad 0.2281 +2026-04-10 10:08:37 - INFO - parrotllm.training - step 61440 | epoch 0 | loss 3.3827 | lr 8.31e-04 | grad 0.2438 +2026-04-10 10:08:40 - INFO - parrotllm.training - step 61450 | epoch 0 | loss 3.3552 | lr 8.31e-04 | grad 0.2550 +2026-04-10 10:08:43 - INFO - parrotllm.training - step 61460 | epoch 0 | loss 3.4023 | lr 8.31e-04 | grad 0.2574 +2026-04-10 10:08:46 - INFO - parrotllm.training - step 61470 | epoch 0 | loss 3.3499 | lr 8.31e-04 | grad 0.2801 +2026-04-10 10:08:49 - INFO - parrotllm.training - step 61480 | epoch 0 | loss 3.3121 | lr 8.30e-04 | grad 0.2496 +2026-04-10 10:08:52 - INFO - parrotllm.training - step 61490 | epoch 0 | loss 3.4146 | lr 8.30e-04 | grad 0.2652 +2026-04-10 10:08:55 - INFO - parrotllm.training - step 61500 | epoch 0 | loss 3.3818 | lr 8.30e-04 | grad 0.2470 +2026-04-10 10:08:55 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:08:55 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:08:58 - INFO - parrotllm.training - Train: loss=3.3818, ppl=29.42 +2026-04-10 10:08:58 - INFO - parrotllm.training - Val: loss=3.3381, ppl=28.17 +2026-04-10 10:08:59 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3381_epoch_0000_step_0061500.pt +2026-04-10 10:09:00 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3356, min_delta=0.001000). +2026-04-10 10:09:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:09:04 - INFO - parrotllm.training - step 61510 | epoch 0 | loss 3.3553 | lr 8.30e-04 | grad 0.2542 +2026-04-10 10:09:07 - INFO - parrotllm.training - step 61520 | epoch 0 | loss 3.3704 | lr 8.30e-04 | grad 0.2394 +2026-04-10 10:09:10 - INFO - parrotllm.training - step 61530 | epoch 0 | loss 3.2948 | lr 8.29e-04 | grad 0.2254 +2026-04-10 10:09:13 - INFO - parrotllm.training - step 61540 | epoch 0 | loss 3.4648 | lr 8.29e-04 | grad 0.2496 +2026-04-10 10:09:16 - INFO - parrotllm.training - step 61550 | epoch 0 | loss 3.4027 | lr 8.29e-04 | grad 0.2345 +2026-04-10 10:09:19 - INFO - parrotllm.training - step 61560 | epoch 0 | loss 3.4193 | lr 8.29e-04 | grad 0.2432 +2026-04-10 10:09:22 - INFO - parrotllm.training - step 61570 | epoch 0 | loss 3.4407 | lr 8.29e-04 | grad 0.2368 +2026-04-10 10:09:25 - INFO - parrotllm.training - step 61580 | epoch 0 | loss 3.4111 | lr 8.29e-04 | grad 0.2634 +2026-04-10 10:09:28 - INFO - parrotllm.training - step 61590 | epoch 0 | loss 3.4188 | lr 8.28e-04 | grad 0.2366 +2026-04-10 10:09:31 - INFO - parrotllm.training - step 61600 | epoch 0 | loss 3.4337 | lr 8.28e-04 | grad 0.2374 +2026-04-10 10:09:34 - INFO - parrotllm.training - step 61610 | epoch 0 | loss 3.2955 | lr 8.28e-04 | grad 0.3190 +2026-04-10 10:09:37 - INFO - parrotllm.training - step 61620 | epoch 0 | loss 3.3073 | lr 8.28e-04 | grad 0.2716 +2026-04-10 10:09:40 - INFO - parrotllm.training - step 61630 | epoch 0 | loss 3.3447 | lr 8.28e-04 | grad 0.2285 +2026-04-10 10:09:43 - INFO - parrotllm.training - step 61640 | epoch 0 | loss 3.3719 | lr 8.28e-04 | grad 0.2458 +2026-04-10 10:09:46 - INFO - parrotllm.training - step 61650 | epoch 0 | loss 3.3926 | lr 8.27e-04 | grad 0.2227 +2026-04-10 10:09:49 - INFO - parrotllm.training - step 61660 | epoch 0 | loss 3.4586 | lr 8.27e-04 | grad 0.2772 +2026-04-10 10:09:52 - INFO - parrotllm.training - step 61670 | epoch 0 | loss 3.3424 | lr 8.27e-04 | grad 0.2277 +2026-04-10 10:09:56 - INFO - parrotllm.training - step 61680 | epoch 0 | loss 3.4538 | lr 8.27e-04 | grad 0.2423 +2026-04-10 10:09:59 - INFO - parrotllm.training - step 61690 | epoch 0 | loss 3.4200 | lr 8.27e-04 | grad 0.2269 +2026-04-10 10:10:02 - INFO - parrotllm.training - step 61700 | epoch 0 | loss 3.3206 | lr 8.26e-04 | grad 0.2402 +2026-04-10 10:10:05 - INFO - parrotllm.training - step 61710 | epoch 0 | loss 3.3719 | lr 8.26e-04 | grad 0.2425 +2026-04-10 10:10:08 - INFO - parrotllm.training - step 61720 | epoch 0 | loss 3.3783 | lr 8.26e-04 | grad 0.2431 +2026-04-10 10:10:11 - INFO - parrotllm.training - step 61730 | epoch 0 | loss 3.3676 | lr 8.26e-04 | grad 0.2445 +2026-04-10 10:10:14 - INFO - parrotllm.training - step 61740 | epoch 0 | loss 3.3571 | lr 8.26e-04 | grad 0.2700 +2026-04-10 10:10:17 - INFO - parrotllm.training - step 61750 | epoch 0 | loss 3.5004 | lr 8.26e-04 | grad 0.2430 +2026-04-10 10:10:20 - INFO - parrotllm.training - step 61760 | epoch 0 | loss 3.3834 | lr 8.25e-04 | grad 0.2463 +2026-04-10 10:10:23 - INFO - parrotllm.training - step 61770 | epoch 0 | loss 3.4145 | lr 8.25e-04 | grad 0.2360 +2026-04-10 10:10:26 - INFO - parrotllm.training - step 61780 | epoch 0 | loss 3.3750 | lr 8.25e-04 | grad 0.2385 +2026-04-10 10:10:29 - INFO - parrotllm.training - step 61790 | epoch 0 | loss 3.3691 | lr 8.25e-04 | grad 0.2469 +2026-04-10 10:10:32 - INFO - parrotllm.training - step 61800 | epoch 0 | loss 3.4440 | lr 8.25e-04 | grad 0.2361 +2026-04-10 10:10:35 - INFO - parrotllm.training - step 61810 | epoch 0 | loss 3.3786 | lr 8.25e-04 | grad 0.2486 +2026-04-10 10:10:38 - INFO - parrotllm.training - step 61820 | epoch 0 | loss 3.4830 | lr 8.24e-04 | grad 0.2359 +2026-04-10 10:10:41 - INFO - parrotllm.training - step 61830 | epoch 0 | loss 3.3891 | lr 8.24e-04 | grad 0.2532 +2026-04-10 10:10:44 - INFO - parrotllm.training - step 61840 | epoch 0 | loss 3.3643 | lr 8.24e-04 | grad 0.2425 +2026-04-10 10:10:47 - INFO - parrotllm.training - step 61850 | epoch 0 | loss 3.5008 | lr 8.24e-04 | grad 0.2703 +2026-04-10 10:10:51 - INFO - parrotllm.training - step 61860 | epoch 0 | loss 3.3485 | lr 8.24e-04 | grad 0.2403 +2026-04-10 10:10:54 - INFO - parrotllm.training - step 61870 | epoch 0 | loss 3.4310 | lr 8.24e-04 | grad 0.2472 +2026-04-10 10:10:57 - INFO - parrotllm.training - step 61880 | epoch 0 | loss 3.2815 | lr 8.23e-04 | grad 0.2420 +2026-04-10 10:11:00 - INFO - parrotllm.training - step 61890 | epoch 0 | loss 3.4063 | lr 8.23e-04 | grad 0.2501 +2026-04-10 10:11:03 - INFO - parrotllm.training - step 61900 | epoch 0 | loss 3.4767 | lr 8.23e-04 | grad 0.2505 +2026-04-10 10:11:06 - INFO - parrotllm.training - step 61910 | epoch 0 | loss 3.3050 | lr 8.23e-04 | grad 0.2321 +2026-04-10 10:11:09 - INFO - parrotllm.training - step 61920 | epoch 0 | loss 3.3164 | lr 8.23e-04 | grad 0.2496 +2026-04-10 10:11:12 - INFO - parrotllm.training - step 61930 | epoch 0 | loss 3.3574 | lr 8.22e-04 | grad 0.2595 +2026-04-10 10:11:15 - INFO - parrotllm.training - step 61940 | epoch 0 | loss 3.2886 | lr 8.22e-04 | grad 0.2567 +2026-04-10 10:11:18 - INFO - parrotllm.training - step 61950 | epoch 0 | loss 3.3365 | lr 8.22e-04 | grad 0.2585 +2026-04-10 10:11:21 - INFO - parrotllm.training - step 61960 | epoch 0 | loss 3.4358 | lr 8.22e-04 | grad 0.2586 +2026-04-10 10:11:24 - INFO - parrotllm.training - step 61970 | epoch 0 | loss 3.5736 | lr 8.22e-04 | grad 0.2408 +2026-04-10 10:11:27 - INFO - parrotllm.training - step 61980 | epoch 0 | loss 3.3796 | lr 8.22e-04 | grad 0.2463 +2026-04-10 10:11:30 - INFO - parrotllm.training - step 61990 | epoch 0 | loss 3.3823 | lr 8.21e-04 | grad 0.2869 +2026-04-10 10:11:33 - INFO - parrotllm.training - step 62000 | epoch 0 | loss 3.2871 | lr 8.21e-04 | grad 0.2359 +2026-04-10 10:11:33 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:11:33 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:11:36 - INFO - parrotllm.training - Train: loss=3.2871, ppl=26.77 +2026-04-10 10:11:36 - INFO - parrotllm.training - Val: loss=3.3354, ppl=28.09 +2026-04-10 10:11:37 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3354_epoch_0000_step_0062000.pt +2026-04-10 10:11:39 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3356, min_delta=0.001000). +2026-04-10 10:11:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:11:42 - INFO - parrotllm.training - step 62010 | epoch 0 | loss 3.3683 | lr 8.21e-04 | grad 0.2945 +2026-04-10 10:11:45 - INFO - parrotllm.training - step 62020 | epoch 0 | loss 3.4092 | lr 8.21e-04 | grad 0.2666 +2026-04-10 10:11:48 - INFO - parrotllm.training - step 62030 | epoch 0 | loss 3.3960 | lr 8.21e-04 | grad 0.2281 +2026-04-10 10:11:51 - INFO - parrotllm.training - step 62040 | epoch 0 | loss 3.4035 | lr 8.21e-04 | grad 0.2266 +2026-04-10 10:11:54 - INFO - parrotllm.training - step 62050 | epoch 0 | loss 3.2666 | lr 8.20e-04 | grad 0.2349 +2026-04-10 10:11:57 - INFO - parrotllm.training - step 62060 | epoch 0 | loss 3.5206 | lr 8.20e-04 | grad 0.2360 +2026-04-10 10:12:00 - INFO - parrotllm.training - step 62070 | epoch 0 | loss 3.3823 | lr 8.20e-04 | grad 0.2359 +2026-04-10 10:12:03 - INFO - parrotllm.training - step 62080 | epoch 0 | loss 3.3642 | lr 8.20e-04 | grad 0.2352 +2026-04-10 10:12:06 - INFO - parrotllm.training - step 62090 | epoch 0 | loss 3.4453 | lr 8.20e-04 | grad 0.2317 +2026-04-10 10:12:09 - INFO - parrotllm.training - step 62100 | epoch 0 | loss 3.3795 | lr 8.19e-04 | grad 0.2528 +2026-04-10 10:12:12 - INFO - parrotllm.training - step 62110 | epoch 0 | loss 3.3731 | lr 8.19e-04 | grad 0.2576 +2026-04-10 10:12:16 - INFO - parrotllm.training - step 62120 | epoch 0 | loss 3.5600 | lr 8.19e-04 | grad 0.2593 +2026-04-10 10:12:19 - INFO - parrotllm.training - step 62130 | epoch 0 | loss 3.4207 | lr 8.19e-04 | grad 0.2435 +2026-04-10 10:12:22 - INFO - parrotllm.training - step 62140 | epoch 0 | loss 3.2748 | lr 8.19e-04 | grad 0.2584 +2026-04-10 10:12:25 - INFO - parrotllm.training - step 62150 | epoch 0 | loss 3.3916 | lr 8.19e-04 | grad 0.2420 +2026-04-10 10:12:28 - INFO - parrotllm.training - step 62160 | epoch 0 | loss 3.4933 | lr 8.18e-04 | grad 0.2433 +2026-04-10 10:12:31 - INFO - parrotllm.training - step 62170 | epoch 0 | loss 3.3717 | lr 8.18e-04 | grad 0.2813 +2026-04-10 10:12:34 - INFO - parrotllm.training - step 62180 | epoch 0 | loss 3.3526 | lr 8.18e-04 | grad 0.2566 +2026-04-10 10:12:37 - INFO - parrotllm.training - step 62190 | epoch 0 | loss 3.3262 | lr 8.18e-04 | grad 0.2718 +2026-04-10 10:12:40 - INFO - parrotllm.training - step 62200 | epoch 0 | loss 3.3961 | lr 8.18e-04 | grad 0.2509 +2026-04-10 10:12:43 - INFO - parrotllm.training - step 62210 | epoch 0 | loss 3.4104 | lr 8.18e-04 | grad 0.2548 +2026-04-10 10:12:46 - INFO - parrotllm.training - step 62220 | epoch 0 | loss 3.3648 | lr 8.17e-04 | grad 0.2783 +2026-04-10 10:12:49 - INFO - parrotllm.training - step 62230 | epoch 0 | loss 3.3178 | lr 8.17e-04 | grad 0.2564 +2026-04-10 10:12:52 - INFO - parrotllm.training - step 62240 | epoch 0 | loss 3.4282 | lr 8.17e-04 | grad 0.2349 +2026-04-10 10:12:55 - INFO - parrotllm.training - step 62250 | epoch 0 | loss 3.3571 | lr 8.17e-04 | grad 0.2363 +2026-04-10 10:12:58 - INFO - parrotllm.training - step 62260 | epoch 0 | loss 3.3555 | lr 8.17e-04 | grad 0.2461 +2026-04-10 10:13:01 - INFO - parrotllm.training - step 62270 | epoch 0 | loss 3.4557 | lr 8.16e-04 | grad 0.2533 +2026-04-10 10:13:04 - INFO - parrotllm.training - step 62280 | epoch 0 | loss 3.5135 | lr 8.16e-04 | grad 0.2355 +2026-04-10 10:13:07 - INFO - parrotllm.training - step 62290 | epoch 0 | loss 3.4694 | lr 8.16e-04 | grad 0.2427 +2026-04-10 10:13:10 - INFO - parrotllm.training - step 62300 | epoch 0 | loss 3.3578 | lr 8.16e-04 | grad 0.2303 +2026-04-10 10:13:14 - INFO - parrotllm.training - step 62310 | epoch 0 | loss 3.4825 | lr 8.16e-04 | grad 0.2214 +2026-04-10 10:13:17 - INFO - parrotllm.training - step 62320 | epoch 0 | loss 3.4384 | lr 8.16e-04 | grad 0.2436 +2026-04-10 10:13:20 - INFO - parrotllm.training - step 62330 | epoch 0 | loss 3.4193 | lr 8.15e-04 | grad 0.2726 +2026-04-10 10:13:23 - INFO - parrotllm.training - step 62340 | epoch 0 | loss 3.3601 | lr 8.15e-04 | grad 0.2496 +2026-04-10 10:13:26 - INFO - parrotllm.training - step 62350 | epoch 0 | loss 3.3813 | lr 8.15e-04 | grad 0.2685 +2026-04-10 10:13:29 - INFO - parrotllm.training - step 62360 | epoch 0 | loss 3.3534 | lr 8.15e-04 | grad 0.2654 +2026-04-10 10:13:32 - INFO - parrotllm.training - step 62370 | epoch 0 | loss 3.3880 | lr 8.15e-04 | grad 0.2770 +2026-04-10 10:13:35 - INFO - parrotllm.training - step 62380 | epoch 0 | loss 3.5338 | lr 8.15e-04 | grad 0.2342 +2026-04-10 10:13:38 - INFO - parrotllm.training - step 62390 | epoch 0 | loss 3.4109 | lr 8.14e-04 | grad 0.2313 +2026-04-10 10:13:41 - INFO - parrotllm.training - step 62400 | epoch 0 | loss 3.4559 | lr 8.14e-04 | grad 0.2662 +2026-04-10 10:13:44 - INFO - parrotllm.training - step 62410 | epoch 0 | loss 3.5475 | lr 8.14e-04 | grad 0.2428 +2026-04-10 10:13:47 - INFO - parrotllm.training - step 62420 | epoch 0 | loss 3.4687 | lr 8.14e-04 | grad 0.2710 +2026-04-10 10:13:50 - INFO - parrotllm.training - step 62430 | epoch 0 | loss 3.3723 | lr 8.14e-04 | grad 0.2398 +2026-04-10 10:13:53 - INFO - parrotllm.training - step 62440 | epoch 0 | loss 3.4763 | lr 8.13e-04 | grad 0.2497 +2026-04-10 10:13:56 - INFO - parrotllm.training - step 62450 | epoch 0 | loss 3.4378 | lr 8.13e-04 | grad 0.2793 +2026-04-10 10:13:59 - INFO - parrotllm.training - step 62460 | epoch 0 | loss 3.4123 | lr 8.13e-04 | grad 0.2157 +2026-04-10 10:14:02 - INFO - parrotllm.training - step 62470 | epoch 0 | loss 3.4400 | lr 8.13e-04 | grad 0.2512 +2026-04-10 10:14:05 - INFO - parrotllm.training - step 62480 | epoch 0 | loss 3.3491 | lr 8.13e-04 | grad 0.2449 +2026-04-10 10:14:09 - INFO - parrotllm.training - step 62490 | epoch 0 | loss 3.3916 | lr 8.13e-04 | grad 0.2597 +2026-04-10 10:14:12 - INFO - parrotllm.training - step 62500 | epoch 0 | loss 3.5191 | lr 8.12e-04 | grad 0.2579 +2026-04-10 10:14:12 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:14:12 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:14:15 - INFO - parrotllm.training - Train: loss=3.5191, ppl=33.75 +2026-04-10 10:14:15 - INFO - parrotllm.training - Val: loss=3.3375, ppl=28.15 +2026-04-10 10:14:16 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3375_epoch_0000_step_0062500.pt +2026-04-10 10:14:18 - INFO - parrotllm.training - No validation improvement for 3/15 evaluation(s) (best=3.3356, min_delta=0.001000). +2026-04-10 10:14:18 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:14:19 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0062500.pt +2026-04-10 10:14:23 - INFO - parrotllm.training - step 62510 | epoch 0 | loss 3.3785 | lr 8.12e-04 | grad 0.2647 +2026-04-10 10:14:26 - INFO - parrotllm.training - step 62520 | epoch 0 | loss 3.3704 | lr 8.12e-04 | grad 0.2244 +2026-04-10 10:14:29 - INFO - parrotllm.training - step 62530 | epoch 0 | loss 3.3322 | lr 8.12e-04 | grad 0.2437 +2026-04-10 10:14:33 - INFO - parrotllm.training - step 62540 | epoch 0 | loss 3.2491 | lr 8.12e-04 | grad 0.2304 +2026-04-10 10:14:36 - INFO - parrotllm.training - step 62550 | epoch 0 | loss 3.3316 | lr 8.12e-04 | grad 0.2409 +2026-04-10 10:14:39 - INFO - parrotllm.training - step 62560 | epoch 0 | loss 3.3923 | lr 8.11e-04 | grad 0.2550 +2026-04-10 10:14:42 - INFO - parrotllm.training - step 62570 | epoch 0 | loss 3.4032 | lr 8.11e-04 | grad 0.3085 +2026-04-10 10:14:45 - INFO - parrotllm.training - step 62580 | epoch 0 | loss 3.3782 | lr 8.11e-04 | grad 0.2710 +2026-04-10 10:14:48 - INFO - parrotllm.training - step 62590 | epoch 0 | loss 3.3853 | lr 8.11e-04 | grad 0.2576 +2026-04-10 10:14:51 - INFO - parrotllm.training - step 62600 | epoch 0 | loss 3.3911 | lr 8.11e-04 | grad 0.2600 +2026-04-10 10:14:54 - INFO - parrotllm.training - step 62610 | epoch 0 | loss 3.2557 | lr 8.10e-04 | grad 0.2432 +2026-04-10 10:14:57 - INFO - parrotllm.training - step 62620 | epoch 0 | loss 3.3917 | lr 8.10e-04 | grad 0.2587 +2026-04-10 10:15:00 - INFO - parrotllm.training - step 62630 | epoch 0 | loss 3.4244 | lr 8.10e-04 | grad 0.2469 +2026-04-10 10:15:03 - INFO - parrotllm.training - step 62640 | epoch 0 | loss 3.4011 | lr 8.10e-04 | grad 0.2606 +2026-04-10 10:15:06 - INFO - parrotllm.training - step 62650 | epoch 0 | loss 3.3994 | lr 8.10e-04 | grad 0.2417 +2026-04-10 10:15:09 - INFO - parrotllm.training - step 62660 | epoch 0 | loss 3.4196 | lr 8.10e-04 | grad 0.2547 +2026-04-10 10:15:12 - INFO - parrotllm.training - step 62670 | epoch 0 | loss 3.4953 | lr 8.09e-04 | grad 0.2485 +2026-04-10 10:15:15 - INFO - parrotllm.training - step 62680 | epoch 0 | loss 3.4100 | lr 8.09e-04 | grad 0.2736 +2026-04-10 10:15:18 - INFO - parrotllm.training - step 62690 | epoch 0 | loss 3.3328 | lr 8.09e-04 | grad 0.2388 +2026-04-10 10:15:21 - INFO - parrotllm.training - step 62700 | epoch 0 | loss 3.3611 | lr 8.09e-04 | grad 0.2540 +2026-04-10 10:15:24 - INFO - parrotllm.training - step 62710 | epoch 0 | loss 3.3881 | lr 8.09e-04 | grad 0.2538 +2026-04-10 10:15:27 - INFO - parrotllm.training - step 62720 | epoch 0 | loss 3.3857 | lr 8.09e-04 | grad 0.2815 +2026-04-10 10:15:30 - INFO - parrotllm.training - step 62730 | epoch 0 | loss 3.3755 | lr 8.08e-04 | grad 0.2594 +2026-04-10 10:15:34 - INFO - parrotllm.training - step 62740 | epoch 0 | loss 3.5255 | lr 8.08e-04 | grad 0.2344 +2026-04-10 10:15:37 - INFO - parrotllm.training - step 62750 | epoch 0 | loss 3.3578 | lr 8.08e-04 | grad 0.2263 +2026-04-10 10:15:40 - INFO - parrotllm.training - step 62760 | epoch 0 | loss 3.3440 | lr 8.08e-04 | grad 0.2539 +2026-04-10 10:15:43 - INFO - parrotllm.training - step 62770 | epoch 0 | loss 3.3772 | lr 8.08e-04 | grad 0.2634 +2026-04-10 10:15:46 - INFO - parrotllm.training - step 62780 | epoch 0 | loss 3.3549 | lr 8.08e-04 | grad 0.2567 +2026-04-10 10:15:49 - INFO - parrotllm.training - step 62790 | epoch 0 | loss 3.3628 | lr 8.07e-04 | grad 0.2507 +2026-04-10 10:15:52 - INFO - parrotllm.training - step 62800 | epoch 0 | loss 3.2830 | lr 8.07e-04 | grad 0.2443 +2026-04-10 10:15:55 - INFO - parrotllm.training - step 62810 | epoch 0 | loss 3.2948 | lr 8.07e-04 | grad 0.2377 +2026-04-10 10:15:58 - INFO - parrotllm.training - step 62820 | epoch 0 | loss 3.4106 | lr 8.07e-04 | grad 0.2425 +2026-04-10 10:16:01 - INFO - parrotllm.training - step 62830 | epoch 0 | loss 3.3822 | lr 8.07e-04 | grad 0.2575 +2026-04-10 10:16:04 - INFO - parrotllm.training - step 62840 | epoch 0 | loss 3.4229 | lr 8.06e-04 | grad 0.2327 +2026-04-10 10:16:07 - INFO - parrotllm.training - step 62850 | epoch 0 | loss 3.4668 | lr 8.06e-04 | grad 0.2305 +2026-04-10 10:16:10 - INFO - parrotllm.training - step 62860 | epoch 0 | loss 3.4152 | lr 8.06e-04 | grad 0.2425 +2026-04-10 10:16:13 - INFO - parrotllm.training - step 62870 | epoch 0 | loss 3.3637 | lr 8.06e-04 | grad 0.2477 +2026-04-10 10:16:16 - INFO - parrotllm.training - step 62880 | epoch 0 | loss 3.3972 | lr 8.06e-04 | grad 0.2521 +2026-04-10 10:16:19 - INFO - parrotllm.training - step 62890 | epoch 0 | loss 3.4134 | lr 8.06e-04 | grad 0.2431 +2026-04-10 10:16:22 - INFO - parrotllm.training - step 62900 | epoch 0 | loss 3.3744 | lr 8.05e-04 | grad 0.2537 +2026-04-10 10:16:25 - INFO - parrotllm.training - step 62910 | epoch 0 | loss 3.4069 | lr 8.05e-04 | grad 0.2355 +2026-04-10 10:16:29 - INFO - parrotllm.training - step 62920 | epoch 0 | loss 3.3950 | lr 8.05e-04 | grad 0.2539 +2026-04-10 10:16:32 - INFO - parrotllm.training - step 62930 | epoch 0 | loss 3.3822 | lr 8.05e-04 | grad 0.2664 +2026-04-10 10:16:35 - INFO - parrotllm.training - step 62940 | epoch 0 | loss 3.3598 | lr 8.05e-04 | grad 0.2464 +2026-04-10 10:16:38 - INFO - parrotllm.training - step 62950 | epoch 0 | loss 3.3262 | lr 8.05e-04 | grad 0.2543 +2026-04-10 10:16:41 - INFO - parrotllm.training - step 62960 | epoch 0 | loss 3.3383 | lr 8.04e-04 | grad 0.2434 +2026-04-10 10:16:44 - INFO - parrotllm.training - step 62970 | epoch 0 | loss 3.4731 | lr 8.04e-04 | grad 0.2399 +2026-04-10 10:16:47 - INFO - parrotllm.training - step 62980 | epoch 0 | loss 3.3524 | lr 8.04e-04 | grad 0.2787 +2026-04-10 10:16:50 - INFO - parrotllm.training - step 62990 | epoch 0 | loss 3.4760 | lr 8.04e-04 | grad 0.2559 +2026-04-10 10:16:53 - INFO - parrotllm.training - step 63000 | epoch 0 | loss 3.3920 | lr 8.04e-04 | grad 0.2760 +2026-04-10 10:16:53 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:16:53 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:16:56 - INFO - parrotllm.training - Train: loss=3.3920, ppl=29.73 +2026-04-10 10:16:56 - INFO - parrotllm.training - Val: loss=3.3357, ppl=28.10 +2026-04-10 10:16:57 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3357_epoch_0000_step_0063000.pt +2026-04-10 10:16:59 - INFO - parrotllm.training - No validation improvement for 4/15 evaluation(s) (best=3.3356, min_delta=0.001000). +2026-04-10 10:16:59 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:17:02 - INFO - parrotllm.training - step 63010 | epoch 0 | loss 3.4404 | lr 8.03e-04 | grad 0.2430 +2026-04-10 10:17:05 - INFO - parrotllm.training - step 63020 | epoch 0 | loss 3.3553 | lr 8.03e-04 | grad 0.2416 +2026-04-10 10:17:08 - INFO - parrotllm.training - step 63030 | epoch 0 | loss 3.3400 | lr 8.03e-04 | grad 0.3113 +2026-04-10 10:17:11 - INFO - parrotllm.training - step 63040 | epoch 0 | loss 3.4320 | lr 8.03e-04 | grad 0.2551 +2026-04-10 10:17:14 - INFO - parrotllm.training - step 63050 | epoch 0 | loss 3.4057 | lr 8.03e-04 | grad 0.2312 +2026-04-10 10:17:17 - INFO - parrotllm.training - step 63060 | epoch 0 | loss 3.3365 | lr 8.03e-04 | grad 0.2446 +2026-04-10 10:17:20 - INFO - parrotllm.training - step 63070 | epoch 0 | loss 3.3464 | lr 8.02e-04 | grad 0.2495 +2026-04-10 10:17:23 - INFO - parrotllm.training - step 63080 | epoch 0 | loss 3.4001 | lr 8.02e-04 | grad 0.2569 +2026-04-10 10:17:26 - INFO - parrotllm.training - step 63090 | epoch 0 | loss 3.4106 | lr 8.02e-04 | grad 0.2668 +2026-04-10 10:17:29 - INFO - parrotllm.training - step 63100 | epoch 0 | loss 3.4417 | lr 8.02e-04 | grad 0.2805 +2026-04-10 10:17:32 - INFO - parrotllm.training - step 63110 | epoch 0 | loss 3.4406 | lr 8.02e-04 | grad 0.2218 +2026-04-10 10:17:35 - INFO - parrotllm.training - step 63120 | epoch 0 | loss 3.4224 | lr 8.02e-04 | grad 0.2436 +2026-04-10 10:17:38 - INFO - parrotllm.training - step 63130 | epoch 0 | loss 3.2798 | lr 8.01e-04 | grad 0.2315 +2026-04-10 10:17:42 - INFO - parrotllm.training - step 63140 | epoch 0 | loss 3.3675 | lr 8.01e-04 | grad 0.2391 +2026-04-10 10:17:45 - INFO - parrotllm.training - step 63150 | epoch 0 | loss 3.4222 | lr 8.01e-04 | grad 0.2365 +2026-04-10 10:17:48 - INFO - parrotllm.training - step 63160 | epoch 0 | loss 3.5166 | lr 8.01e-04 | grad 0.2393 +2026-04-10 10:17:51 - INFO - parrotllm.training - step 63170 | epoch 0 | loss 3.5266 | lr 8.01e-04 | grad 0.2396 +2026-04-10 10:17:54 - INFO - parrotllm.training - step 63180 | epoch 0 | loss 3.3530 | lr 8.00e-04 | grad 0.2666 +2026-04-10 10:17:57 - INFO - parrotllm.training - step 63190 | epoch 0 | loss 3.4411 | lr 8.00e-04 | grad 0.2478 +2026-04-10 10:18:00 - INFO - parrotllm.training - step 63200 | epoch 0 | loss 3.4092 | lr 8.00e-04 | grad 0.2443 +2026-04-10 10:18:03 - INFO - parrotllm.training - step 63210 | epoch 0 | loss 3.3553 | lr 8.00e-04 | grad 0.2405 +2026-04-10 10:18:06 - INFO - parrotllm.training - step 63220 | epoch 0 | loss 3.3910 | lr 8.00e-04 | grad 0.2654 +2026-04-10 10:18:09 - INFO - parrotllm.training - step 63230 | epoch 0 | loss 3.3101 | lr 8.00e-04 | grad 0.2543 +2026-04-10 10:18:12 - INFO - parrotllm.training - step 63240 | epoch 0 | loss 3.3944 | lr 7.99e-04 | grad 0.2489 +2026-04-10 10:18:15 - INFO - parrotllm.training - step 63250 | epoch 0 | loss 3.4544 | lr 7.99e-04 | grad 0.2574 +2026-04-10 10:18:18 - INFO - parrotllm.training - step 63260 | epoch 0 | loss 3.3634 | lr 7.99e-04 | grad 0.2343 +2026-04-10 10:18:21 - INFO - parrotllm.training - step 63270 | epoch 0 | loss 3.4535 | lr 7.99e-04 | grad 0.2241 +2026-04-10 10:18:24 - INFO - parrotllm.training - step 63280 | epoch 0 | loss 3.4001 | lr 7.99e-04 | grad 0.2792 +2026-04-10 10:18:27 - INFO - parrotllm.training - step 63290 | epoch 0 | loss 3.5301 | lr 7.99e-04 | grad 0.2478 +2026-04-10 10:18:30 - INFO - parrotllm.training - step 63300 | epoch 0 | loss 3.3515 | lr 7.98e-04 | grad 0.2296 +2026-04-10 10:18:33 - INFO - parrotllm.training - step 63310 | epoch 0 | loss 3.4459 | lr 7.98e-04 | grad 0.2583 +2026-04-10 10:18:37 - INFO - parrotllm.training - step 63320 | epoch 0 | loss 3.4409 | lr 7.98e-04 | grad 0.2590 +2026-04-10 10:18:40 - INFO - parrotllm.training - step 63330 | epoch 0 | loss 3.3694 | lr 7.98e-04 | grad 0.2516 +2026-04-10 10:18:43 - INFO - parrotllm.training - step 63340 | epoch 0 | loss 3.4654 | lr 7.98e-04 | grad 0.2375 +2026-04-10 10:18:45 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 63371/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 10:18:46 - INFO - parrotllm.training - step 63350 | epoch 0 | loss 3.4025 | lr 7.97e-04 | grad 0.2372 +2026-04-10 10:18:49 - INFO - parrotllm.training - step 63360 | epoch 0 | loss 3.3600 | lr 7.97e-04 | grad 0.2706 +2026-04-10 10:18:52 - INFO - parrotllm.training - step 63370 | epoch 0 | loss 3.4099 | lr 7.97e-04 | grad 0.2426 +2026-04-10 10:18:55 - INFO - parrotllm.training - step 63380 | epoch 0 | loss 3.3280 | lr 7.97e-04 | grad 0.2507 +2026-04-10 10:18:58 - INFO - parrotllm.training - step 63390 | epoch 0 | loss 3.3606 | lr 7.97e-04 | grad 0.2529 +2026-04-10 10:19:01 - INFO - parrotllm.training - step 63400 | epoch 0 | loss 3.3369 | lr 7.97e-04 | grad 0.2289 +2026-04-10 10:19:04 - INFO - parrotllm.training - step 63410 | epoch 0 | loss 3.4014 | lr 7.96e-04 | grad 0.2380 +2026-04-10 10:19:07 - INFO - parrotllm.training - step 63420 | epoch 0 | loss 3.3079 | lr 7.96e-04 | grad 0.2626 +2026-04-10 10:19:10 - INFO - parrotllm.training - step 63430 | epoch 0 | loss 3.4294 | lr 7.96e-04 | grad 0.2523 +2026-04-10 10:19:14 - INFO - parrotllm.training - step 63440 | epoch 0 | loss 3.3230 | lr 7.96e-04 | grad 0.2306 +2026-04-10 10:19:17 - INFO - parrotllm.training - step 63450 | epoch 0 | loss 3.3739 | lr 7.96e-04 | grad 0.2349 +2026-04-10 10:19:20 - INFO - parrotllm.training - step 63460 | epoch 0 | loss 3.4318 | lr 7.96e-04 | grad 0.2383 +2026-04-10 10:19:23 - INFO - parrotllm.training - step 63470 | epoch 0 | loss 3.2822 | lr 7.95e-04 | grad 0.2515 +2026-04-10 10:19:26 - INFO - parrotllm.training - step 63480 | epoch 0 | loss 3.2524 | lr 7.95e-04 | grad 0.2359 +2026-04-10 10:19:29 - INFO - parrotllm.training - step 63490 | epoch 0 | loss 3.3594 | lr 7.95e-04 | grad 0.2505 +2026-04-10 10:19:32 - INFO - parrotllm.training - step 63500 | epoch 0 | loss 3.3703 | lr 7.95e-04 | grad 0.2487 +2026-04-10 10:19:32 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:19:32 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:19:35 - INFO - parrotllm.training - Train: loss=3.3703, ppl=29.09 +2026-04-10 10:19:35 - INFO - parrotllm.training - Val: loss=3.3333, ppl=28.03 +2026-04-10 10:19:35 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:19:36 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3333_epoch_0000_step_0063500.pt +2026-04-10 10:19:38 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:19:41 - INFO - parrotllm.training - step 63510 | epoch 0 | loss 3.3642 | lr 7.95e-04 | grad 0.2395 +2026-04-10 10:19:44 - INFO - parrotllm.training - step 63520 | epoch 0 | loss 3.3898 | lr 7.94e-04 | grad 0.2754 +2026-04-10 10:19:47 - INFO - parrotllm.training - step 63530 | epoch 0 | loss 3.3419 | lr 7.94e-04 | grad 0.2755 +2026-04-10 10:19:50 - INFO - parrotllm.training - step 63540 | epoch 0 | loss 3.3820 | lr 7.94e-04 | grad 0.3169 +2026-04-10 10:19:53 - INFO - parrotllm.training - step 63550 | epoch 0 | loss 3.4196 | lr 7.94e-04 | grad 0.2420 +2026-04-10 10:19:56 - INFO - parrotllm.training - step 63560 | epoch 0 | loss 3.3432 | lr 7.94e-04 | grad 0.2966 +2026-04-10 10:19:59 - INFO - parrotllm.training - step 63570 | epoch 0 | loss 3.3908 | lr 7.94e-04 | grad 0.2786 +2026-04-10 10:20:02 - INFO - parrotllm.training - step 63580 | epoch 0 | loss 3.4922 | lr 7.93e-04 | grad 0.2895 +2026-04-10 10:20:05 - INFO - parrotllm.training - step 63590 | epoch 0 | loss 3.4932 | lr 7.93e-04 | grad 0.2390 +2026-04-10 10:20:08 - INFO - parrotllm.training - step 63600 | epoch 0 | loss 3.4101 | lr 7.93e-04 | grad 0.2570 +2026-04-10 10:20:11 - INFO - parrotllm.training - step 63610 | epoch 0 | loss 3.3129 | lr 7.93e-04 | grad 0.2537 +2026-04-10 10:20:14 - INFO - parrotllm.training - step 63620 | epoch 0 | loss 3.4708 | lr 7.93e-04 | grad 0.2417 +2026-04-10 10:20:17 - INFO - parrotllm.training - step 63630 | epoch 0 | loss 3.2927 | lr 7.93e-04 | grad 0.2631 +2026-04-10 10:20:20 - INFO - parrotllm.training - step 63640 | epoch 0 | loss 3.4405 | lr 7.92e-04 | grad 0.2347 +2026-04-10 10:20:23 - INFO - parrotllm.training - step 63650 | epoch 0 | loss 3.3035 | lr 7.92e-04 | grad 0.2624 +2026-04-10 10:20:27 - INFO - parrotllm.training - step 63660 | epoch 0 | loss 3.4239 | lr 7.92e-04 | grad 0.2367 +2026-04-10 10:20:30 - INFO - parrotllm.training - step 63670 | epoch 0 | loss 3.3596 | lr 7.92e-04 | grad 0.2424 +2026-04-10 10:20:33 - INFO - parrotllm.training - step 63680 | epoch 0 | loss 3.4300 | lr 7.92e-04 | grad 0.2457 +2026-04-10 10:20:36 - INFO - parrotllm.training - step 63690 | epoch 0 | loss 3.3982 | lr 7.92e-04 | grad 0.2452 +2026-04-10 10:20:39 - INFO - parrotllm.training - step 63700 | epoch 0 | loss 3.3556 | lr 7.91e-04 | grad 0.2202 +2026-04-10 10:20:42 - INFO - parrotllm.training - step 63710 | epoch 0 | loss 3.3219 | lr 7.91e-04 | grad 0.2762 +2026-04-10 10:20:45 - INFO - parrotllm.training - step 63720 | epoch 0 | loss 3.3772 | lr 7.91e-04 | grad 0.2507 +2026-04-10 10:20:48 - INFO - parrotllm.training - step 63730 | epoch 0 | loss 3.4231 | lr 7.91e-04 | grad 0.2555 +2026-04-10 10:20:51 - INFO - parrotllm.training - step 63740 | epoch 0 | loss 3.4633 | lr 7.91e-04 | grad 0.2371 +2026-04-10 10:20:54 - INFO - parrotllm.training - step 63750 | epoch 0 | loss 3.4779 | lr 7.90e-04 | grad 0.2307 +2026-04-10 10:20:57 - INFO - parrotllm.training - step 63760 | epoch 0 | loss 3.3480 | lr 7.90e-04 | grad 0.2456 +2026-04-10 10:21:00 - INFO - parrotllm.training - step 63770 | epoch 0 | loss 3.3230 | lr 7.90e-04 | grad 0.2649 +2026-04-10 10:21:03 - INFO - parrotllm.training - step 63780 | epoch 0 | loss 3.2825 | lr 7.90e-04 | grad 0.2438 +2026-04-10 10:21:06 - INFO - parrotllm.training - step 63790 | epoch 0 | loss 3.2982 | lr 7.90e-04 | grad 0.2850 +2026-04-10 10:21:09 - INFO - parrotllm.training - step 63800 | epoch 0 | loss 3.4023 | lr 7.90e-04 | grad 0.2726 +2026-04-10 10:21:12 - INFO - parrotllm.training - step 63810 | epoch 0 | loss 3.4163 | lr 7.89e-04 | grad 0.2600 +2026-04-10 10:21:15 - INFO - parrotllm.training - step 63820 | epoch 0 | loss 3.4483 | lr 7.89e-04 | grad 0.2470 +2026-04-10 10:21:18 - INFO - parrotllm.training - step 63830 | epoch 0 | loss 3.3180 | lr 7.89e-04 | grad 0.2406 +2026-04-10 10:21:21 - INFO - parrotllm.training - step 63840 | epoch 0 | loss 3.4870 | lr 7.89e-04 | grad 0.2461 +2026-04-10 10:21:25 - INFO - parrotllm.training - step 63850 | epoch 0 | loss 3.4242 | lr 7.89e-04 | grad 0.2595 +2026-04-10 10:21:28 - INFO - parrotllm.training - step 63860 | epoch 0 | loss 3.3839 | lr 7.89e-04 | grad 0.2714 +2026-04-10 10:21:31 - INFO - parrotllm.training - step 63870 | epoch 0 | loss 3.3443 | lr 7.88e-04 | grad 0.2430 +2026-04-10 10:21:34 - INFO - parrotllm.training - step 63880 | epoch 0 | loss 3.4037 | lr 7.88e-04 | grad 0.2521 +2026-04-10 10:21:37 - INFO - parrotllm.training - step 63890 | epoch 0 | loss 3.4044 | lr 7.88e-04 | grad 0.2469 +2026-04-10 10:21:40 - INFO - parrotllm.training - step 63900 | epoch 0 | loss 3.2445 | lr 7.88e-04 | grad 0.3017 +2026-04-10 10:21:43 - INFO - parrotllm.training - step 63910 | epoch 0 | loss 3.4727 | lr 7.88e-04 | grad 0.2768 +2026-04-10 10:21:46 - INFO - parrotllm.training - step 63920 | epoch 0 | loss 3.4634 | lr 7.87e-04 | grad 0.2562 +2026-04-10 10:21:49 - INFO - parrotllm.training - step 63930 | epoch 0 | loss 3.3991 | lr 7.87e-04 | grad 0.2961 +2026-04-10 10:21:52 - INFO - parrotllm.training - step 63940 | epoch 0 | loss 3.4656 | lr 7.87e-04 | grad 0.2556 +2026-04-10 10:21:55 - INFO - parrotllm.training - step 63950 | epoch 0 | loss 3.4124 | lr 7.87e-04 | grad 0.2790 +2026-04-10 10:21:58 - INFO - parrotllm.training - step 63960 | epoch 0 | loss 3.3207 | lr 7.87e-04 | grad 0.2758 +2026-04-10 10:22:01 - INFO - parrotllm.training - step 63970 | epoch 0 | loss 3.2844 | lr 7.87e-04 | grad 0.2785 +2026-04-10 10:22:04 - INFO - parrotllm.training - step 63980 | epoch 0 | loss 3.3977 | lr 7.86e-04 | grad 0.2516 +2026-04-10 10:22:07 - INFO - parrotllm.training - step 63990 | epoch 0 | loss 3.4492 | lr 7.86e-04 | grad 0.2704 +2026-04-10 10:22:10 - INFO - parrotllm.training - step 64000 | epoch 0 | loss 3.5149 | lr 7.86e-04 | grad 0.2813 +2026-04-10 10:22:10 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:22:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:22:13 - INFO - parrotllm.training - Train: loss=3.5149, ppl=33.61 +2026-04-10 10:22:13 - INFO - parrotllm.training - Val: loss=3.3331, ppl=28.02 +2026-04-10 10:22:14 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3331_epoch_0000_step_0064000.pt +2026-04-10 10:22:16 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3333, min_delta=0.001000). +2026-04-10 10:22:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:22:19 - INFO - parrotllm.training - step 64010 | epoch 0 | loss 3.3976 | lr 7.86e-04 | grad 0.2279 +2026-04-10 10:22:22 - INFO - parrotllm.training - step 64020 | epoch 0 | loss 3.4811 | lr 7.86e-04 | grad 0.2395 +2026-04-10 10:22:25 - INFO - parrotllm.training - step 64030 | epoch 0 | loss 3.3779 | lr 7.86e-04 | grad 0.2653 +2026-04-10 10:22:28 - INFO - parrotllm.training - step 64040 | epoch 0 | loss 3.2416 | lr 7.85e-04 | grad 0.2609 +2026-04-10 10:22:31 - INFO - parrotllm.training - step 64050 | epoch 0 | loss 3.3506 | lr 7.85e-04 | grad 0.2461 +2026-04-10 10:22:34 - INFO - parrotllm.training - step 64060 | epoch 0 | loss 3.3080 | lr 7.85e-04 | grad 0.2589 +2026-04-10 10:22:38 - INFO - parrotllm.training - step 64070 | epoch 0 | loss 3.3618 | lr 7.85e-04 | grad 0.2483 +2026-04-10 10:22:41 - INFO - parrotllm.training - step 64080 | epoch 0 | loss 3.4734 | lr 7.85e-04 | grad 0.2792 +2026-04-10 10:22:44 - INFO - parrotllm.training - step 64090 | epoch 0 | loss 3.3862 | lr 7.84e-04 | grad 0.2623 +2026-04-10 10:22:47 - INFO - parrotllm.training - step 64100 | epoch 0 | loss 3.3858 | lr 7.84e-04 | grad 0.2308 +2026-04-10 10:22:50 - INFO - parrotllm.training - step 64110 | epoch 0 | loss 3.3283 | lr 7.84e-04 | grad 0.2421 +2026-04-10 10:22:53 - INFO - parrotllm.training - step 64120 | epoch 0 | loss 3.3880 | lr 7.84e-04 | grad 0.2443 +2026-04-10 10:22:56 - INFO - parrotllm.training - step 64130 | epoch 0 | loss 3.3230 | lr 7.84e-04 | grad 0.2898 +2026-04-10 10:22:59 - INFO - parrotllm.training - step 64140 | epoch 0 | loss 3.4006 | lr 7.84e-04 | grad 0.2500 +2026-04-10 10:23:02 - INFO - parrotllm.training - step 64150 | epoch 0 | loss 3.3643 | lr 7.83e-04 | grad 0.2771 +2026-04-10 10:23:05 - INFO - parrotllm.training - step 64160 | epoch 0 | loss 3.3536 | lr 7.83e-04 | grad 0.2545 +2026-04-10 10:23:08 - INFO - parrotllm.training - step 64170 | epoch 0 | loss 3.3442 | lr 7.83e-04 | grad 0.2436 +2026-04-10 10:23:11 - INFO - parrotllm.training - step 64180 | epoch 0 | loss 3.3827 | lr 7.83e-04 | grad 0.2438 +2026-04-10 10:23:14 - INFO - parrotllm.training - step 64190 | epoch 0 | loss 3.3453 | lr 7.83e-04 | grad 0.2379 +2026-04-10 10:23:17 - INFO - parrotllm.training - step 64200 | epoch 0 | loss 3.4129 | lr 7.83e-04 | grad 0.2765 +2026-04-10 10:23:20 - INFO - parrotllm.training - step 64210 | epoch 0 | loss 3.4271 | lr 7.82e-04 | grad 0.2462 +2026-04-10 10:23:23 - INFO - parrotllm.training - step 64220 | epoch 0 | loss 3.4510 | lr 7.82e-04 | grad 0.2585 +2026-04-10 10:23:26 - INFO - parrotllm.training - step 64230 | epoch 0 | loss 3.3924 | lr 7.82e-04 | grad 0.2556 +2026-04-10 10:23:29 - INFO - parrotllm.training - step 64240 | epoch 0 | loss 3.3333 | lr 7.82e-04 | grad 0.2936 +2026-04-10 10:23:33 - INFO - parrotllm.training - step 64250 | epoch 0 | loss 3.3643 | lr 7.82e-04 | grad 0.2411 +2026-04-10 10:23:36 - INFO - parrotllm.training - step 64260 | epoch 0 | loss 3.3174 | lr 7.82e-04 | grad 0.2630 +2026-04-10 10:23:39 - INFO - parrotllm.training - step 64270 | epoch 0 | loss 3.4755 | lr 7.81e-04 | grad 0.2547 +2026-04-10 10:23:42 - INFO - parrotllm.training - step 64280 | epoch 0 | loss 3.3714 | lr 7.81e-04 | grad 0.2374 +2026-04-10 10:23:45 - INFO - parrotllm.training - step 64290 | epoch 0 | loss 3.4911 | lr 7.81e-04 | grad 0.2596 +2026-04-10 10:23:48 - INFO - parrotllm.training - step 64300 | epoch 0 | loss 3.4374 | lr 7.81e-04 | grad 0.2578 +2026-04-10 10:23:51 - INFO - parrotllm.training - step 64310 | epoch 0 | loss 3.3681 | lr 7.81e-04 | grad 0.2659 +2026-04-10 10:23:54 - INFO - parrotllm.training - step 64320 | epoch 0 | loss 3.4334 | lr 7.80e-04 | grad 0.2645 +2026-04-10 10:23:57 - INFO - parrotllm.training - step 64330 | epoch 0 | loss 3.4116 | lr 7.80e-04 | grad 0.2607 +2026-04-10 10:24:00 - INFO - parrotllm.training - step 64340 | epoch 0 | loss 3.4011 | lr 7.80e-04 | grad 0.2608 +2026-04-10 10:24:03 - INFO - parrotllm.training - step 64350 | epoch 0 | loss 3.4833 | lr 7.80e-04 | grad 0.2652 +2026-04-10 10:24:06 - INFO - parrotllm.training - step 64360 | epoch 0 | loss 3.3850 | lr 7.80e-04 | grad 0.2603 +2026-04-10 10:24:09 - INFO - parrotllm.training - step 64370 | epoch 0 | loss 3.3653 | lr 7.80e-04 | grad 0.2465 +2026-04-10 10:24:12 - INFO - parrotllm.training - step 64380 | epoch 0 | loss 3.4822 | lr 7.79e-04 | grad 0.2630 +2026-04-10 10:24:15 - INFO - parrotllm.training - step 64390 | epoch 0 | loss 3.3802 | lr 7.79e-04 | grad 0.2375 +2026-04-10 10:24:18 - INFO - parrotllm.training - step 64400 | epoch 0 | loss 3.3325 | lr 7.79e-04 | grad 0.2414 +2026-04-10 10:24:21 - INFO - parrotllm.training - step 64410 | epoch 0 | loss 3.3123 | lr 7.79e-04 | grad 0.2538 +2026-04-10 10:24:24 - INFO - parrotllm.training - step 64420 | epoch 0 | loss 3.3260 | lr 7.79e-04 | grad 0.2634 +2026-04-10 10:24:28 - INFO - parrotllm.training - step 64430 | epoch 0 | loss 3.3077 | lr 7.79e-04 | grad 0.2274 +2026-04-10 10:24:31 - INFO - parrotllm.training - step 64440 | epoch 0 | loss 3.3360 | lr 7.78e-04 | grad 0.2740 +2026-04-10 10:24:34 - INFO - parrotllm.training - step 64450 | epoch 0 | loss 3.2949 | lr 7.78e-04 | grad 0.2365 +2026-04-10 10:24:37 - INFO - parrotllm.training - step 64460 | epoch 0 | loss 3.4194 | lr 7.78e-04 | grad 0.2568 +2026-04-10 10:24:40 - INFO - parrotllm.training - step 64470 | epoch 0 | loss 3.2943 | lr 7.78e-04 | grad 0.2419 +2026-04-10 10:24:43 - INFO - parrotllm.training - step 64480 | epoch 0 | loss 3.3609 | lr 7.78e-04 | grad 0.2560 +2026-04-10 10:24:46 - INFO - parrotllm.training - step 64490 | epoch 0 | loss 3.3587 | lr 7.77e-04 | grad 0.2373 +2026-04-10 10:24:49 - INFO - parrotllm.training - step 64500 | epoch 0 | loss 3.3586 | lr 7.77e-04 | grad 0.2388 +2026-04-10 10:24:49 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:24:49 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:24:52 - INFO - parrotllm.training - Train: loss=3.3586, ppl=28.75 +2026-04-10 10:24:52 - INFO - parrotllm.training - Val: loss=3.3289, ppl=27.91 +2026-04-10 10:24:52 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:24:53 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3289_epoch_0000_step_0064500.pt +2026-04-10 10:24:55 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:24:58 - INFO - parrotllm.training - step 64510 | epoch 0 | loss 3.2449 | lr 7.77e-04 | grad 0.2477 +2026-04-10 10:25:01 - INFO - parrotllm.training - step 64520 | epoch 0 | loss 3.3814 | lr 7.77e-04 | grad 0.2877 +2026-04-10 10:25:04 - INFO - parrotllm.training - step 64530 | epoch 0 | loss 3.4520 | lr 7.77e-04 | grad 0.2388 +2026-04-10 10:25:07 - INFO - parrotllm.training - step 64540 | epoch 0 | loss 3.3590 | lr 7.77e-04 | grad 0.2587 +2026-04-10 10:25:10 - INFO - parrotllm.training - step 64550 | epoch 0 | loss 3.4845 | lr 7.76e-04 | grad 0.2488 +2026-04-10 10:25:13 - INFO - parrotllm.training - step 64560 | epoch 0 | loss 3.4703 | lr 7.76e-04 | grad 0.2548 +2026-04-10 10:25:16 - INFO - parrotllm.training - step 64570 | epoch 0 | loss 3.3605 | lr 7.76e-04 | grad 0.2769 +2026-04-10 10:25:19 - INFO - parrotllm.training - step 64580 | epoch 0 | loss 3.2972 | lr 7.76e-04 | grad 0.2426 +2026-04-10 10:25:22 - INFO - parrotllm.training - step 64590 | epoch 0 | loss 3.4397 | lr 7.76e-04 | grad 0.2568 +2026-04-10 10:25:25 - INFO - parrotllm.training - step 64600 | epoch 0 | loss 3.4499 | lr 7.76e-04 | grad 0.2678 +2026-04-10 10:25:28 - INFO - parrotllm.training - step 64610 | epoch 0 | loss 3.3867 | lr 7.75e-04 | grad 0.2536 +2026-04-10 10:25:31 - INFO - parrotllm.training - step 64620 | epoch 0 | loss 3.3639 | lr 7.75e-04 | grad 0.2675 +2026-04-10 10:25:34 - INFO - parrotllm.training - step 64630 | epoch 0 | loss 3.4441 | lr 7.75e-04 | grad 0.2448 +2026-04-10 10:25:38 - INFO - parrotllm.training - step 64640 | epoch 0 | loss 3.4391 | lr 7.75e-04 | grad 0.3307 +2026-04-10 10:25:41 - INFO - parrotllm.training - step 64650 | epoch 0 | loss 3.4066 | lr 7.75e-04 | grad 0.2933 +2026-04-10 10:25:44 - INFO - parrotllm.training - step 64660 | epoch 0 | loss 3.4085 | lr 7.74e-04 | grad 0.2535 +2026-04-10 10:25:47 - INFO - parrotllm.training - step 64670 | epoch 0 | loss 3.3357 | lr 7.74e-04 | grad 0.2385 +2026-04-10 10:25:50 - INFO - parrotllm.training - step 64680 | epoch 0 | loss 3.5007 | lr 7.74e-04 | grad 0.2547 +2026-04-10 10:25:53 - INFO - parrotllm.training - step 64690 | epoch 0 | loss 3.4488 | lr 7.74e-04 | grad 0.2870 +2026-04-10 10:25:56 - INFO - parrotllm.training - step 64700 | epoch 0 | loss 3.4112 | lr 7.74e-04 | grad 0.2898 +2026-04-10 10:25:59 - INFO - parrotllm.training - step 64710 | epoch 0 | loss 3.4451 | lr 7.74e-04 | grad 0.2422 +2026-04-10 10:26:02 - INFO - parrotllm.training - step 64720 | epoch 0 | loss 3.4633 | lr 7.73e-04 | grad 0.2252 +2026-04-10 10:26:05 - INFO - parrotllm.training - step 64730 | epoch 0 | loss 3.3391 | lr 7.73e-04 | grad 0.2576 +2026-04-10 10:26:08 - INFO - parrotllm.training - step 64740 | epoch 0 | loss 3.4417 | lr 7.73e-04 | grad 0.2494 +2026-04-10 10:26:11 - INFO - parrotllm.training - step 64750 | epoch 0 | loss 3.4195 | lr 7.73e-04 | grad 0.2375 +2026-04-10 10:26:14 - INFO - parrotllm.training - step 64760 | epoch 0 | loss 3.3714 | lr 7.73e-04 | grad 0.2854 +2026-04-10 10:26:17 - INFO - parrotllm.training - step 64770 | epoch 0 | loss 3.3381 | lr 7.73e-04 | grad 0.2293 +2026-04-10 10:26:20 - INFO - parrotllm.training - step 64780 | epoch 0 | loss 3.3129 | lr 7.72e-04 | grad 0.2690 +2026-04-10 10:26:23 - INFO - parrotllm.training - step 64790 | epoch 0 | loss 3.3207 | lr 7.72e-04 | grad 0.2793 +2026-04-10 10:26:26 - INFO - parrotllm.training - step 64800 | epoch 0 | loss 3.4302 | lr 7.72e-04 | grad 0.2482 +2026-04-10 10:26:29 - INFO - parrotllm.training - step 64810 | epoch 0 | loss 3.4027 | lr 7.72e-04 | grad 0.2465 +2026-04-10 10:26:33 - INFO - parrotllm.training - step 64820 | epoch 0 | loss 3.3777 | lr 7.72e-04 | grad 0.2722 +2026-04-10 10:26:36 - INFO - parrotllm.training - step 64830 | epoch 0 | loss 3.3709 | lr 7.72e-04 | grad 0.2683 +2026-04-10 10:26:39 - INFO - parrotllm.training - step 64840 | epoch 0 | loss 3.3680 | lr 7.71e-04 | grad 0.2654 +2026-04-10 10:26:42 - INFO - parrotllm.training - step 64850 | epoch 0 | loss 3.4632 | lr 7.71e-04 | grad 0.2458 +2026-04-10 10:26:45 - INFO - parrotllm.training - step 64860 | epoch 0 | loss 3.4871 | lr 7.71e-04 | grad 0.2563 +2026-04-10 10:26:48 - INFO - parrotllm.training - step 64870 | epoch 0 | loss 3.4450 | lr 7.71e-04 | grad 0.2519 +2026-04-10 10:26:51 - INFO - parrotllm.training - step 64880 | epoch 0 | loss 3.4397 | lr 7.71e-04 | grad 0.2448 +2026-04-10 10:26:54 - INFO - parrotllm.training - step 64890 | epoch 0 | loss 3.4408 | lr 7.70e-04 | grad 0.2616 +2026-04-10 10:26:57 - INFO - parrotllm.training - step 64900 | epoch 0 | loss 3.3137 | lr 7.70e-04 | grad 0.2637 +2026-04-10 10:27:00 - INFO - parrotllm.training - step 64910 | epoch 0 | loss 3.4340 | lr 7.70e-04 | grad 0.2411 +2026-04-10 10:27:03 - INFO - parrotllm.training - step 64920 | epoch 0 | loss 3.4339 | lr 7.70e-04 | grad 0.2701 +2026-04-10 10:27:06 - INFO - parrotllm.training - step 64930 | epoch 0 | loss 3.3299 | lr 7.70e-04 | grad 0.2374 +2026-04-10 10:27:09 - INFO - parrotllm.training - step 64940 | epoch 0 | loss 3.3394 | lr 7.70e-04 | grad 0.2626 +2026-04-10 10:27:12 - INFO - parrotllm.training - step 64950 | epoch 0 | loss 3.4903 | lr 7.69e-04 | grad 0.2720 +2026-04-10 10:27:15 - INFO - parrotllm.training - step 64960 | epoch 0 | loss 3.4225 | lr 7.69e-04 | grad 0.2530 +2026-04-10 10:27:18 - INFO - parrotllm.training - step 64970 | epoch 0 | loss 3.4355 | lr 7.69e-04 | grad 0.2650 +2026-04-10 10:27:21 - INFO - parrotllm.training - step 64980 | epoch 0 | loss 3.3937 | lr 7.69e-04 | grad 0.2369 +2026-04-10 10:27:24 - INFO - parrotllm.training - step 64990 | epoch 0 | loss 3.4471 | lr 7.69e-04 | grad 0.2576 +2026-04-10 10:27:27 - INFO - parrotllm.training - step 65000 | epoch 0 | loss 3.3546 | lr 7.69e-04 | grad 0.2502 +2026-04-10 10:27:27 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:27:27 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:27:31 - INFO - parrotllm.training - Train: loss=3.3546, ppl=28.63 +2026-04-10 10:27:31 - INFO - parrotllm.training - Val: loss=3.3324, ppl=28.01 +2026-04-10 10:27:32 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3324_epoch_0000_step_0065000.pt +2026-04-10 10:27:33 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3289, min_delta=0.001000). +2026-04-10 10:27:33 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:27:34 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0065000.pt +2026-04-10 10:27:39 - INFO - parrotllm.training - step 65010 | epoch 0 | loss 3.2971 | lr 7.68e-04 | grad 0.2605 +2026-04-10 10:27:42 - INFO - parrotllm.training - step 65020 | epoch 0 | loss 3.3589 | lr 7.68e-04 | grad 0.2527 +2026-04-10 10:27:45 - INFO - parrotllm.training - step 65030 | epoch 0 | loss 3.3661 | lr 7.68e-04 | grad 0.2598 +2026-04-10 10:27:49 - INFO - parrotllm.training - step 65040 | epoch 0 | loss 3.3711 | lr 7.68e-04 | grad 0.2665 +2026-04-10 10:27:52 - INFO - parrotllm.training - step 65050 | epoch 0 | loss 3.1863 | lr 7.68e-04 | grad 0.2510 +2026-04-10 10:27:55 - INFO - parrotllm.training - step 65060 | epoch 0 | loss 3.4144 | lr 7.67e-04 | grad 0.2403 +2026-04-10 10:27:58 - INFO - parrotllm.training - step 65070 | epoch 0 | loss 3.4547 | lr 7.67e-04 | grad 0.2738 +2026-04-10 10:28:01 - INFO - parrotllm.training - step 65080 | epoch 0 | loss 3.3937 | lr 7.67e-04 | grad 0.3252 +2026-04-10 10:28:04 - INFO - parrotllm.training - step 65090 | epoch 0 | loss 3.4841 | lr 7.67e-04 | grad 0.2596 +2026-04-10 10:28:07 - INFO - parrotllm.training - step 65100 | epoch 0 | loss 3.4893 | lr 7.67e-04 | grad 0.2520 +2026-04-10 10:28:10 - INFO - parrotllm.training - step 65110 | epoch 0 | loss 3.4428 | lr 7.67e-04 | grad 0.2704 +2026-04-10 10:28:13 - INFO - parrotllm.training - step 65120 | epoch 0 | loss 3.3677 | lr 7.66e-04 | grad 0.2657 +2026-04-10 10:28:16 - INFO - parrotllm.training - step 65130 | epoch 0 | loss 3.3880 | lr 7.66e-04 | grad 0.2489 +2026-04-10 10:28:19 - INFO - parrotllm.training - step 65140 | epoch 0 | loss 3.3649 | lr 7.66e-04 | grad 0.2491 +2026-04-10 10:28:22 - INFO - parrotllm.training - step 65150 | epoch 0 | loss 3.4120 | lr 7.66e-04 | grad 0.2423 +2026-04-10 10:28:25 - INFO - parrotllm.training - step 65160 | epoch 0 | loss 3.2710 | lr 7.66e-04 | grad 0.2606 +2026-04-10 10:28:28 - INFO - parrotllm.training - step 65170 | epoch 0 | loss 3.3973 | lr 7.66e-04 | grad 0.2434 +2026-04-10 10:28:31 - INFO - parrotllm.training - step 65180 | epoch 0 | loss 3.3947 | lr 7.65e-04 | grad 0.2730 +2026-04-10 10:28:34 - INFO - parrotllm.training - step 65190 | epoch 0 | loss 3.3606 | lr 7.65e-04 | grad 0.2303 +2026-04-10 10:28:37 - INFO - parrotllm.training - step 65200 | epoch 0 | loss 3.4832 | lr 7.65e-04 | grad 0.2384 +2026-04-10 10:28:40 - INFO - parrotllm.training - step 65210 | epoch 0 | loss 3.3138 | lr 7.65e-04 | grad 0.2485 +2026-04-10 10:28:43 - INFO - parrotllm.training - step 65220 | epoch 0 | loss 3.3899 | lr 7.65e-04 | grad 0.2908 +2026-04-10 10:28:47 - INFO - parrotllm.training - step 65230 | epoch 0 | loss 3.4029 | lr 7.64e-04 | grad 0.2365 +2026-04-10 10:28:50 - INFO - parrotllm.training - step 65240 | epoch 0 | loss 3.3781 | lr 7.64e-04 | grad 0.2577 +2026-04-10 10:28:53 - INFO - parrotllm.training - step 65250 | epoch 0 | loss 3.4329 | lr 7.64e-04 | grad 0.2484 +2026-04-10 10:28:56 - INFO - parrotllm.training - step 65260 | epoch 0 | loss 3.3518 | lr 7.64e-04 | grad 0.2543 +2026-04-10 10:28:59 - INFO - parrotllm.training - step 65270 | epoch 0 | loss 3.4106 | lr 7.64e-04 | grad 0.2300 +2026-04-10 10:29:02 - INFO - parrotllm.training - step 65280 | epoch 0 | loss 3.2602 | lr 7.64e-04 | grad 0.2486 +2026-04-10 10:29:05 - INFO - parrotllm.training - step 65290 | epoch 0 | loss 3.4106 | lr 7.63e-04 | grad 0.2321 +2026-04-10 10:29:08 - INFO - parrotllm.training - step 65300 | epoch 0 | loss 3.3270 | lr 7.63e-04 | grad 0.2628 +2026-04-10 10:29:11 - INFO - parrotllm.training - step 65310 | epoch 0 | loss 3.3764 | lr 7.63e-04 | grad 0.2930 +2026-04-10 10:29:14 - INFO - parrotllm.training - step 65320 | epoch 0 | loss 3.3429 | lr 7.63e-04 | grad 0.2265 +2026-04-10 10:29:17 - INFO - parrotllm.training - step 65330 | epoch 0 | loss 3.3570 | lr 7.63e-04 | grad 0.2696 +2026-04-10 10:29:20 - INFO - parrotllm.training - step 65340 | epoch 0 | loss 3.3240 | lr 7.63e-04 | grad 0.2726 +2026-04-10 10:29:23 - INFO - parrotllm.training - step 65350 | epoch 0 | loss 3.3573 | lr 7.62e-04 | grad 0.2788 +2026-04-10 10:29:26 - INFO - parrotllm.training - step 65360 | epoch 0 | loss 3.3759 | lr 7.62e-04 | grad 0.2627 +2026-04-10 10:29:29 - INFO - parrotllm.training - step 65370 | epoch 0 | loss 3.4295 | lr 7.62e-04 | grad 0.2729 +2026-04-10 10:29:32 - INFO - parrotllm.training - step 65380 | epoch 0 | loss 3.3393 | lr 7.62e-04 | grad 0.2282 +2026-04-10 10:29:35 - INFO - parrotllm.training - step 65390 | epoch 0 | loss 3.3581 | lr 7.62e-04 | grad 0.2344 +2026-04-10 10:29:38 - INFO - parrotllm.training - step 65400 | epoch 0 | loss 3.2858 | lr 7.62e-04 | grad 0.2557 +2026-04-10 10:29:41 - INFO - parrotllm.training - step 65410 | epoch 0 | loss 3.3081 | lr 7.61e-04 | grad 0.2370 +2026-04-10 10:29:45 - INFO - parrotllm.training - step 65420 | epoch 0 | loss 3.2755 | lr 7.61e-04 | grad 0.2525 +2026-04-10 10:29:48 - INFO - parrotllm.training - step 65430 | epoch 0 | loss 3.2191 | lr 7.61e-04 | grad 0.2444 +2026-04-10 10:29:51 - INFO - parrotllm.training - step 65440 | epoch 0 | loss 3.3367 | lr 7.61e-04 | grad 0.2586 +2026-04-10 10:29:54 - INFO - parrotllm.training - step 65450 | epoch 0 | loss 3.4009 | lr 7.61e-04 | grad 0.2412 +2026-04-10 10:29:57 - INFO - parrotllm.training - step 65460 | epoch 0 | loss 3.4112 | lr 7.60e-04 | grad 0.2705 +2026-04-10 10:30:00 - INFO - parrotllm.training - step 65470 | epoch 0 | loss 3.3384 | lr 7.60e-04 | grad 0.2829 +2026-04-10 10:30:03 - INFO - parrotllm.training - step 65480 | epoch 0 | loss 3.2916 | lr 7.60e-04 | grad 0.2491 +2026-04-10 10:30:06 - INFO - parrotllm.training - step 65490 | epoch 0 | loss 3.4047 | lr 7.60e-04 | grad 0.2674 +2026-04-10 10:30:09 - INFO - parrotllm.training - step 65500 | epoch 0 | loss 3.3681 | lr 7.60e-04 | grad 0.2545 +2026-04-10 10:30:09 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:30:09 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:30:12 - INFO - parrotllm.training - Train: loss=3.3681, ppl=29.02 +2026-04-10 10:30:12 - INFO - parrotllm.training - Val: loss=3.3303, ppl=27.95 +2026-04-10 10:30:13 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3303_epoch_0000_step_0065500.pt +2026-04-10 10:30:15 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3289, min_delta=0.001000). +2026-04-10 10:30:15 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:30:18 - INFO - parrotllm.training - step 65510 | epoch 0 | loss 3.4450 | lr 7.60e-04 | grad 0.2910 +2026-04-10 10:30:21 - INFO - parrotllm.training - step 65520 | epoch 0 | loss 3.3462 | lr 7.59e-04 | grad 0.2444 +2026-04-10 10:30:24 - INFO - parrotllm.training - step 65530 | epoch 0 | loss 3.3957 | lr 7.59e-04 | grad 0.2483 +2026-04-10 10:30:27 - INFO - parrotllm.training - step 65540 | epoch 0 | loss 3.3811 | lr 7.59e-04 | grad 0.2373 +2026-04-10 10:30:30 - INFO - parrotllm.training - step 65550 | epoch 0 | loss 3.3053 | lr 7.59e-04 | grad 0.2719 +2026-04-10 10:30:33 - INFO - parrotllm.training - step 65560 | epoch 0 | loss 3.3991 | lr 7.59e-04 | grad 0.2679 +2026-04-10 10:30:36 - INFO - parrotllm.training - step 65570 | epoch 0 | loss 3.3885 | lr 7.59e-04 | grad 0.2400 +2026-04-10 10:30:39 - INFO - parrotllm.training - step 65580 | epoch 0 | loss 3.4321 | lr 7.58e-04 | grad 0.2442 +2026-04-10 10:30:42 - INFO - parrotllm.training - step 65590 | epoch 0 | loss 3.2820 | lr 7.58e-04 | grad 0.2533 +2026-04-10 10:30:45 - INFO - parrotllm.training - step 65600 | epoch 0 | loss 3.3780 | lr 7.58e-04 | grad 0.2760 +2026-04-10 10:30:49 - INFO - parrotllm.training - step 65610 | epoch 0 | loss 3.3785 | lr 7.58e-04 | grad 0.2304 +2026-04-10 10:30:52 - INFO - parrotllm.training - step 65620 | epoch 0 | loss 3.4519 | lr 7.58e-04 | grad 0.2853 +2026-04-10 10:30:55 - INFO - parrotllm.training - step 65630 | epoch 0 | loss 3.4183 | lr 7.57e-04 | grad 0.2534 +2026-04-10 10:30:58 - INFO - parrotllm.training - step 65640 | epoch 0 | loss 3.3581 | lr 7.57e-04 | grad 0.2497 +2026-04-10 10:31:01 - INFO - parrotllm.training - step 65650 | epoch 0 | loss 3.4635 | lr 7.57e-04 | grad 0.2587 +2026-04-10 10:31:04 - INFO - parrotllm.training - step 65660 | epoch 0 | loss 3.3553 | lr 7.57e-04 | grad 0.2599 +2026-04-10 10:31:07 - INFO - parrotllm.training - step 65670 | epoch 0 | loss 3.2980 | lr 7.57e-04 | grad 0.2728 +2026-04-10 10:31:10 - INFO - parrotllm.training - step 65680 | epoch 0 | loss 3.3507 | lr 7.57e-04 | grad 0.2547 +2026-04-10 10:31:13 - INFO - parrotllm.training - step 65690 | epoch 0 | loss 3.4199 | lr 7.56e-04 | grad 0.2606 +2026-04-10 10:31:16 - INFO - parrotllm.training - step 65700 | epoch 0 | loss 3.2803 | lr 7.56e-04 | grad 0.2392 +2026-04-10 10:31:19 - INFO - parrotllm.training - step 65710 | epoch 0 | loss 3.3040 | lr 7.56e-04 | grad 0.2421 +2026-04-10 10:31:22 - INFO - parrotllm.training - step 65720 | epoch 0 | loss 3.3581 | lr 7.56e-04 | grad 0.2483 +2026-04-10 10:31:25 - INFO - parrotllm.training - step 65730 | epoch 0 | loss 3.3344 | lr 7.56e-04 | grad 0.2522 +2026-04-10 10:31:28 - INFO - parrotllm.training - step 65740 | epoch 0 | loss 3.3796 | lr 7.56e-04 | grad 0.2488 +2026-04-10 10:31:31 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 65775/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 10:31:32 - INFO - parrotllm.training - step 65750 | epoch 0 | loss 3.3442 | lr 7.55e-04 | grad 0.2426 +2026-04-10 10:31:35 - INFO - parrotllm.training - step 65760 | epoch 0 | loss 3.3512 | lr 7.55e-04 | grad 0.2387 +2026-04-10 10:31:38 - INFO - parrotllm.training - step 65770 | epoch 0 | loss 3.3336 | lr 7.55e-04 | grad 0.2562 +2026-04-10 10:31:41 - INFO - parrotllm.training - step 65780 | epoch 0 | loss 3.3986 | lr 7.55e-04 | grad 0.2806 +2026-04-10 10:31:44 - INFO - parrotllm.training - step 65790 | epoch 0 | loss 3.4315 | lr 7.55e-04 | grad 0.2745 +2026-04-10 10:31:47 - INFO - parrotllm.training - step 65800 | epoch 0 | loss 3.3640 | lr 7.55e-04 | grad 0.3088 +2026-04-10 10:31:50 - INFO - parrotllm.training - step 65810 | epoch 0 | loss 3.4612 | lr 7.54e-04 | grad 0.2412 +2026-04-10 10:31:53 - INFO - parrotllm.training - step 65820 | epoch 0 | loss 3.3610 | lr 7.54e-04 | grad 0.2369 +2026-04-10 10:31:56 - INFO - parrotllm.training - step 65830 | epoch 0 | loss 3.3754 | lr 7.54e-04 | grad 0.2736 +2026-04-10 10:31:59 - INFO - parrotllm.training - step 65840 | epoch 0 | loss 3.3821 | lr 7.54e-04 | grad 0.2598 +2026-04-10 10:32:02 - INFO - parrotllm.training - step 65850 | epoch 0 | loss 3.4259 | lr 7.54e-04 | grad 0.2710 +2026-04-10 10:32:05 - INFO - parrotllm.training - step 65860 | epoch 0 | loss 3.3707 | lr 7.53e-04 | grad 0.2355 +2026-04-10 10:32:08 - INFO - parrotllm.training - step 65870 | epoch 0 | loss 3.3489 | lr 7.53e-04 | grad 0.2751 +2026-04-10 10:32:11 - INFO - parrotllm.training - step 65880 | epoch 0 | loss 3.4431 | lr 7.53e-04 | grad 0.2663 +2026-04-10 10:32:14 - INFO - parrotllm.training - step 65890 | epoch 0 | loss 3.4529 | lr 7.53e-04 | grad 0.2847 +2026-04-10 10:32:17 - INFO - parrotllm.training - step 65900 | epoch 0 | loss 3.3011 | lr 7.53e-04 | grad 0.2957 +2026-04-10 10:32:21 - INFO - parrotllm.training - step 65910 | epoch 0 | loss 3.3449 | lr 7.53e-04 | grad 0.2685 +2026-04-10 10:32:24 - INFO - parrotllm.training - step 65920 | epoch 0 | loss 3.3489 | lr 7.52e-04 | grad 0.2654 +2026-04-10 10:32:27 - INFO - parrotllm.training - step 65930 | epoch 0 | loss 3.3497 | lr 7.52e-04 | grad 0.2612 +2026-04-10 10:32:30 - INFO - parrotllm.training - step 65940 | epoch 0 | loss 3.3509 | lr 7.52e-04 | grad 0.2265 +2026-04-10 10:32:33 - INFO - parrotllm.training - step 65950 | epoch 0 | loss 3.4272 | lr 7.52e-04 | grad 0.2654 +2026-04-10 10:32:36 - INFO - parrotllm.training - step 65960 | epoch 0 | loss 3.4072 | lr 7.52e-04 | grad 0.2334 +2026-04-10 10:32:39 - INFO - parrotllm.training - step 65970 | epoch 0 | loss 3.3380 | lr 7.52e-04 | grad 0.2646 +2026-04-10 10:32:42 - INFO - parrotllm.training - step 65980 | epoch 0 | loss 3.4217 | lr 7.51e-04 | grad 0.2819 +2026-04-10 10:32:45 - INFO - parrotllm.training - step 65990 | epoch 0 | loss 3.2827 | lr 7.51e-04 | grad 0.2555 +2026-04-10 10:32:48 - INFO - parrotllm.training - step 66000 | epoch 0 | loss 3.4339 | lr 7.51e-04 | grad 0.2611 +2026-04-10 10:32:48 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:32:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:32:51 - INFO - parrotllm.training - Train: loss=3.4339, ppl=31.00 +2026-04-10 10:32:51 - INFO - parrotllm.training - Val: loss=3.3294, ppl=27.92 +2026-04-10 10:32:52 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3294_epoch_0000_step_0066000.pt +2026-04-10 10:32:54 - INFO - parrotllm.training - No validation improvement for 3/15 evaluation(s) (best=3.3289, min_delta=0.001000). +2026-04-10 10:32:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:32:57 - INFO - parrotllm.training - step 66010 | epoch 0 | loss 3.3772 | lr 7.51e-04 | grad 0.2510 +2026-04-10 10:33:00 - INFO - parrotllm.training - step 66020 | epoch 0 | loss 3.4270 | lr 7.51e-04 | grad 0.2365 +2026-04-10 10:33:03 - INFO - parrotllm.training - step 66030 | epoch 0 | loss 3.3819 | lr 7.50e-04 | grad 0.2375 +2026-04-10 10:33:06 - INFO - parrotllm.training - step 66040 | epoch 0 | loss 3.5122 | lr 7.50e-04 | grad 0.2764 +2026-04-10 10:33:09 - INFO - parrotllm.training - step 66050 | epoch 0 | loss 3.3894 | lr 7.50e-04 | grad 0.2684 +2026-04-10 10:33:12 - INFO - parrotllm.training - step 66060 | epoch 0 | loss 3.3234 | lr 7.50e-04 | grad 0.2829 +2026-04-10 10:33:15 - INFO - parrotllm.training - step 66070 | epoch 0 | loss 3.3203 | lr 7.50e-04 | grad 0.2513 +2026-04-10 10:33:18 - INFO - parrotllm.training - step 66080 | epoch 0 | loss 3.4225 | lr 7.50e-04 | grad 0.2514 +2026-04-10 10:33:21 - INFO - parrotllm.training - step 66090 | epoch 0 | loss 3.3511 | lr 7.49e-04 | grad 0.2565 +2026-04-10 10:33:24 - INFO - parrotllm.training - step 66100 | epoch 0 | loss 3.4348 | lr 7.49e-04 | grad 0.2937 +2026-04-10 10:33:27 - INFO - parrotllm.training - step 66110 | epoch 0 | loss 3.3441 | lr 7.49e-04 | grad 0.2511 +2026-04-10 10:33:30 - INFO - parrotllm.training - step 66120 | epoch 0 | loss 3.4617 | lr 7.49e-04 | grad 0.2394 +2026-04-10 10:33:33 - INFO - parrotllm.training - step 66130 | epoch 0 | loss 3.3664 | lr 7.49e-04 | grad 0.2514 +2026-04-10 10:33:36 - INFO - parrotllm.training - step 66140 | epoch 0 | loss 3.3707 | lr 7.49e-04 | grad 0.2700 +2026-04-10 10:33:39 - INFO - parrotllm.training - step 66150 | epoch 0 | loss 3.3261 | lr 7.48e-04 | grad 0.2633 +2026-04-10 10:33:42 - INFO - parrotllm.training - step 66160 | epoch 0 | loss 3.3833 | lr 7.48e-04 | grad 0.2790 +2026-04-10 10:33:45 - INFO - parrotllm.training - step 66170 | epoch 0 | loss 3.2442 | lr 7.48e-04 | grad 0.2340 +2026-04-10 10:33:48 - INFO - parrotllm.training - step 66180 | epoch 0 | loss 3.4091 | lr 7.48e-04 | grad 0.2402 +2026-04-10 10:33:52 - INFO - parrotllm.training - step 66190 | epoch 0 | loss 3.3865 | lr 7.48e-04 | grad 0.2569 +2026-04-10 10:33:55 - INFO - parrotllm.training - step 66200 | epoch 0 | loss 3.4550 | lr 7.48e-04 | grad 0.2716 +2026-04-10 10:33:58 - INFO - parrotllm.training - step 66210 | epoch 0 | loss 3.2928 | lr 7.47e-04 | grad 0.2862 +2026-04-10 10:34:01 - INFO - parrotllm.training - step 66220 | epoch 0 | loss 3.3505 | lr 7.47e-04 | grad 0.2767 +2026-04-10 10:34:04 - INFO - parrotllm.training - step 66230 | epoch 0 | loss 3.3710 | lr 7.47e-04 | grad 0.2457 +2026-04-10 10:34:07 - INFO - parrotllm.training - step 66240 | epoch 0 | loss 3.3924 | lr 7.47e-04 | grad 0.2418 +2026-04-10 10:34:10 - INFO - parrotllm.training - step 66250 | epoch 0 | loss 3.4212 | lr 7.47e-04 | grad 0.2640 +2026-04-10 10:34:13 - INFO - parrotllm.training - step 66260 | epoch 0 | loss 3.4255 | lr 7.46e-04 | grad 0.2811 +2026-04-10 10:34:16 - INFO - parrotllm.training - step 66270 | epoch 0 | loss 3.3663 | lr 7.46e-04 | grad 0.2598 +2026-04-10 10:34:19 - INFO - parrotllm.training - step 66280 | epoch 0 | loss 3.4171 | lr 7.46e-04 | grad 0.3051 +2026-04-10 10:34:22 - INFO - parrotllm.training - step 66290 | epoch 0 | loss 3.3581 | lr 7.46e-04 | grad 0.2755 +2026-04-10 10:34:25 - INFO - parrotllm.training - step 66300 | epoch 0 | loss 3.4005 | lr 7.46e-04 | grad 0.2709 +2026-04-10 10:34:28 - INFO - parrotllm.training - step 66310 | epoch 0 | loss 3.3841 | lr 7.46e-04 | grad 0.2693 +2026-04-10 10:34:31 - INFO - parrotllm.training - step 66320 | epoch 0 | loss 3.4639 | lr 7.45e-04 | grad 0.2689 +2026-04-10 10:34:34 - INFO - parrotllm.training - step 66330 | epoch 0 | loss 3.3765 | lr 7.45e-04 | grad 0.2372 +2026-04-10 10:34:37 - INFO - parrotllm.training - step 66340 | epoch 0 | loss 3.3910 | lr 7.45e-04 | grad 0.2456 +2026-04-10 10:34:40 - INFO - parrotllm.training - step 66350 | epoch 0 | loss 3.3437 | lr 7.45e-04 | grad 0.2419 +2026-04-10 10:34:43 - INFO - parrotllm.training - step 66360 | epoch 0 | loss 3.4544 | lr 7.45e-04 | grad 0.2554 +2026-04-10 10:34:46 - INFO - parrotllm.training - step 66370 | epoch 0 | loss 3.3427 | lr 7.45e-04 | grad 0.2189 +2026-04-10 10:34:50 - INFO - parrotllm.training - step 66380 | epoch 0 | loss 3.4949 | lr 7.44e-04 | grad 0.2536 +2026-04-10 10:34:53 - INFO - parrotllm.training - step 66390 | epoch 0 | loss 3.4116 | lr 7.44e-04 | grad 0.2414 +2026-04-10 10:34:56 - INFO - parrotllm.training - step 66400 | epoch 0 | loss 3.4046 | lr 7.44e-04 | grad 0.2468 +2026-04-10 10:34:59 - INFO - parrotllm.training - step 66410 | epoch 0 | loss 3.3433 | lr 7.44e-04 | grad 0.2385 +2026-04-10 10:35:02 - INFO - parrotllm.training - step 66420 | epoch 0 | loss 3.3077 | lr 7.44e-04 | grad 0.2633 +2026-04-10 10:35:05 - INFO - parrotllm.training - step 66430 | epoch 0 | loss 3.4405 | lr 7.44e-04 | grad 0.2533 +2026-04-10 10:35:08 - INFO - parrotllm.training - step 66440 | epoch 0 | loss 3.4083 | lr 7.43e-04 | grad 0.2549 +2026-04-10 10:35:11 - INFO - parrotllm.training - step 66450 | epoch 0 | loss 3.3095 | lr 7.43e-04 | grad 0.2465 +2026-04-10 10:35:14 - INFO - parrotllm.training - step 66460 | epoch 0 | loss 3.5182 | lr 7.43e-04 | grad 0.2526 +2026-04-10 10:35:17 - INFO - parrotllm.training - step 66470 | epoch 0 | loss 3.4257 | lr 7.43e-04 | grad 0.2571 +2026-04-10 10:35:20 - INFO - parrotllm.training - step 66480 | epoch 0 | loss 3.3429 | lr 7.43e-04 | grad 0.2766 +2026-04-10 10:35:23 - INFO - parrotllm.training - step 66490 | epoch 0 | loss 3.2853 | lr 7.42e-04 | grad 0.2390 +2026-04-10 10:35:26 - INFO - parrotllm.training - step 66500 | epoch 0 | loss 3.3975 | lr 7.42e-04 | grad 0.2743 +2026-04-10 10:35:26 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:35:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:35:29 - INFO - parrotllm.training - Train: loss=3.3975, ppl=29.89 +2026-04-10 10:35:29 - INFO - parrotllm.training - Val: loss=3.3286, ppl=27.90 +2026-04-10 10:35:30 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3286_epoch_0000_step_0066500.pt +2026-04-10 10:35:32 - INFO - parrotllm.training - No validation improvement for 4/15 evaluation(s) (best=3.3289, min_delta=0.001000). +2026-04-10 10:35:32 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:35:35 - INFO - parrotllm.training - step 66510 | epoch 0 | loss 3.4664 | lr 7.42e-04 | grad 0.2442 +2026-04-10 10:35:38 - INFO - parrotllm.training - step 66520 | epoch 0 | loss 3.3022 | lr 7.42e-04 | grad 0.2397 +2026-04-10 10:35:41 - INFO - parrotllm.training - step 66530 | epoch 0 | loss 3.4240 | lr 7.42e-04 | grad 0.2615 +2026-04-10 10:35:44 - INFO - parrotllm.training - step 66540 | epoch 0 | loss 3.3398 | lr 7.42e-04 | grad 0.2785 +2026-04-10 10:35:47 - INFO - parrotllm.training - step 66550 | epoch 0 | loss 3.3995 | lr 7.41e-04 | grad 0.2558 +2026-04-10 10:35:50 - INFO - parrotllm.training - step 66560 | epoch 0 | loss 3.4127 | lr 7.41e-04 | grad 0.2436 +2026-04-10 10:35:53 - INFO - parrotllm.training - step 66570 | epoch 0 | loss 3.4655 | lr 7.41e-04 | grad 0.3466 +2026-04-10 10:35:56 - INFO - parrotllm.training - step 66580 | epoch 0 | loss 3.3981 | lr 7.41e-04 | grad 0.2530 +2026-04-10 10:35:59 - INFO - parrotllm.training - step 66590 | epoch 0 | loss 3.3520 | lr 7.41e-04 | grad 0.2628 +2026-04-10 10:36:02 - INFO - parrotllm.training - step 66600 | epoch 0 | loss 3.3829 | lr 7.41e-04 | grad 0.2536 +2026-04-10 10:36:05 - INFO - parrotllm.training - step 66610 | epoch 0 | loss 3.4066 | lr 7.40e-04 | grad 0.2424 +2026-04-10 10:36:08 - INFO - parrotllm.training - step 66620 | epoch 0 | loss 3.3543 | lr 7.40e-04 | grad 0.2406 +2026-04-10 10:36:12 - INFO - parrotllm.training - step 66630 | epoch 0 | loss 3.3587 | lr 7.40e-04 | grad 0.2609 +2026-04-10 10:36:15 - INFO - parrotllm.training - step 66640 | epoch 0 | loss 3.3125 | lr 7.40e-04 | grad 0.2854 +2026-04-10 10:36:18 - INFO - parrotllm.training - step 66650 | epoch 0 | loss 3.3960 | lr 7.40e-04 | grad 0.2365 +2026-04-10 10:36:21 - INFO - parrotllm.training - step 66660 | epoch 0 | loss 3.4725 | lr 7.39e-04 | grad 0.2402 +2026-04-10 10:36:24 - INFO - parrotllm.training - step 66670 | epoch 0 | loss 3.3422 | lr 7.39e-04 | grad 0.2492 +2026-04-10 10:36:27 - INFO - parrotllm.training - step 66680 | epoch 0 | loss 3.4530 | lr 7.39e-04 | grad 0.2502 +2026-04-10 10:36:30 - INFO - parrotllm.training - step 66690 | epoch 0 | loss 3.5060 | lr 7.39e-04 | grad 0.2613 +2026-04-10 10:36:33 - INFO - parrotllm.training - step 66700 | epoch 0 | loss 3.3350 | lr 7.39e-04 | grad 0.2415 +2026-04-10 10:36:36 - INFO - parrotllm.training - step 66710 | epoch 0 | loss 3.4061 | lr 7.39e-04 | grad 0.2427 +2026-04-10 10:36:39 - INFO - parrotllm.training - step 66720 | epoch 0 | loss 3.4686 | lr 7.38e-04 | grad 0.2474 +2026-04-10 10:36:42 - INFO - parrotllm.training - step 66730 | epoch 0 | loss 3.4494 | lr 7.38e-04 | grad 0.2342 +2026-04-10 10:36:45 - INFO - parrotllm.training - step 66740 | epoch 0 | loss 3.4216 | lr 7.38e-04 | grad 0.2815 +2026-04-10 10:36:48 - INFO - parrotllm.training - step 66750 | epoch 0 | loss 3.4661 | lr 7.38e-04 | grad 0.2644 +2026-04-10 10:36:51 - INFO - parrotllm.training - step 66760 | epoch 0 | loss 3.4346 | lr 7.38e-04 | grad 0.2683 +2026-04-10 10:36:54 - INFO - parrotllm.training - step 66770 | epoch 0 | loss 3.4419 | lr 7.38e-04 | grad 0.2478 +2026-04-10 10:36:57 - INFO - parrotllm.training - step 66780 | epoch 0 | loss 3.3145 | lr 7.37e-04 | grad 0.2790 +2026-04-10 10:37:00 - INFO - parrotllm.training - step 66790 | epoch 0 | loss 3.3803 | lr 7.37e-04 | grad 0.2792 +2026-04-10 10:37:03 - INFO - parrotllm.training - step 66800 | epoch 0 | loss 3.4301 | lr 7.37e-04 | grad 0.2546 +2026-04-10 10:37:06 - INFO - parrotllm.training - step 66810 | epoch 0 | loss 3.4032 | lr 7.37e-04 | grad 0.2446 +2026-04-10 10:37:09 - INFO - parrotllm.training - step 66820 | epoch 0 | loss 3.5102 | lr 7.37e-04 | grad 0.2596 +2026-04-10 10:37:12 - INFO - parrotllm.training - step 66830 | epoch 0 | loss 3.3948 | lr 7.37e-04 | grad 0.2519 +2026-04-10 10:37:16 - INFO - parrotllm.training - step 66840 | epoch 0 | loss 3.3516 | lr 7.36e-04 | grad 0.2577 +2026-04-10 10:37:19 - INFO - parrotllm.training - step 66850 | epoch 0 | loss 3.3306 | lr 7.36e-04 | grad 0.2592 +2026-04-10 10:37:22 - INFO - parrotllm.training - step 66860 | epoch 0 | loss 3.3822 | lr 7.36e-04 | grad 0.2493 +2026-04-10 10:37:25 - INFO - parrotllm.training - step 66870 | epoch 0 | loss 3.3428 | lr 7.36e-04 | grad 0.2550 +2026-04-10 10:37:28 - INFO - parrotllm.training - step 66880 | epoch 0 | loss 3.3812 | lr 7.36e-04 | grad 0.2569 +2026-04-10 10:37:31 - INFO - parrotllm.training - step 66890 | epoch 0 | loss 3.4318 | lr 7.35e-04 | grad 0.2644 +2026-04-10 10:37:34 - INFO - parrotllm.training - step 66900 | epoch 0 | loss 3.3858 | lr 7.35e-04 | grad 0.2591 +2026-04-10 10:37:37 - INFO - parrotllm.training - step 66910 | epoch 0 | loss 3.3608 | lr 7.35e-04 | grad 0.2557 +2026-04-10 10:37:40 - INFO - parrotllm.training - step 66920 | epoch 0 | loss 3.3267 | lr 7.35e-04 | grad 0.2542 +2026-04-10 10:37:43 - INFO - parrotllm.training - step 66930 | epoch 0 | loss 3.3567 | lr 7.35e-04 | grad 0.2416 +2026-04-10 10:37:46 - INFO - parrotllm.training - step 66940 | epoch 0 | loss 3.4036 | lr 7.35e-04 | grad 0.2240 +2026-04-10 10:37:49 - INFO - parrotllm.training - step 66950 | epoch 0 | loss 3.4551 | lr 7.34e-04 | grad 0.2819 +2026-04-10 10:37:52 - INFO - parrotllm.training - step 66960 | epoch 0 | loss 3.4627 | lr 7.34e-04 | grad 0.2669 +2026-04-10 10:37:55 - INFO - parrotllm.training - step 66970 | epoch 0 | loss 3.3462 | lr 7.34e-04 | grad 0.2728 +2026-04-10 10:37:58 - INFO - parrotllm.training - step 66980 | epoch 0 | loss 3.4944 | lr 7.34e-04 | grad 0.2470 +2026-04-10 10:38:01 - INFO - parrotllm.training - step 66990 | epoch 0 | loss 3.3333 | lr 7.34e-04 | grad 0.2684 +2026-04-10 10:38:04 - INFO - parrotllm.training - step 67000 | epoch 0 | loss 3.3985 | lr 7.34e-04 | grad 0.2336 +2026-04-10 10:38:04 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:38:04 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:38:07 - INFO - parrotllm.training - Train: loss=3.3985, ppl=29.92 +2026-04-10 10:38:07 - INFO - parrotllm.training - Val: loss=3.3273, ppl=27.86 +2026-04-10 10:38:07 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:38:08 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3273_epoch_0000_step_0067000.pt +2026-04-10 10:38:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:38:13 - INFO - parrotllm.training - step 67010 | epoch 0 | loss 3.3510 | lr 7.33e-04 | grad 0.2575 +2026-04-10 10:38:16 - INFO - parrotllm.training - step 67020 | epoch 0 | loss 3.4094 | lr 7.33e-04 | grad 0.2796 +2026-04-10 10:38:19 - INFO - parrotllm.training - step 67030 | epoch 0 | loss 3.3170 | lr 7.33e-04 | grad 0.3151 +2026-04-10 10:38:22 - INFO - parrotllm.training - step 67040 | epoch 0 | loss 3.3773 | lr 7.33e-04 | grad 0.2362 +2026-04-10 10:38:25 - INFO - parrotllm.training - step 67050 | epoch 0 | loss 3.4238 | lr 7.33e-04 | grad 0.2503 +2026-04-10 10:38:28 - INFO - parrotllm.training - step 67060 | epoch 0 | loss 3.4536 | lr 7.33e-04 | grad 0.2618 +2026-04-10 10:38:31 - INFO - parrotllm.training - step 67070 | epoch 0 | loss 3.3114 | lr 7.32e-04 | grad 0.2561 +2026-04-10 10:38:34 - INFO - parrotllm.training - step 67080 | epoch 0 | loss 3.4002 | lr 7.32e-04 | grad 0.2438 +2026-04-10 10:38:37 - INFO - parrotllm.training - step 67090 | epoch 0 | loss 3.3665 | lr 7.32e-04 | grad 0.2722 +2026-04-10 10:38:40 - INFO - parrotllm.training - step 67100 | epoch 0 | loss 3.3828 | lr 7.32e-04 | grad 0.2614 +2026-04-10 10:38:43 - INFO - parrotllm.training - step 67110 | epoch 0 | loss 3.3900 | lr 7.32e-04 | grad 0.2764 +2026-04-10 10:38:47 - INFO - parrotllm.training - step 67120 | epoch 0 | loss 3.4250 | lr 7.31e-04 | grad 0.2450 +2026-04-10 10:38:50 - INFO - parrotllm.training - step 67130 | epoch 0 | loss 3.4221 | lr 7.31e-04 | grad 0.2676 +2026-04-10 10:38:53 - INFO - parrotllm.training - step 67140 | epoch 0 | loss 3.4668 | lr 7.31e-04 | grad 0.2605 +2026-04-10 10:38:56 - INFO - parrotllm.training - step 67150 | epoch 0 | loss 3.3250 | lr 7.31e-04 | grad 0.2643 +2026-04-10 10:38:59 - INFO - parrotllm.training - step 67160 | epoch 0 | loss 3.4452 | lr 7.31e-04 | grad 0.2389 +2026-04-10 10:39:02 - INFO - parrotllm.training - step 67170 | epoch 0 | loss 3.4749 | lr 7.31e-04 | grad 0.2359 +2026-04-10 10:39:05 - INFO - parrotllm.training - step 67180 | epoch 0 | loss 3.3388 | lr 7.30e-04 | grad 0.2417 +2026-04-10 10:39:08 - INFO - parrotllm.training - step 67190 | epoch 0 | loss 3.3704 | lr 7.30e-04 | grad 0.2511 +2026-04-10 10:39:11 - INFO - parrotllm.training - step 67200 | epoch 0 | loss 3.3961 | lr 7.30e-04 | grad 0.3278 +2026-04-10 10:39:14 - INFO - parrotllm.training - step 67210 | epoch 0 | loss 3.3546 | lr 7.30e-04 | grad 0.2292 +2026-04-10 10:39:17 - INFO - parrotllm.training - step 67220 | epoch 0 | loss 3.4420 | lr 7.30e-04 | grad 0.2420 +2026-04-10 10:39:20 - INFO - parrotllm.training - step 67230 | epoch 0 | loss 3.3114 | lr 7.30e-04 | grad 0.2840 +2026-04-10 10:39:23 - INFO - parrotllm.training - step 67240 | epoch 0 | loss 3.3318 | lr 7.29e-04 | grad 0.2740 +2026-04-10 10:39:26 - INFO - parrotllm.training - step 67250 | epoch 0 | loss 3.2973 | lr 7.29e-04 | grad 0.2986 +2026-04-10 10:39:29 - INFO - parrotllm.training - step 67260 | epoch 0 | loss 3.3914 | lr 7.29e-04 | grad 0.2754 +2026-04-10 10:39:32 - INFO - parrotllm.training - step 67270 | epoch 0 | loss 3.3361 | lr 7.29e-04 | grad 0.2599 +2026-04-10 10:39:35 - INFO - parrotllm.training - step 67280 | epoch 0 | loss 3.3733 | lr 7.29e-04 | grad 0.2634 +2026-04-10 10:39:38 - INFO - parrotllm.training - step 67290 | epoch 0 | loss 3.4649 | lr 7.29e-04 | grad 0.2909 +2026-04-10 10:39:41 - INFO - parrotllm.training - step 67300 | epoch 0 | loss 3.4770 | lr 7.28e-04 | grad 0.2620 +2026-04-10 10:39:44 - INFO - parrotllm.training - step 67310 | epoch 0 | loss 3.3974 | lr 7.28e-04 | grad 0.2581 +2026-04-10 10:39:47 - INFO - parrotllm.training - step 67320 | epoch 0 | loss 3.4132 | lr 7.28e-04 | grad 0.2745 +2026-04-10 10:39:51 - INFO - parrotllm.training - step 67330 | epoch 0 | loss 3.3222 | lr 7.28e-04 | grad 0.2381 +2026-04-10 10:39:54 - INFO - parrotllm.training - step 67340 | epoch 0 | loss 3.2707 | lr 7.28e-04 | grad 0.2579 +2026-04-10 10:39:57 - INFO - parrotllm.training - step 67350 | epoch 0 | loss 3.3138 | lr 7.27e-04 | grad 0.2461 +2026-04-10 10:40:00 - INFO - parrotllm.training - step 67360 | epoch 0 | loss 3.3217 | lr 7.27e-04 | grad 0.2420 +2026-04-10 10:40:03 - INFO - parrotllm.training - step 67370 | epoch 0 | loss 3.4780 | lr 7.27e-04 | grad 0.2386 +2026-04-10 10:40:06 - INFO - parrotllm.training - step 67380 | epoch 0 | loss 3.3357 | lr 7.27e-04 | grad 0.2748 +2026-04-10 10:40:09 - INFO - parrotllm.training - step 67390 | epoch 0 | loss 3.3564 | lr 7.27e-04 | grad 0.2664 +2026-04-10 10:40:12 - INFO - parrotllm.training - step 67400 | epoch 0 | loss 3.3821 | lr 7.27e-04 | grad 0.2408 +2026-04-10 10:40:15 - INFO - parrotllm.training - step 67410 | epoch 0 | loss 3.4474 | lr 7.26e-04 | grad 0.2523 +2026-04-10 10:40:18 - INFO - parrotllm.training - step 67420 | epoch 0 | loss 3.3930 | lr 7.26e-04 | grad 0.2443 +2026-04-10 10:40:21 - INFO - parrotllm.training - step 67430 | epoch 0 | loss 3.3336 | lr 7.26e-04 | grad 0.2779 +2026-04-10 10:40:24 - INFO - parrotllm.training - step 67440 | epoch 0 | loss 3.3622 | lr 7.26e-04 | grad 0.3069 +2026-04-10 10:40:27 - INFO - parrotllm.training - step 67450 | epoch 0 | loss 3.2936 | lr 7.26e-04 | grad 0.2689 +2026-04-10 10:40:30 - INFO - parrotllm.training - step 67460 | epoch 0 | loss 3.4767 | lr 7.26e-04 | grad 0.2676 +2026-04-10 10:40:33 - INFO - parrotllm.training - step 67470 | epoch 0 | loss 3.4403 | lr 7.25e-04 | grad 0.2425 +2026-04-10 10:40:36 - INFO - parrotllm.training - step 67480 | epoch 0 | loss 3.4316 | lr 7.25e-04 | grad 0.2440 +2026-04-10 10:40:39 - INFO - parrotllm.training - step 67490 | epoch 0 | loss 3.3219 | lr 7.25e-04 | grad 0.2997 +2026-04-10 10:40:42 - INFO - parrotllm.training - step 67500 | epoch 0 | loss 3.3284 | lr 7.25e-04 | grad 0.2534 +2026-04-10 10:40:42 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:40:42 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:40:45 - INFO - parrotllm.training - Train: loss=3.3284, ppl=27.89 +2026-04-10 10:40:45 - INFO - parrotllm.training - Val: loss=3.3254, ppl=27.81 +2026-04-10 10:40:45 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:40:47 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3254_epoch_0000_step_0067500.pt +2026-04-10 10:40:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:40:50 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0067500.pt +2026-04-10 10:40:55 - INFO - parrotllm.training - step 67510 | epoch 0 | loss 3.3104 | lr 7.25e-04 | grad 0.2898 +2026-04-10 10:40:58 - INFO - parrotllm.training - step 67520 | epoch 0 | loss 3.4967 | lr 7.25e-04 | grad 0.2742 +2026-04-10 10:41:01 - INFO - parrotllm.training - step 67530 | epoch 0 | loss 3.4375 | lr 7.24e-04 | grad 0.2868 +2026-04-10 10:41:04 - INFO - parrotllm.training - step 67540 | epoch 0 | loss 3.3322 | lr 7.24e-04 | grad 0.2605 +2026-04-10 10:41:07 - INFO - parrotllm.training - step 67550 | epoch 0 | loss 3.3161 | lr 7.24e-04 | grad 0.2754 +2026-04-10 10:41:10 - INFO - parrotllm.training - step 67560 | epoch 0 | loss 3.4401 | lr 7.24e-04 | grad 0.2612 +2026-04-10 10:41:13 - INFO - parrotllm.training - step 67570 | epoch 0 | loss 3.3955 | lr 7.24e-04 | grad 0.2654 +2026-04-10 10:41:16 - INFO - parrotllm.training - step 67580 | epoch 0 | loss 3.3901 | lr 7.23e-04 | grad 0.2628 +2026-04-10 10:41:19 - INFO - parrotllm.training - step 67590 | epoch 0 | loss 3.4013 | lr 7.23e-04 | grad 0.2667 +2026-04-10 10:41:22 - INFO - parrotllm.training - step 67600 | epoch 0 | loss 3.4406 | lr 7.23e-04 | grad 0.2762 +2026-04-10 10:41:25 - INFO - parrotllm.training - step 67610 | epoch 0 | loss 3.3758 | lr 7.23e-04 | grad 0.2611 +2026-04-10 10:41:28 - INFO - parrotllm.training - step 67620 | epoch 0 | loss 3.4404 | lr 7.23e-04 | grad 0.2515 +2026-04-10 10:41:31 - INFO - parrotllm.training - step 67630 | epoch 0 | loss 3.3686 | lr 7.23e-04 | grad 0.2486 +2026-04-10 10:41:34 - INFO - parrotllm.training - step 67640 | epoch 0 | loss 3.4250 | lr 7.22e-04 | grad 0.2528 +2026-04-10 10:41:37 - INFO - parrotllm.training - step 67650 | epoch 0 | loss 3.3685 | lr 7.22e-04 | grad 0.2419 +2026-04-10 10:41:40 - INFO - parrotllm.training - step 67660 | epoch 0 | loss 3.3631 | lr 7.22e-04 | grad 0.2575 +2026-04-10 10:41:43 - INFO - parrotllm.training - step 67670 | epoch 0 | loss 3.4166 | lr 7.22e-04 | grad 0.2462 +2026-04-10 10:41:46 - INFO - parrotllm.training - step 67680 | epoch 0 | loss 3.2947 | lr 7.22e-04 | grad 0.2639 +2026-04-10 10:41:49 - INFO - parrotllm.training - step 67690 | epoch 0 | loss 3.2659 | lr 7.22e-04 | grad 0.2915 +2026-04-10 10:41:53 - INFO - parrotllm.training - step 67700 | epoch 0 | loss 3.3758 | lr 7.21e-04 | grad 0.2681 +2026-04-10 10:41:56 - INFO - parrotllm.training - step 67710 | epoch 0 | loss 3.3805 | lr 7.21e-04 | grad 0.2882 +2026-04-10 10:41:59 - INFO - parrotllm.training - step 67720 | epoch 0 | loss 3.3642 | lr 7.21e-04 | grad 0.2628 +2026-04-10 10:42:02 - INFO - parrotllm.training - step 67730 | epoch 0 | loss 3.3640 | lr 7.21e-04 | grad 0.2491 +2026-04-10 10:42:05 - INFO - parrotllm.training - step 67740 | epoch 0 | loss 3.3287 | lr 7.21e-04 | grad 0.2807 +2026-04-10 10:42:08 - INFO - parrotllm.training - step 67750 | epoch 0 | loss 3.3572 | lr 7.21e-04 | grad 0.2736 +2026-04-10 10:42:11 - INFO - parrotllm.training - step 67760 | epoch 0 | loss 3.2926 | lr 7.20e-04 | grad 0.2871 +2026-04-10 10:42:14 - INFO - parrotllm.training - step 67770 | epoch 0 | loss 3.3841 | lr 7.20e-04 | grad 0.2715 +2026-04-10 10:42:17 - INFO - parrotllm.training - step 67780 | epoch 0 | loss 3.3231 | lr 7.20e-04 | grad 0.2471 +2026-04-10 10:42:20 - INFO - parrotllm.training - step 67790 | epoch 0 | loss 3.3468 | lr 7.20e-04 | grad 0.2883 +2026-04-10 10:42:23 - INFO - parrotllm.training - step 67800 | epoch 0 | loss 3.4258 | lr 7.20e-04 | grad 0.2895 +2026-04-10 10:42:26 - INFO - parrotllm.training - step 67810 | epoch 0 | loss 3.3319 | lr 7.19e-04 | grad 0.2561 +2026-04-10 10:42:29 - INFO - parrotllm.training - step 67820 | epoch 0 | loss 3.4021 | lr 7.19e-04 | grad 0.2752 +2026-04-10 10:42:32 - INFO - parrotllm.training - step 67830 | epoch 0 | loss 3.3081 | lr 7.19e-04 | grad 0.2802 +2026-04-10 10:42:33 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 67861/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 10:42:36 - INFO - parrotllm.training - step 67840 | epoch 0 | loss 3.3974 | lr 7.19e-04 | grad 0.2615 +2026-04-10 10:42:39 - INFO - parrotllm.training - step 67850 | epoch 0 | loss 3.3995 | lr 7.19e-04 | grad 0.2482 +2026-04-10 10:42:42 - INFO - parrotllm.training - step 67860 | epoch 0 | loss 3.3851 | lr 7.19e-04 | grad 0.2373 +2026-04-10 10:42:45 - INFO - parrotllm.training - step 67870 | epoch 0 | loss 3.3530 | lr 7.18e-04 | grad 0.2568 +2026-04-10 10:42:48 - INFO - parrotllm.training - step 67880 | epoch 0 | loss 3.3628 | lr 7.18e-04 | grad 0.2401 +2026-04-10 10:42:51 - INFO - parrotllm.training - step 67890 | epoch 0 | loss 3.4746 | lr 7.18e-04 | grad 0.2712 +2026-04-10 10:42:54 - INFO - parrotllm.training - step 67900 | epoch 0 | loss 3.3639 | lr 7.18e-04 | grad 0.2951 +2026-04-10 10:42:57 - INFO - parrotllm.training - step 67910 | epoch 0 | loss 3.2655 | lr 7.18e-04 | grad 0.2827 +2026-04-10 10:43:00 - INFO - parrotllm.training - step 67920 | epoch 0 | loss 3.3865 | lr 7.18e-04 | grad 0.2549 +2026-04-10 10:43:03 - INFO - parrotllm.training - step 67930 | epoch 0 | loss 3.4443 | lr 7.17e-04 | grad 0.2610 +2026-04-10 10:43:06 - INFO - parrotllm.training - step 67940 | epoch 0 | loss 3.3485 | lr 7.17e-04 | grad 0.2857 +2026-04-10 10:43:09 - INFO - parrotllm.training - step 67950 | epoch 0 | loss 3.4420 | lr 7.17e-04 | grad 0.2592 +2026-04-10 10:43:12 - INFO - parrotllm.training - step 67960 | epoch 0 | loss 3.4302 | lr 7.17e-04 | grad 0.2483 +2026-04-10 10:43:15 - INFO - parrotllm.training - step 67970 | epoch 0 | loss 3.4320 | lr 7.17e-04 | grad 0.2635 +2026-04-10 10:43:18 - INFO - parrotllm.training - step 67980 | epoch 0 | loss 3.4542 | lr 7.17e-04 | grad 0.2563 +2026-04-10 10:43:21 - INFO - parrotllm.training - step 67990 | epoch 0 | loss 3.3838 | lr 7.16e-04 | grad 0.2885 +2026-04-10 10:43:24 - INFO - parrotllm.training - step 68000 | epoch 0 | loss 3.3478 | lr 7.16e-04 | grad 0.2739 +2026-04-10 10:43:24 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:43:24 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:43:27 - INFO - parrotllm.training - Train: loss=3.3478, ppl=28.44 +2026-04-10 10:43:27 - INFO - parrotllm.training - Val: loss=3.3255, ppl=27.81 +2026-04-10 10:43:29 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3255_epoch_0000_step_0068000.pt +2026-04-10 10:43:30 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3254, min_delta=0.001000). +2026-04-10 10:43:30 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:43:34 - INFO - parrotllm.training - step 68010 | epoch 0 | loss 3.4032 | lr 7.16e-04 | grad 0.2484 +2026-04-10 10:43:37 - INFO - parrotllm.training - step 68020 | epoch 0 | loss 3.2862 | lr 7.16e-04 | grad 0.2515 +2026-04-10 10:43:40 - INFO - parrotllm.training - step 68030 | epoch 0 | loss 3.3202 | lr 7.16e-04 | grad 0.2608 +2026-04-10 10:43:43 - INFO - parrotllm.training - step 68040 | epoch 0 | loss 3.3672 | lr 7.15e-04 | grad 0.2815 +2026-04-10 10:43:46 - INFO - parrotllm.training - step 68050 | epoch 0 | loss 3.4192 | lr 7.15e-04 | grad 0.2660 +2026-04-10 10:43:49 - INFO - parrotllm.training - step 68060 | epoch 0 | loss 3.3976 | lr 7.15e-04 | grad 0.2480 +2026-04-10 10:43:52 - INFO - parrotllm.training - step 68070 | epoch 0 | loss 3.4853 | lr 7.15e-04 | grad 0.2452 +2026-04-10 10:43:55 - INFO - parrotllm.training - step 68080 | epoch 0 | loss 3.3408 | lr 7.15e-04 | grad 0.2593 +2026-04-10 10:43:58 - INFO - parrotllm.training - step 68090 | epoch 0 | loss 3.2414 | lr 7.15e-04 | grad 0.2669 +2026-04-10 10:44:01 - INFO - parrotllm.training - step 68100 | epoch 0 | loss 3.4973 | lr 7.14e-04 | grad 0.2538 +2026-04-10 10:44:04 - INFO - parrotllm.training - step 68110 | epoch 0 | loss 3.3677 | lr 7.14e-04 | grad 0.2442 +2026-04-10 10:44:07 - INFO - parrotllm.training - step 68120 | epoch 0 | loss 3.3762 | lr 7.14e-04 | grad 0.2574 +2026-04-10 10:44:10 - INFO - parrotllm.training - step 68130 | epoch 0 | loss 3.3771 | lr 7.14e-04 | grad 0.2590 +2026-04-10 10:44:13 - INFO - parrotllm.training - step 68140 | epoch 0 | loss 3.4133 | lr 7.14e-04 | grad 0.2576 +2026-04-10 10:44:16 - INFO - parrotllm.training - step 68150 | epoch 0 | loss 3.4164 | lr 7.14e-04 | grad 0.2408 +2026-04-10 10:44:19 - INFO - parrotllm.training - step 68160 | epoch 0 | loss 3.3674 | lr 7.13e-04 | grad 0.2555 +2026-04-10 10:44:22 - INFO - parrotllm.training - step 68170 | epoch 0 | loss 3.3614 | lr 7.13e-04 | grad 0.2840 +2026-04-10 10:44:25 - INFO - parrotllm.training - step 68180 | epoch 0 | loss 3.4578 | lr 7.13e-04 | grad 0.2531 +2026-04-10 10:44:28 - INFO - parrotllm.training - step 68190 | epoch 0 | loss 3.2473 | lr 7.13e-04 | grad 0.2452 +2026-04-10 10:44:31 - INFO - parrotllm.training - step 68200 | epoch 0 | loss 3.3659 | lr 7.13e-04 | grad 0.2530 +2026-04-10 10:44:34 - INFO - parrotllm.training - step 68210 | epoch 0 | loss 3.3435 | lr 7.13e-04 | grad 0.2625 +2026-04-10 10:44:38 - INFO - parrotllm.training - step 68220 | epoch 0 | loss 3.4351 | lr 7.12e-04 | grad 0.2619 +2026-04-10 10:44:41 - INFO - parrotllm.training - step 68230 | epoch 0 | loss 3.3554 | lr 7.12e-04 | grad 0.2786 +2026-04-10 10:44:44 - INFO - parrotllm.training - step 68240 | epoch 0 | loss 3.4602 | lr 7.12e-04 | grad 0.2631 +2026-04-10 10:44:47 - INFO - parrotllm.training - step 68250 | epoch 0 | loss 3.4271 | lr 7.12e-04 | grad 0.2623 +2026-04-10 10:44:50 - INFO - parrotllm.training - step 68260 | epoch 0 | loss 3.4046 | lr 7.12e-04 | grad 0.2588 +2026-04-10 10:44:53 - INFO - parrotllm.training - step 68270 | epoch 0 | loss 3.4369 | lr 7.11e-04 | grad 0.2479 +2026-04-10 10:44:56 - INFO - parrotllm.training - step 68280 | epoch 0 | loss 3.3868 | lr 7.11e-04 | grad 0.2399 +2026-04-10 10:44:59 - INFO - parrotllm.training - step 68290 | epoch 0 | loss 3.4725 | lr 7.11e-04 | grad 0.2730 +2026-04-10 10:45:02 - INFO - parrotllm.training - step 68300 | epoch 0 | loss 3.3769 | lr 7.11e-04 | grad 0.2776 +2026-04-10 10:45:05 - INFO - parrotllm.training - step 68310 | epoch 0 | loss 3.4460 | lr 7.11e-04 | grad 0.2442 +2026-04-10 10:45:08 - INFO - parrotllm.training - step 68320 | epoch 0 | loss 3.3436 | lr 7.11e-04 | grad 0.2778 +2026-04-10 10:45:11 - INFO - parrotllm.training - step 68330 | epoch 0 | loss 3.3378 | lr 7.10e-04 | grad 0.2484 +2026-04-10 10:45:14 - INFO - parrotllm.training - step 68340 | epoch 0 | loss 3.3680 | lr 7.10e-04 | grad 0.2719 +2026-04-10 10:45:17 - INFO - parrotllm.training - step 68350 | epoch 0 | loss 3.4161 | lr 7.10e-04 | grad 0.2791 +2026-04-10 10:45:20 - INFO - parrotllm.training - step 68360 | epoch 0 | loss 3.3251 | lr 7.10e-04 | grad 0.2536 +2026-04-10 10:45:23 - INFO - parrotllm.training - step 68370 | epoch 0 | loss 3.3982 | lr 7.10e-04 | grad 0.2513 +2026-04-10 10:45:26 - INFO - parrotllm.training - step 68380 | epoch 0 | loss 3.2789 | lr 7.10e-04 | grad 0.2927 +2026-04-10 10:45:29 - INFO - parrotllm.training - step 68390 | epoch 0 | loss 3.2829 | lr 7.09e-04 | grad 0.2603 +2026-04-10 10:45:32 - INFO - parrotllm.training - step 68400 | epoch 0 | loss 3.3072 | lr 7.09e-04 | grad 0.2537 +2026-04-10 10:45:36 - INFO - parrotllm.training - step 68410 | epoch 0 | loss 3.3002 | lr 7.09e-04 | grad 0.2540 +2026-04-10 10:45:39 - INFO - parrotllm.training - step 68420 | epoch 0 | loss 3.4429 | lr 7.09e-04 | grad 0.2802 +2026-04-10 10:45:42 - INFO - parrotllm.training - step 68430 | epoch 0 | loss 3.4234 | lr 7.09e-04 | grad 0.2663 +2026-04-10 10:45:45 - INFO - parrotllm.training - step 68440 | epoch 0 | loss 3.4034 | lr 7.09e-04 | grad 0.2651 +2026-04-10 10:45:48 - INFO - parrotllm.training - step 68450 | epoch 0 | loss 3.3908 | lr 7.08e-04 | grad 0.2600 +2026-04-10 10:45:51 - INFO - parrotllm.training - step 68460 | epoch 0 | loss 3.4377 | lr 7.08e-04 | grad 0.2685 +2026-04-10 10:45:54 - INFO - parrotllm.training - step 68470 | epoch 0 | loss 3.4955 | lr 7.08e-04 | grad 0.2511 +2026-04-10 10:45:57 - INFO - parrotllm.training - step 68480 | epoch 0 | loss 3.3862 | lr 7.08e-04 | grad 0.2748 +2026-04-10 10:46:00 - INFO - parrotllm.training - step 68490 | epoch 0 | loss 3.4955 | lr 7.08e-04 | grad 0.3027 +2026-04-10 10:46:03 - INFO - parrotllm.training - step 68500 | epoch 0 | loss 3.3850 | lr 7.07e-04 | grad 0.2464 +2026-04-10 10:46:03 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:46:03 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:46:06 - INFO - parrotllm.training - Train: loss=3.3850, ppl=29.52 +2026-04-10 10:46:06 - INFO - parrotllm.training - Val: loss=3.3245, ppl=27.78 +2026-04-10 10:46:07 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3245_epoch_0000_step_0068500.pt +2026-04-10 10:46:09 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3254, min_delta=0.001000). +2026-04-10 10:46:09 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:46:12 - INFO - parrotllm.training - step 68510 | epoch 0 | loss 3.3166 | lr 7.07e-04 | grad 0.2572 +2026-04-10 10:46:15 - INFO - parrotllm.training - step 68520 | epoch 0 | loss 3.4165 | lr 7.07e-04 | grad 0.2515 +2026-04-10 10:46:18 - INFO - parrotllm.training - step 68530 | epoch 0 | loss 3.2916 | lr 7.07e-04 | grad 0.2770 +2026-04-10 10:46:21 - INFO - parrotllm.training - step 68540 | epoch 0 | loss 3.2527 | lr 7.07e-04 | grad 0.2352 +2026-04-10 10:46:24 - INFO - parrotllm.training - step 68550 | epoch 0 | loss 3.4333 | lr 7.07e-04 | grad 0.2388 +2026-04-10 10:46:27 - INFO - parrotllm.training - step 68560 | epoch 0 | loss 3.4188 | lr 7.06e-04 | grad 0.2678 +2026-04-10 10:46:30 - INFO - parrotllm.training - step 68570 | epoch 0 | loss 3.4428 | lr 7.06e-04 | grad 0.2456 +2026-04-10 10:46:33 - INFO - parrotllm.training - step 68580 | epoch 0 | loss 3.4182 | lr 7.06e-04 | grad 0.2404 +2026-04-10 10:46:36 - INFO - parrotllm.training - step 68590 | epoch 0 | loss 3.3446 | lr 7.06e-04 | grad 0.2573 +2026-04-10 10:46:39 - INFO - parrotllm.training - step 68600 | epoch 0 | loss 3.3994 | lr 7.06e-04 | grad 0.2611 +2026-04-10 10:46:42 - INFO - parrotllm.training - step 68610 | epoch 0 | loss 3.3937 | lr 7.06e-04 | grad 0.2587 +2026-04-10 10:46:46 - INFO - parrotllm.training - step 68620 | epoch 0 | loss 3.3103 | lr 7.05e-04 | grad 0.2682 +2026-04-10 10:46:49 - INFO - parrotllm.training - step 68630 | epoch 0 | loss 3.3069 | lr 7.05e-04 | grad 0.2527 +2026-04-10 10:46:52 - INFO - parrotllm.training - step 68640 | epoch 0 | loss 3.3640 | lr 7.05e-04 | grad 0.2449 +2026-04-10 10:46:55 - INFO - parrotllm.training - step 68650 | epoch 0 | loss 3.3233 | lr 7.05e-04 | grad 0.2479 +2026-04-10 10:46:58 - INFO - parrotllm.training - step 68660 | epoch 0 | loss 3.3853 | lr 7.05e-04 | grad 0.2902 +2026-04-10 10:47:01 - INFO - parrotllm.training - step 68670 | epoch 0 | loss 3.4413 | lr 7.05e-04 | grad 0.2654 +2026-04-10 10:47:04 - INFO - parrotllm.training - step 68680 | epoch 0 | loss 3.2646 | lr 7.04e-04 | grad 0.2648 +2026-04-10 10:47:07 - INFO - parrotllm.training - step 68690 | epoch 0 | loss 3.4085 | lr 7.04e-04 | grad 0.2510 +2026-04-10 10:47:10 - INFO - parrotllm.training - step 68700 | epoch 0 | loss 3.4167 | lr 7.04e-04 | grad 0.3008 +2026-04-10 10:47:13 - INFO - parrotllm.training - step 68710 | epoch 0 | loss 3.3436 | lr 7.04e-04 | grad 0.2773 +2026-04-10 10:47:16 - INFO - parrotllm.training - step 68720 | epoch 0 | loss 3.3904 | lr 7.04e-04 | grad 0.2752 +2026-04-10 10:47:19 - INFO - parrotllm.training - step 68730 | epoch 0 | loss 3.3624 | lr 7.04e-04 | grad 0.2804 +2026-04-10 10:47:22 - INFO - parrotllm.training - step 68740 | epoch 0 | loss 3.3731 | lr 7.03e-04 | grad 0.2467 +2026-04-10 10:47:25 - INFO - parrotllm.training - step 68750 | epoch 0 | loss 3.3049 | lr 7.03e-04 | grad 0.2652 +2026-04-10 10:47:28 - INFO - parrotllm.training - step 68760 | epoch 0 | loss 3.3497 | lr 7.03e-04 | grad 0.2612 +2026-04-10 10:47:31 - INFO - parrotllm.training - step 68770 | epoch 0 | loss 3.3139 | lr 7.03e-04 | grad 0.2617 +2026-04-10 10:47:34 - INFO - parrotllm.training - step 68780 | epoch 0 | loss 3.4416 | lr 7.03e-04 | grad 0.2669 +2026-04-10 10:47:37 - INFO - parrotllm.training - step 68790 | epoch 0 | loss 3.4214 | lr 7.02e-04 | grad 0.2757 +2026-04-10 10:47:40 - INFO - parrotllm.training - step 68800 | epoch 0 | loss 3.4111 | lr 7.02e-04 | grad 0.2568 +2026-04-10 10:47:43 - INFO - parrotllm.training - step 68810 | epoch 0 | loss 3.3152 | lr 7.02e-04 | grad 0.2775 +2026-04-10 10:47:46 - INFO - parrotllm.training - step 68820 | epoch 0 | loss 3.2611 | lr 7.02e-04 | grad 0.2578 +2026-04-10 10:47:50 - INFO - parrotllm.training - step 68830 | epoch 0 | loss 3.4885 | lr 7.02e-04 | grad 0.2710 +2026-04-10 10:47:53 - INFO - parrotllm.training - step 68840 | epoch 0 | loss 3.4316 | lr 7.02e-04 | grad 0.2935 +2026-04-10 10:47:56 - INFO - parrotllm.training - step 68850 | epoch 0 | loss 3.4282 | lr 7.01e-04 | grad 0.2502 +2026-04-10 10:47:59 - INFO - parrotllm.training - step 68860 | epoch 0 | loss 3.3595 | lr 7.01e-04 | grad 0.2819 +2026-04-10 10:48:02 - INFO - parrotllm.training - step 68870 | epoch 0 | loss 3.5258 | lr 7.01e-04 | grad 0.2377 +2026-04-10 10:48:05 - INFO - parrotllm.training - step 68880 | epoch 0 | loss 3.3681 | lr 7.01e-04 | grad 0.2749 +2026-04-10 10:48:08 - INFO - parrotllm.training - step 68890 | epoch 0 | loss 3.4323 | lr 7.01e-04 | grad 0.2817 +2026-04-10 10:48:11 - INFO - parrotllm.training - step 68900 | epoch 0 | loss 3.3542 | lr 7.01e-04 | grad 0.2541 +2026-04-10 10:48:14 - INFO - parrotllm.training - step 68910 | epoch 0 | loss 3.3146 | lr 7.00e-04 | grad 0.2503 +2026-04-10 10:48:17 - INFO - parrotllm.training - step 68920 | epoch 0 | loss 3.3884 | lr 7.00e-04 | grad 0.2424 +2026-04-10 10:48:20 - INFO - parrotllm.training - step 68930 | epoch 0 | loss 3.4143 | lr 7.00e-04 | grad 0.2467 +2026-04-10 10:48:23 - INFO - parrotllm.training - step 68940 | epoch 0 | loss 3.3455 | lr 7.00e-04 | grad 0.2327 +2026-04-10 10:48:26 - INFO - parrotllm.training - step 68950 | epoch 0 | loss 3.4061 | lr 7.00e-04 | grad 0.2747 +2026-04-10 10:48:29 - INFO - parrotllm.training - step 68960 | epoch 0 | loss 3.3553 | lr 7.00e-04 | grad 0.2725 +2026-04-10 10:48:32 - INFO - parrotllm.training - step 68970 | epoch 0 | loss 3.3698 | lr 6.99e-04 | grad 0.2685 +2026-04-10 10:48:35 - INFO - parrotllm.training - step 68980 | epoch 0 | loss 3.2821 | lr 6.99e-04 | grad 0.2390 +2026-04-10 10:48:38 - INFO - parrotllm.training - step 68990 | epoch 0 | loss 3.3372 | lr 6.99e-04 | grad 0.2676 +2026-04-10 10:48:41 - INFO - parrotllm.training - step 69000 | epoch 0 | loss 3.3745 | lr 6.99e-04 | grad 0.2586 +2026-04-10 10:48:41 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:48:41 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:48:44 - INFO - parrotllm.training - Train: loss=3.3745, ppl=29.21 +2026-04-10 10:48:44 - INFO - parrotllm.training - Val: loss=3.3243, ppl=27.78 +2026-04-10 10:48:44 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:48:45 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3243_epoch_0000_step_0069000.pt +2026-04-10 10:48:47 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:48:50 - INFO - parrotllm.training - step 69010 | epoch 0 | loss 3.3782 | lr 6.99e-04 | grad 0.2732 +2026-04-10 10:48:53 - INFO - parrotllm.training - step 69020 | epoch 0 | loss 3.3306 | lr 6.99e-04 | grad 0.2694 +2026-04-10 10:48:56 - INFO - parrotllm.training - step 69030 | epoch 0 | loss 3.3177 | lr 6.98e-04 | grad 0.2324 +2026-04-10 10:48:59 - INFO - parrotllm.training - step 69040 | epoch 0 | loss 3.3532 | lr 6.98e-04 | grad 0.2745 +2026-04-10 10:49:03 - INFO - parrotllm.training - step 69050 | epoch 0 | loss 3.3525 | lr 6.98e-04 | grad 0.3138 +2026-04-10 10:49:06 - INFO - parrotllm.training - step 69060 | epoch 0 | loss 3.3982 | lr 6.98e-04 | grad 0.2499 +2026-04-10 10:49:09 - INFO - parrotllm.training - step 69070 | epoch 0 | loss 3.3593 | lr 6.98e-04 | grad 0.2457 +2026-04-10 10:49:12 - INFO - parrotllm.training - step 69080 | epoch 0 | loss 3.4646 | lr 6.97e-04 | grad 0.2727 +2026-04-10 10:49:15 - INFO - parrotllm.training - step 69090 | epoch 0 | loss 3.4214 | lr 6.97e-04 | grad 0.2951 +2026-04-10 10:49:18 - INFO - parrotllm.training - step 69100 | epoch 0 | loss 3.3979 | lr 6.97e-04 | grad 0.2428 +2026-04-10 10:49:21 - INFO - parrotllm.training - step 69110 | epoch 0 | loss 3.3795 | lr 6.97e-04 | grad 0.2491 +2026-04-10 10:49:24 - INFO - parrotllm.training - step 69120 | epoch 0 | loss 3.4110 | lr 6.97e-04 | grad 0.2594 +2026-04-10 10:49:27 - INFO - parrotllm.training - step 69130 | epoch 0 | loss 3.4184 | lr 6.97e-04 | grad 0.2566 +2026-04-10 10:49:30 - INFO - parrotllm.training - step 69140 | epoch 0 | loss 3.3502 | lr 6.96e-04 | grad 0.2628 +2026-04-10 10:49:33 - INFO - parrotllm.training - step 69150 | epoch 0 | loss 3.3517 | lr 6.96e-04 | grad 0.2371 +2026-04-10 10:49:36 - INFO - parrotllm.training - step 69160 | epoch 0 | loss 3.3892 | lr 6.96e-04 | grad 0.2613 +2026-04-10 10:49:39 - INFO - parrotllm.training - step 69170 | epoch 0 | loss 3.4667 | lr 6.96e-04 | grad 0.2673 +2026-04-10 10:49:42 - INFO - parrotllm.training - step 69180 | epoch 0 | loss 3.3886 | lr 6.96e-04 | grad 0.2654 +2026-04-10 10:49:45 - INFO - parrotllm.training - step 69190 | epoch 0 | loss 3.4481 | lr 6.96e-04 | grad 0.2436 +2026-04-10 10:49:48 - INFO - parrotllm.training - step 69200 | epoch 0 | loss 3.3003 | lr 6.95e-04 | grad 0.2728 +2026-04-10 10:49:51 - INFO - parrotllm.training - step 69210 | epoch 0 | loss 3.3748 | lr 6.95e-04 | grad 0.2936 +2026-04-10 10:49:54 - INFO - parrotllm.training - step 69220 | epoch 0 | loss 3.3750 | lr 6.95e-04 | grad 0.2443 +2026-04-10 10:49:57 - INFO - parrotllm.training - step 69230 | epoch 0 | loss 3.4006 | lr 6.95e-04 | grad 0.2549 +2026-04-10 10:50:00 - INFO - parrotllm.training - step 69240 | epoch 0 | loss 3.4065 | lr 6.95e-04 | grad 0.2468 +2026-04-10 10:50:04 - INFO - parrotllm.training - step 69250 | epoch 0 | loss 3.3462 | lr 6.95e-04 | grad 0.3395 +2026-04-10 10:50:07 - INFO - parrotllm.training - step 69260 | epoch 0 | loss 3.4671 | lr 6.94e-04 | grad 0.2633 +2026-04-10 10:50:10 - INFO - parrotllm.training - step 69270 | epoch 0 | loss 3.2773 | lr 6.94e-04 | grad 0.2579 +2026-04-10 10:50:13 - INFO - parrotllm.training - step 69280 | epoch 0 | loss 3.3696 | lr 6.94e-04 | grad 0.2582 +2026-04-10 10:50:16 - INFO - parrotllm.training - step 69290 | epoch 0 | loss 3.4651 | lr 6.94e-04 | grad 0.2423 +2026-04-10 10:50:19 - INFO - parrotllm.training - step 69300 | epoch 0 | loss 3.4019 | lr 6.94e-04 | grad 0.2545 +2026-04-10 10:50:22 - INFO - parrotllm.training - step 69310 | epoch 0 | loss 3.4374 | lr 6.93e-04 | grad 0.2695 +2026-04-10 10:50:25 - INFO - parrotllm.training - step 69320 | epoch 0 | loss 3.4506 | lr 6.93e-04 | grad 0.2681 +2026-04-10 10:50:28 - INFO - parrotllm.training - step 69330 | epoch 0 | loss 3.4140 | lr 6.93e-04 | grad 0.2445 +2026-04-10 10:50:31 - INFO - parrotllm.training - step 69340 | epoch 0 | loss 3.4090 | lr 6.93e-04 | grad 0.2545 +2026-04-10 10:50:34 - INFO - parrotllm.training - step 69350 | epoch 0 | loss 3.3339 | lr 6.93e-04 | grad 0.2744 +2026-04-10 10:50:37 - INFO - parrotllm.training - step 69360 | epoch 0 | loss 3.4574 | lr 6.93e-04 | grad 0.2591 +2026-04-10 10:50:40 - INFO - parrotllm.training - step 69370 | epoch 0 | loss 3.4084 | lr 6.92e-04 | grad 0.2429 +2026-04-10 10:50:43 - INFO - parrotllm.training - step 69380 | epoch 0 | loss 3.3513 | lr 6.92e-04 | grad 0.2629 +2026-04-10 10:50:46 - INFO - parrotllm.training - step 69390 | epoch 0 | loss 3.3933 | lr 6.92e-04 | grad 0.2547 +2026-04-10 10:50:49 - INFO - parrotllm.training - step 69400 | epoch 0 | loss 3.4952 | lr 6.92e-04 | grad 0.2782 +2026-04-10 10:50:52 - INFO - parrotllm.training - step 69410 | epoch 0 | loss 3.4247 | lr 6.92e-04 | grad 0.2669 +2026-04-10 10:50:55 - INFO - parrotllm.training - step 69420 | epoch 0 | loss 3.3572 | lr 6.92e-04 | grad 0.2497 +2026-04-10 10:50:58 - INFO - parrotllm.training - step 69430 | epoch 0 | loss 3.2981 | lr 6.91e-04 | grad 0.2461 +2026-04-10 10:51:01 - INFO - parrotllm.training - step 69440 | epoch 0 | loss 3.3277 | lr 6.91e-04 | grad 0.2854 +2026-04-10 10:51:05 - INFO - parrotllm.training - step 69450 | epoch 0 | loss 3.3776 | lr 6.91e-04 | grad 0.2516 +2026-04-10 10:51:08 - INFO - parrotllm.training - step 69460 | epoch 0 | loss 3.3707 | lr 6.91e-04 | grad 0.2694 +2026-04-10 10:51:11 - INFO - parrotllm.training - step 69470 | epoch 0 | loss 3.4422 | lr 6.91e-04 | grad 0.2577 +2026-04-10 10:51:14 - INFO - parrotllm.training - step 69480 | epoch 0 | loss 3.3957 | lr 6.91e-04 | grad 0.2760 +2026-04-10 10:51:17 - INFO - parrotllm.training - step 69490 | epoch 0 | loss 3.3532 | lr 6.90e-04 | grad 0.2335 +2026-04-10 10:51:20 - INFO - parrotllm.training - step 69500 | epoch 0 | loss 3.2672 | lr 6.90e-04 | grad 0.2489 +2026-04-10 10:51:20 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:51:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:51:23 - INFO - parrotllm.training - Train: loss=3.2672, ppl=26.24 +2026-04-10 10:51:23 - INFO - parrotllm.training - Val: loss=3.3219, ppl=27.71 +2026-04-10 10:51:23 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:51:24 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3219_epoch_0000_step_0069500.pt +2026-04-10 10:51:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:51:29 - INFO - parrotllm.training - step 69510 | epoch 0 | loss 3.3753 | lr 6.90e-04 | grad 0.2698 +2026-04-10 10:51:32 - INFO - parrotllm.training - step 69520 | epoch 0 | loss 3.2761 | lr 6.90e-04 | grad 0.2674 +2026-04-10 10:51:35 - INFO - parrotllm.training - step 69530 | epoch 0 | loss 3.4079 | lr 6.90e-04 | grad 0.2713 +2026-04-10 10:51:38 - INFO - parrotllm.training - step 69540 | epoch 0 | loss 3.4707 | lr 6.90e-04 | grad 0.2577 +2026-04-10 10:51:41 - INFO - parrotllm.training - step 69550 | epoch 0 | loss 3.3393 | lr 6.89e-04 | grad 0.2463 +2026-04-10 10:51:44 - INFO - parrotllm.training - step 69560 | epoch 0 | loss 3.3468 | lr 6.89e-04 | grad 0.2701 +2026-04-10 10:51:47 - INFO - parrotllm.training - step 69570 | epoch 0 | loss 3.3022 | lr 6.89e-04 | grad 0.2513 +2026-04-10 10:51:50 - INFO - parrotllm.training - step 69580 | epoch 0 | loss 3.4000 | lr 6.89e-04 | grad 0.2394 +2026-04-10 10:51:53 - INFO - parrotllm.training - step 69590 | epoch 0 | loss 3.4069 | lr 6.89e-04 | grad 0.3377 +2026-04-10 10:51:56 - INFO - parrotllm.training - step 69600 | epoch 0 | loss 3.3240 | lr 6.88e-04 | grad 0.2646 +2026-04-10 10:51:59 - INFO - parrotllm.training - step 69610 | epoch 0 | loss 3.3291 | lr 6.88e-04 | grad 0.2573 +2026-04-10 10:52:02 - INFO - parrotllm.training - step 69620 | epoch 0 | loss 3.3652 | lr 6.88e-04 | grad 0.2536 +2026-04-10 10:52:05 - INFO - parrotllm.training - step 69630 | epoch 0 | loss 3.4268 | lr 6.88e-04 | grad 0.2480 +2026-04-10 10:52:08 - INFO - parrotllm.training - step 69640 | epoch 0 | loss 3.3471 | lr 6.88e-04 | grad 0.2806 +2026-04-10 10:52:11 - INFO - parrotllm.training - step 69650 | epoch 0 | loss 3.3851 | lr 6.88e-04 | grad 0.2299 +2026-04-10 10:52:15 - INFO - parrotllm.training - step 69660 | epoch 0 | loss 3.3636 | lr 6.87e-04 | grad 0.2579 +2026-04-10 10:52:18 - INFO - parrotllm.training - step 69670 | epoch 0 | loss 3.4291 | lr 6.87e-04 | grad 0.2545 +2026-04-10 10:52:21 - INFO - parrotllm.training - step 69680 | epoch 0 | loss 3.3741 | lr 6.87e-04 | grad 0.2858 +2026-04-10 10:52:24 - INFO - parrotllm.training - step 69690 | epoch 0 | loss 3.3905 | lr 6.87e-04 | grad 0.2791 +2026-04-10 10:52:27 - INFO - parrotllm.training - step 69700 | epoch 0 | loss 3.4424 | lr 6.87e-04 | grad 0.2812 +2026-04-10 10:52:30 - INFO - parrotllm.training - step 69710 | epoch 0 | loss 3.4159 | lr 6.87e-04 | grad 0.3144 +2026-04-10 10:52:33 - INFO - parrotllm.training - step 69720 | epoch 0 | loss 3.3406 | lr 6.86e-04 | grad 0.2559 +2026-04-10 10:52:36 - INFO - parrotllm.training - step 69730 | epoch 0 | loss 3.3715 | lr 6.86e-04 | grad 0.2810 +2026-04-10 10:52:39 - INFO - parrotllm.training - step 69740 | epoch 0 | loss 3.3132 | lr 6.86e-04 | grad 0.2532 +2026-04-10 10:52:42 - INFO - parrotllm.training - step 69750 | epoch 0 | loss 3.4525 | lr 6.86e-04 | grad 0.2569 +2026-04-10 10:52:45 - INFO - parrotllm.training - step 69760 | epoch 0 | loss 3.3700 | lr 6.86e-04 | grad 0.2620 +2026-04-10 10:52:48 - INFO - parrotllm.training - step 69770 | epoch 0 | loss 3.2958 | lr 6.86e-04 | grad 0.2983 +2026-04-10 10:52:51 - INFO - parrotllm.training - step 69780 | epoch 0 | loss 3.3892 | lr 6.85e-04 | grad 0.2722 +2026-04-10 10:52:54 - INFO - parrotllm.training - step 69790 | epoch 0 | loss 3.3852 | lr 6.85e-04 | grad 0.2939 +2026-04-10 10:52:57 - INFO - parrotllm.training - step 69800 | epoch 0 | loss 3.4709 | lr 6.85e-04 | grad 0.2689 +2026-04-10 10:53:00 - INFO - parrotllm.training - step 69810 | epoch 0 | loss 3.3283 | lr 6.85e-04 | grad 0.2765 +2026-04-10 10:53:03 - INFO - parrotllm.training - step 69820 | epoch 0 | loss 3.4271 | lr 6.85e-04 | grad 0.2828 +2026-04-10 10:53:06 - INFO - parrotllm.training - step 69830 | epoch 0 | loss 3.3672 | lr 6.85e-04 | grad 0.2869 +2026-04-10 10:53:09 - INFO - parrotllm.training - step 69840 | epoch 0 | loss 3.4060 | lr 6.84e-04 | grad 0.2883 +2026-04-10 10:53:12 - INFO - parrotllm.training - step 69850 | epoch 0 | loss 3.3109 | lr 6.84e-04 | grad 0.2582 +2026-04-10 10:53:16 - INFO - parrotllm.training - step 69860 | epoch 0 | loss 3.3724 | lr 6.84e-04 | grad 0.2642 +2026-04-10 10:53:19 - INFO - parrotllm.training - step 69870 | epoch 0 | loss 3.3970 | lr 6.84e-04 | grad 0.2617 +2026-04-10 10:53:22 - INFO - parrotllm.training - step 69880 | epoch 0 | loss 3.3438 | lr 6.84e-04 | grad 0.2799 +2026-04-10 10:53:25 - INFO - parrotllm.training - step 69890 | epoch 0 | loss 3.3868 | lr 6.84e-04 | grad 0.2623 +2026-04-10 10:53:28 - INFO - parrotllm.training - step 69900 | epoch 0 | loss 3.3453 | lr 6.83e-04 | grad 0.2532 +2026-04-10 10:53:31 - INFO - parrotllm.training - step 69910 | epoch 0 | loss 3.4076 | lr 6.83e-04 | grad 0.2707 +2026-04-10 10:53:34 - INFO - parrotllm.training - step 69920 | epoch 0 | loss 3.4950 | lr 6.83e-04 | grad 0.2520 +2026-04-10 10:53:37 - INFO - parrotllm.training - step 69930 | epoch 0 | loss 3.3634 | lr 6.83e-04 | grad 0.2324 +2026-04-10 10:53:40 - INFO - parrotllm.training - step 69940 | epoch 0 | loss 3.4767 | lr 6.83e-04 | grad 0.2435 +2026-04-10 10:53:43 - INFO - parrotllm.training - step 69950 | epoch 0 | loss 3.3413 | lr 6.82e-04 | grad 0.2514 +2026-04-10 10:53:46 - INFO - parrotllm.training - step 69960 | epoch 0 | loss 3.3360 | lr 6.82e-04 | grad 0.2588 +2026-04-10 10:53:49 - INFO - parrotllm.training - step 69970 | epoch 0 | loss 3.4265 | lr 6.82e-04 | grad 0.2605 +2026-04-10 10:53:52 - INFO - parrotllm.training - step 69980 | epoch 0 | loss 3.3486 | lr 6.82e-04 | grad 0.2628 +2026-04-10 10:53:55 - INFO - parrotllm.training - step 69990 | epoch 0 | loss 3.3716 | lr 6.82e-04 | grad 0.2578 +2026-04-10 10:53:58 - INFO - parrotllm.training - step 70000 | epoch 0 | loss 3.4242 | lr 6.82e-04 | grad 0.3087 +2026-04-10 10:53:58 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:53:58 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:54:01 - INFO - parrotllm.training - Train: loss=3.4242, ppl=30.70 +2026-04-10 10:54:01 - INFO - parrotllm.training - Val: loss=3.3209, ppl=27.69 +2026-04-10 10:54:02 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3209_epoch_0000_step_0070000.pt +2026-04-10 10:54:04 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3219, min_delta=0.001000). +2026-04-10 10:54:04 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:54:05 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0070000.pt +2026-04-10 10:54:10 - INFO - parrotllm.training - step 70010 | epoch 0 | loss 3.3932 | lr 6.81e-04 | grad 0.2709 +2026-04-10 10:54:14 - INFO - parrotllm.training - step 70020 | epoch 0 | loss 3.3199 | lr 6.81e-04 | grad 0.2695 +2026-04-10 10:54:17 - INFO - parrotllm.training - step 70030 | epoch 0 | loss 3.3754 | lr 6.81e-04 | grad 0.3057 +2026-04-10 10:54:20 - INFO - parrotllm.training - step 70040 | epoch 0 | loss 3.5582 | lr 6.81e-04 | grad 0.2698 +2026-04-10 10:54:23 - INFO - parrotllm.training - step 70050 | epoch 0 | loss 3.3692 | lr 6.81e-04 | grad 0.2523 +2026-04-10 10:54:26 - INFO - parrotllm.training - step 70060 | epoch 0 | loss 3.3518 | lr 6.81e-04 | grad 0.2543 +2026-04-10 10:54:29 - INFO - parrotllm.training - step 70070 | epoch 0 | loss 3.3963 | lr 6.80e-04 | grad 0.2690 +2026-04-10 10:54:32 - INFO - parrotllm.training - step 70080 | epoch 0 | loss 3.4099 | lr 6.80e-04 | grad 0.2852 +2026-04-10 10:54:35 - INFO - parrotllm.training - step 70090 | epoch 0 | loss 3.3666 | lr 6.80e-04 | grad 0.2468 +2026-04-10 10:54:35 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 70120/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 10:54:38 - INFO - parrotllm.training - step 70100 | epoch 0 | loss 3.3045 | lr 6.80e-04 | grad 0.2669 +2026-04-10 10:54:41 - INFO - parrotllm.training - step 70110 | epoch 0 | loss 3.3465 | lr 6.80e-04 | grad 0.2673 +2026-04-10 10:54:44 - INFO - parrotllm.training - step 70120 | epoch 0 | loss 3.3913 | lr 6.80e-04 | grad 0.2652 +2026-04-10 10:54:47 - INFO - parrotllm.training - step 70130 | epoch 0 | loss 3.3860 | lr 6.79e-04 | grad 0.2522 +2026-04-10 10:54:50 - INFO - parrotllm.training - step 70140 | epoch 0 | loss 3.2948 | lr 6.79e-04 | grad 0.2457 +2026-04-10 10:54:53 - INFO - parrotllm.training - step 70150 | epoch 0 | loss 3.2686 | lr 6.79e-04 | grad 0.2404 +2026-04-10 10:54:57 - INFO - parrotllm.training - step 70160 | epoch 0 | loss 3.3879 | lr 6.79e-04 | grad 0.2443 +2026-04-10 10:55:00 - INFO - parrotllm.training - step 70170 | epoch 0 | loss 3.2863 | lr 6.79e-04 | grad 0.2651 +2026-04-10 10:55:03 - INFO - parrotllm.training - step 70180 | epoch 0 | loss 3.3387 | lr 6.79e-04 | grad 0.2660 +2026-04-10 10:55:06 - INFO - parrotllm.training - step 70190 | epoch 0 | loss 3.4369 | lr 6.78e-04 | grad 0.2441 +2026-04-10 10:55:09 - INFO - parrotllm.training - step 70200 | epoch 0 | loss 3.3927 | lr 6.78e-04 | grad 0.2811 +2026-04-10 10:55:12 - INFO - parrotllm.training - step 70210 | epoch 0 | loss 3.4494 | lr 6.78e-04 | grad 0.2704 +2026-04-10 10:55:15 - INFO - parrotllm.training - step 70220 | epoch 0 | loss 3.4286 | lr 6.78e-04 | grad 0.2508 +2026-04-10 10:55:18 - INFO - parrotllm.training - step 70230 | epoch 0 | loss 3.4129 | lr 6.78e-04 | grad 0.2508 +2026-04-10 10:55:21 - INFO - parrotllm.training - step 70240 | epoch 0 | loss 3.4524 | lr 6.77e-04 | grad 0.2537 +2026-04-10 10:55:24 - INFO - parrotllm.training - step 70250 | epoch 0 | loss 3.4333 | lr 6.77e-04 | grad 0.2823 +2026-04-10 10:55:27 - INFO - parrotllm.training - step 70260 | epoch 0 | loss 3.3195 | lr 6.77e-04 | grad 0.2410 +2026-04-10 10:55:30 - INFO - parrotllm.training - step 70270 | epoch 0 | loss 3.4309 | lr 6.77e-04 | grad 0.2670 +2026-04-10 10:55:33 - INFO - parrotllm.training - step 70280 | epoch 0 | loss 3.2408 | lr 6.77e-04 | grad 0.2733 +2026-04-10 10:55:36 - INFO - parrotllm.training - step 70290 | epoch 0 | loss 3.2907 | lr 6.77e-04 | grad 0.2800 +2026-04-10 10:55:39 - INFO - parrotllm.training - step 70300 | epoch 0 | loss 3.3432 | lr 6.76e-04 | grad 0.2285 +2026-04-10 10:55:42 - INFO - parrotllm.training - step 70310 | epoch 0 | loss 3.4165 | lr 6.76e-04 | grad 0.2627 +2026-04-10 10:55:45 - INFO - parrotllm.training - step 70320 | epoch 0 | loss 3.3553 | lr 6.76e-04 | grad 0.2566 +2026-04-10 10:55:48 - INFO - parrotllm.training - step 70330 | epoch 0 | loss 3.3628 | lr 6.76e-04 | grad 0.2673 +2026-04-10 10:55:52 - INFO - parrotllm.training - step 70340 | epoch 0 | loss 3.3523 | lr 6.76e-04 | grad 0.2712 +2026-04-10 10:55:55 - INFO - parrotllm.training - step 70350 | epoch 0 | loss 3.3619 | lr 6.76e-04 | grad 0.2615 +2026-04-10 10:55:58 - INFO - parrotllm.training - step 70360 | epoch 0 | loss 3.4150 | lr 6.75e-04 | grad 0.2678 +2026-04-10 10:56:01 - INFO - parrotllm.training - step 70370 | epoch 0 | loss 3.2690 | lr 6.75e-04 | grad 0.2596 +2026-04-10 10:56:04 - INFO - parrotllm.training - step 70380 | epoch 0 | loss 3.4216 | lr 6.75e-04 | grad 0.2500 +2026-04-10 10:56:07 - INFO - parrotllm.training - step 70390 | epoch 0 | loss 3.3634 | lr 6.75e-04 | grad 0.2681 +2026-04-10 10:56:10 - INFO - parrotllm.training - step 70400 | epoch 0 | loss 3.3135 | lr 6.75e-04 | grad 0.3074 +2026-04-10 10:56:13 - INFO - parrotllm.training - step 70410 | epoch 0 | loss 3.3615 | lr 6.75e-04 | grad 0.2810 +2026-04-10 10:56:16 - INFO - parrotllm.training - step 70420 | epoch 0 | loss 3.4034 | lr 6.74e-04 | grad 0.2484 +2026-04-10 10:56:19 - INFO - parrotllm.training - step 70430 | epoch 0 | loss 3.3279 | lr 6.74e-04 | grad 0.2532 +2026-04-10 10:56:22 - INFO - parrotllm.training - step 70440 | epoch 0 | loss 3.3042 | lr 6.74e-04 | grad 0.2624 +2026-04-10 10:56:25 - INFO - parrotllm.training - step 70450 | epoch 0 | loss 3.3671 | lr 6.74e-04 | grad 0.2749 +2026-04-10 10:56:28 - INFO - parrotllm.training - step 70460 | epoch 0 | loss 3.3556 | lr 6.74e-04 | grad 0.2944 +2026-04-10 10:56:31 - INFO - parrotllm.training - step 70470 | epoch 0 | loss 3.4069 | lr 6.74e-04 | grad 0.2644 +2026-04-10 10:56:34 - INFO - parrotllm.training - step 70480 | epoch 0 | loss 3.2811 | lr 6.73e-04 | grad 0.2586 +2026-04-10 10:56:37 - INFO - parrotllm.training - step 70490 | epoch 0 | loss 3.4605 | lr 6.73e-04 | grad 0.2783 +2026-04-10 10:56:40 - INFO - parrotllm.training - step 70500 | epoch 0 | loss 3.3340 | lr 6.73e-04 | grad 0.2809 +2026-04-10 10:56:40 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:56:40 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:56:43 - INFO - parrotllm.training - Train: loss=3.3340, ppl=28.05 +2026-04-10 10:56:43 - INFO - parrotllm.training - Val: loss=3.3206, ppl=27.68 +2026-04-10 10:56:43 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:56:45 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3206_epoch_0000_step_0070500.pt +2026-04-10 10:56:46 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:56:49 - INFO - parrotllm.training - step 70510 | epoch 0 | loss 3.3653 | lr 6.73e-04 | grad 0.2397 +2026-04-10 10:56:52 - INFO - parrotllm.training - step 70520 | epoch 0 | loss 3.3415 | lr 6.73e-04 | grad 0.2591 +2026-04-10 10:56:55 - INFO - parrotllm.training - step 70530 | epoch 0 | loss 3.2281 | lr 6.73e-04 | grad 0.2732 +2026-04-10 10:56:58 - INFO - parrotllm.training - step 70540 | epoch 0 | loss 3.4271 | lr 6.72e-04 | grad 0.2915 +2026-04-10 10:57:01 - INFO - parrotllm.training - step 70550 | epoch 0 | loss 3.3522 | lr 6.72e-04 | grad 0.2626 +2026-04-10 10:57:05 - INFO - parrotllm.training - step 70560 | epoch 0 | loss 3.3413 | lr 6.72e-04 | grad 0.3030 +2026-04-10 10:57:08 - INFO - parrotllm.training - step 70570 | epoch 0 | loss 3.3350 | lr 6.72e-04 | grad 0.2525 +2026-04-10 10:57:11 - INFO - parrotllm.training - step 70580 | epoch 0 | loss 3.2847 | lr 6.72e-04 | grad 0.2628 +2026-04-10 10:57:14 - INFO - parrotllm.training - step 70590 | epoch 0 | loss 3.3264 | lr 6.71e-04 | grad 0.2394 +2026-04-10 10:57:17 - INFO - parrotllm.training - step 70600 | epoch 0 | loss 3.5230 | lr 6.71e-04 | grad 0.2655 +2026-04-10 10:57:20 - INFO - parrotllm.training - step 70610 | epoch 0 | loss 3.4546 | lr 6.71e-04 | grad 0.2517 +2026-04-10 10:57:23 - INFO - parrotllm.training - step 70620 | epoch 0 | loss 3.2520 | lr 6.71e-04 | grad 0.2558 +2026-04-10 10:57:26 - INFO - parrotllm.training - step 70630 | epoch 0 | loss 3.3192 | lr 6.71e-04 | grad 0.2895 +2026-04-10 10:57:29 - INFO - parrotllm.training - step 70640 | epoch 0 | loss 3.3290 | lr 6.71e-04 | grad 0.2704 +2026-04-10 10:57:32 - INFO - parrotllm.training - step 70650 | epoch 0 | loss 3.3722 | lr 6.70e-04 | grad 0.2978 +2026-04-10 10:57:35 - INFO - parrotllm.training - step 70660 | epoch 0 | loss 3.4218 | lr 6.70e-04 | grad 0.2573 +2026-04-10 10:57:38 - INFO - parrotllm.training - step 70670 | epoch 0 | loss 3.3283 | lr 6.70e-04 | grad 0.2468 +2026-04-10 10:57:41 - INFO - parrotllm.training - step 70680 | epoch 0 | loss 3.4191 | lr 6.70e-04 | grad 0.2439 +2026-04-10 10:57:44 - INFO - parrotllm.training - step 70690 | epoch 0 | loss 3.2848 | lr 6.70e-04 | grad 0.2420 +2026-04-10 10:57:47 - INFO - parrotllm.training - step 70700 | epoch 0 | loss 3.3636 | lr 6.70e-04 | grad 0.2862 +2026-04-10 10:57:51 - INFO - parrotllm.training - step 70710 | epoch 0 | loss 3.3549 | lr 6.69e-04 | grad 0.2422 +2026-04-10 10:57:54 - INFO - parrotllm.training - step 70720 | epoch 0 | loss 3.3808 | lr 6.69e-04 | grad 0.2695 +2026-04-10 10:57:57 - INFO - parrotllm.training - step 70730 | epoch 0 | loss 3.3014 | lr 6.69e-04 | grad 0.2607 +2026-04-10 10:58:00 - INFO - parrotllm.training - step 70740 | epoch 0 | loss 3.4670 | lr 6.69e-04 | grad 0.2389 +2026-04-10 10:58:03 - INFO - parrotllm.training - step 70750 | epoch 0 | loss 3.2828 | lr 6.69e-04 | grad 0.2848 +2026-04-10 10:58:06 - INFO - parrotllm.training - step 70760 | epoch 0 | loss 3.4966 | lr 6.69e-04 | grad 0.2751 +2026-04-10 10:58:09 - INFO - parrotllm.training - step 70770 | epoch 0 | loss 3.3668 | lr 6.68e-04 | grad 0.2696 +2026-04-10 10:58:12 - INFO - parrotllm.training - step 70780 | epoch 0 | loss 3.3964 | lr 6.68e-04 | grad 0.2483 +2026-04-10 10:58:15 - INFO - parrotllm.training - step 70790 | epoch 0 | loss 3.3121 | lr 6.68e-04 | grad 0.2893 +2026-04-10 10:58:18 - INFO - parrotllm.training - step 70800 | epoch 0 | loss 3.3441 | lr 6.68e-04 | grad 0.2607 +2026-04-10 10:58:21 - INFO - parrotllm.training - step 70810 | epoch 0 | loss 3.3312 | lr 6.68e-04 | grad 0.2597 +2026-04-10 10:58:24 - INFO - parrotllm.training - step 70820 | epoch 0 | loss 3.4528 | lr 6.68e-04 | grad 0.2539 +2026-04-10 10:58:27 - INFO - parrotllm.training - step 70830 | epoch 0 | loss 3.2926 | lr 6.67e-04 | grad 0.2585 +2026-04-10 10:58:30 - INFO - parrotllm.training - step 70840 | epoch 0 | loss 3.4127 | lr 6.67e-04 | grad 0.2419 +2026-04-10 10:58:33 - INFO - parrotllm.training - step 70850 | epoch 0 | loss 3.3505 | lr 6.67e-04 | grad 0.2551 +2026-04-10 10:58:37 - INFO - parrotllm.training - step 70860 | epoch 0 | loss 3.3926 | lr 6.67e-04 | grad 0.2482 +2026-04-10 10:58:40 - INFO - parrotllm.training - step 70870 | epoch 0 | loss 3.3184 | lr 6.67e-04 | grad 0.2459 +2026-04-10 10:58:43 - INFO - parrotllm.training - step 70880 | epoch 0 | loss 3.4743 | lr 6.67e-04 | grad 0.2991 +2026-04-10 10:58:46 - INFO - parrotllm.training - step 70890 | epoch 0 | loss 3.3599 | lr 6.66e-04 | grad 0.2753 +2026-04-10 10:58:49 - INFO - parrotllm.training - step 70900 | epoch 0 | loss 3.2970 | lr 6.66e-04 | grad 0.2714 +2026-04-10 10:58:52 - INFO - parrotllm.training - step 70910 | epoch 0 | loss 3.3978 | lr 6.66e-04 | grad 0.2360 +2026-04-10 10:58:55 - INFO - parrotllm.training - step 70920 | epoch 0 | loss 3.3142 | lr 6.66e-04 | grad 0.2736 +2026-04-10 10:58:58 - INFO - parrotllm.training - step 70930 | epoch 0 | loss 3.4317 | lr 6.66e-04 | grad 0.2600 +2026-04-10 10:59:01 - INFO - parrotllm.training - step 70940 | epoch 0 | loss 3.3643 | lr 6.66e-04 | grad 0.2689 +2026-04-10 10:59:04 - INFO - parrotllm.training - step 70950 | epoch 0 | loss 3.4184 | lr 6.65e-04 | grad 0.2381 +2026-04-10 10:59:07 - INFO - parrotllm.training - step 70960 | epoch 0 | loss 3.3425 | lr 6.65e-04 | grad 0.2501 +2026-04-10 10:59:10 - INFO - parrotllm.training - step 70970 | epoch 0 | loss 3.4087 | lr 6.65e-04 | grad 0.2941 +2026-04-10 10:59:13 - INFO - parrotllm.training - step 70980 | epoch 0 | loss 3.3537 | lr 6.65e-04 | grad 0.2512 +2026-04-10 10:59:16 - INFO - parrotllm.training - step 70990 | epoch 0 | loss 3.3632 | lr 6.65e-04 | grad 0.2696 +2026-04-10 10:59:20 - INFO - parrotllm.training - step 71000 | epoch 0 | loss 3.3596 | lr 6.64e-04 | grad 0.2663 +2026-04-10 10:59:20 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 10:59:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:59:23 - INFO - parrotllm.training - Train: loss=3.3596, ppl=28.78 +2026-04-10 10:59:23 - INFO - parrotllm.training - Val: loss=3.3173, ppl=27.59 +2026-04-10 10:59:23 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 10:59:24 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3173_epoch_0000_step_0071000.pt +2026-04-10 10:59:25 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 10:59:29 - INFO - parrotllm.training - step 71010 | epoch 0 | loss 3.3275 | lr 6.64e-04 | grad 0.2919 +2026-04-10 10:59:32 - INFO - parrotllm.training - step 71020 | epoch 0 | loss 3.4625 | lr 6.64e-04 | grad 0.2845 +2026-04-10 10:59:35 - INFO - parrotllm.training - step 71030 | epoch 0 | loss 3.4191 | lr 6.64e-04 | grad 0.2528 +2026-04-10 10:59:38 - INFO - parrotllm.training - step 71040 | epoch 0 | loss 3.3375 | lr 6.64e-04 | grad 0.2583 +2026-04-10 10:59:41 - INFO - parrotllm.training - step 71050 | epoch 0 | loss 3.3568 | lr 6.64e-04 | grad 0.3000 +2026-04-10 10:59:44 - INFO - parrotllm.training - step 71060 | epoch 0 | loss 3.4050 | lr 6.63e-04 | grad 0.2621 +2026-04-10 10:59:47 - INFO - parrotllm.training - step 71070 | epoch 0 | loss 3.3817 | lr 6.63e-04 | grad 0.2656 +2026-04-10 10:59:50 - INFO - parrotllm.training - step 71080 | epoch 0 | loss 3.4235 | lr 6.63e-04 | grad 0.2749 +2026-04-10 10:59:53 - INFO - parrotllm.training - step 71090 | epoch 0 | loss 3.3419 | lr 6.63e-04 | grad 0.2597 +2026-04-10 10:59:56 - INFO - parrotllm.training - step 71100 | epoch 0 | loss 3.3374 | lr 6.63e-04 | grad 0.2571 +2026-04-10 11:00:00 - INFO - parrotllm.training - step 71110 | epoch 0 | loss 3.4181 | lr 6.63e-04 | grad 0.2903 +2026-04-10 11:00:03 - INFO - parrotllm.training - step 71120 | epoch 0 | loss 3.4030 | lr 6.62e-04 | grad 0.2632 +2026-04-10 11:00:06 - INFO - parrotllm.training - step 71130 | epoch 0 | loss 3.4526 | lr 6.62e-04 | grad 0.2447 +2026-04-10 11:00:09 - INFO - parrotllm.training - step 71140 | epoch 0 | loss 3.2993 | lr 6.62e-04 | grad 0.2717 +2026-04-10 11:00:12 - INFO - parrotllm.training - step 71150 | epoch 0 | loss 3.3750 | lr 6.62e-04 | grad 0.2749 +2026-04-10 11:00:15 - INFO - parrotllm.training - step 71160 | epoch 0 | loss 3.3438 | lr 6.62e-04 | grad 0.2491 +2026-04-10 11:00:18 - INFO - parrotllm.training - step 71170 | epoch 0 | loss 3.3632 | lr 6.62e-04 | grad 0.2667 +2026-04-10 11:00:21 - INFO - parrotllm.training - step 71180 | epoch 0 | loss 3.3737 | lr 6.61e-04 | grad 0.2869 +2026-04-10 11:00:24 - INFO - parrotllm.training - step 71190 | epoch 0 | loss 3.3048 | lr 6.61e-04 | grad 0.2825 +2026-04-10 11:00:28 - INFO - parrotllm.training - step 71200 | epoch 0 | loss 3.2963 | lr 6.61e-04 | grad 0.3033 +2026-04-10 11:00:31 - INFO - parrotllm.training - step 71210 | epoch 0 | loss 3.3688 | lr 6.61e-04 | grad 0.2761 +2026-04-10 11:00:34 - INFO - parrotllm.training - step 71220 | epoch 0 | loss 3.4571 | lr 6.61e-04 | grad 0.2822 +2026-04-10 11:00:37 - INFO - parrotllm.training - step 71230 | epoch 0 | loss 3.4071 | lr 6.61e-04 | grad 0.2504 +2026-04-10 11:00:40 - INFO - parrotllm.training - step 71240 | epoch 0 | loss 3.3474 | lr 6.60e-04 | grad 0.2514 +2026-04-10 11:00:43 - INFO - parrotllm.training - step 71250 | epoch 0 | loss 3.2271 | lr 6.60e-04 | grad 0.2744 +2026-04-10 11:00:46 - INFO - parrotllm.training - step 71260 | epoch 0 | loss 3.3456 | lr 6.60e-04 | grad 0.2475 +2026-04-10 11:00:49 - INFO - parrotllm.training - step 71270 | epoch 0 | loss 3.4365 | lr 6.60e-04 | grad 0.2606 +2026-04-10 11:00:52 - INFO - parrotllm.training - step 71280 | epoch 0 | loss 3.3960 | lr 6.60e-04 | grad 0.2531 +2026-04-10 11:00:55 - INFO - parrotllm.training - step 71290 | epoch 0 | loss 3.3450 | lr 6.60e-04 | grad 0.2972 +2026-04-10 11:00:58 - INFO - parrotllm.training - step 71300 | epoch 0 | loss 3.3599 | lr 6.59e-04 | grad 0.2632 +2026-04-10 11:01:02 - INFO - parrotllm.training - step 71310 | epoch 0 | loss 3.3514 | lr 6.59e-04 | grad 0.2493 +2026-04-10 11:01:05 - INFO - parrotllm.training - step 71320 | epoch 0 | loss 3.3713 | lr 6.59e-04 | grad 0.2760 +2026-04-10 11:01:08 - INFO - parrotllm.training - step 71330 | epoch 0 | loss 3.4388 | lr 6.59e-04 | grad 0.2850 +2026-04-10 11:01:11 - INFO - parrotllm.training - step 71340 | epoch 0 | loss 3.3675 | lr 6.59e-04 | grad 0.2502 +2026-04-10 11:01:14 - INFO - parrotllm.training - step 71350 | epoch 0 | loss 3.3490 | lr 6.59e-04 | grad 0.2546 +2026-04-10 11:01:17 - INFO - parrotllm.training - step 71360 | epoch 0 | loss 3.3230 | lr 6.58e-04 | grad 0.2901 +2026-04-10 11:01:20 - INFO - parrotllm.training - step 71370 | epoch 0 | loss 3.3330 | lr 6.58e-04 | grad 0.2517 +2026-04-10 11:01:23 - INFO - parrotllm.training - step 71380 | epoch 0 | loss 3.3955 | lr 6.58e-04 | grad 0.2556 +2026-04-10 11:01:26 - INFO - parrotllm.training - step 71390 | epoch 0 | loss 3.4213 | lr 6.58e-04 | grad 0.2555 +2026-04-10 11:01:29 - INFO - parrotllm.training - step 71400 | epoch 0 | loss 3.4092 | lr 6.58e-04 | grad 0.2483 +2026-04-10 11:01:32 - INFO - parrotllm.training - step 71410 | epoch 0 | loss 3.4330 | lr 6.57e-04 | grad 0.2442 +2026-04-10 11:01:36 - INFO - parrotllm.training - step 71420 | epoch 0 | loss 3.4067 | lr 6.57e-04 | grad 0.2932 +2026-04-10 11:01:39 - INFO - parrotllm.training - step 71430 | epoch 0 | loss 3.3139 | lr 6.57e-04 | grad 0.2706 +2026-04-10 11:01:42 - INFO - parrotllm.training - step 71440 | epoch 0 | loss 3.3573 | lr 6.57e-04 | grad 0.2678 +2026-04-10 11:01:45 - INFO - parrotllm.training - step 71450 | epoch 0 | loss 3.2642 | lr 6.57e-04 | grad 0.2653 +2026-04-10 11:01:48 - INFO - parrotllm.training - step 71460 | epoch 0 | loss 3.3854 | lr 6.57e-04 | grad 0.2523 +2026-04-10 11:01:51 - INFO - parrotllm.training - step 71470 | epoch 0 | loss 3.2801 | lr 6.56e-04 | grad 0.2694 +2026-04-10 11:01:54 - INFO - parrotllm.training - step 71480 | epoch 0 | loss 3.4249 | lr 6.56e-04 | grad 0.2572 +2026-04-10 11:01:57 - INFO - parrotllm.training - step 71490 | epoch 0 | loss 3.2741 | lr 6.56e-04 | grad 0.2494 +2026-04-10 11:02:00 - INFO - parrotllm.training - step 71500 | epoch 0 | loss 3.3134 | lr 6.56e-04 | grad 0.2839 +2026-04-10 11:02:00 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:02:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:02:03 - INFO - parrotllm.training - Train: loss=3.3134, ppl=27.48 +2026-04-10 11:02:03 - INFO - parrotllm.training - Val: loss=3.3197, ppl=27.65 +2026-04-10 11:02:05 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3197_epoch_0000_step_0071500.pt +2026-04-10 11:02:06 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3173, min_delta=0.001000). +2026-04-10 11:02:06 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:02:10 - INFO - parrotllm.training - step 71510 | epoch 0 | loss 3.2746 | lr 6.56e-04 | grad 0.2464 +2026-04-10 11:02:13 - INFO - parrotllm.training - step 71520 | epoch 0 | loss 3.3618 | lr 6.56e-04 | grad 0.2751 +2026-04-10 11:02:16 - INFO - parrotllm.training - step 71530 | epoch 0 | loss 3.3673 | lr 6.55e-04 | grad 0.2941 +2026-04-10 11:02:19 - INFO - parrotllm.training - step 71540 | epoch 0 | loss 3.2933 | lr 6.55e-04 | grad 0.2780 +2026-04-10 11:02:22 - INFO - parrotllm.training - step 71550 | epoch 0 | loss 3.3639 | lr 6.55e-04 | grad 0.2566 +2026-04-10 11:02:25 - INFO - parrotllm.training - step 71560 | epoch 0 | loss 3.3486 | lr 6.55e-04 | grad 0.2557 +2026-04-10 11:02:28 - INFO - parrotllm.training - step 71570 | epoch 0 | loss 3.3873 | lr 6.55e-04 | grad 0.2777 +2026-04-10 11:02:31 - INFO - parrotllm.training - step 71580 | epoch 0 | loss 3.3226 | lr 6.55e-04 | grad 0.2512 +2026-04-10 11:02:35 - INFO - parrotllm.training - step 71590 | epoch 0 | loss 3.3659 | lr 6.54e-04 | grad 0.2878 +2026-04-10 11:02:38 - INFO - parrotllm.training - step 71600 | epoch 0 | loss 3.3920 | lr 6.54e-04 | grad 0.2732 +2026-04-10 11:02:41 - INFO - parrotllm.training - step 71610 | epoch 0 | loss 3.4451 | lr 6.54e-04 | grad 0.2476 +2026-04-10 11:02:44 - INFO - parrotllm.training - step 71620 | epoch 0 | loss 3.4109 | lr 6.54e-04 | grad 0.2824 +2026-04-10 11:02:47 - INFO - parrotllm.training - step 71630 | epoch 0 | loss 3.4384 | lr 6.54e-04 | grad 0.2678 +2026-04-10 11:02:50 - INFO - parrotllm.training - step 71640 | epoch 0 | loss 3.3705 | lr 6.54e-04 | grad 0.2790 +2026-04-10 11:02:53 - INFO - parrotllm.training - step 71650 | epoch 0 | loss 3.4337 | lr 6.53e-04 | grad 0.2718 +2026-04-10 11:02:56 - INFO - parrotllm.training - step 71660 | epoch 0 | loss 3.3281 | lr 6.53e-04 | grad 0.2556 +2026-04-10 11:02:59 - INFO - parrotllm.training - step 71670 | epoch 0 | loss 3.3419 | lr 6.53e-04 | grad 0.2790 +2026-04-10 11:03:02 - INFO - parrotllm.training - step 71680 | epoch 0 | loss 3.3444 | lr 6.53e-04 | grad 0.2783 +2026-04-10 11:03:05 - INFO - parrotllm.training - step 71690 | epoch 0 | loss 3.3827 | lr 6.53e-04 | grad 0.2700 +2026-04-10 11:03:09 - INFO - parrotllm.training - step 71700 | epoch 0 | loss 3.3212 | lr 6.53e-04 | grad 0.2643 +2026-04-10 11:03:12 - INFO - parrotllm.training - step 71710 | epoch 0 | loss 3.3596 | lr 6.52e-04 | grad 0.2663 +2026-04-10 11:03:15 - INFO - parrotllm.training - step 71720 | epoch 0 | loss 3.4641 | lr 6.52e-04 | grad 0.2588 +2026-04-10 11:03:18 - INFO - parrotllm.training - step 71730 | epoch 0 | loss 3.4193 | lr 6.52e-04 | grad 0.2837 +2026-04-10 11:03:21 - INFO - parrotllm.training - step 71740 | epoch 0 | loss 3.4850 | lr 6.52e-04 | grad 0.2565 +2026-04-10 11:03:24 - INFO - parrotllm.training - step 71750 | epoch 0 | loss 3.3388 | lr 6.52e-04 | grad 0.3277 +2026-04-10 11:03:27 - INFO - parrotllm.training - step 71760 | epoch 0 | loss 3.3251 | lr 6.52e-04 | grad 0.2752 +2026-04-10 11:03:30 - INFO - parrotllm.training - step 71770 | epoch 0 | loss 3.4150 | lr 6.51e-04 | grad 0.2774 +2026-04-10 11:03:33 - INFO - parrotllm.training - step 71780 | epoch 0 | loss 3.3506 | lr 6.51e-04 | grad 0.2651 +2026-04-10 11:03:37 - INFO - parrotllm.training - step 71790 | epoch 0 | loss 3.3678 | lr 6.51e-04 | grad 0.2490 +2026-04-10 11:03:40 - INFO - parrotllm.training - step 71800 | epoch 0 | loss 3.2744 | lr 6.51e-04 | grad 0.2704 +2026-04-10 11:03:43 - INFO - parrotllm.training - step 71810 | epoch 0 | loss 3.3241 | lr 6.51e-04 | grad 0.2662 +2026-04-10 11:03:46 - INFO - parrotllm.training - step 71820 | epoch 0 | loss 3.2447 | lr 6.51e-04 | grad 0.2625 +2026-04-10 11:03:49 - INFO - parrotllm.training - step 71830 | epoch 0 | loss 3.3108 | lr 6.50e-04 | grad 0.2587 +2026-04-10 11:03:52 - INFO - parrotllm.training - step 71840 | epoch 0 | loss 3.2685 | lr 6.50e-04 | grad 0.2738 +2026-04-10 11:03:55 - INFO - parrotllm.training - step 71850 | epoch 0 | loss 3.3199 | lr 6.50e-04 | grad 0.2645 +2026-04-10 11:03:58 - INFO - parrotllm.training - step 71860 | epoch 0 | loss 3.4428 | lr 6.50e-04 | grad 0.2552 +2026-04-10 11:04:02 - INFO - parrotllm.training - step 71870 | epoch 0 | loss 3.3689 | lr 6.50e-04 | grad 0.2639 +2026-04-10 11:04:05 - INFO - parrotllm.training - step 71880 | epoch 0 | loss 3.2799 | lr 6.49e-04 | grad 0.2639 +2026-04-10 11:04:08 - INFO - parrotllm.training - step 71890 | epoch 0 | loss 3.3835 | lr 6.49e-04 | grad 0.2984 +2026-04-10 11:04:11 - INFO - parrotllm.training - step 71900 | epoch 0 | loss 3.4369 | lr 6.49e-04 | grad 0.2676 +2026-04-10 11:04:14 - INFO - parrotllm.training - step 71910 | epoch 0 | loss 3.3868 | lr 6.49e-04 | grad 0.2858 +2026-04-10 11:04:17 - INFO - parrotllm.training - step 71920 | epoch 0 | loss 3.3446 | lr 6.49e-04 | grad 0.2467 +2026-04-10 11:04:20 - INFO - parrotllm.training - step 71930 | epoch 0 | loss 3.3348 | lr 6.49e-04 | grad 0.2567 +2026-04-10 11:04:23 - INFO - parrotllm.training - step 71940 | epoch 0 | loss 3.3978 | lr 6.48e-04 | grad 0.2698 +2026-04-10 11:04:27 - INFO - parrotllm.training - step 71950 | epoch 0 | loss 3.2702 | lr 6.48e-04 | grad 0.2605 +2026-04-10 11:04:30 - INFO - parrotllm.training - step 71960 | epoch 0 | loss 3.3533 | lr 6.48e-04 | grad 0.2718 +2026-04-10 11:04:33 - INFO - parrotllm.training - step 71970 | epoch 0 | loss 3.2582 | lr 6.48e-04 | grad 0.2657 +2026-04-10 11:04:36 - INFO - parrotllm.training - step 71980 | epoch 0 | loss 3.4671 | lr 6.48e-04 | grad 0.2886 +2026-04-10 11:04:39 - INFO - parrotllm.training - step 71990 | epoch 0 | loss 3.4962 | lr 6.48e-04 | grad 0.2579 +2026-04-10 11:04:42 - INFO - parrotllm.training - step 72000 | epoch 0 | loss 3.3643 | lr 6.47e-04 | grad 0.2436 +2026-04-10 11:04:42 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:04:42 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:04:45 - INFO - parrotllm.training - Train: loss=3.3643, ppl=28.91 +2026-04-10 11:04:45 - INFO - parrotllm.training - Val: loss=3.3150, ppl=27.52 +2026-04-10 11:04:45 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:04:47 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3150_epoch_0000_step_0072000.pt +2026-04-10 11:04:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:04:51 - INFO - parrotllm.training - step 72010 | epoch 0 | loss 3.4650 | lr 6.47e-04 | grad 0.2520 +2026-04-10 11:04:55 - INFO - parrotllm.training - step 72020 | epoch 0 | loss 3.4085 | lr 6.47e-04 | grad 0.2650 +2026-04-10 11:04:58 - INFO - parrotllm.training - step 72030 | epoch 0 | loss 3.4219 | lr 6.47e-04 | grad 0.2581 +2026-04-10 11:05:01 - INFO - parrotllm.training - step 72040 | epoch 0 | loss 3.3330 | lr 6.47e-04 | grad 0.2934 +2026-04-10 11:05:04 - INFO - parrotllm.training - step 72050 | epoch 0 | loss 3.2770 | lr 6.47e-04 | grad 0.3039 +2026-04-10 11:05:07 - INFO - parrotllm.training - step 72060 | epoch 0 | loss 3.4143 | lr 6.46e-04 | grad 0.2873 +2026-04-10 11:05:10 - INFO - parrotllm.training - step 72070 | epoch 0 | loss 3.3514 | lr 6.46e-04 | grad 0.2870 +2026-04-10 11:05:13 - INFO - parrotllm.training - step 72080 | epoch 0 | loss 3.2912 | lr 6.46e-04 | grad 0.2733 +2026-04-10 11:05:17 - INFO - parrotllm.training - step 72090 | epoch 0 | loss 3.4152 | lr 6.46e-04 | grad 0.2613 +2026-04-10 11:05:20 - INFO - parrotllm.training - step 72100 | epoch 0 | loss 3.3839 | lr 6.46e-04 | grad 0.2872 +2026-04-10 11:05:23 - INFO - parrotllm.training - step 72110 | epoch 0 | loss 3.2699 | lr 6.46e-04 | grad 0.2669 +2026-04-10 11:05:26 - INFO - parrotllm.training - step 72120 | epoch 0 | loss 3.3489 | lr 6.45e-04 | grad 0.2689 +2026-04-10 11:05:29 - INFO - parrotllm.training - step 72130 | epoch 0 | loss 3.3572 | lr 6.45e-04 | grad 0.2761 +2026-04-10 11:05:32 - INFO - parrotllm.training - step 72140 | epoch 0 | loss 3.3103 | lr 6.45e-04 | grad 0.2952 +2026-04-10 11:05:35 - INFO - parrotllm.training - step 72150 | epoch 0 | loss 3.3800 | lr 6.45e-04 | grad 0.2620 +2026-04-10 11:05:39 - INFO - parrotllm.training - step 72160 | epoch 0 | loss 3.3905 | lr 6.45e-04 | grad 0.2622 +2026-04-10 11:05:42 - INFO - parrotllm.training - step 72170 | epoch 0 | loss 3.3565 | lr 6.45e-04 | grad 0.2675 +2026-04-10 11:05:45 - INFO - parrotllm.training - step 72180 | epoch 0 | loss 3.3669 | lr 6.44e-04 | grad 0.2915 +2026-04-10 11:05:47 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 72217/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 11:05:48 - INFO - parrotllm.training - step 72190 | epoch 0 | loss 3.2801 | lr 6.44e-04 | grad 0.2703 +2026-04-10 11:05:51 - INFO - parrotllm.training - step 72200 | epoch 0 | loss 3.3476 | lr 6.44e-04 | grad 0.2908 +2026-04-10 11:05:55 - INFO - parrotllm.training - step 72210 | epoch 0 | loss 3.4115 | lr 6.44e-04 | grad 0.3168 +2026-04-10 11:05:58 - INFO - parrotllm.training - step 72220 | epoch 0 | loss 3.3287 | lr 6.44e-04 | grad 0.2593 +2026-04-10 11:06:01 - INFO - parrotllm.training - step 72230 | epoch 0 | loss 3.2732 | lr 6.44e-04 | grad 0.2734 +2026-04-10 11:06:04 - INFO - parrotllm.training - step 72240 | epoch 0 | loss 3.2766 | lr 6.43e-04 | grad 0.2448 +2026-04-10 11:06:07 - INFO - parrotllm.training - step 72250 | epoch 0 | loss 3.3213 | lr 6.43e-04 | grad 0.2664 +2026-04-10 11:06:11 - INFO - parrotllm.training - step 72260 | epoch 0 | loss 3.3986 | lr 6.43e-04 | grad 0.2988 +2026-04-10 11:06:14 - INFO - parrotllm.training - step 72270 | epoch 0 | loss 3.3080 | lr 6.43e-04 | grad 0.2504 +2026-04-10 11:06:17 - INFO - parrotllm.training - step 72280 | epoch 0 | loss 3.4499 | lr 6.43e-04 | grad 0.2600 +2026-04-10 11:06:20 - INFO - parrotllm.training - step 72290 | epoch 0 | loss 3.3278 | lr 6.43e-04 | grad 0.2877 +2026-04-10 11:06:23 - INFO - parrotllm.training - step 72300 | epoch 0 | loss 3.3584 | lr 6.42e-04 | grad 0.2562 +2026-04-10 11:06:27 - INFO - parrotllm.training - step 72310 | epoch 0 | loss 3.3553 | lr 6.42e-04 | grad 0.2686 +2026-04-10 11:06:30 - INFO - parrotllm.training - step 72320 | epoch 0 | loss 3.3262 | lr 6.42e-04 | grad 0.3042 +2026-04-10 11:06:33 - INFO - parrotllm.training - step 72330 | epoch 0 | loss 3.3446 | lr 6.42e-04 | grad 0.2793 +2026-04-10 11:06:36 - INFO - parrotllm.training - step 72340 | epoch 0 | loss 3.4562 | lr 6.42e-04 | grad 0.2644 +2026-04-10 11:06:39 - INFO - parrotllm.training - step 72350 | epoch 0 | loss 3.3895 | lr 6.42e-04 | grad 0.2529 +2026-04-10 11:06:43 - INFO - parrotllm.training - step 72360 | epoch 0 | loss 3.4257 | lr 6.41e-04 | grad 0.2549 +2026-04-10 11:06:46 - INFO - parrotllm.training - step 72370 | epoch 0 | loss 3.3417 | lr 6.41e-04 | grad 0.2606 +2026-04-10 11:06:49 - INFO - parrotllm.training - step 72380 | epoch 0 | loss 3.4020 | lr 6.41e-04 | grad 0.2712 +2026-04-10 11:06:52 - INFO - parrotllm.training - step 72390 | epoch 0 | loss 3.4272 | lr 6.41e-04 | grad 0.2740 +2026-04-10 11:06:55 - INFO - parrotllm.training - step 72400 | epoch 0 | loss 3.3204 | lr 6.41e-04 | grad 0.2772 +2026-04-10 11:06:58 - INFO - parrotllm.training - step 72410 | epoch 0 | loss 3.2842 | lr 6.41e-04 | grad 0.2890 +2026-04-10 11:07:02 - INFO - parrotllm.training - step 72420 | epoch 0 | loss 3.3102 | lr 6.40e-04 | grad 0.2892 +2026-04-10 11:07:05 - INFO - parrotllm.training - step 72430 | epoch 0 | loss 3.2438 | lr 6.40e-04 | grad 0.2881 +2026-04-10 11:07:08 - INFO - parrotllm.training - step 72440 | epoch 0 | loss 3.3156 | lr 6.40e-04 | grad 0.2619 +2026-04-10 11:07:12 - INFO - parrotllm.training - step 72450 | epoch 0 | loss 3.2777 | lr 6.40e-04 | grad 0.2508 +2026-04-10 11:07:15 - INFO - parrotllm.training - step 72460 | epoch 0 | loss 3.4074 | lr 6.40e-04 | grad 0.2939 +2026-04-10 11:07:18 - INFO - parrotllm.training - step 72470 | epoch 0 | loss 3.2584 | lr 6.39e-04 | grad 0.3079 +2026-04-10 11:07:21 - INFO - parrotllm.training - step 72480 | epoch 0 | loss 3.4167 | lr 6.39e-04 | grad 0.3248 +2026-04-10 11:07:25 - INFO - parrotllm.training - step 72490 | epoch 0 | loss 3.3723 | lr 6.39e-04 | grad 0.2697 +2026-04-10 11:07:28 - INFO - parrotllm.training - step 72500 | epoch 0 | loss 3.2612 | lr 6.39e-04 | grad 0.2424 +2026-04-10 11:07:28 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:07:28 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:07:31 - INFO - parrotllm.training - Train: loss=3.2612, ppl=26.08 +2026-04-10 11:07:31 - INFO - parrotllm.training - Val: loss=3.3155, ppl=27.54 +2026-04-10 11:07:32 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3155_epoch_0000_step_0072500.pt +2026-04-10 11:07:34 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3150, min_delta=0.001000). +2026-04-10 11:07:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:07:36 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0072500.pt +2026-04-10 11:07:41 - INFO - parrotllm.training - step 72510 | epoch 0 | loss 3.4736 | lr 6.39e-04 | grad 0.2711 +2026-04-10 11:07:44 - INFO - parrotllm.training - step 72520 | epoch 0 | loss 3.4087 | lr 6.39e-04 | grad 0.2770 +2026-04-10 11:07:47 - INFO - parrotllm.training - step 72530 | epoch 0 | loss 3.3864 | lr 6.38e-04 | grad 0.2818 +2026-04-10 11:07:51 - INFO - parrotllm.training - step 72540 | epoch 0 | loss 3.3181 | lr 6.38e-04 | grad 0.2842 +2026-04-10 11:07:54 - INFO - parrotllm.training - step 72550 | epoch 0 | loss 3.3560 | lr 6.38e-04 | grad 0.2567 +2026-04-10 11:07:57 - INFO - parrotllm.training - step 72560 | epoch 0 | loss 3.2987 | lr 6.38e-04 | grad 0.2709 +2026-04-10 11:08:00 - INFO - parrotllm.training - step 72570 | epoch 0 | loss 3.4004 | lr 6.38e-04 | grad 0.2515 +2026-04-10 11:08:03 - INFO - parrotllm.training - step 72580 | epoch 0 | loss 3.4482 | lr 6.38e-04 | grad 0.2891 +2026-04-10 11:08:07 - INFO - parrotllm.training - step 72590 | epoch 0 | loss 3.4131 | lr 6.37e-04 | grad 0.2576 +2026-04-10 11:08:10 - INFO - parrotllm.training - step 72600 | epoch 0 | loss 3.4407 | lr 6.37e-04 | grad 0.2932 +2026-04-10 11:08:13 - INFO - parrotllm.training - step 72610 | epoch 0 | loss 3.3032 | lr 6.37e-04 | grad 0.2745 +2026-04-10 11:08:16 - INFO - parrotllm.training - step 72620 | epoch 0 | loss 3.4098 | lr 6.37e-04 | grad 0.2555 +2026-04-10 11:08:20 - INFO - parrotllm.training - step 72630 | epoch 0 | loss 3.4002 | lr 6.37e-04 | grad 0.2539 +2026-04-10 11:08:23 - INFO - parrotllm.training - step 72640 | epoch 0 | loss 3.4520 | lr 6.37e-04 | grad 0.2854 +2026-04-10 11:08:26 - INFO - parrotllm.training - step 72650 | epoch 0 | loss 3.3922 | lr 6.36e-04 | grad 0.2819 +2026-04-10 11:08:29 - INFO - parrotllm.training - step 72660 | epoch 0 | loss 3.2875 | lr 6.36e-04 | grad 0.2615 +2026-04-10 11:08:33 - INFO - parrotllm.training - step 72670 | epoch 0 | loss 3.4463 | lr 6.36e-04 | grad 0.2864 +2026-04-10 11:08:36 - INFO - parrotllm.training - step 72680 | epoch 0 | loss 3.4196 | lr 6.36e-04 | grad 0.2616 +2026-04-10 11:08:39 - INFO - parrotllm.training - step 72690 | epoch 0 | loss 3.4570 | lr 6.36e-04 | grad 0.2725 +2026-04-10 11:08:42 - INFO - parrotllm.training - step 72700 | epoch 0 | loss 3.4480 | lr 6.36e-04 | grad 0.2445 +2026-04-10 11:08:46 - INFO - parrotllm.training - step 72710 | epoch 0 | loss 3.3325 | lr 6.35e-04 | grad 0.2461 +2026-04-10 11:08:49 - INFO - parrotllm.training - step 72720 | epoch 0 | loss 3.4534 | lr 6.35e-04 | grad 0.2864 +2026-04-10 11:08:52 - INFO - parrotllm.training - step 72730 | epoch 0 | loss 3.3220 | lr 6.35e-04 | grad 0.2559 +2026-04-10 11:08:56 - INFO - parrotllm.training - step 72740 | epoch 0 | loss 3.3346 | lr 6.35e-04 | grad 0.2602 +2026-04-10 11:08:59 - INFO - parrotllm.training - step 72750 | epoch 0 | loss 3.4001 | lr 6.35e-04 | grad 0.2493 +2026-04-10 11:09:02 - INFO - parrotllm.training - step 72760 | epoch 0 | loss 3.3532 | lr 6.35e-04 | grad 0.2755 +2026-04-10 11:09:05 - INFO - parrotllm.training - step 72770 | epoch 0 | loss 3.5053 | lr 6.34e-04 | grad 0.2485 +2026-04-10 11:09:09 - INFO - parrotllm.training - step 72780 | epoch 0 | loss 3.2797 | lr 6.34e-04 | grad 0.2546 +2026-04-10 11:09:12 - INFO - parrotllm.training - step 72790 | epoch 0 | loss 3.3455 | lr 6.34e-04 | grad 0.2658 +2026-04-10 11:09:15 - INFO - parrotllm.training - step 72800 | epoch 0 | loss 3.3111 | lr 6.34e-04 | grad 0.2530 +2026-04-10 11:09:18 - INFO - parrotllm.training - step 72810 | epoch 0 | loss 3.4854 | lr 6.34e-04 | grad 0.2719 +2026-04-10 11:09:22 - INFO - parrotllm.training - step 72820 | epoch 0 | loss 3.2809 | lr 6.34e-04 | grad 0.2459 +2026-04-10 11:09:25 - INFO - parrotllm.training - step 72830 | epoch 0 | loss 3.3548 | lr 6.33e-04 | grad 0.2911 +2026-04-10 11:09:28 - INFO - parrotllm.training - step 72840 | epoch 0 | loss 3.3434 | lr 6.33e-04 | grad 0.2849 +2026-04-10 11:09:31 - INFO - parrotllm.training - step 72850 | epoch 0 | loss 3.3927 | lr 6.33e-04 | grad 0.2533 +2026-04-10 11:09:35 - INFO - parrotllm.training - step 72860 | epoch 0 | loss 3.3649 | lr 6.33e-04 | grad 0.2528 +2026-04-10 11:09:38 - INFO - parrotllm.training - step 72870 | epoch 0 | loss 3.3879 | lr 6.33e-04 | grad 0.2770 +2026-04-10 11:09:41 - INFO - parrotllm.training - step 72880 | epoch 0 | loss 3.3346 | lr 6.33e-04 | grad 0.2700 +2026-04-10 11:09:44 - INFO - parrotllm.training - step 72890 | epoch 0 | loss 3.3289 | lr 6.32e-04 | grad 0.2883 +2026-04-10 11:09:48 - INFO - parrotllm.training - step 72900 | epoch 0 | loss 3.2693 | lr 6.32e-04 | grad 0.2647 +2026-04-10 11:09:51 - INFO - parrotllm.training - step 72910 | epoch 0 | loss 3.3110 | lr 6.32e-04 | grad 0.2615 +2026-04-10 11:09:54 - INFO - parrotllm.training - step 72920 | epoch 0 | loss 3.3892 | lr 6.32e-04 | grad 0.2731 +2026-04-10 11:09:58 - INFO - parrotllm.training - step 72930 | epoch 0 | loss 3.3994 | lr 6.32e-04 | grad 0.2687 +2026-04-10 11:10:01 - INFO - parrotllm.training - step 72940 | epoch 0 | loss 3.4209 | lr 6.32e-04 | grad 0.2584 +2026-04-10 11:10:04 - INFO - parrotllm.training - step 72950 | epoch 0 | loss 3.4546 | lr 6.31e-04 | grad 0.2645 +2026-04-10 11:10:08 - INFO - parrotllm.training - step 72960 | epoch 0 | loss 3.3992 | lr 6.31e-04 | grad 0.2598 +2026-04-10 11:10:11 - INFO - parrotllm.training - step 72970 | epoch 0 | loss 3.3790 | lr 6.31e-04 | grad 0.2751 +2026-04-10 11:10:14 - INFO - parrotllm.training - step 72980 | epoch 0 | loss 3.3068 | lr 6.31e-04 | grad 0.2817 +2026-04-10 11:10:18 - INFO - parrotllm.training - step 72990 | epoch 0 | loss 3.4213 | lr 6.31e-04 | grad 0.2616 +2026-04-10 11:10:21 - INFO - parrotllm.training - step 73000 | epoch 0 | loss 3.3150 | lr 6.31e-04 | grad 0.2709 +2026-04-10 11:10:21 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:10:21 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:10:24 - INFO - parrotllm.training - Train: loss=3.3150, ppl=27.52 +2026-04-10 11:10:24 - INFO - parrotllm.training - Val: loss=3.3136, ppl=27.48 +2026-04-10 11:10:24 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:10:25 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3136_epoch_0000_step_0073000.pt +2026-04-10 11:10:27 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:10:31 - INFO - parrotllm.training - step 73010 | epoch 0 | loss 3.3203 | lr 6.30e-04 | grad 0.2780 +2026-04-10 11:10:34 - INFO - parrotllm.training - step 73020 | epoch 0 | loss 3.3603 | lr 6.30e-04 | grad 0.2765 +2026-04-10 11:10:37 - INFO - parrotllm.training - step 73030 | epoch 0 | loss 3.4020 | lr 6.30e-04 | grad 0.2881 +2026-04-10 11:10:41 - INFO - parrotllm.training - step 73040 | epoch 0 | loss 3.3204 | lr 6.30e-04 | grad 0.2742 +2026-04-10 11:10:44 - INFO - parrotllm.training - step 73050 | epoch 0 | loss 3.4519 | lr 6.30e-04 | grad 0.2591 +2026-04-10 11:10:47 - INFO - parrotllm.training - step 73060 | epoch 0 | loss 3.3247 | lr 6.30e-04 | grad 0.2710 +2026-04-10 11:10:51 - INFO - parrotllm.training - step 73070 | epoch 0 | loss 3.3712 | lr 6.29e-04 | grad 0.2670 +2026-04-10 11:10:54 - INFO - parrotllm.training - step 73080 | epoch 0 | loss 3.3804 | lr 6.29e-04 | grad 0.2735 +2026-04-10 11:10:57 - INFO - parrotllm.training - step 73090 | epoch 0 | loss 3.4386 | lr 6.29e-04 | grad 0.2721 +2026-04-10 11:11:01 - INFO - parrotllm.training - step 73100 | epoch 0 | loss 3.3942 | lr 6.29e-04 | grad 0.2623 +2026-04-10 11:11:04 - INFO - parrotllm.training - step 73110 | epoch 0 | loss 3.3539 | lr 6.29e-04 | grad 0.2861 +2026-04-10 11:11:07 - INFO - parrotllm.training - step 73120 | epoch 0 | loss 3.4530 | lr 6.29e-04 | grad 0.2792 +2026-04-10 11:11:11 - INFO - parrotllm.training - step 73130 | epoch 0 | loss 3.3898 | lr 6.28e-04 | grad 0.2646 +2026-04-10 11:11:14 - INFO - parrotllm.training - step 73140 | epoch 0 | loss 3.3731 | lr 6.28e-04 | grad 0.2814 +2026-04-10 11:11:18 - INFO - parrotllm.training - step 73150 | epoch 0 | loss 3.3165 | lr 6.28e-04 | grad 0.3018 +2026-04-10 11:11:21 - INFO - parrotllm.training - step 73160 | epoch 0 | loss 3.4061 | lr 6.28e-04 | grad 0.2487 +2026-04-10 11:11:24 - INFO - parrotllm.training - step 73170 | epoch 0 | loss 3.3117 | lr 6.28e-04 | grad 0.2734 +2026-04-10 11:11:27 - INFO - parrotllm.training - step 73180 | epoch 0 | loss 3.4679 | lr 6.28e-04 | grad 0.2835 +2026-04-10 11:11:31 - INFO - parrotllm.training - step 73190 | epoch 0 | loss 3.3708 | lr 6.27e-04 | grad 0.2681 +2026-04-10 11:11:34 - INFO - parrotllm.training - step 73200 | epoch 0 | loss 3.3861 | lr 6.27e-04 | grad 0.2630 +2026-04-10 11:11:37 - INFO - parrotllm.training - step 73210 | epoch 0 | loss 3.3617 | lr 6.27e-04 | grad 0.2642 +2026-04-10 11:11:41 - INFO - parrotllm.training - step 73220 | epoch 0 | loss 3.3827 | lr 6.27e-04 | grad 0.2983 +2026-04-10 11:11:44 - INFO - parrotllm.training - step 73230 | epoch 0 | loss 3.4193 | lr 6.27e-04 | grad 0.2735 +2026-04-10 11:11:47 - INFO - parrotllm.training - step 73240 | epoch 0 | loss 3.3207 | lr 6.27e-04 | grad 0.2702 +2026-04-10 11:11:50 - INFO - parrotllm.training - step 73250 | epoch 0 | loss 3.3970 | lr 6.26e-04 | grad 0.2576 +2026-04-10 11:11:54 - INFO - parrotllm.training - step 73260 | epoch 0 | loss 3.4341 | lr 6.26e-04 | grad 0.2714 +2026-04-10 11:11:57 - INFO - parrotllm.training - step 73270 | epoch 0 | loss 3.4914 | lr 6.26e-04 | grad 0.2809 +2026-04-10 11:12:00 - INFO - parrotllm.training - step 73280 | epoch 0 | loss 3.3875 | lr 6.26e-04 | grad 0.2789 +2026-04-10 11:12:04 - INFO - parrotllm.training - step 73290 | epoch 0 | loss 3.3470 | lr 6.26e-04 | grad 0.2722 +2026-04-10 11:12:07 - INFO - parrotllm.training - step 73300 | epoch 0 | loss 3.3357 | lr 6.26e-04 | grad 0.3106 +2026-04-10 11:12:10 - INFO - parrotllm.training - step 73310 | epoch 0 | loss 3.3964 | lr 6.25e-04 | grad 0.3346 +2026-04-10 11:12:14 - INFO - parrotllm.training - step 73320 | epoch 0 | loss 3.3283 | lr 6.25e-04 | grad 0.2667 +2026-04-10 11:12:17 - INFO - parrotllm.training - step 73330 | epoch 0 | loss 3.3646 | lr 6.25e-04 | grad 0.2808 +2026-04-10 11:12:20 - INFO - parrotllm.training - step 73340 | epoch 0 | loss 3.3719 | lr 6.25e-04 | grad 0.2532 +2026-04-10 11:12:24 - INFO - parrotllm.training - step 73350 | epoch 0 | loss 3.4289 | lr 6.25e-04 | grad 0.2544 +2026-04-10 11:12:27 - INFO - parrotllm.training - step 73360 | epoch 0 | loss 3.3824 | lr 6.24e-04 | grad 0.2827 +2026-04-10 11:12:30 - INFO - parrotllm.training - step 73370 | epoch 0 | loss 3.2843 | lr 6.24e-04 | grad 0.2470 +2026-04-10 11:12:34 - INFO - parrotllm.training - step 73380 | epoch 0 | loss 3.3413 | lr 6.24e-04 | grad 0.2956 +2026-04-10 11:12:37 - INFO - parrotllm.training - step 73390 | epoch 0 | loss 3.3267 | lr 6.24e-04 | grad 0.2431 +2026-04-10 11:12:40 - INFO - parrotllm.training - step 73400 | epoch 0 | loss 3.4003 | lr 6.24e-04 | grad 0.2725 +2026-04-10 11:12:44 - INFO - parrotllm.training - step 73410 | epoch 0 | loss 3.3555 | lr 6.24e-04 | grad 0.2416 +2026-04-10 11:12:47 - INFO - parrotllm.training - step 73420 | epoch 0 | loss 3.4605 | lr 6.23e-04 | grad 0.2972 +2026-04-10 11:12:50 - INFO - parrotllm.training - step 73430 | epoch 0 | loss 3.4145 | lr 6.23e-04 | grad 0.2570 +2026-04-10 11:12:54 - INFO - parrotllm.training - step 73440 | epoch 0 | loss 3.4240 | lr 6.23e-04 | grad 0.2606 +2026-04-10 11:12:57 - INFO - parrotllm.training - step 73450 | epoch 0 | loss 3.3402 | lr 6.23e-04 | grad 0.2707 +2026-04-10 11:13:01 - INFO - parrotllm.training - step 73460 | epoch 0 | loss 3.3833 | lr 6.23e-04 | grad 0.2874 +2026-04-10 11:13:04 - INFO - parrotllm.training - step 73470 | epoch 0 | loss 3.3984 | lr 6.23e-04 | grad 0.2908 +2026-04-10 11:13:07 - INFO - parrotllm.training - step 73480 | epoch 0 | loss 3.2992 | lr 6.22e-04 | grad 0.2635 +2026-04-10 11:13:10 - INFO - parrotllm.training - step 73490 | epoch 0 | loss 3.4305 | lr 6.22e-04 | grad 0.2571 +2026-04-10 11:13:14 - INFO - parrotllm.training - step 73500 | epoch 0 | loss 3.3287 | lr 6.22e-04 | grad 0.2571 +2026-04-10 11:13:14 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:13:14 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:13:17 - INFO - parrotllm.training - Train: loss=3.3287, ppl=27.90 +2026-04-10 11:13:17 - INFO - parrotllm.training - Val: loss=3.3127, ppl=27.46 +2026-04-10 11:13:18 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3127_epoch_0000_step_0073500.pt +2026-04-10 11:13:20 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3136, min_delta=0.001000). +2026-04-10 11:13:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:13:24 - INFO - parrotllm.training - step 73510 | epoch 0 | loss 3.4238 | lr 6.22e-04 | grad 0.3124 +2026-04-10 11:13:27 - INFO - parrotllm.training - step 73520 | epoch 0 | loss 3.3207 | lr 6.22e-04 | grad 0.2914 +2026-04-10 11:13:30 - INFO - parrotllm.training - step 73530 | epoch 0 | loss 3.3686 | lr 6.22e-04 | grad 0.2605 +2026-04-10 11:13:34 - INFO - parrotllm.training - step 73540 | epoch 0 | loss 3.3085 | lr 6.21e-04 | grad 0.2618 +2026-04-10 11:13:37 - INFO - parrotllm.training - step 73550 | epoch 0 | loss 3.3999 | lr 6.21e-04 | grad 0.2700 +2026-04-10 11:13:40 - INFO - parrotllm.training - step 73560 | epoch 0 | loss 3.5079 | lr 6.21e-04 | grad 0.2497 +2026-04-10 11:13:44 - INFO - parrotllm.training - step 73570 | epoch 0 | loss 3.4456 | lr 6.21e-04 | grad 0.2677 +2026-04-10 11:13:47 - INFO - parrotllm.training - step 73580 | epoch 0 | loss 3.4886 | lr 6.21e-04 | grad 0.2602 +2026-04-10 11:13:50 - INFO - parrotllm.training - step 73590 | epoch 0 | loss 3.2975 | lr 6.21e-04 | grad 0.3058 +2026-04-10 11:13:54 - INFO - parrotllm.training - step 73600 | epoch 0 | loss 3.3422 | lr 6.20e-04 | grad 0.2792 +2026-04-10 11:13:57 - INFO - parrotllm.training - step 73610 | epoch 0 | loss 3.4330 | lr 6.20e-04 | grad 0.2725 +2026-04-10 11:14:00 - INFO - parrotllm.training - step 73620 | epoch 0 | loss 3.3882 | lr 6.20e-04 | grad 0.2752 +2026-04-10 11:14:04 - INFO - parrotllm.training - step 73630 | epoch 0 | loss 3.3378 | lr 6.20e-04 | grad 0.2947 +2026-04-10 11:14:07 - INFO - parrotllm.training - step 73640 | epoch 0 | loss 3.4887 | lr 6.20e-04 | grad 0.3076 +2026-04-10 11:14:11 - INFO - parrotllm.training - step 73650 | epoch 0 | loss 3.4451 | lr 6.20e-04 | grad 0.2773 +2026-04-10 11:14:14 - INFO - parrotllm.training - step 73660 | epoch 0 | loss 3.3287 | lr 6.19e-04 | grad 0.2857 +2026-04-10 11:14:17 - INFO - parrotllm.training - step 73670 | epoch 0 | loss 3.3396 | lr 6.19e-04 | grad 0.2620 +2026-04-10 11:14:21 - INFO - parrotllm.training - step 73680 | epoch 0 | loss 3.3278 | lr 6.19e-04 | grad 0.2591 +2026-04-10 11:14:24 - INFO - parrotllm.training - step 73690 | epoch 0 | loss 3.3855 | lr 6.19e-04 | grad 0.2559 +2026-04-10 11:14:27 - INFO - parrotllm.training - step 73700 | epoch 0 | loss 3.3776 | lr 6.19e-04 | grad 0.2781 +2026-04-10 11:14:31 - INFO - parrotllm.training - step 73710 | epoch 0 | loss 3.4625 | lr 6.19e-04 | grad 0.2617 +2026-04-10 11:14:34 - INFO - parrotllm.training - step 73720 | epoch 0 | loss 3.3596 | lr 6.18e-04 | grad 0.2562 +2026-04-10 11:14:37 - INFO - parrotllm.training - step 73730 | epoch 0 | loss 3.3229 | lr 6.18e-04 | grad 0.2609 +2026-04-10 11:14:41 - INFO - parrotllm.training - step 73740 | epoch 0 | loss 3.3176 | lr 6.18e-04 | grad 0.2397 +2026-04-10 11:14:44 - INFO - parrotllm.training - step 73750 | epoch 0 | loss 3.4436 | lr 6.18e-04 | grad 0.2652 +2026-04-10 11:14:47 - INFO - parrotllm.training - step 73760 | epoch 0 | loss 3.3541 | lr 6.18e-04 | grad 0.2951 +2026-04-10 11:14:51 - INFO - parrotllm.training - step 73770 | epoch 0 | loss 3.3215 | lr 6.18e-04 | grad 0.3071 +2026-04-10 11:14:54 - INFO - parrotllm.training - step 73780 | epoch 0 | loss 3.3558 | lr 6.17e-04 | grad 0.2715 +2026-04-10 11:14:57 - INFO - parrotllm.training - step 73790 | epoch 0 | loss 3.3268 | lr 6.17e-04 | grad 0.2956 +2026-04-10 11:15:01 - INFO - parrotllm.training - step 73800 | epoch 0 | loss 3.4648 | lr 6.17e-04 | grad 0.2853 +2026-04-10 11:15:04 - INFO - parrotllm.training - step 73810 | epoch 0 | loss 3.3219 | lr 6.17e-04 | grad 0.2882 +2026-04-10 11:15:07 - INFO - parrotllm.training - step 73820 | epoch 0 | loss 3.4303 | lr 6.17e-04 | grad 0.2727 +2026-04-10 11:15:11 - INFO - parrotllm.training - step 73830 | epoch 0 | loss 3.5019 | lr 6.17e-04 | grad 0.2765 +2026-04-10 11:15:14 - INFO - parrotllm.training - step 73840 | epoch 0 | loss 3.4223 | lr 6.16e-04 | grad 0.2515 +2026-04-10 11:15:17 - INFO - parrotllm.training - step 73850 | epoch 0 | loss 3.4096 | lr 6.16e-04 | grad 0.2865 +2026-04-10 11:15:20 - INFO - parrotllm.training - step 73860 | epoch 0 | loss 3.3855 | lr 6.16e-04 | grad 0.2653 +2026-04-10 11:15:24 - INFO - parrotllm.training - step 73870 | epoch 0 | loss 3.2845 | lr 6.16e-04 | grad 0.2843 +2026-04-10 11:15:27 - INFO - parrotllm.training - step 73880 | epoch 0 | loss 3.3677 | lr 6.16e-04 | grad 0.2866 +2026-04-10 11:15:30 - INFO - parrotllm.training - step 73890 | epoch 0 | loss 3.4389 | lr 6.16e-04 | grad 0.2715 +2026-04-10 11:15:34 - INFO - parrotllm.training - step 73900 | epoch 0 | loss 3.3587 | lr 6.15e-04 | grad 0.2532 +2026-04-10 11:15:37 - INFO - parrotllm.training - step 73910 | epoch 0 | loss 3.2872 | lr 6.15e-04 | grad 0.2726 +2026-04-10 11:15:40 - INFO - parrotllm.training - step 73920 | epoch 0 | loss 3.3747 | lr 6.15e-04 | grad 0.3058 +2026-04-10 11:15:44 - INFO - parrotllm.training - step 73930 | epoch 0 | loss 3.4872 | lr 6.15e-04 | grad 0.2852 +2026-04-10 11:15:47 - INFO - parrotllm.training - step 73940 | epoch 0 | loss 3.3169 | lr 6.15e-04 | grad 0.2704 +2026-04-10 11:15:50 - INFO - parrotllm.training - step 73950 | epoch 0 | loss 3.4216 | lr 6.15e-04 | grad 0.3063 +2026-04-10 11:15:54 - INFO - parrotllm.training - step 73960 | epoch 0 | loss 3.3264 | lr 6.14e-04 | grad 0.2519 +2026-04-10 11:15:57 - INFO - parrotllm.training - step 73970 | epoch 0 | loss 3.4105 | lr 6.14e-04 | grad 0.3086 +2026-04-10 11:16:00 - INFO - parrotllm.training - step 73980 | epoch 0 | loss 3.3352 | lr 6.14e-04 | grad 0.2811 +2026-04-10 11:16:04 - INFO - parrotllm.training - step 73990 | epoch 0 | loss 3.4361 | lr 6.14e-04 | grad 0.2700 +2026-04-10 11:16:07 - INFO - parrotllm.training - step 74000 | epoch 0 | loss 3.4750 | lr 6.14e-04 | grad 0.2613 +2026-04-10 11:16:07 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:16:07 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:16:10 - INFO - parrotllm.training - Train: loss=3.4750, ppl=32.30 +2026-04-10 11:16:10 - INFO - parrotllm.training - Val: loss=3.3126, ppl=27.46 +2026-04-10 11:16:10 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:16:11 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3126_epoch_0000_step_0074000.pt +2026-04-10 11:16:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:16:17 - INFO - parrotllm.training - step 74010 | epoch 0 | loss 3.3564 | lr 6.14e-04 | grad 0.2574 +2026-04-10 11:16:20 - INFO - parrotllm.training - step 74020 | epoch 0 | loss 3.3986 | lr 6.13e-04 | grad 0.2629 +2026-04-10 11:16:24 - INFO - parrotllm.training - step 74030 | epoch 0 | loss 3.4193 | lr 6.13e-04 | grad 0.2811 +2026-04-10 11:16:27 - INFO - parrotllm.training - step 74040 | epoch 0 | loss 3.3165 | lr 6.13e-04 | grad 0.2969 +2026-04-10 11:16:30 - INFO - parrotllm.training - step 74050 | epoch 0 | loss 3.2468 | lr 6.13e-04 | grad 0.3052 +2026-04-10 11:16:34 - INFO - parrotllm.training - step 74060 | epoch 0 | loss 3.4283 | lr 6.13e-04 | grad 0.2486 +2026-04-10 11:16:37 - INFO - parrotllm.training - step 74070 | epoch 0 | loss 3.3774 | lr 6.13e-04 | grad 0.2716 +2026-04-10 11:16:40 - INFO - parrotllm.training - step 74080 | epoch 0 | loss 3.3211 | lr 6.12e-04 | grad 0.2636 +2026-04-10 11:16:43 - INFO - parrotllm.training - step 74090 | epoch 0 | loss 3.4617 | lr 6.12e-04 | grad 0.2658 +2026-04-10 11:16:47 - INFO - parrotllm.training - step 74100 | epoch 0 | loss 3.2813 | lr 6.12e-04 | grad 0.2471 +2026-04-10 11:16:50 - INFO - parrotllm.training - step 74110 | epoch 0 | loss 3.3262 | lr 6.12e-04 | grad 0.2767 +2026-04-10 11:16:54 - INFO - parrotllm.training - step 74120 | epoch 0 | loss 3.3357 | lr 6.12e-04 | grad 0.2864 +2026-04-10 11:16:57 - INFO - parrotllm.training - step 74130 | epoch 0 | loss 3.3633 | lr 6.12e-04 | grad 0.2735 +2026-04-10 11:17:00 - INFO - parrotllm.training - step 74140 | epoch 0 | loss 3.4101 | lr 6.11e-04 | grad 0.2779 +2026-04-10 11:17:04 - INFO - parrotllm.training - step 74150 | epoch 0 | loss 3.2835 | lr 6.11e-04 | grad 0.2606 +2026-04-10 11:17:07 - INFO - parrotllm.training - step 74160 | epoch 0 | loss 3.3240 | lr 6.11e-04 | grad 0.2501 +2026-04-10 11:17:10 - INFO - parrotllm.training - step 74170 | epoch 0 | loss 3.3328 | lr 6.11e-04 | grad 0.2634 +2026-04-10 11:17:14 - INFO - parrotllm.training - step 74180 | epoch 0 | loss 3.3274 | lr 6.11e-04 | grad 0.2664 +2026-04-10 11:17:17 - INFO - parrotllm.training - step 74190 | epoch 0 | loss 3.3014 | lr 6.11e-04 | grad 0.2671 +2026-04-10 11:17:20 - INFO - parrotllm.training - step 74200 | epoch 0 | loss 3.3407 | lr 6.10e-04 | grad 0.2553 +2026-04-10 11:17:24 - INFO - parrotllm.training - step 74210 | epoch 0 | loss 3.3334 | lr 6.10e-04 | grad 0.2855 +2026-04-10 11:17:27 - INFO - parrotllm.training - step 74220 | epoch 0 | loss 3.4133 | lr 6.10e-04 | grad 0.2819 +2026-04-10 11:17:31 - INFO - parrotllm.training - step 74230 | epoch 0 | loss 3.3067 | lr 6.10e-04 | grad 0.2767 +2026-04-10 11:17:34 - INFO - parrotllm.training - step 74240 | epoch 0 | loss 3.4280 | lr 6.10e-04 | grad 0.2660 +2026-04-10 11:17:37 - INFO - parrotllm.training - step 74250 | epoch 0 | loss 3.3474 | lr 6.10e-04 | grad 0.2711 +2026-04-10 11:17:41 - INFO - parrotllm.training - step 74260 | epoch 0 | loss 3.3734 | lr 6.09e-04 | grad 0.2948 +2026-04-10 11:17:44 - INFO - parrotllm.training - step 74270 | epoch 0 | loss 3.3620 | lr 6.09e-04 | grad 0.2669 +2026-04-10 11:17:47 - INFO - parrotllm.training - step 74280 | epoch 0 | loss 3.4402 | lr 6.09e-04 | grad 0.2527 +2026-04-10 11:17:51 - INFO - parrotllm.training - step 74290 | epoch 0 | loss 3.3811 | lr 6.09e-04 | grad 0.2664 +2026-04-10 11:17:54 - INFO - parrotllm.training - step 74300 | epoch 0 | loss 3.3449 | lr 6.09e-04 | grad 0.2796 +2026-04-10 11:17:57 - INFO - parrotllm.training - step 74310 | epoch 0 | loss 3.3785 | lr 6.09e-04 | grad 0.3349 +2026-04-10 11:18:01 - INFO - parrotllm.training - step 74320 | epoch 0 | loss 3.3939 | lr 6.08e-04 | grad 0.2887 +2026-04-10 11:18:04 - INFO - parrotllm.training - step 74330 | epoch 0 | loss 3.3405 | lr 6.08e-04 | grad 0.2745 +2026-04-10 11:18:07 - INFO - parrotllm.training - step 74340 | epoch 0 | loss 3.3827 | lr 6.08e-04 | grad 0.2743 +2026-04-10 11:18:11 - INFO - parrotllm.training - step 74350 | epoch 0 | loss 3.3987 | lr 6.08e-04 | grad 0.2750 +2026-04-10 11:18:14 - INFO - parrotllm.training - step 74360 | epoch 0 | loss 3.4041 | lr 6.08e-04 | grad 0.2958 +2026-04-10 11:18:17 - INFO - parrotllm.training - step 74370 | epoch 0 | loss 3.3413 | lr 6.08e-04 | grad 0.2539 +2026-04-10 11:18:21 - INFO - parrotllm.training - step 74380 | epoch 0 | loss 3.3343 | lr 6.07e-04 | grad 0.2633 +2026-04-10 11:18:24 - INFO - parrotllm.training - step 74390 | epoch 0 | loss 3.4272 | lr 6.07e-04 | grad 0.2920 +2026-04-10 11:18:27 - INFO - parrotllm.training - step 74400 | epoch 0 | loss 3.4188 | lr 6.07e-04 | grad 0.2674 +2026-04-10 11:18:30 - INFO - parrotllm.training - step 74410 | epoch 0 | loss 3.4040 | lr 6.07e-04 | grad 0.2936 +2026-04-10 11:18:34 - INFO - parrotllm.training - step 74420 | epoch 0 | loss 3.3372 | lr 6.07e-04 | grad 0.2587 +2026-04-10 11:18:37 - INFO - parrotllm.training - step 74430 | epoch 0 | loss 3.3675 | lr 6.07e-04 | grad 0.2587 +2026-04-10 11:18:40 - INFO - parrotllm.training - step 74440 | epoch 0 | loss 3.3800 | lr 6.06e-04 | grad 0.2698 +2026-04-10 11:18:43 - INFO - parrotllm.training - step 74450 | epoch 0 | loss 3.4125 | lr 6.06e-04 | grad 0.2421 +2026-04-10 11:18:47 - INFO - parrotllm.training - step 74460 | epoch 0 | loss 3.3086 | lr 6.06e-04 | grad 0.2643 +2026-04-10 11:18:50 - INFO - parrotllm.training - step 74470 | epoch 0 | loss 3.3867 | lr 6.06e-04 | grad 0.2946 +2026-04-10 11:18:53 - INFO - parrotllm.training - step 74480 | epoch 0 | loss 3.3584 | lr 6.06e-04 | grad 0.2641 +2026-04-10 11:18:56 - INFO - parrotllm.training - step 74490 | epoch 0 | loss 3.2672 | lr 6.06e-04 | grad 0.2631 +2026-04-10 11:19:00 - INFO - parrotllm.training - step 74500 | epoch 0 | loss 3.3857 | lr 6.05e-04 | grad 0.2758 +2026-04-10 11:19:00 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:19:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:19:03 - INFO - parrotllm.training - Train: loss=3.3857, ppl=29.54 +2026-04-10 11:19:03 - INFO - parrotllm.training - Val: loss=3.3093, ppl=27.37 +2026-04-10 11:19:03 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:19:04 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3093_epoch_0000_step_0074500.pt +2026-04-10 11:19:06 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:19:09 - INFO - parrotllm.training - step 74510 | epoch 0 | loss 3.3052 | lr 6.05e-04 | grad 0.2533 +2026-04-10 11:19:13 - INFO - parrotllm.training - step 74520 | epoch 0 | loss 3.4580 | lr 6.05e-04 | grad 0.2565 +2026-04-10 11:19:16 - INFO - parrotllm.training - step 74530 | epoch 0 | loss 3.3885 | lr 6.05e-04 | grad 0.2738 +2026-04-10 11:19:19 - INFO - parrotllm.training - step 74540 | epoch 0 | loss 3.3803 | lr 6.05e-04 | grad 0.2793 +2026-04-10 11:19:23 - INFO - parrotllm.training - step 74550 | epoch 0 | loss 3.3629 | lr 6.05e-04 | grad 0.2752 +2026-04-10 11:19:26 - INFO - parrotllm.training - step 74560 | epoch 0 | loss 3.2720 | lr 6.04e-04 | grad 0.2701 +2026-04-10 11:19:29 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 74600/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 11:19:30 - INFO - parrotllm.training - step 74570 | epoch 0 | loss 3.3658 | lr 6.04e-04 | grad 0.2629 +2026-04-10 11:19:33 - INFO - parrotllm.training - step 74580 | epoch 0 | loss 3.3901 | lr 6.04e-04 | grad 0.2846 +2026-04-10 11:19:36 - INFO - parrotllm.training - step 74590 | epoch 0 | loss 3.3934 | lr 6.04e-04 | grad 0.2750 +2026-04-10 11:19:40 - INFO - parrotllm.training - step 74600 | epoch 0 | loss 3.4434 | lr 6.04e-04 | grad 0.2617 +2026-04-10 11:19:43 - INFO - parrotllm.training - step 74610 | epoch 0 | loss 3.2521 | lr 6.04e-04 | grad 0.2769 +2026-04-10 11:19:47 - INFO - parrotllm.training - step 74620 | epoch 0 | loss 3.3660 | lr 6.03e-04 | grad 0.2561 +2026-04-10 11:19:50 - INFO - parrotllm.training - step 74630 | epoch 0 | loss 3.4014 | lr 6.03e-04 | grad 0.2845 +2026-04-10 11:19:53 - INFO - parrotllm.training - step 74640 | epoch 0 | loss 3.2625 | lr 6.03e-04 | grad 0.2781 +2026-04-10 11:19:57 - INFO - parrotllm.training - step 74650 | epoch 0 | loss 3.4055 | lr 6.03e-04 | grad 0.2884 +2026-04-10 11:20:00 - INFO - parrotllm.training - step 74660 | epoch 0 | loss 3.3351 | lr 6.03e-04 | grad 0.2839 +2026-04-10 11:20:03 - INFO - parrotllm.training - step 74670 | epoch 0 | loss 3.3492 | lr 6.03e-04 | grad 0.2777 +2026-04-10 11:20:06 - INFO - parrotllm.training - step 74680 | epoch 0 | loss 3.3818 | lr 6.02e-04 | grad 0.2642 +2026-04-10 11:20:10 - INFO - parrotllm.training - step 74690 | epoch 0 | loss 3.2834 | lr 6.02e-04 | grad 0.2526 +2026-04-10 11:20:13 - INFO - parrotllm.training - step 74700 | epoch 0 | loss 3.3956 | lr 6.02e-04 | grad 0.2699 +2026-04-10 11:20:17 - INFO - parrotllm.training - step 74710 | epoch 0 | loss 3.4160 | lr 6.02e-04 | grad 0.2537 +2026-04-10 11:20:20 - INFO - parrotllm.training - step 74720 | epoch 0 | loss 3.3854 | lr 6.02e-04 | grad 0.2554 +2026-04-10 11:20:23 - INFO - parrotllm.training - step 74730 | epoch 0 | loss 3.3436 | lr 6.02e-04 | grad 0.2834 +2026-04-10 11:20:27 - INFO - parrotllm.training - step 74740 | epoch 0 | loss 3.3817 | lr 6.01e-04 | grad 0.2958 +2026-04-10 11:20:30 - INFO - parrotllm.training - step 74750 | epoch 0 | loss 3.4272 | lr 6.01e-04 | grad 0.2593 +2026-04-10 11:20:33 - INFO - parrotllm.training - step 74760 | epoch 0 | loss 3.4095 | lr 6.01e-04 | grad 0.2700 +2026-04-10 11:20:37 - INFO - parrotllm.training - step 74770 | epoch 0 | loss 3.3598 | lr 6.01e-04 | grad 0.2840 +2026-04-10 11:20:40 - INFO - parrotllm.training - step 74780 | epoch 0 | loss 3.3866 | lr 6.01e-04 | grad 0.2586 +2026-04-10 11:20:43 - INFO - parrotllm.training - step 74790 | epoch 0 | loss 3.4062 | lr 6.01e-04 | grad 0.2817 +2026-04-10 11:20:47 - INFO - parrotllm.training - step 74800 | epoch 0 | loss 3.3150 | lr 6.00e-04 | grad 0.2648 +2026-04-10 11:20:50 - INFO - parrotllm.training - step 74810 | epoch 0 | loss 3.3474 | lr 6.00e-04 | grad 0.2680 +2026-04-10 11:20:53 - INFO - parrotllm.training - step 74820 | epoch 0 | loss 3.3968 | lr 6.00e-04 | grad 0.2611 +2026-04-10 11:20:57 - INFO - parrotllm.training - step 74830 | epoch 0 | loss 3.3320 | lr 6.00e-04 | grad 0.2657 +2026-04-10 11:21:00 - INFO - parrotllm.training - step 74840 | epoch 0 | loss 3.3451 | lr 6.00e-04 | grad 0.2478 +2026-04-10 11:21:04 - INFO - parrotllm.training - step 74850 | epoch 0 | loss 3.4309 | lr 6.00e-04 | grad 0.2832 +2026-04-10 11:21:07 - INFO - parrotllm.training - step 74860 | epoch 0 | loss 3.5195 | lr 5.99e-04 | grad 0.2678 +2026-04-10 11:21:10 - INFO - parrotllm.training - step 74870 | epoch 0 | loss 3.4186 | lr 5.99e-04 | grad 0.2559 +2026-04-10 11:21:14 - INFO - parrotllm.training - step 74880 | epoch 0 | loss 3.3634 | lr 5.99e-04 | grad 0.2639 +2026-04-10 11:21:24 - INFO - parrotllm.training - step 74890 | epoch 0 | loss 3.4293 | lr 5.99e-04 | grad 0.2628 +2026-04-10 11:21:42 - INFO - parrotllm.training - step 74900 | epoch 0 | loss 3.3583 | lr 5.99e-04 | grad 0.2755 +2026-04-10 11:21:45 - INFO - parrotllm.training - step 74910 | epoch 0 | loss 3.3367 | lr 5.99e-04 | grad 0.2524 +2026-04-10 11:21:49 - INFO - parrotllm.training - step 74920 | epoch 0 | loss 3.2733 | lr 5.98e-04 | grad 0.2848 +2026-04-10 11:21:52 - INFO - parrotllm.training - step 74930 | epoch 0 | loss 3.3580 | lr 5.98e-04 | grad 0.2674 +2026-04-10 11:21:55 - INFO - parrotllm.training - step 74940 | epoch 0 | loss 3.4345 | lr 5.98e-04 | grad 0.2688 +2026-04-10 11:21:58 - INFO - parrotllm.training - step 74950 | epoch 0 | loss 3.3200 | lr 5.98e-04 | grad 0.2777 +2026-04-10 11:22:02 - INFO - parrotllm.training - step 74960 | epoch 0 | loss 3.3560 | lr 5.98e-04 | grad 0.3106 +2026-04-10 11:22:05 - INFO - parrotllm.training - step 74970 | epoch 0 | loss 3.3436 | lr 5.98e-04 | grad 0.3031 +2026-04-10 11:22:08 - INFO - parrotllm.training - step 74980 | epoch 0 | loss 3.4191 | lr 5.97e-04 | grad 0.2558 +2026-04-10 11:22:12 - INFO - parrotllm.training - step 74990 | epoch 0 | loss 3.3799 | lr 5.97e-04 | grad 0.2741 +2026-04-10 11:22:15 - INFO - parrotllm.training - step 75000 | epoch 0 | loss 3.4153 | lr 5.97e-04 | grad 0.2739 +2026-04-10 11:22:15 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:22:15 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:22:18 - INFO - parrotllm.training - Train: loss=3.4153, ppl=30.43 +2026-04-10 11:22:18 - INFO - parrotllm.training - Val: loss=3.3095, ppl=27.37 +2026-04-10 11:22:19 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3095_epoch_0000_step_0075000.pt +2026-04-10 11:22:21 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3093, min_delta=0.001000). +2026-04-10 11:22:21 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:22:22 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0075000.pt +2026-04-10 11:22:27 - INFO - parrotllm.training - step 75010 | epoch 0 | loss 3.4003 | lr 5.97e-04 | grad 0.2765 +2026-04-10 11:22:31 - INFO - parrotllm.training - step 75020 | epoch 0 | loss 3.3748 | lr 5.97e-04 | grad 0.2548 +2026-04-10 11:22:34 - INFO - parrotllm.training - step 75030 | epoch 0 | loss 3.3044 | lr 5.97e-04 | grad 0.2964 +2026-04-10 11:22:37 - INFO - parrotllm.training - step 75040 | epoch 0 | loss 3.2998 | lr 5.96e-04 | grad 0.2729 +2026-04-10 11:22:40 - INFO - parrotllm.training - step 75050 | epoch 0 | loss 3.3809 | lr 5.96e-04 | grad 0.2989 +2026-04-10 11:22:43 - INFO - parrotllm.training - step 75060 | epoch 0 | loss 3.3741 | lr 5.96e-04 | grad 0.2552 +2026-04-10 11:22:46 - INFO - parrotllm.training - step 75070 | epoch 0 | loss 3.4143 | lr 5.96e-04 | grad 0.2703 +2026-04-10 11:22:49 - INFO - parrotllm.training - step 75080 | epoch 0 | loss 3.3658 | lr 5.96e-04 | grad 0.2796 +2026-04-10 11:22:52 - INFO - parrotllm.training - step 75090 | epoch 0 | loss 3.3136 | lr 5.96e-04 | grad 0.2791 +2026-04-10 11:22:55 - INFO - parrotllm.training - step 75100 | epoch 0 | loss 3.3182 | lr 5.95e-04 | grad 0.2654 +2026-04-10 11:22:58 - INFO - parrotllm.training - step 75110 | epoch 0 | loss 3.4041 | lr 5.95e-04 | grad 0.2663 +2026-04-10 11:23:01 - INFO - parrotllm.training - step 75120 | epoch 0 | loss 3.3959 | lr 5.95e-04 | grad 0.2821 +2026-04-10 11:23:04 - INFO - parrotllm.training - step 75130 | epoch 0 | loss 3.2914 | lr 5.95e-04 | grad 0.2929 +2026-04-10 11:23:07 - INFO - parrotllm.training - step 75140 | epoch 0 | loss 3.3569 | lr 5.95e-04 | grad 0.2585 +2026-04-10 11:23:10 - INFO - parrotllm.training - step 75150 | epoch 0 | loss 3.2455 | lr 5.95e-04 | grad 0.2714 +2026-04-10 11:23:13 - INFO - parrotllm.training - step 75160 | epoch 0 | loss 3.3590 | lr 5.94e-04 | grad 0.2642 +2026-04-10 11:23:17 - INFO - parrotllm.training - step 75170 | epoch 0 | loss 3.3482 | lr 5.94e-04 | grad 0.2531 +2026-04-10 11:23:20 - INFO - parrotllm.training - step 75180 | epoch 0 | loss 3.4018 | lr 5.94e-04 | grad 0.2778 +2026-04-10 11:23:23 - INFO - parrotllm.training - step 75190 | epoch 0 | loss 3.4234 | lr 5.94e-04 | grad 0.2735 +2026-04-10 11:23:26 - INFO - parrotllm.training - step 75200 | epoch 0 | loss 3.3874 | lr 5.94e-04 | grad 0.2619 +2026-04-10 11:23:30 - INFO - parrotllm.training - step 75210 | epoch 0 | loss 3.2633 | lr 5.94e-04 | grad 0.2852 +2026-04-10 11:23:33 - INFO - parrotllm.training - step 75220 | epoch 0 | loss 3.4260 | lr 5.94e-04 | grad 0.2850 +2026-04-10 11:23:36 - INFO - parrotllm.training - step 75230 | epoch 0 | loss 3.3530 | lr 5.93e-04 | grad 0.2818 +2026-04-10 11:23:39 - INFO - parrotllm.training - step 75240 | epoch 0 | loss 3.4388 | lr 5.93e-04 | grad 0.2725 +2026-04-10 11:23:42 - INFO - parrotllm.training - step 75250 | epoch 0 | loss 3.3047 | lr 5.93e-04 | grad 0.2601 +2026-04-10 11:23:46 - INFO - parrotllm.training - step 75260 | epoch 0 | loss 3.3944 | lr 5.93e-04 | grad 0.2924 +2026-04-10 11:23:49 - INFO - parrotllm.training - step 75270 | epoch 0 | loss 3.3494 | lr 5.93e-04 | grad 0.2755 +2026-04-10 11:23:52 - INFO - parrotllm.training - step 75280 | epoch 0 | loss 3.4363 | lr 5.93e-04 | grad 0.2738 +2026-04-10 11:23:55 - INFO - parrotllm.training - step 75290 | epoch 0 | loss 3.4592 | lr 5.92e-04 | grad 0.2743 +2026-04-10 11:23:58 - INFO - parrotllm.training - step 75300 | epoch 0 | loss 3.3117 | lr 5.92e-04 | grad 0.2491 +2026-04-10 11:24:01 - INFO - parrotllm.training - step 75310 | epoch 0 | loss 3.3960 | lr 5.92e-04 | grad 0.2706 +2026-04-10 11:24:04 - INFO - parrotllm.training - step 75320 | epoch 0 | loss 3.4097 | lr 5.92e-04 | grad 0.2634 +2026-04-10 11:24:07 - INFO - parrotllm.training - step 75330 | epoch 0 | loss 3.3963 | lr 5.92e-04 | grad 0.2853 +2026-04-10 11:24:11 - INFO - parrotllm.training - step 75340 | epoch 0 | loss 3.4171 | lr 5.92e-04 | grad 0.2961 +2026-04-10 11:24:14 - INFO - parrotllm.training - step 75350 | epoch 0 | loss 3.4259 | lr 5.91e-04 | grad 0.2710 +2026-04-10 11:24:17 - INFO - parrotllm.training - step 75360 | epoch 0 | loss 3.3837 | lr 5.91e-04 | grad 0.2622 +2026-04-10 11:24:20 - INFO - parrotllm.training - step 75370 | epoch 0 | loss 3.3190 | lr 5.91e-04 | grad 0.2902 +2026-04-10 11:24:23 - INFO - parrotllm.training - step 75380 | epoch 0 | loss 3.4175 | lr 5.91e-04 | grad 0.2599 +2026-04-10 11:24:26 - INFO - parrotllm.training - step 75390 | epoch 0 | loss 3.4455 | lr 5.91e-04 | grad 0.2641 +2026-04-10 11:24:29 - INFO - parrotllm.training - step 75400 | epoch 0 | loss 3.3884 | lr 5.91e-04 | grad 0.2714 +2026-04-10 11:24:32 - INFO - parrotllm.training - step 75410 | epoch 0 | loss 3.2743 | lr 5.90e-04 | grad 0.3029 +2026-04-10 11:24:35 - INFO - parrotllm.training - step 75420 | epoch 0 | loss 3.3811 | lr 5.90e-04 | grad 0.2630 +2026-04-10 11:24:38 - INFO - parrotllm.training - step 75430 | epoch 0 | loss 3.2741 | lr 5.90e-04 | grad 0.2678 +2026-04-10 11:24:41 - INFO - parrotllm.training - step 75440 | epoch 0 | loss 3.4177 | lr 5.90e-04 | grad 0.2908 +2026-04-10 11:24:44 - INFO - parrotllm.training - step 75450 | epoch 0 | loss 3.4586 | lr 5.90e-04 | grad 0.2652 +2026-04-10 11:24:47 - INFO - parrotllm.training - step 75460 | epoch 0 | loss 3.3912 | lr 5.90e-04 | grad 0.2739 +2026-04-10 11:24:51 - INFO - parrotllm.training - step 75470 | epoch 0 | loss 3.2591 | lr 5.89e-04 | grad 0.2611 +2026-04-10 11:24:54 - INFO - parrotllm.training - step 75480 | epoch 0 | loss 3.3347 | lr 5.89e-04 | grad 0.3089 +2026-04-10 11:24:57 - INFO - parrotllm.training - step 75490 | epoch 0 | loss 3.4360 | lr 5.89e-04 | grad 0.2857 +2026-04-10 11:25:00 - INFO - parrotllm.training - step 75500 | epoch 0 | loss 3.3808 | lr 5.89e-04 | grad 0.2604 +2026-04-10 11:25:00 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:25:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:25:03 - INFO - parrotllm.training - Train: loss=3.3808, ppl=29.39 +2026-04-10 11:25:03 - INFO - parrotllm.training - Val: loss=3.3091, ppl=27.36 +2026-04-10 11:25:04 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3091_epoch_0000_step_0075500.pt +2026-04-10 11:25:05 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3093, min_delta=0.001000). +2026-04-10 11:25:05 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:25:08 - INFO - parrotllm.training - step 75510 | epoch 0 | loss 3.3589 | lr 5.89e-04 | grad 0.2917 +2026-04-10 11:25:12 - INFO - parrotllm.training - step 75520 | epoch 0 | loss 3.3351 | lr 5.89e-04 | grad 0.2896 +2026-04-10 11:25:15 - INFO - parrotllm.training - step 75530 | epoch 0 | loss 3.3465 | lr 5.88e-04 | grad 0.2621 +2026-04-10 11:25:18 - INFO - parrotllm.training - step 75540 | epoch 0 | loss 3.5334 | lr 5.88e-04 | grad 0.2664 +2026-04-10 11:25:21 - INFO - parrotllm.training - step 75550 | epoch 0 | loss 3.3904 | lr 5.88e-04 | grad 0.3021 +2026-04-10 11:25:24 - INFO - parrotllm.training - step 75560 | epoch 0 | loss 3.4208 | lr 5.88e-04 | grad 0.2721 +2026-04-10 11:25:27 - INFO - parrotllm.training - step 75570 | epoch 0 | loss 3.3337 | lr 5.88e-04 | grad 0.2702 +2026-04-10 11:25:30 - INFO - parrotllm.training - step 75580 | epoch 0 | loss 3.3012 | lr 5.88e-04 | grad 0.2861 +2026-04-10 11:25:33 - INFO - parrotllm.training - step 75590 | epoch 0 | loss 3.3435 | lr 5.87e-04 | grad 0.3007 +2026-04-10 11:25:36 - INFO - parrotllm.training - step 75600 | epoch 0 | loss 3.3866 | lr 5.87e-04 | grad 0.2684 +2026-04-10 11:25:39 - INFO - parrotllm.training - step 75610 | epoch 0 | loss 3.3561 | lr 5.87e-04 | grad 0.3223 +2026-04-10 11:25:42 - INFO - parrotllm.training - step 75620 | epoch 0 | loss 3.2585 | lr 5.87e-04 | grad 0.2889 +2026-04-10 11:25:45 - INFO - parrotllm.training - step 75630 | epoch 0 | loss 3.3900 | lr 5.87e-04 | grad 0.2841 +2026-04-10 11:25:48 - INFO - parrotllm.training - step 75640 | epoch 0 | loss 3.3790 | lr 5.87e-04 | grad 0.2636 +2026-04-10 11:25:51 - INFO - parrotllm.training - step 75650 | epoch 0 | loss 3.3603 | lr 5.86e-04 | grad 0.2841 +2026-04-10 11:25:54 - INFO - parrotllm.training - step 75660 | epoch 0 | loss 3.2962 | lr 5.86e-04 | grad 0.2824 +2026-04-10 11:25:57 - INFO - parrotllm.training - step 75670 | epoch 0 | loss 3.2805 | lr 5.86e-04 | grad 0.2824 +2026-04-10 11:26:00 - INFO - parrotllm.training - step 75680 | epoch 0 | loss 3.4690 | lr 5.86e-04 | grad 0.2960 +2026-04-10 11:26:03 - INFO - parrotllm.training - step 75690 | epoch 0 | loss 3.3698 | lr 5.86e-04 | grad 0.2878 +2026-04-10 11:26:06 - INFO - parrotllm.training - step 75700 | epoch 0 | loss 3.3239 | lr 5.86e-04 | grad 0.2758 +2026-04-10 11:26:09 - INFO - parrotllm.training - step 75710 | epoch 0 | loss 3.3348 | lr 5.85e-04 | grad 0.2592 +2026-04-10 11:26:13 - INFO - parrotllm.training - step 75720 | epoch 0 | loss 3.3969 | lr 5.85e-04 | grad 0.2597 +2026-04-10 11:26:16 - INFO - parrotllm.training - step 75730 | epoch 0 | loss 3.3970 | lr 5.85e-04 | grad 0.2683 +2026-04-10 11:26:19 - INFO - parrotllm.training - step 75740 | epoch 0 | loss 3.3370 | lr 5.85e-04 | grad 0.3010 +2026-04-10 11:26:22 - INFO - parrotllm.training - step 75750 | epoch 0 | loss 3.4180 | lr 5.85e-04 | grad 0.2912 +2026-04-10 11:26:25 - INFO - parrotllm.training - step 75760 | epoch 0 | loss 3.3688 | lr 5.85e-04 | grad 0.2721 +2026-04-10 11:26:28 - INFO - parrotllm.training - step 75770 | epoch 0 | loss 3.4261 | lr 5.84e-04 | grad 0.2527 +2026-04-10 11:26:31 - INFO - parrotllm.training - step 75780 | epoch 0 | loss 3.3978 | lr 5.84e-04 | grad 0.2723 +2026-04-10 11:26:34 - INFO - parrotllm.training - step 75790 | epoch 0 | loss 3.3215 | lr 5.84e-04 | grad 0.2871 +2026-04-10 11:26:37 - INFO - parrotllm.training - step 75800 | epoch 0 | loss 3.3468 | lr 5.84e-04 | grad 0.2892 +2026-04-10 11:26:40 - INFO - parrotllm.training - step 75810 | epoch 0 | loss 3.3199 | lr 5.84e-04 | grad 0.2925 +2026-04-10 11:26:43 - INFO - parrotllm.training - step 75820 | epoch 0 | loss 3.2350 | lr 5.84e-04 | grad 0.2579 +2026-04-10 11:26:46 - INFO - parrotllm.training - step 75830 | epoch 0 | loss 3.3358 | lr 5.83e-04 | grad 0.2572 +2026-04-10 11:26:49 - INFO - parrotllm.training - step 75840 | epoch 0 | loss 3.3480 | lr 5.83e-04 | grad 0.2570 +2026-04-10 11:26:52 - INFO - parrotllm.training - step 75850 | epoch 0 | loss 3.4443 | lr 5.83e-04 | grad 0.3159 +2026-04-10 11:26:55 - INFO - parrotllm.training - step 75860 | epoch 0 | loss 3.4048 | lr 5.83e-04 | grad 0.2937 +2026-04-10 11:26:58 - INFO - parrotllm.training - step 75870 | epoch 0 | loss 3.3552 | lr 5.83e-04 | grad 0.3067 +2026-04-10 11:27:01 - INFO - parrotllm.training - step 75880 | epoch 0 | loss 3.3882 | lr 5.83e-04 | grad 0.2940 +2026-04-10 11:27:04 - INFO - parrotllm.training - step 75890 | epoch 0 | loss 3.4933 | lr 5.82e-04 | grad 0.2737 +2026-04-10 11:27:07 - INFO - parrotllm.training - step 75900 | epoch 0 | loss 3.4247 | lr 5.82e-04 | grad 0.2703 +2026-04-10 11:27:10 - INFO - parrotllm.training - step 75910 | epoch 0 | loss 3.3904 | lr 5.82e-04 | grad 0.2655 +2026-04-10 11:27:14 - INFO - parrotllm.training - step 75920 | epoch 0 | loss 3.3922 | lr 5.82e-04 | grad 0.2701 +2026-04-10 11:27:17 - INFO - parrotllm.training - step 75930 | epoch 0 | loss 3.4419 | lr 5.82e-04 | grad 0.2949 +2026-04-10 11:27:20 - INFO - parrotllm.training - step 75940 | epoch 0 | loss 3.3816 | lr 5.82e-04 | grad 0.2765 +2026-04-10 11:27:23 - INFO - parrotllm.training - step 75950 | epoch 0 | loss 3.4368 | lr 5.81e-04 | grad 0.2604 +2026-04-10 11:27:26 - INFO - parrotllm.training - step 75960 | epoch 0 | loss 3.4834 | lr 5.81e-04 | grad 0.2776 +2026-04-10 11:27:29 - INFO - parrotllm.training - step 75970 | epoch 0 | loss 3.3325 | lr 5.81e-04 | grad 0.3029 +2026-04-10 11:27:32 - INFO - parrotllm.training - step 75980 | epoch 0 | loss 3.2748 | lr 5.81e-04 | grad 0.2940 +2026-04-10 11:27:35 - INFO - parrotllm.training - step 75990 | epoch 0 | loss 3.3322 | lr 5.81e-04 | grad 0.2845 +2026-04-10 11:27:38 - INFO - parrotllm.training - step 76000 | epoch 0 | loss 3.2948 | lr 5.81e-04 | grad 0.2721 +2026-04-10 11:27:38 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:27:38 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:27:41 - INFO - parrotllm.training - Train: loss=3.2948, ppl=26.97 +2026-04-10 11:27:41 - INFO - parrotllm.training - Val: loss=3.3063, ppl=27.28 +2026-04-10 11:27:41 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:27:42 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3063_epoch_0000_step_0076000.pt +2026-04-10 11:27:44 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:27:47 - INFO - parrotllm.training - step 76010 | epoch 0 | loss 3.3243 | lr 5.81e-04 | grad 0.2704 +2026-04-10 11:27:50 - INFO - parrotllm.training - step 76020 | epoch 0 | loss 3.3344 | lr 5.80e-04 | grad 0.2782 +2026-04-10 11:27:53 - INFO - parrotllm.training - step 76030 | epoch 0 | loss 3.4337 | lr 5.80e-04 | grad 0.3386 +2026-04-10 11:27:56 - INFO - parrotllm.training - step 76040 | epoch 0 | loss 3.5015 | lr 5.80e-04 | grad 0.2846 +2026-04-10 11:27:59 - INFO - parrotllm.training - step 76050 | epoch 0 | loss 3.3910 | lr 5.80e-04 | grad 0.3022 +2026-04-10 11:28:02 - INFO - parrotllm.training - step 76060 | epoch 0 | loss 3.3546 | lr 5.80e-04 | grad 0.2750 +2026-04-10 11:28:05 - INFO - parrotllm.training - step 76070 | epoch 0 | loss 3.3105 | lr 5.80e-04 | grad 0.2857 +2026-04-10 11:28:08 - INFO - parrotllm.training - step 76080 | epoch 0 | loss 3.2856 | lr 5.79e-04 | grad 0.2787 +2026-04-10 11:28:11 - INFO - parrotllm.training - step 76090 | epoch 0 | loss 3.4453 | lr 5.79e-04 | grad 0.2715 +2026-04-10 11:28:14 - INFO - parrotllm.training - step 76100 | epoch 0 | loss 3.4461 | lr 5.79e-04 | grad 0.2831 +2026-04-10 11:28:17 - INFO - parrotllm.training - step 76110 | epoch 0 | loss 3.4068 | lr 5.79e-04 | grad 0.2656 +2026-04-10 11:28:20 - INFO - parrotllm.training - step 76120 | epoch 0 | loss 3.3037 | lr 5.79e-04 | grad 0.2818 +2026-04-10 11:28:23 - INFO - parrotllm.training - step 76130 | epoch 0 | loss 3.2994 | lr 5.79e-04 | grad 0.2682 +2026-04-10 11:28:26 - INFO - parrotllm.training - step 76140 | epoch 0 | loss 3.3211 | lr 5.78e-04 | grad 0.2897 +2026-04-10 11:28:29 - INFO - parrotllm.training - step 76150 | epoch 0 | loss 3.3582 | lr 5.78e-04 | grad 0.2730 +2026-04-10 11:28:32 - INFO - parrotllm.training - step 76160 | epoch 0 | loss 3.4056 | lr 5.78e-04 | grad 0.2768 +2026-04-10 11:28:35 - INFO - parrotllm.training - step 76170 | epoch 0 | loss 3.3338 | lr 5.78e-04 | grad 0.2730 +2026-04-10 11:28:39 - INFO - parrotllm.training - step 76180 | epoch 0 | loss 3.3512 | lr 5.78e-04 | grad 0.2703 +2026-04-10 11:28:42 - INFO - parrotllm.training - step 76190 | epoch 0 | loss 3.3116 | lr 5.78e-04 | grad 0.2685 +2026-04-10 11:28:45 - INFO - parrotllm.training - step 76200 | epoch 0 | loss 3.4306 | lr 5.77e-04 | grad 0.2809 +2026-04-10 11:28:48 - INFO - parrotllm.training - step 76210 | epoch 0 | loss 3.4123 | lr 5.77e-04 | grad 0.2851 +2026-04-10 11:28:51 - INFO - parrotllm.training - step 76220 | epoch 0 | loss 3.3845 | lr 5.77e-04 | grad 0.2491 +2026-04-10 11:28:54 - INFO - parrotllm.training - step 76230 | epoch 0 | loss 3.3291 | lr 5.77e-04 | grad 0.3093 +2026-04-10 11:28:57 - INFO - parrotllm.training - step 76240 | epoch 0 | loss 3.2366 | lr 5.77e-04 | grad 0.2806 +2026-04-10 11:29:00 - INFO - parrotllm.training - step 76250 | epoch 0 | loss 3.3319 | lr 5.77e-04 | grad 0.2743 +2026-04-10 11:29:03 - INFO - parrotllm.training - step 76260 | epoch 0 | loss 3.4266 | lr 5.76e-04 | grad 0.2574 +2026-04-10 11:29:06 - INFO - parrotllm.training - step 76270 | epoch 0 | loss 3.4501 | lr 5.76e-04 | grad 0.2916 +2026-04-10 11:29:09 - INFO - parrotllm.training - step 76280 | epoch 0 | loss 3.3122 | lr 5.76e-04 | grad 0.2713 +2026-04-10 11:29:12 - INFO - parrotllm.training - step 76290 | epoch 0 | loss 3.3523 | lr 5.76e-04 | grad 0.3107 +2026-04-10 11:29:15 - INFO - parrotllm.training - step 76300 | epoch 0 | loss 3.4170 | lr 5.76e-04 | grad 0.2914 +2026-04-10 11:29:18 - INFO - parrotllm.training - step 76310 | epoch 0 | loss 3.4201 | lr 5.76e-04 | grad 0.2799 +2026-04-10 11:29:21 - INFO - parrotllm.training - step 76320 | epoch 0 | loss 3.4963 | lr 5.75e-04 | grad 0.2916 +2026-04-10 11:29:24 - INFO - parrotllm.training - step 76330 | epoch 0 | loss 3.3370 | lr 5.75e-04 | grad 0.2805 +2026-04-10 11:29:27 - INFO - parrotllm.training - step 76340 | epoch 0 | loss 3.4112 | lr 5.75e-04 | grad 0.2750 +2026-04-10 11:29:30 - INFO - parrotllm.training - step 76350 | epoch 0 | loss 3.3924 | lr 5.75e-04 | grad 0.2607 +2026-04-10 11:29:33 - INFO - parrotllm.training - step 76360 | epoch 0 | loss 3.3650 | lr 5.75e-04 | grad 0.2764 +2026-04-10 11:29:36 - INFO - parrotllm.training - step 76370 | epoch 0 | loss 3.3226 | lr 5.75e-04 | grad 0.2653 +2026-04-10 11:29:40 - INFO - parrotllm.training - step 76380 | epoch 0 | loss 3.3345 | lr 5.74e-04 | grad 0.2823 +2026-04-10 11:29:43 - INFO - parrotllm.training - step 76390 | epoch 0 | loss 3.4192 | lr 5.74e-04 | grad 0.2824 +2026-04-10 11:29:46 - INFO - parrotllm.training - step 76400 | epoch 0 | loss 3.3211 | lr 5.74e-04 | grad 0.2983 +2026-04-10 11:29:49 - INFO - parrotllm.training - step 76410 | epoch 0 | loss 3.2627 | lr 5.74e-04 | grad 0.2876 +2026-04-10 11:29:52 - INFO - parrotllm.training - step 76420 | epoch 0 | loss 3.4574 | lr 5.74e-04 | grad 0.2884 +2026-04-10 11:29:55 - INFO - parrotllm.training - step 76430 | epoch 0 | loss 3.3912 | lr 5.74e-04 | grad 0.2650 +2026-04-10 11:29:58 - INFO - parrotllm.training - step 76440 | epoch 0 | loss 3.3702 | lr 5.73e-04 | grad 0.3018 +2026-04-10 11:30:01 - INFO - parrotllm.training - step 76450 | epoch 0 | loss 3.3558 | lr 5.73e-04 | grad 0.2646 +2026-04-10 11:30:04 - INFO - parrotllm.training - step 76460 | epoch 0 | loss 3.2758 | lr 5.73e-04 | grad 0.2796 +2026-04-10 11:30:07 - INFO - parrotllm.training - step 76470 | epoch 0 | loss 3.3736 | lr 5.73e-04 | grad 0.2739 +2026-04-10 11:30:10 - INFO - parrotllm.training - step 76480 | epoch 0 | loss 3.4056 | lr 5.73e-04 | grad 0.2923 +2026-04-10 11:30:13 - INFO - parrotllm.training - step 76490 | epoch 0 | loss 3.3489 | lr 5.73e-04 | grad 0.3084 +2026-04-10 11:30:16 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 76529/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 11:30:16 - INFO - parrotllm.training - step 76500 | epoch 0 | loss 3.4264 | lr 5.72e-04 | grad 0.2882 +2026-04-10 11:30:16 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:30:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:30:20 - INFO - parrotllm.training - Train: loss=3.4264, ppl=30.77 +2026-04-10 11:30:20 - INFO - parrotllm.training - Val: loss=3.3044, ppl=27.23 +2026-04-10 11:30:20 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:30:20 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3044_epoch_0000_step_0076500.pt +2026-04-10 11:30:22 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:30:25 - INFO - parrotllm.training - step 76510 | epoch 0 | loss 3.3285 | lr 5.72e-04 | grad 0.2792 +2026-04-10 11:30:28 - INFO - parrotllm.training - step 76520 | epoch 0 | loss 3.3164 | lr 5.72e-04 | grad 0.2730 +2026-04-10 11:30:31 - INFO - parrotllm.training - step 76530 | epoch 0 | loss 3.3365 | lr 5.72e-04 | grad 0.3101 +2026-04-10 11:30:34 - INFO - parrotllm.training - step 76540 | epoch 0 | loss 3.3019 | lr 5.72e-04 | grad 0.2704 +2026-04-10 11:30:37 - INFO - parrotllm.training - step 76550 | epoch 0 | loss 3.2669 | lr 5.72e-04 | grad 0.2935 +2026-04-10 11:30:40 - INFO - parrotllm.training - step 76560 | epoch 0 | loss 3.4435 | lr 5.72e-04 | grad 0.2792 +2026-04-10 11:30:43 - INFO - parrotllm.training - step 76570 | epoch 0 | loss 3.3987 | lr 5.71e-04 | grad 0.2887 +2026-04-10 11:30:46 - INFO - parrotllm.training - step 76580 | epoch 0 | loss 3.2126 | lr 5.71e-04 | grad 0.2899 +2026-04-10 11:30:49 - INFO - parrotllm.training - step 76590 | epoch 0 | loss 3.4975 | lr 5.71e-04 | grad 0.2952 +2026-04-10 11:30:53 - INFO - parrotllm.training - step 76600 | epoch 0 | loss 3.3940 | lr 5.71e-04 | grad 0.2700 +2026-04-10 11:30:56 - INFO - parrotllm.training - step 76610 | epoch 0 | loss 3.3500 | lr 5.71e-04 | grad 0.3065 +2026-04-10 11:30:59 - INFO - parrotllm.training - step 76620 | epoch 0 | loss 3.3826 | lr 5.71e-04 | grad 0.2992 +2026-04-10 11:31:02 - INFO - parrotllm.training - step 76630 | epoch 0 | loss 3.4448 | lr 5.70e-04 | grad 0.2870 +2026-04-10 11:31:05 - INFO - parrotllm.training - step 76640 | epoch 0 | loss 3.3835 | lr 5.70e-04 | grad 0.2806 +2026-04-10 11:31:08 - INFO - parrotllm.training - step 76650 | epoch 0 | loss 3.2413 | lr 5.70e-04 | grad 0.2826 +2026-04-10 11:31:11 - INFO - parrotllm.training - step 76660 | epoch 0 | loss 3.2427 | lr 5.70e-04 | grad 0.2694 +2026-04-10 11:31:14 - INFO - parrotllm.training - step 76670 | epoch 0 | loss 3.3159 | lr 5.70e-04 | grad 0.2509 +2026-04-10 11:31:17 - INFO - parrotllm.training - step 76680 | epoch 0 | loss 3.2953 | lr 5.70e-04 | grad 0.2958 +2026-04-10 11:31:20 - INFO - parrotllm.training - step 76690 | epoch 0 | loss 3.3595 | lr 5.69e-04 | grad 0.2581 +2026-04-10 11:31:23 - INFO - parrotllm.training - step 76700 | epoch 0 | loss 3.3487 | lr 5.69e-04 | grad 0.2830 +2026-04-10 11:31:26 - INFO - parrotllm.training - step 76710 | epoch 0 | loss 3.4336 | lr 5.69e-04 | grad 0.2816 +2026-04-10 11:31:29 - INFO - parrotllm.training - step 76720 | epoch 0 | loss 3.2682 | lr 5.69e-04 | grad 0.2733 +2026-04-10 11:31:32 - INFO - parrotllm.training - step 76730 | epoch 0 | loss 3.3604 | lr 5.69e-04 | grad 0.3043 +2026-04-10 11:31:35 - INFO - parrotllm.training - step 76740 | epoch 0 | loss 3.3192 | lr 5.69e-04 | grad 0.2834 +2026-04-10 11:31:38 - INFO - parrotllm.training - step 76750 | epoch 0 | loss 3.3822 | lr 5.68e-04 | grad 0.2738 +2026-04-10 11:31:41 - INFO - parrotllm.training - step 76760 | epoch 0 | loss 3.3701 | lr 5.68e-04 | grad 0.2657 +2026-04-10 11:31:44 - INFO - parrotllm.training - step 76770 | epoch 0 | loss 3.3661 | lr 5.68e-04 | grad 0.2775 +2026-04-10 11:31:47 - INFO - parrotllm.training - step 76780 | epoch 0 | loss 3.3593 | lr 5.68e-04 | grad 0.2653 +2026-04-10 11:31:50 - INFO - parrotllm.training - step 76790 | epoch 0 | loss 3.2022 | lr 5.68e-04 | grad 0.3193 +2026-04-10 11:31:53 - INFO - parrotllm.training - step 76800 | epoch 0 | loss 3.4165 | lr 5.68e-04 | grad 0.2734 +2026-04-10 11:31:56 - INFO - parrotllm.training - step 76810 | epoch 0 | loss 3.4193 | lr 5.67e-04 | grad 0.2799 +2026-04-10 11:32:00 - INFO - parrotllm.training - step 76820 | epoch 0 | loss 3.3778 | lr 5.67e-04 | grad 0.2715 +2026-04-10 11:32:03 - INFO - parrotllm.training - step 76830 | epoch 0 | loss 3.3577 | lr 5.67e-04 | grad 0.2656 +2026-04-10 11:32:06 - INFO - parrotllm.training - step 76840 | epoch 0 | loss 3.3933 | lr 5.67e-04 | grad 0.2961 +2026-04-10 11:32:09 - INFO - parrotllm.training - step 76850 | epoch 0 | loss 3.3565 | lr 5.67e-04 | grad 0.2584 +2026-04-10 11:32:12 - INFO - parrotllm.training - step 76860 | epoch 0 | loss 3.3104 | lr 5.67e-04 | grad 0.2649 +2026-04-10 11:32:15 - INFO - parrotllm.training - step 76870 | epoch 0 | loss 3.3688 | lr 5.66e-04 | grad 0.2931 +2026-04-10 11:32:18 - INFO - parrotllm.training - step 76880 | epoch 0 | loss 3.3623 | lr 5.66e-04 | grad 0.2705 +2026-04-10 11:32:21 - INFO - parrotllm.training - step 76890 | epoch 0 | loss 3.4134 | lr 5.66e-04 | grad 0.2710 +2026-04-10 11:32:24 - INFO - parrotllm.training - step 76900 | epoch 0 | loss 3.3810 | lr 5.66e-04 | grad 0.2643 +2026-04-10 11:32:27 - INFO - parrotllm.training - step 76910 | epoch 0 | loss 3.3980 | lr 5.66e-04 | grad 0.2687 +2026-04-10 11:32:30 - INFO - parrotllm.training - step 76920 | epoch 0 | loss 3.4634 | lr 5.66e-04 | grad 0.2654 +2026-04-10 11:32:33 - INFO - parrotllm.training - step 76930 | epoch 0 | loss 3.2149 | lr 5.65e-04 | grad 0.3059 +2026-04-10 11:32:36 - INFO - parrotllm.training - step 76940 | epoch 0 | loss 3.4795 | lr 5.65e-04 | grad 0.2724 +2026-04-10 11:32:39 - INFO - parrotllm.training - step 76950 | epoch 0 | loss 3.3160 | lr 5.65e-04 | grad 0.2599 +2026-04-10 11:32:42 - INFO - parrotllm.training - step 76960 | epoch 0 | loss 3.3738 | lr 5.65e-04 | grad 0.2900 +2026-04-10 11:32:45 - INFO - parrotllm.training - step 76970 | epoch 0 | loss 3.3240 | lr 5.65e-04 | grad 0.2694 +2026-04-10 11:32:48 - INFO - parrotllm.training - step 76980 | epoch 0 | loss 3.3377 | lr 5.65e-04 | grad 0.2522 +2026-04-10 11:32:51 - INFO - parrotllm.training - step 76990 | epoch 0 | loss 3.3379 | lr 5.65e-04 | grad 0.2489 +2026-04-10 11:32:54 - INFO - parrotllm.training - step 77000 | epoch 0 | loss 3.3429 | lr 5.64e-04 | grad 0.2925 +2026-04-10 11:32:54 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:32:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:32:57 - INFO - parrotllm.training - Train: loss=3.3429, ppl=28.30 +2026-04-10 11:32:57 - INFO - parrotllm.training - Val: loss=3.3032, ppl=27.20 +2026-04-10 11:32:57 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:32:58 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3032_epoch_0000_step_0077000.pt +2026-04-10 11:33:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:33:03 - INFO - parrotllm.training - step 77010 | epoch 0 | loss 3.4033 | lr 5.64e-04 | grad 0.2778 +2026-04-10 11:33:07 - INFO - parrotllm.training - step 77020 | epoch 0 | loss 3.3194 | lr 5.64e-04 | grad 0.2591 +2026-04-10 11:33:10 - INFO - parrotllm.training - step 77030 | epoch 0 | loss 3.3088 | lr 5.64e-04 | grad 0.2749 +2026-04-10 11:33:13 - INFO - parrotllm.training - step 77040 | epoch 0 | loss 3.4966 | lr 5.64e-04 | grad 0.2921 +2026-04-10 11:33:16 - INFO - parrotllm.training - step 77050 | epoch 0 | loss 3.3211 | lr 5.64e-04 | grad 0.2935 +2026-04-10 11:33:19 - INFO - parrotllm.training - step 77060 | epoch 0 | loss 3.4834 | lr 5.63e-04 | grad 0.2918 +2026-04-10 11:33:22 - INFO - parrotllm.training - step 77070 | epoch 0 | loss 3.4108 | lr 5.63e-04 | grad 0.2739 +2026-04-10 11:33:25 - INFO - parrotllm.training - step 77080 | epoch 0 | loss 3.3439 | lr 5.63e-04 | grad 0.2670 +2026-04-10 11:33:28 - INFO - parrotllm.training - step 77090 | epoch 0 | loss 3.3617 | lr 5.63e-04 | grad 0.2842 +2026-04-10 11:33:31 - INFO - parrotllm.training - step 77100 | epoch 0 | loss 3.3823 | lr 5.63e-04 | grad 0.3234 +2026-04-10 11:33:34 - INFO - parrotllm.training - step 77110 | epoch 0 | loss 3.4620 | lr 5.63e-04 | grad 0.2803 +2026-04-10 11:33:37 - INFO - parrotllm.training - step 77120 | epoch 0 | loss 3.2992 | lr 5.62e-04 | grad 0.2961 +2026-04-10 11:33:40 - INFO - parrotllm.training - step 77130 | epoch 0 | loss 3.2477 | lr 5.62e-04 | grad 0.3259 +2026-04-10 11:33:43 - INFO - parrotllm.training - step 77140 | epoch 0 | loss 3.3312 | lr 5.62e-04 | grad 0.2810 +2026-04-10 11:33:46 - INFO - parrotllm.training - step 77150 | epoch 0 | loss 3.3796 | lr 5.62e-04 | grad 0.2773 +2026-04-10 11:33:49 - INFO - parrotllm.training - step 77160 | epoch 0 | loss 3.3172 | lr 5.62e-04 | grad 0.3043 +2026-04-10 11:33:52 - INFO - parrotllm.training - step 77170 | epoch 0 | loss 3.3980 | lr 5.62e-04 | grad 0.3011 +2026-04-10 11:33:55 - INFO - parrotllm.training - step 77180 | epoch 0 | loss 3.2956 | lr 5.61e-04 | grad 0.2801 +2026-04-10 11:33:58 - INFO - parrotllm.training - step 77190 | epoch 0 | loss 3.2861 | lr 5.61e-04 | grad 0.3000 +2026-04-10 11:34:01 - INFO - parrotllm.training - step 77200 | epoch 0 | loss 3.3535 | lr 5.61e-04 | grad 0.2814 +2026-04-10 11:34:04 - INFO - parrotllm.training - step 77210 | epoch 0 | loss 3.3812 | lr 5.61e-04 | grad 0.2904 +2026-04-10 11:34:07 - INFO - parrotllm.training - step 77220 | epoch 0 | loss 3.4457 | lr 5.61e-04 | grad 0.2858 +2026-04-10 11:34:10 - INFO - parrotllm.training - step 77230 | epoch 0 | loss 3.3699 | lr 5.61e-04 | grad 0.2815 +2026-04-10 11:34:14 - INFO - parrotllm.training - step 77240 | epoch 0 | loss 3.3340 | lr 5.60e-04 | grad 0.3817 +2026-04-10 11:34:17 - INFO - parrotllm.training - step 77250 | epoch 0 | loss 3.3284 | lr 5.60e-04 | grad 0.2973 +2026-04-10 11:34:20 - INFO - parrotllm.training - step 77260 | epoch 0 | loss 3.4047 | lr 5.60e-04 | grad 0.2903 +2026-04-10 11:34:23 - INFO - parrotllm.training - step 77270 | epoch 0 | loss 3.4305 | lr 5.60e-04 | grad 0.2658 +2026-04-10 11:34:26 - INFO - parrotllm.training - step 77280 | epoch 0 | loss 3.4264 | lr 5.60e-04 | grad 0.2803 +2026-04-10 11:34:29 - INFO - parrotllm.training - step 77290 | epoch 0 | loss 3.3692 | lr 5.60e-04 | grad 0.2722 +2026-04-10 11:34:32 - INFO - parrotllm.training - step 77300 | epoch 0 | loss 3.2439 | lr 5.59e-04 | grad 0.2659 +2026-04-10 11:34:35 - INFO - parrotllm.training - step 77310 | epoch 0 | loss 3.3869 | lr 5.59e-04 | grad 0.2640 +2026-04-10 11:34:38 - INFO - parrotllm.training - step 77320 | epoch 0 | loss 3.3224 | lr 5.59e-04 | grad 0.2713 +2026-04-10 11:34:41 - INFO - parrotllm.training - step 77330 | epoch 0 | loss 3.2684 | lr 5.59e-04 | grad 0.2769 +2026-04-10 11:34:44 - INFO - parrotllm.training - step 77340 | epoch 0 | loss 3.3737 | lr 5.59e-04 | grad 0.2586 +2026-04-10 11:34:47 - INFO - parrotllm.training - step 77350 | epoch 0 | loss 3.4002 | lr 5.59e-04 | grad 0.2666 +2026-04-10 11:34:50 - INFO - parrotllm.training - step 77360 | epoch 0 | loss 3.3245 | lr 5.59e-04 | grad 0.2969 +2026-04-10 11:34:53 - INFO - parrotllm.training - step 77370 | epoch 0 | loss 3.4102 | lr 5.58e-04 | grad 0.3207 +2026-04-10 11:34:56 - INFO - parrotllm.training - step 77380 | epoch 0 | loss 3.3392 | lr 5.58e-04 | grad 0.2789 +2026-04-10 11:34:59 - INFO - parrotllm.training - step 77390 | epoch 0 | loss 3.3713 | lr 5.58e-04 | grad 0.2935 +2026-04-10 11:35:02 - INFO - parrotllm.training - step 77400 | epoch 0 | loss 3.2526 | lr 5.58e-04 | grad 0.2818 +2026-04-10 11:35:05 - INFO - parrotllm.training - step 77410 | epoch 0 | loss 3.3259 | lr 5.58e-04 | grad 0.2661 +2026-04-10 11:35:08 - INFO - parrotllm.training - step 77420 | epoch 0 | loss 3.4010 | lr 5.58e-04 | grad 0.2854 +2026-04-10 11:35:11 - INFO - parrotllm.training - step 77430 | epoch 0 | loss 3.3878 | lr 5.57e-04 | grad 0.2819 +2026-04-10 11:35:14 - INFO - parrotllm.training - step 77440 | epoch 0 | loss 3.3362 | lr 5.57e-04 | grad 0.3203 +2026-04-10 11:35:18 - INFO - parrotllm.training - step 77450 | epoch 0 | loss 3.3156 | lr 5.57e-04 | grad 0.2810 +2026-04-10 11:35:21 - INFO - parrotllm.training - step 77460 | epoch 0 | loss 3.4061 | lr 5.57e-04 | grad 0.3119 +2026-04-10 11:35:24 - INFO - parrotllm.training - step 77470 | epoch 0 | loss 3.4626 | lr 5.57e-04 | grad 0.2942 +2026-04-10 11:35:27 - INFO - parrotllm.training - step 77480 | epoch 0 | loss 3.3549 | lr 5.57e-04 | grad 0.2581 +2026-04-10 11:35:30 - INFO - parrotllm.training - step 77490 | epoch 0 | loss 3.3022 | lr 5.56e-04 | grad 0.3066 +2026-04-10 11:35:33 - INFO - parrotllm.training - step 77500 | epoch 0 | loss 3.4286 | lr 5.56e-04 | grad 0.2709 +2026-04-10 11:35:33 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:35:33 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:35:36 - INFO - parrotllm.training - Train: loss=3.4286, ppl=30.83 +2026-04-10 11:35:36 - INFO - parrotllm.training - Val: loss=3.3032, ppl=27.20 +2026-04-10 11:35:37 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3032_epoch_0000_step_0077500.pt +2026-04-10 11:35:39 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.3032, min_delta=0.001000). +2026-04-10 11:35:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:35:40 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0077500.pt +2026-04-10 11:35:44 - INFO - parrotllm.training - step 77510 | epoch 0 | loss 3.4413 | lr 5.56e-04 | grad 0.3067 +2026-04-10 11:35:47 - INFO - parrotllm.training - step 77520 | epoch 0 | loss 3.3801 | lr 5.56e-04 | grad 0.2661 +2026-04-10 11:35:51 - INFO - parrotllm.training - step 77530 | epoch 0 | loss 3.3537 | lr 5.56e-04 | grad 0.2899 +2026-04-10 11:35:54 - INFO - parrotllm.training - step 77540 | epoch 0 | loss 3.3543 | lr 5.56e-04 | grad 0.2660 +2026-04-10 11:35:57 - INFO - parrotllm.training - step 77550 | epoch 0 | loss 3.3202 | lr 5.55e-04 | grad 0.2534 +2026-04-10 11:36:00 - INFO - parrotllm.training - step 77560 | epoch 0 | loss 3.3205 | lr 5.55e-04 | grad 0.2823 +2026-04-10 11:36:03 - INFO - parrotllm.training - step 77570 | epoch 0 | loss 3.2670 | lr 5.55e-04 | grad 0.2633 +2026-04-10 11:36:06 - INFO - parrotllm.training - step 77580 | epoch 0 | loss 3.3809 | lr 5.55e-04 | grad 0.2902 +2026-04-10 11:36:09 - INFO - parrotllm.training - step 77590 | epoch 0 | loss 3.4335 | lr 5.55e-04 | grad 0.2826 +2026-04-10 11:36:12 - INFO - parrotllm.training - step 77600 | epoch 0 | loss 3.3114 | lr 5.55e-04 | grad 0.3041 +2026-04-10 11:36:15 - INFO - parrotllm.training - step 77610 | epoch 0 | loss 3.3496 | lr 5.54e-04 | grad 0.3149 +2026-04-10 11:36:18 - INFO - parrotllm.training - step 77620 | epoch 0 | loss 3.4269 | lr 5.54e-04 | grad 0.2573 +2026-04-10 11:36:21 - INFO - parrotllm.training - step 77630 | epoch 0 | loss 3.4021 | lr 5.54e-04 | grad 0.2958 +2026-04-10 11:36:24 - INFO - parrotllm.training - step 77640 | epoch 0 | loss 3.3687 | lr 5.54e-04 | grad 0.2797 +2026-04-10 11:36:27 - INFO - parrotllm.training - step 77650 | epoch 0 | loss 3.4098 | lr 5.54e-04 | grad 0.2719 +2026-04-10 11:36:30 - INFO - parrotllm.training - step 77660 | epoch 0 | loss 3.3292 | lr 5.54e-04 | grad 0.2545 +2026-04-10 11:36:33 - INFO - parrotllm.training - step 77670 | epoch 0 | loss 3.3928 | lr 5.54e-04 | grad 0.2883 +2026-04-10 11:36:36 - INFO - parrotllm.training - step 77680 | epoch 0 | loss 3.3833 | lr 5.53e-04 | grad 0.2653 +2026-04-10 11:36:39 - INFO - parrotllm.training - step 77690 | epoch 0 | loss 3.4131 | lr 5.53e-04 | grad 0.2718 +2026-04-10 11:36:42 - INFO - parrotllm.training - step 77700 | epoch 0 | loss 3.3856 | lr 5.53e-04 | grad 0.2702 +2026-04-10 11:36:45 - INFO - parrotllm.training - step 77710 | epoch 0 | loss 3.4023 | lr 5.53e-04 | grad 0.2638 +2026-04-10 11:36:48 - INFO - parrotllm.training - step 77720 | epoch 0 | loss 3.3742 | lr 5.53e-04 | grad 0.2702 +2026-04-10 11:36:51 - INFO - parrotllm.training - step 77730 | epoch 0 | loss 3.3315 | lr 5.53e-04 | grad 0.2802 +2026-04-10 11:36:54 - INFO - parrotllm.training - step 77740 | epoch 0 | loss 3.2478 | lr 5.52e-04 | grad 0.2998 +2026-04-10 11:36:57 - INFO - parrotllm.training - step 77750 | epoch 0 | loss 3.3555 | lr 5.52e-04 | grad 0.2663 +2026-04-10 11:37:01 - INFO - parrotllm.training - step 77760 | epoch 0 | loss 3.2971 | lr 5.52e-04 | grad 0.2606 +2026-04-10 11:37:04 - INFO - parrotllm.training - step 77770 | epoch 0 | loss 3.4261 | lr 5.52e-04 | grad 0.2665 +2026-04-10 11:37:07 - INFO - parrotllm.training - step 77780 | epoch 0 | loss 3.3746 | lr 5.52e-04 | grad 0.2930 +2026-04-10 11:37:10 - INFO - parrotllm.training - step 77790 | epoch 0 | loss 3.3424 | lr 5.52e-04 | grad 0.2638 +2026-04-10 11:37:13 - INFO - parrotllm.training - step 77800 | epoch 0 | loss 3.4219 | lr 5.51e-04 | grad 0.2780 +2026-04-10 11:37:16 - INFO - parrotllm.training - step 77810 | epoch 0 | loss 3.3930 | lr 5.51e-04 | grad 0.2753 +2026-04-10 11:37:19 - INFO - parrotllm.training - step 77820 | epoch 0 | loss 3.3697 | lr 5.51e-04 | grad 0.2691 +2026-04-10 11:37:22 - INFO - parrotllm.training - step 77830 | epoch 0 | loss 3.2743 | lr 5.51e-04 | grad 0.2956 +2026-04-10 11:37:25 - INFO - parrotllm.training - step 77840 | epoch 0 | loss 3.2871 | lr 5.51e-04 | grad 0.3154 +2026-04-10 11:37:28 - INFO - parrotllm.training - step 77850 | epoch 0 | loss 3.4814 | lr 5.51e-04 | grad 0.2835 +2026-04-10 11:37:31 - INFO - parrotllm.training - step 77860 | epoch 0 | loss 3.3408 | lr 5.50e-04 | grad 0.2958 +2026-04-10 11:37:34 - INFO - parrotllm.training - step 77870 | epoch 0 | loss 3.2872 | lr 5.50e-04 | grad 0.2683 +2026-04-10 11:37:37 - INFO - parrotllm.training - step 77880 | epoch 0 | loss 3.4180 | lr 5.50e-04 | grad 0.2693 +2026-04-10 11:37:40 - INFO - parrotllm.training - step 77890 | epoch 0 | loss 3.3564 | lr 5.50e-04 | grad 0.2678 +2026-04-10 11:37:43 - INFO - parrotllm.training - step 77900 | epoch 0 | loss 3.3241 | lr 5.50e-04 | grad 0.2895 +2026-04-10 11:37:46 - INFO - parrotllm.training - step 77910 | epoch 0 | loss 3.3049 | lr 5.50e-04 | grad 0.2821 +2026-04-10 11:37:49 - INFO - parrotllm.training - step 77920 | epoch 0 | loss 3.3957 | lr 5.50e-04 | grad 0.3025 +2026-04-10 11:37:52 - INFO - parrotllm.training - step 77930 | epoch 0 | loss 3.4367 | lr 5.49e-04 | grad 0.2934 +2026-04-10 11:37:55 - INFO - parrotllm.training - step 77940 | epoch 0 | loss 3.3807 | lr 5.49e-04 | grad 0.2918 +2026-04-10 11:37:58 - INFO - parrotllm.training - step 77950 | epoch 0 | loss 3.3968 | lr 5.49e-04 | grad 0.3014 +2026-04-10 11:38:02 - INFO - parrotllm.training - step 77960 | epoch 0 | loss 3.3722 | lr 5.49e-04 | grad 0.2811 +2026-04-10 11:38:05 - INFO - parrotllm.training - step 77970 | epoch 0 | loss 3.4953 | lr 5.49e-04 | grad 0.2972 +2026-04-10 11:38:08 - INFO - parrotllm.training - step 77980 | epoch 0 | loss 3.3329 | lr 5.49e-04 | grad 0.2858 +2026-04-10 11:38:11 - INFO - parrotllm.training - step 77990 | epoch 0 | loss 3.3901 | lr 5.48e-04 | grad 0.2741 +2026-04-10 11:38:14 - INFO - parrotllm.training - step 78000 | epoch 0 | loss 3.4668 | lr 5.48e-04 | grad 0.2620 +2026-04-10 11:38:14 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:38:14 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:38:17 - INFO - parrotllm.training - Train: loss=3.4668, ppl=32.03 +2026-04-10 11:38:17 - INFO - parrotllm.training - Val: loss=3.3025, ppl=27.18 +2026-04-10 11:38:18 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3025_epoch_0000_step_0078000.pt +2026-04-10 11:38:20 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.3032, min_delta=0.001000). +2026-04-10 11:38:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:38:23 - INFO - parrotllm.training - step 78010 | epoch 0 | loss 3.4519 | lr 5.48e-04 | grad 0.2724 +2026-04-10 11:38:26 - INFO - parrotllm.training - step 78020 | epoch 0 | loss 3.4175 | lr 5.48e-04 | grad 0.2628 +2026-04-10 11:38:29 - INFO - parrotllm.training - step 78030 | epoch 0 | loss 3.4131 | lr 5.48e-04 | grad 0.2670 +2026-04-10 11:38:32 - INFO - parrotllm.training - step 78040 | epoch 0 | loss 3.3367 | lr 5.48e-04 | grad 0.2671 +2026-04-10 11:38:35 - INFO - parrotllm.training - step 78050 | epoch 0 | loss 3.2959 | lr 5.47e-04 | grad 0.2869 +2026-04-10 11:38:38 - INFO - parrotllm.training - step 78060 | epoch 0 | loss 3.3995 | lr 5.47e-04 | grad 0.2710 +2026-04-10 11:38:41 - INFO - parrotllm.training - step 78070 | epoch 0 | loss 3.4480 | lr 5.47e-04 | grad 0.2872 +2026-04-10 11:38:44 - INFO - parrotllm.training - step 78080 | epoch 0 | loss 3.4213 | lr 5.47e-04 | grad 0.2742 +2026-04-10 11:38:47 - INFO - parrotllm.training - step 78090 | epoch 0 | loss 3.4038 | lr 5.47e-04 | grad 0.2584 +2026-04-10 11:38:50 - INFO - parrotllm.training - step 78100 | epoch 0 | loss 3.2962 | lr 5.47e-04 | grad 0.2888 +2026-04-10 11:38:53 - INFO - parrotllm.training - step 78110 | epoch 0 | loss 3.3054 | lr 5.46e-04 | grad 0.2741 +2026-04-10 11:38:57 - INFO - parrotllm.training - step 78120 | epoch 0 | loss 3.3017 | lr 5.46e-04 | grad 0.3142 +2026-04-10 11:39:00 - INFO - parrotllm.training - step 78130 | epoch 0 | loss 3.2654 | lr 5.46e-04 | grad 0.2828 +2026-04-10 11:39:03 - INFO - parrotllm.training - step 78140 | epoch 0 | loss 3.3583 | lr 5.46e-04 | grad 0.2747 +2026-04-10 11:39:06 - INFO - parrotllm.training - step 78150 | epoch 0 | loss 3.4463 | lr 5.46e-04 | grad 0.2961 +2026-04-10 11:39:09 - INFO - parrotllm.training - step 78160 | epoch 0 | loss 3.3109 | lr 5.46e-04 | grad 0.2805 +2026-04-10 11:39:12 - INFO - parrotllm.training - step 78170 | epoch 0 | loss 3.4098 | lr 5.45e-04 | grad 0.2700 +2026-04-10 11:39:15 - INFO - parrotllm.training - step 78180 | epoch 0 | loss 3.3204 | lr 5.45e-04 | grad 0.2690 +2026-04-10 11:39:18 - INFO - parrotllm.training - step 78190 | epoch 0 | loss 3.3462 | lr 5.45e-04 | grad 0.2644 +2026-04-10 11:39:21 - INFO - parrotllm.training - step 78200 | epoch 0 | loss 3.3465 | lr 5.45e-04 | grad 0.2565 +2026-04-10 11:39:24 - INFO - parrotllm.training - step 78210 | epoch 0 | loss 3.3034 | lr 5.45e-04 | grad 0.2701 +2026-04-10 11:39:27 - INFO - parrotllm.training - step 78220 | epoch 0 | loss 3.4573 | lr 5.45e-04 | grad 0.3012 +2026-04-10 11:39:30 - INFO - parrotllm.training - step 78230 | epoch 0 | loss 3.3317 | lr 5.45e-04 | grad 0.2902 +2026-04-10 11:39:33 - INFO - parrotllm.training - step 78240 | epoch 0 | loss 3.3930 | lr 5.44e-04 | grad 0.2834 +2026-04-10 11:39:36 - INFO - parrotllm.training - step 78250 | epoch 0 | loss 3.3671 | lr 5.44e-04 | grad 0.2632 +2026-04-10 11:39:39 - INFO - parrotllm.training - step 78260 | epoch 0 | loss 3.4468 | lr 5.44e-04 | grad 0.2841 +2026-04-10 11:39:42 - INFO - parrotllm.training - step 78270 | epoch 0 | loss 3.4428 | lr 5.44e-04 | grad 0.3005 +2026-04-10 11:39:45 - INFO - parrotllm.training - step 78280 | epoch 0 | loss 3.3279 | lr 5.44e-04 | grad 0.2934 +2026-04-10 11:39:48 - INFO - parrotllm.training - step 78290 | epoch 0 | loss 3.4055 | lr 5.44e-04 | grad 0.2746 +2026-04-10 11:39:51 - INFO - parrotllm.training - step 78300 | epoch 0 | loss 3.3285 | lr 5.43e-04 | grad 0.2920 +2026-04-10 11:39:54 - INFO - parrotllm.training - step 78310 | epoch 0 | loss 3.4025 | lr 5.43e-04 | grad 0.2872 +2026-04-10 11:39:58 - INFO - parrotllm.training - step 78320 | epoch 0 | loss 3.3335 | lr 5.43e-04 | grad 0.2928 +2026-04-10 11:40:01 - INFO - parrotllm.training - step 78330 | epoch 0 | loss 3.3225 | lr 5.43e-04 | grad 0.3025 +2026-04-10 11:40:04 - INFO - parrotllm.training - step 78340 | epoch 0 | loss 3.4034 | lr 5.43e-04 | grad 0.3049 +2026-04-10 11:40:07 - INFO - parrotllm.training - step 78350 | epoch 0 | loss 3.3751 | lr 5.43e-04 | grad 0.2965 +2026-04-10 11:40:10 - INFO - parrotllm.training - step 78360 | epoch 0 | loss 3.3222 | lr 5.42e-04 | grad 0.2925 +2026-04-10 11:40:13 - INFO - parrotllm.training - step 78370 | epoch 0 | loss 3.4529 | lr 5.42e-04 | grad 0.2881 +2026-04-10 11:40:16 - INFO - parrotllm.training - step 78380 | epoch 0 | loss 3.4290 | lr 5.42e-04 | grad 0.2670 +2026-04-10 11:40:19 - INFO - parrotllm.training - step 78390 | epoch 0 | loss 3.4379 | lr 5.42e-04 | grad 0.2700 +2026-04-10 11:40:22 - INFO - parrotllm.training - step 78400 | epoch 0 | loss 3.3891 | lr 5.42e-04 | grad 0.2710 +2026-04-10 11:40:25 - INFO - parrotllm.training - step 78410 | epoch 0 | loss 3.3161 | lr 5.42e-04 | grad 0.2526 +2026-04-10 11:40:28 - INFO - parrotllm.training - step 78420 | epoch 0 | loss 3.4292 | lr 5.42e-04 | grad 0.2796 +2026-04-10 11:40:31 - INFO - parrotllm.training - step 78430 | epoch 0 | loss 3.4433 | lr 5.41e-04 | grad 0.2709 +2026-04-10 11:40:34 - INFO - parrotllm.training - step 78440 | epoch 0 | loss 3.3792 | lr 5.41e-04 | grad 0.2894 +2026-04-10 11:40:37 - INFO - parrotllm.training - step 78450 | epoch 0 | loss 3.3586 | lr 5.41e-04 | grad 0.2936 +2026-04-10 11:40:40 - INFO - parrotllm.training - step 78460 | epoch 0 | loss 3.2910 | lr 5.41e-04 | grad 0.2821 +2026-04-10 11:40:43 - INFO - parrotllm.training - step 78470 | epoch 0 | loss 3.2958 | lr 5.41e-04 | grad 0.2842 +2026-04-10 11:40:46 - INFO - parrotllm.training - step 78480 | epoch 0 | loss 3.3548 | lr 5.41e-04 | grad 0.2824 +2026-04-10 11:40:49 - INFO - parrotllm.training - step 78490 | epoch 0 | loss 3.3103 | lr 5.40e-04 | grad 0.3261 +2026-04-10 11:40:52 - INFO - parrotllm.training - step 78500 | epoch 0 | loss 3.2973 | lr 5.40e-04 | grad 0.2939 +2026-04-10 11:40:52 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:40:52 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:40:55 - INFO - parrotllm.training - Train: loss=3.2973, ppl=27.04 +2026-04-10 11:40:55 - INFO - parrotllm.training - Val: loss=3.3029, ppl=27.19 +2026-04-10 11:40:56 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3029_epoch_0000_step_0078500.pt +2026-04-10 11:40:58 - INFO - parrotllm.training - No validation improvement for 3/15 evaluation(s) (best=3.3032, min_delta=0.001000). +2026-04-10 11:40:58 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:41:01 - INFO - parrotllm.training - step 78510 | epoch 0 | loss 3.3371 | lr 5.40e-04 | grad 0.2842 +2026-04-10 11:41:05 - INFO - parrotllm.training - step 78520 | epoch 0 | loss 3.5026 | lr 5.40e-04 | grad 0.3156 +2026-04-10 11:41:08 - INFO - parrotllm.training - step 78530 | epoch 0 | loss 3.3386 | lr 5.40e-04 | grad 0.2602 +2026-04-10 11:41:11 - INFO - parrotllm.training - step 78540 | epoch 0 | loss 3.3709 | lr 5.40e-04 | grad 0.2688 +2026-04-10 11:41:14 - INFO - parrotllm.training - step 78550 | epoch 0 | loss 3.3841 | lr 5.39e-04 | grad 0.2648 +2026-04-10 11:41:17 - INFO - parrotllm.training - step 78560 | epoch 0 | loss 3.3089 | lr 5.39e-04 | grad 0.2799 +2026-04-10 11:41:20 - INFO - parrotllm.training - step 78570 | epoch 0 | loss 3.3176 | lr 5.39e-04 | grad 0.2734 +2026-04-10 11:41:23 - INFO - parrotllm.training - step 78580 | epoch 0 | loss 3.4115 | lr 5.39e-04 | grad 0.2524 +2026-04-10 11:41:26 - INFO - parrotllm.training - step 78590 | epoch 0 | loss 3.3329 | lr 5.39e-04 | grad 0.2908 +2026-04-10 11:41:29 - INFO - parrotllm.training - step 78600 | epoch 0 | loss 3.2330 | lr 5.39e-04 | grad 0.2700 +2026-04-10 11:41:32 - INFO - parrotllm.training - step 78610 | epoch 0 | loss 3.3194 | lr 5.38e-04 | grad 0.2910 +2026-04-10 11:41:35 - INFO - parrotllm.training - step 78620 | epoch 0 | loss 3.3836 | lr 5.38e-04 | grad 0.2814 +2026-04-10 11:41:38 - INFO - parrotllm.training - step 78630 | epoch 0 | loss 3.2684 | lr 5.38e-04 | grad 0.3102 +2026-04-10 11:41:41 - INFO - parrotllm.training - step 78640 | epoch 0 | loss 3.4185 | lr 5.38e-04 | grad 0.2687 +2026-04-10 11:41:44 - INFO - parrotllm.training - step 78650 | epoch 0 | loss 3.3735 | lr 5.38e-04 | grad 0.2608 +2026-04-10 11:41:47 - INFO - parrotllm.training - step 78660 | epoch 0 | loss 3.3998 | lr 5.38e-04 | grad 0.2800 +2026-04-10 11:41:50 - INFO - parrotllm.training - step 78670 | epoch 0 | loss 3.2801 | lr 5.38e-04 | grad 0.2684 +2026-04-10 11:41:53 - INFO - parrotllm.training - step 78680 | epoch 0 | loss 3.3570 | lr 5.37e-04 | grad 0.2603 +2026-04-10 11:41:56 - INFO - parrotllm.training - step 78690 | epoch 0 | loss 3.4267 | lr 5.37e-04 | grad 0.3200 +2026-04-10 11:42:00 - INFO - parrotllm.training - step 78700 | epoch 0 | loss 3.2911 | lr 5.37e-04 | grad 0.3019 +2026-04-10 11:42:03 - INFO - parrotllm.training - step 78710 | epoch 0 | loss 3.3049 | lr 5.37e-04 | grad 0.2770 +2026-04-10 11:42:06 - INFO - parrotllm.training - step 78720 | epoch 0 | loss 3.3660 | lr 5.37e-04 | grad 0.3007 +2026-04-10 11:42:09 - INFO - parrotllm.training - step 78730 | epoch 0 | loss 3.3182 | lr 5.37e-04 | grad 0.2710 +2026-04-10 11:42:12 - INFO - parrotllm.training - step 78740 | epoch 0 | loss 3.3622 | lr 5.36e-04 | grad 0.3055 +2026-04-10 11:42:15 - INFO - parrotllm.training - step 78750 | epoch 0 | loss 3.4862 | lr 5.36e-04 | grad 0.2799 +2026-04-10 11:42:18 - INFO - parrotllm.training - step 78760 | epoch 0 | loss 3.3868 | lr 5.36e-04 | grad 0.2606 +2026-04-10 11:42:21 - INFO - parrotllm.training - step 78770 | epoch 0 | loss 3.4129 | lr 5.36e-04 | grad 0.2630 +2026-04-10 11:42:24 - INFO - parrotllm.training - step 78780 | epoch 0 | loss 3.3728 | lr 5.36e-04 | grad 0.2802 +2026-04-10 11:42:27 - INFO - parrotllm.training - step 78790 | epoch 0 | loss 3.2671 | lr 5.36e-04 | grad 0.2926 +2026-04-10 11:42:30 - INFO - parrotllm.training - step 78800 | epoch 0 | loss 3.2908 | lr 5.35e-04 | grad 0.2548 +2026-04-10 11:42:33 - INFO - parrotllm.training - step 78810 | epoch 0 | loss 3.3536 | lr 5.35e-04 | grad 0.2891 +2026-04-10 11:42:36 - INFO - parrotllm.training - step 78820 | epoch 0 | loss 3.4378 | lr 5.35e-04 | grad 0.3235 +2026-04-10 11:42:39 - INFO - parrotllm.training - step 78830 | epoch 0 | loss 3.3646 | lr 5.35e-04 | grad 0.2876 +2026-04-10 11:42:42 - INFO - parrotllm.training - step 78840 | epoch 0 | loss 3.4629 | lr 5.35e-04 | grad 0.2672 +2026-04-10 11:42:45 - INFO - parrotllm.training - step 78850 | epoch 0 | loss 3.3246 | lr 5.35e-04 | grad 0.2764 +2026-04-10 11:42:48 - INFO - parrotllm.training - step 78860 | epoch 0 | loss 3.3325 | lr 5.34e-04 | grad 0.3234 +2026-04-10 11:42:51 - INFO - parrotllm.training - step 78870 | epoch 0 | loss 3.3411 | lr 5.34e-04 | grad 0.2928 +2026-04-10 11:42:54 - INFO - parrotllm.training - step 78880 | epoch 0 | loss 3.2821 | lr 5.34e-04 | grad 0.2782 +2026-04-10 11:42:57 - INFO - parrotllm.training - step 78890 | epoch 0 | loss 3.4088 | lr 5.34e-04 | grad 0.2838 +2026-04-10 11:43:01 - INFO - parrotllm.training - step 78900 | epoch 0 | loss 3.3069 | lr 5.34e-04 | grad 0.2730 +2026-04-10 11:43:04 - INFO - parrotllm.training - step 78910 | epoch 0 | loss 3.3769 | lr 5.34e-04 | grad 0.3012 +2026-04-10 11:43:07 - INFO - parrotllm.training - step 78920 | epoch 0 | loss 3.3397 | lr 5.34e-04 | grad 0.2774 +2026-04-10 11:43:10 - INFO - parrotllm.training - step 78930 | epoch 0 | loss 3.3188 | lr 5.33e-04 | grad 0.2767 +2026-04-10 11:43:13 - INFO - parrotllm.training - step 78940 | epoch 0 | loss 3.3705 | lr 5.33e-04 | grad 0.2929 +2026-04-10 11:43:16 - INFO - parrotllm.training - step 78950 | epoch 0 | loss 3.4152 | lr 5.33e-04 | grad 0.2573 +2026-04-10 11:43:19 - INFO - parrotllm.training - step 78960 | epoch 0 | loss 3.3677 | lr 5.33e-04 | grad 0.2743 +2026-04-10 11:43:22 - INFO - parrotllm.training - step 78970 | epoch 0 | loss 3.3207 | lr 5.33e-04 | grad 0.2606 +2026-04-10 11:43:25 - INFO - parrotllm.training - step 78980 | epoch 0 | loss 3.4162 | lr 5.33e-04 | grad 0.2698 +2026-04-10 11:43:28 - INFO - parrotllm.training - step 78990 | epoch 0 | loss 3.3325 | lr 5.32e-04 | grad 0.2789 +2026-04-10 11:43:31 - INFO - parrotllm.training - step 79000 | epoch 0 | loss 3.3215 | lr 5.32e-04 | grad 0.2888 +2026-04-10 11:43:31 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:43:31 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:43:34 - INFO - parrotllm.training - Train: loss=3.3215, ppl=27.70 +2026-04-10 11:43:34 - INFO - parrotllm.training - Val: loss=3.3028, ppl=27.19 +2026-04-10 11:43:35 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3028_epoch_0000_step_0079000.pt +2026-04-10 11:43:37 - INFO - parrotllm.training - No validation improvement for 4/15 evaluation(s) (best=3.3032, min_delta=0.001000). +2026-04-10 11:43:37 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:43:40 - INFO - parrotllm.training - step 79010 | epoch 0 | loss 3.4537 | lr 5.32e-04 | grad 0.2770 +2026-04-10 11:43:43 - INFO - parrotllm.training - step 79020 | epoch 0 | loss 3.3342 | lr 5.32e-04 | grad 0.2790 +2026-04-10 11:43:46 - INFO - parrotllm.training - step 79030 | epoch 0 | loss 3.3539 | lr 5.32e-04 | grad 0.3064 +2026-04-10 11:43:49 - INFO - parrotllm.training - step 79040 | epoch 0 | loss 3.2802 | lr 5.32e-04 | grad 0.2928 +2026-04-10 11:43:52 - INFO - parrotllm.training - step 79050 | epoch 0 | loss 3.3277 | lr 5.31e-04 | grad 0.2670 +2026-04-10 11:43:55 - INFO - parrotllm.training - step 79060 | epoch 0 | loss 3.4208 | lr 5.31e-04 | grad 0.2680 +2026-04-10 11:43:58 - INFO - parrotllm.training - step 79070 | epoch 0 | loss 3.3319 | lr 5.31e-04 | grad 0.3091 +2026-04-10 11:44:01 - INFO - parrotllm.training - step 79080 | epoch 0 | loss 3.3286 | lr 5.31e-04 | grad 0.2698 +2026-04-10 11:44:04 - INFO - parrotllm.training - step 79090 | epoch 0 | loss 3.3533 | lr 5.31e-04 | grad 0.2824 +2026-04-10 11:44:08 - INFO - parrotllm.training - step 79100 | epoch 0 | loss 3.4939 | lr 5.31e-04 | grad 0.2864 +2026-04-10 11:44:11 - INFO - parrotllm.training - step 79110 | epoch 0 | loss 3.3199 | lr 5.31e-04 | grad 0.2940 +2026-04-10 11:44:14 - INFO - parrotllm.training - step 79120 | epoch 0 | loss 3.3538 | lr 5.30e-04 | grad 0.2611 +2026-04-10 11:44:17 - INFO - parrotllm.training - step 79130 | epoch 0 | loss 3.3573 | lr 5.30e-04 | grad 0.2767 +2026-04-10 11:44:20 - INFO - parrotllm.training - step 79140 | epoch 0 | loss 3.3055 | lr 5.30e-04 | grad 0.2858 +2026-04-10 11:44:23 - INFO - parrotllm.training - step 79150 | epoch 0 | loss 3.2918 | lr 5.30e-04 | grad 0.2731 +2026-04-10 11:44:26 - INFO - parrotllm.training - step 79160 | epoch 0 | loss 3.3786 | lr 5.30e-04 | grad 0.2964 +2026-04-10 11:44:29 - INFO - parrotllm.training - step 79170 | epoch 0 | loss 3.3850 | lr 5.30e-04 | grad 0.2769 +2026-04-10 11:44:32 - INFO - parrotllm.training - step 79180 | epoch 0 | loss 3.3493 | lr 5.29e-04 | grad 0.2923 +2026-04-10 11:44:35 - INFO - parrotllm.training - step 79190 | epoch 0 | loss 3.4408 | lr 5.29e-04 | grad 0.2965 +2026-04-10 11:44:38 - INFO - parrotllm.training - step 79200 | epoch 0 | loss 3.4840 | lr 5.29e-04 | grad 0.2897 +2026-04-10 11:44:41 - INFO - parrotllm.training - step 79210 | epoch 0 | loss 3.4378 | lr 5.29e-04 | grad 0.2835 +2026-04-10 11:44:44 - INFO - parrotllm.training - step 79220 | epoch 0 | loss 3.3525 | lr 5.29e-04 | grad 0.2731 +2026-04-10 11:44:47 - INFO - parrotllm.training - step 79230 | epoch 0 | loss 3.2717 | lr 5.29e-04 | grad 0.2804 +2026-04-10 11:44:50 - INFO - parrotllm.training - step 79240 | epoch 0 | loss 3.4894 | lr 5.28e-04 | grad 0.2851 +2026-04-10 11:44:53 - INFO - parrotllm.training - step 79250 | epoch 0 | loss 3.3555 | lr 5.28e-04 | grad 0.2887 +2026-04-10 11:44:56 - INFO - parrotllm.training - step 79260 | epoch 0 | loss 3.4680 | lr 5.28e-04 | grad 0.2981 +2026-04-10 11:44:59 - INFO - parrotllm.training - step 79270 | epoch 0 | loss 3.3639 | lr 5.28e-04 | grad 0.2666 +2026-04-10 11:45:02 - INFO - parrotllm.training - step 79280 | epoch 0 | loss 3.3285 | lr 5.28e-04 | grad 0.2997 +2026-04-10 11:45:06 - INFO - parrotllm.training - step 79290 | epoch 0 | loss 3.3555 | lr 5.28e-04 | grad 0.2850 +2026-04-10 11:45:09 - INFO - parrotllm.training - step 79300 | epoch 0 | loss 3.4240 | lr 5.28e-04 | grad 0.3008 +2026-04-10 11:45:12 - INFO - parrotllm.training - step 79310 | epoch 0 | loss 3.3005 | lr 5.27e-04 | grad 0.2971 +2026-04-10 11:45:15 - INFO - parrotllm.training - step 79320 | epoch 0 | loss 3.3538 | lr 5.27e-04 | grad 0.3022 +2026-04-10 11:45:18 - INFO - parrotllm.training - step 79330 | epoch 0 | loss 3.3321 | lr 5.27e-04 | grad 0.2678 +2026-04-10 11:45:21 - INFO - parrotllm.training - step 79340 | epoch 0 | loss 3.3851 | lr 5.27e-04 | grad 0.2810 +2026-04-10 11:45:24 - INFO - parrotllm.training - step 79350 | epoch 0 | loss 3.4214 | lr 5.27e-04 | grad 0.2672 +2026-04-10 11:45:27 - INFO - parrotllm.training - step 79360 | epoch 0 | loss 3.3288 | lr 5.27e-04 | grad 0.2457 +2026-04-10 11:45:30 - INFO - parrotllm.training - step 79370 | epoch 0 | loss 3.3744 | lr 5.26e-04 | grad 0.2931 +2026-04-10 11:45:33 - INFO - parrotllm.training - step 79380 | epoch 0 | loss 3.3745 | lr 5.26e-04 | grad 0.2773 +2026-04-10 11:45:36 - INFO - parrotllm.training - step 79390 | epoch 0 | loss 3.3661 | lr 5.26e-04 | grad 0.3087 +2026-04-10 11:45:39 - INFO - parrotllm.training - step 79400 | epoch 0 | loss 3.3067 | lr 5.26e-04 | grad 0.2938 +2026-04-10 11:45:42 - INFO - parrotllm.training - step 79410 | epoch 0 | loss 3.3812 | lr 5.26e-04 | grad 0.2890 +2026-04-10 11:45:45 - INFO - parrotllm.training - step 79420 | epoch 0 | loss 3.3597 | lr 5.26e-04 | grad 0.2613 +2026-04-10 11:45:48 - INFO - parrotllm.training - step 79430 | epoch 0 | loss 3.2962 | lr 5.25e-04 | grad 0.2935 +2026-04-10 11:45:51 - INFO - parrotllm.training - step 79440 | epoch 0 | loss 3.2963 | lr 5.25e-04 | grad 0.3162 +2026-04-10 11:45:54 - INFO - parrotllm.training - step 79450 | epoch 0 | loss 3.4201 | lr 5.25e-04 | grad 0.3034 +2026-04-10 11:45:57 - INFO - parrotllm.training - step 79460 | epoch 0 | loss 3.4172 | lr 5.25e-04 | grad 0.2756 +2026-04-10 11:46:00 - INFO - parrotllm.training - step 79470 | epoch 0 | loss 3.4018 | lr 5.25e-04 | grad 0.2725 +2026-04-10 11:46:04 - INFO - parrotllm.training - step 79480 | epoch 0 | loss 3.3126 | lr 5.25e-04 | grad 0.2638 +2026-04-10 11:46:07 - INFO - parrotllm.training - step 79490 | epoch 0 | loss 3.4252 | lr 5.25e-04 | grad 0.2757 +2026-04-10 11:46:10 - INFO - parrotllm.training - step 79500 | epoch 0 | loss 3.4003 | lr 5.24e-04 | grad 0.2626 +2026-04-10 11:46:10 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:46:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:46:13 - INFO - parrotllm.training - Train: loss=3.4003, ppl=29.97 +2026-04-10 11:46:13 - INFO - parrotllm.training - Val: loss=3.3018, ppl=27.16 +2026-04-10 11:46:13 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:46:14 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p3018_epoch_0000_step_0079500.pt +2026-04-10 11:46:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:46:19 - INFO - parrotllm.training - step 79510 | epoch 0 | loss 3.3443 | lr 5.24e-04 | grad 0.2824 +2026-04-10 11:46:22 - INFO - parrotllm.training - step 79520 | epoch 0 | loss 3.2984 | lr 5.24e-04 | grad 0.2809 +2026-04-10 11:46:25 - INFO - parrotllm.training - step 79530 | epoch 0 | loss 3.4564 | lr 5.24e-04 | grad 0.2761 +2026-04-10 11:46:28 - INFO - parrotllm.training - step 79540 | epoch 0 | loss 3.2951 | lr 5.24e-04 | grad 0.2534 +2026-04-10 11:46:31 - INFO - parrotllm.training - step 79550 | epoch 0 | loss 3.2958 | lr 5.24e-04 | grad 0.2667 +2026-04-10 11:46:34 - INFO - parrotllm.training - step 79560 | epoch 0 | loss 3.4292 | lr 5.23e-04 | grad 0.3120 +2026-04-10 11:46:37 - INFO - parrotllm.training - step 79570 | epoch 0 | loss 3.3728 | lr 5.23e-04 | grad 0.2982 +2026-04-10 11:46:40 - INFO - parrotllm.training - step 79580 | epoch 0 | loss 3.4367 | lr 5.23e-04 | grad 0.2703 +2026-04-10 11:46:43 - INFO - parrotllm.training - step 79590 | epoch 0 | loss 3.2882 | lr 5.23e-04 | grad 0.2906 +2026-04-10 11:46:46 - INFO - parrotllm.training - step 79600 | epoch 0 | loss 3.4308 | lr 5.23e-04 | grad 0.2969 +2026-04-10 11:46:49 - INFO - parrotllm.training - step 79610 | epoch 0 | loss 3.3238 | lr 5.23e-04 | grad 0.3113 +2026-04-10 11:46:52 - INFO - parrotllm.training - step 79620 | epoch 0 | loss 3.3974 | lr 5.22e-04 | grad 0.2838 +2026-04-10 11:46:55 - INFO - parrotllm.training - step 79630 | epoch 0 | loss 3.3125 | lr 5.22e-04 | grad 0.2766 +2026-04-10 11:46:58 - INFO - parrotllm.training - step 79640 | epoch 0 | loss 3.3250 | lr 5.22e-04 | grad 0.2529 +2026-04-10 11:47:01 - INFO - parrotllm.training - step 79650 | epoch 0 | loss 3.2701 | lr 5.22e-04 | grad 0.2776 +2026-04-10 11:47:04 - INFO - parrotllm.training - step 79660 | epoch 0 | loss 3.2310 | lr 5.22e-04 | grad 0.3052 +2026-04-10 11:47:07 - INFO - parrotllm.training - step 79670 | epoch 0 | loss 3.3704 | lr 5.22e-04 | grad 0.2651 +2026-04-10 11:47:10 - INFO - parrotllm.training - step 79680 | epoch 0 | loss 3.4411 | lr 5.22e-04 | grad 0.2692 +2026-04-10 11:47:14 - INFO - parrotllm.training - step 79690 | epoch 0 | loss 3.3185 | lr 5.21e-04 | grad 0.2902 +2026-04-10 11:47:17 - INFO - parrotllm.training - step 79700 | epoch 0 | loss 3.3903 | lr 5.21e-04 | grad 0.3136 +2026-04-10 11:47:17 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 79734/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 11:47:20 - INFO - parrotllm.training - step 79710 | epoch 0 | loss 3.3694 | lr 5.21e-04 | grad 0.2705 +2026-04-10 11:47:23 - INFO - parrotllm.training - step 79720 | epoch 0 | loss 3.2946 | lr 5.21e-04 | grad 0.3013 +2026-04-10 11:47:26 - INFO - parrotllm.training - step 79730 | epoch 0 | loss 3.4777 | lr 5.21e-04 | grad 0.2675 +2026-04-10 11:47:29 - INFO - parrotllm.training - step 79740 | epoch 0 | loss 3.4457 | lr 5.21e-04 | grad 0.2918 +2026-04-10 11:47:32 - INFO - parrotllm.training - step 79750 | epoch 0 | loss 3.4171 | lr 5.20e-04 | grad 0.2569 +2026-04-10 11:47:35 - INFO - parrotllm.training - step 79760 | epoch 0 | loss 3.4403 | lr 5.20e-04 | grad 0.2766 +2026-04-10 11:47:38 - INFO - parrotllm.training - step 79770 | epoch 0 | loss 3.3303 | lr 5.20e-04 | grad 0.2916 +2026-04-10 11:47:41 - INFO - parrotllm.training - step 79780 | epoch 0 | loss 3.4239 | lr 5.20e-04 | grad 0.2945 +2026-04-10 11:47:44 - INFO - parrotllm.training - step 79790 | epoch 0 | loss 3.3421 | lr 5.20e-04 | grad 0.2800 +2026-04-10 11:47:47 - INFO - parrotllm.training - step 79800 | epoch 0 | loss 3.2242 | lr 5.20e-04 | grad 0.2934 +2026-04-10 11:47:50 - INFO - parrotllm.training - step 79810 | epoch 0 | loss 3.4722 | lr 5.20e-04 | grad 0.3031 +2026-04-10 11:47:53 - INFO - parrotllm.training - step 79820 | epoch 0 | loss 3.2718 | lr 5.19e-04 | grad 0.2917 +2026-04-10 11:47:57 - INFO - parrotllm.training - step 79830 | epoch 0 | loss 3.3671 | lr 5.19e-04 | grad 0.2905 +2026-04-10 11:48:00 - INFO - parrotllm.training - step 79840 | epoch 0 | loss 3.2742 | lr 5.19e-04 | grad 0.2630 +2026-04-10 11:48:03 - INFO - parrotllm.training - step 79850 | epoch 0 | loss 3.3569 | lr 5.19e-04 | grad 0.3319 +2026-04-10 11:48:06 - INFO - parrotllm.training - step 79860 | epoch 0 | loss 3.3345 | lr 5.19e-04 | grad 0.2939 +2026-04-10 11:48:09 - INFO - parrotllm.training - step 79870 | epoch 0 | loss 3.3491 | lr 5.19e-04 | grad 0.2676 +2026-04-10 11:48:12 - INFO - parrotllm.training - step 79880 | epoch 0 | loss 3.3625 | lr 5.18e-04 | grad 0.3028 +2026-04-10 11:48:15 - INFO - parrotllm.training - step 79890 | epoch 0 | loss 3.3840 | lr 5.18e-04 | grad 0.2979 +2026-04-10 11:48:18 - INFO - parrotllm.training - step 79900 | epoch 0 | loss 3.4300 | lr 5.18e-04 | grad 0.2797 +2026-04-10 11:48:21 - INFO - parrotllm.training - step 79910 | epoch 0 | loss 3.3646 | lr 5.18e-04 | grad 0.2738 +2026-04-10 11:48:24 - INFO - parrotllm.training - step 79920 | epoch 0 | loss 3.3555 | lr 5.18e-04 | grad 0.2700 +2026-04-10 11:48:27 - INFO - parrotllm.training - step 79930 | epoch 0 | loss 3.3708 | lr 5.18e-04 | grad 0.2693 +2026-04-10 11:48:30 - INFO - parrotllm.training - step 79940 | epoch 0 | loss 3.3355 | lr 5.17e-04 | grad 0.2811 +2026-04-10 11:48:33 - INFO - parrotllm.training - step 79950 | epoch 0 | loss 3.3697 | lr 5.17e-04 | grad 0.2739 +2026-04-10 11:48:36 - INFO - parrotllm.training - step 79960 | epoch 0 | loss 3.3720 | lr 5.17e-04 | grad 0.2547 +2026-04-10 11:48:39 - INFO - parrotllm.training - step 79970 | epoch 0 | loss 3.3467 | lr 5.17e-04 | grad 0.3209 +2026-04-10 11:48:42 - INFO - parrotllm.training - step 79980 | epoch 0 | loss 3.3860 | lr 5.17e-04 | grad 0.3323 +2026-04-10 11:48:45 - INFO - parrotllm.training - step 79990 | epoch 0 | loss 3.4209 | lr 5.17e-04 | grad 0.2804 +2026-04-10 11:48:48 - INFO - parrotllm.training - step 80000 | epoch 0 | loss 3.2870 | lr 5.17e-04 | grad 0.2594 +2026-04-10 11:48:48 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:48:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:48:51 - INFO - parrotllm.training - Train: loss=3.2870, ppl=26.76 +2026-04-10 11:48:51 - INFO - parrotllm.training - Val: loss=3.2976, ppl=27.05 +2026-04-10 11:48:51 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:48:52 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2976_epoch_0000_step_0080000.pt +2026-04-10 11:48:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:48:56 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0080000.pt +2026-04-10 11:49:02 - INFO - parrotllm.training - step 80010 | epoch 0 | loss 3.3889 | lr 5.16e-04 | grad 0.3195 +2026-04-10 11:49:05 - INFO - parrotllm.training - step 80020 | epoch 0 | loss 3.3270 | lr 5.16e-04 | grad 0.2637 +2026-04-10 11:49:08 - INFO - parrotllm.training - step 80030 | epoch 0 | loss 3.3479 | lr 5.16e-04 | grad 0.2696 +2026-04-10 11:49:11 - INFO - parrotllm.training - step 80040 | epoch 0 | loss 3.3452 | lr 5.16e-04 | grad 0.2858 +2026-04-10 11:49:14 - INFO - parrotllm.training - step 80050 | epoch 0 | loss 3.2724 | lr 5.16e-04 | grad 0.2906 +2026-04-10 11:49:17 - INFO - parrotllm.training - step 80060 | epoch 0 | loss 3.3116 | lr 5.16e-04 | grad 0.2652 +2026-04-10 11:49:20 - INFO - parrotllm.training - step 80070 | epoch 0 | loss 3.2724 | lr 5.15e-04 | grad 0.3204 +2026-04-10 11:49:23 - INFO - parrotllm.training - step 80080 | epoch 0 | loss 3.3650 | lr 5.15e-04 | grad 0.2835 +2026-04-10 11:49:26 - INFO - parrotllm.training - step 80090 | epoch 0 | loss 3.3206 | lr 5.15e-04 | grad 0.2716 +2026-04-10 11:49:29 - INFO - parrotllm.training - step 80100 | epoch 0 | loss 3.3809 | lr 5.15e-04 | grad 0.2855 +2026-04-10 11:49:32 - INFO - parrotllm.training - step 80110 | epoch 0 | loss 3.3438 | lr 5.15e-04 | grad 0.2979 +2026-04-10 11:49:35 - INFO - parrotllm.training - step 80120 | epoch 0 | loss 3.3205 | lr 5.15e-04 | grad 0.3154 +2026-04-10 11:49:38 - INFO - parrotllm.training - step 80130 | epoch 0 | loss 3.5151 | lr 5.14e-04 | grad 0.2979 +2026-04-10 11:49:41 - INFO - parrotllm.training - step 80140 | epoch 0 | loss 3.2591 | lr 5.14e-04 | grad 0.2783 +2026-04-10 11:49:44 - INFO - parrotllm.training - step 80150 | epoch 0 | loss 3.3247 | lr 5.14e-04 | grad 0.2804 +2026-04-10 11:49:47 - INFO - parrotllm.training - step 80160 | epoch 0 | loss 3.3506 | lr 5.14e-04 | grad 0.2953 +2026-04-10 11:49:50 - INFO - parrotllm.training - step 80170 | epoch 0 | loss 3.3741 | lr 5.14e-04 | grad 0.2922 +2026-04-10 11:49:53 - INFO - parrotllm.training - step 80180 | epoch 0 | loss 3.2112 | lr 5.14e-04 | grad 0.2925 +2026-04-10 11:49:57 - INFO - parrotllm.training - step 80190 | epoch 0 | loss 3.3664 | lr 5.14e-04 | grad 0.2709 +2026-04-10 11:50:00 - INFO - parrotllm.training - step 80200 | epoch 0 | loss 3.3078 | lr 5.13e-04 | grad 0.2666 +2026-04-10 11:50:03 - INFO - parrotllm.training - step 80210 | epoch 0 | loss 3.4044 | lr 5.13e-04 | grad 0.2737 +2026-04-10 11:50:06 - INFO - parrotllm.training - step 80220 | epoch 0 | loss 3.3483 | lr 5.13e-04 | grad 0.2767 +2026-04-10 11:50:09 - INFO - parrotllm.training - step 80230 | epoch 0 | loss 3.3119 | lr 5.13e-04 | grad 0.2646 +2026-04-10 11:50:12 - INFO - parrotllm.training - step 80240 | epoch 0 | loss 3.4051 | lr 5.13e-04 | grad 0.2793 +2026-04-10 11:50:15 - INFO - parrotllm.training - step 80250 | epoch 0 | loss 3.3190 | lr 5.13e-04 | grad 0.3063 +2026-04-10 11:50:18 - INFO - parrotllm.training - step 80260 | epoch 0 | loss 3.3006 | lr 5.12e-04 | grad 0.2762 +2026-04-10 11:50:21 - INFO - parrotllm.training - step 80270 | epoch 0 | loss 3.3468 | lr 5.12e-04 | grad 0.2916 +2026-04-10 11:50:24 - INFO - parrotllm.training - step 80280 | epoch 0 | loss 3.3494 | lr 5.12e-04 | grad 0.3075 +2026-04-10 11:50:27 - INFO - parrotllm.training - step 80290 | epoch 0 | loss 3.3269 | lr 5.12e-04 | grad 0.2848 +2026-04-10 11:50:30 - INFO - parrotllm.training - step 80300 | epoch 0 | loss 3.3751 | lr 5.12e-04 | grad 0.2826 +2026-04-10 11:50:33 - INFO - parrotllm.training - step 80310 | epoch 0 | loss 3.2513 | lr 5.12e-04 | grad 0.2602 +2026-04-10 11:50:36 - INFO - parrotllm.training - step 80320 | epoch 0 | loss 3.1685 | lr 5.12e-04 | grad 0.2942 +2026-04-10 11:50:39 - INFO - parrotllm.training - step 80330 | epoch 0 | loss 3.3915 | lr 5.11e-04 | grad 0.3411 +2026-04-10 11:50:42 - INFO - parrotllm.training - step 80340 | epoch 0 | loss 3.4068 | lr 5.11e-04 | grad 0.2633 +2026-04-10 11:50:45 - INFO - parrotllm.training - step 80350 | epoch 0 | loss 3.3632 | lr 5.11e-04 | grad 0.2827 +2026-04-10 11:50:48 - INFO - parrotllm.training - step 80360 | epoch 0 | loss 3.2693 | lr 5.11e-04 | grad 0.2741 +2026-04-10 11:50:51 - INFO - parrotllm.training - step 80370 | epoch 0 | loss 3.3803 | lr 5.11e-04 | grad 0.3092 +2026-04-10 11:50:54 - INFO - parrotllm.training - step 80380 | epoch 0 | loss 3.2941 | lr 5.11e-04 | grad 0.2932 +2026-04-10 11:50:58 - INFO - parrotllm.training - step 80390 | epoch 0 | loss 3.4000 | lr 5.10e-04 | grad 0.2864 +2026-04-10 11:51:01 - INFO - parrotllm.training - step 80400 | epoch 0 | loss 3.3443 | lr 5.10e-04 | grad 0.2719 +2026-04-10 11:51:04 - INFO - parrotllm.training - step 80410 | epoch 0 | loss 3.3538 | lr 5.10e-04 | grad 0.2976 +2026-04-10 11:51:07 - INFO - parrotllm.training - step 80420 | epoch 0 | loss 3.3221 | lr 5.10e-04 | grad 0.3065 +2026-04-10 11:51:10 - INFO - parrotllm.training - step 80430 | epoch 0 | loss 3.3957 | lr 5.10e-04 | grad 0.2770 +2026-04-10 11:51:13 - INFO - parrotllm.training - step 80440 | epoch 0 | loss 3.3997 | lr 5.10e-04 | grad 0.2736 +2026-04-10 11:51:16 - INFO - parrotllm.training - step 80450 | epoch 0 | loss 3.2366 | lr 5.10e-04 | grad 0.2804 +2026-04-10 11:51:19 - INFO - parrotllm.training - step 80460 | epoch 0 | loss 3.3706 | lr 5.09e-04 | grad 0.2853 +2026-04-10 11:51:22 - INFO - parrotllm.training - step 80470 | epoch 0 | loss 3.3575 | lr 5.09e-04 | grad 0.3013 +2026-04-10 11:51:25 - INFO - parrotllm.training - step 80480 | epoch 0 | loss 3.4162 | lr 5.09e-04 | grad 0.2844 +2026-04-10 11:51:28 - INFO - parrotllm.training - step 80490 | epoch 0 | loss 3.2992 | lr 5.09e-04 | grad 0.2919 +2026-04-10 11:51:31 - INFO - parrotllm.training - step 80500 | epoch 0 | loss 3.3842 | lr 5.09e-04 | grad 0.2803 +2026-04-10 11:51:31 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:51:31 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:51:34 - INFO - parrotllm.training - Train: loss=3.3842, ppl=29.50 +2026-04-10 11:51:34 - INFO - parrotllm.training - Val: loss=3.2986, ppl=27.07 +2026-04-10 11:51:35 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2986_epoch_0000_step_0080500.pt +2026-04-10 11:51:37 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2976, min_delta=0.001000). +2026-04-10 11:51:37 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:51:40 - INFO - parrotllm.training - step 80510 | epoch 0 | loss 3.3723 | lr 5.09e-04 | grad 0.2617 +2026-04-10 11:51:43 - INFO - parrotllm.training - step 80520 | epoch 0 | loss 3.3958 | lr 5.08e-04 | grad 0.2673 +2026-04-10 11:51:46 - INFO - parrotllm.training - step 80530 | epoch 0 | loss 3.4399 | lr 5.08e-04 | grad 0.2781 +2026-04-10 11:51:49 - INFO - parrotllm.training - step 80540 | epoch 0 | loss 3.4196 | lr 5.08e-04 | grad 0.2821 +2026-04-10 11:51:52 - INFO - parrotllm.training - step 80550 | epoch 0 | loss 3.3419 | lr 5.08e-04 | grad 0.2942 +2026-04-10 11:51:55 - INFO - parrotllm.training - step 80560 | epoch 0 | loss 3.3524 | lr 5.08e-04 | grad 0.3026 +2026-04-10 11:51:58 - INFO - parrotllm.training - step 80570 | epoch 0 | loss 3.3450 | lr 5.08e-04 | grad 0.2811 +2026-04-10 11:52:01 - INFO - parrotllm.training - step 80580 | epoch 0 | loss 3.3586 | lr 5.07e-04 | grad 0.2887 +2026-04-10 11:52:04 - INFO - parrotllm.training - step 80590 | epoch 0 | loss 3.4153 | lr 5.07e-04 | grad 0.2766 +2026-04-10 11:52:08 - INFO - parrotllm.training - step 80600 | epoch 0 | loss 3.3336 | lr 5.07e-04 | grad 0.2765 +2026-04-10 11:52:11 - INFO - parrotllm.training - step 80610 | epoch 0 | loss 3.3654 | lr 5.07e-04 | grad 0.2554 +2026-04-10 11:52:14 - INFO - parrotllm.training - step 80620 | epoch 0 | loss 3.3226 | lr 5.07e-04 | grad 0.2753 +2026-04-10 11:52:17 - INFO - parrotllm.training - step 80630 | epoch 0 | loss 3.2939 | lr 5.07e-04 | grad 0.2886 +2026-04-10 11:52:20 - INFO - parrotllm.training - step 80640 | epoch 0 | loss 3.4245 | lr 5.07e-04 | grad 0.2824 +2026-04-10 11:52:23 - INFO - parrotllm.training - step 80650 | epoch 0 | loss 3.4359 | lr 5.06e-04 | grad 0.2709 +2026-04-10 11:52:26 - INFO - parrotllm.training - step 80660 | epoch 0 | loss 3.2996 | lr 5.06e-04 | grad 0.2680 +2026-04-10 11:52:29 - INFO - parrotllm.training - step 80670 | epoch 0 | loss 3.2887 | lr 5.06e-04 | grad 0.3395 +2026-04-10 11:52:32 - INFO - parrotllm.training - step 80680 | epoch 0 | loss 3.3714 | lr 5.06e-04 | grad 0.2835 +2026-04-10 11:52:35 - INFO - parrotllm.training - step 80690 | epoch 0 | loss 3.4000 | lr 5.06e-04 | grad 0.2846 +2026-04-10 11:52:38 - INFO - parrotllm.training - step 80700 | epoch 0 | loss 3.2489 | lr 5.06e-04 | grad 0.2622 +2026-04-10 11:52:41 - INFO - parrotllm.training - step 80710 | epoch 0 | loss 3.2616 | lr 5.05e-04 | grad 0.2665 +2026-04-10 11:52:44 - INFO - parrotllm.training - step 80720 | epoch 0 | loss 3.3671 | lr 5.05e-04 | grad 0.2768 +2026-04-10 11:52:47 - INFO - parrotllm.training - step 80730 | epoch 0 | loss 3.3420 | lr 5.05e-04 | grad 0.2939 +2026-04-10 11:52:50 - INFO - parrotllm.training - step 80740 | epoch 0 | loss 3.3166 | lr 5.05e-04 | grad 0.2787 +2026-04-10 11:52:53 - INFO - parrotllm.training - step 80750 | epoch 0 | loss 3.3360 | lr 5.05e-04 | grad 0.2897 +2026-04-10 11:52:56 - INFO - parrotllm.training - step 80760 | epoch 0 | loss 3.4347 | lr 5.05e-04 | grad 0.3017 +2026-04-10 11:52:59 - INFO - parrotllm.training - step 80770 | epoch 0 | loss 3.3442 | lr 5.05e-04 | grad 0.2927 +2026-04-10 11:53:03 - INFO - parrotllm.training - step 80780 | epoch 0 | loss 3.3731 | lr 5.04e-04 | grad 0.2769 +2026-04-10 11:53:06 - INFO - parrotllm.training - step 80790 | epoch 0 | loss 3.3658 | lr 5.04e-04 | grad 0.2702 +2026-04-10 11:53:09 - INFO - parrotllm.training - step 80800 | epoch 0 | loss 3.3031 | lr 5.04e-04 | grad 0.2831 +2026-04-10 11:53:12 - INFO - parrotllm.training - step 80810 | epoch 0 | loss 3.3918 | lr 5.04e-04 | grad 0.3314 +2026-04-10 11:53:15 - INFO - parrotllm.training - step 80820 | epoch 0 | loss 3.3456 | lr 5.04e-04 | grad 0.2721 +2026-04-10 11:53:18 - INFO - parrotllm.training - step 80830 | epoch 0 | loss 3.3305 | lr 5.04e-04 | grad 0.2863 +2026-04-10 11:53:21 - INFO - parrotllm.training - step 80840 | epoch 0 | loss 3.3326 | lr 5.03e-04 | grad 0.3471 +2026-04-10 11:53:24 - INFO - parrotllm.training - step 80850 | epoch 0 | loss 3.3048 | lr 5.03e-04 | grad 0.2876 +2026-04-10 11:53:27 - INFO - parrotllm.training - step 80860 | epoch 0 | loss 3.3217 | lr 5.03e-04 | grad 0.2906 +2026-04-10 11:53:30 - INFO - parrotllm.training - step 80870 | epoch 0 | loss 3.2805 | lr 5.03e-04 | grad 0.2836 +2026-04-10 11:53:33 - INFO - parrotllm.training - step 80880 | epoch 0 | loss 3.3536 | lr 5.03e-04 | grad 0.2939 +2026-04-10 11:53:36 - INFO - parrotllm.training - step 80890 | epoch 0 | loss 3.4073 | lr 5.03e-04 | grad 0.3005 +2026-04-10 11:53:39 - INFO - parrotllm.training - step 80900 | epoch 0 | loss 3.4177 | lr 5.03e-04 | grad 0.2732 +2026-04-10 11:53:42 - INFO - parrotllm.training - step 80910 | epoch 0 | loss 3.2767 | lr 5.02e-04 | grad 0.2843 +2026-04-10 11:53:45 - INFO - parrotllm.training - step 80920 | epoch 0 | loss 3.2614 | lr 5.02e-04 | grad 0.3156 +2026-04-10 11:53:48 - INFO - parrotllm.training - step 80930 | epoch 0 | loss 3.3409 | lr 5.02e-04 | grad 0.2808 +2026-04-10 11:53:51 - INFO - parrotllm.training - step 80940 | epoch 0 | loss 3.3144 | lr 5.02e-04 | grad 0.3016 +2026-04-10 11:53:54 - INFO - parrotllm.training - step 80950 | epoch 0 | loss 3.4483 | lr 5.02e-04 | grad 0.2917 +2026-04-10 11:53:57 - INFO - parrotllm.training - step 80960 | epoch 0 | loss 3.2875 | lr 5.02e-04 | grad 0.2758 +2026-04-10 11:54:01 - INFO - parrotllm.training - step 80970 | epoch 0 | loss 3.3191 | lr 5.01e-04 | grad 0.2834 +2026-04-10 11:54:04 - INFO - parrotllm.training - step 80980 | epoch 0 | loss 3.4103 | lr 5.01e-04 | grad 0.2819 +2026-04-10 11:54:07 - INFO - parrotllm.training - step 80990 | epoch 0 | loss 3.3983 | lr 5.01e-04 | grad 0.2783 +2026-04-10 11:54:10 - INFO - parrotllm.training - step 81000 | epoch 0 | loss 3.4089 | lr 5.01e-04 | grad 0.2773 +2026-04-10 11:54:10 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:54:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:54:13 - INFO - parrotllm.training - Train: loss=3.4089, ppl=30.23 +2026-04-10 11:54:13 - INFO - parrotllm.training - Val: loss=3.2965, ppl=27.02 +2026-04-10 11:54:13 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:54:14 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2965_epoch_0000_step_0081000.pt +2026-04-10 11:54:16 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:54:19 - INFO - parrotllm.training - step 81010 | epoch 0 | loss 3.3250 | lr 5.01e-04 | grad 0.2760 +2026-04-10 11:54:22 - INFO - parrotllm.training - step 81020 | epoch 0 | loss 3.3522 | lr 5.01e-04 | grad 0.2816 +2026-04-10 11:54:25 - INFO - parrotllm.training - step 81030 | epoch 0 | loss 3.3517 | lr 5.01e-04 | grad 0.2960 +2026-04-10 11:54:28 - INFO - parrotllm.training - step 81040 | epoch 0 | loss 3.3214 | lr 5.00e-04 | grad 0.2945 +2026-04-10 11:54:31 - INFO - parrotllm.training - step 81050 | epoch 0 | loss 3.3121 | lr 5.00e-04 | grad 0.2935 +2026-04-10 11:54:34 - INFO - parrotllm.training - step 81060 | epoch 0 | loss 3.3644 | lr 5.00e-04 | grad 0.2882 +2026-04-10 11:54:37 - INFO - parrotllm.training - step 81070 | epoch 0 | loss 3.4112 | lr 5.00e-04 | grad 0.2765 +2026-04-10 11:54:40 - INFO - parrotllm.training - step 81080 | epoch 0 | loss 3.3101 | lr 5.00e-04 | grad 0.2640 +2026-04-10 11:54:43 - INFO - parrotllm.training - step 81090 | epoch 0 | loss 3.4415 | lr 5.00e-04 | grad 0.2876 +2026-04-10 11:54:46 - INFO - parrotllm.training - step 81100 | epoch 0 | loss 3.3124 | lr 4.99e-04 | grad 0.3258 +2026-04-10 11:54:49 - INFO - parrotllm.training - step 81110 | epoch 0 | loss 3.3194 | lr 4.99e-04 | grad 0.2755 +2026-04-10 11:54:52 - INFO - parrotllm.training - step 81120 | epoch 0 | loss 3.2984 | lr 4.99e-04 | grad 0.3051 +2026-04-10 11:54:55 - INFO - parrotllm.training - step 81130 | epoch 0 | loss 3.2518 | lr 4.99e-04 | grad 0.3036 +2026-04-10 11:54:58 - INFO - parrotllm.training - step 81140 | epoch 0 | loss 3.3453 | lr 4.99e-04 | grad 0.2998 +2026-04-10 11:55:01 - INFO - parrotllm.training - step 81150 | epoch 0 | loss 3.3893 | lr 4.99e-04 | grad 0.2978 +2026-04-10 11:55:04 - INFO - parrotllm.training - step 81160 | epoch 0 | loss 3.3473 | lr 4.99e-04 | grad 0.2865 +2026-04-10 11:55:07 - INFO - parrotllm.training - step 81170 | epoch 0 | loss 3.4178 | lr 4.98e-04 | grad 0.3163 +2026-04-10 11:55:10 - INFO - parrotllm.training - step 81180 | epoch 0 | loss 3.3162 | lr 4.98e-04 | grad 0.2937 +2026-04-10 11:55:14 - INFO - parrotllm.training - step 81190 | epoch 0 | loss 3.3586 | lr 4.98e-04 | grad 0.2857 +2026-04-10 11:55:17 - INFO - parrotllm.training - step 81200 | epoch 0 | loss 3.3536 | lr 4.98e-04 | grad 0.3104 +2026-04-10 11:55:20 - INFO - parrotllm.training - step 81210 | epoch 0 | loss 3.4363 | lr 4.98e-04 | grad 0.2659 +2026-04-10 11:55:23 - INFO - parrotllm.training - step 81220 | epoch 0 | loss 3.3414 | lr 4.98e-04 | grad 0.2909 +2026-04-10 11:55:26 - INFO - parrotllm.training - step 81230 | epoch 0 | loss 3.3123 | lr 4.97e-04 | grad 0.2769 +2026-04-10 11:55:29 - INFO - parrotllm.training - step 81240 | epoch 0 | loss 3.2898 | lr 4.97e-04 | grad 0.2742 +2026-04-10 11:55:32 - INFO - parrotllm.training - step 81250 | epoch 0 | loss 3.3055 | lr 4.97e-04 | grad 0.2875 +2026-04-10 11:55:35 - INFO - parrotllm.training - step 81260 | epoch 0 | loss 3.3539 | lr 4.97e-04 | grad 0.2736 +2026-04-10 11:55:38 - INFO - parrotllm.training - step 81270 | epoch 0 | loss 3.3896 | lr 4.97e-04 | grad 0.2713 +2026-04-10 11:55:41 - INFO - parrotllm.training - step 81280 | epoch 0 | loss 3.2849 | lr 4.97e-04 | grad 0.2691 +2026-04-10 11:55:44 - INFO - parrotllm.training - step 81290 | epoch 0 | loss 3.3277 | lr 4.97e-04 | grad 0.2886 +2026-04-10 11:55:47 - INFO - parrotllm.training - step 81300 | epoch 0 | loss 3.4284 | lr 4.96e-04 | grad 0.2716 +2026-04-10 11:55:50 - INFO - parrotllm.training - step 81310 | epoch 0 | loss 3.4683 | lr 4.96e-04 | grad 0.2977 +2026-04-10 11:55:53 - INFO - parrotllm.training - step 81320 | epoch 0 | loss 3.3622 | lr 4.96e-04 | grad 0.2972 +2026-04-10 11:55:56 - INFO - parrotllm.training - step 81330 | epoch 0 | loss 3.3060 | lr 4.96e-04 | grad 0.3040 +2026-04-10 11:55:59 - INFO - parrotllm.training - step 81340 | epoch 0 | loss 3.3994 | lr 4.96e-04 | grad 0.3184 +2026-04-10 11:56:02 - INFO - parrotllm.training - step 81350 | epoch 0 | loss 3.3628 | lr 4.96e-04 | grad 0.3268 +2026-04-10 11:56:05 - INFO - parrotllm.training - step 81360 | epoch 0 | loss 3.3916 | lr 4.95e-04 | grad 0.2929 +2026-04-10 11:56:08 - INFO - parrotllm.training - step 81370 | epoch 0 | loss 3.3377 | lr 4.95e-04 | grad 0.3025 +2026-04-10 11:56:12 - INFO - parrotllm.training - step 81380 | epoch 0 | loss 3.3110 | lr 4.95e-04 | grad 0.2958 +2026-04-10 11:56:15 - INFO - parrotllm.training - step 81390 | epoch 0 | loss 3.3529 | lr 4.95e-04 | grad 0.2879 +2026-04-10 11:56:18 - INFO - parrotllm.training - step 81400 | epoch 0 | loss 3.3438 | lr 4.95e-04 | grad 0.3071 +2026-04-10 11:56:21 - INFO - parrotllm.training - step 81410 | epoch 0 | loss 3.3279 | lr 4.95e-04 | grad 0.2967 +2026-04-10 11:56:24 - INFO - parrotllm.training - step 81420 | epoch 0 | loss 3.4395 | lr 4.95e-04 | grad 0.2745 +2026-04-10 11:56:27 - INFO - parrotllm.training - step 81430 | epoch 0 | loss 3.3702 | lr 4.94e-04 | grad 0.3203 +2026-04-10 11:56:30 - INFO - parrotllm.training - step 81440 | epoch 0 | loss 3.3640 | lr 4.94e-04 | grad 0.2835 +2026-04-10 11:56:33 - INFO - parrotllm.training - step 81450 | epoch 0 | loss 3.2746 | lr 4.94e-04 | grad 0.3022 +2026-04-10 11:56:36 - INFO - parrotllm.training - step 81460 | epoch 0 | loss 3.3459 | lr 4.94e-04 | grad 0.3123 +2026-04-10 11:56:39 - INFO - parrotllm.training - step 81470 | epoch 0 | loss 3.3922 | lr 4.94e-04 | grad 0.3297 +2026-04-10 11:56:42 - INFO - parrotllm.training - step 81480 | epoch 0 | loss 3.2341 | lr 4.94e-04 | grad 0.2953 +2026-04-10 11:56:45 - INFO - parrotllm.training - step 81490 | epoch 0 | loss 3.3409 | lr 4.93e-04 | grad 0.2857 +2026-04-10 11:56:48 - INFO - parrotllm.training - step 81500 | epoch 0 | loss 3.4605 | lr 4.93e-04 | grad 0.2776 +2026-04-10 11:56:48 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:56:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:56:51 - INFO - parrotllm.training - Train: loss=3.4605, ppl=31.83 +2026-04-10 11:56:51 - INFO - parrotllm.training - Val: loss=3.2966, ppl=27.02 +2026-04-10 11:56:52 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2966_epoch_0000_step_0081500.pt +2026-04-10 11:56:54 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2965, min_delta=0.001000). +2026-04-10 11:56:54 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:56:57 - INFO - parrotllm.training - step 81510 | epoch 0 | loss 3.2546 | lr 4.93e-04 | grad 0.2849 +2026-04-10 11:57:00 - INFO - parrotllm.training - step 81520 | epoch 0 | loss 3.3288 | lr 4.93e-04 | grad 0.2752 +2026-04-10 11:57:03 - INFO - parrotllm.training - step 81530 | epoch 0 | loss 3.4933 | lr 4.93e-04 | grad 0.2796 +2026-04-10 11:57:06 - INFO - parrotllm.training - step 81540 | epoch 0 | loss 3.3598 | lr 4.93e-04 | grad 0.2901 +2026-04-10 11:57:09 - INFO - parrotllm.training - step 81550 | epoch 0 | loss 3.2206 | lr 4.93e-04 | grad 0.3009 +2026-04-10 11:57:12 - INFO - parrotllm.training - step 81560 | epoch 0 | loss 3.3685 | lr 4.92e-04 | grad 0.3037 +2026-04-10 11:57:15 - INFO - parrotllm.training - step 81570 | epoch 0 | loss 3.3380 | lr 4.92e-04 | grad 0.3050 +2026-04-10 11:57:18 - INFO - parrotllm.training - step 81580 | epoch 0 | loss 3.3745 | lr 4.92e-04 | grad 0.2917 +2026-04-10 11:57:21 - INFO - parrotllm.training - step 81590 | epoch 0 | loss 3.2889 | lr 4.92e-04 | grad 0.3182 +2026-04-10 11:57:24 - INFO - parrotllm.training - step 81600 | epoch 0 | loss 3.3145 | lr 4.92e-04 | grad 0.2974 +2026-04-10 11:57:28 - INFO - parrotllm.training - step 81610 | epoch 0 | loss 3.3485 | lr 4.92e-04 | grad 0.2727 +2026-04-10 11:57:31 - INFO - parrotllm.training - step 81620 | epoch 0 | loss 3.3174 | lr 4.91e-04 | grad 0.2734 +2026-04-10 11:57:34 - INFO - parrotllm.training - step 81630 | epoch 0 | loss 3.3919 | lr 4.91e-04 | grad 0.3028 +2026-04-10 11:57:37 - INFO - parrotllm.training - step 81640 | epoch 0 | loss 3.3570 | lr 4.91e-04 | grad 0.2825 +2026-04-10 11:57:40 - INFO - parrotllm.training - step 81650 | epoch 0 | loss 3.3238 | lr 4.91e-04 | grad 0.2749 +2026-04-10 11:57:43 - INFO - parrotllm.training - step 81660 | epoch 0 | loss 3.4188 | lr 4.91e-04 | grad 0.2989 +2026-04-10 11:57:46 - INFO - parrotllm.training - step 81670 | epoch 0 | loss 3.3093 | lr 4.91e-04 | grad 0.2760 +2026-04-10 11:57:49 - INFO - parrotllm.training - step 81680 | epoch 0 | loss 3.3107 | lr 4.91e-04 | grad 0.3062 +2026-04-10 11:57:52 - INFO - parrotllm.training - step 81690 | epoch 0 | loss 3.3980 | lr 4.90e-04 | grad 0.2803 +2026-04-10 11:57:55 - INFO - parrotllm.training - step 81700 | epoch 0 | loss 3.4237 | lr 4.90e-04 | grad 0.2899 +2026-04-10 11:57:58 - INFO - parrotllm.training - step 81710 | epoch 0 | loss 3.3544 | lr 4.90e-04 | grad 0.3141 +2026-04-10 11:58:01 - INFO - parrotllm.training - step 81720 | epoch 0 | loss 3.3515 | lr 4.90e-04 | grad 0.2737 +2026-04-10 11:58:04 - INFO - parrotllm.training - step 81730 | epoch 0 | loss 3.4177 | lr 4.90e-04 | grad 0.2830 +2026-04-10 11:58:07 - INFO - parrotllm.training - step 81740 | epoch 0 | loss 3.2980 | lr 4.90e-04 | grad 0.2918 +2026-04-10 11:58:10 - INFO - parrotllm.training - step 81750 | epoch 0 | loss 3.3032 | lr 4.89e-04 | grad 0.2616 +2026-04-10 11:58:13 - INFO - parrotllm.training - step 81760 | epoch 0 | loss 3.3378 | lr 4.89e-04 | grad 0.2818 +2026-04-10 11:58:16 - INFO - parrotllm.training - step 81770 | epoch 0 | loss 3.4313 | lr 4.89e-04 | grad 0.2619 +2026-04-10 11:58:19 - INFO - parrotllm.training - step 81780 | epoch 0 | loss 3.3062 | lr 4.89e-04 | grad 0.3106 +2026-04-10 11:58:22 - INFO - parrotllm.training - step 81790 | epoch 0 | loss 3.2822 | lr 4.89e-04 | grad 0.2844 +2026-04-10 11:58:25 - INFO - parrotllm.training - step 81800 | epoch 0 | loss 3.4257 | lr 4.89e-04 | grad 0.3060 +2026-04-10 11:58:29 - INFO - parrotllm.training - step 81810 | epoch 0 | loss 3.3995 | lr 4.89e-04 | grad 0.2633 +2026-04-10 11:58:32 - INFO - parrotllm.training - step 81820 | epoch 0 | loss 3.4372 | lr 4.88e-04 | grad 0.2673 +2026-04-10 11:58:35 - INFO - parrotllm.training - step 81830 | epoch 0 | loss 3.3352 | lr 4.88e-04 | grad 0.3002 +2026-04-10 11:58:38 - INFO - parrotllm.training - step 81840 | epoch 0 | loss 3.4923 | lr 4.88e-04 | grad 0.2786 +2026-04-10 11:58:41 - INFO - parrotllm.training - step 81850 | epoch 0 | loss 3.3778 | lr 4.88e-04 | grad 0.2715 +2026-04-10 11:58:44 - INFO - parrotllm.training - step 81860 | epoch 0 | loss 3.3400 | lr 4.88e-04 | grad 0.3057 +2026-04-10 11:58:47 - INFO - parrotllm.training - step 81870 | epoch 0 | loss 3.3254 | lr 4.88e-04 | grad 0.2808 +2026-04-10 11:58:50 - INFO - parrotllm.training - step 81880 | epoch 0 | loss 3.3380 | lr 4.88e-04 | grad 0.2972 +2026-04-10 11:58:53 - INFO - parrotllm.training - step 81890 | epoch 0 | loss 3.3330 | lr 4.87e-04 | grad 0.3333 +2026-04-10 11:58:56 - INFO - parrotllm.training - step 81900 | epoch 0 | loss 3.2649 | lr 4.87e-04 | grad 0.2932 +2026-04-10 11:58:59 - INFO - parrotllm.training - step 81910 | epoch 0 | loss 3.3488 | lr 4.87e-04 | grad 0.2900 +2026-04-10 11:59:02 - INFO - parrotllm.training - step 81920 | epoch 0 | loss 3.4317 | lr 4.87e-04 | grad 0.3147 +2026-04-10 11:59:05 - INFO - parrotllm.training - step 81930 | epoch 0 | loss 3.3480 | lr 4.87e-04 | grad 0.2751 +2026-04-10 11:59:08 - INFO - parrotllm.training - step 81940 | epoch 0 | loss 3.3038 | lr 4.87e-04 | grad 0.3343 +2026-04-10 11:59:11 - INFO - parrotllm.training - step 81950 | epoch 0 | loss 3.3854 | lr 4.86e-04 | grad 0.2745 +2026-04-10 11:59:14 - INFO - parrotllm.training - step 81960 | epoch 0 | loss 3.3247 | lr 4.86e-04 | grad 0.2876 +2026-04-10 11:59:17 - INFO - parrotllm.training - step 81970 | epoch 0 | loss 3.3602 | lr 4.86e-04 | grad 0.2746 +2026-04-10 11:59:20 - INFO - parrotllm.training - step 81980 | epoch 0 | loss 3.3165 | lr 4.86e-04 | grad 0.2890 +2026-04-10 11:59:23 - INFO - parrotllm.training - step 81990 | epoch 0 | loss 3.3775 | lr 4.86e-04 | grad 0.2809 +2026-04-10 11:59:26 - INFO - parrotllm.training - step 82000 | epoch 0 | loss 3.3791 | lr 4.86e-04 | grad 0.3004 +2026-04-10 11:59:26 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 11:59:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:59:29 - INFO - parrotllm.training - Train: loss=3.3791, ppl=29.35 +2026-04-10 11:59:29 - INFO - parrotllm.training - Val: loss=3.2954, ppl=26.99 +2026-04-10 11:59:29 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 11:59:30 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2954_epoch_0000_step_0082000.pt +2026-04-10 11:59:33 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 11:59:36 - INFO - parrotllm.training - step 82010 | epoch 0 | loss 3.2785 | lr 4.86e-04 | grad 0.2810 +2026-04-10 11:59:39 - INFO - parrotllm.training - step 82020 | epoch 0 | loss 3.3745 | lr 4.85e-04 | grad 0.3057 +2026-04-10 11:59:42 - INFO - parrotllm.training - step 82030 | epoch 0 | loss 3.3216 | lr 4.85e-04 | grad 0.2756 +2026-04-10 11:59:45 - INFO - parrotllm.training - step 82040 | epoch 0 | loss 3.4540 | lr 4.85e-04 | grad 0.2756 +2026-04-10 11:59:48 - INFO - parrotllm.training - step 82050 | epoch 0 | loss 3.2470 | lr 4.85e-04 | grad 0.2729 +2026-04-10 11:59:51 - INFO - parrotllm.training - step 82060 | epoch 0 | loss 3.4331 | lr 4.85e-04 | grad 0.2869 +2026-04-10 11:59:54 - INFO - parrotllm.training - step 82070 | epoch 0 | loss 3.3610 | lr 4.85e-04 | grad 0.3083 +2026-04-10 11:59:57 - INFO - parrotllm.training - step 82080 | epoch 0 | loss 3.2960 | lr 4.84e-04 | grad 0.3264 +2026-04-10 12:00:00 - INFO - parrotllm.training - step 82090 | epoch 0 | loss 3.3568 | lr 4.84e-04 | grad 0.3127 +2026-04-10 12:00:03 - INFO - parrotllm.training - step 82100 | epoch 0 | loss 3.3671 | lr 4.84e-04 | grad 0.3239 +2026-04-10 12:00:06 - INFO - parrotllm.training - step 82110 | epoch 0 | loss 3.3408 | lr 4.84e-04 | grad 0.2877 +2026-04-10 12:00:09 - INFO - parrotllm.training - step 82120 | epoch 0 | loss 3.3117 | lr 4.84e-04 | grad 0.2792 +2026-04-10 12:00:12 - INFO - parrotllm.training - step 82130 | epoch 0 | loss 3.2707 | lr 4.84e-04 | grad 0.2717 +2026-04-10 12:00:15 - INFO - parrotllm.training - step 82140 | epoch 0 | loss 3.3623 | lr 4.84e-04 | grad 0.2990 +2026-04-10 12:00:18 - INFO - parrotllm.training - step 82150 | epoch 0 | loss 3.3053 | lr 4.83e-04 | grad 0.2867 +2026-04-10 12:00:21 - INFO - parrotllm.training - step 82160 | epoch 0 | loss 3.3249 | lr 4.83e-04 | grad 0.2799 +2026-04-10 12:00:25 - INFO - parrotllm.training - step 82170 | epoch 0 | loss 3.2844 | lr 4.83e-04 | grad 0.2843 +2026-04-10 12:00:28 - INFO - parrotllm.training - step 82180 | epoch 0 | loss 3.2887 | lr 4.83e-04 | grad 0.2843 +2026-04-10 12:00:31 - INFO - parrotllm.training - step 82190 | epoch 0 | loss 3.3967 | lr 4.83e-04 | grad 0.2953 +2026-04-10 12:00:34 - INFO - parrotllm.training - step 82200 | epoch 0 | loss 3.3423 | lr 4.83e-04 | grad 0.2818 +2026-04-10 12:00:37 - INFO - parrotllm.training - step 82210 | epoch 0 | loss 3.4301 | lr 4.82e-04 | grad 0.2868 +2026-04-10 12:00:40 - INFO - parrotllm.training - step 82220 | epoch 0 | loss 3.3244 | lr 4.82e-04 | grad 0.2913 +2026-04-10 12:00:43 - INFO - parrotllm.training - step 82230 | epoch 0 | loss 3.3807 | lr 4.82e-04 | grad 0.3256 +2026-04-10 12:00:46 - INFO - parrotllm.training - step 82240 | epoch 0 | loss 3.4473 | lr 4.82e-04 | grad 0.2790 +2026-04-10 12:00:49 - INFO - parrotllm.training - step 82250 | epoch 0 | loss 3.4923 | lr 4.82e-04 | grad 0.2888 +2026-04-10 12:00:52 - INFO - parrotllm.training - step 82260 | epoch 0 | loss 3.4151 | lr 4.82e-04 | grad 0.2931 +2026-04-10 12:00:55 - INFO - parrotllm.training - step 82270 | epoch 0 | loss 3.3677 | lr 4.82e-04 | grad 0.2925 +2026-04-10 12:00:58 - INFO - parrotllm.training - step 82280 | epoch 0 | loss 3.4274 | lr 4.81e-04 | grad 0.2849 +2026-04-10 12:01:01 - INFO - parrotllm.training - step 82290 | epoch 0 | loss 3.2964 | lr 4.81e-04 | grad 0.3028 +2026-04-10 12:01:04 - INFO - parrotllm.training - step 82300 | epoch 0 | loss 3.2682 | lr 4.81e-04 | grad 0.3109 +2026-04-10 12:01:07 - INFO - parrotllm.training - step 82310 | epoch 0 | loss 3.4803 | lr 4.81e-04 | grad 0.2968 +2026-04-10 12:01:10 - INFO - parrotllm.training - step 82320 | epoch 0 | loss 3.4211 | lr 4.81e-04 | grad 0.2782 +2026-04-10 12:01:13 - INFO - parrotllm.training - step 82330 | epoch 0 | loss 3.3551 | lr 4.81e-04 | grad 0.2908 +2026-04-10 12:01:16 - INFO - parrotllm.training - step 82340 | epoch 0 | loss 3.2303 | lr 4.81e-04 | grad 0.3070 +2026-04-10 12:01:19 - INFO - parrotllm.training - step 82350 | epoch 0 | loss 3.3429 | lr 4.80e-04 | grad 0.2957 +2026-04-10 12:01:22 - INFO - parrotllm.training - step 82360 | epoch 0 | loss 3.3735 | lr 4.80e-04 | grad 0.3205 +2026-04-10 12:01:26 - INFO - parrotllm.training - step 82370 | epoch 0 | loss 3.3423 | lr 4.80e-04 | grad 0.3263 +2026-04-10 12:01:29 - INFO - parrotllm.training - step 82380 | epoch 0 | loss 3.3420 | lr 4.80e-04 | grad 0.2809 +2026-04-10 12:01:32 - INFO - parrotllm.training - step 82390 | epoch 0 | loss 3.3533 | lr 4.80e-04 | grad 0.2784 +2026-04-10 12:01:35 - INFO - parrotllm.training - step 82400 | epoch 0 | loss 3.3376 | lr 4.80e-04 | grad 0.3057 +2026-04-10 12:01:38 - INFO - parrotllm.training - step 82410 | epoch 0 | loss 3.3037 | lr 4.79e-04 | grad 0.2877 +2026-04-10 12:01:41 - INFO - parrotllm.training - step 82420 | epoch 0 | loss 3.3069 | lr 4.79e-04 | grad 0.2731 +2026-04-10 12:01:44 - INFO - parrotllm.training - step 82430 | epoch 0 | loss 3.4034 | lr 4.79e-04 | grad 0.3260 +2026-04-10 12:01:47 - INFO - parrotllm.training - step 82440 | epoch 0 | loss 3.4438 | lr 4.79e-04 | grad 0.2835 +2026-04-10 12:01:50 - INFO - parrotllm.training - step 82450 | epoch 0 | loss 3.3303 | lr 4.79e-04 | grad 0.2734 +2026-04-10 12:01:53 - INFO - parrotllm.training - step 82460 | epoch 0 | loss 3.3221 | lr 4.79e-04 | grad 0.2789 +2026-04-10 12:01:56 - INFO - parrotllm.training - step 82470 | epoch 0 | loss 3.3215 | lr 4.79e-04 | grad 0.2712 +2026-04-10 12:01:59 - INFO - parrotllm.training - step 82480 | epoch 0 | loss 3.3050 | lr 4.78e-04 | grad 0.2814 +2026-04-10 12:02:02 - INFO - parrotllm.training - step 82490 | epoch 0 | loss 3.3384 | lr 4.78e-04 | grad 0.2961 +2026-04-10 12:02:05 - INFO - parrotllm.training - step 82500 | epoch 0 | loss 3.3692 | lr 4.78e-04 | grad 0.3124 +2026-04-10 12:02:05 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:02:05 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:02:08 - INFO - parrotllm.training - Train: loss=3.3692, ppl=29.05 +2026-04-10 12:02:08 - INFO - parrotllm.training - Val: loss=3.2954, ppl=26.99 +2026-04-10 12:02:09 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2954_epoch_0000_step_0082500.pt +2026-04-10 12:02:11 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2954, min_delta=0.001000). +2026-04-10 12:02:11 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:02:12 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0082500.pt +2026-04-10 12:02:17 - INFO - parrotllm.training - step 82510 | epoch 0 | loss 3.2755 | lr 4.78e-04 | grad 0.2981 +2026-04-10 12:02:20 - INFO - parrotllm.training - step 82520 | epoch 0 | loss 3.2719 | lr 4.78e-04 | grad 0.3018 +2026-04-10 12:02:23 - INFO - parrotllm.training - step 82530 | epoch 0 | loss 3.2775 | lr 4.78e-04 | grad 0.2777 +2026-04-10 12:02:26 - INFO - parrotllm.training - step 82540 | epoch 0 | loss 3.3708 | lr 4.77e-04 | grad 0.2959 +2026-04-10 12:02:29 - INFO - parrotllm.training - step 82550 | epoch 0 | loss 3.2615 | lr 4.77e-04 | grad 0.2895 +2026-04-10 12:02:32 - INFO - parrotllm.training - step 82560 | epoch 0 | loss 3.4021 | lr 4.77e-04 | grad 0.2951 +2026-04-10 12:02:35 - INFO - parrotllm.training - step 82570 | epoch 0 | loss 3.3533 | lr 4.77e-04 | grad 0.2820 +2026-04-10 12:02:39 - INFO - parrotllm.training - step 82580 | epoch 0 | loss 3.2699 | lr 4.77e-04 | grad 0.2861 +2026-04-10 12:02:42 - INFO - parrotllm.training - step 82590 | epoch 0 | loss 3.3326 | lr 4.77e-04 | grad 0.2842 +2026-04-10 12:02:45 - INFO - parrotllm.training - step 82600 | epoch 0 | loss 3.3745 | lr 4.77e-04 | grad 0.2715 +2026-04-10 12:02:48 - INFO - parrotllm.training - step 82610 | epoch 0 | loss 3.4889 | lr 4.76e-04 | grad 0.2969 +2026-04-10 12:02:51 - INFO - parrotllm.training - step 82620 | epoch 0 | loss 3.3221 | lr 4.76e-04 | grad 0.2857 +2026-04-10 12:02:54 - INFO - parrotllm.training - step 82630 | epoch 0 | loss 3.3384 | lr 4.76e-04 | grad 0.2926 +2026-04-10 12:02:57 - INFO - parrotllm.training - step 82640 | epoch 0 | loss 3.4045 | lr 4.76e-04 | grad 0.2771 +2026-04-10 12:03:00 - INFO - parrotllm.training - step 82650 | epoch 0 | loss 3.3777 | lr 4.76e-04 | grad 0.2729 +2026-04-10 12:03:03 - INFO - parrotllm.training - step 82660 | epoch 0 | loss 3.3753 | lr 4.76e-04 | grad 0.2748 +2026-04-10 12:03:06 - INFO - parrotllm.training - step 82670 | epoch 0 | loss 3.3724 | lr 4.76e-04 | grad 0.2886 +2026-04-10 12:03:09 - INFO - parrotllm.training - step 82680 | epoch 0 | loss 3.3546 | lr 4.75e-04 | grad 0.3007 +2026-04-10 12:03:12 - INFO - parrotllm.training - step 82690 | epoch 0 | loss 3.3878 | lr 4.75e-04 | grad 0.2814 +2026-04-10 12:03:15 - INFO - parrotllm.training - step 82700 | epoch 0 | loss 3.2902 | lr 4.75e-04 | grad 0.2863 +2026-04-10 12:03:18 - INFO - parrotllm.training - step 82710 | epoch 0 | loss 3.3591 | lr 4.75e-04 | grad 0.3381 +2026-04-10 12:03:21 - INFO - parrotllm.training - step 82720 | epoch 0 | loss 3.3615 | lr 4.75e-04 | grad 0.2675 +2026-04-10 12:03:24 - INFO - parrotllm.training - step 82730 | epoch 0 | loss 3.3401 | lr 4.75e-04 | grad 0.3030 +2026-04-10 12:03:27 - INFO - parrotllm.training - step 82740 | epoch 0 | loss 3.3505 | lr 4.74e-04 | grad 0.3165 +2026-04-10 12:03:30 - INFO - parrotllm.training - step 82750 | epoch 0 | loss 3.4185 | lr 4.74e-04 | grad 0.2967 +2026-04-10 12:03:33 - INFO - parrotllm.training - step 82760 | epoch 0 | loss 3.3269 | lr 4.74e-04 | grad 0.3092 +2026-04-10 12:03:37 - INFO - parrotllm.training - step 82770 | epoch 0 | loss 3.2831 | lr 4.74e-04 | grad 0.3191 +2026-04-10 12:03:40 - INFO - parrotllm.training - step 82780 | epoch 0 | loss 3.2999 | lr 4.74e-04 | grad 0.2689 +2026-04-10 12:03:43 - INFO - parrotllm.training - step 82790 | epoch 0 | loss 3.2809 | lr 4.74e-04 | grad 0.2797 +2026-04-10 12:03:46 - INFO - parrotllm.training - step 82800 | epoch 0 | loss 3.4049 | lr 4.74e-04 | grad 0.2859 +2026-04-10 12:03:49 - INFO - parrotllm.training - step 82810 | epoch 0 | loss 3.3369 | lr 4.73e-04 | grad 0.2887 +2026-04-10 12:03:52 - INFO - parrotllm.training - step 82820 | epoch 0 | loss 3.3637 | lr 4.73e-04 | grad 0.3000 +2026-04-10 12:03:55 - INFO - parrotllm.training - step 82830 | epoch 0 | loss 3.4408 | lr 4.73e-04 | grad 0.2886 +2026-04-10 12:03:58 - INFO - parrotllm.training - step 82840 | epoch 0 | loss 3.3618 | lr 4.73e-04 | grad 0.2791 +2026-04-10 12:04:01 - INFO - parrotllm.training - step 82850 | epoch 0 | loss 3.2773 | lr 4.73e-04 | grad 0.2910 +2026-04-10 12:04:04 - INFO - parrotllm.training - step 82860 | epoch 0 | loss 3.4429 | lr 4.73e-04 | grad 0.2729 +2026-04-10 12:04:07 - INFO - parrotllm.training - step 82870 | epoch 0 | loss 3.2540 | lr 4.73e-04 | grad 0.2806 +2026-04-10 12:04:10 - INFO - parrotllm.training - step 82880 | epoch 0 | loss 3.3434 | lr 4.72e-04 | grad 0.2957 +2026-04-10 12:04:13 - INFO - parrotllm.training - step 82890 | epoch 0 | loss 3.2944 | lr 4.72e-04 | grad 0.2803 +2026-04-10 12:04:16 - INFO - parrotllm.training - step 82900 | epoch 0 | loss 3.3201 | lr 4.72e-04 | grad 0.3032 +2026-04-10 12:04:19 - INFO - parrotllm.training - step 82910 | epoch 0 | loss 3.2378 | lr 4.72e-04 | grad 0.3118 +2026-04-10 12:04:22 - INFO - parrotllm.training - step 82920 | epoch 0 | loss 3.2726 | lr 4.72e-04 | grad 0.3051 +2026-04-10 12:04:25 - INFO - parrotllm.training - step 82930 | epoch 0 | loss 3.3863 | lr 4.72e-04 | grad 0.2938 +2026-04-10 12:04:28 - INFO - parrotllm.training - step 82940 | epoch 0 | loss 3.3422 | lr 4.71e-04 | grad 0.2790 +2026-04-10 12:04:31 - INFO - parrotllm.training - step 82950 | epoch 0 | loss 3.2563 | lr 4.71e-04 | grad 0.2868 +2026-04-10 12:04:34 - INFO - parrotllm.training - step 82960 | epoch 0 | loss 3.4010 | lr 4.71e-04 | grad 0.2860 +2026-04-10 12:04:38 - INFO - parrotllm.training - step 82970 | epoch 0 | loss 3.4804 | lr 4.71e-04 | grad 0.2783 +2026-04-10 12:04:41 - INFO - parrotllm.training - step 82980 | epoch 0 | loss 3.3703 | lr 4.71e-04 | grad 0.3232 +2026-04-10 12:04:44 - INFO - parrotllm.training - step 82990 | epoch 0 | loss 3.3589 | lr 4.71e-04 | grad 0.2657 +2026-04-10 12:04:47 - INFO - parrotllm.training - step 83000 | epoch 0 | loss 3.4067 | lr 4.71e-04 | grad 0.3521 +2026-04-10 12:04:47 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:04:47 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:04:50 - INFO - parrotllm.training - Train: loss=3.4067, ppl=30.16 +2026-04-10 12:04:50 - INFO - parrotllm.training - Val: loss=3.2927, ppl=26.92 +2026-04-10 12:04:50 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:04:51 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2927_epoch_0000_step_0083000.pt +2026-04-10 12:04:53 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:04:56 - INFO - parrotllm.training - step 83010 | epoch 0 | loss 3.3856 | lr 4.70e-04 | grad 0.3008 +2026-04-10 12:04:59 - INFO - parrotllm.training - step 83020 | epoch 0 | loss 3.2833 | lr 4.70e-04 | grad 0.2979 +2026-04-10 12:05:02 - INFO - parrotllm.training - step 83030 | epoch 0 | loss 3.2507 | lr 4.70e-04 | grad 0.2629 +2026-04-10 12:05:05 - INFO - parrotllm.training - step 83040 | epoch 0 | loss 3.3404 | lr 4.70e-04 | grad 0.2844 +2026-04-10 12:05:08 - INFO - parrotllm.training - step 83050 | epoch 0 | loss 3.2924 | lr 4.70e-04 | grad 0.2743 +2026-04-10 12:05:11 - INFO - parrotllm.training - step 83060 | epoch 0 | loss 3.4254 | lr 4.70e-04 | grad 0.3054 +2026-04-10 12:05:14 - INFO - parrotllm.training - step 83070 | epoch 0 | loss 3.3116 | lr 4.70e-04 | grad 0.2767 +2026-04-10 12:05:17 - INFO - parrotllm.training - step 83080 | epoch 0 | loss 3.4816 | lr 4.69e-04 | grad 0.2885 +2026-04-10 12:05:20 - INFO - parrotllm.training - step 83090 | epoch 0 | loss 3.4183 | lr 4.69e-04 | grad 0.3111 +2026-04-10 12:05:23 - INFO - parrotllm.training - step 83100 | epoch 0 | loss 3.3365 | lr 4.69e-04 | grad 0.3132 +2026-04-10 12:05:26 - INFO - parrotllm.training - step 83110 | epoch 0 | loss 3.4094 | lr 4.69e-04 | grad 0.2934 +2026-04-10 12:05:30 - INFO - parrotllm.training - step 83120 | epoch 0 | loss 3.3603 | lr 4.69e-04 | grad 0.3231 +2026-04-10 12:05:33 - INFO - parrotllm.training - step 83130 | epoch 0 | loss 3.4184 | lr 4.69e-04 | grad 0.2809 +2026-04-10 12:05:36 - INFO - parrotllm.training - step 83140 | epoch 0 | loss 3.3252 | lr 4.68e-04 | grad 0.2948 +2026-04-10 12:05:39 - INFO - parrotllm.training - step 83150 | epoch 0 | loss 3.3799 | lr 4.68e-04 | grad 0.2918 +2026-04-10 12:05:42 - INFO - parrotllm.training - step 83160 | epoch 0 | loss 3.2820 | lr 4.68e-04 | grad 0.2767 +2026-04-10 12:05:45 - INFO - parrotllm.training - step 83170 | epoch 0 | loss 3.3921 | lr 4.68e-04 | grad 0.2804 +2026-04-10 12:05:48 - INFO - parrotllm.training - step 83180 | epoch 0 | loss 3.3445 | lr 4.68e-04 | grad 0.2791 +2026-04-10 12:05:51 - INFO - parrotllm.training - step 83190 | epoch 0 | loss 3.3066 | lr 4.68e-04 | grad 0.2805 +2026-04-10 12:05:54 - INFO - parrotllm.training - step 83200 | epoch 0 | loss 3.2999 | lr 4.68e-04 | grad 0.3086 +2026-04-10 12:05:57 - INFO - parrotllm.training - step 83210 | epoch 0 | loss 3.4826 | lr 4.67e-04 | grad 0.2984 +2026-04-10 12:06:00 - INFO - parrotllm.training - step 83220 | epoch 0 | loss 3.3877 | lr 4.67e-04 | grad 0.2855 +2026-04-10 12:06:03 - INFO - parrotllm.training - step 83230 | epoch 0 | loss 3.2438 | lr 4.67e-04 | grad 0.2887 +2026-04-10 12:06:06 - INFO - parrotllm.training - step 83240 | epoch 0 | loss 3.3248 | lr 4.67e-04 | grad 0.3044 +2026-04-10 12:06:09 - INFO - parrotllm.training - step 83250 | epoch 0 | loss 3.3542 | lr 4.67e-04 | grad 0.2944 +2026-04-10 12:06:12 - INFO - parrotllm.training - step 83260 | epoch 0 | loss 3.3215 | lr 4.67e-04 | grad 0.3056 +2026-04-10 12:06:13 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 83296/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 12:06:16 - INFO - parrotllm.training - step 83270 | epoch 0 | loss 3.4609 | lr 4.67e-04 | grad 0.2875 +2026-04-10 12:06:19 - INFO - parrotllm.training - step 83280 | epoch 0 | loss 3.4144 | lr 4.66e-04 | grad 0.2833 +2026-04-10 12:06:22 - INFO - parrotllm.training - step 83290 | epoch 0 | loss 3.3175 | lr 4.66e-04 | grad 0.2878 +2026-04-10 12:06:25 - INFO - parrotllm.training - step 83300 | epoch 0 | loss 3.3578 | lr 4.66e-04 | grad 0.2972 +2026-04-10 12:06:28 - INFO - parrotllm.training - step 83310 | epoch 0 | loss 3.2956 | lr 4.66e-04 | grad 0.3262 +2026-04-10 12:06:31 - INFO - parrotllm.training - step 83320 | epoch 0 | loss 3.2563 | lr 4.66e-04 | grad 0.3426 +2026-04-10 12:06:34 - INFO - parrotllm.training - step 83330 | epoch 0 | loss 3.4149 | lr 4.66e-04 | grad 0.3095 +2026-04-10 12:06:37 - INFO - parrotllm.training - step 83340 | epoch 0 | loss 3.3984 | lr 4.66e-04 | grad 0.2996 +2026-04-10 12:06:40 - INFO - parrotllm.training - step 83350 | epoch 0 | loss 3.4022 | lr 4.65e-04 | grad 0.2837 +2026-04-10 12:06:43 - INFO - parrotllm.training - step 83360 | epoch 0 | loss 3.3025 | lr 4.65e-04 | grad 0.3008 +2026-04-10 12:06:46 - INFO - parrotllm.training - step 83370 | epoch 0 | loss 3.2650 | lr 4.65e-04 | grad 0.2806 +2026-04-10 12:06:49 - INFO - parrotllm.training - step 83380 | epoch 0 | loss 3.3900 | lr 4.65e-04 | grad 0.2873 +2026-04-10 12:06:52 - INFO - parrotllm.training - step 83390 | epoch 0 | loss 3.3227 | lr 4.65e-04 | grad 0.2785 +2026-04-10 12:06:55 - INFO - parrotllm.training - step 83400 | epoch 0 | loss 3.3820 | lr 4.65e-04 | grad 0.2829 +2026-04-10 12:06:58 - INFO - parrotllm.training - step 83410 | epoch 0 | loss 3.4169 | lr 4.64e-04 | grad 0.2916 +2026-04-10 12:07:01 - INFO - parrotllm.training - step 83420 | epoch 0 | loss 3.2988 | lr 4.64e-04 | grad 0.2951 +2026-04-10 12:07:05 - INFO - parrotllm.training - step 83430 | epoch 0 | loss 3.3535 | lr 4.64e-04 | grad 0.2960 +2026-04-10 12:07:08 - INFO - parrotllm.training - step 83440 | epoch 0 | loss 3.3890 | lr 4.64e-04 | grad 0.2878 +2026-04-10 12:07:11 - INFO - parrotllm.training - step 83450 | epoch 0 | loss 3.3490 | lr 4.64e-04 | grad 0.3261 +2026-04-10 12:07:14 - INFO - parrotllm.training - step 83460 | epoch 0 | loss 3.3389 | lr 4.64e-04 | grad 0.2822 +2026-04-10 12:07:17 - INFO - parrotllm.training - step 83470 | epoch 0 | loss 3.3367 | lr 4.64e-04 | grad 0.3220 +2026-04-10 12:07:20 - INFO - parrotllm.training - step 83480 | epoch 0 | loss 3.4683 | lr 4.63e-04 | grad 0.3234 +2026-04-10 12:07:23 - INFO - parrotllm.training - step 83490 | epoch 0 | loss 3.3974 | lr 4.63e-04 | grad 0.2885 +2026-04-10 12:07:26 - INFO - parrotllm.training - step 83500 | epoch 0 | loss 3.3593 | lr 4.63e-04 | grad 0.2789 +2026-04-10 12:07:26 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:07:26 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:07:29 - INFO - parrotllm.training - Train: loss=3.3593, ppl=28.77 +2026-04-10 12:07:29 - INFO - parrotllm.training - Val: loss=3.2901, ppl=26.85 +2026-04-10 12:07:29 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:07:30 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2901_epoch_0000_step_0083500.pt +2026-04-10 12:07:32 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:07:35 - INFO - parrotllm.training - step 83510 | epoch 0 | loss 3.4036 | lr 4.63e-04 | grad 0.3089 +2026-04-10 12:07:38 - INFO - parrotllm.training - step 83520 | epoch 0 | loss 3.2264 | lr 4.63e-04 | grad 0.2891 +2026-04-10 12:07:41 - INFO - parrotllm.training - step 83530 | epoch 0 | loss 3.3115 | lr 4.63e-04 | grad 0.3011 +2026-04-10 12:07:44 - INFO - parrotllm.training - step 83540 | epoch 0 | loss 3.4999 | lr 4.63e-04 | grad 0.3376 +2026-04-10 12:07:47 - INFO - parrotllm.training - step 83550 | epoch 0 | loss 3.4001 | lr 4.62e-04 | grad 0.2929 +2026-04-10 12:07:50 - INFO - parrotllm.training - step 83560 | epoch 0 | loss 3.2638 | lr 4.62e-04 | grad 0.3057 +2026-04-10 12:07:53 - INFO - parrotllm.training - step 83570 | epoch 0 | loss 3.3024 | lr 4.62e-04 | grad 0.2773 +2026-04-10 12:07:56 - INFO - parrotllm.training - step 83580 | epoch 0 | loss 3.3867 | lr 4.62e-04 | grad 0.2862 +2026-04-10 12:07:59 - INFO - parrotllm.training - step 83590 | epoch 0 | loss 3.2549 | lr 4.62e-04 | grad 0.2872 +2026-04-10 12:08:02 - INFO - parrotllm.training - step 83600 | epoch 0 | loss 3.3715 | lr 4.62e-04 | grad 0.3076 +2026-04-10 12:08:06 - INFO - parrotllm.training - step 83610 | epoch 0 | loss 3.2224 | lr 4.62e-04 | grad 0.3072 +2026-04-10 12:08:09 - INFO - parrotllm.training - step 83620 | epoch 0 | loss 3.3170 | lr 4.61e-04 | grad 0.3006 +2026-04-10 12:08:12 - INFO - parrotllm.training - step 83630 | epoch 0 | loss 3.2619 | lr 4.61e-04 | grad 0.2906 +2026-04-10 12:08:15 - INFO - parrotllm.training - step 83640 | epoch 0 | loss 3.3678 | lr 4.61e-04 | grad 0.2866 +2026-04-10 12:08:18 - INFO - parrotllm.training - step 83650 | epoch 0 | loss 3.3798 | lr 4.61e-04 | grad 0.3055 +2026-04-10 12:08:21 - INFO - parrotllm.training - step 83660 | epoch 0 | loss 3.2695 | lr 4.61e-04 | grad 0.3020 +2026-04-10 12:08:24 - INFO - parrotllm.training - step 83670 | epoch 0 | loss 3.3067 | lr 4.61e-04 | grad 0.3120 +2026-04-10 12:08:27 - INFO - parrotllm.training - step 83680 | epoch 0 | loss 3.3356 | lr 4.60e-04 | grad 0.2930 +2026-04-10 12:08:30 - INFO - parrotllm.training - step 83690 | epoch 0 | loss 3.3487 | lr 4.60e-04 | grad 0.3049 +2026-04-10 12:08:33 - INFO - parrotllm.training - step 83700 | epoch 0 | loss 3.3473 | lr 4.60e-04 | grad 0.2928 +2026-04-10 12:08:36 - INFO - parrotllm.training - step 83710 | epoch 0 | loss 3.3819 | lr 4.60e-04 | grad 0.2876 +2026-04-10 12:08:39 - INFO - parrotllm.training - step 83720 | epoch 0 | loss 3.3652 | lr 4.60e-04 | grad 0.2950 +2026-04-10 12:08:42 - INFO - parrotllm.training - step 83730 | epoch 0 | loss 3.2595 | lr 4.60e-04 | grad 0.3081 +2026-04-10 12:08:45 - INFO - parrotllm.training - step 83740 | epoch 0 | loss 3.3865 | lr 4.60e-04 | grad 0.3090 +2026-04-10 12:08:48 - INFO - parrotllm.training - step 83750 | epoch 0 | loss 3.3848 | lr 4.59e-04 | grad 0.2877 +2026-04-10 12:08:51 - INFO - parrotllm.training - step 83760 | epoch 0 | loss 3.3466 | lr 4.59e-04 | grad 0.2800 +2026-04-10 12:08:54 - INFO - parrotllm.training - step 83770 | epoch 0 | loss 3.3260 | lr 4.59e-04 | grad 0.3085 +2026-04-10 12:08:57 - INFO - parrotllm.training - step 83780 | epoch 0 | loss 3.3640 | lr 4.59e-04 | grad 0.2988 +2026-04-10 12:09:00 - INFO - parrotllm.training - step 83790 | epoch 0 | loss 3.4973 | lr 4.59e-04 | grad 0.2924 +2026-04-10 12:09:04 - INFO - parrotllm.training - step 83800 | epoch 0 | loss 3.3227 | lr 4.59e-04 | grad 0.3044 +2026-04-10 12:09:07 - INFO - parrotllm.training - step 83810 | epoch 0 | loss 3.2905 | lr 4.59e-04 | grad 0.3021 +2026-04-10 12:09:10 - INFO - parrotllm.training - step 83820 | epoch 0 | loss 3.2534 | lr 4.58e-04 | grad 0.2675 +2026-04-10 12:09:13 - INFO - parrotllm.training - step 83830 | epoch 0 | loss 3.2895 | lr 4.58e-04 | grad 0.2905 +2026-04-10 12:09:16 - INFO - parrotllm.training - step 83840 | epoch 0 | loss 3.3594 | lr 4.58e-04 | grad 0.3082 +2026-04-10 12:09:19 - INFO - parrotllm.training - step 83850 | epoch 0 | loss 3.2173 | lr 4.58e-04 | grad 0.2887 +2026-04-10 12:09:22 - INFO - parrotllm.training - step 83860 | epoch 0 | loss 3.4347 | lr 4.58e-04 | grad 0.3140 +2026-04-10 12:09:25 - INFO - parrotllm.training - step 83870 | epoch 0 | loss 3.3886 | lr 4.58e-04 | grad 0.2770 +2026-04-10 12:09:28 - INFO - parrotllm.training - step 83880 | epoch 0 | loss 3.3926 | lr 4.58e-04 | grad 0.2791 +2026-04-10 12:09:31 - INFO - parrotllm.training - step 83890 | epoch 0 | loss 3.3764 | lr 4.57e-04 | grad 0.3031 +2026-04-10 12:09:34 - INFO - parrotllm.training - step 83900 | epoch 0 | loss 3.2748 | lr 4.57e-04 | grad 0.2792 +2026-04-10 12:09:37 - INFO - parrotllm.training - step 83910 | epoch 0 | loss 3.3056 | lr 4.57e-04 | grad 0.2918 +2026-04-10 12:09:40 - INFO - parrotllm.training - step 83920 | epoch 0 | loss 3.4004 | lr 4.57e-04 | grad 0.2792 +2026-04-10 12:09:43 - INFO - parrotllm.training - step 83930 | epoch 0 | loss 3.4270 | lr 4.57e-04 | grad 0.2877 +2026-04-10 12:09:46 - INFO - parrotllm.training - step 83940 | epoch 0 | loss 3.3666 | lr 4.57e-04 | grad 0.3476 +2026-04-10 12:09:49 - INFO - parrotllm.training - step 83950 | epoch 0 | loss 3.3562 | lr 4.56e-04 | grad 0.2862 +2026-04-10 12:09:52 - INFO - parrotllm.training - step 83960 | epoch 0 | loss 3.4667 | lr 4.56e-04 | grad 0.2879 +2026-04-10 12:09:55 - INFO - parrotllm.training - step 83970 | epoch 0 | loss 3.4160 | lr 4.56e-04 | grad 0.3265 +2026-04-10 12:09:59 - INFO - parrotllm.training - step 83980 | epoch 0 | loss 3.2403 | lr 4.56e-04 | grad 0.3252 +2026-04-10 12:10:02 - INFO - parrotllm.training - step 83990 | epoch 0 | loss 3.2705 | lr 4.56e-04 | grad 0.3054 +2026-04-10 12:10:05 - INFO - parrotllm.training - step 84000 | epoch 0 | loss 3.3329 | lr 4.56e-04 | grad 0.2975 +2026-04-10 12:10:05 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:10:05 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:10:08 - INFO - parrotllm.training - Train: loss=3.3329, ppl=28.02 +2026-04-10 12:10:08 - INFO - parrotllm.training - Val: loss=3.2880, ppl=26.79 +2026-04-10 12:10:08 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:10:09 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2880_epoch_0000_step_0084000.pt +2026-04-10 12:10:11 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:10:14 - INFO - parrotllm.training - step 84010 | epoch 0 | loss 3.3370 | lr 4.56e-04 | grad 0.2896 +2026-04-10 12:10:17 - INFO - parrotllm.training - step 84020 | epoch 0 | loss 3.3611 | lr 4.55e-04 | grad 0.3171 +2026-04-10 12:10:20 - INFO - parrotllm.training - step 84030 | epoch 0 | loss 3.3378 | lr 4.55e-04 | grad 0.2740 +2026-04-10 12:10:23 - INFO - parrotllm.training - step 84040 | epoch 0 | loss 3.3827 | lr 4.55e-04 | grad 0.2894 +2026-04-10 12:10:26 - INFO - parrotllm.training - step 84050 | epoch 0 | loss 3.3628 | lr 4.55e-04 | grad 0.3036 +2026-04-10 12:10:29 - INFO - parrotllm.training - step 84060 | epoch 0 | loss 3.2925 | lr 4.55e-04 | grad 0.3408 +2026-04-10 12:10:32 - INFO - parrotllm.training - step 84070 | epoch 0 | loss 3.4094 | lr 4.55e-04 | grad 0.3117 +2026-04-10 12:10:35 - INFO - parrotllm.training - step 84080 | epoch 0 | loss 3.4035 | lr 4.55e-04 | grad 0.2795 +2026-04-10 12:10:38 - INFO - parrotllm.training - step 84090 | epoch 0 | loss 3.3517 | lr 4.54e-04 | grad 0.2991 +2026-04-10 12:10:41 - INFO - parrotllm.training - step 84100 | epoch 0 | loss 3.3373 | lr 4.54e-04 | grad 0.2872 +2026-04-10 12:10:44 - INFO - parrotllm.training - step 84110 | epoch 0 | loss 3.3470 | lr 4.54e-04 | grad 0.2884 +2026-04-10 12:10:47 - INFO - parrotllm.training - step 84120 | epoch 0 | loss 3.3256 | lr 4.54e-04 | grad 0.3049 +2026-04-10 12:10:50 - INFO - parrotllm.training - step 84130 | epoch 0 | loss 3.3221 | lr 4.54e-04 | grad 0.2854 +2026-04-10 12:10:54 - INFO - parrotllm.training - step 84140 | epoch 0 | loss 3.2897 | lr 4.54e-04 | grad 0.3144 +2026-04-10 12:10:57 - INFO - parrotllm.training - step 84150 | epoch 0 | loss 3.2014 | lr 4.54e-04 | grad 0.2754 +2026-04-10 12:11:00 - INFO - parrotllm.training - step 84160 | epoch 0 | loss 3.1819 | lr 4.53e-04 | grad 0.3226 +2026-04-10 12:11:03 - INFO - parrotllm.training - step 84170 | epoch 0 | loss 3.3747 | lr 4.53e-04 | grad 0.2909 +2026-04-10 12:11:06 - INFO - parrotllm.training - step 84180 | epoch 0 | loss 3.3204 | lr 4.53e-04 | grad 0.2789 +2026-04-10 12:11:09 - INFO - parrotllm.training - step 84190 | epoch 0 | loss 3.3843 | lr 4.53e-04 | grad 0.2790 +2026-04-10 12:11:12 - INFO - parrotllm.training - step 84200 | epoch 0 | loss 3.4607 | lr 4.53e-04 | grad 0.2827 +2026-04-10 12:11:15 - INFO - parrotllm.training - step 84210 | epoch 0 | loss 3.3584 | lr 4.53e-04 | grad 0.2811 +2026-04-10 12:11:18 - INFO - parrotllm.training - step 84220 | epoch 0 | loss 3.3311 | lr 4.53e-04 | grad 0.2793 +2026-04-10 12:11:21 - INFO - parrotllm.training - step 84230 | epoch 0 | loss 3.3471 | lr 4.52e-04 | grad 0.2845 +2026-04-10 12:11:24 - INFO - parrotllm.training - step 84240 | epoch 0 | loss 3.3944 | lr 4.52e-04 | grad 0.3021 +2026-04-10 12:11:27 - INFO - parrotllm.training - step 84250 | epoch 0 | loss 3.3094 | lr 4.52e-04 | grad 0.3389 +2026-04-10 12:11:30 - INFO - parrotllm.training - step 84260 | epoch 0 | loss 3.4465 | lr 4.52e-04 | grad 0.3656 +2026-04-10 12:11:33 - INFO - parrotllm.training - step 84270 | epoch 0 | loss 3.3793 | lr 4.52e-04 | grad 0.2830 +2026-04-10 12:11:36 - INFO - parrotllm.training - step 84280 | epoch 0 | loss 3.3750 | lr 4.52e-04 | grad 0.2941 +2026-04-10 12:11:39 - INFO - parrotllm.training - step 84290 | epoch 0 | loss 3.4147 | lr 4.51e-04 | grad 0.2799 +2026-04-10 12:11:42 - INFO - parrotllm.training - step 84300 | epoch 0 | loss 3.4202 | lr 4.51e-04 | grad 0.2923 +2026-04-10 12:11:45 - INFO - parrotllm.training - step 84310 | epoch 0 | loss 3.3461 | lr 4.51e-04 | grad 0.2828 +2026-04-10 12:11:49 - INFO - parrotllm.training - step 84320 | epoch 0 | loss 3.3224 | lr 4.51e-04 | grad 0.3349 +2026-04-10 12:11:52 - INFO - parrotllm.training - step 84330 | epoch 0 | loss 3.3335 | lr 4.51e-04 | grad 0.3418 +2026-04-10 12:11:55 - INFO - parrotllm.training - step 84340 | epoch 0 | loss 3.3817 | lr 4.51e-04 | grad 0.2866 +2026-04-10 12:11:58 - INFO - parrotllm.training - step 84350 | epoch 0 | loss 3.4712 | lr 4.51e-04 | grad 0.3105 +2026-04-10 12:12:01 - INFO - parrotllm.training - step 84360 | epoch 0 | loss 3.2746 | lr 4.50e-04 | grad 0.2928 +2026-04-10 12:12:04 - INFO - parrotllm.training - step 84370 | epoch 0 | loss 3.3728 | lr 4.50e-04 | grad 0.3056 +2026-04-10 12:12:07 - INFO - parrotllm.training - step 84380 | epoch 0 | loss 3.3476 | lr 4.50e-04 | grad 0.2908 +2026-04-10 12:12:10 - INFO - parrotllm.training - step 84390 | epoch 0 | loss 3.3717 | lr 4.50e-04 | grad 0.3085 +2026-04-10 12:12:13 - INFO - parrotllm.training - step 84400 | epoch 0 | loss 3.2738 | lr 4.50e-04 | grad 0.3078 +2026-04-10 12:12:16 - INFO - parrotllm.training - step 84410 | epoch 0 | loss 3.3792 | lr 4.50e-04 | grad 0.2887 +2026-04-10 12:12:19 - INFO - parrotllm.training - step 84420 | epoch 0 | loss 3.3624 | lr 4.50e-04 | grad 0.2750 +2026-04-10 12:12:22 - INFO - parrotllm.training - step 84430 | epoch 0 | loss 3.4268 | lr 4.49e-04 | grad 0.2809 +2026-04-10 12:12:25 - INFO - parrotllm.training - step 84440 | epoch 0 | loss 3.2990 | lr 4.49e-04 | grad 0.3413 +2026-04-10 12:12:28 - INFO - parrotllm.training - step 84450 | epoch 0 | loss 3.3840 | lr 4.49e-04 | grad 0.3303 +2026-04-10 12:12:31 - INFO - parrotllm.training - step 84460 | epoch 0 | loss 3.3010 | lr 4.49e-04 | grad 0.2872 +2026-04-10 12:12:34 - INFO - parrotllm.training - step 84470 | epoch 0 | loss 3.3355 | lr 4.49e-04 | grad 0.2658 +2026-04-10 12:12:37 - INFO - parrotllm.training - step 84480 | epoch 0 | loss 3.4253 | lr 4.49e-04 | grad 0.2896 +2026-04-10 12:12:40 - INFO - parrotllm.training - step 84490 | epoch 0 | loss 3.3270 | lr 4.49e-04 | grad 0.3058 +2026-04-10 12:12:43 - INFO - parrotllm.training - step 84500 | epoch 0 | loss 3.3515 | lr 4.48e-04 | grad 0.2877 +2026-04-10 12:12:43 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:12:43 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:12:47 - INFO - parrotllm.training - Train: loss=3.3515, ppl=28.55 +2026-04-10 12:12:47 - INFO - parrotllm.training - Val: loss=3.2894, ppl=26.83 +2026-04-10 12:12:48 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2894_epoch_0000_step_0084500.pt +2026-04-10 12:12:49 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2880, min_delta=0.001000). +2026-04-10 12:12:49 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:12:53 - INFO - parrotllm.training - step 84510 | epoch 0 | loss 3.3441 | lr 4.48e-04 | grad 0.3109 +2026-04-10 12:12:56 - INFO - parrotllm.training - step 84520 | epoch 0 | loss 3.3174 | lr 4.48e-04 | grad 0.3334 +2026-04-10 12:12:59 - INFO - parrotllm.training - step 84530 | epoch 0 | loss 3.3280 | lr 4.48e-04 | grad 0.3023 +2026-04-10 12:13:02 - INFO - parrotllm.training - step 84540 | epoch 0 | loss 3.4368 | lr 4.48e-04 | grad 0.3024 +2026-04-10 12:13:05 - INFO - parrotllm.training - step 84550 | epoch 0 | loss 3.4074 | lr 4.48e-04 | grad 0.3049 +2026-04-10 12:13:08 - INFO - parrotllm.training - step 84560 | epoch 0 | loss 3.3713 | lr 4.48e-04 | grad 0.2914 +2026-04-10 12:13:11 - INFO - parrotllm.training - step 84570 | epoch 0 | loss 3.2442 | lr 4.47e-04 | grad 0.3158 +2026-04-10 12:13:14 - INFO - parrotllm.training - step 84580 | epoch 0 | loss 3.2780 | lr 4.47e-04 | grad 0.2994 +2026-04-10 12:13:17 - INFO - parrotllm.training - step 84590 | epoch 0 | loss 3.3234 | lr 4.47e-04 | grad 0.3037 +2026-04-10 12:13:20 - INFO - parrotllm.training - step 84600 | epoch 0 | loss 3.3775 | lr 4.47e-04 | grad 0.3146 +2026-04-10 12:13:23 - INFO - parrotllm.training - step 84610 | epoch 0 | loss 3.4034 | lr 4.47e-04 | grad 0.2893 +2026-04-10 12:13:26 - INFO - parrotllm.training - step 84620 | epoch 0 | loss 3.2738 | lr 4.47e-04 | grad 0.2758 +2026-04-10 12:13:29 - INFO - parrotllm.training - step 84630 | epoch 0 | loss 3.3413 | lr 4.47e-04 | grad 0.2970 +2026-04-10 12:13:32 - INFO - parrotllm.training - step 84640 | epoch 0 | loss 3.2610 | lr 4.46e-04 | grad 0.3318 +2026-04-10 12:13:35 - INFO - parrotllm.training - step 84650 | epoch 0 | loss 3.3271 | lr 4.46e-04 | grad 0.3178 +2026-04-10 12:13:38 - INFO - parrotllm.training - step 84660 | epoch 0 | loss 3.3435 | lr 4.46e-04 | grad 0.2995 +2026-04-10 12:13:41 - INFO - parrotllm.training - step 84670 | epoch 0 | loss 3.3304 | lr 4.46e-04 | grad 0.2825 +2026-04-10 12:13:44 - INFO - parrotllm.training - step 84680 | epoch 0 | loss 3.3620 | lr 4.46e-04 | grad 0.2989 +2026-04-10 12:13:48 - INFO - parrotllm.training - step 84690 | epoch 0 | loss 3.4226 | lr 4.46e-04 | grad 0.3017 +2026-04-10 12:13:51 - INFO - parrotllm.training - step 84700 | epoch 0 | loss 3.3214 | lr 4.45e-04 | grad 0.3043 +2026-04-10 12:13:54 - INFO - parrotllm.training - step 84710 | epoch 0 | loss 3.3171 | lr 4.45e-04 | grad 0.2947 +2026-04-10 12:13:57 - INFO - parrotllm.training - step 84720 | epoch 0 | loss 3.4485 | lr 4.45e-04 | grad 0.3226 +2026-04-10 12:14:00 - INFO - parrotllm.training - step 84730 | epoch 0 | loss 3.3509 | lr 4.45e-04 | grad 0.2922 +2026-04-10 12:14:03 - INFO - parrotllm.training - step 84740 | epoch 0 | loss 3.3672 | lr 4.45e-04 | grad 0.3124 +2026-04-10 12:14:06 - INFO - parrotllm.training - step 84750 | epoch 0 | loss 3.3108 | lr 4.45e-04 | grad 0.3101 +2026-04-10 12:14:09 - INFO - parrotllm.training - step 84760 | epoch 0 | loss 3.3118 | lr 4.45e-04 | grad 0.3167 +2026-04-10 12:14:12 - INFO - parrotllm.training - step 84770 | epoch 0 | loss 3.2984 | lr 4.44e-04 | grad 0.2834 +2026-04-10 12:14:15 - INFO - parrotllm.training - step 84780 | epoch 0 | loss 3.3501 | lr 4.44e-04 | grad 0.2860 +2026-04-10 12:14:18 - INFO - parrotllm.training - step 84790 | epoch 0 | loss 3.3087 | lr 4.44e-04 | grad 0.3055 +2026-04-10 12:14:21 - INFO - parrotllm.training - step 84800 | epoch 0 | loss 3.3008 | lr 4.44e-04 | grad 0.3117 +2026-04-10 12:14:24 - INFO - parrotllm.training - step 84810 | epoch 0 | loss 3.3224 | lr 4.44e-04 | grad 0.2923 +2026-04-10 12:14:27 - INFO - parrotllm.training - step 84820 | epoch 0 | loss 3.4038 | lr 4.44e-04 | grad 0.3011 +2026-04-10 12:14:30 - INFO - parrotllm.training - step 84830 | epoch 0 | loss 3.2760 | lr 4.44e-04 | grad 0.3038 +2026-04-10 12:14:33 - INFO - parrotllm.training - step 84840 | epoch 0 | loss 3.3236 | lr 4.43e-04 | grad 0.2942 +2026-04-10 12:14:36 - INFO - parrotllm.training - step 84850 | epoch 0 | loss 3.2998 | lr 4.43e-04 | grad 0.2876 +2026-04-10 12:14:39 - INFO - parrotllm.training - step 84860 | epoch 0 | loss 3.4424 | lr 4.43e-04 | grad 0.2998 +2026-04-10 12:14:43 - INFO - parrotllm.training - step 84870 | epoch 0 | loss 3.3119 | lr 4.43e-04 | grad 0.2862 +2026-04-10 12:14:46 - INFO - parrotllm.training - step 84880 | epoch 0 | loss 3.3201 | lr 4.43e-04 | grad 0.2866 +2026-04-10 12:14:49 - INFO - parrotllm.training - step 84890 | epoch 0 | loss 3.2688 | lr 4.43e-04 | grad 0.2780 +2026-04-10 12:14:52 - INFO - parrotllm.training - step 84900 | epoch 0 | loss 3.3529 | lr 4.43e-04 | grad 0.2936 +2026-04-10 12:14:55 - INFO - parrotllm.training - step 84910 | epoch 0 | loss 3.3650 | lr 4.42e-04 | grad 0.3125 +2026-04-10 12:14:58 - INFO - parrotllm.training - step 84920 | epoch 0 | loss 3.3773 | lr 4.42e-04 | grad 0.2963 +2026-04-10 12:15:01 - INFO - parrotllm.training - step 84930 | epoch 0 | loss 3.3973 | lr 4.42e-04 | grad 0.2917 +2026-04-10 12:15:04 - INFO - parrotllm.training - step 84940 | epoch 0 | loss 3.3489 | lr 4.42e-04 | grad 0.2947 +2026-04-10 12:15:07 - INFO - parrotllm.training - step 84950 | epoch 0 | loss 3.4112 | lr 4.42e-04 | grad 0.2894 +2026-04-10 12:15:10 - INFO - parrotllm.training - step 84960 | epoch 0 | loss 3.3168 | lr 4.42e-04 | grad 0.2788 +2026-04-10 12:15:13 - INFO - parrotllm.training - step 84970 | epoch 0 | loss 3.3304 | lr 4.42e-04 | grad 0.2830 +2026-04-10 12:15:16 - INFO - parrotllm.training - step 84980 | epoch 0 | loss 3.3249 | lr 4.41e-04 | grad 0.2829 +2026-04-10 12:15:19 - INFO - parrotllm.training - step 84990 | epoch 0 | loss 3.3767 | lr 4.41e-04 | grad 0.3050 +2026-04-10 12:15:22 - INFO - parrotllm.training - step 85000 | epoch 0 | loss 3.4128 | lr 4.41e-04 | grad 0.3249 +2026-04-10 12:15:22 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:15:22 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:15:25 - INFO - parrotllm.training - Train: loss=3.4128, ppl=30.35 +2026-04-10 12:15:25 - INFO - parrotllm.training - Val: loss=3.2892, ppl=26.82 +2026-04-10 12:15:26 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2892_epoch_0000_step_0085000.pt +2026-04-10 12:15:28 - INFO - parrotllm.training - No validation improvement for 2/15 evaluation(s) (best=3.2880, min_delta=0.001000). +2026-04-10 12:15:28 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:15:29 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0085000.pt +2026-04-10 12:15:34 - INFO - parrotllm.training - step 85010 | epoch 0 | loss 3.3622 | lr 4.41e-04 | grad 0.3367 +2026-04-10 12:15:38 - INFO - parrotllm.training - step 85020 | epoch 0 | loss 3.3284 | lr 4.41e-04 | grad 0.2804 +2026-04-10 12:15:41 - INFO - parrotllm.training - step 85030 | epoch 0 | loss 3.4141 | lr 4.41e-04 | grad 0.2751 +2026-04-10 12:15:44 - INFO - parrotllm.training - step 85040 | epoch 0 | loss 3.4396 | lr 4.41e-04 | grad 0.3137 +2026-04-10 12:15:47 - INFO - parrotllm.training - step 85050 | epoch 0 | loss 3.3945 | lr 4.40e-04 | grad 0.2842 +2026-04-10 12:15:50 - INFO - parrotllm.training - step 85060 | epoch 0 | loss 3.2923 | lr 4.40e-04 | grad 0.3069 +2026-04-10 12:15:53 - INFO - parrotllm.training - step 85070 | epoch 0 | loss 3.3666 | lr 4.40e-04 | grad 0.2918 +2026-04-10 12:15:56 - INFO - parrotllm.training - step 85080 | epoch 0 | loss 3.3525 | lr 4.40e-04 | grad 0.3133 +2026-04-10 12:15:59 - INFO - parrotllm.training - step 85090 | epoch 0 | loss 3.3821 | lr 4.40e-04 | grad 0.3016 +2026-04-10 12:16:02 - INFO - parrotllm.training - step 85100 | epoch 0 | loss 3.2944 | lr 4.40e-04 | grad 0.3019 +2026-04-10 12:16:05 - INFO - parrotllm.training - step 85110 | epoch 0 | loss 3.4252 | lr 4.40e-04 | grad 0.3127 +2026-04-10 12:16:08 - INFO - parrotllm.training - step 85120 | epoch 0 | loss 3.4701 | lr 4.39e-04 | grad 0.2909 +2026-04-10 12:16:11 - INFO - parrotllm.training - step 85130 | epoch 0 | loss 3.3524 | lr 4.39e-04 | grad 0.2895 +2026-04-10 12:16:14 - INFO - parrotllm.training - step 85140 | epoch 0 | loss 3.3543 | lr 4.39e-04 | grad 0.2916 +2026-04-10 12:16:17 - INFO - parrotllm.training - step 85150 | epoch 0 | loss 3.3642 | lr 4.39e-04 | grad 0.3009 +2026-04-10 12:16:20 - INFO - parrotllm.training - step 85160 | epoch 0 | loss 3.4581 | lr 4.39e-04 | grad 0.3267 +2026-04-10 12:16:23 - INFO - parrotllm.training - step 85170 | epoch 0 | loss 3.3393 | lr 4.39e-04 | grad 0.2995 +2026-04-10 12:16:26 - INFO - parrotllm.training - step 85180 | epoch 0 | loss 3.3789 | lr 4.39e-04 | grad 0.2923 +2026-04-10 12:16:29 - INFO - parrotllm.training - step 85190 | epoch 0 | loss 3.2973 | lr 4.38e-04 | grad 0.2900 +2026-04-10 12:16:32 - INFO - parrotllm.training - step 85200 | epoch 0 | loss 3.3974 | lr 4.38e-04 | grad 0.2925 +2026-04-10 12:16:36 - INFO - parrotllm.training - step 85210 | epoch 0 | loss 3.2894 | lr 4.38e-04 | grad 0.3008 +2026-04-10 12:16:39 - INFO - parrotllm.training - step 85220 | epoch 0 | loss 3.2821 | lr 4.38e-04 | grad 0.3004 +2026-04-10 12:16:42 - INFO - parrotllm.training - step 85230 | epoch 0 | loss 3.3906 | lr 4.38e-04 | grad 0.3202 +2026-04-10 12:16:45 - INFO - parrotllm.training - step 85240 | epoch 0 | loss 3.2617 | lr 4.38e-04 | grad 0.3223 +2026-04-10 12:16:48 - INFO - parrotllm.training - step 85250 | epoch 0 | loss 3.4099 | lr 4.38e-04 | grad 0.3063 +2026-04-10 12:16:51 - INFO - parrotllm.training - step 85260 | epoch 0 | loss 3.3828 | lr 4.37e-04 | grad 0.2933 +2026-04-10 12:16:54 - INFO - parrotllm.training - step 85270 | epoch 0 | loss 3.3231 | lr 4.37e-04 | grad 0.2913 +2026-04-10 12:16:57 - INFO - parrotllm.training - step 85280 | epoch 0 | loss 3.3976 | lr 4.37e-04 | grad 0.2829 +2026-04-10 12:17:00 - INFO - parrotllm.training - step 85290 | epoch 0 | loss 3.3629 | lr 4.37e-04 | grad 0.2922 +2026-04-10 12:17:03 - INFO - parrotllm.training - step 85300 | epoch 0 | loss 3.3608 | lr 4.37e-04 | grad 0.3202 +2026-04-10 12:17:06 - INFO - parrotllm.training - step 85310 | epoch 0 | loss 3.4114 | lr 4.37e-04 | grad 0.3272 +2026-04-10 12:17:09 - INFO - parrotllm.training - step 85320 | epoch 0 | loss 3.3765 | lr 4.37e-04 | grad 0.2648 +2026-04-10 12:17:12 - INFO - parrotllm.training - step 85330 | epoch 0 | loss 3.3714 | lr 4.36e-04 | grad 0.2994 +2026-04-10 12:17:15 - INFO - parrotllm.training - step 85340 | epoch 0 | loss 3.4074 | lr 4.36e-04 | grad 0.3069 +2026-04-10 12:17:18 - INFO - parrotllm.training - step 85350 | epoch 0 | loss 3.4767 | lr 4.36e-04 | grad 0.2962 +2026-04-10 12:17:21 - INFO - parrotllm.training - step 85360 | epoch 0 | loss 3.2500 | lr 4.36e-04 | grad 0.3059 +2026-04-10 12:17:24 - INFO - parrotllm.training - step 85370 | epoch 0 | loss 3.3310 | lr 4.36e-04 | grad 0.3209 +2026-04-10 12:17:27 - INFO - parrotllm.training - step 85380 | epoch 0 | loss 3.2875 | lr 4.36e-04 | grad 0.2973 +2026-04-10 12:17:30 - INFO - parrotllm.training - step 85390 | epoch 0 | loss 3.3580 | lr 4.36e-04 | grad 0.2926 +2026-04-10 12:17:34 - INFO - parrotllm.training - step 85400 | epoch 0 | loss 3.2053 | lr 4.35e-04 | grad 0.2914 +2026-04-10 12:17:37 - INFO - parrotllm.training - step 85410 | epoch 0 | loss 3.3574 | lr 4.35e-04 | grad 0.3158 +2026-04-10 12:17:40 - INFO - parrotllm.training - step 85420 | epoch 0 | loss 3.3503 | lr 4.35e-04 | grad 0.3071 +2026-04-10 12:17:43 - INFO - parrotllm.training - step 85430 | epoch 0 | loss 3.4550 | lr 4.35e-04 | grad 0.2928 +2026-04-10 12:17:46 - INFO - parrotllm.training - step 85440 | epoch 0 | loss 3.3849 | lr 4.35e-04 | grad 0.2951 +2026-04-10 12:17:49 - INFO - parrotllm.training - step 85450 | epoch 0 | loss 3.2479 | lr 4.35e-04 | grad 0.3034 +2026-04-10 12:17:52 - INFO - parrotllm.training - step 85460 | epoch 0 | loss 3.3765 | lr 4.35e-04 | grad 0.3192 +2026-04-10 12:17:55 - INFO - parrotllm.training - step 85470 | epoch 0 | loss 3.3626 | lr 4.34e-04 | grad 0.3011 +2026-04-10 12:17:58 - INFO - parrotllm.training - step 85480 | epoch 0 | loss 3.3438 | lr 4.34e-04 | grad 0.3041 +2026-04-10 12:18:01 - INFO - parrotllm.training - step 85490 | epoch 0 | loss 3.3367 | lr 4.34e-04 | grad 0.2982 +2026-04-10 12:18:04 - INFO - parrotllm.training - step 85500 | epoch 0 | loss 3.3226 | lr 4.34e-04 | grad 0.3214 +2026-04-10 12:18:04 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:18:04 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:18:07 - INFO - parrotllm.training - Train: loss=3.3226, ppl=27.73 +2026-04-10 12:18:07 - INFO - parrotllm.training - Val: loss=3.2879, ppl=26.79 +2026-04-10 12:18:08 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2879_epoch_0000_step_0085500.pt +2026-04-10 12:18:10 - INFO - parrotllm.training - No validation improvement for 3/15 evaluation(s) (best=3.2880, min_delta=0.001000). +2026-04-10 12:18:10 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:18:13 - INFO - parrotllm.training - step 85510 | epoch 0 | loss 3.2557 | lr 4.34e-04 | grad 0.3288 +2026-04-10 12:18:16 - INFO - parrotllm.training - step 85520 | epoch 0 | loss 3.3659 | lr 4.34e-04 | grad 0.3108 +2026-04-10 12:18:19 - INFO - parrotllm.training - step 85530 | epoch 0 | loss 3.3687 | lr 4.34e-04 | grad 0.2874 +2026-04-10 12:18:22 - INFO - parrotllm.training - step 85540 | epoch 0 | loss 3.4435 | lr 4.33e-04 | grad 0.2947 +2026-04-10 12:18:25 - INFO - parrotllm.training - step 85550 | epoch 0 | loss 3.4084 | lr 4.33e-04 | grad 0.3044 +2026-04-10 12:18:28 - INFO - parrotllm.training - step 85560 | epoch 0 | loss 3.3366 | lr 4.33e-04 | grad 0.3089 +2026-04-10 12:18:31 - INFO - parrotllm.training - step 85570 | epoch 0 | loss 3.3018 | lr 4.33e-04 | grad 0.2965 +2026-04-10 12:18:34 - INFO - parrotllm.training - step 85580 | epoch 0 | loss 3.4228 | lr 4.33e-04 | grad 0.3006 +2026-04-10 12:18:37 - INFO - parrotllm.training - step 85590 | epoch 0 | loss 3.2736 | lr 4.33e-04 | grad 0.2899 +2026-04-10 12:18:41 - INFO - parrotllm.training - step 85600 | epoch 0 | loss 3.3306 | lr 4.33e-04 | grad 0.2939 +2026-04-10 12:18:44 - INFO - parrotllm.training - step 85610 | epoch 0 | loss 3.2618 | lr 4.32e-04 | grad 0.2987 +2026-04-10 12:18:47 - INFO - parrotllm.training - step 85620 | epoch 0 | loss 3.4181 | lr 4.32e-04 | grad 0.2810 +2026-04-10 12:18:50 - INFO - parrotllm.training - step 85630 | epoch 0 | loss 3.3787 | lr 4.32e-04 | grad 0.3440 +2026-04-10 12:18:53 - INFO - parrotllm.training - step 85640 | epoch 0 | loss 3.3118 | lr 4.32e-04 | grad 0.3049 +2026-04-10 12:18:56 - INFO - parrotllm.training - step 85650 | epoch 0 | loss 3.3787 | lr 4.32e-04 | grad 0.2882 +2026-04-10 12:18:59 - INFO - parrotllm.training - step 85660 | epoch 0 | loss 3.3561 | lr 4.32e-04 | grad 0.2831 +2026-04-10 12:19:02 - INFO - parrotllm.training - step 85670 | epoch 0 | loss 3.3127 | lr 4.32e-04 | grad 0.3239 +2026-04-10 12:19:05 - INFO - parrotllm.training - step 85680 | epoch 0 | loss 3.3065 | lr 4.31e-04 | grad 0.2929 +2026-04-10 12:19:08 - INFO - parrotllm.training - step 85690 | epoch 0 | loss 3.3129 | lr 4.31e-04 | grad 0.2941 +2026-04-10 12:19:11 - INFO - parrotllm.training - step 85700 | epoch 0 | loss 3.2852 | lr 4.31e-04 | grad 0.2791 +2026-04-10 12:19:14 - INFO - parrotllm.training - step 85710 | epoch 0 | loss 3.2568 | lr 4.31e-04 | grad 0.3104 +2026-04-10 12:19:17 - INFO - parrotllm.training - step 85720 | epoch 0 | loss 3.4242 | lr 4.31e-04 | grad 0.2823 +2026-04-10 12:19:20 - INFO - parrotllm.training - step 85730 | epoch 0 | loss 3.4008 | lr 4.31e-04 | grad 0.2880 +2026-04-10 12:19:23 - INFO - parrotllm.training - step 85740 | epoch 0 | loss 3.4837 | lr 4.31e-04 | grad 0.3179 +2026-04-10 12:19:26 - INFO - parrotllm.training - step 85750 | epoch 0 | loss 3.2402 | lr 4.30e-04 | grad 0.3285 +2026-04-10 12:19:29 - INFO - parrotllm.training - step 85760 | epoch 0 | loss 3.3388 | lr 4.30e-04 | grad 0.2930 +2026-04-10 12:19:32 - INFO - parrotllm.training - step 85770 | epoch 0 | loss 3.3305 | lr 4.30e-04 | grad 0.3065 +2026-04-10 12:19:35 - INFO - parrotllm.training - step 85780 | epoch 0 | loss 3.3234 | lr 4.30e-04 | grad 0.2825 +2026-04-10 12:19:38 - INFO - parrotllm.training - step 85790 | epoch 0 | loss 3.3912 | lr 4.30e-04 | grad 0.3071 +2026-04-10 12:19:42 - INFO - parrotllm.training - step 85800 | epoch 0 | loss 3.4111 | lr 4.30e-04 | grad 0.3618 +2026-04-10 12:19:45 - INFO - parrotllm.training - step 85810 | epoch 0 | loss 3.3910 | lr 4.30e-04 | grad 0.3041 +2026-04-10 12:19:48 - INFO - parrotllm.training - step 85820 | epoch 0 | loss 3.4128 | lr 4.29e-04 | grad 0.2714 +2026-04-10 12:19:51 - INFO - parrotllm.training - step 85830 | epoch 0 | loss 3.2555 | lr 4.29e-04 | grad 0.2862 +2026-04-10 12:19:54 - INFO - parrotllm.training - step 85840 | epoch 0 | loss 3.3223 | lr 4.29e-04 | grad 0.3050 +2026-04-10 12:19:57 - INFO - parrotllm.training - step 85850 | epoch 0 | loss 3.3183 | lr 4.29e-04 | grad 0.3108 +2026-04-10 12:20:00 - INFO - parrotllm.training - step 85860 | epoch 0 | loss 3.3842 | lr 4.29e-04 | grad 0.2895 +2026-04-10 12:20:03 - INFO - parrotllm.training - step 85870 | epoch 0 | loss 3.3344 | lr 4.29e-04 | grad 0.2874 +2026-04-10 12:20:06 - INFO - parrotllm.training - step 85880 | epoch 0 | loss 3.3283 | lr 4.29e-04 | grad 0.2919 +2026-04-10 12:20:09 - INFO - parrotllm.training - step 85890 | epoch 0 | loss 3.4486 | lr 4.28e-04 | grad 0.2915 +2026-04-10 12:20:12 - INFO - parrotllm.training - step 85900 | epoch 0 | loss 3.2885 | lr 4.28e-04 | grad 0.2937 +2026-04-10 12:20:15 - INFO - parrotllm.training - step 85910 | epoch 0 | loss 3.3671 | lr 4.28e-04 | grad 0.3023 +2026-04-10 12:20:18 - INFO - parrotllm.training - step 85920 | epoch 0 | loss 3.3154 | lr 4.28e-04 | grad 0.3046 +2026-04-10 12:20:21 - INFO - parrotllm.training - step 85930 | epoch 0 | loss 3.2754 | lr 4.28e-04 | grad 0.2840 +2026-04-10 12:20:24 - INFO - parrotllm.training - step 85940 | epoch 0 | loss 3.3391 | lr 4.28e-04 | grad 0.2996 +2026-04-10 12:20:27 - INFO - parrotllm.training - step 85950 | epoch 0 | loss 3.4093 | lr 4.28e-04 | grad 0.3008 +2026-04-10 12:20:30 - INFO - parrotllm.training - step 85960 | epoch 0 | loss 3.4017 | lr 4.27e-04 | grad 0.2785 +2026-04-10 12:20:33 - INFO - parrotllm.training - step 85970 | epoch 0 | loss 3.3016 | lr 4.27e-04 | grad 0.2990 +2026-04-10 12:20:36 - INFO - parrotllm.training - step 85980 | epoch 0 | loss 3.4243 | lr 4.27e-04 | grad 0.3140 +2026-04-10 12:20:39 - INFO - parrotllm.training - step 85990 | epoch 0 | loss 3.2534 | lr 4.27e-04 | grad 0.2842 +2026-04-10 12:20:42 - INFO - parrotllm.training - step 86000 | epoch 0 | loss 3.2574 | lr 4.27e-04 | grad 0.3105 +2026-04-10 12:20:42 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:20:42 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:20:46 - INFO - parrotllm.training - Train: loss=3.2574, ppl=25.98 +2026-04-10 12:20:46 - INFO - parrotllm.training - Val: loss=3.2864, ppl=26.75 +2026-04-10 12:20:46 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:20:47 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2864_epoch_0000_step_0086000.pt +2026-04-10 12:20:48 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:20:51 - INFO - parrotllm.training - step 86010 | epoch 0 | loss 3.3571 | lr 4.27e-04 | grad 0.2790 +2026-04-10 12:20:54 - INFO - parrotllm.training - step 86020 | epoch 0 | loss 3.2754 | lr 4.27e-04 | grad 0.2974 +2026-04-10 12:20:58 - INFO - parrotllm.training - step 86030 | epoch 0 | loss 3.3864 | lr 4.26e-04 | grad 0.3082 +2026-04-10 12:21:01 - INFO - parrotllm.training - step 86040 | epoch 0 | loss 3.3718 | lr 4.26e-04 | grad 0.2874 +2026-04-10 12:21:04 - INFO - parrotllm.training - step 86050 | epoch 0 | loss 3.4036 | lr 4.26e-04 | grad 0.3012 +2026-04-10 12:21:07 - INFO - parrotllm.training - step 86060 | epoch 0 | loss 3.4159 | lr 4.26e-04 | grad 0.3053 +2026-04-10 12:21:10 - INFO - parrotllm.training - step 86070 | epoch 0 | loss 3.3369 | lr 4.26e-04 | grad 0.3001 +2026-04-10 12:21:13 - INFO - parrotllm.training - step 86080 | epoch 0 | loss 3.2230 | lr 4.26e-04 | grad 0.2743 +2026-04-10 12:21:16 - INFO - parrotllm.training - step 86090 | epoch 0 | loss 3.3022 | lr 4.26e-04 | grad 0.2860 +2026-04-10 12:21:19 - INFO - parrotllm.training - step 86100 | epoch 0 | loss 3.2459 | lr 4.25e-04 | grad 0.3117 +2026-04-10 12:21:22 - INFO - parrotllm.training - step 86110 | epoch 0 | loss 3.3831 | lr 4.25e-04 | grad 0.2932 +2026-04-10 12:21:25 - INFO - parrotllm.training - step 86120 | epoch 0 | loss 3.2941 | lr 4.25e-04 | grad 0.2753 +2026-04-10 12:21:28 - INFO - parrotllm.training - step 86130 | epoch 0 | loss 3.3249 | lr 4.25e-04 | grad 0.2892 +2026-04-10 12:21:31 - INFO - parrotllm.training - step 86140 | epoch 0 | loss 3.4420 | lr 4.25e-04 | grad 0.3122 +2026-04-10 12:21:34 - INFO - parrotllm.training - step 86150 | epoch 0 | loss 3.4263 | lr 4.25e-04 | grad 0.2727 +2026-04-10 12:21:37 - INFO - parrotllm.training - step 86160 | epoch 0 | loss 3.3262 | lr 4.25e-04 | grad 0.3054 +2026-04-10 12:21:40 - INFO - parrotllm.training - step 86170 | epoch 0 | loss 3.4716 | lr 4.24e-04 | grad 0.2846 +2026-04-10 12:21:43 - INFO - parrotllm.training - step 86180 | epoch 0 | loss 3.2861 | lr 4.24e-04 | grad 0.2930 +2026-04-10 12:21:46 - INFO - parrotllm.training - step 86190 | epoch 0 | loss 3.3876 | lr 4.24e-04 | grad 0.3004 +2026-04-10 12:21:49 - INFO - parrotllm.training - step 86200 | epoch 0 | loss 3.2492 | lr 4.24e-04 | grad 0.2973 +2026-04-10 12:21:52 - INFO - parrotllm.training - step 86210 | epoch 0 | loss 3.4038 | lr 4.24e-04 | grad 0.3124 +2026-04-10 12:21:55 - INFO - parrotllm.training - step 86220 | epoch 0 | loss 3.5090 | lr 4.24e-04 | grad 0.3099 +2026-04-10 12:21:58 - INFO - parrotllm.training - step 86230 | epoch 0 | loss 3.3409 | lr 4.24e-04 | grad 0.3361 +2026-04-10 12:22:02 - INFO - parrotllm.training - step 86240 | epoch 0 | loss 3.4101 | lr 4.23e-04 | grad 0.3375 +2026-04-10 12:22:05 - INFO - parrotllm.training - step 86250 | epoch 0 | loss 3.2860 | lr 4.23e-04 | grad 0.3269 +2026-04-10 12:22:08 - INFO - parrotllm.training - step 86260 | epoch 0 | loss 3.3381 | lr 4.23e-04 | grad 0.3087 +2026-04-10 12:22:11 - INFO - parrotllm.training - step 86270 | epoch 0 | loss 3.3759 | lr 4.23e-04 | grad 0.3010 +2026-04-10 12:22:14 - INFO - parrotllm.training - step 86280 | epoch 0 | loss 3.2441 | lr 4.23e-04 | grad 0.2980 +2026-04-10 12:22:17 - INFO - parrotllm.training - step 86290 | epoch 0 | loss 3.4259 | lr 4.23e-04 | grad 0.2953 +2026-04-10 12:22:20 - INFO - parrotllm.training - step 86300 | epoch 0 | loss 3.3960 | lr 4.23e-04 | grad 0.3080 +2026-04-10 12:22:23 - INFO - parrotllm.training - step 86310 | epoch 0 | loss 3.4587 | lr 4.22e-04 | grad 0.2936 +2026-04-10 12:22:26 - INFO - parrotllm.training - step 86320 | epoch 0 | loss 3.3406 | lr 4.22e-04 | grad 0.3127 +2026-04-10 12:22:29 - INFO - parrotllm.training - step 86330 | epoch 0 | loss 3.2423 | lr 4.22e-04 | grad 0.2985 +2026-04-10 12:22:32 - INFO - parrotllm.training - step 86340 | epoch 0 | loss 3.3938 | lr 4.22e-04 | grad 0.3048 +2026-04-10 12:22:35 - INFO - parrotllm.training - step 86350 | epoch 0 | loss 3.2793 | lr 4.22e-04 | grad 0.2797 +2026-04-10 12:22:38 - INFO - parrotllm.training - step 86360 | epoch 0 | loss 3.3807 | lr 4.22e-04 | grad 0.3015 +2026-04-10 12:22:41 - INFO - parrotllm.training - step 86370 | epoch 0 | loss 3.3165 | lr 4.22e-04 | grad 0.3341 +2026-04-10 12:22:44 - INFO - parrotllm.training - step 86380 | epoch 0 | loss 3.4630 | lr 4.21e-04 | grad 0.2963 +2026-04-10 12:22:47 - INFO - parrotllm.training - step 86390 | epoch 0 | loss 3.3932 | lr 4.21e-04 | grad 0.2816 +2026-04-10 12:22:50 - INFO - parrotllm.training - step 86400 | epoch 0 | loss 3.1942 | lr 4.21e-04 | grad 0.3001 +2026-04-10 12:22:53 - INFO - parrotllm.training - step 86410 | epoch 0 | loss 3.3853 | lr 4.21e-04 | grad 0.2980 +2026-04-10 12:22:56 - INFO - parrotllm.training - step 86420 | epoch 0 | loss 3.4136 | lr 4.21e-04 | grad 0.3214 +2026-04-10 12:23:00 - INFO - parrotllm.training - step 86430 | epoch 0 | loss 3.2943 | lr 4.21e-04 | grad 0.2880 +2026-04-10 12:23:03 - INFO - parrotllm.training - step 86440 | epoch 0 | loss 3.3656 | lr 4.21e-04 | grad 0.3043 +2026-04-10 12:23:06 - INFO - parrotllm.training - step 86450 | epoch 0 | loss 3.3312 | lr 4.20e-04 | grad 0.2885 +2026-04-10 12:23:09 - INFO - parrotllm.training - step 86460 | epoch 0 | loss 3.3778 | lr 4.20e-04 | grad 0.3106 +2026-04-10 12:23:12 - INFO - parrotllm.training - step 86470 | epoch 0 | loss 3.3472 | lr 4.20e-04 | grad 0.3017 +2026-04-10 12:23:15 - INFO - parrotllm.training - step 86480 | epoch 0 | loss 3.3924 | lr 4.20e-04 | grad 0.2912 +2026-04-10 12:23:18 - INFO - parrotllm.training - step 86490 | epoch 0 | loss 3.3136 | lr 4.20e-04 | grad 0.2862 +2026-04-10 12:23:21 - INFO - parrotllm.training - step 86500 | epoch 0 | loss 3.3039 | lr 4.20e-04 | grad 0.3293 +2026-04-10 12:23:21 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:23:21 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:23:24 - INFO - parrotllm.training - Train: loss=3.3039, ppl=27.22 +2026-04-10 12:23:24 - INFO - parrotllm.training - Val: loss=3.2853, ppl=26.72 +2026-04-10 12:23:24 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:23:25 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2853_epoch_0000_step_0086500.pt +2026-04-10 12:23:27 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:23:30 - INFO - parrotllm.training - step 86510 | epoch 0 | loss 3.2370 | lr 4.20e-04 | grad 0.2779 +2026-04-10 12:23:33 - INFO - parrotllm.training - step 86520 | epoch 0 | loss 3.2694 | lr 4.19e-04 | grad 0.2782 +2026-04-10 12:23:36 - INFO - parrotllm.training - step 86530 | epoch 0 | loss 3.4541 | lr 4.19e-04 | grad 0.3620 +2026-04-10 12:23:39 - INFO - parrotllm.training - step 86540 | epoch 0 | loss 3.3565 | lr 4.19e-04 | grad 0.2971 +2026-04-10 12:23:42 - INFO - parrotllm.training - step 86550 | epoch 0 | loss 3.3096 | lr 4.19e-04 | grad 0.3370 +2026-04-10 12:23:45 - INFO - parrotllm.training - step 86560 | epoch 0 | loss 3.3788 | lr 4.19e-04 | grad 0.2992 +2026-04-10 12:23:48 - INFO - parrotllm.training - step 86570 | epoch 0 | loss 3.3708 | lr 4.19e-04 | grad 0.2884 +2026-04-10 12:23:51 - INFO - parrotllm.training - step 86580 | epoch 0 | loss 3.3506 | lr 4.19e-04 | grad 0.2854 +2026-04-10 12:23:55 - INFO - parrotllm.training - step 86590 | epoch 0 | loss 3.3030 | lr 4.18e-04 | grad 0.2980 +2026-04-10 12:23:58 - INFO - parrotllm.training - step 86600 | epoch 0 | loss 3.3603 | lr 4.18e-04 | grad 0.2929 +2026-04-10 12:24:01 - INFO - parrotllm.training - step 86610 | epoch 0 | loss 3.3767 | lr 4.18e-04 | grad 0.2934 +2026-04-10 12:24:04 - INFO - parrotllm.training - step 86620 | epoch 0 | loss 3.4290 | lr 4.18e-04 | grad 0.3116 +2026-04-10 12:24:07 - INFO - parrotllm.training - step 86630 | epoch 0 | loss 3.3806 | lr 4.18e-04 | grad 0.3214 +2026-04-10 12:24:10 - INFO - parrotllm.training - step 86640 | epoch 0 | loss 3.3376 | lr 4.18e-04 | grad 0.2890 +2026-04-10 12:24:13 - INFO - parrotllm.training - step 86650 | epoch 0 | loss 3.3576 | lr 4.18e-04 | grad 0.2912 +2026-04-10 12:24:16 - INFO - parrotllm.training - step 86660 | epoch 0 | loss 3.4213 | lr 4.17e-04 | grad 0.2899 +2026-04-10 12:24:19 - INFO - parrotllm.training - step 86670 | epoch 0 | loss 3.3933 | lr 4.17e-04 | grad 0.2828 +2026-04-10 12:24:22 - INFO - parrotllm.training - step 86680 | epoch 0 | loss 3.2856 | lr 4.17e-04 | grad 0.3618 +2026-04-10 12:24:25 - INFO - parrotllm.training - step 86690 | epoch 0 | loss 3.3149 | lr 4.17e-04 | grad 0.2814 +2026-04-10 12:24:28 - INFO - parrotllm.training - step 86700 | epoch 0 | loss 3.3821 | lr 4.17e-04 | grad 0.3305 +2026-04-10 12:24:31 - INFO - parrotllm.training - step 86710 | epoch 0 | loss 3.4362 | lr 4.17e-04 | grad 0.3156 +2026-04-10 12:24:34 - INFO - parrotllm.training - step 86720 | epoch 0 | loss 3.4081 | lr 4.17e-04 | grad 0.2968 +2026-04-10 12:24:37 - INFO - parrotllm.training - step 86730 | epoch 0 | loss 3.3760 | lr 4.17e-04 | grad 0.3088 +2026-04-10 12:24:40 - INFO - parrotllm.training - step 86740 | epoch 0 | loss 3.3300 | lr 4.16e-04 | grad 0.2902 +2026-04-10 12:24:43 - INFO - parrotllm.training - step 86750 | epoch 0 | loss 3.2619 | lr 4.16e-04 | grad 0.3243 +2026-04-10 12:24:46 - INFO - parrotllm.training - step 86760 | epoch 0 | loss 3.3315 | lr 4.16e-04 | grad 0.3627 +2026-04-10 12:24:49 - INFO - parrotllm.training - step 86770 | epoch 0 | loss 3.3350 | lr 4.16e-04 | grad 0.3035 +2026-04-10 12:24:52 - INFO - parrotllm.training - step 86780 | epoch 0 | loss 3.3019 | lr 4.16e-04 | grad 0.3035 +2026-04-10 12:24:55 - INFO - parrotllm.training - step 86790 | epoch 0 | loss 3.3813 | lr 4.16e-04 | grad 0.3062 +2026-04-10 12:24:59 - INFO - parrotllm.training - step 86800 | epoch 0 | loss 3.3238 | lr 4.16e-04 | grad 0.2885 +2026-04-10 12:25:02 - INFO - parrotllm.training - step 86810 | epoch 0 | loss 3.3179 | lr 4.15e-04 | grad 0.3271 +2026-04-10 12:25:05 - INFO - parrotllm.training - step 86820 | epoch 0 | loss 3.2526 | lr 4.15e-04 | grad 0.2971 +2026-04-10 12:25:08 - INFO - parrotllm.training - step 86830 | epoch 0 | loss 3.3418 | lr 4.15e-04 | grad 0.2902 +2026-04-10 12:25:11 - INFO - parrotllm.training - step 86840 | epoch 0 | loss 3.3615 | lr 4.15e-04 | grad 0.2876 +2026-04-10 12:25:14 - INFO - parrotllm.training - step 86850 | epoch 0 | loss 3.3387 | lr 4.15e-04 | grad 0.3072 +2026-04-10 12:25:17 - INFO - parrotllm.training - step 86860 | epoch 0 | loss 3.3669 | lr 4.15e-04 | grad 0.3008 +2026-04-10 12:25:20 - INFO - parrotllm.training - step 86870 | epoch 0 | loss 3.3526 | lr 4.15e-04 | grad 0.2928 +2026-04-10 12:25:23 - INFO - parrotllm.training - step 86880 | epoch 0 | loss 3.3294 | lr 4.14e-04 | grad 0.3137 +2026-04-10 12:25:26 - INFO - parrotllm.training - step 86890 | epoch 0 | loss 3.3458 | lr 4.14e-04 | grad 0.2851 +2026-04-10 12:25:29 - INFO - parrotllm.training - step 86900 | epoch 0 | loss 3.2793 | lr 4.14e-04 | grad 0.3177 +2026-04-10 12:25:32 - INFO - parrotllm.training - step 86910 | epoch 0 | loss 3.3275 | lr 4.14e-04 | grad 0.2775 +2026-04-10 12:25:35 - INFO - parrotllm.training - step 86920 | epoch 0 | loss 3.2825 | lr 4.14e-04 | grad 0.2898 +2026-04-10 12:25:38 - INFO - parrotllm.training - step 86930 | epoch 0 | loss 3.3275 | lr 4.14e-04 | grad 0.3111 +2026-04-10 12:25:41 - INFO - parrotllm.training - step 86940 | epoch 0 | loss 3.3116 | lr 4.14e-04 | grad 0.2987 +2026-04-10 12:25:44 - INFO - parrotllm.training - step 86950 | epoch 0 | loss 3.3752 | lr 4.13e-04 | grad 0.3130 +2026-04-10 12:25:47 - INFO - parrotllm.training - step 86960 | epoch 0 | loss 3.4001 | lr 4.13e-04 | grad 0.3001 +2026-04-10 12:25:50 - INFO - parrotllm.training - step 86970 | epoch 0 | loss 3.3682 | lr 4.13e-04 | grad 0.3276 +2026-04-10 12:25:53 - INFO - parrotllm.training - step 86980 | epoch 0 | loss 3.2506 | lr 4.13e-04 | grad 0.2842 +2026-04-10 12:25:57 - INFO - parrotllm.training - step 86990 | epoch 0 | loss 3.2965 | lr 4.13e-04 | grad 0.2912 +2026-04-10 12:26:00 - INFO - parrotllm.training - step 87000 | epoch 0 | loss 3.3330 | lr 4.13e-04 | grad 0.2945 +2026-04-10 12:26:00 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:26:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:26:03 - INFO - parrotllm.training - Train: loss=3.3330, ppl=28.02 +2026-04-10 12:26:03 - INFO - parrotllm.training - Val: loss=3.2836, ppl=26.67 +2026-04-10 12:26:03 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:26:04 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2836_epoch_0000_step_0087000.pt +2026-04-10 12:26:06 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:26:09 - INFO - parrotllm.training - step 87010 | epoch 0 | loss 3.1963 | lr 4.13e-04 | grad 0.2956 +2026-04-10 12:26:12 - INFO - parrotllm.training - step 87020 | epoch 0 | loss 3.3324 | lr 4.12e-04 | grad 0.2786 +2026-04-10 12:26:15 - INFO - parrotllm.training - step 87030 | epoch 0 | loss 3.3891 | lr 4.12e-04 | grad 0.2910 +2026-04-10 12:26:18 - INFO - parrotllm.training - step 87040 | epoch 0 | loss 3.4712 | lr 4.12e-04 | grad 0.3062 +2026-04-10 12:26:21 - INFO - parrotllm.training - step 87050 | epoch 0 | loss 3.4676 | lr 4.12e-04 | grad 0.3205 +2026-04-10 12:26:24 - INFO - parrotllm.training - step 87060 | epoch 0 | loss 3.4782 | lr 4.12e-04 | grad 0.3038 +2026-04-10 12:26:27 - INFO - parrotllm.training - step 87070 | epoch 0 | loss 3.3405 | lr 4.12e-04 | grad 0.3172 +2026-04-10 12:26:30 - INFO - parrotllm.training - step 87080 | epoch 0 | loss 3.2717 | lr 4.12e-04 | grad 0.3335 +2026-04-10 12:26:33 - INFO - parrotllm.training - step 87090 | epoch 0 | loss 3.2728 | lr 4.11e-04 | grad 0.3008 +2026-04-10 12:26:36 - INFO - parrotllm.training - step 87100 | epoch 0 | loss 3.2687 | lr 4.11e-04 | grad 0.3032 +2026-04-10 12:26:39 - INFO - parrotllm.training - step 87110 | epoch 0 | loss 3.3596 | lr 4.11e-04 | grad 0.3203 +2026-04-10 12:26:42 - INFO - parrotllm.training - step 87120 | epoch 0 | loss 3.2643 | lr 4.11e-04 | grad 0.2804 +2026-04-10 12:26:45 - INFO - parrotllm.training - step 87130 | epoch 0 | loss 3.3609 | lr 4.11e-04 | grad 0.2862 +2026-04-10 12:26:48 - INFO - parrotllm.training - step 87140 | epoch 0 | loss 3.3271 | lr 4.11e-04 | grad 0.2867 +2026-04-10 12:26:51 - INFO - parrotllm.training - step 87150 | epoch 0 | loss 3.3432 | lr 4.11e-04 | grad 0.2912 +2026-04-10 12:26:54 - INFO - parrotllm.training - step 87160 | epoch 0 | loss 3.2727 | lr 4.11e-04 | grad 0.2853 +2026-04-10 12:26:57 - INFO - parrotllm.training - step 87170 | epoch 0 | loss 3.3510 | lr 4.10e-04 | grad 0.2937 +2026-04-10 12:27:01 - INFO - parrotllm.training - step 87180 | epoch 0 | loss 3.2578 | lr 4.10e-04 | grad 0.3038 +2026-04-10 12:27:04 - INFO - parrotllm.training - step 87190 | epoch 0 | loss 3.3053 | lr 4.10e-04 | grad 0.3072 +2026-04-10 12:27:07 - INFO - parrotllm.training - step 87200 | epoch 0 | loss 3.3941 | lr 4.10e-04 | grad 0.3108 +2026-04-10 12:27:10 - INFO - parrotllm.training - step 87210 | epoch 0 | loss 3.3255 | lr 4.10e-04 | grad 0.3340 +2026-04-10 12:27:13 - INFO - parrotllm.training - step 87220 | epoch 0 | loss 3.3273 | lr 4.10e-04 | grad 0.2985 +2026-04-10 12:27:16 - INFO - parrotllm.training - step 87230 | epoch 0 | loss 3.2493 | lr 4.10e-04 | grad 0.2855 +2026-04-10 12:27:19 - INFO - parrotllm.training - step 87240 | epoch 0 | loss 3.3396 | lr 4.09e-04 | grad 0.3059 +2026-04-10 12:27:22 - INFO - parrotllm.training - step 87250 | epoch 0 | loss 3.3344 | lr 4.09e-04 | grad 0.2968 +2026-04-10 12:27:25 - INFO - parrotllm.training - step 87260 | epoch 0 | loss 3.2937 | lr 4.09e-04 | grad 0.3036 +2026-04-10 12:27:28 - INFO - parrotllm.training - step 87270 | epoch 0 | loss 3.4439 | lr 4.09e-04 | grad 0.3022 +2026-04-10 12:27:31 - INFO - parrotllm.training - step 87280 | epoch 0 | loss 3.3775 | lr 4.09e-04 | grad 0.2959 +2026-04-10 12:27:34 - INFO - parrotllm.training - step 87290 | epoch 0 | loss 3.3110 | lr 4.09e-04 | grad 0.2875 +2026-04-10 12:27:37 - INFO - parrotllm.training - step 87300 | epoch 0 | loss 3.3809 | lr 4.09e-04 | grad 0.3008 +2026-04-10 12:27:40 - INFO - parrotllm.training - step 87310 | epoch 0 | loss 3.2374 | lr 4.08e-04 | grad 0.2827 +2026-04-10 12:27:43 - INFO - parrotllm.training - step 87320 | epoch 0 | loss 3.5091 | lr 4.08e-04 | grad 0.3051 +2026-04-10 12:27:46 - INFO - parrotllm.training - step 87330 | epoch 0 | loss 3.3019 | lr 4.08e-04 | grad 0.2884 +2026-04-10 12:27:49 - INFO - parrotllm.training - step 87340 | epoch 0 | loss 3.1733 | lr 4.08e-04 | grad 0.2920 +2026-04-10 12:27:52 - INFO - parrotllm.training - step 87350 | epoch 0 | loss 3.4193 | lr 4.08e-04 | grad 0.3019 +2026-04-10 12:27:55 - INFO - parrotllm.training - step 87360 | epoch 0 | loss 3.3734 | lr 4.08e-04 | grad 0.3210 +2026-04-10 12:27:59 - INFO - parrotllm.training - step 87370 | epoch 0 | loss 3.4252 | lr 4.08e-04 | grad 0.3493 +2026-04-10 12:28:02 - INFO - parrotllm.training - step 87380 | epoch 0 | loss 3.2051 | lr 4.07e-04 | grad 0.3389 +2026-04-10 12:28:05 - INFO - parrotllm.training - step 87390 | epoch 0 | loss 3.3722 | lr 4.07e-04 | grad 0.3239 +2026-04-10 12:28:08 - INFO - parrotllm.training - step 87400 | epoch 0 | loss 3.3026 | lr 4.07e-04 | grad 0.3070 +2026-04-10 12:28:11 - INFO - parrotllm.training - step 87410 | epoch 0 | loss 3.4262 | lr 4.07e-04 | grad 0.3115 +2026-04-10 12:28:14 - INFO - parrotllm.training - step 87420 | epoch 0 | loss 3.3796 | lr 4.07e-04 | grad 0.3017 +2026-04-10 12:28:17 - INFO - parrotllm.training - step 87430 | epoch 0 | loss 3.2894 | lr 4.07e-04 | grad 0.2935 +2026-04-10 12:28:20 - INFO - parrotllm.training - step 87440 | epoch 0 | loss 3.3550 | lr 4.07e-04 | grad 0.2846 +2026-04-10 12:28:23 - INFO - parrotllm.training - step 87450 | epoch 0 | loss 3.3527 | lr 4.07e-04 | grad 0.2823 +2026-04-10 12:28:26 - INFO - parrotllm.training - step 87460 | epoch 0 | loss 3.3901 | lr 4.06e-04 | grad 0.3377 +2026-04-10 12:28:29 - INFO - parrotllm.training - step 87470 | epoch 0 | loss 3.2996 | lr 4.06e-04 | grad 0.3091 +2026-04-10 12:28:32 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 87513/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 12:28:32 - INFO - parrotllm.training - step 87480 | epoch 0 | loss 3.2464 | lr 4.06e-04 | grad 0.3210 +2026-04-10 12:28:35 - INFO - parrotllm.training - step 87490 | epoch 0 | loss 3.4054 | lr 4.06e-04 | grad 0.3724 +2026-04-10 12:28:39 - INFO - parrotllm.training - step 87500 | epoch 0 | loss 3.3249 | lr 4.06e-04 | grad 0.2953 +2026-04-10 12:28:39 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:28:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:28:42 - INFO - parrotllm.training - Train: loss=3.3249, ppl=27.80 +2026-04-10 12:28:42 - INFO - parrotllm.training - Val: loss=3.2828, ppl=26.65 +2026-04-10 12:28:43 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2828_epoch_0000_step_0087500.pt +2026-04-10 12:28:45 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2836, min_delta=0.001000). +2026-04-10 12:28:45 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:28:46 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0087500.pt +2026-04-10 12:28:51 - INFO - parrotllm.training - step 87510 | epoch 0 | loss 3.2251 | lr 4.06e-04 | grad 0.3031 +2026-04-10 12:28:54 - INFO - parrotllm.training - step 87520 | epoch 0 | loss 3.2747 | lr 4.06e-04 | grad 0.3014 +2026-04-10 12:28:57 - INFO - parrotllm.training - step 87530 | epoch 0 | loss 3.3579 | lr 4.05e-04 | grad 0.3414 +2026-04-10 12:29:00 - INFO - parrotllm.training - step 87540 | epoch 0 | loss 3.3792 | lr 4.05e-04 | grad 0.3145 +2026-04-10 12:29:03 - INFO - parrotllm.training - step 87550 | epoch 0 | loss 3.3548 | lr 4.05e-04 | grad 0.3128 +2026-04-10 12:29:07 - INFO - parrotllm.training - step 87560 | epoch 0 | loss 3.3456 | lr 4.05e-04 | grad 0.3072 +2026-04-10 12:29:10 - INFO - parrotllm.training - step 87570 | epoch 0 | loss 3.4023 | lr 4.05e-04 | grad 0.3096 +2026-04-10 12:29:13 - INFO - parrotllm.training - step 87580 | epoch 0 | loss 3.2875 | lr 4.05e-04 | grad 0.2997 +2026-04-10 12:29:16 - INFO - parrotllm.training - step 87590 | epoch 0 | loss 3.2671 | lr 4.05e-04 | grad 0.2780 +2026-04-10 12:29:19 - INFO - parrotllm.training - step 87600 | epoch 0 | loss 3.2991 | lr 4.04e-04 | grad 0.3049 +2026-04-10 12:29:22 - INFO - parrotllm.training - step 87610 | epoch 0 | loss 3.3491 | lr 4.04e-04 | grad 0.3019 +2026-04-10 12:29:25 - INFO - parrotllm.training - step 87620 | epoch 0 | loss 3.2214 | lr 4.04e-04 | grad 0.2825 +2026-04-10 12:29:28 - INFO - parrotllm.training - step 87630 | epoch 0 | loss 3.3890 | lr 4.04e-04 | grad 0.3428 +2026-04-10 12:29:31 - INFO - parrotllm.training - step 87640 | epoch 0 | loss 3.3671 | lr 4.04e-04 | grad 0.3031 +2026-04-10 12:29:32 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 87677/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 12:29:34 - INFO - parrotllm.training - step 87650 | epoch 0 | loss 3.3436 | lr 4.04e-04 | grad 0.3138 +2026-04-10 12:29:37 - INFO - parrotllm.training - step 87660 | epoch 0 | loss 3.4012 | lr 4.04e-04 | grad 0.3033 +2026-04-10 12:29:40 - INFO - parrotllm.training - step 87670 | epoch 0 | loss 3.2800 | lr 4.03e-04 | grad 0.3075 +2026-04-10 12:29:43 - INFO - parrotllm.training - step 87680 | epoch 0 | loss 3.1905 | lr 4.03e-04 | grad 0.3182 +2026-04-10 12:29:46 - INFO - parrotllm.training - step 87690 | epoch 0 | loss 3.3809 | lr 4.03e-04 | grad 0.3399 +2026-04-10 12:29:50 - INFO - parrotllm.training - step 87700 | epoch 0 | loss 3.4535 | lr 4.03e-04 | grad 0.3169 +2026-04-10 12:29:53 - INFO - parrotllm.training - step 87710 | epoch 0 | loss 3.4409 | lr 4.03e-04 | grad 0.3123 +2026-04-10 12:29:56 - INFO - parrotllm.training - step 87720 | epoch 0 | loss 3.4185 | lr 4.03e-04 | grad 0.2973 +2026-04-10 12:29:59 - INFO - parrotllm.training - step 87730 | epoch 0 | loss 3.3998 | lr 4.03e-04 | grad 0.2885 +2026-04-10 12:30:02 - INFO - parrotllm.training - step 87740 | epoch 0 | loss 3.3036 | lr 4.03e-04 | grad 0.3428 +2026-04-10 12:30:05 - INFO - parrotllm.training - step 87750 | epoch 0 | loss 3.2482 | lr 4.02e-04 | grad 0.3217 +2026-04-10 12:30:08 - INFO - parrotllm.training - step 87760 | epoch 0 | loss 3.3515 | lr 4.02e-04 | grad 0.2775 +2026-04-10 12:30:11 - INFO - parrotllm.training - step 87770 | epoch 0 | loss 3.4487 | lr 4.02e-04 | grad 0.2917 +2026-04-10 12:30:14 - INFO - parrotllm.training - step 87780 | epoch 0 | loss 3.2985 | lr 4.02e-04 | grad 0.2884 +2026-04-10 12:30:17 - INFO - parrotllm.training - step 87790 | epoch 0 | loss 3.3591 | lr 4.02e-04 | grad 0.3137 +2026-04-10 12:30:20 - INFO - parrotllm.training - step 87800 | epoch 0 | loss 3.2929 | lr 4.02e-04 | grad 0.3058 +2026-04-10 12:30:23 - INFO - parrotllm.training - step 87810 | epoch 0 | loss 3.3280 | lr 4.02e-04 | grad 0.2946 +2026-04-10 12:30:26 - INFO - parrotllm.training - step 87820 | epoch 0 | loss 3.3615 | lr 4.01e-04 | grad 0.2892 +2026-04-10 12:30:29 - INFO - parrotllm.training - step 87830 | epoch 0 | loss 3.3599 | lr 4.01e-04 | grad 0.3117 +2026-04-10 12:30:32 - INFO - parrotllm.training - step 87840 | epoch 0 | loss 3.2578 | lr 4.01e-04 | grad 0.3080 +2026-04-10 12:30:35 - INFO - parrotllm.training - step 87850 | epoch 0 | loss 3.3204 | lr 4.01e-04 | grad 0.3122 +2026-04-10 12:30:38 - INFO - parrotllm.training - step 87860 | epoch 0 | loss 3.4222 | lr 4.01e-04 | grad 0.3439 +2026-04-10 12:30:41 - INFO - parrotllm.training - step 87870 | epoch 0 | loss 3.3221 | lr 4.01e-04 | grad 0.3538 +2026-04-10 12:30:44 - INFO - parrotllm.training - step 87880 | epoch 0 | loss 3.3464 | lr 4.01e-04 | grad 0.3120 +2026-04-10 12:30:47 - INFO - parrotllm.training - step 87890 | epoch 0 | loss 3.3918 | lr 4.00e-04 | grad 0.3109 +2026-04-10 12:30:51 - INFO - parrotllm.training - step 87900 | epoch 0 | loss 3.3173 | lr 4.00e-04 | grad 0.3025 +2026-04-10 12:30:54 - INFO - parrotllm.training - step 87910 | epoch 0 | loss 3.3808 | lr 4.00e-04 | grad 0.2973 +2026-04-10 12:30:57 - INFO - parrotllm.training - step 87920 | epoch 0 | loss 3.3331 | lr 4.00e-04 | grad 0.2933 +2026-04-10 12:31:00 - INFO - parrotllm.training - step 87930 | epoch 0 | loss 3.3125 | lr 4.00e-04 | grad 0.3124 +2026-04-10 12:31:03 - INFO - parrotllm.training - step 87940 | epoch 0 | loss 3.3926 | lr 4.00e-04 | grad 0.3174 +2026-04-10 12:31:06 - INFO - parrotllm.training - step 87950 | epoch 0 | loss 3.3920 | lr 4.00e-04 | grad 0.3639 +2026-04-10 12:31:09 - INFO - parrotllm.training - step 87960 | epoch 0 | loss 3.4173 | lr 4.00e-04 | grad 0.3191 +2026-04-10 12:31:12 - INFO - parrotllm.training - step 87970 | epoch 0 | loss 3.3250 | lr 3.99e-04 | grad 0.2970 +2026-04-10 12:31:15 - INFO - parrotllm.training - step 87980 | epoch 0 | loss 3.4290 | lr 3.99e-04 | grad 0.2872 +2026-04-10 12:31:18 - INFO - parrotllm.training - step 87990 | epoch 0 | loss 3.3197 | lr 3.99e-04 | grad 0.2933 +2026-04-10 12:31:21 - INFO - parrotllm.training - step 88000 | epoch 0 | loss 3.2700 | lr 3.99e-04 | grad 0.3033 +2026-04-10 12:31:21 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:31:21 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:31:24 - INFO - parrotllm.training - Train: loss=3.2700, ppl=26.31 +2026-04-10 12:31:24 - INFO - parrotllm.training - Val: loss=3.2813, ppl=26.61 +2026-04-10 12:31:24 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:31:25 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2813_epoch_0000_step_0088000.pt +2026-04-10 12:31:27 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:31:30 - INFO - parrotllm.training - step 88010 | epoch 0 | loss 3.3657 | lr 3.99e-04 | grad 0.3275 +2026-04-10 12:31:33 - INFO - parrotllm.training - step 88020 | epoch 0 | loss 3.2630 | lr 3.99e-04 | grad 0.3049 +2026-04-10 12:31:36 - INFO - parrotllm.training - step 88030 | epoch 0 | loss 3.2883 | lr 3.99e-04 | grad 0.3225 +2026-04-10 12:31:39 - INFO - parrotllm.training - step 88040 | epoch 0 | loss 3.3055 | lr 3.98e-04 | grad 0.3348 +2026-04-10 12:31:42 - INFO - parrotllm.training - step 88050 | epoch 0 | loss 3.4194 | lr 3.98e-04 | grad 0.2871 +2026-04-10 12:31:45 - INFO - parrotllm.training - step 88060 | epoch 0 | loss 3.3557 | lr 3.98e-04 | grad 0.3307 +2026-04-10 12:31:48 - INFO - parrotllm.training - step 88070 | epoch 0 | loss 3.2996 | lr 3.98e-04 | grad 0.3152 +2026-04-10 12:31:51 - INFO - parrotllm.training - step 88080 | epoch 0 | loss 3.4212 | lr 3.98e-04 | grad 0.3038 +2026-04-10 12:31:55 - INFO - parrotllm.training - step 88090 | epoch 0 | loss 3.3448 | lr 3.98e-04 | grad 0.3192 +2026-04-10 12:31:58 - INFO - parrotllm.training - step 88100 | epoch 0 | loss 3.4274 | lr 3.98e-04 | grad 0.2782 +2026-04-10 12:32:01 - INFO - parrotllm.training - step 88110 | epoch 0 | loss 3.3948 | lr 3.97e-04 | grad 0.3017 +2026-04-10 12:32:04 - INFO - parrotllm.training - step 88120 | epoch 0 | loss 3.4034 | lr 3.97e-04 | grad 0.3488 +2026-04-10 12:32:07 - INFO - parrotllm.training - step 88130 | epoch 0 | loss 3.3795 | lr 3.97e-04 | grad 0.3082 +2026-04-10 12:32:10 - INFO - parrotllm.training - step 88140 | epoch 0 | loss 3.2572 | lr 3.97e-04 | grad 0.2934 +2026-04-10 12:32:13 - INFO - parrotllm.training - step 88150 | epoch 0 | loss 3.4051 | lr 3.97e-04 | grad 0.2953 +2026-04-10 12:32:16 - INFO - parrotllm.training - step 88160 | epoch 0 | loss 3.3231 | lr 3.97e-04 | grad 0.2754 +2026-04-10 12:32:19 - INFO - parrotllm.training - step 88170 | epoch 0 | loss 3.3195 | lr 3.97e-04 | grad 0.2952 +2026-04-10 12:32:22 - INFO - parrotllm.training - step 88180 | epoch 0 | loss 3.3336 | lr 3.97e-04 | grad 0.3212 +2026-04-10 12:32:25 - INFO - parrotllm.training - step 88190 | epoch 0 | loss 3.3786 | lr 3.96e-04 | grad 0.3056 +2026-04-10 12:32:28 - INFO - parrotllm.training - step 88200 | epoch 0 | loss 3.3110 | lr 3.96e-04 | grad 0.3168 +2026-04-10 12:32:31 - INFO - parrotllm.training - step 88210 | epoch 0 | loss 3.2999 | lr 3.96e-04 | grad 0.2983 +2026-04-10 12:32:34 - INFO - parrotllm.training - step 88220 | epoch 0 | loss 3.3495 | lr 3.96e-04 | grad 0.3233 +2026-04-10 12:32:37 - INFO - parrotllm.training - step 88230 | epoch 0 | loss 3.2776 | lr 3.96e-04 | grad 0.2958 +2026-04-10 12:32:40 - INFO - parrotllm.training - step 88240 | epoch 0 | loss 3.4331 | lr 3.96e-04 | grad 0.3231 +2026-04-10 12:32:43 - INFO - parrotllm.training - step 88250 | epoch 0 | loss 3.3742 | lr 3.96e-04 | grad 0.3040 +2026-04-10 12:32:46 - INFO - parrotllm.training - step 88260 | epoch 0 | loss 3.3569 | lr 3.95e-04 | grad 0.3063 +2026-04-10 12:32:49 - INFO - parrotllm.training - step 88270 | epoch 0 | loss 3.3260 | lr 3.95e-04 | grad 0.3162 +2026-04-10 12:32:52 - INFO - parrotllm.training - step 88280 | epoch 0 | loss 3.4759 | lr 3.95e-04 | grad 0.2990 +2026-04-10 12:32:56 - INFO - parrotllm.training - step 88290 | epoch 0 | loss 3.3047 | lr 3.95e-04 | grad 0.2815 +2026-04-10 12:32:59 - INFO - parrotllm.training - step 88300 | epoch 0 | loss 3.2839 | lr 3.95e-04 | grad 0.2946 +2026-04-10 12:33:02 - INFO - parrotllm.training - step 88310 | epoch 0 | loss 3.3141 | lr 3.95e-04 | grad 0.3224 +2026-04-10 12:33:05 - INFO - parrotllm.training - step 88320 | epoch 0 | loss 3.2876 | lr 3.95e-04 | grad 0.2838 +2026-04-10 12:33:08 - INFO - parrotllm.training - step 88330 | epoch 0 | loss 3.3793 | lr 3.94e-04 | grad 0.3079 +2026-04-10 12:33:11 - INFO - parrotllm.training - step 88340 | epoch 0 | loss 3.4189 | lr 3.94e-04 | grad 0.2823 +2026-04-10 12:33:14 - INFO - parrotllm.training - step 88350 | epoch 0 | loss 3.3216 | lr 3.94e-04 | grad 0.3134 +2026-04-10 12:33:17 - INFO - parrotllm.training - step 88360 | epoch 0 | loss 3.2673 | lr 3.94e-04 | grad 0.3434 +2026-04-10 12:33:20 - INFO - parrotllm.training - step 88370 | epoch 0 | loss 3.5053 | lr 3.94e-04 | grad 0.3223 +2026-04-10 12:33:23 - INFO - parrotllm.training - step 88380 | epoch 0 | loss 3.4163 | lr 3.94e-04 | grad 0.2922 +2026-04-10 12:33:26 - INFO - parrotllm.training - step 88390 | epoch 0 | loss 3.3737 | lr 3.94e-04 | grad 0.3263 +2026-04-10 12:33:29 - INFO - parrotllm.training - step 88400 | epoch 0 | loss 3.2914 | lr 3.94e-04 | grad 0.3224 +2026-04-10 12:33:32 - INFO - parrotllm.training - step 88410 | epoch 0 | loss 3.3024 | lr 3.93e-04 | grad 0.3032 +2026-04-10 12:33:35 - INFO - parrotllm.training - step 88420 | epoch 0 | loss 3.3977 | lr 3.93e-04 | grad 0.2920 +2026-04-10 12:33:38 - INFO - parrotllm.training - step 88430 | epoch 0 | loss 3.4386 | lr 3.93e-04 | grad 0.2882 +2026-04-10 12:33:41 - INFO - parrotllm.training - step 88440 | epoch 0 | loss 3.4565 | lr 3.93e-04 | grad 0.3042 +2026-04-10 12:33:44 - INFO - parrotllm.training - step 88450 | epoch 0 | loss 3.3334 | lr 3.93e-04 | grad 0.3022 +2026-04-10 12:33:47 - INFO - parrotllm.training - step 88460 | epoch 0 | loss 3.3830 | lr 3.93e-04 | grad 0.3182 +2026-04-10 12:33:50 - INFO - parrotllm.training - step 88470 | epoch 0 | loss 3.3720 | lr 3.93e-04 | grad 0.2922 +2026-04-10 12:33:54 - INFO - parrotllm.training - step 88480 | epoch 0 | loss 3.3176 | lr 3.92e-04 | grad 0.2974 +2026-04-10 12:33:57 - INFO - parrotllm.training - step 88490 | epoch 0 | loss 3.2927 | lr 3.92e-04 | grad 0.3133 +2026-04-10 12:34:00 - INFO - parrotllm.training - step 88500 | epoch 0 | loss 3.2195 | lr 3.92e-04 | grad 0.3242 +2026-04-10 12:34:00 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:34:00 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:34:03 - INFO - parrotllm.training - Train: loss=3.2195, ppl=25.02 +2026-04-10 12:34:03 - INFO - parrotllm.training - Val: loss=3.2815, ppl=26.62 +2026-04-10 12:34:04 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2815_epoch_0000_step_0088500.pt +2026-04-10 12:34:06 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2813, min_delta=0.001000). +2026-04-10 12:34:06 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:34:09 - INFO - parrotllm.training - step 88510 | epoch 0 | loss 3.3756 | lr 3.92e-04 | grad 0.3134 +2026-04-10 12:34:12 - INFO - parrotllm.training - step 88520 | epoch 0 | loss 3.3419 | lr 3.92e-04 | grad 0.3078 +2026-04-10 12:34:15 - INFO - parrotllm.training - step 88530 | epoch 0 | loss 3.3986 | lr 3.92e-04 | grad 0.2861 +2026-04-10 12:34:18 - INFO - parrotllm.training - step 88540 | epoch 0 | loss 3.4016 | lr 3.92e-04 | grad 0.3003 +2026-04-10 12:34:21 - INFO - parrotllm.training - step 88550 | epoch 0 | loss 3.2876 | lr 3.92e-04 | grad 0.3142 +2026-04-10 12:34:24 - INFO - parrotllm.training - step 88560 | epoch 0 | loss 3.3295 | lr 3.91e-04 | grad 0.3082 +2026-04-10 12:34:27 - INFO - parrotllm.training - step 88570 | epoch 0 | loss 3.2450 | lr 3.91e-04 | grad 0.3068 +2026-04-10 12:34:30 - INFO - parrotllm.training - step 88580 | epoch 0 | loss 3.2277 | lr 3.91e-04 | grad 0.2868 +2026-04-10 12:34:33 - INFO - parrotllm.training - step 88590 | epoch 0 | loss 3.3304 | lr 3.91e-04 | grad 0.2914 +2026-04-10 12:34:36 - INFO - parrotllm.training - step 88600 | epoch 0 | loss 3.3506 | lr 3.91e-04 | grad 0.3033 +2026-04-10 12:34:39 - INFO - parrotllm.training - step 88610 | epoch 0 | loss 3.3648 | lr 3.91e-04 | grad 0.3022 +2026-04-10 12:34:43 - INFO - parrotllm.training - step 88620 | epoch 0 | loss 3.3097 | lr 3.91e-04 | grad 0.2904 +2026-04-10 12:34:46 - INFO - parrotllm.training - step 88630 | epoch 0 | loss 3.3153 | lr 3.90e-04 | grad 0.3030 +2026-04-10 12:34:49 - INFO - parrotllm.training - step 88640 | epoch 0 | loss 3.3341 | lr 3.90e-04 | grad 0.3153 +2026-04-10 12:34:52 - INFO - parrotllm.training - step 88650 | epoch 0 | loss 3.3528 | lr 3.90e-04 | grad 0.3109 +2026-04-10 12:34:55 - INFO - parrotllm.training - step 88660 | epoch 0 | loss 3.3110 | lr 3.90e-04 | grad 0.3080 +2026-04-10 12:34:58 - INFO - parrotllm.training - step 88670 | epoch 0 | loss 3.3524 | lr 3.90e-04 | grad 0.3054 +2026-04-10 12:35:01 - INFO - parrotllm.training - step 88680 | epoch 0 | loss 3.3972 | lr 3.90e-04 | grad 0.3553 +2026-04-10 12:35:04 - INFO - parrotllm.training - step 88690 | epoch 0 | loss 3.3506 | lr 3.90e-04 | grad 0.3193 +2026-04-10 12:35:07 - INFO - parrotllm.training - step 88700 | epoch 0 | loss 3.4573 | lr 3.89e-04 | grad 0.2974 +2026-04-10 12:35:10 - INFO - parrotllm.training - step 88710 | epoch 0 | loss 3.3390 | lr 3.89e-04 | grad 0.3046 +2026-04-10 12:35:13 - INFO - parrotllm.training - step 88720 | epoch 0 | loss 3.3899 | lr 3.89e-04 | grad 0.3107 +2026-04-10 12:35:16 - INFO - parrotllm.training - step 88730 | epoch 0 | loss 3.2821 | lr 3.89e-04 | grad 0.3065 +2026-04-10 12:35:19 - INFO - parrotllm.training - step 88740 | epoch 0 | loss 3.3863 | lr 3.89e-04 | grad 0.3048 +2026-04-10 12:35:22 - INFO - parrotllm.training - step 88750 | epoch 0 | loss 3.3179 | lr 3.89e-04 | grad 0.3425 +2026-04-10 12:35:25 - INFO - parrotllm.training - step 88760 | epoch 0 | loss 3.3195 | lr 3.89e-04 | grad 0.3043 +2026-04-10 12:35:28 - INFO - parrotllm.training - step 88770 | epoch 0 | loss 3.3553 | lr 3.89e-04 | grad 0.3025 +2026-04-10 12:35:31 - INFO - parrotllm.training - step 88780 | epoch 0 | loss 3.3130 | lr 3.88e-04 | grad 0.3053 +2026-04-10 12:35:34 - INFO - parrotllm.training - step 88790 | epoch 0 | loss 3.2322 | lr 3.88e-04 | grad 0.2981 +2026-04-10 12:35:37 - INFO - parrotllm.training - step 88800 | epoch 0 | loss 3.3567 | lr 3.88e-04 | grad 0.2835 +2026-04-10 12:35:41 - INFO - parrotllm.training - step 88810 | epoch 0 | loss 3.3631 | lr 3.88e-04 | grad 0.3257 +2026-04-10 12:35:44 - INFO - parrotllm.training - step 88820 | epoch 0 | loss 3.3407 | lr 3.88e-04 | grad 0.2898 +2026-04-10 12:35:47 - INFO - parrotllm.training - step 88830 | epoch 0 | loss 3.3517 | lr 3.88e-04 | grad 0.3084 +2026-04-10 12:35:50 - INFO - parrotllm.training - step 88840 | epoch 0 | loss 3.3075 | lr 3.88e-04 | grad 0.3102 +2026-04-10 12:35:53 - INFO - parrotllm.training - step 88850 | epoch 0 | loss 3.3647 | lr 3.87e-04 | grad 0.3078 +2026-04-10 12:35:56 - INFO - parrotllm.training - step 88860 | epoch 0 | loss 3.3189 | lr 3.87e-04 | grad 0.2849 +2026-04-10 12:35:59 - INFO - parrotllm.training - step 88870 | epoch 0 | loss 3.3046 | lr 3.87e-04 | grad 0.3447 +2026-04-10 12:36:02 - INFO - parrotllm.training - step 88880 | epoch 0 | loss 3.2386 | lr 3.87e-04 | grad 0.3171 +2026-04-10 12:36:05 - INFO - parrotllm.training - step 88890 | epoch 0 | loss 3.3390 | lr 3.87e-04 | grad 0.3238 +2026-04-10 12:36:08 - INFO - parrotllm.training - step 88900 | epoch 0 | loss 3.3459 | lr 3.87e-04 | grad 0.3166 +2026-04-10 12:36:11 - INFO - parrotllm.training - step 88910 | epoch 0 | loss 3.2949 | lr 3.87e-04 | grad 0.2961 +2026-04-10 12:36:14 - INFO - parrotllm.training - step 88920 | epoch 0 | loss 3.3345 | lr 3.87e-04 | grad 0.3160 +2026-04-10 12:36:17 - INFO - parrotllm.training - step 88930 | epoch 0 | loss 3.2779 | lr 3.86e-04 | grad 0.2792 +2026-04-10 12:36:20 - INFO - parrotllm.training - step 88940 | epoch 0 | loss 3.3709 | lr 3.86e-04 | grad 0.3004 +2026-04-10 12:36:23 - INFO - parrotllm.training - step 88950 | epoch 0 | loss 3.3564 | lr 3.86e-04 | grad 0.3200 +2026-04-10 12:36:26 - INFO - parrotllm.training - step 88960 | epoch 0 | loss 3.2533 | lr 3.86e-04 | grad 0.3169 +2026-04-10 12:36:29 - INFO - parrotllm.training - step 88970 | epoch 0 | loss 3.2690 | lr 3.86e-04 | grad 0.3156 +2026-04-10 12:36:32 - INFO - parrotllm.training - step 88980 | epoch 0 | loss 3.2940 | lr 3.86e-04 | grad 0.2990 +2026-04-10 12:36:35 - INFO - parrotllm.training - step 88990 | epoch 0 | loss 3.3611 | lr 3.86e-04 | grad 0.3021 +2026-04-10 12:36:39 - INFO - parrotllm.training - step 89000 | epoch 0 | loss 3.3500 | lr 3.85e-04 | grad 0.3069 +2026-04-10 12:36:39 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:36:39 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:36:42 - INFO - parrotllm.training - Train: loss=3.3500, ppl=28.50 +2026-04-10 12:36:42 - INFO - parrotllm.training - Val: loss=3.2799, ppl=26.57 +2026-04-10 12:36:42 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:36:43 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2799_epoch_0000_step_0089000.pt +2026-04-10 12:36:45 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:36:48 - INFO - parrotllm.training - step 89010 | epoch 0 | loss 3.3150 | lr 3.85e-04 | grad 0.2979 +2026-04-10 12:36:51 - INFO - parrotllm.training - step 89020 | epoch 0 | loss 3.4198 | lr 3.85e-04 | grad 0.3368 +2026-04-10 12:36:54 - INFO - parrotllm.training - step 89030 | epoch 0 | loss 3.3113 | lr 3.85e-04 | grad 0.3090 +2026-04-10 12:36:57 - INFO - parrotllm.training - step 89040 | epoch 0 | loss 3.3054 | lr 3.85e-04 | grad 0.2914 +2026-04-10 12:37:00 - INFO - parrotllm.training - step 89050 | epoch 0 | loss 3.2955 | lr 3.85e-04 | grad 0.2956 +2026-04-10 12:37:03 - INFO - parrotllm.training - step 89060 | epoch 0 | loss 3.4070 | lr 3.85e-04 | grad 0.3330 +2026-04-10 12:37:06 - INFO - parrotllm.training - step 89070 | epoch 0 | loss 3.4073 | lr 3.85e-04 | grad 0.3102 +2026-04-10 12:37:09 - INFO - parrotllm.training - step 89080 | epoch 0 | loss 3.3000 | lr 3.84e-04 | grad 0.3123 +2026-04-10 12:37:12 - INFO - parrotllm.training - step 89090 | epoch 0 | loss 3.3445 | lr 3.84e-04 | grad 0.3362 +2026-04-10 12:37:15 - INFO - parrotllm.training - step 89100 | epoch 0 | loss 3.3436 | lr 3.84e-04 | grad 0.3060 +2026-04-10 12:37:18 - INFO - parrotllm.training - step 89110 | epoch 0 | loss 3.3956 | lr 3.84e-04 | grad 0.3052 +2026-04-10 12:37:21 - INFO - parrotllm.training - step 89120 | epoch 0 | loss 3.3593 | lr 3.84e-04 | grad 0.3016 +2026-04-10 12:37:24 - INFO - parrotllm.training - step 89130 | epoch 0 | loss 3.2615 | lr 3.84e-04 | grad 0.3141 +2026-04-10 12:37:27 - INFO - parrotllm.training - step 89140 | epoch 0 | loss 3.2362 | lr 3.84e-04 | grad 0.3177 +2026-04-10 12:37:30 - INFO - parrotllm.training - step 89150 | epoch 0 | loss 3.3669 | lr 3.83e-04 | grad 0.3250 +2026-04-10 12:37:33 - INFO - parrotllm.training - step 89160 | epoch 0 | loss 3.2898 | lr 3.83e-04 | grad 0.3274 +2026-04-10 12:37:37 - INFO - parrotllm.training - step 89170 | epoch 0 | loss 3.2998 | lr 3.83e-04 | grad 0.3084 +2026-04-10 12:37:40 - INFO - parrotllm.training - step 89180 | epoch 0 | loss 3.2306 | lr 3.83e-04 | grad 0.2894 +2026-04-10 12:37:43 - INFO - parrotllm.training - step 89190 | epoch 0 | loss 3.3394 | lr 3.83e-04 | grad 0.2854 +2026-04-10 12:37:46 - INFO - parrotllm.training - step 89200 | epoch 0 | loss 3.3769 | lr 3.83e-04 | grad 0.2955 +2026-04-10 12:37:49 - INFO - parrotllm.training - step 89210 | epoch 0 | loss 3.3662 | lr 3.83e-04 | grad 0.3260 +2026-04-10 12:37:52 - INFO - parrotllm.training - step 89220 | epoch 0 | loss 3.2194 | lr 3.83e-04 | grad 0.2970 +2026-04-10 12:37:55 - INFO - parrotllm.training - step 89230 | epoch 0 | loss 3.3490 | lr 3.82e-04 | grad 0.2915 +2026-04-10 12:37:58 - INFO - parrotllm.training - step 89240 | epoch 0 | loss 3.3222 | lr 3.82e-04 | grad 0.3016 +2026-04-10 12:38:01 - INFO - parrotllm.training - step 89250 | epoch 0 | loss 3.2665 | lr 3.82e-04 | grad 0.3045 +2026-04-10 12:38:04 - INFO - parrotllm.training - step 89260 | epoch 0 | loss 3.1861 | lr 3.82e-04 | grad 0.3302 +2026-04-10 12:38:07 - INFO - parrotllm.training - step 89270 | epoch 0 | loss 3.2663 | lr 3.82e-04 | grad 0.2979 +2026-04-10 12:38:10 - INFO - parrotllm.training - step 89280 | epoch 0 | loss 3.2570 | lr 3.82e-04 | grad 0.3038 +2026-04-10 12:38:13 - INFO - parrotllm.training - step 89290 | epoch 0 | loss 3.2529 | lr 3.82e-04 | grad 0.3247 +2026-04-10 12:38:16 - INFO - parrotllm.training - step 89300 | epoch 0 | loss 3.3745 | lr 3.81e-04 | grad 0.3142 +2026-04-10 12:38:19 - INFO - parrotllm.training - step 89310 | epoch 0 | loss 3.3640 | lr 3.81e-04 | grad 0.2959 +2026-04-10 12:38:22 - INFO - parrotllm.training - step 89320 | epoch 0 | loss 3.2979 | lr 3.81e-04 | grad 0.2953 +2026-04-10 12:38:25 - INFO - parrotllm.training - step 89330 | epoch 0 | loss 3.3217 | lr 3.81e-04 | grad 0.2930 +2026-04-10 12:38:28 - INFO - parrotllm.training - step 89340 | epoch 0 | loss 3.4120 | lr 3.81e-04 | grad 0.3006 +2026-04-10 12:38:31 - INFO - parrotllm.training - step 89350 | epoch 0 | loss 3.2872 | lr 3.81e-04 | grad 0.3023 +2026-04-10 12:38:35 - INFO - parrotllm.training - step 89360 | epoch 0 | loss 3.3213 | lr 3.81e-04 | grad 0.3359 +2026-04-10 12:38:38 - INFO - parrotllm.training - step 89370 | epoch 0 | loss 3.2886 | lr 3.81e-04 | grad 0.3216 +2026-04-10 12:38:41 - INFO - parrotllm.training - step 89380 | epoch 0 | loss 3.2427 | lr 3.80e-04 | grad 0.3070 +2026-04-10 12:38:44 - INFO - parrotllm.training - step 89390 | epoch 0 | loss 3.3136 | lr 3.80e-04 | grad 0.2987 +2026-04-10 12:38:47 - INFO - parrotllm.training - step 89400 | epoch 0 | loss 3.3617 | lr 3.80e-04 | grad 0.2980 +2026-04-10 12:38:50 - INFO - parrotllm.training - step 89410 | epoch 0 | loss 3.3389 | lr 3.80e-04 | grad 0.3007 +2026-04-10 12:38:53 - INFO - parrotllm.training - step 89420 | epoch 0 | loss 3.3652 | lr 3.80e-04 | grad 0.3024 +2026-04-10 12:38:56 - INFO - parrotllm.training - step 89430 | epoch 0 | loss 3.3639 | lr 3.80e-04 | grad 0.3310 +2026-04-10 12:38:59 - INFO - parrotllm.training - step 89440 | epoch 0 | loss 3.3064 | lr 3.80e-04 | grad 0.2972 +2026-04-10 12:39:02 - INFO - parrotllm.training - step 89450 | epoch 0 | loss 3.3771 | lr 3.79e-04 | grad 0.2970 +2026-04-10 12:39:05 - INFO - parrotllm.training - step 89460 | epoch 0 | loss 3.2965 | lr 3.79e-04 | grad 0.2958 +2026-04-10 12:39:08 - INFO - parrotllm.training - step 89470 | epoch 0 | loss 3.2912 | lr 3.79e-04 | grad 0.3342 +2026-04-10 12:39:11 - INFO - parrotllm.training - step 89480 | epoch 0 | loss 3.3060 | lr 3.79e-04 | grad 0.3040 +2026-04-10 12:39:14 - INFO - parrotllm.training - step 89490 | epoch 0 | loss 3.3659 | lr 3.79e-04 | grad 0.3402 +2026-04-10 12:39:17 - INFO - parrotllm.training - step 89500 | epoch 0 | loss 3.3480 | lr 3.79e-04 | grad 0.3111 +2026-04-10 12:39:17 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:39:17 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:39:20 - INFO - parrotllm.training - Train: loss=3.3480, ppl=28.44 +2026-04-10 12:39:20 - INFO - parrotllm.training - Val: loss=3.2789, ppl=26.55 +2026-04-10 12:39:20 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:39:21 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2789_epoch_0000_step_0089500.pt +2026-04-10 12:39:24 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:39:27 - INFO - parrotllm.training - step 89510 | epoch 0 | loss 3.3606 | lr 3.79e-04 | grad 0.3085 +2026-04-10 12:39:30 - INFO - parrotllm.training - step 89520 | epoch 0 | loss 3.2127 | lr 3.79e-04 | grad 0.3312 +2026-04-10 12:39:33 - INFO - parrotllm.training - step 89530 | epoch 0 | loss 3.3205 | lr 3.78e-04 | grad 0.3231 +2026-04-10 12:39:36 - INFO - parrotllm.training - step 89540 | epoch 0 | loss 3.3248 | lr 3.78e-04 | grad 0.2730 +2026-04-10 12:39:39 - INFO - parrotllm.training - step 89550 | epoch 0 | loss 3.3009 | lr 3.78e-04 | grad 0.3015 +2026-04-10 12:39:42 - INFO - parrotllm.training - step 89560 | epoch 0 | loss 3.3516 | lr 3.78e-04 | grad 0.3173 +2026-04-10 12:39:45 - INFO - parrotllm.training - step 89570 | epoch 0 | loss 3.3495 | lr 3.78e-04 | grad 0.3309 +2026-04-10 12:39:48 - INFO - parrotllm.training - step 89580 | epoch 0 | loss 3.2859 | lr 3.78e-04 | grad 0.3293 +2026-04-10 12:39:51 - INFO - parrotllm.training - step 89590 | epoch 0 | loss 3.3196 | lr 3.78e-04 | grad 0.3157 +2026-04-10 12:39:54 - INFO - parrotllm.training - step 89600 | epoch 0 | loss 3.4039 | lr 3.78e-04 | grad 0.3170 +2026-04-10 12:39:57 - INFO - parrotllm.training - step 89610 | epoch 0 | loss 3.2896 | lr 3.77e-04 | grad 0.3111 +2026-04-10 12:40:00 - INFO - parrotllm.training - step 89620 | epoch 0 | loss 3.2661 | lr 3.77e-04 | grad 0.2830 +2026-04-10 12:40:03 - INFO - parrotllm.training - step 89630 | epoch 0 | loss 3.3479 | lr 3.77e-04 | grad 0.3070 +2026-04-10 12:40:06 - INFO - parrotllm.training - step 89640 | epoch 0 | loss 3.3531 | lr 3.77e-04 | grad 0.3114 +2026-04-10 12:40:09 - INFO - parrotllm.training - step 89650 | epoch 0 | loss 3.2970 | lr 3.77e-04 | grad 0.3546 +2026-04-10 12:40:12 - INFO - parrotllm.training - step 89660 | epoch 0 | loss 3.3902 | lr 3.77e-04 | grad 0.3379 +2026-04-10 12:40:15 - INFO - parrotllm.training - step 89670 | epoch 0 | loss 3.3837 | lr 3.77e-04 | grad 0.3303 +2026-04-10 12:40:18 - INFO - parrotllm.training - step 89680 | epoch 0 | loss 3.4204 | lr 3.76e-04 | grad 0.3397 +2026-04-10 12:40:22 - INFO - parrotllm.training - step 89690 | epoch 0 | loss 3.3226 | lr 3.76e-04 | grad 0.3097 +2026-04-10 12:40:25 - INFO - parrotllm.training - step 89700 | epoch 0 | loss 3.3573 | lr 3.76e-04 | grad 0.3323 +2026-04-10 12:40:28 - INFO - parrotllm.training - step 89710 | epoch 0 | loss 3.4023 | lr 3.76e-04 | grad 0.3425 +2026-04-10 12:40:31 - INFO - parrotllm.training - step 89720 | epoch 0 | loss 3.3353 | lr 3.76e-04 | grad 0.3269 +2026-04-10 12:40:34 - INFO - parrotllm.training - step 89730 | epoch 0 | loss 3.4831 | lr 3.76e-04 | grad 0.3292 +2026-04-10 12:40:37 - INFO - parrotllm.training - step 89740 | epoch 0 | loss 3.3161 | lr 3.76e-04 | grad 0.3058 +2026-04-10 12:40:40 - INFO - parrotllm.training - step 89750 | epoch 0 | loss 3.3019 | lr 3.76e-04 | grad 0.3427 +2026-04-10 12:40:43 - INFO - parrotllm.training - step 89760 | epoch 0 | loss 3.3008 | lr 3.75e-04 | grad 0.3075 +2026-04-10 12:40:46 - INFO - parrotllm.training - step 89770 | epoch 0 | loss 3.4116 | lr 3.75e-04 | grad 0.3428 +2026-04-10 12:40:49 - INFO - parrotllm.training - step 89780 | epoch 0 | loss 3.3057 | lr 3.75e-04 | grad 0.2993 +2026-04-10 12:40:52 - INFO - parrotllm.training - step 89790 | epoch 0 | loss 3.3593 | lr 3.75e-04 | grad 0.3124 +2026-04-10 12:40:55 - INFO - parrotllm.training - step 89800 | epoch 0 | loss 3.3133 | lr 3.75e-04 | grad 0.3009 +2026-04-10 12:40:58 - INFO - parrotllm.training - step 89810 | epoch 0 | loss 3.3031 | lr 3.75e-04 | grad 0.2996 +2026-04-10 12:41:01 - INFO - parrotllm.training - step 89820 | epoch 0 | loss 3.3532 | lr 3.75e-04 | grad 0.3019 +2026-04-10 12:41:04 - INFO - parrotllm.training - step 89830 | epoch 0 | loss 3.3672 | lr 3.74e-04 | grad 0.2898 +2026-04-10 12:41:07 - INFO - parrotllm.training - step 89840 | epoch 0 | loss 3.3438 | lr 3.74e-04 | grad 0.2997 +2026-04-10 12:41:10 - INFO - parrotllm.training - step 89850 | epoch 0 | loss 3.2791 | lr 3.74e-04 | grad 0.3178 +2026-04-10 12:41:13 - INFO - parrotllm.training - step 89860 | epoch 0 | loss 3.3118 | lr 3.74e-04 | grad 0.2988 +2026-04-10 12:41:16 - INFO - parrotllm.training - step 89870 | epoch 0 | loss 3.3128 | lr 3.74e-04 | grad 0.3077 +2026-04-10 12:41:20 - INFO - parrotllm.training - step 89880 | epoch 0 | loss 3.3225 | lr 3.74e-04 | grad 0.3156 +2026-04-10 12:41:23 - INFO - parrotllm.training - step 89890 | epoch 0 | loss 3.3587 | lr 3.74e-04 | grad 0.3399 +2026-04-10 12:41:26 - INFO - parrotllm.training - step 89900 | epoch 0 | loss 3.2558 | lr 3.74e-04 | grad 0.3215 +2026-04-10 12:41:29 - INFO - parrotllm.training - step 89910 | epoch 0 | loss 3.2668 | lr 3.73e-04 | grad 0.3128 +2026-04-10 12:41:32 - INFO - parrotllm.training - step 89920 | epoch 0 | loss 3.3601 | lr 3.73e-04 | grad 0.3076 +2026-04-10 12:41:35 - INFO - parrotllm.training - step 89930 | epoch 0 | loss 3.2830 | lr 3.73e-04 | grad 0.3057 +2026-04-10 12:41:38 - INFO - parrotllm.training - step 89940 | epoch 0 | loss 3.3551 | lr 3.73e-04 | grad 0.3045 +2026-04-10 12:41:41 - INFO - parrotllm.training - step 89950 | epoch 0 | loss 3.2346 | lr 3.73e-04 | grad 0.3079 +2026-04-10 12:41:44 - INFO - parrotllm.training - step 89960 | epoch 0 | loss 3.4302 | lr 3.73e-04 | grad 0.3124 +2026-04-10 12:41:47 - INFO - parrotllm.training - step 89970 | epoch 0 | loss 3.3184 | lr 3.73e-04 | grad 0.3262 +2026-04-10 12:41:50 - INFO - parrotllm.training - step 89980 | epoch 0 | loss 3.3249 | lr 3.73e-04 | grad 0.2956 +2026-04-10 12:41:53 - INFO - parrotllm.training - step 89990 | epoch 0 | loss 3.3484 | lr 3.72e-04 | grad 0.3188 +2026-04-10 12:41:56 - INFO - parrotllm.training - step 90000 | epoch 0 | loss 3.3113 | lr 3.72e-04 | grad 0.3045 +2026-04-10 12:41:56 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:41:56 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:41:59 - INFO - parrotllm.training - Train: loss=3.3113, ppl=27.42 +2026-04-10 12:41:59 - INFO - parrotllm.training - Val: loss=3.2765, ppl=26.48 +2026-04-10 12:41:59 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:42:00 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2765_epoch_0000_step_0090000.pt +2026-04-10 12:42:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:42:03 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0090000.pt +2026-04-10 12:42:08 - INFO - parrotllm.training - step 90010 | epoch 0 | loss 3.2861 | lr 3.72e-04 | grad 0.3082 +2026-04-10 12:42:12 - INFO - parrotllm.training - step 90020 | epoch 0 | loss 3.2350 | lr 3.72e-04 | grad 0.3083 +2026-04-10 12:42:15 - INFO - parrotllm.training - step 90030 | epoch 0 | loss 3.3269 | lr 3.72e-04 | grad 0.3194 +2026-04-10 12:42:18 - INFO - parrotllm.training - step 90040 | epoch 0 | loss 3.2742 | lr 3.72e-04 | grad 0.3074 +2026-04-10 12:42:21 - INFO - parrotllm.training - step 90050 | epoch 0 | loss 3.3726 | lr 3.72e-04 | grad 0.3184 +2026-04-10 12:42:24 - INFO - parrotllm.training - step 90060 | epoch 0 | loss 3.2854 | lr 3.71e-04 | grad 0.3215 +2026-04-10 12:42:27 - INFO - parrotllm.training - step 90070 | epoch 0 | loss 3.4229 | lr 3.71e-04 | grad 0.3134 +2026-04-10 12:42:30 - INFO - parrotllm.training - step 90080 | epoch 0 | loss 3.2660 | lr 3.71e-04 | grad 0.3095 +2026-04-10 12:42:33 - INFO - parrotllm.training - step 90090 | epoch 0 | loss 3.4795 | lr 3.71e-04 | grad 0.3039 +2026-04-10 12:42:36 - INFO - parrotllm.training - step 90100 | epoch 0 | loss 3.2979 | lr 3.71e-04 | grad 0.2989 +2026-04-10 12:42:39 - INFO - parrotllm.training - step 90110 | epoch 0 | loss 3.3602 | lr 3.71e-04 | grad 0.3311 +2026-04-10 12:42:42 - INFO - parrotllm.training - step 90120 | epoch 0 | loss 3.2670 | lr 3.71e-04 | grad 0.3294 +2026-04-10 12:42:45 - INFO - parrotllm.training - step 90130 | epoch 0 | loss 3.3153 | lr 3.71e-04 | grad 0.3112 +2026-04-10 12:42:48 - INFO - parrotllm.training - step 90140 | epoch 0 | loss 3.3468 | lr 3.70e-04 | grad 0.2975 +2026-04-10 12:42:51 - INFO - parrotllm.training - step 90150 | epoch 0 | loss 3.3312 | lr 3.70e-04 | grad 0.3138 +2026-04-10 12:42:54 - INFO - parrotllm.training - step 90160 | epoch 0 | loss 3.2796 | lr 3.70e-04 | grad 0.3120 +2026-04-10 12:42:57 - INFO - parrotllm.training - step 90170 | epoch 0 | loss 3.3445 | lr 3.70e-04 | grad 0.2944 +2026-04-10 12:43:00 - INFO - parrotllm.training - step 90180 | epoch 0 | loss 3.4009 | lr 3.70e-04 | grad 0.3075 +2026-04-10 12:43:04 - INFO - parrotllm.training - step 90190 | epoch 0 | loss 3.3682 | lr 3.70e-04 | grad 0.2964 +2026-04-10 12:43:07 - INFO - parrotllm.training - step 90200 | epoch 0 | loss 3.3469 | lr 3.70e-04 | grad 0.3141 +2026-04-10 12:43:10 - INFO - parrotllm.training - step 90210 | epoch 0 | loss 3.3485 | lr 3.70e-04 | grad 0.2916 +2026-04-10 12:43:13 - INFO - parrotllm.training - step 90220 | epoch 0 | loss 3.3431 | lr 3.69e-04 | grad 0.3008 +2026-04-10 12:43:16 - INFO - parrotllm.training - step 90230 | epoch 0 | loss 3.3016 | lr 3.69e-04 | grad 0.2980 +2026-04-10 12:43:19 - INFO - parrotllm.training - step 90240 | epoch 0 | loss 3.4636 | lr 3.69e-04 | grad 0.2960 +2026-04-10 12:43:22 - INFO - parrotllm.training - step 90250 | epoch 0 | loss 3.3487 | lr 3.69e-04 | grad 0.3196 +2026-04-10 12:43:25 - INFO - parrotllm.training - step 90260 | epoch 0 | loss 3.3210 | lr 3.69e-04 | grad 0.3025 +2026-04-10 12:43:28 - INFO - parrotllm.training - step 90270 | epoch 0 | loss 3.4200 | lr 3.69e-04 | grad 0.3252 +2026-04-10 12:43:31 - INFO - parrotllm.training - step 90280 | epoch 0 | loss 3.3346 | lr 3.69e-04 | grad 0.3025 +2026-04-10 12:43:34 - INFO - parrotllm.training - step 90290 | epoch 0 | loss 3.3685 | lr 3.68e-04 | grad 0.3016 +2026-04-10 12:43:37 - INFO - parrotllm.training - step 90300 | epoch 0 | loss 3.3392 | lr 3.68e-04 | grad 0.2920 +2026-04-10 12:43:40 - INFO - parrotllm.training - step 90310 | epoch 0 | loss 3.2425 | lr 3.68e-04 | grad 0.3040 +2026-04-10 12:43:43 - INFO - parrotllm.training - step 90320 | epoch 0 | loss 3.2893 | lr 3.68e-04 | grad 0.3121 +2026-04-10 12:43:46 - INFO - parrotllm.training - step 90330 | epoch 0 | loss 3.3848 | lr 3.68e-04 | grad 0.2900 +2026-04-10 12:43:49 - INFO - parrotllm.training - step 90340 | epoch 0 | loss 3.3143 | lr 3.68e-04 | grad 0.2986 +2026-04-10 12:43:52 - INFO - parrotllm.training - step 90350 | epoch 0 | loss 3.4247 | lr 3.68e-04 | grad 0.3167 +2026-04-10 12:43:55 - INFO - parrotllm.training - step 90360 | epoch 0 | loss 3.3761 | lr 3.68e-04 | grad 0.3096 +2026-04-10 12:43:58 - INFO - parrotllm.training - step 90370 | epoch 0 | loss 3.3009 | lr 3.67e-04 | grad 0.3130 +2026-04-10 12:44:02 - INFO - parrotllm.training - step 90380 | epoch 0 | loss 3.4350 | lr 3.67e-04 | grad 0.2900 +2026-04-10 12:44:05 - INFO - parrotllm.training - step 90390 | epoch 0 | loss 3.3726 | lr 3.67e-04 | grad 0.2986 +2026-04-10 12:44:08 - INFO - parrotllm.training - step 90400 | epoch 0 | loss 3.2802 | lr 3.67e-04 | grad 0.3481 +2026-04-10 12:44:11 - INFO - parrotllm.training - step 90410 | epoch 0 | loss 3.3225 | lr 3.67e-04 | grad 0.3273 +2026-04-10 12:44:14 - INFO - parrotllm.training - step 90420 | epoch 0 | loss 3.3558 | lr 3.67e-04 | grad 0.3084 +2026-04-10 12:44:17 - INFO - parrotllm.training - step 90430 | epoch 0 | loss 3.2993 | lr 3.67e-04 | grad 0.2970 +2026-04-10 12:44:20 - INFO - parrotllm.training - step 90440 | epoch 0 | loss 3.3156 | lr 3.67e-04 | grad 0.3241 +2026-04-10 12:44:23 - INFO - parrotllm.training - step 90450 | epoch 0 | loss 3.3550 | lr 3.66e-04 | grad 0.3146 +2026-04-10 12:44:26 - INFO - parrotllm.training - step 90460 | epoch 0 | loss 3.4276 | lr 3.66e-04 | grad 0.3042 +2026-04-10 12:44:29 - INFO - parrotllm.training - step 90470 | epoch 0 | loss 3.3292 | lr 3.66e-04 | grad 0.2962 +2026-04-10 12:44:32 - INFO - parrotllm.training - step 90480 | epoch 0 | loss 3.2800 | lr 3.66e-04 | grad 0.2866 +2026-04-10 12:44:35 - INFO - parrotllm.training - step 90490 | epoch 0 | loss 3.2205 | lr 3.66e-04 | grad 0.3212 +2026-04-10 12:44:38 - INFO - parrotllm.training - step 90500 | epoch 0 | loss 3.4449 | lr 3.66e-04 | grad 0.2984 +2026-04-10 12:44:38 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:44:38 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:44:41 - INFO - parrotllm.training - Train: loss=3.4449, ppl=31.34 +2026-04-10 12:44:41 - INFO - parrotllm.training - Val: loss=3.2754, ppl=26.45 +2026-04-10 12:44:41 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:44:42 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2754_epoch_0000_step_0090500.pt +2026-04-10 12:44:44 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:44:47 - INFO - parrotllm.training - step 90510 | epoch 0 | loss 3.2919 | lr 3.66e-04 | grad 0.2932 +2026-04-10 12:44:50 - INFO - parrotllm.training - step 90520 | epoch 0 | loss 3.3860 | lr 3.66e-04 | grad 0.2946 +2026-04-10 12:44:53 - INFO - parrotllm.training - step 90530 | epoch 0 | loss 3.3770 | lr 3.65e-04 | grad 0.2881 +2026-04-10 12:44:56 - INFO - parrotllm.training - step 90540 | epoch 0 | loss 3.3444 | lr 3.65e-04 | grad 0.3057 +2026-04-10 12:44:59 - INFO - parrotllm.training - step 90550 | epoch 0 | loss 3.3358 | lr 3.65e-04 | grad 0.3080 +2026-04-10 12:45:02 - INFO - parrotllm.training - step 90560 | epoch 0 | loss 3.3444 | lr 3.65e-04 | grad 0.3110 +2026-04-10 12:45:05 - INFO - parrotllm.training - step 90570 | epoch 0 | loss 3.2778 | lr 3.65e-04 | grad 0.3108 +2026-04-10 12:45:09 - INFO - parrotllm.training - step 90580 | epoch 0 | loss 3.3448 | lr 3.65e-04 | grad 0.3250 +2026-04-10 12:45:12 - INFO - parrotllm.training - step 90590 | epoch 0 | loss 3.3228 | lr 3.65e-04 | grad 0.3737 +2026-04-10 12:45:15 - INFO - parrotllm.training - step 90600 | epoch 0 | loss 3.3280 | lr 3.64e-04 | grad 0.3038 +2026-04-10 12:45:18 - INFO - parrotllm.training - step 90610 | epoch 0 | loss 3.2835 | lr 3.64e-04 | grad 0.3209 +2026-04-10 12:45:21 - INFO - parrotllm.training - step 90620 | epoch 0 | loss 3.2720 | lr 3.64e-04 | grad 0.3062 +2026-04-10 12:45:24 - INFO - parrotllm.training - step 90630 | epoch 0 | loss 3.4091 | lr 3.64e-04 | grad 0.2945 +2026-04-10 12:45:27 - INFO - parrotllm.training - step 90640 | epoch 0 | loss 3.3949 | lr 3.64e-04 | grad 0.2884 +2026-04-10 12:45:30 - INFO - parrotllm.training - step 90650 | epoch 0 | loss 3.2945 | lr 3.64e-04 | grad 0.3264 +2026-04-10 12:45:33 - INFO - parrotllm.training - step 90660 | epoch 0 | loss 3.2882 | lr 3.64e-04 | grad 0.3331 +2026-04-10 12:45:36 - INFO - parrotllm.training - step 90670 | epoch 0 | loss 3.2798 | lr 3.64e-04 | grad 0.3421 +2026-04-10 12:45:39 - INFO - parrotllm.training - step 90680 | epoch 0 | loss 3.3478 | lr 3.63e-04 | grad 0.3111 +2026-04-10 12:45:42 - INFO - parrotllm.training - step 90690 | epoch 0 | loss 3.4046 | lr 3.63e-04 | grad 0.3189 +2026-04-10 12:45:45 - INFO - parrotllm.training - step 90700 | epoch 0 | loss 3.3371 | lr 3.63e-04 | grad 0.3200 +2026-04-10 12:45:48 - INFO - parrotllm.training - step 90710 | epoch 0 | loss 3.4367 | lr 3.63e-04 | grad 0.3194 +2026-04-10 12:45:51 - INFO - parrotllm.training - step 90720 | epoch 0 | loss 3.4251 | lr 3.63e-04 | grad 0.3061 +2026-04-10 12:45:54 - INFO - parrotllm.training - step 90730 | epoch 0 | loss 3.2838 | lr 3.63e-04 | grad 0.2906 +2026-04-10 12:45:57 - INFO - parrotllm.training - step 90740 | epoch 0 | loss 3.4164 | lr 3.63e-04 | grad 0.2897 +2026-04-10 12:46:00 - INFO - parrotllm.training - step 90750 | epoch 0 | loss 3.3249 | lr 3.63e-04 | grad 0.2849 +2026-04-10 12:46:03 - INFO - parrotllm.training - step 90760 | epoch 0 | loss 3.4081 | lr 3.62e-04 | grad 0.3339 +2026-04-10 12:46:06 - INFO - parrotllm.training - step 90770 | epoch 0 | loss 3.3756 | lr 3.62e-04 | grad 0.3502 +2026-04-10 12:46:10 - INFO - parrotllm.training - step 90780 | epoch 0 | loss 3.4454 | lr 3.62e-04 | grad 0.2966 +2026-04-10 12:46:13 - INFO - parrotllm.training - step 90790 | epoch 0 | loss 3.3049 | lr 3.62e-04 | grad 0.3261 +2026-04-10 12:46:16 - INFO - parrotllm.training - step 90800 | epoch 0 | loss 3.3240 | lr 3.62e-04 | grad 0.2916 +2026-04-10 12:46:19 - INFO - parrotllm.training - step 90810 | epoch 0 | loss 3.1935 | lr 3.62e-04 | grad 0.3075 +2026-04-10 12:46:22 - INFO - parrotllm.training - step 90820 | epoch 0 | loss 3.3451 | lr 3.62e-04 | grad 0.3155 +2026-04-10 12:46:25 - INFO - parrotllm.training - step 90830 | epoch 0 | loss 3.2993 | lr 3.62e-04 | grad 0.3030 +2026-04-10 12:46:28 - INFO - parrotllm.training - step 90840 | epoch 0 | loss 3.2646 | lr 3.61e-04 | grad 0.2944 +2026-04-10 12:46:31 - INFO - parrotllm.training - step 90850 | epoch 0 | loss 3.3518 | lr 3.61e-04 | grad 0.3034 +2026-04-10 12:46:33 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 90893/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 12:46:34 - INFO - parrotllm.training - step 90860 | epoch 0 | loss 3.2973 | lr 3.61e-04 | grad 0.3179 +2026-04-10 12:46:37 - INFO - parrotllm.training - step 90870 | epoch 0 | loss 3.3443 | lr 3.61e-04 | grad 0.3073 +2026-04-10 12:46:40 - INFO - parrotllm.training - step 90880 | epoch 0 | loss 3.3847 | lr 3.61e-04 | grad 0.3586 +2026-04-10 12:46:43 - INFO - parrotllm.training - step 90890 | epoch 0 | loss 3.3244 | lr 3.61e-04 | grad 0.3294 +2026-04-10 12:46:46 - INFO - parrotllm.training - step 90900 | epoch 0 | loss 3.4450 | lr 3.61e-04 | grad 0.3523 +2026-04-10 12:46:49 - INFO - parrotllm.training - step 90910 | epoch 0 | loss 3.3553 | lr 3.61e-04 | grad 0.3390 +2026-04-10 12:46:52 - INFO - parrotllm.training - step 90920 | epoch 0 | loss 3.3915 | lr 3.60e-04 | grad 0.3255 +2026-04-10 12:46:56 - INFO - parrotllm.training - step 90930 | epoch 0 | loss 3.2583 | lr 3.60e-04 | grad 0.3237 +2026-04-10 12:46:59 - INFO - parrotllm.training - step 90940 | epoch 0 | loss 3.2997 | lr 3.60e-04 | grad 0.3401 +2026-04-10 12:47:02 - INFO - parrotllm.training - step 90950 | epoch 0 | loss 3.3534 | lr 3.60e-04 | grad 0.3001 +2026-04-10 12:47:05 - INFO - parrotllm.training - step 90960 | epoch 0 | loss 3.3511 | lr 3.60e-04 | grad 0.2756 +2026-04-10 12:47:08 - INFO - parrotllm.training - step 90970 | epoch 0 | loss 3.3919 | lr 3.60e-04 | grad 0.3143 +2026-04-10 12:47:11 - INFO - parrotllm.training - step 90980 | epoch 0 | loss 3.3808 | lr 3.60e-04 | grad 0.3195 +2026-04-10 12:47:14 - INFO - parrotllm.training - step 90990 | epoch 0 | loss 3.3487 | lr 3.60e-04 | grad 0.3161 +2026-04-10 12:47:17 - INFO - parrotllm.training - step 91000 | epoch 0 | loss 3.3366 | lr 3.59e-04 | grad 0.3260 +2026-04-10 12:47:17 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:47:17 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:47:20 - INFO - parrotllm.training - Train: loss=3.3366, ppl=28.12 +2026-04-10 12:47:20 - INFO - parrotllm.training - Val: loss=3.2760, ppl=26.47 +2026-04-10 12:47:21 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2760_epoch_0000_step_0091000.pt +2026-04-10 12:47:23 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2754, min_delta=0.001000). +2026-04-10 12:47:23 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:47:26 - INFO - parrotllm.training - step 91010 | epoch 0 | loss 3.3254 | lr 3.59e-04 | grad 0.3014 +2026-04-10 12:47:29 - INFO - parrotllm.training - step 91020 | epoch 0 | loss 3.4184 | lr 3.59e-04 | grad 0.3131 +2026-04-10 12:47:32 - INFO - parrotllm.training - step 91030 | epoch 0 | loss 3.2404 | lr 3.59e-04 | grad 0.3153 +2026-04-10 12:47:35 - INFO - parrotllm.training - step 91040 | epoch 0 | loss 3.3325 | lr 3.59e-04 | grad 0.3181 +2026-04-10 12:47:38 - INFO - parrotllm.training - step 91050 | epoch 0 | loss 3.3550 | lr 3.59e-04 | grad 0.3466 +2026-04-10 12:47:41 - INFO - parrotllm.training - step 91060 | epoch 0 | loss 3.3138 | lr 3.59e-04 | grad 0.3087 +2026-04-10 12:47:44 - INFO - parrotllm.training - step 91070 | epoch 0 | loss 3.2954 | lr 3.58e-04 | grad 0.3415 +2026-04-10 12:47:47 - INFO - parrotllm.training - step 91080 | epoch 0 | loss 3.3424 | lr 3.58e-04 | grad 0.3666 +2026-04-10 12:47:50 - INFO - parrotllm.training - step 91090 | epoch 0 | loss 3.2082 | lr 3.58e-04 | grad 0.2979 +2026-04-10 12:47:53 - INFO - parrotllm.training - step 91100 | epoch 0 | loss 3.2618 | lr 3.58e-04 | grad 0.3273 +2026-04-10 12:47:56 - INFO - parrotllm.training - step 91110 | epoch 0 | loss 3.3024 | lr 3.58e-04 | grad 0.3054 +2026-04-10 12:47:59 - INFO - parrotllm.training - step 91120 | epoch 0 | loss 3.3274 | lr 3.58e-04 | grad 0.3506 +2026-04-10 12:48:02 - INFO - parrotllm.training - step 91130 | epoch 0 | loss 3.3456 | lr 3.58e-04 | grad 0.3050 +2026-04-10 12:48:05 - INFO - parrotllm.training - step 91140 | epoch 0 | loss 3.3225 | lr 3.58e-04 | grad 0.3493 +2026-04-10 12:48:09 - INFO - parrotllm.training - step 91150 | epoch 0 | loss 3.2878 | lr 3.57e-04 | grad 0.3115 +2026-04-10 12:48:12 - INFO - parrotllm.training - step 91160 | epoch 0 | loss 3.3275 | lr 3.57e-04 | grad 0.3160 +2026-04-10 12:48:15 - INFO - parrotllm.training - step 91170 | epoch 0 | loss 3.1856 | lr 3.57e-04 | grad 0.3179 +2026-04-10 12:48:18 - INFO - parrotllm.training - step 91180 | epoch 0 | loss 3.3533 | lr 3.57e-04 | grad 0.3189 +2026-04-10 12:48:21 - INFO - parrotllm.training - step 91190 | epoch 0 | loss 3.3833 | lr 3.57e-04 | grad 0.3274 +2026-04-10 12:48:24 - INFO - parrotllm.training - step 91200 | epoch 0 | loss 3.4566 | lr 3.57e-04 | grad 0.3231 +2026-04-10 12:48:27 - INFO - parrotllm.training - step 91210 | epoch 0 | loss 3.3167 | lr 3.57e-04 | grad 0.2995 +2026-04-10 12:48:30 - INFO - parrotllm.training - step 91220 | epoch 0 | loss 3.3972 | lr 3.57e-04 | grad 0.3009 +2026-04-10 12:48:33 - INFO - parrotllm.training - step 91230 | epoch 0 | loss 3.3446 | lr 3.56e-04 | grad 0.3002 +2026-04-10 12:48:36 - INFO - parrotllm.training - step 91240 | epoch 0 | loss 3.2450 | lr 3.56e-04 | grad 0.3168 +2026-04-10 12:48:39 - INFO - parrotllm.training - step 91250 | epoch 0 | loss 3.2548 | lr 3.56e-04 | grad 0.2980 +2026-04-10 12:48:42 - INFO - parrotllm.training - step 91260 | epoch 0 | loss 3.3534 | lr 3.56e-04 | grad 0.3246 +2026-04-10 12:48:45 - INFO - parrotllm.training - step 91270 | epoch 0 | loss 3.3880 | lr 3.56e-04 | grad 0.3323 +2026-04-10 12:48:48 - INFO - parrotllm.training - step 91280 | epoch 0 | loss 3.3477 | lr 3.56e-04 | grad 0.3165 +2026-04-10 12:48:51 - INFO - parrotllm.training - step 91290 | epoch 0 | loss 3.2669 | lr 3.56e-04 | grad 0.3077 +2026-04-10 12:48:54 - INFO - parrotllm.training - step 91300 | epoch 0 | loss 3.2263 | lr 3.56e-04 | grad 0.3105 +2026-04-10 12:48:57 - INFO - parrotllm.training - step 91310 | epoch 0 | loss 3.4009 | lr 3.55e-04 | grad 0.3264 +2026-04-10 12:49:00 - INFO - parrotllm.training - step 91320 | epoch 0 | loss 3.3175 | lr 3.55e-04 | grad 0.3244 +2026-04-10 12:49:03 - INFO - parrotllm.training - step 91330 | epoch 0 | loss 3.2747 | lr 3.55e-04 | grad 0.3164 +2026-04-10 12:49:06 - INFO - parrotllm.training - step 91340 | epoch 0 | loss 3.3506 | lr 3.55e-04 | grad 0.3170 +2026-04-10 12:49:10 - INFO - parrotllm.training - step 91350 | epoch 0 | loss 3.2809 | lr 3.55e-04 | grad 0.3344 +2026-04-10 12:49:13 - INFO - parrotllm.training - step 91360 | epoch 0 | loss 3.3208 | lr 3.55e-04 | grad 0.3048 +2026-04-10 12:49:16 - INFO - parrotllm.training - step 91370 | epoch 0 | loss 3.3267 | lr 3.55e-04 | grad 0.3180 +2026-04-10 12:49:19 - INFO - parrotllm.training - step 91380 | epoch 0 | loss 3.3775 | lr 3.55e-04 | grad 0.3263 +2026-04-10 12:49:22 - INFO - parrotllm.training - step 91390 | epoch 0 | loss 3.2442 | lr 3.54e-04 | grad 0.3194 +2026-04-10 12:49:25 - INFO - parrotllm.training - step 91400 | epoch 0 | loss 3.3569 | lr 3.54e-04 | grad 0.3283 +2026-04-10 12:49:28 - INFO - parrotllm.training - step 91410 | epoch 0 | loss 3.3525 | lr 3.54e-04 | grad 0.3048 +2026-04-10 12:49:31 - INFO - parrotllm.training - step 91420 | epoch 0 | loss 3.2219 | lr 3.54e-04 | grad 0.3274 +2026-04-10 12:49:34 - INFO - parrotllm.training - step 91430 | epoch 0 | loss 3.3556 | lr 3.54e-04 | grad 0.3187 +2026-04-10 12:49:37 - INFO - parrotllm.training - step 91440 | epoch 0 | loss 3.3439 | lr 3.54e-04 | grad 0.3292 +2026-04-10 12:49:40 - INFO - parrotllm.training - step 91450 | epoch 0 | loss 3.2516 | lr 3.54e-04 | grad 0.3006 +2026-04-10 12:49:43 - INFO - parrotllm.training - step 91460 | epoch 0 | loss 3.2994 | lr 3.54e-04 | grad 0.2924 +2026-04-10 12:49:46 - INFO - parrotllm.training - step 91470 | epoch 0 | loss 3.2959 | lr 3.53e-04 | grad 0.3212 +2026-04-10 12:49:49 - INFO - parrotllm.training - step 91480 | epoch 0 | loss 3.2876 | lr 3.53e-04 | grad 0.2997 +2026-04-10 12:49:52 - INFO - parrotllm.training - step 91490 | epoch 0 | loss 3.3683 | lr 3.53e-04 | grad 0.3155 +2026-04-10 12:49:55 - INFO - parrotllm.training - step 91500 | epoch 0 | loss 3.3760 | lr 3.53e-04 | grad 0.3319 +2026-04-10 12:49:55 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:49:55 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:49:58 - INFO - parrotllm.training - Train: loss=3.3760, ppl=29.25 +2026-04-10 12:49:58 - INFO - parrotllm.training - Val: loss=3.2739, ppl=26.41 +2026-04-10 12:49:58 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:49:59 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2739_epoch_0000_step_0091500.pt +2026-04-10 12:50:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:50:05 - INFO - parrotllm.training - step 91510 | epoch 0 | loss 3.3932 | lr 3.53e-04 | grad 0.3243 +2026-04-10 12:50:08 - INFO - parrotllm.training - step 91520 | epoch 0 | loss 3.2912 | lr 3.53e-04 | grad 0.3341 +2026-04-10 12:50:11 - INFO - parrotllm.training - step 91530 | epoch 0 | loss 3.3632 | lr 3.53e-04 | grad 0.3132 +2026-04-10 12:50:14 - INFO - parrotllm.training - step 91540 | epoch 0 | loss 3.3465 | lr 3.53e-04 | grad 0.3008 +2026-04-10 12:50:17 - INFO - parrotllm.training - step 91550 | epoch 0 | loss 3.2726 | lr 3.52e-04 | grad 0.3422 +2026-04-10 12:50:20 - INFO - parrotllm.training - step 91560 | epoch 0 | loss 3.3588 | lr 3.52e-04 | grad 0.3312 +2026-04-10 12:50:23 - INFO - parrotllm.training - step 91570 | epoch 0 | loss 3.2181 | lr 3.52e-04 | grad 0.2911 +2026-04-10 12:50:26 - INFO - parrotllm.training - step 91580 | epoch 0 | loss 3.2445 | lr 3.52e-04 | grad 0.3023 +2026-04-10 12:50:29 - INFO - parrotllm.training - step 91590 | epoch 0 | loss 3.2891 | lr 3.52e-04 | grad 0.3398 +2026-04-10 12:50:32 - INFO - parrotllm.training - step 91600 | epoch 0 | loss 3.4453 | lr 3.52e-04 | grad 0.3337 +2026-04-10 12:50:35 - INFO - parrotllm.training - step 91610 | epoch 0 | loss 3.2629 | lr 3.52e-04 | grad 0.3101 +2026-04-10 12:50:38 - INFO - parrotllm.training - step 91620 | epoch 0 | loss 3.3470 | lr 3.52e-04 | grad 0.3100 +2026-04-10 12:50:41 - INFO - parrotllm.training - step 91630 | epoch 0 | loss 3.4106 | lr 3.51e-04 | grad 0.3277 +2026-04-10 12:50:44 - INFO - parrotllm.training - step 91640 | epoch 0 | loss 3.2534 | lr 3.51e-04 | grad 0.3475 +2026-04-10 12:50:47 - INFO - parrotllm.training - step 91650 | epoch 0 | loss 3.3647 | lr 3.51e-04 | grad 0.3414 +2026-04-10 12:50:50 - INFO - parrotllm.training - step 91660 | epoch 0 | loss 3.3797 | lr 3.51e-04 | grad 0.3477 +2026-04-10 12:50:53 - INFO - parrotllm.training - step 91670 | epoch 0 | loss 3.3891 | lr 3.51e-04 | grad 0.3123 +2026-04-10 12:50:56 - INFO - parrotllm.training - step 91680 | epoch 0 | loss 3.3594 | lr 3.51e-04 | grad 0.3543 +2026-04-10 12:51:00 - INFO - parrotllm.training - step 91690 | epoch 0 | loss 3.4338 | lr 3.51e-04 | grad 0.3386 +2026-04-10 12:51:03 - INFO - parrotllm.training - step 91700 | epoch 0 | loss 3.4177 | lr 3.51e-04 | grad 0.2998 +2026-04-10 12:51:06 - INFO - parrotllm.training - step 91710 | epoch 0 | loss 3.3438 | lr 3.50e-04 | grad 0.3525 +2026-04-10 12:51:09 - INFO - parrotllm.training - step 91720 | epoch 0 | loss 3.3428 | lr 3.50e-04 | grad 0.2941 +2026-04-10 12:51:12 - INFO - parrotllm.training - step 91730 | epoch 0 | loss 3.4058 | lr 3.50e-04 | grad 0.3085 +2026-04-10 12:51:15 - INFO - parrotllm.training - step 91740 | epoch 0 | loss 3.2870 | lr 3.50e-04 | grad 0.3067 +2026-04-10 12:51:18 - INFO - parrotllm.training - step 91750 | epoch 0 | loss 3.3335 | lr 3.50e-04 | grad 0.3140 +2026-04-10 12:51:21 - INFO - parrotllm.training - step 91760 | epoch 0 | loss 3.3064 | lr 3.50e-04 | grad 0.3032 +2026-04-10 12:51:24 - INFO - parrotllm.training - step 91770 | epoch 0 | loss 3.3019 | lr 3.50e-04 | grad 0.3147 +2026-04-10 12:51:27 - INFO - parrotllm.training - step 91780 | epoch 0 | loss 3.3785 | lr 3.50e-04 | grad 0.2929 +2026-04-10 12:51:30 - INFO - parrotllm.training - step 91790 | epoch 0 | loss 3.3071 | lr 3.49e-04 | grad 0.3232 +2026-04-10 12:51:33 - INFO - parrotllm.training - step 91800 | epoch 0 | loss 3.2810 | lr 3.49e-04 | grad 0.3348 +2026-04-10 12:51:36 - INFO - parrotllm.training - step 91810 | epoch 0 | loss 3.4179 | lr 3.49e-04 | grad 0.3261 +2026-04-10 12:51:39 - INFO - parrotllm.training - step 91820 | epoch 0 | loss 3.3206 | lr 3.49e-04 | grad 0.3158 +2026-04-10 12:51:42 - INFO - parrotllm.training - step 91830 | epoch 0 | loss 3.3586 | lr 3.49e-04 | grad 0.3365 +2026-04-10 12:51:45 - INFO - parrotllm.training - step 91840 | epoch 0 | loss 3.3223 | lr 3.49e-04 | grad 0.3087 +2026-04-10 12:51:48 - INFO - parrotllm.training - step 91850 | epoch 0 | loss 3.2971 | lr 3.49e-04 | grad 0.3314 +2026-04-10 12:51:51 - INFO - parrotllm.training - step 91860 | epoch 0 | loss 3.2777 | lr 3.49e-04 | grad 0.2929 +2026-04-10 12:51:54 - INFO - parrotllm.training - step 91870 | epoch 0 | loss 3.3034 | lr 3.48e-04 | grad 0.3036 +2026-04-10 12:51:57 - INFO - parrotllm.training - step 91880 | epoch 0 | loss 3.3896 | lr 3.48e-04 | grad 0.3220 +2026-04-10 12:52:00 - INFO - parrotllm.training - step 91890 | epoch 0 | loss 3.3569 | lr 3.48e-04 | grad 0.3051 +2026-04-10 12:52:03 - INFO - parrotllm.training - step 91900 | epoch 0 | loss 3.3936 | lr 3.48e-04 | grad 0.2969 +2026-04-10 12:52:06 - INFO - parrotllm.training - step 91910 | epoch 0 | loss 3.2625 | lr 3.48e-04 | grad 0.3138 +2026-04-10 12:52:10 - INFO - parrotllm.training - step 91920 | epoch 0 | loss 3.3122 | lr 3.48e-04 | grad 0.3668 +2026-04-10 12:52:13 - INFO - parrotllm.training - step 91930 | epoch 0 | loss 3.2967 | lr 3.48e-04 | grad 0.3095 +2026-04-10 12:52:16 - INFO - parrotllm.training - step 91940 | epoch 0 | loss 3.3506 | lr 3.48e-04 | grad 0.3240 +2026-04-10 12:52:19 - INFO - parrotllm.training - step 91950 | epoch 0 | loss 3.3982 | lr 3.47e-04 | grad 0.3280 +2026-04-10 12:52:22 - INFO - parrotllm.training - step 91960 | epoch 0 | loss 3.3846 | lr 3.47e-04 | grad 0.3293 +2026-04-10 12:52:25 - INFO - parrotllm.training - step 91970 | epoch 0 | loss 3.2952 | lr 3.47e-04 | grad 0.2831 +2026-04-10 12:52:28 - INFO - parrotllm.training - step 91980 | epoch 0 | loss 3.2842 | lr 3.47e-04 | grad 0.3340 +2026-04-10 12:52:31 - INFO - parrotllm.training - step 91990 | epoch 0 | loss 3.3920 | lr 3.47e-04 | grad 0.3289 +2026-04-10 12:52:34 - INFO - parrotllm.training - step 92000 | epoch 0 | loss 3.3950 | lr 3.47e-04 | grad 0.3185 +2026-04-10 12:52:34 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:52:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:52:37 - INFO - parrotllm.training - Train: loss=3.3950, ppl=29.81 +2026-04-10 12:52:37 - INFO - parrotllm.training - Val: loss=3.2753, ppl=26.45 +2026-04-10 12:52:38 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2753_epoch_0000_step_0092000.pt +2026-04-10 12:52:40 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2739, min_delta=0.001000). +2026-04-10 12:52:40 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:52:43 - INFO - parrotllm.training - step 92010 | epoch 0 | loss 3.3546 | lr 3.47e-04 | grad 0.2920 +2026-04-10 12:52:46 - INFO - parrotllm.training - step 92020 | epoch 0 | loss 3.3707 | lr 3.47e-04 | grad 0.3247 +2026-04-10 12:52:49 - INFO - parrotllm.training - step 92030 | epoch 0 | loss 3.5513 | lr 3.46e-04 | grad 0.2877 +2026-04-10 12:52:52 - INFO - parrotllm.training - step 92040 | epoch 0 | loss 3.2788 | lr 3.46e-04 | grad 0.3271 +2026-04-10 12:52:55 - INFO - parrotllm.training - step 92050 | epoch 0 | loss 3.3843 | lr 3.46e-04 | grad 0.3074 +2026-04-10 12:52:58 - INFO - parrotllm.training - step 92060 | epoch 0 | loss 3.3858 | lr 3.46e-04 | grad 0.3237 +2026-04-10 12:53:02 - INFO - parrotllm.training - step 92070 | epoch 0 | loss 3.3840 | lr 3.46e-04 | grad 0.3038 +2026-04-10 12:53:05 - INFO - parrotllm.training - step 92080 | epoch 0 | loss 3.2446 | lr 3.46e-04 | grad 0.3436 +2026-04-10 12:53:08 - INFO - parrotllm.training - step 92090 | epoch 0 | loss 3.3645 | lr 3.46e-04 | grad 0.3249 +2026-04-10 12:53:11 - INFO - parrotllm.training - step 92100 | epoch 0 | loss 3.3174 | lr 3.46e-04 | grad 0.3020 +2026-04-10 12:53:14 - INFO - parrotllm.training - step 92110 | epoch 0 | loss 3.3530 | lr 3.45e-04 | grad 0.3064 +2026-04-10 12:53:17 - INFO - parrotllm.training - step 92120 | epoch 0 | loss 3.3448 | lr 3.45e-04 | grad 0.3071 +2026-04-10 12:53:20 - INFO - parrotllm.training - step 92130 | epoch 0 | loss 3.2852 | lr 3.45e-04 | grad 0.3109 +2026-04-10 12:53:23 - INFO - parrotllm.training - step 92140 | epoch 0 | loss 3.3284 | lr 3.45e-04 | grad 0.3070 +2026-04-10 12:53:26 - INFO - parrotllm.training - step 92150 | epoch 0 | loss 3.3197 | lr 3.45e-04 | grad 0.3341 +2026-04-10 12:53:29 - INFO - parrotllm.training - step 92160 | epoch 0 | loss 3.2977 | lr 3.45e-04 | grad 0.2953 +2026-04-10 12:53:32 - INFO - parrotllm.training - step 92170 | epoch 0 | loss 3.3236 | lr 3.45e-04 | grad 0.3227 +2026-04-10 12:53:35 - INFO - parrotllm.training - step 92180 | epoch 0 | loss 3.2910 | lr 3.45e-04 | grad 0.3566 +2026-04-10 12:53:38 - INFO - parrotllm.training - step 92190 | epoch 0 | loss 3.2217 | lr 3.44e-04 | grad 0.2955 +2026-04-10 12:53:41 - INFO - parrotllm.training - step 92200 | epoch 0 | loss 3.2088 | lr 3.44e-04 | grad 0.3471 +2026-04-10 12:53:44 - INFO - parrotllm.training - step 92210 | epoch 0 | loss 3.2480 | lr 3.44e-04 | grad 0.3537 +2026-04-10 12:53:47 - INFO - parrotllm.training - step 92220 | epoch 0 | loss 3.3297 | lr 3.44e-04 | grad 0.3175 +2026-04-10 12:53:50 - INFO - parrotllm.training - step 92230 | epoch 0 | loss 3.3903 | lr 3.44e-04 | grad 0.3097 +2026-04-10 12:53:53 - INFO - parrotllm.training - step 92240 | epoch 0 | loss 3.3782 | lr 3.44e-04 | grad 0.3296 +2026-04-10 12:53:57 - INFO - parrotllm.training - step 92250 | epoch 0 | loss 3.3612 | lr 3.44e-04 | grad 0.3149 +2026-04-10 12:54:00 - INFO - parrotllm.training - step 92260 | epoch 0 | loss 3.3742 | lr 3.44e-04 | grad 0.3114 +2026-04-10 12:54:03 - INFO - parrotllm.training - step 92270 | epoch 0 | loss 3.3661 | lr 3.43e-04 | grad 0.3139 +2026-04-10 12:54:06 - INFO - parrotllm.training - step 92280 | epoch 0 | loss 3.2709 | lr 3.43e-04 | grad 0.3047 +2026-04-10 12:54:09 - INFO - parrotllm.training - step 92290 | epoch 0 | loss 3.3090 | lr 3.43e-04 | grad 0.3276 +2026-04-10 12:54:12 - INFO - parrotllm.training - step 92300 | epoch 0 | loss 3.3100 | lr 3.43e-04 | grad 0.3171 +2026-04-10 12:54:15 - INFO - parrotllm.training - step 92310 | epoch 0 | loss 3.3344 | lr 3.43e-04 | grad 0.3568 +2026-04-10 12:54:18 - INFO - parrotllm.training - step 92320 | epoch 0 | loss 3.2514 | lr 3.43e-04 | grad 0.2992 +2026-04-10 12:54:21 - INFO - parrotllm.training - step 92330 | epoch 0 | loss 3.3390 | lr 3.43e-04 | grad 0.3231 +2026-04-10 12:54:24 - INFO - parrotllm.training - step 92340 | epoch 0 | loss 3.2912 | lr 3.43e-04 | grad 0.3131 +2026-04-10 12:54:27 - INFO - parrotllm.training - step 92350 | epoch 0 | loss 3.3133 | lr 3.42e-04 | grad 0.3071 +2026-04-10 12:54:30 - INFO - parrotllm.training - step 92360 | epoch 0 | loss 3.2250 | lr 3.42e-04 | grad 0.3397 +2026-04-10 12:54:33 - INFO - parrotllm.training - step 92370 | epoch 0 | loss 3.3112 | lr 3.42e-04 | grad 0.3413 +2026-04-10 12:54:36 - INFO - parrotllm.training - step 92380 | epoch 0 | loss 3.2669 | lr 3.42e-04 | grad 0.3057 +2026-04-10 12:54:39 - INFO - parrotllm.training - step 92390 | epoch 0 | loss 3.2962 | lr 3.42e-04 | grad 0.3357 +2026-04-10 12:54:42 - INFO - parrotllm.training - step 92400 | epoch 0 | loss 3.2638 | lr 3.42e-04 | grad 0.3223 +2026-04-10 12:54:45 - INFO - parrotllm.training - step 92410 | epoch 0 | loss 3.2949 | lr 3.42e-04 | grad 0.3191 +2026-04-10 12:54:48 - INFO - parrotllm.training - step 92420 | epoch 0 | loss 3.3621 | lr 3.42e-04 | grad 0.3051 +2026-04-10 12:54:51 - INFO - parrotllm.training - step 92430 | epoch 0 | loss 3.2667 | lr 3.41e-04 | grad 0.3343 +2026-04-10 12:54:55 - INFO - parrotllm.training - step 92440 | epoch 0 | loss 3.2818 | lr 3.41e-04 | grad 0.3032 +2026-04-10 12:54:58 - INFO - parrotllm.training - step 92450 | epoch 0 | loss 3.3401 | lr 3.41e-04 | grad 0.3406 +2026-04-10 12:55:01 - INFO - parrotllm.training - step 92460 | epoch 0 | loss 3.2825 | lr 3.41e-04 | grad 0.3436 +2026-04-10 12:55:04 - INFO - parrotllm.training - step 92470 | epoch 0 | loss 3.3472 | lr 3.41e-04 | grad 0.3045 +2026-04-10 12:55:07 - INFO - parrotllm.training - step 92480 | epoch 0 | loss 3.3579 | lr 3.41e-04 | grad 0.3190 +2026-04-10 12:55:10 - INFO - parrotllm.training - step 92490 | epoch 0 | loss 3.2724 | lr 3.41e-04 | grad 0.3212 +2026-04-10 12:55:13 - INFO - parrotllm.training - step 92500 | epoch 0 | loss 3.3626 | lr 3.41e-04 | grad 0.3408 +2026-04-10 12:55:13 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:55:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:55:16 - INFO - parrotllm.training - Train: loss=3.3626, ppl=28.86 +2026-04-10 12:55:16 - INFO - parrotllm.training - Val: loss=3.2729, ppl=26.39 +2026-04-10 12:55:16 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 12:55:17 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2729_epoch_0000_step_0092500.pt +2026-04-10 12:55:19 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:55:20 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0092500.pt +2026-04-10 12:55:25 - INFO - parrotllm.training - step 92510 | epoch 0 | loss 3.2994 | lr 3.41e-04 | grad 0.3155 +2026-04-10 12:55:29 - INFO - parrotllm.training - step 92520 | epoch 0 | loss 3.3150 | lr 3.40e-04 | grad 0.3487 +2026-04-10 12:55:32 - INFO - parrotllm.training - step 92530 | epoch 0 | loss 3.4190 | lr 3.40e-04 | grad 0.3137 +2026-04-10 12:55:35 - INFO - parrotllm.training - step 92540 | epoch 0 | loss 3.4370 | lr 3.40e-04 | grad 0.3430 +2026-04-10 12:55:38 - INFO - parrotllm.training - step 92550 | epoch 0 | loss 3.2802 | lr 3.40e-04 | grad 0.3410 +2026-04-10 12:55:41 - INFO - parrotllm.training - step 92560 | epoch 0 | loss 3.2963 | lr 3.40e-04 | grad 0.2948 +2026-04-10 12:55:44 - INFO - parrotllm.training - step 92570 | epoch 0 | loss 3.3558 | lr 3.40e-04 | grad 0.3188 +2026-04-10 12:55:47 - INFO - parrotllm.training - step 92580 | epoch 0 | loss 3.2540 | lr 3.40e-04 | grad 0.3070 +2026-04-10 12:55:50 - INFO - parrotllm.training - step 92590 | epoch 0 | loss 3.2972 | lr 3.40e-04 | grad 0.3153 +2026-04-10 12:55:53 - INFO - parrotllm.training - step 92600 | epoch 0 | loss 3.2308 | lr 3.39e-04 | grad 0.2977 +2026-04-10 12:55:56 - INFO - parrotllm.training - step 92610 | epoch 0 | loss 3.3388 | lr 3.39e-04 | grad 0.3110 +2026-04-10 12:55:59 - INFO - parrotllm.training - step 92620 | epoch 0 | loss 3.4237 | lr 3.39e-04 | grad 0.3314 +2026-04-10 12:56:02 - INFO - parrotllm.training - step 92630 | epoch 0 | loss 3.3021 | lr 3.39e-04 | grad 0.3152 +2026-04-10 12:56:05 - INFO - parrotllm.training - step 92640 | epoch 0 | loss 3.4070 | lr 3.39e-04 | grad 0.2927 +2026-04-10 12:56:08 - INFO - parrotllm.training - step 92650 | epoch 0 | loss 3.4571 | lr 3.39e-04 | grad 0.3351 +2026-04-10 12:56:11 - INFO - parrotllm.training - step 92660 | epoch 0 | loss 3.3766 | lr 3.39e-04 | grad 0.3080 +2026-04-10 12:56:15 - INFO - parrotllm.training - step 92670 | epoch 0 | loss 3.3825 | lr 3.39e-04 | grad 0.3382 +2026-04-10 12:56:18 - INFO - parrotllm.training - step 92680 | epoch 0 | loss 3.3011 | lr 3.38e-04 | grad 0.3543 +2026-04-10 12:56:21 - INFO - parrotllm.training - step 92690 | epoch 0 | loss 3.2831 | lr 3.38e-04 | grad 0.2943 +2026-04-10 12:56:24 - INFO - parrotllm.training - step 92700 | epoch 0 | loss 3.3510 | lr 3.38e-04 | grad 0.3233 +2026-04-10 12:56:27 - INFO - parrotllm.training - step 92710 | epoch 0 | loss 3.3321 | lr 3.38e-04 | grad 0.3506 +2026-04-10 12:56:30 - INFO - parrotllm.training - step 92720 | epoch 0 | loss 3.3212 | lr 3.38e-04 | grad 0.3420 +2026-04-10 12:56:33 - INFO - parrotllm.training - step 92730 | epoch 0 | loss 3.3498 | lr 3.38e-04 | grad 0.3335 +2026-04-10 12:56:36 - INFO - parrotllm.training - step 92740 | epoch 0 | loss 3.4053 | lr 3.38e-04 | grad 0.3322 +2026-04-10 12:56:39 - INFO - parrotllm.training - step 92750 | epoch 0 | loss 3.3423 | lr 3.38e-04 | grad 0.3057 +2026-04-10 12:56:42 - INFO - parrotllm.training - step 92760 | epoch 0 | loss 3.2734 | lr 3.37e-04 | grad 0.3065 +2026-04-10 12:56:45 - INFO - parrotllm.training - step 92770 | epoch 0 | loss 3.3776 | lr 3.37e-04 | grad 0.3011 +2026-04-10 12:56:48 - INFO - parrotllm.training - step 92780 | epoch 0 | loss 3.2609 | lr 3.37e-04 | grad 0.3338 +2026-04-10 12:56:51 - INFO - parrotllm.training - step 92790 | epoch 0 | loss 3.2936 | lr 3.37e-04 | grad 0.3402 +2026-04-10 12:56:54 - INFO - parrotllm.training - step 92800 | epoch 0 | loss 3.3190 | lr 3.37e-04 | grad 0.3153 +2026-04-10 12:56:57 - INFO - parrotllm.training - step 92810 | epoch 0 | loss 3.3643 | lr 3.37e-04 | grad 0.3313 +2026-04-10 12:57:00 - INFO - parrotllm.training - step 92820 | epoch 0 | loss 3.3555 | lr 3.37e-04 | grad 0.3273 +2026-04-10 12:57:03 - INFO - parrotllm.training - step 92830 | epoch 0 | loss 3.3138 | lr 3.37e-04 | grad 0.3028 +2026-04-10 12:57:06 - INFO - parrotllm.training - step 92840 | epoch 0 | loss 3.3552 | lr 3.36e-04 | grad 0.3223 +2026-04-10 12:57:09 - INFO - parrotllm.training - step 92850 | epoch 0 | loss 3.3968 | lr 3.36e-04 | grad 0.3358 +2026-04-10 12:57:13 - INFO - parrotllm.training - step 92860 | epoch 0 | loss 3.2472 | lr 3.36e-04 | grad 0.3173 +2026-04-10 12:57:16 - INFO - parrotllm.training - step 92870 | epoch 0 | loss 3.2347 | lr 3.36e-04 | grad 0.3214 +2026-04-10 12:57:19 - INFO - parrotllm.training - step 92880 | epoch 0 | loss 3.3011 | lr 3.36e-04 | grad 0.3331 +2026-04-10 12:57:22 - INFO - parrotllm.training - step 92890 | epoch 0 | loss 3.3534 | lr 3.36e-04 | grad 0.3367 +2026-04-10 12:57:25 - INFO - parrotllm.training - step 92900 | epoch 0 | loss 3.3469 | lr 3.36e-04 | grad 0.3323 +2026-04-10 12:57:28 - INFO - parrotllm.training - step 92910 | epoch 0 | loss 3.2997 | lr 3.36e-04 | grad 0.3295 +2026-04-10 12:57:31 - INFO - parrotllm.training - step 92920 | epoch 0 | loss 3.2879 | lr 3.36e-04 | grad 0.3084 +2026-04-10 12:57:34 - INFO - parrotllm.training - step 92930 | epoch 0 | loss 3.2393 | lr 3.35e-04 | grad 0.3151 +2026-04-10 12:57:37 - INFO - parrotllm.training - step 92940 | epoch 0 | loss 3.3336 | lr 3.35e-04 | grad 0.3520 +2026-04-10 12:57:40 - INFO - parrotllm.training - step 92950 | epoch 0 | loss 3.3651 | lr 3.35e-04 | grad 0.3148 +2026-04-10 12:57:43 - INFO - parrotllm.training - step 92960 | epoch 0 | loss 3.3424 | lr 3.35e-04 | grad 0.3071 +2026-04-10 12:57:46 - INFO - parrotllm.training - step 92970 | epoch 0 | loss 3.3076 | lr 3.35e-04 | grad 0.3188 +2026-04-10 12:57:49 - INFO - parrotllm.training - step 92980 | epoch 0 | loss 3.2013 | lr 3.35e-04 | grad 0.3250 +2026-04-10 12:57:50 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 93019/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 12:57:53 - INFO - parrotllm.training - step 92990 | epoch 0 | loss 3.2028 | lr 3.35e-04 | grad 0.3238 +2026-04-10 12:57:56 - INFO - parrotllm.training - step 93000 | epoch 0 | loss 3.4155 | lr 3.35e-04 | grad 0.3128 +2026-04-10 12:57:56 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 12:57:56 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:57:59 - INFO - parrotllm.training - Train: loss=3.4155, ppl=30.43 +2026-04-10 12:57:59 - INFO - parrotllm.training - Val: loss=3.2719, ppl=26.36 +2026-04-10 12:58:00 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2719_epoch_0000_step_0093000.pt +2026-04-10 12:58:02 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2729, min_delta=0.001000). +2026-04-10 12:58:02 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 12:58:05 - INFO - parrotllm.training - step 93010 | epoch 0 | loss 3.2847 | lr 3.34e-04 | grad 0.3340 +2026-04-10 12:58:08 - INFO - parrotllm.training - step 93020 | epoch 0 | loss 3.2234 | lr 3.34e-04 | grad 0.3336 +2026-04-10 12:58:11 - INFO - parrotllm.training - step 93030 | epoch 0 | loss 3.3151 | lr 3.34e-04 | grad 0.3171 +2026-04-10 12:58:14 - INFO - parrotllm.training - step 93040 | epoch 0 | loss 3.3655 | lr 3.34e-04 | grad 0.3030 +2026-04-10 12:58:17 - INFO - parrotllm.training - step 93050 | epoch 0 | loss 3.3739 | lr 3.34e-04 | grad 0.3439 +2026-04-10 12:58:20 - INFO - parrotllm.training - step 93060 | epoch 0 | loss 3.3169 | lr 3.34e-04 | grad 0.3408 +2026-04-10 12:58:23 - INFO - parrotllm.training - step 93070 | epoch 0 | loss 3.3096 | lr 3.34e-04 | grad 0.2947 +2026-04-10 12:58:26 - INFO - parrotllm.training - step 93080 | epoch 0 | loss 3.3002 | lr 3.34e-04 | grad 0.3119 +2026-04-10 12:58:29 - INFO - parrotllm.training - step 93090 | epoch 0 | loss 3.2762 | lr 3.33e-04 | grad 0.3206 +2026-04-10 12:58:32 - INFO - parrotllm.training - step 93100 | epoch 0 | loss 3.2423 | lr 3.33e-04 | grad 0.3388 +2026-04-10 12:58:35 - INFO - parrotllm.training - step 93110 | epoch 0 | loss 3.3547 | lr 3.33e-04 | grad 0.3291 +2026-04-10 12:58:38 - INFO - parrotllm.training - step 93120 | epoch 0 | loss 3.2203 | lr 3.33e-04 | grad 0.3147 +2026-04-10 12:58:41 - INFO - parrotllm.training - step 93130 | epoch 0 | loss 3.3428 | lr 3.33e-04 | grad 0.3389 +2026-04-10 12:58:45 - INFO - parrotllm.training - step 93140 | epoch 0 | loss 3.3641 | lr 3.33e-04 | grad 0.3350 +2026-04-10 12:58:48 - INFO - parrotllm.training - step 93150 | epoch 0 | loss 3.3177 | lr 3.33e-04 | grad 0.3451 +2026-04-10 12:58:51 - INFO - parrotllm.training - step 93160 | epoch 0 | loss 3.1878 | lr 3.33e-04 | grad 0.2881 +2026-04-10 12:58:54 - INFO - parrotllm.training - step 93170 | epoch 0 | loss 3.2982 | lr 3.33e-04 | grad 0.3344 +2026-04-10 12:58:57 - INFO - parrotllm.training - step 93180 | epoch 0 | loss 3.2640 | lr 3.32e-04 | grad 0.3344 +2026-04-10 12:59:00 - INFO - parrotllm.training - step 93190 | epoch 0 | loss 3.2789 | lr 3.32e-04 | grad 0.3041 +2026-04-10 12:59:03 - INFO - parrotllm.training - step 93200 | epoch 0 | loss 3.2539 | lr 3.32e-04 | grad 0.3255 +2026-04-10 12:59:06 - INFO - parrotllm.training - step 93210 | epoch 0 | loss 3.3906 | lr 3.32e-04 | grad 0.3126 +2026-04-10 12:59:09 - INFO - parrotllm.training - step 93220 | epoch 0 | loss 3.2711 | lr 3.32e-04 | grad 0.3357 +2026-04-10 12:59:12 - INFO - parrotllm.training - step 93230 | epoch 0 | loss 3.3453 | lr 3.32e-04 | grad 0.3002 +2026-04-10 12:59:15 - INFO - parrotllm.training - step 93240 | epoch 0 | loss 3.2979 | lr 3.32e-04 | grad 0.3275 +2026-04-10 12:59:18 - INFO - parrotllm.training - step 93250 | epoch 0 | loss 3.3272 | lr 3.32e-04 | grad 0.2993 +2026-04-10 12:59:21 - INFO - parrotllm.training - step 93260 | epoch 0 | loss 3.3733 | lr 3.31e-04 | grad 0.3222 +2026-04-10 12:59:24 - INFO - parrotllm.training - step 93270 | epoch 0 | loss 3.3776 | lr 3.31e-04 | grad 0.3434 +2026-04-10 12:59:27 - INFO - parrotllm.training - step 93280 | epoch 0 | loss 3.4030 | lr 3.31e-04 | grad 0.2987 +2026-04-10 12:59:30 - INFO - parrotllm.training - step 93290 | epoch 0 | loss 3.2660 | lr 3.31e-04 | grad 0.3228 +2026-04-10 12:59:33 - INFO - parrotllm.training - step 93300 | epoch 0 | loss 3.3628 | lr 3.31e-04 | grad 0.3130 +2026-04-10 12:59:36 - INFO - parrotllm.training - step 93310 | epoch 0 | loss 3.2918 | lr 3.31e-04 | grad 0.3176 +2026-04-10 12:59:39 - INFO - parrotllm.training - step 93320 | epoch 0 | loss 3.2462 | lr 3.31e-04 | grad 0.3112 +2026-04-10 12:59:43 - INFO - parrotllm.training - step 93330 | epoch 0 | loss 3.3401 | lr 3.31e-04 | grad 0.3476 +2026-04-10 12:59:46 - INFO - parrotllm.training - step 93340 | epoch 0 | loss 3.2753 | lr 3.30e-04 | grad 0.3029 +2026-04-10 12:59:49 - INFO - parrotllm.training - step 93350 | epoch 0 | loss 3.2674 | lr 3.30e-04 | grad 0.3078 +2026-04-10 12:59:52 - INFO - parrotllm.training - step 93360 | epoch 0 | loss 3.3876 | lr 3.30e-04 | grad 0.3074 +2026-04-10 12:59:55 - INFO - parrotllm.training - step 93370 | epoch 0 | loss 3.4029 | lr 3.30e-04 | grad 0.3093 +2026-04-10 12:59:58 - INFO - parrotllm.training - step 93380 | epoch 0 | loss 3.3625 | lr 3.30e-04 | grad 0.3252 +2026-04-10 13:00:01 - INFO - parrotllm.training - step 93390 | epoch 0 | loss 3.3935 | lr 3.30e-04 | grad 0.3101 +2026-04-10 13:00:04 - INFO - parrotllm.training - step 93400 | epoch 0 | loss 3.3153 | lr 3.30e-04 | grad 0.3150 +2026-04-10 13:00:07 - INFO - parrotllm.training - step 93410 | epoch 0 | loss 3.3646 | lr 3.30e-04 | grad 0.3552 +2026-04-10 13:00:10 - INFO - parrotllm.training - step 93420 | epoch 0 | loss 3.3472 | lr 3.30e-04 | grad 0.3296 +2026-04-10 13:00:13 - INFO - parrotllm.training - step 93430 | epoch 0 | loss 3.3024 | lr 3.29e-04 | grad 0.3007 +2026-04-10 13:00:16 - INFO - parrotllm.training - step 93440 | epoch 0 | loss 3.4679 | lr 3.29e-04 | grad 0.3011 +2026-04-10 13:00:19 - INFO - parrotllm.training - step 93450 | epoch 0 | loss 3.2554 | lr 3.29e-04 | grad 0.3342 +2026-04-10 13:00:22 - INFO - parrotllm.training - step 93460 | epoch 0 | loss 3.2077 | lr 3.29e-04 | grad 0.3257 +2026-04-10 13:00:25 - INFO - parrotllm.training - step 93470 | epoch 0 | loss 3.2823 | lr 3.29e-04 | grad 0.3146 +2026-04-10 13:00:28 - INFO - parrotllm.training - step 93480 | epoch 0 | loss 3.2347 | lr 3.29e-04 | grad 0.3293 +2026-04-10 13:00:31 - INFO - parrotllm.training - step 93490 | epoch 0 | loss 3.2760 | lr 3.29e-04 | grad 0.3184 +2026-04-10 13:00:34 - INFO - parrotllm.training - step 93500 | epoch 0 | loss 3.2747 | lr 3.29e-04 | grad 0.3080 +2026-04-10 13:00:34 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 13:00:34 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:00:37 - INFO - parrotllm.training - Train: loss=3.2747, ppl=26.44 +2026-04-10 13:00:37 - INFO - parrotllm.training - Val: loss=3.2688, ppl=26.28 +2026-04-10 13:00:37 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 13:00:39 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2688_epoch_0000_step_0093500.pt +2026-04-10 13:00:41 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:00:44 - INFO - parrotllm.training - step 93510 | epoch 0 | loss 3.3031 | lr 3.28e-04 | grad 0.3028 +2026-04-10 13:00:47 - INFO - parrotllm.training - step 93520 | epoch 0 | loss 3.2896 | lr 3.28e-04 | grad 0.3146 +2026-04-10 13:00:50 - INFO - parrotllm.training - step 93530 | epoch 0 | loss 3.3214 | lr 3.28e-04 | grad 0.3252 +2026-04-10 13:00:53 - INFO - parrotllm.training - step 93540 | epoch 0 | loss 3.2921 | lr 3.28e-04 | grad 0.3252 +2026-04-10 13:00:56 - INFO - parrotllm.training - step 93550 | epoch 0 | loss 3.3343 | lr 3.28e-04 | grad 0.3146 +2026-04-10 13:00:59 - INFO - parrotllm.training - step 93560 | epoch 0 | loss 3.3242 | lr 3.28e-04 | grad 0.3630 +2026-04-10 13:01:02 - INFO - parrotllm.training - step 93570 | epoch 0 | loss 3.2837 | lr 3.28e-04 | grad 0.3429 +2026-04-10 13:01:05 - INFO - parrotllm.training - step 93580 | epoch 0 | loss 3.3628 | lr 3.28e-04 | grad 0.3251 +2026-04-10 13:01:08 - INFO - parrotllm.training - step 93590 | epoch 0 | loss 3.2312 | lr 3.27e-04 | grad 0.3255 +2026-04-10 13:01:11 - INFO - parrotllm.training - step 93600 | epoch 0 | loss 3.3775 | lr 3.27e-04 | grad 0.3396 +2026-04-10 13:01:14 - INFO - parrotllm.training - step 93610 | epoch 0 | loss 3.3667 | lr 3.27e-04 | grad 0.3493 +2026-04-10 13:01:18 - INFO - parrotllm.training - step 93620 | epoch 0 | loss 3.4258 | lr 3.27e-04 | grad 0.3159 +2026-04-10 13:01:21 - INFO - parrotllm.training - step 93630 | epoch 0 | loss 3.3235 | lr 3.27e-04 | grad 0.3010 +2026-04-10 13:01:24 - INFO - parrotllm.training - step 93640 | epoch 0 | loss 3.4558 | lr 3.27e-04 | grad 0.3085 +2026-04-10 13:01:27 - INFO - parrotllm.training - step 93650 | epoch 0 | loss 3.2542 | lr 3.27e-04 | grad 0.3204 +2026-04-10 13:01:30 - INFO - parrotllm.training - step 93660 | epoch 0 | loss 3.2995 | lr 3.27e-04 | grad 0.3180 +2026-04-10 13:01:33 - INFO - parrotllm.training - step 93670 | epoch 0 | loss 3.2687 | lr 3.27e-04 | grad 0.3232 +2026-04-10 13:01:36 - INFO - parrotllm.training - step 93680 | epoch 0 | loss 3.2845 | lr 3.26e-04 | grad 0.3042 +2026-04-10 13:01:39 - INFO - parrotllm.training - step 93690 | epoch 0 | loss 3.4137 | lr 3.26e-04 | grad 0.3195 +2026-04-10 13:01:42 - INFO - parrotllm.training - step 93700 | epoch 0 | loss 3.3597 | lr 3.26e-04 | grad 0.3269 +2026-04-10 13:01:45 - INFO - parrotllm.training - step 93710 | epoch 0 | loss 3.1930 | lr 3.26e-04 | grad 0.3277 +2026-04-10 13:01:48 - INFO - parrotllm.training - step 93720 | epoch 0 | loss 3.4246 | lr 3.26e-04 | grad 0.3504 +2026-04-10 13:01:51 - INFO - parrotllm.training - step 93730 | epoch 0 | loss 3.3784 | lr 3.26e-04 | grad 0.3273 +2026-04-10 13:01:54 - INFO - parrotllm.training - step 93740 | epoch 0 | loss 3.4854 | lr 3.26e-04 | grad 0.3189 +2026-04-10 13:01:57 - INFO - parrotllm.training - step 93750 | epoch 0 | loss 3.3663 | lr 3.26e-04 | grad 0.3136 +2026-04-10 13:02:00 - INFO - parrotllm.training - step 93760 | epoch 0 | loss 3.2433 | lr 3.25e-04 | grad 0.3098 +2026-04-10 13:02:03 - INFO - parrotllm.training - step 93770 | epoch 0 | loss 3.3035 | lr 3.25e-04 | grad 0.3016 +2026-04-10 13:02:06 - INFO - parrotllm.training - step 93780 | epoch 0 | loss 3.3426 | lr 3.25e-04 | grad 0.3141 +2026-04-10 13:02:09 - INFO - parrotllm.training - step 93790 | epoch 0 | loss 3.3168 | lr 3.25e-04 | grad 0.3365 +2026-04-10 13:02:12 - INFO - parrotllm.training - step 93800 | epoch 0 | loss 3.3260 | lr 3.25e-04 | grad 0.3445 +2026-04-10 13:02:15 - INFO - parrotllm.training - step 93810 | epoch 0 | loss 3.3587 | lr 3.25e-04 | grad 0.3171 +2026-04-10 13:02:19 - INFO - parrotllm.training - step 93820 | epoch 0 | loss 3.3520 | lr 3.25e-04 | grad 0.3124 +2026-04-10 13:02:22 - INFO - parrotllm.training - step 93830 | epoch 0 | loss 3.3200 | lr 3.25e-04 | grad 0.3300 +2026-04-10 13:02:25 - INFO - parrotllm.training - step 93840 | epoch 0 | loss 3.3071 | lr 3.25e-04 | grad 0.3083 +2026-04-10 13:02:28 - INFO - parrotllm.training - step 93850 | epoch 0 | loss 3.3242 | lr 3.24e-04 | grad 0.3012 +2026-04-10 13:02:31 - INFO - parrotllm.training - step 93860 | epoch 0 | loss 3.3208 | lr 3.24e-04 | grad 0.3328 +2026-04-10 13:02:34 - INFO - parrotllm.training - step 93870 | epoch 0 | loss 3.3607 | lr 3.24e-04 | grad 0.3280 +2026-04-10 13:02:37 - INFO - parrotllm.training - step 93880 | epoch 0 | loss 3.3701 | lr 3.24e-04 | grad 0.3040 +2026-04-10 13:02:40 - INFO - parrotllm.training - step 93890 | epoch 0 | loss 3.3227 | lr 3.24e-04 | grad 0.3358 +2026-04-10 13:02:43 - INFO - parrotllm.training - step 93900 | epoch 0 | loss 3.3345 | lr 3.24e-04 | grad 0.3088 +2026-04-10 13:02:46 - INFO - parrotllm.training - step 93910 | epoch 0 | loss 3.2895 | lr 3.24e-04 | grad 0.3132 +2026-04-10 13:02:49 - INFO - parrotllm.training - step 93920 | epoch 0 | loss 3.2844 | lr 3.24e-04 | grad 0.3109 +2026-04-10 13:02:52 - INFO - parrotllm.training - step 93930 | epoch 0 | loss 3.2489 | lr 3.23e-04 | grad 0.3202 +2026-04-10 13:02:55 - INFO - parrotllm.training - step 93940 | epoch 0 | loss 3.3341 | lr 3.23e-04 | grad 0.2976 +2026-04-10 13:02:58 - INFO - parrotllm.training - step 93950 | epoch 0 | loss 3.4015 | lr 3.23e-04 | grad 0.3089 +2026-04-10 13:03:01 - INFO - parrotllm.training - step 93960 | epoch 0 | loss 3.3202 | lr 3.23e-04 | grad 0.2965 +2026-04-10 13:03:04 - INFO - parrotllm.training - step 93970 | epoch 0 | loss 3.4267 | lr 3.23e-04 | grad 0.3054 +2026-04-10 13:03:07 - INFO - parrotllm.training - step 93980 | epoch 0 | loss 3.3764 | lr 3.23e-04 | grad 0.2988 +2026-04-10 13:03:10 - INFO - parrotllm.training - step 93990 | epoch 0 | loss 3.2898 | lr 3.23e-04 | grad 0.3114 +2026-04-10 13:03:13 - INFO - parrotllm.training - step 94000 | epoch 0 | loss 3.3352 | lr 3.23e-04 | grad 0.3113 +2026-04-10 13:03:13 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 13:03:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:03:17 - INFO - parrotllm.training - Train: loss=3.3352, ppl=28.08 +2026-04-10 13:03:17 - INFO - parrotllm.training - Val: loss=3.2682, ppl=26.26 +2026-04-10 13:03:18 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2682_epoch_0000_step_0094000.pt +2026-04-10 13:03:20 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2688, min_delta=0.001000). +2026-04-10 13:03:20 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:03:23 - INFO - parrotllm.training - step 94010 | epoch 0 | loss 3.3431 | lr 3.23e-04 | grad 0.3405 +2026-04-10 13:03:26 - INFO - parrotllm.training - step 94020 | epoch 0 | loss 3.2925 | lr 3.22e-04 | grad 0.3246 +2026-04-10 13:03:29 - INFO - parrotllm.training - step 94030 | epoch 0 | loss 3.2931 | lr 3.22e-04 | grad 0.3227 +2026-04-10 13:03:32 - INFO - parrotllm.training - step 94040 | epoch 0 | loss 3.3420 | lr 3.22e-04 | grad 0.3163 +2026-04-10 13:03:35 - INFO - parrotllm.training - step 94050 | epoch 0 | loss 3.3898 | lr 3.22e-04 | grad 0.3301 +2026-04-10 13:03:38 - INFO - parrotllm.training - step 94060 | epoch 0 | loss 3.2991 | lr 3.22e-04 | grad 0.3535 +2026-04-10 13:03:41 - INFO - parrotllm.training - step 94070 | epoch 0 | loss 3.4351 | lr 3.22e-04 | grad 0.3200 +2026-04-10 13:03:44 - INFO - parrotllm.training - step 94080 | epoch 0 | loss 3.2904 | lr 3.22e-04 | grad 0.3540 +2026-04-10 13:03:47 - INFO - parrotllm.training - step 94090 | epoch 0 | loss 3.3713 | lr 3.22e-04 | grad 0.3315 +2026-04-10 13:03:50 - INFO - parrotllm.training - step 94100 | epoch 0 | loss 3.3994 | lr 3.21e-04 | grad 0.3208 +2026-04-10 13:03:53 - INFO - parrotllm.training - step 94110 | epoch 0 | loss 3.2493 | lr 3.21e-04 | grad 0.3273 +2026-04-10 13:03:56 - INFO - parrotllm.training - step 94120 | epoch 0 | loss 3.2841 | lr 3.21e-04 | grad 0.3178 +2026-04-10 13:03:59 - INFO - parrotllm.training - step 94130 | epoch 0 | loss 3.4013 | lr 3.21e-04 | grad 0.3447 +2026-04-10 13:04:02 - INFO - parrotllm.training - step 94140 | epoch 0 | loss 3.2358 | lr 3.21e-04 | grad 0.3328 +2026-04-10 13:04:06 - INFO - parrotllm.training - step 94150 | epoch 0 | loss 3.4015 | lr 3.21e-04 | grad 0.3706 +2026-04-10 13:04:09 - INFO - parrotllm.training - step 94160 | epoch 0 | loss 3.4361 | lr 3.21e-04 | grad 0.3661 +2026-04-10 13:04:12 - INFO - parrotllm.training - step 94170 | epoch 0 | loss 3.2149 | lr 3.21e-04 | grad 0.3077 +2026-04-10 13:04:15 - INFO - parrotllm.training - step 94180 | epoch 0 | loss 3.2752 | lr 3.21e-04 | grad 0.3389 +2026-04-10 13:04:18 - INFO - parrotllm.training - step 94190 | epoch 0 | loss 3.2361 | lr 3.20e-04 | grad 0.3355 +2026-04-10 13:04:21 - INFO - parrotllm.training - step 94200 | epoch 0 | loss 3.2993 | lr 3.20e-04 | grad 0.3253 +2026-04-10 13:04:24 - INFO - parrotllm.training - step 94210 | epoch 0 | loss 3.3233 | lr 3.20e-04 | grad 0.3214 +2026-04-10 13:04:27 - INFO - parrotllm.training - step 94220 | epoch 0 | loss 3.3275 | lr 3.20e-04 | grad 0.3348 +2026-04-10 13:04:30 - INFO - parrotllm.training - step 94230 | epoch 0 | loss 3.4079 | lr 3.20e-04 | grad 0.3318 +2026-04-10 13:04:33 - INFO - parrotllm.training - step 94240 | epoch 0 | loss 3.2221 | lr 3.20e-04 | grad 0.3226 +2026-04-10 13:04:36 - INFO - parrotllm.training - step 94250 | epoch 0 | loss 3.3867 | lr 3.20e-04 | grad 0.3428 +2026-04-10 13:04:39 - INFO - parrotllm.training - step 94260 | epoch 0 | loss 3.2619 | lr 3.20e-04 | grad 0.3308 +2026-04-10 13:04:42 - INFO - parrotllm.training - step 94270 | epoch 0 | loss 3.3663 | lr 3.19e-04 | grad 0.3490 +2026-04-10 13:04:45 - INFO - parrotllm.training - step 94280 | epoch 0 | loss 3.3210 | lr 3.19e-04 | grad 0.3200 +2026-04-10 13:04:48 - INFO - parrotllm.training - step 94290 | epoch 0 | loss 3.2330 | lr 3.19e-04 | grad 0.3217 +2026-04-10 13:04:51 - INFO - parrotllm.training - step 94300 | epoch 0 | loss 3.3588 | lr 3.19e-04 | grad 0.3636 +2026-04-10 13:04:54 - INFO - parrotllm.training - step 94310 | epoch 0 | loss 3.3033 | lr 3.19e-04 | grad 0.3082 +2026-04-10 13:04:57 - INFO - parrotllm.training - step 94320 | epoch 0 | loss 3.3791 | lr 3.19e-04 | grad 0.3045 +2026-04-10 13:05:00 - INFO - parrotllm.training - step 94330 | epoch 0 | loss 3.3638 | lr 3.19e-04 | grad 0.3045 +2026-04-10 13:05:04 - INFO - parrotllm.training - step 94340 | epoch 0 | loss 3.3304 | lr 3.19e-04 | grad 0.3319 +2026-04-10 13:05:07 - INFO - parrotllm.training - step 94350 | epoch 0 | loss 3.3114 | lr 3.19e-04 | grad 0.3243 +2026-04-10 13:05:10 - INFO - parrotllm.training - step 94360 | epoch 0 | loss 3.3910 | lr 3.18e-04 | grad 0.3237 +2026-04-10 13:05:13 - INFO - parrotllm.training - step 94370 | epoch 0 | loss 3.3827 | lr 3.18e-04 | grad 0.3159 +2026-04-10 13:05:16 - INFO - parrotllm.training - step 94380 | epoch 0 | loss 3.3792 | lr 3.18e-04 | grad 0.3273 +2026-04-10 13:05:19 - INFO - parrotllm.training - step 94390 | epoch 0 | loss 3.3930 | lr 3.18e-04 | grad 0.3357 +2026-04-10 13:05:22 - INFO - parrotllm.training - step 94400 | epoch 0 | loss 3.2951 | lr 3.18e-04 | grad 0.3271 +2026-04-10 13:05:25 - INFO - parrotllm.training - step 94410 | epoch 0 | loss 3.2874 | lr 3.18e-04 | grad 0.3730 +2026-04-10 13:05:28 - INFO - parrotllm.training - step 94420 | epoch 0 | loss 3.3843 | lr 3.18e-04 | grad 0.3113 +2026-04-10 13:05:31 - INFO - parrotllm.training - step 94430 | epoch 0 | loss 3.3071 | lr 3.18e-04 | grad 0.3407 +2026-04-10 13:05:34 - INFO - parrotllm.training - step 94440 | epoch 0 | loss 3.2675 | lr 3.18e-04 | grad 0.3359 +2026-04-10 13:05:37 - INFO - parrotllm.training - step 94450 | epoch 0 | loss 3.2911 | lr 3.17e-04 | grad 0.3157 +2026-04-10 13:05:40 - INFO - parrotllm.training - step 94460 | epoch 0 | loss 3.3479 | lr 3.17e-04 | grad 0.3350 +2026-04-10 13:05:43 - INFO - parrotllm.training - step 94470 | epoch 0 | loss 3.3949 | lr 3.17e-04 | grad 0.3291 +2026-04-10 13:05:46 - INFO - parrotllm.training - step 94480 | epoch 0 | loss 3.2930 | lr 3.17e-04 | grad 0.3364 +2026-04-10 13:05:49 - INFO - parrotllm.training - step 94490 | epoch 0 | loss 3.4850 | lr 3.17e-04 | grad 0.3380 +2026-04-10 13:05:52 - INFO - parrotllm.training - step 94500 | epoch 0 | loss 3.2702 | lr 3.17e-04 | grad 0.3435 +2026-04-10 13:05:52 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 13:05:52 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:05:55 - INFO - parrotllm.training - Train: loss=3.2702, ppl=26.32 +2026-04-10 13:05:55 - INFO - parrotllm.training - Val: loss=3.2677, ppl=26.25 +2026-04-10 13:05:55 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 13:05:56 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2677_epoch_0000_step_0094500.pt +2026-04-10 13:05:58 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:06:01 - INFO - parrotllm.training - step 94510 | epoch 0 | loss 3.2288 | lr 3.17e-04 | grad 0.3303 +2026-04-10 13:06:05 - INFO - parrotllm.training - step 94520 | epoch 0 | loss 3.3607 | lr 3.17e-04 | grad 0.3033 +2026-04-10 13:06:08 - INFO - parrotllm.training - step 94530 | epoch 0 | loss 3.3115 | lr 3.16e-04 | grad 0.3964 +2026-04-10 13:06:11 - INFO - parrotllm.training - step 94540 | epoch 0 | loss 3.3807 | lr 3.16e-04 | grad 0.3431 +2026-04-10 13:06:14 - INFO - parrotllm.training - step 94550 | epoch 0 | loss 3.4081 | lr 3.16e-04 | grad 0.3136 +2026-04-10 13:06:17 - INFO - parrotllm.training - step 94560 | epoch 0 | loss 3.2285 | lr 3.16e-04 | grad 0.3170 +2026-04-10 13:06:20 - INFO - parrotllm.training - step 94570 | epoch 0 | loss 3.3469 | lr 3.16e-04 | grad 0.3133 +2026-04-10 13:06:23 - INFO - parrotllm.training - step 94580 | epoch 0 | loss 3.2925 | lr 3.16e-04 | grad 0.3156 +2026-04-10 13:06:26 - INFO - parrotllm.training - step 94590 | epoch 0 | loss 3.4061 | lr 3.16e-04 | grad 0.3070 +2026-04-10 13:06:29 - INFO - parrotllm.training - step 94600 | epoch 0 | loss 3.3128 | lr 3.16e-04 | grad 0.3264 +2026-04-10 13:06:32 - INFO - parrotllm.training - step 94610 | epoch 0 | loss 3.3757 | lr 3.16e-04 | grad 0.3319 +2026-04-10 13:06:35 - INFO - parrotllm.training - step 94620 | epoch 0 | loss 3.2846 | lr 3.15e-04 | grad 0.3162 +2026-04-10 13:06:38 - INFO - parrotllm.training - step 94630 | epoch 0 | loss 3.1767 | lr 3.15e-04 | grad 0.3128 +2026-04-10 13:06:41 - INFO - parrotllm.training - step 94640 | epoch 0 | loss 3.2828 | lr 3.15e-04 | grad 0.3357 +2026-04-10 13:06:44 - INFO - parrotllm.training - step 94650 | epoch 0 | loss 3.2434 | lr 3.15e-04 | grad 0.3277 +2026-04-10 13:06:47 - INFO - parrotllm.training - step 94660 | epoch 0 | loss 3.3365 | lr 3.15e-04 | grad 0.3527 +2026-04-10 13:06:50 - INFO - parrotllm.training - step 94670 | epoch 0 | loss 3.2267 | lr 3.15e-04 | grad 0.3203 +2026-04-10 13:06:53 - INFO - parrotllm.training - step 94680 | epoch 0 | loss 3.4042 | lr 3.15e-04 | grad 0.3026 +2026-04-10 13:06:56 - INFO - parrotllm.training - step 94690 | epoch 0 | loss 3.2612 | lr 3.15e-04 | grad 0.3084 +2026-04-10 13:06:59 - INFO - parrotllm.training - step 94700 | epoch 0 | loss 3.3865 | lr 3.14e-04 | grad 0.3654 +2026-04-10 13:07:02 - INFO - parrotllm.training - step 94710 | epoch 0 | loss 3.3355 | lr 3.14e-04 | grad 0.3537 +2026-04-10 13:07:06 - INFO - parrotllm.training - step 94720 | epoch 0 | loss 3.2189 | lr 3.14e-04 | grad 0.3383 +2026-04-10 13:07:09 - INFO - parrotllm.training - step 94730 | epoch 0 | loss 3.3502 | lr 3.14e-04 | grad 0.3130 +2026-04-10 13:07:12 - INFO - parrotllm.training - step 94740 | epoch 0 | loss 3.4018 | lr 3.14e-04 | grad 0.3054 +2026-04-10 13:07:15 - INFO - parrotllm.training - step 94750 | epoch 0 | loss 3.3873 | lr 3.14e-04 | grad 0.3547 +2026-04-10 13:07:18 - INFO - parrotllm.training - step 94760 | epoch 0 | loss 3.3719 | lr 3.14e-04 | grad 0.3281 +2026-04-10 13:07:21 - INFO - parrotllm.training - step 94770 | epoch 0 | loss 3.2412 | lr 3.14e-04 | grad 0.3089 +2026-04-10 13:07:24 - INFO - parrotllm.training - step 94780 | epoch 0 | loss 3.3777 | lr 3.14e-04 | grad 0.3160 +2026-04-10 13:07:27 - INFO - parrotllm.training - step 94790 | epoch 0 | loss 3.2956 | lr 3.13e-04 | grad 0.3243 +2026-04-10 13:07:30 - INFO - parrotllm.training - step 94800 | epoch 0 | loss 3.3335 | lr 3.13e-04 | grad 0.3354 +2026-04-10 13:07:33 - INFO - parrotllm.training - step 94810 | epoch 0 | loss 3.2353 | lr 3.13e-04 | grad 0.3244 +2026-04-10 13:07:36 - INFO - parrotllm.training - step 94820 | epoch 0 | loss 3.3897 | lr 3.13e-04 | grad 0.3135 +2026-04-10 13:07:39 - INFO - parrotllm.training - step 94830 | epoch 0 | loss 3.3305 | lr 3.13e-04 | grad 0.3285 +2026-04-10 13:07:42 - INFO - parrotllm.training - step 94840 | epoch 0 | loss 3.2496 | lr 3.13e-04 | grad 0.3273 +2026-04-10 13:07:45 - INFO - parrotllm.training - step 94850 | epoch 0 | loss 3.3445 | lr 3.13e-04 | grad 0.3367 +2026-04-10 13:07:48 - INFO - parrotllm.training - step 94860 | epoch 0 | loss 3.2433 | lr 3.13e-04 | grad 0.3370 +2026-04-10 13:07:51 - INFO - parrotllm.training - step 94870 | epoch 0 | loss 3.3005 | lr 3.13e-04 | grad 0.3200 +2026-04-10 13:07:54 - INFO - parrotllm.training - step 94880 | epoch 0 | loss 3.2685 | lr 3.12e-04 | grad 0.3115 +2026-04-10 13:07:57 - INFO - parrotllm.training - step 94890 | epoch 0 | loss 3.3149 | lr 3.12e-04 | grad 0.2972 +2026-04-10 13:08:00 - INFO - parrotllm.training - step 94900 | epoch 0 | loss 3.3988 | lr 3.12e-04 | grad 0.3046 +2026-04-10 13:08:04 - INFO - parrotllm.training - step 94910 | epoch 0 | loss 3.3431 | lr 3.12e-04 | grad 0.3220 +2026-04-10 13:08:07 - INFO - parrotllm.training - step 94920 | epoch 0 | loss 3.3643 | lr 3.12e-04 | grad 0.3174 +2026-04-10 13:08:10 - INFO - parrotllm.training - step 94930 | epoch 0 | loss 3.3154 | lr 3.12e-04 | grad 0.3261 +2026-04-10 13:08:13 - INFO - parrotllm.training - step 94940 | epoch 0 | loss 3.3211 | lr 3.12e-04 | grad 0.3157 +2026-04-10 13:08:16 - INFO - parrotllm.training - step 94950 | epoch 0 | loss 3.2701 | lr 3.12e-04 | grad 0.3266 +2026-04-10 13:08:19 - INFO - parrotllm.training - step 94960 | epoch 0 | loss 3.2721 | lr 3.12e-04 | grad 0.3097 +2026-04-10 13:08:22 - INFO - parrotllm.training - step 94970 | epoch 0 | loss 3.2914 | lr 3.11e-04 | grad 0.3480 +2026-04-10 13:08:25 - INFO - parrotllm.training - step 94980 | epoch 0 | loss 3.2773 | lr 3.11e-04 | grad 0.3338 +2026-04-10 13:08:28 - INFO - parrotllm.training - step 94990 | epoch 0 | loss 3.2601 | lr 3.11e-04 | grad 0.3064 +2026-04-10 13:08:31 - INFO - parrotllm.training - step 95000 | epoch 0 | loss 3.3380 | lr 3.11e-04 | grad 0.3077 +2026-04-10 13:08:31 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 13:08:31 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:08:34 - INFO - parrotllm.training - Train: loss=3.3380, ppl=28.16 +2026-04-10 13:08:34 - INFO - parrotllm.training - Val: loss=3.2668, ppl=26.23 +2026-04-10 13:08:35 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2668_epoch_0000_step_0095000.pt +2026-04-10 13:08:37 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2677, min_delta=0.001000). +2026-04-10 13:08:37 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:08:38 - INFO - parrotllm.training - Saved checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/last_epoch_0000_step_0095000.pt +2026-04-10 13:08:43 - INFO - parrotllm.training - step 95010 | epoch 0 | loss 3.3694 | lr 3.11e-04 | grad 0.3189 +2026-04-10 13:08:46 - INFO - parrotllm.training - step 95020 | epoch 0 | loss 3.2837 | lr 3.11e-04 | grad 0.3033 +2026-04-10 13:08:49 - INFO - parrotllm.training - step 95030 | epoch 0 | loss 3.3831 | lr 3.11e-04 | grad 0.3163 +2026-04-10 13:08:52 - INFO - parrotllm.training - step 95040 | epoch 0 | loss 3.3756 | lr 3.11e-04 | grad 0.3436 +2026-04-10 13:08:55 - INFO - parrotllm.training - step 95050 | epoch 0 | loss 3.3234 | lr 3.10e-04 | grad 0.3217 +2026-04-10 13:08:58 - INFO - parrotllm.training - step 95060 | epoch 0 | loss 3.2632 | lr 3.10e-04 | grad 0.3282 +2026-04-10 13:09:01 - INFO - parrotllm.training - step 95070 | epoch 0 | loss 3.3868 | lr 3.10e-04 | grad 0.3240 +2026-04-10 13:09:04 - INFO - parrotllm.training - step 95080 | epoch 0 | loss 3.3831 | lr 3.10e-04 | grad 0.3318 +2026-04-10 13:09:07 - INFO - parrotllm.training - step 95090 | epoch 0 | loss 3.3187 | lr 3.10e-04 | grad 0.3137 +2026-04-10 13:09:11 - INFO - parrotllm.training - step 95100 | epoch 0 | loss 3.2184 | lr 3.10e-04 | grad 0.3227 +2026-04-10 13:09:14 - INFO - parrotllm.training - step 95110 | epoch 0 | loss 3.3146 | lr 3.10e-04 | grad 0.3106 +2026-04-10 13:09:17 - INFO - parrotllm.training - step 95120 | epoch 0 | loss 3.3054 | lr 3.10e-04 | grad 0.3086 +2026-04-10 13:09:20 - INFO - parrotllm.training - step 95130 | epoch 0 | loss 3.2305 | lr 3.10e-04 | grad 0.2998 +2026-04-10 13:09:23 - INFO - parrotllm.training - step 95140 | epoch 0 | loss 3.3345 | lr 3.09e-04 | grad 0.3232 +2026-04-10 13:09:26 - INFO - parrotllm.training - step 95150 | epoch 0 | loss 3.3785 | lr 3.09e-04 | grad 0.3256 +2026-04-10 13:09:29 - INFO - parrotllm.training - step 95160 | epoch 0 | loss 3.3136 | lr 3.09e-04 | grad 0.3354 +2026-04-10 13:09:32 - INFO - parrotllm.training - step 95170 | epoch 0 | loss 3.2318 | lr 3.09e-04 | grad 0.3446 +2026-04-10 13:09:35 - INFO - parrotllm.training - step 95180 | epoch 0 | loss 3.3804 | lr 3.09e-04 | grad 0.3073 +2026-04-10 13:09:38 - INFO - parrotllm.training - step 95190 | epoch 0 | loss 3.5026 | lr 3.09e-04 | grad 0.3317 +2026-04-10 13:09:41 - INFO - parrotllm.training - step 95200 | epoch 0 | loss 3.2887 | lr 3.09e-04 | grad 0.3315 +2026-04-10 13:09:44 - INFO - parrotllm.training - step 95210 | epoch 0 | loss 3.3082 | lr 3.09e-04 | grad 0.3155 +2026-04-10 13:09:47 - INFO - parrotllm.training - step 95220 | epoch 0 | loss 3.3885 | lr 3.09e-04 | grad 0.3497 +2026-04-10 13:09:50 - INFO - parrotllm.training - step 95230 | epoch 0 | loss 3.3074 | lr 3.08e-04 | grad 0.3486 +2026-04-10 13:09:53 - INFO - parrotllm.training - step 95240 | epoch 0 | loss 3.2663 | lr 3.08e-04 | grad 0.3494 +2026-04-10 13:09:56 - INFO - parrotllm.training - step 95250 | epoch 0 | loss 3.2630 | lr 3.08e-04 | grad 0.3020 +2026-04-10 13:09:59 - INFO - parrotllm.training - step 95260 | epoch 0 | loss 3.2584 | lr 3.08e-04 | grad 0.3181 +2026-04-10 13:10:02 - INFO - parrotllm.training - step 95270 | epoch 0 | loss 3.2702 | lr 3.08e-04 | grad 0.3356 +2026-04-10 13:10:05 - INFO - parrotllm.training - step 95280 | epoch 0 | loss 3.3482 | lr 3.08e-04 | grad 0.3702 +2026-04-10 13:10:09 - INFO - parrotllm.training - step 95290 | epoch 0 | loss 3.2712 | lr 3.08e-04 | grad 0.3299 +2026-04-10 13:10:12 - INFO - parrotllm.training - step 95300 | epoch 0 | loss 3.3142 | lr 3.08e-04 | grad 0.3128 +2026-04-10 13:10:15 - INFO - parrotllm.training - step 95310 | epoch 0 | loss 3.2987 | lr 3.08e-04 | grad 0.3364 +2026-04-10 13:10:18 - INFO - parrotllm.training - step 95320 | epoch 0 | loss 3.2899 | lr 3.07e-04 | grad 0.3400 +2026-04-10 13:10:21 - INFO - parrotllm.training - step 95330 | epoch 0 | loss 3.4185 | lr 3.07e-04 | grad 0.3242 +2026-04-10 13:10:24 - INFO - parrotllm.training - step 95340 | epoch 0 | loss 3.4175 | lr 3.07e-04 | grad 0.3092 +2026-04-10 13:10:27 - INFO - parrotllm.training - step 95350 | epoch 0 | loss 3.3876 | lr 3.07e-04 | grad 0.3221 +2026-04-10 13:10:30 - INFO - parrotllm.training - step 95360 | epoch 0 | loss 3.2986 | lr 3.07e-04 | grad 0.3096 +2026-04-10 13:10:33 - INFO - parrotllm.training - step 95370 | epoch 0 | loss 3.2702 | lr 3.07e-04 | grad 0.3307 +2026-04-10 13:10:36 - INFO - parrotllm.training - step 95380 | epoch 0 | loss 3.2897 | lr 3.07e-04 | grad 0.3326 +2026-04-10 13:10:39 - INFO - parrotllm.training - step 95390 | epoch 0 | loss 3.3379 | lr 3.07e-04 | grad 0.3118 +2026-04-10 13:10:42 - INFO - parrotllm.training - step 95400 | epoch 0 | loss 3.2858 | lr 3.07e-04 | grad 0.3190 +2026-04-10 13:10:45 - INFO - parrotllm.training - step 95410 | epoch 0 | loss 3.4790 | lr 3.06e-04 | grad 0.3455 +2026-04-10 13:10:48 - INFO - parrotllm.training - step 95420 | epoch 0 | loss 3.3330 | lr 3.06e-04 | grad 0.3267 +2026-04-10 13:10:51 - INFO - parrotllm.training - step 95430 | epoch 0 | loss 3.2373 | lr 3.06e-04 | grad 0.3115 +2026-04-10 13:10:54 - INFO - parrotllm.training - step 95440 | epoch 0 | loss 3.2978 | lr 3.06e-04 | grad 0.3361 +2026-04-10 13:10:57 - INFO - parrotllm.training - step 95450 | epoch 0 | loss 3.3881 | lr 3.06e-04 | grad 0.3258 +2026-04-10 13:11:00 - INFO - parrotllm.training - step 95460 | epoch 0 | loss 3.2670 | lr 3.06e-04 | grad 0.3352 +2026-04-10 13:11:03 - INFO - parrotllm.training - step 95470 | epoch 0 | loss 3.3337 | lr 3.06e-04 | grad 0.3096 +2026-04-10 13:11:06 - INFO - parrotllm.training - step 95480 | epoch 0 | loss 3.3281 | lr 3.06e-04 | grad 0.3424 +2026-04-10 13:11:10 - INFO - parrotllm.training - step 95490 | epoch 0 | loss 3.3153 | lr 3.06e-04 | grad 0.3035 +2026-04-10 13:11:13 - INFO - parrotllm.training - step 95500 | epoch 0 | loss 3.3507 | lr 3.05e-04 | grad 0.3276 +2026-04-10 13:11:13 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 13:11:13 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:11:16 - INFO - parrotllm.training - Train: loss=3.3507, ppl=28.52 +2026-04-10 13:11:16 - INFO - parrotllm.training - Val: loss=3.2650, ppl=26.18 +2026-04-10 13:11:16 - INFO - parrotllm.training - ** New best validation loss! ** +2026-04-10 13:11:17 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2650_epoch_0000_step_0095500.pt +2026-04-10 13:11:19 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:11:22 - INFO - parrotllm.training - step 95510 | epoch 0 | loss 3.2985 | lr 3.05e-04 | grad 0.3194 +2026-04-10 13:11:25 - INFO - parrotllm.training - step 95520 | epoch 0 | loss 3.3891 | lr 3.05e-04 | grad 0.3108 +2026-04-10 13:11:28 - INFO - parrotllm.training - step 95530 | epoch 0 | loss 3.3687 | lr 3.05e-04 | grad 0.2950 +2026-04-10 13:11:31 - INFO - parrotllm.training - step 95540 | epoch 0 | loss 3.2627 | lr 3.05e-04 | grad 0.3285 +2026-04-10 13:11:34 - INFO - parrotllm.training - step 95550 | epoch 0 | loss 3.2911 | lr 3.05e-04 | grad 0.3121 +2026-04-10 13:11:37 - INFO - parrotllm.training - step 95560 | epoch 0 | loss 3.2520 | lr 3.05e-04 | grad 0.3229 +2026-04-10 13:11:40 - INFO - parrotllm.training - step 95570 | epoch 0 | loss 3.2529 | lr 3.05e-04 | grad 0.3592 +2026-04-10 13:11:43 - INFO - parrotllm.training - step 95580 | epoch 0 | loss 3.3820 | lr 3.05e-04 | grad 0.3608 +2026-04-10 13:11:46 - INFO - parrotllm.training - step 95590 | epoch 0 | loss 3.2768 | lr 3.04e-04 | grad 0.3776 +2026-04-10 13:11:49 - INFO - parrotllm.training - step 95600 | epoch 0 | loss 3.3469 | lr 3.04e-04 | grad 0.3277 +2026-04-10 13:11:52 - INFO - parrotllm.training - step 95610 | epoch 0 | loss 3.3597 | lr 3.04e-04 | grad 0.3013 +2026-04-10 13:11:55 - INFO - parrotllm.training - step 95620 | epoch 0 | loss 3.3679 | lr 3.04e-04 | grad 0.3060 +2026-04-10 13:11:58 - INFO - parrotllm.training - step 95630 | epoch 0 | loss 3.2850 | lr 3.04e-04 | grad 0.3086 +2026-04-10 13:12:01 - INFO - parrotllm.training - step 95640 | epoch 0 | loss 3.2803 | lr 3.04e-04 | grad 0.3397 +2026-04-10 13:12:04 - INFO - parrotllm.training - step 95650 | epoch 0 | loss 3.3118 | lr 3.04e-04 | grad 0.3750 +2026-04-10 13:12:07 - INFO - parrotllm.training - step 95660 | epoch 0 | loss 3.3378 | lr 3.04e-04 | grad 0.3438 +2026-04-10 13:12:11 - INFO - parrotllm.training - step 95670 | epoch 0 | loss 3.4163 | lr 3.04e-04 | grad 0.3744 +2026-04-10 13:12:14 - INFO - parrotllm.training - step 95680 | epoch 0 | loss 3.3192 | lr 3.03e-04 | grad 0.3726 +2026-04-10 13:12:17 - INFO - parrotllm.training - step 95690 | epoch 0 | loss 3.4475 | lr 3.03e-04 | grad 0.3484 +2026-04-10 13:12:20 - INFO - parrotllm.training - step 95700 | epoch 0 | loss 3.3156 | lr 3.03e-04 | grad 0.3309 +2026-04-10 13:12:23 - INFO - parrotllm.training - step 95710 | epoch 0 | loss 3.3081 | lr 3.03e-04 | grad 0.3260 +2026-04-10 13:12:26 - WARNING - parrotllm.training - Optimizer step skipped at epoch 0 after 95757/31574926 micro-batches; LR schedule and completed_steps remain unchanged. +2026-04-10 13:12:26 - INFO - parrotllm.training - step 95720 | epoch 0 | loss 3.3665 | lr 3.03e-04 | grad 0.3067 +2026-04-10 13:12:29 - INFO - parrotllm.training - step 95730 | epoch 0 | loss 3.2788 | lr 3.03e-04 | grad 0.3397 +2026-04-10 13:12:32 - INFO - parrotllm.training - step 95740 | epoch 0 | loss 3.3298 | lr 3.03e-04 | grad 0.3101 +2026-04-10 13:12:35 - INFO - parrotllm.training - step 95750 | epoch 0 | loss 3.3281 | lr 3.03e-04 | grad 0.3235 +2026-04-10 13:12:38 - INFO - parrotllm.training - step 95760 | epoch 0 | loss 3.3354 | lr 3.02e-04 | grad 0.3402 +2026-04-10 13:12:41 - INFO - parrotllm.training - step 95770 | epoch 0 | loss 3.2762 | lr 3.02e-04 | grad 0.3380 +2026-04-10 13:12:44 - INFO - parrotllm.training - step 95780 | epoch 0 | loss 3.3122 | lr 3.02e-04 | grad 0.3404 +2026-04-10 13:12:47 - INFO - parrotllm.training - step 95790 | epoch 0 | loss 3.2479 | lr 3.02e-04 | grad 0.3208 +2026-04-10 13:12:51 - INFO - parrotllm.training - step 95800 | epoch 0 | loss 3.3577 | lr 3.02e-04 | grad 0.3236 +2026-04-10 13:12:54 - INFO - parrotllm.training - step 95810 | epoch 0 | loss 3.3288 | lr 3.02e-04 | grad 0.3259 +2026-04-10 13:12:57 - INFO - parrotllm.training - step 95820 | epoch 0 | loss 3.3052 | lr 3.02e-04 | grad 0.3132 +2026-04-10 13:13:00 - INFO - parrotllm.training - step 95830 | epoch 0 | loss 3.2605 | lr 3.02e-04 | grad 0.3165 +2026-04-10 13:13:03 - INFO - parrotllm.training - step 95840 | epoch 0 | loss 3.2998 | lr 3.02e-04 | grad 0.3121 +2026-04-10 13:13:06 - INFO - parrotllm.training - step 95850 | epoch 0 | loss 3.3981 | lr 3.01e-04 | grad 0.3366 +2026-04-10 13:13:09 - INFO - parrotllm.training - step 95860 | epoch 0 | loss 3.3698 | lr 3.01e-04 | grad 0.3135 +2026-04-10 13:13:12 - INFO - parrotllm.training - step 95870 | epoch 0 | loss 3.3277 | lr 3.01e-04 | grad 0.3260 +2026-04-10 13:13:15 - INFO - parrotllm.training - step 95880 | epoch 0 | loss 3.2712 | lr 3.01e-04 | grad 0.3539 +2026-04-10 13:13:18 - INFO - parrotllm.training - step 95890 | epoch 0 | loss 3.3124 | lr 3.01e-04 | grad 0.3204 +2026-04-10 13:13:21 - INFO - parrotllm.training - step 95900 | epoch 0 | loss 3.1771 | lr 3.01e-04 | grad 0.3184 +2026-04-10 13:13:24 - INFO - parrotllm.training - step 95910 | epoch 0 | loss 3.3082 | lr 3.01e-04 | grad 0.3268 +2026-04-10 13:13:27 - INFO - parrotllm.training - step 95920 | epoch 0 | loss 3.3281 | lr 3.01e-04 | grad 0.3351 +2026-04-10 13:13:30 - INFO - parrotllm.training - step 95930 | epoch 0 | loss 3.3401 | lr 3.01e-04 | grad 0.3307 +2026-04-10 13:13:33 - INFO - parrotllm.training - step 95940 | epoch 0 | loss 3.2767 | lr 3.00e-04 | grad 0.3398 +2026-04-10 13:13:36 - INFO - parrotllm.training - step 95950 | epoch 0 | loss 3.3869 | lr 3.00e-04 | grad 0.3307 +2026-04-10 13:13:39 - INFO - parrotllm.training - step 95960 | epoch 0 | loss 3.3453 | lr 3.00e-04 | grad 0.3155 +2026-04-10 13:13:42 - INFO - parrotllm.training - step 95970 | epoch 0 | loss 3.3611 | lr 3.00e-04 | grad 0.3197 +2026-04-10 13:13:45 - INFO - parrotllm.training - step 95980 | epoch 0 | loss 3.3046 | lr 3.00e-04 | grad 0.3347 +2026-04-10 13:13:48 - INFO - parrotllm.training - step 95990 | epoch 0 | loss 3.3373 | lr 3.00e-04 | grad 0.3204 +2026-04-10 13:13:51 - INFO - parrotllm.training - step 96000 | epoch 0 | loss 3.3042 | lr 3.00e-04 | grad 0.3484 +2026-04-10 13:13:51 - INFO - parrotllm.training - Starting evaluation... +2026-04-10 13:13:51 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:13:55 - INFO - parrotllm.training - Train: loss=3.3042, ppl=27.23 +2026-04-10 13:13:55 - INFO - parrotllm.training - Val: loss=3.2650, ppl=26.18 +2026-04-10 13:13:56 - INFO - parrotllm.training - Saved best checkpoint: /home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260410_044337/checkpoints/best_loss_3p2650_epoch_0000_step_0096000.pt +2026-04-10 13:13:58 - INFO - parrotllm.training - No validation improvement for 1/15 evaluation(s) (best=3.2650, min_delta=0.001000). +2026-04-10 13:13:58 - INFO - parrotllm.training - ------------------------------------------------------------ +2026-04-10 13:14:01 - INFO - parrotllm.training - step 96010 | epoch 0 | loss 3.2164 | lr 3.00e-04 | grad 0.3267