diff --git "a/runs/big_run/exp_c/run_20260408_151821/metrics.jsonl" "b/runs/big_run/exp_c/run_20260408_151821/metrics.jsonl" new file mode 100644--- /dev/null +++ "b/runs/big_run/exp_c/run_20260408_151821/metrics.jsonl" @@ -0,0 +1,3132 @@ +{"stage": "pretraining", "type": "config", "timestamp": "2026-04-08T15:18:22.263239+00:00", "device": "auto", "train_bin": "data/exp_c/train.bin", "val_bin": "data/exp_a/val.bin", "num_workers": 4, "pin_memory": true, "batch_size": 32, "gradient_accumulation_steps": 1, "train_samples_per_epoch": null, "val_sequence_stride": null, "learning_rate": 0.001305, "min_lr": 0.0001305, "weight_decay": 0.063, "beta1": 0.9, "beta2": 0.9978, "grad_clip": 1.786, "warmup_steps": 305, "max_steps": 3052, "lr_schedule": "wsd", "lr_decay_ratio": 0.179, "z_loss_coeff": 0.000275, "save_every": 250, "eval_every": 100, "checkpoint_dir": "checkpoints", "keep_last_checkpoints": 10, "keep_best_checkpoints": 10, "early_stopping_patience": 0, "early_stopping_min_delta": 0.001, "runs_dir": "runs/big_run/exp_c", "log_every": 10, "hf_upload": {"repo_id": "ParrotLabs/Preprocessed", "repo_type": "dataset", "path_in_repo": "", "private": null}, "compile": true} +{"stage": "pretraining", "type": "checkpoint_policy", "timestamp": "2026-04-08T15:18:22.263322+00:00", "checkpoint_dir": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints", "keep_last": 10, "keep_best": 10} +{"stage": "pretraining", "type": "model_architecture", "timestamp": "2026-04-08T15:18:24.082297+00:00", "vocab_size": 50258, "context_length": 1024, "n_layers": 14, "n_heads": 6, "d_model": 384, "d_ff": 768, "dropout": 0.005, "bias": false, "total_params": 39966592, "total_params_non_embedding": 39966592, "trainable_params": 39966592, "non_trainable_params": 0, "params_size_mb": 159.87} +{"stage": "pretraining", "type": "initial_validation", "timestamp": "2026-04-08T15:18:32.916887+00:00", "val_loss": 10.892181491851806, "val_ppl": 53754.4370232034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:38.439571+00:00", "epoch": 0, "step": 1, "train_loss": 10.925138473510742, "perplexity": 55555.53741348256, "lr": 4.278688524590164e-06, "grad_norm": 7.191328, "tokens_per_sec": 60} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:38.770393+00:00", "epoch": 0, "step": 2, "train_loss": 10.887932777404785, "perplexity": 53526.534259746775, "lr": 8.557377049180328e-06, "grad_norm": 7.041317, "tokens_per_sec": 98891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:39.076355+00:00", "epoch": 0, "step": 3, "train_loss": 10.813186645507812, "perplexity": 49671.501914683584, "lr": 1.2836065573770493e-05, "grad_norm": 6.803489, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:39.382732+00:00", "epoch": 0, "step": 4, "train_loss": 10.707441329956055, "perplexity": 44687.15280269748, "lr": 1.7114754098360656e-05, "grad_norm": 6.585573, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:39.687998+00:00", "epoch": 0, "step": 5, "train_loss": 10.601113319396973, "perplexity": 40179.54520628084, "lr": 2.1393442622950817e-05, "grad_norm": 5.849924, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:39.992179+00:00", "epoch": 0, "step": 6, "train_loss": 10.502140998840332, "perplexity": 36393.33741567178, "lr": 2.5672131147540985e-05, "grad_norm": 4.206415, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:40.296581+00:00", "epoch": 0, "step": 7, "train_loss": 10.420210838317871, "perplexity": 33530.502915466794, "lr": 2.9950819672131143e-05, "grad_norm": 2.950861, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:40.602027+00:00", "epoch": 0, "step": 8, "train_loss": 10.350740432739258, "perplexity": 31280.195127751336, "lr": 3.422950819672131e-05, "grad_norm": 2.334904, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:40.906293+00:00", "epoch": 0, "step": 9, "train_loss": 10.308854103088379, "perplexity": 29997.043419021684, "lr": 3.850819672131147e-05, "grad_norm": 1.995344, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:41.211301+00:00", "epoch": 0, "step": 10, "train_loss": 10.240361213684082, "perplexity": 28011.24214303376, "lr": 4.2786885245901634e-05, "grad_norm": 1.825583, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:41.515512+00:00", "epoch": 0, "step": 11, "train_loss": 10.214096069335938, "perplexity": 27285.100674076017, "lr": 4.70655737704918e-05, "grad_norm": 1.765255, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:41.819575+00:00", "epoch": 0, "step": 12, "train_loss": 10.215258598327637, "perplexity": 27316.838839345255, "lr": 5.134426229508197e-05, "grad_norm": 1.740026, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:42.125155+00:00", "epoch": 0, "step": 13, "train_loss": 10.191203117370605, "perplexity": 26667.559811828603, "lr": 5.562295081967213e-05, "grad_norm": 1.730685, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:42.429022+00:00", "epoch": 0, "step": 14, "train_loss": 10.177030563354492, "perplexity": 26292.278012316045, "lr": 5.990163934426229e-05, "grad_norm": 1.708565, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:42.732729+00:00", "epoch": 0, "step": 15, "train_loss": 10.106958389282227, "perplexity": 24512.988263899322, "lr": 6.418032786885245e-05, "grad_norm": 1.71246, "tokens_per_sec": 107893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:43.036516+00:00", "epoch": 0, "step": 16, "train_loss": 10.095608711242676, "perplexity": 24236.346605823495, "lr": 6.845901639344262e-05, "grad_norm": 1.689027, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:43.341836+00:00", "epoch": 0, "step": 17, "train_loss": 10.040718078613281, "perplexity": 22941.85105579028, "lr": 7.273770491803279e-05, "grad_norm": 1.700617, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:43.646296+00:00", "epoch": 0, "step": 18, "train_loss": 9.999994277954102, "perplexity": 22026.339758719052, "lr": 7.701639344262295e-05, "grad_norm": 1.706825, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:43.951526+00:00", "epoch": 0, "step": 19, "train_loss": 9.95140266418457, "perplexity": 20981.63193474218, "lr": 8.12950819672131e-05, "grad_norm": 1.681398, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:44.255752+00:00", "epoch": 0, "step": 20, "train_loss": 9.883414268493652, "perplexity": 19602.536860351705, "lr": 8.557377049180327e-05, "grad_norm": 1.708031, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:44.559689+00:00", "epoch": 0, "step": 21, "train_loss": 9.812990188598633, "perplexity": 18269.534837591345, "lr": 8.985245901639344e-05, "grad_norm": 1.708455, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:44.864827+00:00", "epoch": 0, "step": 22, "train_loss": 9.799002647399902, "perplexity": 18015.767891647298, "lr": 9.41311475409836e-05, "grad_norm": 1.760283, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:45.170049+00:00", "epoch": 0, "step": 23, "train_loss": 9.712845802307129, "perplexity": 16528.572059643444, "lr": 9.840983606557377e-05, "grad_norm": 1.68651, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:45.475027+00:00", "epoch": 0, "step": 24, "train_loss": 9.667057037353516, "perplexity": 15788.814690830517, "lr": 0.00010268852459016394, "grad_norm": 1.681567, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:45.778246+00:00", "epoch": 0, "step": 25, "train_loss": 9.624343872070312, "perplexity": 15128.624220360103, "lr": 0.0001069672131147541, "grad_norm": 1.650385, "tokens_per_sec": 108066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:46.082059+00:00", "epoch": 0, "step": 26, "train_loss": 9.526634216308594, "perplexity": 13720.333606249811, "lr": 0.00011124590163934426, "grad_norm": 1.652497, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:46.386909+00:00", "epoch": 0, "step": 27, "train_loss": 9.529011726379395, "perplexity": 13752.992645787202, "lr": 0.0001155245901639344, "grad_norm": 1.631601, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:46.692165+00:00", "epoch": 0, "step": 28, "train_loss": 9.46126937866211, "perplexity": 12852.18839147634, "lr": 0.00011980327868852457, "grad_norm": 1.625283, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:46.998269+00:00", "epoch": 0, "step": 29, "train_loss": 9.362519264221191, "perplexity": 11643.685142994496, "lr": 0.00012408196721311476, "grad_norm": 1.625786, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:47.302651+00:00", "epoch": 0, "step": 30, "train_loss": 9.331765174865723, "perplexity": 11291.04457340158, "lr": 0.0001283606557377049, "grad_norm": 1.601478, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:47.606214+00:00", "epoch": 0, "step": 31, "train_loss": 9.210372924804688, "perplexity": 10000.325533583538, "lr": 0.00013263934426229506, "grad_norm": 1.588702, "tokens_per_sec": 107852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:47.911798+00:00", "epoch": 0, "step": 32, "train_loss": 9.185861587524414, "perplexity": 9758.183912209452, "lr": 0.00013691803278688525, "grad_norm": 1.590044, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:48.216986+00:00", "epoch": 0, "step": 33, "train_loss": 9.11579704284668, "perplexity": 9097.8831343634, "lr": 0.0001411967213114754, "grad_norm": 1.581696, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:48.521857+00:00", "epoch": 0, "step": 34, "train_loss": 9.076313972473145, "perplexity": 8745.669761650293, "lr": 0.00014547540983606558, "grad_norm": 1.703359, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:48.827871+00:00", "epoch": 0, "step": 35, "train_loss": 8.966943740844727, "perplexity": 7839.605089548538, "lr": 0.00014975409836065574, "grad_norm": 2.058831, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:49.131340+00:00", "epoch": 0, "step": 36, "train_loss": 8.933473587036133, "perplexity": 7581.554881123981, "lr": 0.0001540327868852459, "grad_norm": 2.043529, "tokens_per_sec": 107978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:49.435474+00:00", "epoch": 0, "step": 37, "train_loss": 8.871153831481934, "perplexity": 7123.495546564767, "lr": 0.00015831147540983607, "grad_norm": 1.593132, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:49.740316+00:00", "epoch": 0, "step": 38, "train_loss": 8.76101303100586, "perplexity": 6380.572022417468, "lr": 0.0001625901639344262, "grad_norm": 1.747585, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:50.046105+00:00", "epoch": 0, "step": 39, "train_loss": 8.749940872192383, "perplexity": 6310.31498196781, "lr": 0.00016686885245901638, "grad_norm": 1.585425, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:50.351206+00:00", "epoch": 0, "step": 40, "train_loss": 8.670637130737305, "perplexity": 5829.212136860332, "lr": 0.00017114754098360654, "grad_norm": 1.612467, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:50.655265+00:00", "epoch": 0, "step": 41, "train_loss": 8.625904083251953, "perplexity": 5574.199972007869, "lr": 0.00017542622950819672, "grad_norm": 1.559344, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:50.960150+00:00", "epoch": 0, "step": 42, "train_loss": 8.546354293823242, "perplexity": 5147.9522530734175, "lr": 0.00017970491803278687, "grad_norm": 1.538993, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:51.264868+00:00", "epoch": 0, "step": 43, "train_loss": 8.49260139465332, "perplexity": 4878.540589802637, "lr": 0.00018398360655737706, "grad_norm": 1.524956, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:51.570412+00:00", "epoch": 0, "step": 44, "train_loss": 8.439778327941895, "perplexity": 4627.529076960957, "lr": 0.0001882622950819672, "grad_norm": 1.495461, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:51.875311+00:00", "epoch": 0, "step": 45, "train_loss": 8.358461380004883, "perplexity": 4266.12576768719, "lr": 0.00019254098360655736, "grad_norm": 1.43945, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:52.180419+00:00", "epoch": 0, "step": 46, "train_loss": 8.277644157409668, "perplexity": 3934.9134174089377, "lr": 0.00019681967213114755, "grad_norm": 1.454227, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:52.484400+00:00", "epoch": 0, "step": 47, "train_loss": 8.258899688720703, "perplexity": 3861.8425329882466, "lr": 0.0002010983606557377, "grad_norm": 1.445024, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:52.789079+00:00", "epoch": 0, "step": 48, "train_loss": 8.112327575683594, "perplexity": 3335.33224035663, "lr": 0.00020537704918032788, "grad_norm": 1.440589, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:53.094100+00:00", "epoch": 0, "step": 49, "train_loss": 8.103015899658203, "perplexity": 3304.4188581649933, "lr": 0.00020965573770491804, "grad_norm": 1.371716, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:53.399578+00:00", "epoch": 0, "step": 50, "train_loss": 8.044576644897461, "perplexity": 3116.8452919500846, "lr": 0.0002139344262295082, "grad_norm": 1.35098, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:53.704149+00:00", "epoch": 0, "step": 51, "train_loss": 7.997970104217529, "perplexity": 2974.913090326938, "lr": 0.00021821311475409837, "grad_norm": 1.320078, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:54.007699+00:00", "epoch": 0, "step": 52, "train_loss": 7.88278865814209, "perplexity": 2651.2557061939433, "lr": 0.00022249180327868853, "grad_norm": 1.294531, "tokens_per_sec": 107949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:54.312245+00:00", "epoch": 0, "step": 53, "train_loss": 7.935197830200195, "perplexity": 2793.9114047355756, "lr": 0.0002267704918032787, "grad_norm": 1.254014, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:54.616622+00:00", "epoch": 0, "step": 54, "train_loss": 7.792810916900635, "perplexity": 2423.1192050439827, "lr": 0.0002310491803278688, "grad_norm": 1.222776, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:54.921651+00:00", "epoch": 0, "step": 55, "train_loss": 7.756931304931641, "perplexity": 2337.719837646585, "lr": 0.000235327868852459, "grad_norm": 1.20459, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:55.227340+00:00", "epoch": 0, "step": 56, "train_loss": 7.800248146057129, "perplexity": 2441.207678530305, "lr": 0.00023960655737704915, "grad_norm": 1.150545, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:55.532008+00:00", "epoch": 0, "step": 57, "train_loss": 7.701826095581055, "perplexity": 2212.38433064517, "lr": 0.00024388524590163933, "grad_norm": 1.100038, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:55.836650+00:00", "epoch": 0, "step": 58, "train_loss": 7.71302604675293, "perplexity": 2237.3022061442225, "lr": 0.0002481639344262295, "grad_norm": 1.062448, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:56.141547+00:00", "epoch": 0, "step": 59, "train_loss": 7.63408088684082, "perplexity": 2067.469938650646, "lr": 0.00025244262295081964, "grad_norm": 0.995213, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:56.447181+00:00", "epoch": 0, "step": 60, "train_loss": 7.593369007110596, "perplexity": 1984.9897057954631, "lr": 0.0002567213114754098, "grad_norm": 0.959199, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:56.752176+00:00", "epoch": 0, "step": 61, "train_loss": 7.509592533111572, "perplexity": 1825.4695728485706, "lr": 0.000261, "grad_norm": 0.946008, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:57.056698+00:00", "epoch": 0, "step": 62, "train_loss": 7.51112174987793, "perplexity": 1828.2632470481406, "lr": 0.00026527868852459013, "grad_norm": 0.921708, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:57.360818+00:00", "epoch": 0, "step": 63, "train_loss": 7.476842403411865, "perplexity": 1766.6535802020478, "lr": 0.0002695573770491803, "grad_norm": 0.960456, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:57.665950+00:00", "epoch": 0, "step": 64, "train_loss": 7.465190887451172, "perplexity": 1746.1888419476797, "lr": 0.0002738360655737705, "grad_norm": 0.87881, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:57.970770+00:00", "epoch": 0, "step": 65, "train_loss": 7.364450931549072, "perplexity": 1578.8482926946672, "lr": 0.0002781147540983607, "grad_norm": 0.88895, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:58.275934+00:00", "epoch": 0, "step": 66, "train_loss": 7.424074172973633, "perplexity": 1675.8473098331278, "lr": 0.0002823934426229508, "grad_norm": 0.853021, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:58.580849+00:00", "epoch": 0, "step": 67, "train_loss": 7.355688095092773, "perplexity": 1565.0735443934734, "lr": 0.000286672131147541, "grad_norm": 0.845793, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:58.885546+00:00", "epoch": 0, "step": 68, "train_loss": 7.302865982055664, "perplexity": 1484.5485259091702, "lr": 0.00029095081967213117, "grad_norm": 0.866788, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:59.189381+00:00", "epoch": 0, "step": 69, "train_loss": 7.270268440246582, "perplexity": 1436.9361327581064, "lr": 0.0002952295081967213, "grad_norm": 0.657188, "tokens_per_sec": 107845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:59.494215+00:00", "epoch": 0, "step": 70, "train_loss": 7.325019836425781, "perplexity": 1517.8040061715565, "lr": 0.0002995081967213115, "grad_norm": 0.589919, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:27:59.799035+00:00", "epoch": 0, "step": 71, "train_loss": 7.285610675811768, "perplexity": 1459.1519296244974, "lr": 0.00030378688524590166, "grad_norm": 0.726828, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:00.104669+00:00", "epoch": 0, "step": 72, "train_loss": 7.320457935333252, "perplexity": 1510.8957038940962, "lr": 0.0003080655737704918, "grad_norm": 1.016021, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:00.409222+00:00", "epoch": 0, "step": 73, "train_loss": 7.208889961242676, "perplexity": 1351.3913377221486, "lr": 0.00031234426229508197, "grad_norm": 2.192906, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:00.713708+00:00", "epoch": 0, "step": 74, "train_loss": 7.266481399536133, "perplexity": 1431.504688167261, "lr": 0.00031662295081967215, "grad_norm": 1.497879, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:01.020483+00:00", "epoch": 0, "step": 75, "train_loss": 7.2602105140686035, "perplexity": 1422.555973728119, "lr": 0.0003209016393442623, "grad_norm": 0.89688, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:01.327516+00:00", "epoch": 0, "step": 76, "train_loss": 7.173984050750732, "perplexity": 1305.0335793791035, "lr": 0.0003251803278688524, "grad_norm": 1.038913, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:01.634368+00:00", "epoch": 0, "step": 77, "train_loss": 7.057409286499023, "perplexity": 1161.4323263477079, "lr": 0.0003294590163934426, "grad_norm": 1.00398, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:01.939962+00:00", "epoch": 0, "step": 78, "train_loss": 7.095025539398193, "perplexity": 1205.9531624154838, "lr": 0.00033373770491803277, "grad_norm": 0.921667, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:02.245818+00:00", "epoch": 0, "step": 79, "train_loss": 7.099234580993652, "perplexity": 1211.0397667934346, "lr": 0.00033801639344262295, "grad_norm": 1.089738, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:02.551458+00:00", "epoch": 0, "step": 80, "train_loss": 7.072169780731201, "perplexity": 1178.7027882165887, "lr": 0.0003422950819672131, "grad_norm": 0.686696, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:02.857067+00:00", "epoch": 0, "step": 81, "train_loss": 7.038724899291992, "perplexity": 1139.9331499700663, "lr": 0.00034657377049180326, "grad_norm": 0.87763, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:03.165392+00:00", "epoch": 0, "step": 82, "train_loss": 7.1029744148254395, "perplexity": 1215.577333869693, "lr": 0.00035085245901639344, "grad_norm": 0.913861, "tokens_per_sec": 106277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:03.471861+00:00", "epoch": 0, "step": 83, "train_loss": 7.099661827087402, "perplexity": 1211.557289350046, "lr": 0.00035513114754098357, "grad_norm": 0.625135, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:03.778254+00:00", "epoch": 0, "step": 84, "train_loss": 7.008434295654297, "perplexity": 1105.9216023893466, "lr": 0.00035940983606557375, "grad_norm": 0.533216, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:04.083991+00:00", "epoch": 0, "step": 85, "train_loss": 6.945107460021973, "perplexity": 1038.0585410061244, "lr": 0.00036368852459016393, "grad_norm": 0.817739, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:04.390148+00:00", "epoch": 0, "step": 86, "train_loss": 7.024871349334717, "perplexity": 1124.249914201628, "lr": 0.0003679672131147541, "grad_norm": 0.56518, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:04.695985+00:00", "epoch": 0, "step": 87, "train_loss": 7.032032012939453, "perplexity": 1132.3291815807404, "lr": 0.00037224590163934424, "grad_norm": 0.473061, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:05.004360+00:00", "epoch": 0, "step": 88, "train_loss": 6.966794490814209, "perplexity": 1060.8168364962826, "lr": 0.0003765245901639344, "grad_norm": 0.541486, "tokens_per_sec": 106261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:05.311353+00:00", "epoch": 0, "step": 89, "train_loss": 6.9189629554748535, "perplexity": 1011.2707177942743, "lr": 0.0003808032786885246, "grad_norm": 0.434592, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:05.618262+00:00", "epoch": 0, "step": 90, "train_loss": 6.853714466094971, "perplexity": 947.393439851192, "lr": 0.00038508196721311473, "grad_norm": 0.811723, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:05.923370+00:00", "epoch": 0, "step": 91, "train_loss": 6.944664001464844, "perplexity": 1037.5983071181847, "lr": 0.0003893606557377049, "grad_norm": 1.152428, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:06.230009+00:00", "epoch": 0, "step": 92, "train_loss": 6.859500408172607, "perplexity": 952.8908920523429, "lr": 0.0003936393442622951, "grad_norm": 0.71565, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:06.536809+00:00", "epoch": 0, "step": 93, "train_loss": 6.902803421020508, "perplexity": 995.0603822747017, "lr": 0.0003979180327868853, "grad_norm": 1.847557, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:06.844482+00:00", "epoch": 0, "step": 94, "train_loss": 6.8546648025512695, "perplexity": 948.2942103252577, "lr": 0.0004021967213114754, "grad_norm": 1.648718, "tokens_per_sec": 106503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:07.149838+00:00", "epoch": 0, "step": 95, "train_loss": 6.91689395904541, "perplexity": 1009.1805652946044, "lr": 0.0004064754098360656, "grad_norm": 1.120593, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:07.456503+00:00", "epoch": 0, "step": 96, "train_loss": 6.863960266113281, "perplexity": 957.1501408262706, "lr": 0.00041075409836065577, "grad_norm": 1.730887, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:07.762592+00:00", "epoch": 0, "step": 97, "train_loss": 6.889654159545898, "perplexity": 982.0617218096111, "lr": 0.0004150327868852459, "grad_norm": 1.334174, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:08.070198+00:00", "epoch": 0, "step": 98, "train_loss": 6.814297676086426, "perplexity": 910.7766308994477, "lr": 0.0004193114754098361, "grad_norm": 1.42969, "tokens_per_sec": 106526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:08.376974+00:00", "epoch": 0, "step": 99, "train_loss": 6.807150840759277, "perplexity": 904.2906649708966, "lr": 0.00042359016393442626, "grad_norm": 0.998837, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:08.684547+00:00", "epoch": 0, "step": 100, "train_loss": 6.771239280700684, "perplexity": 872.3923633694122, "lr": 0.0004278688524590164, "grad_norm": 1.082256, "tokens_per_sec": 106591} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:28:14.545661+00:00", "step": 100, "epoch": 0, "val_loss": 6.785474085807801, "val_ppl": 884.8995058140137, "eval_train_loss": 6.771239280700684, "eval_train_ppl": 872.3923633694122} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:28:15.515847+00:00", "step": 100, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_6p7855_epoch_0000_step_0000100.pt", "val_loss": 6.785474085807801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:18.442408+00:00", "epoch": 0, "step": 101, "train_loss": 6.725833892822266, "perplexity": 833.666875594194, "lr": 0.00043214754098360657, "grad_norm": 1.015533, "tokens_per_sec": 3358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:18.750371+00:00", "epoch": 0, "step": 102, "train_loss": 6.848966598510742, "perplexity": 942.9060025573837, "lr": 0.00043642622950819675, "grad_norm": 0.588392, "tokens_per_sec": 106401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:19.057641+00:00", "epoch": 0, "step": 103, "train_loss": 6.685342788696289, "perplexity": 800.5850626548853, "lr": 0.00044070491803278693, "grad_norm": 0.931246, "tokens_per_sec": 106642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:19.364173+00:00", "epoch": 0, "step": 104, "train_loss": 6.667148590087891, "perplexity": 786.1507674176823, "lr": 0.00044498360655737706, "grad_norm": 0.729425, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:19.671988+00:00", "epoch": 0, "step": 105, "train_loss": 6.708486080169678, "perplexity": 819.3293011677958, "lr": 0.00044926229508196724, "grad_norm": 0.585896, "tokens_per_sec": 106453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:19.978433+00:00", "epoch": 0, "step": 106, "train_loss": 6.806981563568115, "perplexity": 904.1376021425245, "lr": 0.0004535409836065574, "grad_norm": 0.748112, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:20.285506+00:00", "epoch": 0, "step": 107, "train_loss": 6.822669982910156, "perplexity": 918.4339422527181, "lr": 0.0004578196721311475, "grad_norm": 0.551387, "tokens_per_sec": 106711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:20.592888+00:00", "epoch": 0, "step": 108, "train_loss": 6.70697021484375, "perplexity": 818.088249161137, "lr": 0.0004620983606557376, "grad_norm": 0.516456, "tokens_per_sec": 106605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:20.985706+00:00", "epoch": 0, "step": 109, "train_loss": 6.79592752456665, "perplexity": 894.1982659393273, "lr": 0.0004663770491803278, "grad_norm": 0.456011, "tokens_per_sec": 83417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:21.292399+00:00", "epoch": 0, "step": 110, "train_loss": 6.644448757171631, "perplexity": 768.5062972631058, "lr": 0.000470655737704918, "grad_norm": 0.444469, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:21.600160+00:00", "epoch": 0, "step": 111, "train_loss": 6.683117389678955, "perplexity": 798.8054223824751, "lr": 0.00047493442622950817, "grad_norm": 0.428392, "tokens_per_sec": 106419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:21.907503+00:00", "epoch": 0, "step": 112, "train_loss": 6.788802623748779, "perplexity": 887.8498348101385, "lr": 0.0004792131147540983, "grad_norm": 0.400437, "tokens_per_sec": 106617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:22.213434+00:00", "epoch": 0, "step": 113, "train_loss": 6.541093349456787, "perplexity": 693.0439031255733, "lr": 0.0004834918032786885, "grad_norm": 0.483718, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:22.519420+00:00", "epoch": 0, "step": 114, "train_loss": 6.5053863525390625, "perplexity": 668.7339865241433, "lr": 0.00048777049180327866, "grad_norm": 0.479501, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:22.826327+00:00", "epoch": 0, "step": 115, "train_loss": 6.532106399536133, "perplexity": 686.8434555629129, "lr": 0.0004920491803278688, "grad_norm": 0.573113, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:23.133763+00:00", "epoch": 0, "step": 116, "train_loss": 6.6530585289001465, "perplexity": 775.151526960013, "lr": 0.000496327868852459, "grad_norm": 0.761126, "tokens_per_sec": 106576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:23.440837+00:00", "epoch": 0, "step": 117, "train_loss": 6.548292636871338, "perplexity": 698.0513286952397, "lr": 0.0005006065573770491, "grad_norm": 1.223415, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:23.747999+00:00", "epoch": 0, "step": 118, "train_loss": 6.585346221923828, "perplexity": 724.4018080427708, "lr": 0.0005048852459016393, "grad_norm": 1.279161, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:24.054692+00:00", "epoch": 0, "step": 119, "train_loss": 6.50838565826416, "perplexity": 670.7427351182079, "lr": 0.0005091639344262295, "grad_norm": 0.726596, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:24.362401+00:00", "epoch": 0, "step": 120, "train_loss": 6.51278018951416, "perplexity": 673.6968211860813, "lr": 0.0005134426229508196, "grad_norm": 0.814118, "tokens_per_sec": 106555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:24.670089+00:00", "epoch": 0, "step": 121, "train_loss": 6.60355281829834, "perplexity": 737.7114937565815, "lr": 0.0005177213114754098, "grad_norm": 0.668717, "tokens_per_sec": 106433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:24.977465+00:00", "epoch": 0, "step": 122, "train_loss": 6.554230690002441, "perplexity": 702.2087257796809, "lr": 0.000522, "grad_norm": 0.61099, "tokens_per_sec": 106605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:25.284180+00:00", "epoch": 0, "step": 123, "train_loss": 6.487945556640625, "perplexity": 657.1718530405856, "lr": 0.0005262786885245901, "grad_norm": 0.473037, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:25.590274+00:00", "epoch": 0, "step": 124, "train_loss": 6.474989414215088, "perplexity": 648.7123604450135, "lr": 0.0005305573770491803, "grad_norm": 0.586231, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:25.896586+00:00", "epoch": 0, "step": 125, "train_loss": 6.356937885284424, "perplexity": 576.4784079285281, "lr": 0.0005348360655737705, "grad_norm": 0.646042, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:26.203654+00:00", "epoch": 0, "step": 126, "train_loss": 6.5539021492004395, "perplexity": 701.9780594554649, "lr": 0.0005391147540983606, "grad_norm": 0.709659, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:26.510893+00:00", "epoch": 0, "step": 127, "train_loss": 6.366359233856201, "perplexity": 581.93527712372, "lr": 0.0005433934426229507, "grad_norm": 0.711121, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:26.819564+00:00", "epoch": 0, "step": 128, "train_loss": 6.4569478034973145, "perplexity": 637.1134903370552, "lr": 0.000547672131147541, "grad_norm": 0.468847, "tokens_per_sec": 106158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:27.125855+00:00", "epoch": 0, "step": 129, "train_loss": 6.443160057067871, "perplexity": 628.3894119805376, "lr": 0.0005519508196721311, "grad_norm": 0.46702, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:27.432937+00:00", "epoch": 0, "step": 130, "train_loss": 6.362724781036377, "perplexity": 579.8240996262484, "lr": 0.0005562295081967213, "grad_norm": 0.526456, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:27.739751+00:00", "epoch": 0, "step": 131, "train_loss": 6.496746063232422, "perplexity": 662.980821707255, "lr": 0.0005605081967213115, "grad_norm": 0.551893, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:28.047460+00:00", "epoch": 0, "step": 132, "train_loss": 6.425428867340088, "perplexity": 617.3455201087676, "lr": 0.0005647868852459016, "grad_norm": 0.498765, "tokens_per_sec": 106489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:28.355919+00:00", "epoch": 0, "step": 133, "train_loss": 6.37173318862915, "perplexity": 585.0709890242836, "lr": 0.0005690655737704918, "grad_norm": 0.477964, "tokens_per_sec": 106231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:28.664219+00:00", "epoch": 0, "step": 134, "train_loss": 6.331409931182861, "perplexity": 561.9483440586238, "lr": 0.000573344262295082, "grad_norm": 0.532881, "tokens_per_sec": 106287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:28.971232+00:00", "epoch": 0, "step": 135, "train_loss": 6.405792236328125, "perplexity": 605.3411820363333, "lr": 0.0005776229508196721, "grad_norm": 1.052963, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:29.277699+00:00", "epoch": 0, "step": 136, "train_loss": 6.43749475479126, "perplexity": 624.839461280962, "lr": 0.0005819016393442623, "grad_norm": 1.456434, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:29.585154+00:00", "epoch": 0, "step": 137, "train_loss": 6.262004852294922, "perplexity": 524.2689690255554, "lr": 0.0005861803278688525, "grad_norm": 1.016246, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:29.893824+00:00", "epoch": 0, "step": 138, "train_loss": 6.38751745223999, "perplexity": 594.3791718973031, "lr": 0.0005904590163934426, "grad_norm": 1.424461, "tokens_per_sec": 106159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:30.202459+00:00", "epoch": 0, "step": 139, "train_loss": 6.349056243896484, "perplexity": 571.9526703977745, "lr": 0.0005947377049180328, "grad_norm": 1.359715, "tokens_per_sec": 106171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:30.509778+00:00", "epoch": 0, "step": 140, "train_loss": 6.387722015380859, "perplexity": 594.5007724046403, "lr": 0.000599016393442623, "grad_norm": 1.003077, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:30.815898+00:00", "epoch": 0, "step": 141, "train_loss": 6.266111850738525, "perplexity": 526.426568461372, "lr": 0.0006032950819672131, "grad_norm": 0.932018, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:31.123229+00:00", "epoch": 0, "step": 142, "train_loss": 6.333727836608887, "perplexity": 563.2523979273765, "lr": 0.0006075737704918033, "grad_norm": 1.020456, "tokens_per_sec": 106622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:31.431337+00:00", "epoch": 0, "step": 143, "train_loss": 6.376130104064941, "perplexity": 587.6491605343934, "lr": 0.0006118524590163934, "grad_norm": 1.262024, "tokens_per_sec": 106352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:31.739404+00:00", "epoch": 0, "step": 144, "train_loss": 6.275888919830322, "perplexity": 531.5987204356788, "lr": 0.0006161311475409836, "grad_norm": 0.931131, "tokens_per_sec": 106368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:32.047629+00:00", "epoch": 0, "step": 145, "train_loss": 6.475006580352783, "perplexity": 648.7234964262982, "lr": 0.0006204098360655738, "grad_norm": 0.830157, "tokens_per_sec": 106311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:32.354875+00:00", "epoch": 0, "step": 146, "train_loss": 6.360959529876709, "perplexity": 578.8014673289572, "lr": 0.0006246885245901639, "grad_norm": 1.34333, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:32.661227+00:00", "epoch": 0, "step": 147, "train_loss": 6.336133003234863, "perplexity": 564.6087442624896, "lr": 0.0006289672131147542, "grad_norm": 1.022, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:32.969252+00:00", "epoch": 0, "step": 148, "train_loss": 6.201017379760742, "perplexity": 493.2506090940932, "lr": 0.0006332459016393443, "grad_norm": 0.62446, "tokens_per_sec": 106381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:33.278189+00:00", "epoch": 0, "step": 149, "train_loss": 6.267780780792236, "perplexity": 527.3058711257751, "lr": 0.0006375245901639344, "grad_norm": 0.663116, "tokens_per_sec": 106072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:33.586443+00:00", "epoch": 0, "step": 150, "train_loss": 6.336348533630371, "perplexity": 564.7304477233741, "lr": 0.0006418032786885245, "grad_norm": 0.529942, "tokens_per_sec": 106359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:33.893851+00:00", "epoch": 0, "step": 151, "train_loss": 6.38627815246582, "perplexity": 593.6430141780669, "lr": 0.0006460819672131147, "grad_norm": 0.397419, "tokens_per_sec": 106532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:34.201778+00:00", "epoch": 0, "step": 152, "train_loss": 6.189460277557373, "perplexity": 487.5828757644489, "lr": 0.0006503606557377048, "grad_norm": 0.484852, "tokens_per_sec": 106415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:34.508797+00:00", "epoch": 0, "step": 153, "train_loss": 6.2576704025268555, "perplexity": 522.0014692463488, "lr": 0.000654639344262295, "grad_norm": 0.375337, "tokens_per_sec": 106730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:34.816543+00:00", "epoch": 0, "step": 154, "train_loss": 6.104379653930664, "perplexity": 447.8147551483071, "lr": 0.0006589180327868852, "grad_norm": 0.412102, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:35.126516+00:00", "epoch": 0, "step": 155, "train_loss": 6.289748668670654, "perplexity": 539.0178399887634, "lr": 0.0006631967213114754, "grad_norm": 0.390686, "tokens_per_sec": 105712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:35.436514+00:00", "epoch": 0, "step": 156, "train_loss": 6.284816265106201, "perplexity": 536.3657324825948, "lr": 0.0006674754098360655, "grad_norm": 0.427617, "tokens_per_sec": 105704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:35.757371+00:00", "epoch": 0, "step": 157, "train_loss": 6.122264862060547, "perplexity": 455.8960678070387, "lr": 0.0006717540983606557, "grad_norm": 0.415847, "tokens_per_sec": 102126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:36.065683+00:00", "epoch": 0, "step": 158, "train_loss": 6.192530155181885, "perplexity": 489.0819954044459, "lr": 0.0006760327868852459, "grad_norm": 0.439991, "tokens_per_sec": 106282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:36.373496+00:00", "epoch": 0, "step": 159, "train_loss": 6.27088737487793, "perplexity": 528.9465435585032, "lr": 0.000680311475409836, "grad_norm": 0.503619, "tokens_per_sec": 106455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:36.698326+00:00", "epoch": 0, "step": 160, "train_loss": 6.0753607749938965, "perplexity": 435.0064136460484, "lr": 0.0006845901639344261, "grad_norm": 0.463327, "tokens_per_sec": 100924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:37.006627+00:00", "epoch": 0, "step": 161, "train_loss": 6.316779136657715, "perplexity": 553.786446420825, "lr": 0.0006888688524590164, "grad_norm": 0.424349, "tokens_per_sec": 106236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:37.315241+00:00", "epoch": 0, "step": 162, "train_loss": 6.159759998321533, "perplexity": 473.3144649360696, "lr": 0.0006931475409836065, "grad_norm": 0.412972, "tokens_per_sec": 106177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:37.632533+00:00", "epoch": 0, "step": 163, "train_loss": 6.121644973754883, "perplexity": 455.61355073956673, "lr": 0.0006974262295081966, "grad_norm": 0.569191, "tokens_per_sec": 103273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:37.954542+00:00", "epoch": 0, "step": 164, "train_loss": 6.164870738983154, "perplexity": 475.7396443705874, "lr": 0.0007017049180327869, "grad_norm": 0.637422, "tokens_per_sec": 101762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:38.275195+00:00", "epoch": 0, "step": 165, "train_loss": 6.165546894073486, "perplexity": 476.06142692799887, "lr": 0.000705983606557377, "grad_norm": 0.684936, "tokens_per_sec": 102191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:38.583409+00:00", "epoch": 0, "step": 166, "train_loss": 6.209374904632568, "perplexity": 497.39023775694835, "lr": 0.0007102622950819671, "grad_norm": 0.682883, "tokens_per_sec": 106316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:38.904730+00:00", "epoch": 0, "step": 167, "train_loss": 6.110626220703125, "perplexity": 450.6208149136093, "lr": 0.0007145409836065574, "grad_norm": 0.830094, "tokens_per_sec": 101979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:39.213542+00:00", "epoch": 0, "step": 168, "train_loss": 6.348473072052002, "perplexity": 571.619220942639, "lr": 0.0007188196721311475, "grad_norm": 1.490445, "tokens_per_sec": 106110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:39.533239+00:00", "epoch": 0, "step": 169, "train_loss": 6.210010051727295, "perplexity": 497.7062540692026, "lr": 0.0007230983606557376, "grad_norm": 1.580899, "tokens_per_sec": 102497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:39.843187+00:00", "epoch": 0, "step": 170, "train_loss": 6.1720967292785645, "perplexity": 479.18978475201635, "lr": 0.0007273770491803279, "grad_norm": 1.06335, "tokens_per_sec": 105780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:40.162560+00:00", "epoch": 0, "step": 171, "train_loss": 6.036995887756348, "perplexity": 418.6335218859867, "lr": 0.000731655737704918, "grad_norm": 0.885438, "tokens_per_sec": 102547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:40.483668+00:00", "epoch": 0, "step": 172, "train_loss": 6.242539405822754, "perplexity": 514.1625218374437, "lr": 0.0007359344262295082, "grad_norm": 0.830745, "tokens_per_sec": 102046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:40.807780+00:00", "epoch": 0, "step": 173, "train_loss": 6.120323181152344, "perplexity": 455.01172195262495, "lr": 0.0007402131147540983, "grad_norm": 0.948741, "tokens_per_sec": 101101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:41.116040+00:00", "epoch": 0, "step": 174, "train_loss": 6.114569187164307, "perplexity": 452.40110517941713, "lr": 0.0007444918032786885, "grad_norm": 0.910505, "tokens_per_sec": 106301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:41.434505+00:00", "epoch": 0, "step": 175, "train_loss": 6.203191757202148, "perplexity": 494.32428896108587, "lr": 0.0007487704918032787, "grad_norm": 1.014194, "tokens_per_sec": 102894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:41.754535+00:00", "epoch": 0, "step": 176, "train_loss": 6.182302951812744, "perplexity": 484.1055453329214, "lr": 0.0007530491803278688, "grad_norm": 0.88443, "tokens_per_sec": 102391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:42.062870+00:00", "epoch": 0, "step": 177, "train_loss": 5.9922356605529785, "perplexity": 400.30856433176126, "lr": 0.000757327868852459, "grad_norm": 0.673865, "tokens_per_sec": 106273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:42.381168+00:00", "epoch": 0, "step": 178, "train_loss": 6.039786338806152, "perplexity": 419.80332962316066, "lr": 0.0007616065573770492, "grad_norm": 0.618465, "tokens_per_sec": 102949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:42.700446+00:00", "epoch": 0, "step": 179, "train_loss": 6.018190383911133, "perplexity": 410.8344699720406, "lr": 0.0007658852459016393, "grad_norm": 0.531957, "tokens_per_sec": 102632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:43.019954+00:00", "epoch": 0, "step": 180, "train_loss": 5.97731351852417, "perplexity": 394.3794506371787, "lr": 0.0007701639344262295, "grad_norm": 0.60478, "tokens_per_sec": 102607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:43.339839+00:00", "epoch": 0, "step": 181, "train_loss": 6.093380451202393, "perplexity": 442.91613969108363, "lr": 0.0007744426229508197, "grad_norm": 0.905373, "tokens_per_sec": 102388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:43.660757+00:00", "epoch": 0, "step": 182, "train_loss": 5.954471111297607, "perplexity": 385.47298444908205, "lr": 0.0007787213114754098, "grad_norm": 0.939298, "tokens_per_sec": 102107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:43.970483+00:00", "epoch": 0, "step": 183, "train_loss": 6.001686096191406, "perplexity": 404.1095870253513, "lr": 0.000783, "grad_norm": 0.623524, "tokens_per_sec": 105797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:44.288500+00:00", "epoch": 0, "step": 184, "train_loss": 6.06959867477417, "perplexity": 432.50707074176455, "lr": 0.0007872786885245902, "grad_norm": 0.543238, "tokens_per_sec": 103039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:44.609642+00:00", "epoch": 0, "step": 185, "train_loss": 6.100262641906738, "perplexity": 445.97488639653824, "lr": 0.0007915573770491803, "grad_norm": 0.463595, "tokens_per_sec": 102036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:44.928793+00:00", "epoch": 0, "step": 186, "train_loss": 6.04084587097168, "perplexity": 420.24836047470984, "lr": 0.0007958360655737705, "grad_norm": 0.489347, "tokens_per_sec": 102672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:45.246459+00:00", "epoch": 0, "step": 187, "train_loss": 6.041751384735107, "perplexity": 420.62907349360296, "lr": 0.0008001147540983607, "grad_norm": 0.631166, "tokens_per_sec": 103152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:45.567365+00:00", "epoch": 0, "step": 188, "train_loss": 5.990536689758301, "perplexity": 399.6290291907381, "lr": 0.0008043934426229508, "grad_norm": 1.090023, "tokens_per_sec": 102111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:45.887873+00:00", "epoch": 0, "step": 189, "train_loss": 6.023647308349609, "perplexity": 413.08249069253833, "lr": 0.000808672131147541, "grad_norm": 1.428311, "tokens_per_sec": 102238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:46.206304+00:00", "epoch": 0, "step": 190, "train_loss": 6.06870698928833, "perplexity": 432.1215823570017, "lr": 0.0008129508196721312, "grad_norm": 2.00682, "tokens_per_sec": 102963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:46.525966+00:00", "epoch": 0, "step": 191, "train_loss": 6.038980960845947, "perplexity": 419.4653653866031, "lr": 0.0008172295081967213, "grad_norm": 1.198126, "tokens_per_sec": 102450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:46.835474+00:00", "epoch": 0, "step": 192, "train_loss": 6.123938083648682, "perplexity": 456.6595214851088, "lr": 0.0008215081967213115, "grad_norm": 1.184506, "tokens_per_sec": 105872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:47.159007+00:00", "epoch": 0, "step": 193, "train_loss": 6.102535724639893, "perplexity": 446.9897772386643, "lr": 0.0008257868852459017, "grad_norm": 1.087739, "tokens_per_sec": 101282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:47.478658+00:00", "epoch": 0, "step": 194, "train_loss": 6.128190517425537, "perplexity": 458.60557064882806, "lr": 0.0008300655737704918, "grad_norm": 0.76499, "tokens_per_sec": 102513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:47.800385+00:00", "epoch": 0, "step": 195, "train_loss": 6.024497032165527, "perplexity": 413.4336458941587, "lr": 0.000834344262295082, "grad_norm": 0.730426, "tokens_per_sec": 101849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:48.119258+00:00", "epoch": 0, "step": 196, "train_loss": 5.9736151695251465, "perplexity": 392.92359158751907, "lr": 0.0008386229508196722, "grad_norm": 0.563324, "tokens_per_sec": 102762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:48.437847+00:00", "epoch": 0, "step": 197, "train_loss": 5.89792537689209, "perplexity": 364.28093772804067, "lr": 0.0008429016393442623, "grad_norm": 0.515608, "tokens_per_sec": 102855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:48.757938+00:00", "epoch": 0, "step": 198, "train_loss": 5.952120304107666, "perplexity": 384.567876070178, "lr": 0.0008471803278688525, "grad_norm": 0.480206, "tokens_per_sec": 102371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:49.078852+00:00", "epoch": 0, "step": 199, "train_loss": 5.965709686279297, "perplexity": 389.8295866332218, "lr": 0.0008514590163934426, "grad_norm": 0.450325, "tokens_per_sec": 102109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:49.397741+00:00", "epoch": 0, "step": 200, "train_loss": 5.9640960693359375, "perplexity": 389.2010582456136, "lr": 0.0008557377049180328, "grad_norm": 0.39048, "tokens_per_sec": 102806} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:28:55.131309+00:00", "step": 200, "epoch": 0, "val_loss": 5.9414715051651, "val_ppl": 380.4944172929635, "eval_train_loss": 5.9640960693359375, "eval_train_ppl": 389.2010582456136} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:28:55.990489+00:00", "step": 200, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_5p9415_epoch_0000_step_0000200.pt", "val_loss": 5.9414715051651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:56.996509+00:00", "epoch": 0, "step": 201, "train_loss": 5.917051315307617, "perplexity": 371.3152065778415, "lr": 0.000860016393442623, "grad_norm": 0.36709, "tokens_per_sec": 4312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:57.307375+00:00", "epoch": 0, "step": 202, "train_loss": 5.795741558074951, "perplexity": 328.8959890494693, "lr": 0.0008642950819672131, "grad_norm": 0.341079, "tokens_per_sec": 105406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:57.613800+00:00", "epoch": 0, "step": 203, "train_loss": 5.909712314605713, "perplexity": 368.6000992897917, "lr": 0.0008685737704918034, "grad_norm": 0.365966, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:57.921691+00:00", "epoch": 0, "step": 204, "train_loss": 5.9002814292907715, "perplexity": 365.14021455830147, "lr": 0.0008728524590163935, "grad_norm": 0.335596, "tokens_per_sec": 106427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:58.227095+00:00", "epoch": 0, "step": 205, "train_loss": 5.928938865661621, "perplexity": 375.7555750548314, "lr": 0.0008771311475409836, "grad_norm": 0.365707, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:58.533088+00:00", "epoch": 0, "step": 206, "train_loss": 5.932735443115234, "perplexity": 377.1848717000398, "lr": 0.0008814098360655739, "grad_norm": 0.437492, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:58.840917+00:00", "epoch": 0, "step": 207, "train_loss": 5.840609550476074, "perplexity": 343.9889554046253, "lr": 0.000885688524590164, "grad_norm": 0.646556, "tokens_per_sec": 106449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:59.148057+00:00", "epoch": 0, "step": 208, "train_loss": 5.954983234405518, "perplexity": 385.6704446295385, "lr": 0.0008899672131147541, "grad_norm": 0.803797, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:59.454834+00:00", "epoch": 0, "step": 209, "train_loss": 5.914871692657471, "perplexity": 370.5067609165386, "lr": 0.0008942459016393444, "grad_norm": 0.557969, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:28:59.761241+00:00", "epoch": 0, "step": 210, "train_loss": 5.899311542510986, "perplexity": 364.78624157603934, "lr": 0.0008985245901639345, "grad_norm": 0.370159, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:00.068440+00:00", "epoch": 0, "step": 211, "train_loss": 5.74595832824707, "perplexity": 312.9233674848546, "lr": 0.0009028032786885246, "grad_norm": 0.446432, "tokens_per_sec": 106603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:00.375560+00:00", "epoch": 0, "step": 212, "train_loss": 5.829489231109619, "perplexity": 340.18487883656286, "lr": 0.0009070819672131148, "grad_norm": 0.417486, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:00.681636+00:00", "epoch": 0, "step": 213, "train_loss": 5.915893077850342, "perplexity": 370.8853843632726, "lr": 0.000911360655737705, "grad_norm": 0.472851, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:00.989524+00:00", "epoch": 0, "step": 214, "train_loss": 5.6843156814575195, "perplexity": 294.2164383812006, "lr": 0.000915639344262295, "grad_norm": 0.496098, "tokens_per_sec": 106429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:01.368498+00:00", "epoch": 0, "step": 215, "train_loss": 5.759499549865723, "perplexity": 317.18955163740037, "lr": 0.0009199180327868851, "grad_norm": 0.479528, "tokens_per_sec": 86465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:01.675200+00:00", "epoch": 0, "step": 216, "train_loss": 5.675970077514648, "perplexity": 291.7712420252918, "lr": 0.0009241967213114752, "grad_norm": 0.358081, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:01.983279+00:00", "epoch": 0, "step": 217, "train_loss": 5.897132873535156, "perplexity": 363.9923582272396, "lr": 0.0009284754098360655, "grad_norm": 0.431653, "tokens_per_sec": 106362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:02.290833+00:00", "epoch": 0, "step": 218, "train_loss": 5.702713489532471, "perplexity": 299.6794758111984, "lr": 0.0009327540983606556, "grad_norm": 0.351647, "tokens_per_sec": 106545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:02.598106+00:00", "epoch": 0, "step": 219, "train_loss": 5.756419658660889, "perplexity": 316.21414517098236, "lr": 0.0009370327868852458, "grad_norm": 0.343929, "tokens_per_sec": 106642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:02.905228+00:00", "epoch": 0, "step": 220, "train_loss": 5.789309024810791, "perplexity": 326.7871445372483, "lr": 0.000941311475409836, "grad_norm": 0.306056, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:03.212876+00:00", "epoch": 0, "step": 221, "train_loss": 5.772482872009277, "perplexity": 321.3345756318798, "lr": 0.0009455901639344261, "grad_norm": 0.374727, "tokens_per_sec": 106457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:03.519571+00:00", "epoch": 0, "step": 222, "train_loss": 5.7875542640686035, "perplexity": 326.214214109887, "lr": 0.0009498688524590163, "grad_norm": 0.503075, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:03.826784+00:00", "epoch": 0, "step": 223, "train_loss": 5.804753303527832, "perplexity": 331.87331126885584, "lr": 0.0009541475409836065, "grad_norm": 0.738292, "tokens_per_sec": 106663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:04.132639+00:00", "epoch": 0, "step": 224, "train_loss": 5.905909538269043, "perplexity": 367.20105736226054, "lr": 0.0009584262295081966, "grad_norm": 0.807664, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:04.438908+00:00", "epoch": 0, "step": 225, "train_loss": 5.70789909362793, "perplexity": 301.23753116599994, "lr": 0.0009627049180327868, "grad_norm": 0.686956, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:04.745849+00:00", "epoch": 0, "step": 226, "train_loss": 5.858423709869385, "perplexity": 350.1717365047636, "lr": 0.000966983606557377, "grad_norm": 1.026547, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:05.051835+00:00", "epoch": 0, "step": 227, "train_loss": 5.686145782470703, "perplexity": 294.75537718898545, "lr": 0.0009712622950819671, "grad_norm": 0.915425, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:05.359070+00:00", "epoch": 0, "step": 228, "train_loss": 5.722306728363037, "perplexity": 305.6090676137195, "lr": 0.0009755409836065573, "grad_norm": 0.644443, "tokens_per_sec": 106655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:05.665869+00:00", "epoch": 0, "step": 229, "train_loss": 5.741686820983887, "perplexity": 311.58956375201234, "lr": 0.0009798196721311474, "grad_norm": 0.579221, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:05.971996+00:00", "epoch": 0, "step": 230, "train_loss": 5.767723083496094, "perplexity": 319.80872525320297, "lr": 0.0009840983606557376, "grad_norm": 0.572385, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:06.277642+00:00", "epoch": 0, "step": 231, "train_loss": 5.708439350128174, "perplexity": 301.4003206704143, "lr": 0.0009883770491803277, "grad_norm": 0.654091, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:06.584009+00:00", "epoch": 0, "step": 232, "train_loss": 5.8420939445495605, "perplexity": 344.49994973602304, "lr": 0.000992655737704918, "grad_norm": 1.258792, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:06.890869+00:00", "epoch": 0, "step": 233, "train_loss": 5.895840644836426, "perplexity": 363.5223006321877, "lr": 0.0009969344262295082, "grad_norm": 1.536317, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:07.199474+00:00", "epoch": 0, "step": 234, "train_loss": 5.774986743927002, "perplexity": 322.1401643766379, "lr": 0.0010012131147540983, "grad_norm": 1.406694, "tokens_per_sec": 106181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:07.506524+00:00", "epoch": 0, "step": 235, "train_loss": 5.916654586791992, "perplexity": 371.16792446454576, "lr": 0.0010054918032786884, "grad_norm": 1.010735, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:07.813042+00:00", "epoch": 0, "step": 236, "train_loss": 5.698245048522949, "perplexity": 298.3433631452622, "lr": 0.0010097704918032786, "grad_norm": 0.716267, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:08.119891+00:00", "epoch": 0, "step": 237, "train_loss": 5.725967884063721, "perplexity": 306.7300006969803, "lr": 0.0010140491803278687, "grad_norm": 0.61352, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:08.426863+00:00", "epoch": 0, "step": 238, "train_loss": 5.840866565704346, "perplexity": 344.07737716690445, "lr": 0.001018327868852459, "grad_norm": 0.581431, "tokens_per_sec": 106747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:08.733877+00:00", "epoch": 0, "step": 239, "train_loss": 5.75484561920166, "perplexity": 315.7168031495848, "lr": 0.0010226065573770492, "grad_norm": 0.712832, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:09.041459+00:00", "epoch": 0, "step": 240, "train_loss": 5.800710678100586, "perplexity": 330.5343800043758, "lr": 0.0010268852459016393, "grad_norm": 0.715425, "tokens_per_sec": 106590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:09.349725+00:00", "epoch": 0, "step": 241, "train_loss": 5.732552528381348, "perplexity": 308.75637280135936, "lr": 0.0010311639344262294, "grad_norm": 0.509155, "tokens_per_sec": 106243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:09.656452+00:00", "epoch": 0, "step": 242, "train_loss": 5.673046112060547, "perplexity": 290.9193590382584, "lr": 0.0010354426229508195, "grad_norm": 0.496264, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:09.962284+00:00", "epoch": 0, "step": 243, "train_loss": 5.776470184326172, "perplexity": 322.61839473608785, "lr": 0.0010397213114754099, "grad_norm": 0.453773, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:10.268492+00:00", "epoch": 0, "step": 244, "train_loss": 5.628250598907471, "perplexity": 278.1750519808164, "lr": 0.001044, "grad_norm": 0.354641, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:10.575472+00:00", "epoch": 0, "step": 245, "train_loss": 5.750302791595459, "perplexity": 314.28580898138614, "lr": 0.0010482786885245901, "grad_norm": 0.442569, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:10.883481+00:00", "epoch": 0, "step": 246, "train_loss": 5.7321600914001465, "perplexity": 308.6352291546688, "lr": 0.0010525573770491803, "grad_norm": 0.351705, "tokens_per_sec": 106388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:11.189904+00:00", "epoch": 0, "step": 247, "train_loss": 5.646759986877441, "perplexity": 283.3718483348738, "lr": 0.0010568360655737704, "grad_norm": 0.319691, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:11.496365+00:00", "epoch": 0, "step": 248, "train_loss": 5.6255879402160645, "perplexity": 277.4353519822664, "lr": 0.0010611147540983605, "grad_norm": 0.320884, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:11.802633+00:00", "epoch": 0, "step": 249, "train_loss": 5.648036003112793, "perplexity": 283.73366620763744, "lr": 0.0010653934426229509, "grad_norm": 0.363256, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:12.110345+00:00", "epoch": 0, "step": 250, "train_loss": 5.7076215744018555, "perplexity": 301.1539435586068, "lr": 0.001069672131147541, "grad_norm": 0.339853, "tokens_per_sec": 106546} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:29:13.030108+00:00", "step": 250, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0000_step_0000250.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:14.001910+00:00", "epoch": 0, "step": 251, "train_loss": 5.6217360496521, "perplexity": 276.3687568904017, "lr": 0.0010739508196721311, "grad_norm": 0.379154, "tokens_per_sec": 17322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:14.307732+00:00", "epoch": 0, "step": 252, "train_loss": 5.621762275695801, "perplexity": 276.3760050445423, "lr": 0.0010782295081967212, "grad_norm": 0.372821, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:14.616109+00:00", "epoch": 0, "step": 253, "train_loss": 5.693219184875488, "perplexity": 296.84769175051525, "lr": 0.0010825081967213114, "grad_norm": 0.362462, "tokens_per_sec": 106266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:14.923512+00:00", "epoch": 0, "step": 254, "train_loss": 5.6325788497924805, "perplexity": 279.38167279082336, "lr": 0.0010867868852459015, "grad_norm": 0.489657, "tokens_per_sec": 106592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:15.231314+00:00", "epoch": 0, "step": 255, "train_loss": 5.744588375091553, "perplexity": 312.49497063893926, "lr": 0.0010910655737704918, "grad_norm": 0.810231, "tokens_per_sec": 106458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:15.538471+00:00", "epoch": 0, "step": 256, "train_loss": 5.644151210784912, "perplexity": 282.63355806735746, "lr": 0.001095344262295082, "grad_norm": 0.917594, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:15.845193+00:00", "epoch": 0, "step": 257, "train_loss": 5.562156677246094, "perplexity": 260.38379500804996, "lr": 0.001099622950819672, "grad_norm": 0.865791, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:16.151595+00:00", "epoch": 0, "step": 258, "train_loss": 5.68347692489624, "perplexity": 293.9697658766244, "lr": 0.0011039016393442622, "grad_norm": 0.890475, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:16.577195+00:00", "epoch": 0, "step": 259, "train_loss": 5.691441059112549, "perplexity": 296.32032822043266, "lr": 0.0011081803278688524, "grad_norm": 0.77637, "tokens_per_sec": 76993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:16.884474+00:00", "epoch": 0, "step": 260, "train_loss": 5.636532306671143, "perplexity": 280.48838240979586, "lr": 0.0011124590163934427, "grad_norm": 0.919643, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:17.191467+00:00", "epoch": 0, "step": 261, "train_loss": 5.608102798461914, "perplexity": 272.6265195996605, "lr": 0.0011167377049180328, "grad_norm": 0.900317, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:17.498523+00:00", "epoch": 0, "step": 262, "train_loss": 5.60368013381958, "perplexity": 271.4234462867712, "lr": 0.001121016393442623, "grad_norm": 1.080681, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:17.806267+00:00", "epoch": 0, "step": 263, "train_loss": 5.673232078552246, "perplexity": 290.973465321648, "lr": 0.001125295081967213, "grad_norm": 0.816187, "tokens_per_sec": 106479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:18.114135+00:00", "epoch": 0, "step": 264, "train_loss": 5.607629776000977, "perplexity": 272.49759162775996, "lr": 0.0011295737704918032, "grad_norm": 0.737561, "tokens_per_sec": 106434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:18.422275+00:00", "epoch": 0, "step": 265, "train_loss": 5.571247577667236, "perplexity": 262.76171047935253, "lr": 0.0011338524590163933, "grad_norm": 0.651096, "tokens_per_sec": 106342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:18.728667+00:00", "epoch": 0, "step": 266, "train_loss": 5.472677707672119, "perplexity": 238.09689381035327, "lr": 0.0011381311475409837, "grad_norm": 0.703147, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:19.035008+00:00", "epoch": 0, "step": 267, "train_loss": 5.60360860824585, "perplexity": 271.4040332633237, "lr": 0.0011424098360655738, "grad_norm": 0.571776, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:19.342186+00:00", "epoch": 0, "step": 268, "train_loss": 5.480945110321045, "perplexity": 240.0734961261771, "lr": 0.001146688524590164, "grad_norm": 0.414412, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:19.649713+00:00", "epoch": 0, "step": 269, "train_loss": 5.42655086517334, "perplexity": 227.3636834493069, "lr": 0.001150967213114754, "grad_norm": 0.410879, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:19.959200+00:00", "epoch": 0, "step": 270, "train_loss": 5.542973041534424, "perplexity": 255.43629439146852, "lr": 0.0011552459016393442, "grad_norm": 0.384877, "tokens_per_sec": 105935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:20.267783+00:00", "epoch": 0, "step": 271, "train_loss": 5.617738246917725, "perplexity": 275.2660946999705, "lr": 0.0011595245901639343, "grad_norm": 0.306918, "tokens_per_sec": 106132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:20.573826+00:00", "epoch": 0, "step": 272, "train_loss": 5.6065754890441895, "perplexity": 272.2104523613393, "lr": 0.0011638032786885247, "grad_norm": 0.320302, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:20.880342+00:00", "epoch": 0, "step": 273, "train_loss": 5.523648262023926, "perplexity": 250.54743453602245, "lr": 0.0011680819672131148, "grad_norm": 0.280813, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:21.187202+00:00", "epoch": 0, "step": 274, "train_loss": 5.462286949157715, "perplexity": 235.6356954866128, "lr": 0.001172360655737705, "grad_norm": 0.317324, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:21.494387+00:00", "epoch": 0, "step": 275, "train_loss": 5.537036895751953, "perplexity": 253.92447892774277, "lr": 0.001176639344262295, "grad_norm": 0.295794, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:21.801400+00:00", "epoch": 0, "step": 276, "train_loss": 5.648141384124756, "perplexity": 283.76356792401924, "lr": 0.0011809180327868852, "grad_norm": 0.286432, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:22.107287+00:00", "epoch": 0, "step": 277, "train_loss": 5.463951587677002, "perplexity": 236.02827039884636, "lr": 0.0011851967213114755, "grad_norm": 0.362837, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:22.414434+00:00", "epoch": 0, "step": 278, "train_loss": 5.545236110687256, "perplexity": 256.0150189897188, "lr": 0.0011894754098360656, "grad_norm": 0.441764, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:22.720574+00:00", "epoch": 0, "step": 279, "train_loss": 5.4636969566345215, "perplexity": 235.9681779253291, "lr": 0.0011937540983606558, "grad_norm": 0.54508, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:23.027980+00:00", "epoch": 0, "step": 280, "train_loss": 5.52549934387207, "perplexity": 251.01164786006777, "lr": 0.001198032786885246, "grad_norm": 0.471359, "tokens_per_sec": 106652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:23.335257+00:00", "epoch": 0, "step": 281, "train_loss": 5.484488010406494, "perplexity": 240.92556103526653, "lr": 0.001202311475409836, "grad_norm": 0.322832, "tokens_per_sec": 106583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:23.642180+00:00", "epoch": 0, "step": 282, "train_loss": 5.5019378662109375, "perplexity": 245.1665722376921, "lr": 0.0012065901639344262, "grad_norm": 0.478606, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:23.948204+00:00", "epoch": 0, "step": 283, "train_loss": 5.453249454498291, "perplexity": 233.51573315043402, "lr": 0.0012108688524590165, "grad_norm": 0.590681, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:24.255121+00:00", "epoch": 0, "step": 284, "train_loss": 5.433923244476318, "perplexity": 229.04608880847763, "lr": 0.0012151475409836066, "grad_norm": 0.671674, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:24.561761+00:00", "epoch": 0, "step": 285, "train_loss": 5.392948150634766, "perplexity": 219.8505837042719, "lr": 0.0012194262295081968, "grad_norm": 0.624678, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:24.868726+00:00", "epoch": 0, "step": 286, "train_loss": 5.413325309753418, "perplexity": 224.3764697756914, "lr": 0.0012237049180327869, "grad_norm": 0.634454, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:25.175871+00:00", "epoch": 0, "step": 287, "train_loss": 5.429811954498291, "perplexity": 228.10634701823622, "lr": 0.001227983606557377, "grad_norm": 0.535472, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:25.481753+00:00", "epoch": 0, "step": 288, "train_loss": 5.4483537673950195, "perplexity": 232.37530704970183, "lr": 0.0012322622950819671, "grad_norm": 0.492643, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:25.787920+00:00", "epoch": 0, "step": 289, "train_loss": 5.52730131149292, "perplexity": 251.46437049523013, "lr": 0.0012365409836065575, "grad_norm": 0.518219, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:26.094396+00:00", "epoch": 0, "step": 290, "train_loss": 5.442154884338379, "perplexity": 230.9392951293954, "lr": 0.0012408196721311476, "grad_norm": 0.600871, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:26.400290+00:00", "epoch": 0, "step": 291, "train_loss": 5.524394512176514, "perplexity": 250.7344753782239, "lr": 0.0012450983606557377, "grad_norm": 0.570183, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:26.707825+00:00", "epoch": 0, "step": 292, "train_loss": 5.455496788024902, "perplexity": 234.04111101506012, "lr": 0.0012493770491803279, "grad_norm": 0.50665, "tokens_per_sec": 106550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:27.015048+00:00", "epoch": 0, "step": 293, "train_loss": 5.396918296813965, "perplexity": 220.72515760351547, "lr": 0.001253655737704918, "grad_norm": 0.516978, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:27.321842+00:00", "epoch": 0, "step": 294, "train_loss": 5.442235946655273, "perplexity": 230.9580163625036, "lr": 0.0012579344262295083, "grad_norm": 0.423651, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:27.628604+00:00", "epoch": 0, "step": 295, "train_loss": 5.394186019897461, "perplexity": 220.12289849450656, "lr": 0.0012622131147540985, "grad_norm": 0.476298, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:27.934493+00:00", "epoch": 0, "step": 296, "train_loss": 5.490837097167969, "perplexity": 242.46008458207214, "lr": 0.0012664918032786886, "grad_norm": 0.424806, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:28.241503+00:00", "epoch": 0, "step": 297, "train_loss": 5.338465690612793, "perplexity": 208.19303258531156, "lr": 0.0012707704918032787, "grad_norm": 0.466365, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:28.549067+00:00", "epoch": 0, "step": 298, "train_loss": 5.430662631988525, "perplexity": 228.30047451126177, "lr": 0.0012750491803278688, "grad_norm": 0.506145, "tokens_per_sec": 106541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:28.855706+00:00", "epoch": 0, "step": 299, "train_loss": 5.43773889541626, "perplexity": 229.92171821824869, "lr": 0.001279327868852459, "grad_norm": 0.581819, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:29.162991+00:00", "epoch": 0, "step": 300, "train_loss": 5.400849342346191, "perplexity": 221.59454593115018, "lr": 0.001283606557377049, "grad_norm": 0.487616, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:29:34.766349+00:00", "step": 300, "epoch": 0, "val_loss": 5.397777009010315, "val_ppl": 220.91477839158057, "eval_train_loss": 5.400849342346191, "eval_train_ppl": 221.59454593115018} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:29:35.660183+00:00", "step": 300, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_5p3978_epoch_0000_step_0000300.pt", "val_loss": 5.397777009010315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:36.625826+00:00", "epoch": 0, "step": 301, "train_loss": 5.369535446166992, "perplexity": 214.7630755167886, "lr": 0.0012878852459016392, "grad_norm": 0.41025, "tokens_per_sec": 4391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:36.931987+00:00", "epoch": 0, "step": 302, "train_loss": 5.411109924316406, "perplexity": 223.87993961806038, "lr": 0.0012921639344262294, "grad_norm": 0.520481, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:37.237794+00:00", "epoch": 0, "step": 303, "train_loss": 5.385076999664307, "perplexity": 218.12689915886622, "lr": 0.0012964426229508195, "grad_norm": 0.551208, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:37.543136+00:00", "epoch": 0, "step": 304, "train_loss": 5.390411853790283, "perplexity": 219.29368389271798, "lr": 0.0013007213114754096, "grad_norm": 0.605496, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:37.847959+00:00", "epoch": 0, "step": 305, "train_loss": 5.315545082092285, "perplexity": 203.47539369992202, "lr": 0.001305, "grad_norm": 0.514944, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:38.153221+00:00", "epoch": 0, "step": 306, "train_loss": 5.3159871101379395, "perplexity": 203.565355411873, "lr": 0.001305, "grad_norm": 0.570389, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:38.459908+00:00", "epoch": 0, "step": 307, "train_loss": 5.41316556930542, "perplexity": 224.34063064044517, "lr": 0.001305, "grad_norm": 0.620235, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:38.764944+00:00", "epoch": 0, "step": 308, "train_loss": 5.550622463226318, "perplexity": 257.39772666978774, "lr": 0.001305, "grad_norm": 0.721353, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:39.069535+00:00", "epoch": 0, "step": 309, "train_loss": 5.14976167678833, "perplexity": 172.3904007867778, "lr": 0.001305, "grad_norm": 0.656452, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:39.375846+00:00", "epoch": 0, "step": 310, "train_loss": 5.359043121337891, "perplexity": 212.52149184360027, "lr": 0.001305, "grad_norm": 0.501726, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:39.681210+00:00", "epoch": 0, "step": 311, "train_loss": 5.4487738609313965, "perplexity": 232.47294692171064, "lr": 0.001305, "grad_norm": 0.395493, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:39.988270+00:00", "epoch": 0, "step": 312, "train_loss": 5.236572742462158, "perplexity": 188.02458817863672, "lr": 0.001305, "grad_norm": 0.459733, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:40.294333+00:00", "epoch": 0, "step": 313, "train_loss": 5.324303150177002, "perplexity": 205.2652715466849, "lr": 0.001305, "grad_norm": 0.444949, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:40.600682+00:00", "epoch": 0, "step": 314, "train_loss": 5.421109199523926, "perplexity": 226.12980652077258, "lr": 0.001305, "grad_norm": 0.459191, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:40.934956+00:00", "epoch": 0, "step": 315, "train_loss": 5.419651985168457, "perplexity": 225.80052689427853, "lr": 0.001305, "grad_norm": 0.370119, "tokens_per_sec": 98028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:41.241469+00:00", "epoch": 0, "step": 316, "train_loss": 5.375874042510986, "perplexity": 216.12869544653353, "lr": 0.001305, "grad_norm": 0.323517, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:41.546573+00:00", "epoch": 0, "step": 317, "train_loss": 5.252215385437012, "perplexity": 190.98891418329092, "lr": 0.001305, "grad_norm": 0.363472, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:41.851429+00:00", "epoch": 0, "step": 318, "train_loss": 5.295059680938721, "perplexity": 199.34952297821968, "lr": 0.001305, "grad_norm": 0.330259, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:42.158136+00:00", "epoch": 0, "step": 319, "train_loss": 5.311233043670654, "perplexity": 202.59988894607048, "lr": 0.001305, "grad_norm": 0.364842, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:42.465563+00:00", "epoch": 0, "step": 320, "train_loss": 5.375011920928955, "perplexity": 215.942446529834, "lr": 0.001305, "grad_norm": 0.299963, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:42.771664+00:00", "epoch": 0, "step": 321, "train_loss": 5.163254737854004, "perplexity": 174.73223874300106, "lr": 0.001305, "grad_norm": 0.300611, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:43.077608+00:00", "epoch": 0, "step": 322, "train_loss": 5.2910566329956055, "perplexity": 198.55311237870626, "lr": 0.001305, "grad_norm": 0.289004, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:43.383138+00:00", "epoch": 0, "step": 323, "train_loss": 5.291439056396484, "perplexity": 198.62905825601197, "lr": 0.001305, "grad_norm": 0.263723, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:43.690087+00:00", "epoch": 0, "step": 324, "train_loss": 5.246492862701416, "perplexity": 189.899097005462, "lr": 0.001305, "grad_norm": 0.317969, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:43.996547+00:00", "epoch": 0, "step": 325, "train_loss": 5.207749843597412, "perplexity": 182.68253110019705, "lr": 0.001305, "grad_norm": 0.365473, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:44.302028+00:00", "epoch": 0, "step": 326, "train_loss": 5.252755641937256, "perplexity": 191.092125063317, "lr": 0.001305, "grad_norm": 0.429435, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:44.608055+00:00", "epoch": 0, "step": 327, "train_loss": 5.129665851593018, "perplexity": 168.96065067890532, "lr": 0.001305, "grad_norm": 0.553609, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:44.913754+00:00", "epoch": 0, "step": 328, "train_loss": 5.249813556671143, "perplexity": 190.53074196111277, "lr": 0.001305, "grad_norm": 0.60285, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:45.220817+00:00", "epoch": 0, "step": 329, "train_loss": 5.293896198272705, "perplexity": 199.11771814036078, "lr": 0.001305, "grad_norm": 0.652878, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:45.528440+00:00", "epoch": 0, "step": 330, "train_loss": 5.298732280731201, "perplexity": 200.08300005439204, "lr": 0.001305, "grad_norm": 0.784399, "tokens_per_sec": 106575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:45.835414+00:00", "epoch": 0, "step": 331, "train_loss": 5.296490669250488, "perplexity": 199.6349940197059, "lr": 0.001305, "grad_norm": 0.863864, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:46.142656+00:00", "epoch": 0, "step": 332, "train_loss": 5.206589698791504, "perplexity": 182.47071580257716, "lr": 0.001305, "grad_norm": 0.666718, "tokens_per_sec": 106656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:46.447972+00:00", "epoch": 0, "step": 333, "train_loss": 5.273014068603516, "perplexity": 195.0028294782828, "lr": 0.001305, "grad_norm": 0.491257, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:46.754932+00:00", "epoch": 0, "step": 334, "train_loss": 5.251365661621094, "perplexity": 190.8266952847244, "lr": 0.001305, "grad_norm": 0.506868, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:47.063646+00:00", "epoch": 0, "step": 335, "train_loss": 5.288103103637695, "perplexity": 197.96754510293624, "lr": 0.001305, "grad_norm": 0.630917, "tokens_per_sec": 106145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:47.371181+00:00", "epoch": 0, "step": 336, "train_loss": 5.1732306480407715, "perplexity": 176.4840754177347, "lr": 0.001305, "grad_norm": 0.781518, "tokens_per_sec": 106550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:47.677935+00:00", "epoch": 0, "step": 337, "train_loss": 5.20140266418457, "perplexity": 181.52668436345164, "lr": 0.001305, "grad_norm": 0.669613, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:47.984174+00:00", "epoch": 0, "step": 338, "train_loss": 5.309063911437988, "perplexity": 202.16089928206344, "lr": 0.001305, "grad_norm": 0.615389, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:48.291179+00:00", "epoch": 0, "step": 339, "train_loss": 5.128115177154541, "perplexity": 168.6988507524763, "lr": 0.001305, "grad_norm": 0.600022, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:48.599725+00:00", "epoch": 0, "step": 340, "train_loss": 5.143986225128174, "perplexity": 171.39763794679016, "lr": 0.001305, "grad_norm": 0.535998, "tokens_per_sec": 106262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:48.907781+00:00", "epoch": 0, "step": 341, "train_loss": 5.281744480133057, "perplexity": 196.71273766938089, "lr": 0.001305, "grad_norm": 0.438038, "tokens_per_sec": 106312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:49.214867+00:00", "epoch": 0, "step": 342, "train_loss": 5.310948371887207, "perplexity": 202.54222268272642, "lr": 0.001305, "grad_norm": 0.430064, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:49.521876+00:00", "epoch": 0, "step": 343, "train_loss": 5.287246227264404, "perplexity": 197.797984047665, "lr": 0.001305, "grad_norm": 0.444349, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:49.828878+00:00", "epoch": 0, "step": 344, "train_loss": 5.223834037780762, "perplexity": 185.64458971015793, "lr": 0.001305, "grad_norm": 0.359551, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:50.136320+00:00", "epoch": 0, "step": 345, "train_loss": 5.243664741516113, "perplexity": 189.36279806280695, "lr": 0.001305, "grad_norm": 0.354177, "tokens_per_sec": 106583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:50.444331+00:00", "epoch": 0, "step": 346, "train_loss": 5.3524699211120605, "perplexity": 211.1291266862223, "lr": 0.001305, "grad_norm": 0.320606, "tokens_per_sec": 106386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:50.750932+00:00", "epoch": 0, "step": 347, "train_loss": 5.1406097412109375, "perplexity": 170.8198925018012, "lr": 0.001305, "grad_norm": 0.342729, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:51.057780+00:00", "epoch": 0, "step": 348, "train_loss": 5.2321248054504395, "perplexity": 187.19012385190877, "lr": 0.001305, "grad_norm": 0.397066, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:51.365438+00:00", "epoch": 0, "step": 349, "train_loss": 5.230154037475586, "perplexity": 186.82157882826098, "lr": 0.001305, "grad_norm": 0.565987, "tokens_per_sec": 106508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:51.674282+00:00", "epoch": 0, "step": 350, "train_loss": 5.224435806274414, "perplexity": 185.75633839530923, "lr": 0.001305, "grad_norm": 0.621658, "tokens_per_sec": 106159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:51.981619+00:00", "epoch": 0, "step": 351, "train_loss": 5.230326175689697, "perplexity": 186.85374072926462, "lr": 0.001305, "grad_norm": 0.523797, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:52.288345+00:00", "epoch": 0, "step": 352, "train_loss": 5.263272762298584, "perplexity": 193.1124694226495, "lr": 0.001305, "grad_norm": 0.535088, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:52.594755+00:00", "epoch": 0, "step": 353, "train_loss": 5.327576160430908, "perplexity": 205.93820754770294, "lr": 0.001305, "grad_norm": 0.53339, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:52.902281+00:00", "epoch": 0, "step": 354, "train_loss": 5.170456409454346, "perplexity": 175.99514500421438, "lr": 0.001305, "grad_norm": 0.507948, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:53.209815+00:00", "epoch": 0, "step": 355, "train_loss": 5.2052388191223145, "perplexity": 182.22438624063565, "lr": 0.001305, "grad_norm": 0.503044, "tokens_per_sec": 106550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:53.517581+00:00", "epoch": 0, "step": 356, "train_loss": 5.1249518394470215, "perplexity": 168.1660424872185, "lr": 0.001305, "grad_norm": 0.439799, "tokens_per_sec": 106471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:53.823911+00:00", "epoch": 0, "step": 357, "train_loss": 5.158883094787598, "perplexity": 173.9700390094956, "lr": 0.001305, "grad_norm": 0.409005, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:54.129898+00:00", "epoch": 0, "step": 358, "train_loss": 5.139888763427734, "perplexity": 170.69677954056613, "lr": 0.001305, "grad_norm": 0.4094, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:54.436732+00:00", "epoch": 0, "step": 359, "train_loss": 5.148899078369141, "perplexity": 172.2417612169213, "lr": 0.001305, "grad_norm": 0.399403, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:54.744275+00:00", "epoch": 0, "step": 360, "train_loss": 5.204927921295166, "perplexity": 182.16774188066046, "lr": 0.001305, "grad_norm": 0.320467, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:55.074510+00:00", "epoch": 0, "step": 361, "train_loss": 5.150020599365234, "perplexity": 172.43504233268575, "lr": 0.001305, "grad_norm": 0.327559, "tokens_per_sec": 99172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:55.380818+00:00", "epoch": 0, "step": 362, "train_loss": 5.132461071014404, "perplexity": 169.43359345261126, "lr": 0.001305, "grad_norm": 0.314848, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:55.686395+00:00", "epoch": 0, "step": 363, "train_loss": 5.106797218322754, "perplexity": 165.1405976978111, "lr": 0.001305, "grad_norm": 0.331731, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:55.992689+00:00", "epoch": 0, "step": 364, "train_loss": 5.09335994720459, "perplexity": 162.93640107350814, "lr": 0.001305, "grad_norm": 0.285634, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:56.299725+00:00", "epoch": 0, "step": 365, "train_loss": 5.014570713043213, "perplexity": 150.59147593146272, "lr": 0.001305, "grad_norm": 0.262543, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:56.606840+00:00", "epoch": 0, "step": 366, "train_loss": 5.070409774780273, "perplexity": 159.23956633404302, "lr": 0.001305, "grad_norm": 0.282073, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:56.913605+00:00", "epoch": 0, "step": 367, "train_loss": 5.0728302001953125, "perplexity": 159.62546065329116, "lr": 0.001305, "grad_norm": 0.238711, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:57.219981+00:00", "epoch": 0, "step": 368, "train_loss": 5.095550537109375, "perplexity": 163.29371913460247, "lr": 0.001305, "grad_norm": 0.277515, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:57.526165+00:00", "epoch": 0, "step": 369, "train_loss": 5.094960689544678, "perplexity": 163.19742913304378, "lr": 0.001305, "grad_norm": 0.300718, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:57.833238+00:00", "epoch": 0, "step": 370, "train_loss": 5.262484550476074, "perplexity": 192.96031586367098, "lr": 0.001305, "grad_norm": 0.288823, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:58.140702+00:00", "epoch": 0, "step": 371, "train_loss": 5.178603172302246, "perplexity": 177.43479198179298, "lr": 0.001305, "grad_norm": 0.315879, "tokens_per_sec": 106516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:58.447935+00:00", "epoch": 0, "step": 372, "train_loss": 5.070028781890869, "perplexity": 159.17890874734627, "lr": 0.001305, "grad_norm": 0.343292, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:58.754204+00:00", "epoch": 0, "step": 373, "train_loss": 5.019587993621826, "perplexity": 151.34893422037536, "lr": 0.001305, "grad_norm": 0.349811, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:59.059809+00:00", "epoch": 0, "step": 374, "train_loss": 5.06754207611084, "perplexity": 158.78356938496609, "lr": 0.001305, "grad_norm": 0.434066, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:59.365016+00:00", "epoch": 0, "step": 375, "train_loss": 5.199193000793457, "perplexity": 181.12601433048178, "lr": 0.001305, "grad_norm": 0.489004, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:59.672404+00:00", "epoch": 0, "step": 376, "train_loss": 5.1364240646362305, "perplexity": 170.10638996639344, "lr": 0.001305, "grad_norm": 0.420261, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:29:59.978712+00:00", "epoch": 0, "step": 377, "train_loss": 5.1619553565979, "perplexity": 174.50534239149866, "lr": 0.001305, "grad_norm": 0.430214, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:00.286500+00:00", "epoch": 0, "step": 378, "train_loss": 5.059307098388672, "perplexity": 157.48135942253998, "lr": 0.001305, "grad_norm": 0.456524, "tokens_per_sec": 106462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:00.592531+00:00", "epoch": 0, "step": 379, "train_loss": 4.933588981628418, "perplexity": 138.87704615330378, "lr": 0.001305, "grad_norm": 0.595397, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:00.898972+00:00", "epoch": 0, "step": 380, "train_loss": 5.055272102355957, "perplexity": 156.8472030317057, "lr": 0.001305, "grad_norm": 0.617592, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:01.205254+00:00", "epoch": 0, "step": 381, "train_loss": 5.083837509155273, "perplexity": 161.39221317360096, "lr": 0.001305, "grad_norm": 0.575959, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:01.511844+00:00", "epoch": 0, "step": 382, "train_loss": 5.036146640777588, "perplexity": 153.87593191182526, "lr": 0.001305, "grad_norm": 0.526874, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:01.819940+00:00", "epoch": 0, "step": 383, "train_loss": 5.07717227935791, "perplexity": 160.32007398151035, "lr": 0.001305, "grad_norm": 0.489248, "tokens_per_sec": 106356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:02.127376+00:00", "epoch": 0, "step": 384, "train_loss": 5.184142589569092, "perplexity": 178.42040467190884, "lr": 0.001305, "grad_norm": 0.406799, "tokens_per_sec": 106585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:02.435080+00:00", "epoch": 0, "step": 385, "train_loss": 5.11104679107666, "perplexity": 165.84386792299023, "lr": 0.001305, "grad_norm": 0.391353, "tokens_per_sec": 106493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:02.742350+00:00", "epoch": 0, "step": 386, "train_loss": 5.0082597732543945, "perplexity": 149.64409477123252, "lr": 0.001305, "grad_norm": 0.343838, "tokens_per_sec": 106643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:03.048513+00:00", "epoch": 0, "step": 387, "train_loss": 4.95857048034668, "perplexity": 142.39010089100796, "lr": 0.001305, "grad_norm": 0.326096, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:03.355535+00:00", "epoch": 0, "step": 388, "train_loss": 5.000909805297852, "perplexity": 148.54824762380693, "lr": 0.001305, "grad_norm": 0.394616, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:03.661964+00:00", "epoch": 0, "step": 389, "train_loss": 5.077030658721924, "perplexity": 160.29737095831794, "lr": 0.001305, "grad_norm": 0.419185, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:03.970611+00:00", "epoch": 0, "step": 390, "train_loss": 4.997586250305176, "perplexity": 148.0553588788928, "lr": 0.001305, "grad_norm": 0.515075, "tokens_per_sec": 106225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:04.277413+00:00", "epoch": 0, "step": 391, "train_loss": 5.027230262756348, "perplexity": 152.5100145019105, "lr": 0.001305, "grad_norm": 0.585764, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:04.583776+00:00", "epoch": 0, "step": 392, "train_loss": 5.03786039352417, "perplexity": 154.13986330475072, "lr": 0.001305, "grad_norm": 0.565193, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:04.890369+00:00", "epoch": 0, "step": 393, "train_loss": 5.032658576965332, "perplexity": 153.340137827357, "lr": 0.001305, "grad_norm": 0.579636, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:05.198216+00:00", "epoch": 0, "step": 394, "train_loss": 5.101363658905029, "perplexity": 164.24572980850527, "lr": 0.001305, "grad_norm": 0.539153, "tokens_per_sec": 106443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:05.504757+00:00", "epoch": 0, "step": 395, "train_loss": 5.137851238250732, "perplexity": 170.3493346386884, "lr": 0.001305, "grad_norm": 0.473229, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:05.812287+00:00", "epoch": 0, "step": 396, "train_loss": 5.208595275878906, "perplexity": 182.83704211435068, "lr": 0.001305, "grad_norm": 0.382746, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:06.118653+00:00", "epoch": 0, "step": 397, "train_loss": 5.095718860626221, "perplexity": 163.32120762110048, "lr": 0.001305, "grad_norm": 0.370499, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:06.425322+00:00", "epoch": 0, "step": 398, "train_loss": 4.910874843597412, "perplexity": 135.75812960275942, "lr": 0.001305, "grad_norm": 0.343987, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:06.731098+00:00", "epoch": 0, "step": 399, "train_loss": 4.980903625488281, "perplexity": 145.60589542341253, "lr": 0.001305, "grad_norm": 0.350807, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:07.039238+00:00", "epoch": 0, "step": 400, "train_loss": 5.036612033843994, "perplexity": 153.94756137025763, "lr": 0.001305, "grad_norm": 0.349042, "tokens_per_sec": 106398} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:30:12.646200+00:00", "step": 400, "epoch": 0, "val_loss": 5.004489278793335, "val_ppl": 149.08092492236304, "eval_train_loss": 5.036612033843994, "eval_train_ppl": 153.94756137025763} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:30:13.532646+00:00", "step": 400, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_5p0045_epoch_0000_step_0000400.pt", "val_loss": 5.004489278793335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:14.533767+00:00", "epoch": 0, "step": 401, "train_loss": 5.061088562011719, "perplexity": 157.76215677653317, "lr": 0.001305, "grad_norm": 0.329009, "tokens_per_sec": 4372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:14.838807+00:00", "epoch": 0, "step": 402, "train_loss": 4.968611240386963, "perplexity": 143.8270074712384, "lr": 0.001305, "grad_norm": 0.360605, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:15.143786+00:00", "epoch": 0, "step": 403, "train_loss": 5.060128688812256, "perplexity": 157.6107977647047, "lr": 0.001305, "grad_norm": 0.375976, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:15.449150+00:00", "epoch": 0, "step": 404, "train_loss": 4.967255592346191, "perplexity": 143.63216077196353, "lr": 0.001305, "grad_norm": 0.336568, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:15.755503+00:00", "epoch": 0, "step": 405, "train_loss": 4.916141510009766, "perplexity": 136.47500850811713, "lr": 0.001305, "grad_norm": 0.303957, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:16.062346+00:00", "epoch": 0, "step": 406, "train_loss": 5.064091682434082, "perplexity": 158.2366476514344, "lr": 0.001305, "grad_norm": 0.335191, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:16.368068+00:00", "epoch": 0, "step": 407, "train_loss": 4.903985500335693, "perplexity": 134.82605960899346, "lr": 0.001305, "grad_norm": 0.361104, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:16.673769+00:00", "epoch": 0, "step": 408, "train_loss": 4.9972991943359375, "perplexity": 148.01286480372966, "lr": 0.001305, "grad_norm": 0.435464, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:16.979195+00:00", "epoch": 0, "step": 409, "train_loss": 4.994763374328613, "perplexity": 147.6380063075299, "lr": 0.001305, "grad_norm": 0.619206, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:17.285790+00:00", "epoch": 0, "step": 410, "train_loss": 4.982238292694092, "perplexity": 145.80036058127394, "lr": 0.001305, "grad_norm": 0.603568, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:17.592886+00:00", "epoch": 0, "step": 411, "train_loss": 5.034398555755615, "perplexity": 153.6071786702043, "lr": 0.001305, "grad_norm": 0.453228, "tokens_per_sec": 106642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:17.899151+00:00", "epoch": 0, "step": 412, "train_loss": 4.956480503082275, "perplexity": 142.09281958126928, "lr": 0.001305, "grad_norm": 0.461213, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:18.205462+00:00", "epoch": 0, "step": 413, "train_loss": 4.962649345397949, "perplexity": 142.9720769912992, "lr": 0.001305, "grad_norm": 0.402146, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:18.512851+00:00", "epoch": 0, "step": 414, "train_loss": 4.96638822555542, "perplexity": 143.50763301904897, "lr": 0.001305, "grad_norm": 0.406164, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:18.884885+00:00", "epoch": 0, "step": 415, "train_loss": 4.999311923980713, "perplexity": 148.31107469179662, "lr": 0.001305, "grad_norm": 0.400849, "tokens_per_sec": 88078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:19.190988+00:00", "epoch": 0, "step": 416, "train_loss": 5.01227331161499, "perplexity": 150.24590397033373, "lr": 0.001305, "grad_norm": 0.392054, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:19.496457+00:00", "epoch": 0, "step": 417, "train_loss": 4.999813079833984, "perplexity": 148.38542028279142, "lr": 0.001305, "grad_norm": 0.43299, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:19.802997+00:00", "epoch": 0, "step": 418, "train_loss": 4.903947830200195, "perplexity": 134.82098078871985, "lr": 0.001305, "grad_norm": 0.460237, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:20.110317+00:00", "epoch": 0, "step": 419, "train_loss": 5.112758159637451, "perplexity": 166.12793090345076, "lr": 0.001305, "grad_norm": 0.453139, "tokens_per_sec": 106627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:20.419924+00:00", "epoch": 0, "step": 420, "train_loss": 5.0989251136779785, "perplexity": 163.84569711515772, "lr": 0.001305, "grad_norm": 0.522754, "tokens_per_sec": 105895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:20.727397+00:00", "epoch": 0, "step": 421, "train_loss": 4.893754959106445, "perplexity": 133.45374776312593, "lr": 0.001305, "grad_norm": 0.545551, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:21.034427+00:00", "epoch": 0, "step": 422, "train_loss": 4.8761186599731445, "perplexity": 131.12075073463538, "lr": 0.001305, "grad_norm": 0.560093, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:21.341522+00:00", "epoch": 0, "step": 423, "train_loss": 4.904018402099609, "perplexity": 134.83049569715362, "lr": 0.001305, "grad_norm": 0.624236, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:21.648574+00:00", "epoch": 0, "step": 424, "train_loss": 4.8869194984436035, "perplexity": 132.54464054302446, "lr": 0.001305, "grad_norm": 0.507982, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:21.955650+00:00", "epoch": 0, "step": 425, "train_loss": 5.0080718994140625, "perplexity": 149.61598320126183, "lr": 0.001305, "grad_norm": 0.407223, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:22.263533+00:00", "epoch": 0, "step": 426, "train_loss": 4.970626354217529, "perplexity": 144.11712747744738, "lr": 0.001305, "grad_norm": 0.383835, "tokens_per_sec": 106430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:22.571610+00:00", "epoch": 0, "step": 427, "train_loss": 4.8935441970825195, "perplexity": 133.42562374498848, "lr": 0.001305, "grad_norm": 0.355676, "tokens_per_sec": 106367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:22.879308+00:00", "epoch": 0, "step": 428, "train_loss": 4.887360095977783, "perplexity": 132.60305225189896, "lr": 0.001305, "grad_norm": 0.312443, "tokens_per_sec": 106495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:23.185178+00:00", "epoch": 0, "step": 429, "train_loss": 4.8831281661987305, "perplexity": 132.0430711822985, "lr": 0.001305, "grad_norm": 0.330127, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:23.493541+00:00", "epoch": 0, "step": 430, "train_loss": 4.906830310821533, "perplexity": 135.2101602849067, "lr": 0.001305, "grad_norm": 0.316833, "tokens_per_sec": 106327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:23.800998+00:00", "epoch": 0, "step": 431, "train_loss": 4.893272399902344, "perplexity": 133.38936396456765, "lr": 0.001305, "grad_norm": 0.323906, "tokens_per_sec": 106515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:24.108126+00:00", "epoch": 0, "step": 432, "train_loss": 4.787836074829102, "perplexity": 120.04132696033676, "lr": 0.001305, "grad_norm": 0.332696, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:24.414393+00:00", "epoch": 0, "step": 433, "train_loss": 4.8892951011657715, "perplexity": 132.85988825531095, "lr": 0.001305, "grad_norm": 0.310975, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:24.720256+00:00", "epoch": 0, "step": 434, "train_loss": 4.860841751098633, "perplexity": 129.1328540938014, "lr": 0.001305, "grad_norm": 0.308904, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:25.027115+00:00", "epoch": 0, "step": 435, "train_loss": 4.89216947555542, "perplexity": 133.24232668783185, "lr": 0.001305, "grad_norm": 0.299063, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:25.334036+00:00", "epoch": 0, "step": 436, "train_loss": 4.943689346313477, "perplexity": 140.28686281156808, "lr": 0.001305, "grad_norm": 0.336001, "tokens_per_sec": 106763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:25.640882+00:00", "epoch": 0, "step": 437, "train_loss": 4.840086460113525, "perplexity": 126.4802867573379, "lr": 0.001305, "grad_norm": 0.369402, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:25.947398+00:00", "epoch": 0, "step": 438, "train_loss": 4.817862510681152, "perplexity": 123.70039971031703, "lr": 0.001305, "grad_norm": 0.405661, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:26.254093+00:00", "epoch": 0, "step": 439, "train_loss": 4.930787086486816, "perplexity": 138.48847185883307, "lr": 0.001305, "grad_norm": 0.402309, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:26.561909+00:00", "epoch": 0, "step": 440, "train_loss": 4.906621932983398, "perplexity": 135.1819884193108, "lr": 0.001305, "grad_norm": 0.4042, "tokens_per_sec": 106510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:26.868637+00:00", "epoch": 0, "step": 441, "train_loss": 4.784417152404785, "perplexity": 119.63161576024594, "lr": 0.001305, "grad_norm": 0.444502, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:27.176465+00:00", "epoch": 0, "step": 442, "train_loss": 4.763276100158691, "perplexity": 117.12902443730094, "lr": 0.001305, "grad_norm": 0.487897, "tokens_per_sec": 106449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:27.483609+00:00", "epoch": 0, "step": 443, "train_loss": 4.86063814163208, "perplexity": 129.10656409880167, "lr": 0.001305, "grad_norm": 0.446439, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:27.789717+00:00", "epoch": 0, "step": 444, "train_loss": 4.799468517303467, "perplexity": 121.4458539930933, "lr": 0.001305, "grad_norm": 0.396756, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:28.096352+00:00", "epoch": 0, "step": 445, "train_loss": 4.9530415534973145, "perplexity": 141.60500879744328, "lr": 0.001305, "grad_norm": 0.369024, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:28.403927+00:00", "epoch": 0, "step": 446, "train_loss": 4.817770004272461, "perplexity": 123.68895715984893, "lr": 0.001305, "grad_norm": 0.368429, "tokens_per_sec": 106536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:28.712208+00:00", "epoch": 0, "step": 447, "train_loss": 4.802752494812012, "perplexity": 121.8453350304664, "lr": 0.001305, "grad_norm": 0.322927, "tokens_per_sec": 106297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:29.020452+00:00", "epoch": 0, "step": 448, "train_loss": 4.887153625488281, "perplexity": 132.5756764610348, "lr": 0.001305, "grad_norm": 0.348683, "tokens_per_sec": 106303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:29.327283+00:00", "epoch": 0, "step": 449, "train_loss": 4.824774742126465, "perplexity": 124.55840746084657, "lr": 0.001305, "grad_norm": 0.345016, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:29.633327+00:00", "epoch": 0, "step": 450, "train_loss": 4.832603454589844, "perplexity": 125.5373664118997, "lr": 0.001305, "grad_norm": 0.332616, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:29.939783+00:00", "epoch": 0, "step": 451, "train_loss": 4.796695232391357, "perplexity": 121.10951663380521, "lr": 0.001305, "grad_norm": 0.403405, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:30.246459+00:00", "epoch": 0, "step": 452, "train_loss": 4.827460289001465, "perplexity": 124.89336447282275, "lr": 0.001305, "grad_norm": 0.532724, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:30.553711+00:00", "epoch": 0, "step": 453, "train_loss": 4.769301891326904, "perplexity": 117.83695024275616, "lr": 0.001305, "grad_norm": 0.511801, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:30.860309+00:00", "epoch": 0, "step": 454, "train_loss": 4.80272102355957, "perplexity": 121.84150046550845, "lr": 0.001305, "grad_norm": 0.444983, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:31.166883+00:00", "epoch": 0, "step": 455, "train_loss": 4.798538684844971, "perplexity": 121.33298218017069, "lr": 0.001305, "grad_norm": 0.483089, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:31.472549+00:00", "epoch": 0, "step": 456, "train_loss": 4.91769552230835, "perplexity": 136.68725722563252, "lr": 0.001305, "grad_norm": 0.482934, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:31.778953+00:00", "epoch": 0, "step": 457, "train_loss": 4.712034702301025, "perplexity": 111.27834802570212, "lr": 0.001305, "grad_norm": 0.416233, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:32.085570+00:00", "epoch": 0, "step": 458, "train_loss": 4.642213821411133, "perplexity": 103.77383020000326, "lr": 0.001305, "grad_norm": 0.411002, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:32.392583+00:00", "epoch": 0, "step": 459, "train_loss": 4.838492393493652, "perplexity": 126.27882936505763, "lr": 0.001305, "grad_norm": 0.401581, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:32.700272+00:00", "epoch": 0, "step": 460, "train_loss": 4.804568290710449, "perplexity": 122.06678228071809, "lr": 0.001305, "grad_norm": 0.363299, "tokens_per_sec": 106561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:33.007502+00:00", "epoch": 0, "step": 461, "train_loss": 4.813382625579834, "perplexity": 123.14747557566047, "lr": 0.001305, "grad_norm": 0.343062, "tokens_per_sec": 106592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:33.313822+00:00", "epoch": 0, "step": 462, "train_loss": 4.806334972381592, "perplexity": 122.28262603509081, "lr": 0.001305, "grad_norm": 0.354891, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:33.619471+00:00", "epoch": 0, "step": 463, "train_loss": 4.681637287139893, "perplexity": 107.946667654266, "lr": 0.001305, "grad_norm": 0.320575, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:33.925960+00:00", "epoch": 0, "step": 464, "train_loss": 4.690989971160889, "perplexity": 108.9609946741955, "lr": 0.001305, "grad_norm": 0.297874, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:34.234579+00:00", "epoch": 0, "step": 465, "train_loss": 4.788372993469238, "perplexity": 120.10579669232014, "lr": 0.001305, "grad_norm": 0.285566, "tokens_per_sec": 106176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:34.542080+00:00", "epoch": 0, "step": 466, "train_loss": 4.827771186828613, "perplexity": 124.9321995850258, "lr": 0.001305, "grad_norm": 0.291314, "tokens_per_sec": 106565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:34.848486+00:00", "epoch": 0, "step": 467, "train_loss": 4.738306522369385, "perplexity": 114.24057391785296, "lr": 0.001305, "grad_norm": 0.291067, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:35.154545+00:00", "epoch": 0, "step": 468, "train_loss": 4.704989433288574, "perplexity": 110.49711735015069, "lr": 0.001305, "grad_norm": 0.279949, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:35.461033+00:00", "epoch": 0, "step": 469, "train_loss": 4.7555718421936035, "perplexity": 116.23009943922276, "lr": 0.001305, "grad_norm": 0.31437, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:35.767985+00:00", "epoch": 0, "step": 470, "train_loss": 4.719208240509033, "perplexity": 112.0794775380877, "lr": 0.001305, "grad_norm": 0.373493, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:36.074992+00:00", "epoch": 0, "step": 471, "train_loss": 4.773927211761475, "perplexity": 118.38324631993017, "lr": 0.001305, "grad_norm": 0.45724, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:36.382383+00:00", "epoch": 0, "step": 472, "train_loss": 4.762250900268555, "perplexity": 117.00900530662781, "lr": 0.001305, "grad_norm": 0.60703, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:36.688847+00:00", "epoch": 0, "step": 473, "train_loss": 4.862874507904053, "perplexity": 129.39561675758023, "lr": 0.001305, "grad_norm": 0.633051, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:36.995478+00:00", "epoch": 0, "step": 474, "train_loss": 4.735198497772217, "perplexity": 113.88606260443888, "lr": 0.001305, "grad_norm": 0.68835, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:37.302416+00:00", "epoch": 0, "step": 475, "train_loss": 4.81614875793457, "perplexity": 123.48858935768664, "lr": 0.001305, "grad_norm": 0.541508, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:37.610546+00:00", "epoch": 0, "step": 476, "train_loss": 4.765495300292969, "perplexity": 117.38924581892658, "lr": 0.001305, "grad_norm": 0.502195, "tokens_per_sec": 106344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:37.918561+00:00", "epoch": 0, "step": 477, "train_loss": 4.714959621429443, "perplexity": 111.60430466043522, "lr": 0.001305, "grad_norm": 0.493966, "tokens_per_sec": 106384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:38.226215+00:00", "epoch": 0, "step": 478, "train_loss": 4.800655364990234, "perplexity": 121.5900772925877, "lr": 0.001305, "grad_norm": 0.415557, "tokens_per_sec": 106509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:38.533395+00:00", "epoch": 0, "step": 479, "train_loss": 4.833443641662598, "perplexity": 125.64288560603067, "lr": 0.001305, "grad_norm": 0.347696, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:38.839710+00:00", "epoch": 0, "step": 480, "train_loss": 4.75838041305542, "perplexity": 116.55699875483637, "lr": 0.001305, "grad_norm": 0.360175, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:39.147364+00:00", "epoch": 0, "step": 481, "train_loss": 4.716558456420898, "perplexity": 111.78288424952972, "lr": 0.001305, "grad_norm": 0.392436, "tokens_per_sec": 106442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:39.455210+00:00", "epoch": 0, "step": 482, "train_loss": 4.791011810302734, "perplexity": 120.42315242810619, "lr": 0.001305, "grad_norm": 0.459296, "tokens_per_sec": 106443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:39.762865+00:00", "epoch": 0, "step": 483, "train_loss": 4.6928606033325195, "perplexity": 109.1650113768974, "lr": 0.001305, "grad_norm": 0.511562, "tokens_per_sec": 106509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:40.069633+00:00", "epoch": 0, "step": 484, "train_loss": 4.772721767425537, "perplexity": 118.24062788274742, "lr": 0.001305, "grad_norm": 0.552951, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:40.376008+00:00", "epoch": 0, "step": 485, "train_loss": 4.660079479217529, "perplexity": 105.64447836346972, "lr": 0.001305, "grad_norm": 0.491717, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:40.684252+00:00", "epoch": 0, "step": 486, "train_loss": 4.612979888916016, "perplexity": 100.78402782006066, "lr": 0.001305, "grad_norm": 0.427445, "tokens_per_sec": 106308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:40.991707+00:00", "epoch": 0, "step": 487, "train_loss": 4.790670871734619, "perplexity": 120.38210252909396, "lr": 0.001305, "grad_norm": 0.345839, "tokens_per_sec": 106575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:41.299621+00:00", "epoch": 0, "step": 488, "train_loss": 4.812117099761963, "perplexity": 122.99172783809185, "lr": 0.001305, "grad_norm": 0.389238, "tokens_per_sec": 106420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:41.607300+00:00", "epoch": 0, "step": 489, "train_loss": 4.767768383026123, "perplexity": 117.65638478611898, "lr": 0.001305, "grad_norm": 0.401689, "tokens_per_sec": 106500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:41.914759+00:00", "epoch": 0, "step": 490, "train_loss": 4.834953308105469, "perplexity": 125.83270770218809, "lr": 0.001305, "grad_norm": 0.45533, "tokens_per_sec": 106636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:42.221303+00:00", "epoch": 0, "step": 491, "train_loss": 4.733360290527344, "perplexity": 113.6769087120638, "lr": 0.001305, "grad_norm": 0.507234, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:42.527355+00:00", "epoch": 0, "step": 492, "train_loss": 4.716344356536865, "perplexity": 111.75895410878658, "lr": 0.001305, "grad_norm": 0.464569, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:42.834323+00:00", "epoch": 0, "step": 493, "train_loss": 4.613525390625, "perplexity": 100.83902067745997, "lr": 0.001305, "grad_norm": 0.419576, "tokens_per_sec": 106747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:43.142733+00:00", "epoch": 0, "step": 494, "train_loss": 4.7046380043029785, "perplexity": 110.45829228281568, "lr": 0.001305, "grad_norm": 0.414579, "tokens_per_sec": 106257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:43.449463+00:00", "epoch": 0, "step": 495, "train_loss": 4.707397937774658, "perplexity": 110.76357090146901, "lr": 0.001305, "grad_norm": 0.421119, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:43.755461+00:00", "epoch": 0, "step": 496, "train_loss": 4.711212158203125, "perplexity": 111.18685431127543, "lr": 0.001305, "grad_norm": 0.391925, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:44.061242+00:00", "epoch": 0, "step": 497, "train_loss": 4.65458345413208, "perplexity": 105.06544630513737, "lr": 0.001305, "grad_norm": 0.354361, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:44.368048+00:00", "epoch": 0, "step": 498, "train_loss": 4.652238845825195, "perplexity": 104.8193975436635, "lr": 0.001305, "grad_norm": 0.316282, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:44.676337+00:00", "epoch": 0, "step": 499, "train_loss": 4.729290008544922, "perplexity": 113.21515201619634, "lr": 0.001305, "grad_norm": 0.30561, "tokens_per_sec": 106291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:44.985571+00:00", "epoch": 0, "step": 500, "train_loss": 4.660125732421875, "perplexity": 105.64936487212293, "lr": 0.001305, "grad_norm": 0.301724, "tokens_per_sec": 106018} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:30:50.567447+00:00", "step": 500, "epoch": 0, "val_loss": 4.689027142524719, "val_ppl": 108.74733267316843, "eval_train_loss": 4.660125732421875, "eval_train_ppl": 105.64936487212293} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:30:51.484663+00:00", "step": 500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_4p6890_epoch_0000_step_0000500.pt", "val_loss": 4.689027142524719} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:30:53.285234+00:00", "step": 500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0000_step_0000500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:54.304919+00:00", "epoch": 0, "step": 501, "train_loss": 4.802831649780273, "perplexity": 121.85498007581518, "lr": 0.001305, "grad_norm": 0.332137, "tokens_per_sec": 3516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:54.610379+00:00", "epoch": 0, "step": 502, "train_loss": 4.774619102478027, "perplexity": 118.46518293137602, "lr": 0.001305, "grad_norm": 0.370751, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:54.915702+00:00", "epoch": 0, "step": 503, "train_loss": 4.703454494476318, "perplexity": 110.32764113717045, "lr": 0.001305, "grad_norm": 0.396925, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:55.221854+00:00", "epoch": 0, "step": 504, "train_loss": 4.753448009490967, "perplexity": 115.98350810506376, "lr": 0.001305, "grad_norm": 0.418651, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:55.528154+00:00", "epoch": 0, "step": 505, "train_loss": 4.680428981781006, "perplexity": 107.81631388670306, "lr": 0.001305, "grad_norm": 0.385551, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:55.833663+00:00", "epoch": 0, "step": 506, "train_loss": 4.72589111328125, "perplexity": 112.83099879034962, "lr": 0.001305, "grad_norm": 0.399662, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:56.139614+00:00", "epoch": 0, "step": 507, "train_loss": 4.69400691986084, "perplexity": 109.29022078484905, "lr": 0.001305, "grad_norm": 0.454082, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:56.446058+00:00", "epoch": 0, "step": 508, "train_loss": 4.722081184387207, "perplexity": 112.40193857138122, "lr": 0.001305, "grad_norm": 0.511974, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:56.752692+00:00", "epoch": 0, "step": 509, "train_loss": 4.603366374969482, "perplexity": 99.81978148707405, "lr": 0.001305, "grad_norm": 0.542573, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:57.058841+00:00", "epoch": 0, "step": 510, "train_loss": 4.630163669586182, "perplexity": 102.53084391807886, "lr": 0.001305, "grad_norm": 0.487145, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:57.364459+00:00", "epoch": 0, "step": 511, "train_loss": 4.6594719886779785, "perplexity": 105.58031983213166, "lr": 0.001305, "grad_norm": 0.49268, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:57.671409+00:00", "epoch": 0, "step": 512, "train_loss": 4.638516426086426, "perplexity": 103.39084578409559, "lr": 0.001305, "grad_norm": 0.505047, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:57.977976+00:00", "epoch": 0, "step": 513, "train_loss": 4.6741485595703125, "perplexity": 107.14130380718184, "lr": 0.001305, "grad_norm": 0.386419, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:58.285392+00:00", "epoch": 0, "step": 514, "train_loss": 4.6650166511535645, "perplexity": 106.16735301615581, "lr": 0.001305, "grad_norm": 0.343821, "tokens_per_sec": 106593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:58.819026+00:00", "epoch": 0, "step": 515, "train_loss": 4.579275131225586, "perplexity": 97.44373468017365, "lr": 0.001305, "grad_norm": 0.34331, "tokens_per_sec": 61405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:59.125274+00:00", "epoch": 0, "step": 516, "train_loss": 4.542873382568359, "perplexity": 93.96039677302548, "lr": 0.001305, "grad_norm": 0.325196, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:59.431993+00:00", "epoch": 0, "step": 517, "train_loss": 4.644171237945557, "perplexity": 103.97715774450556, "lr": 0.001305, "grad_norm": 0.328694, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:30:59.738406+00:00", "epoch": 0, "step": 518, "train_loss": 4.5858869552612305, "perplexity": 98.09015014500717, "lr": 0.001305, "grad_norm": 0.303377, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:00.045935+00:00", "epoch": 0, "step": 519, "train_loss": 4.704150199890137, "perplexity": 110.40442338021754, "lr": 0.001305, "grad_norm": 0.325022, "tokens_per_sec": 106552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:00.353966+00:00", "epoch": 0, "step": 520, "train_loss": 4.752544403076172, "perplexity": 115.87875199940196, "lr": 0.001305, "grad_norm": 0.307511, "tokens_per_sec": 106433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:00.660300+00:00", "epoch": 0, "step": 521, "train_loss": 4.651666641235352, "perplexity": 104.75943655989545, "lr": 0.001305, "grad_norm": 0.30764, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:00.967385+00:00", "epoch": 0, "step": 522, "train_loss": 4.651470184326172, "perplexity": 104.73885786626094, "lr": 0.001305, "grad_norm": 0.307848, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:01.273382+00:00", "epoch": 0, "step": 523, "train_loss": 4.536831855773926, "perplexity": 93.39444384962246, "lr": 0.001305, "grad_norm": 0.307333, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:01.579979+00:00", "epoch": 0, "step": 524, "train_loss": 4.535733699798584, "perplexity": 93.29193847679045, "lr": 0.001305, "grad_norm": 0.32676, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:01.887152+00:00", "epoch": 0, "step": 525, "train_loss": 4.642602443695068, "perplexity": 103.81416686026319, "lr": 0.001305, "grad_norm": 0.340492, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:02.193599+00:00", "epoch": 0, "step": 526, "train_loss": 4.5286736488342285, "perplexity": 92.6356122111987, "lr": 0.001305, "grad_norm": 0.363895, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:02.499675+00:00", "epoch": 0, "step": 527, "train_loss": 4.609057426452637, "perplexity": 100.3894805583062, "lr": 0.001305, "grad_norm": 0.383271, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:02.805735+00:00", "epoch": 0, "step": 528, "train_loss": 4.5208516120910645, "perplexity": 91.91383959480982, "lr": 0.001305, "grad_norm": 0.309412, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:03.111890+00:00", "epoch": 0, "step": 529, "train_loss": 4.498709201812744, "perplexity": 89.9010123098312, "lr": 0.001305, "grad_norm": 0.273634, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:03.419800+00:00", "epoch": 0, "step": 530, "train_loss": 4.635737419128418, "perplexity": 103.10392077222673, "lr": 0.001305, "grad_norm": 0.29168, "tokens_per_sec": 106480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:03.725784+00:00", "epoch": 0, "step": 531, "train_loss": 4.633050918579102, "perplexity": 102.82730376467548, "lr": 0.001305, "grad_norm": 0.259532, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:04.031337+00:00", "epoch": 0, "step": 532, "train_loss": 4.5242743492126465, "perplexity": 92.22897551162373, "lr": 0.001305, "grad_norm": 0.275017, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:04.336895+00:00", "epoch": 0, "step": 533, "train_loss": 4.54630708694458, "perplexity": 94.28358354499336, "lr": 0.001305, "grad_norm": 0.327154, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:04.643425+00:00", "epoch": 0, "step": 534, "train_loss": 4.59013557434082, "perplexity": 98.50778438461452, "lr": 0.001305, "grad_norm": 0.326123, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:04.950813+00:00", "epoch": 0, "step": 535, "train_loss": 4.632064342498779, "perplexity": 102.72590683249557, "lr": 0.001305, "grad_norm": 0.368391, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:05.257346+00:00", "epoch": 0, "step": 536, "train_loss": 4.7286906242370605, "perplexity": 113.14731296352852, "lr": 0.001305, "grad_norm": 0.379853, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:05.562902+00:00", "epoch": 0, "step": 537, "train_loss": 4.5913166999816895, "perplexity": 98.62420319366237, "lr": 0.001305, "grad_norm": 0.378424, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:05.869088+00:00", "epoch": 0, "step": 538, "train_loss": 4.443244934082031, "perplexity": 85.0504775775165, "lr": 0.001305, "grad_norm": 0.416733, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:06.175926+00:00", "epoch": 0, "step": 539, "train_loss": 4.525904655456543, "perplexity": 92.37945962052873, "lr": 0.001305, "grad_norm": 0.461038, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:06.482711+00:00", "epoch": 0, "step": 540, "train_loss": 4.664113998413086, "perplexity": 106.07156400261815, "lr": 0.001305, "grad_norm": 0.447608, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:06.790236+00:00", "epoch": 0, "step": 541, "train_loss": 4.483829975128174, "perplexity": 88.57325724880563, "lr": 0.001305, "grad_norm": 0.417083, "tokens_per_sec": 106496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:07.097665+00:00", "epoch": 0, "step": 542, "train_loss": 4.548230171203613, "perplexity": 94.46507327448205, "lr": 0.001305, "grad_norm": 0.401439, "tokens_per_sec": 106593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:07.404649+00:00", "epoch": 0, "step": 543, "train_loss": 4.619208812713623, "perplexity": 101.41376309978672, "lr": 0.001305, "grad_norm": 0.374475, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:07.710812+00:00", "epoch": 0, "step": 544, "train_loss": 4.552611351013184, "perplexity": 94.8798496877854, "lr": 0.001305, "grad_norm": 0.369771, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:08.016276+00:00", "epoch": 0, "step": 545, "train_loss": 4.708656311035156, "perplexity": 110.9030405513635, "lr": 0.001305, "grad_norm": 0.440925, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:08.323488+00:00", "epoch": 0, "step": 546, "train_loss": 4.502340793609619, "perplexity": 90.22808963486119, "lr": 0.001305, "grad_norm": 0.382443, "tokens_per_sec": 106664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:08.630829+00:00", "epoch": 0, "step": 547, "train_loss": 4.647242546081543, "perplexity": 104.29699454230607, "lr": 0.001305, "grad_norm": 0.310679, "tokens_per_sec": 106617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:08.937976+00:00", "epoch": 0, "step": 548, "train_loss": 4.450527191162109, "perplexity": 85.67209767157888, "lr": 0.001305, "grad_norm": 0.324965, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:09.243503+00:00", "epoch": 0, "step": 549, "train_loss": 4.434769630432129, "perplexity": 84.3326949638487, "lr": 0.001305, "grad_norm": 0.329219, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:09.549926+00:00", "epoch": 0, "step": 550, "train_loss": 4.5351176261901855, "perplexity": 93.23448147630668, "lr": 0.001305, "grad_norm": 0.281463, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:09.856828+00:00", "epoch": 0, "step": 551, "train_loss": 4.552611351013184, "perplexity": 94.8798496877854, "lr": 0.001305, "grad_norm": 0.305675, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:10.163857+00:00", "epoch": 0, "step": 552, "train_loss": 4.605669975280762, "perplexity": 100.04999142081485, "lr": 0.001305, "grad_norm": 0.302583, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:10.471344+00:00", "epoch": 0, "step": 553, "train_loss": 4.495558261871338, "perplexity": 89.61818543861582, "lr": 0.001305, "grad_norm": 0.306047, "tokens_per_sec": 106566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:10.778030+00:00", "epoch": 0, "step": 554, "train_loss": 4.522019863128662, "perplexity": 92.02128078020445, "lr": 0.001305, "grad_norm": 0.331841, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:11.084378+00:00", "epoch": 0, "step": 555, "train_loss": 4.525327205657959, "perplexity": 92.32613051909888, "lr": 0.001305, "grad_norm": 0.417536, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:11.390779+00:00", "epoch": 0, "step": 556, "train_loss": 4.494176387786865, "perplexity": 89.49442991762776, "lr": 0.001305, "grad_norm": 0.492936, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:11.698167+00:00", "epoch": 0, "step": 557, "train_loss": 4.481117248535156, "perplexity": 88.33330782417588, "lr": 0.001305, "grad_norm": 0.494595, "tokens_per_sec": 106600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:12.005393+00:00", "epoch": 0, "step": 558, "train_loss": 4.425510406494141, "perplexity": 83.55544358120294, "lr": 0.001305, "grad_norm": 0.480832, "tokens_per_sec": 106658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:12.310913+00:00", "epoch": 0, "step": 559, "train_loss": 4.504085063934326, "perplexity": 90.38560915243885, "lr": 0.001305, "grad_norm": 0.49084, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:12.617250+00:00", "epoch": 0, "step": 560, "train_loss": 4.485213279724121, "perplexity": 88.69586582558853, "lr": 0.001305, "grad_norm": 0.506865, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:12.923268+00:00", "epoch": 0, "step": 561, "train_loss": 4.615288257598877, "perplexity": 101.01694323751539, "lr": 0.001305, "grad_norm": 0.40108, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:13.230356+00:00", "epoch": 0, "step": 562, "train_loss": 4.4293928146362305, "perplexity": 83.88047045089793, "lr": 0.001305, "grad_norm": 0.42903, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:13.537644+00:00", "epoch": 0, "step": 563, "train_loss": 4.446342945098877, "perplexity": 85.31437345914846, "lr": 0.001305, "grad_norm": 0.417806, "tokens_per_sec": 106636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:13.844762+00:00", "epoch": 0, "step": 564, "train_loss": 4.51588249206543, "perplexity": 91.45824159190049, "lr": 0.001305, "grad_norm": 0.356757, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:14.151844+00:00", "epoch": 0, "step": 565, "train_loss": 4.487513065338135, "perplexity": 88.90008203858258, "lr": 0.001305, "grad_norm": 0.315643, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:14.459278+00:00", "epoch": 0, "step": 566, "train_loss": 4.562902450561523, "perplexity": 95.86130915266688, "lr": 0.001305, "grad_norm": 0.358658, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:14.766073+00:00", "epoch": 0, "step": 567, "train_loss": 4.562671661376953, "perplexity": 95.83918795206081, "lr": 0.001305, "grad_norm": 0.372995, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:15.073651+00:00", "epoch": 0, "step": 568, "train_loss": 4.512844085693359, "perplexity": 91.18077602789963, "lr": 0.001305, "grad_norm": 0.4064, "tokens_per_sec": 106536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:15.381488+00:00", "epoch": 0, "step": 569, "train_loss": 4.546321868896484, "perplexity": 94.28497725069151, "lr": 0.001305, "grad_norm": 0.384431, "tokens_per_sec": 106445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:15.689681+00:00", "epoch": 0, "step": 570, "train_loss": 4.537327289581299, "perplexity": 93.44072607847012, "lr": 0.001305, "grad_norm": 0.396378, "tokens_per_sec": 106380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:15.996453+00:00", "epoch": 0, "step": 571, "train_loss": 4.600103378295898, "perplexity": 99.49460069256983, "lr": 0.001305, "grad_norm": 0.391086, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:16.303265+00:00", "epoch": 0, "step": 572, "train_loss": 4.480236530303955, "perplexity": 88.25554531800773, "lr": 0.001305, "grad_norm": 0.334952, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:16.609594+00:00", "epoch": 0, "step": 573, "train_loss": 4.449276924133301, "perplexity": 85.56505160459955, "lr": 0.001305, "grad_norm": 0.301828, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:16.918417+00:00", "epoch": 0, "step": 574, "train_loss": 4.410360813140869, "perplexity": 82.29915276354161, "lr": 0.001305, "grad_norm": 0.292272, "tokens_per_sec": 106106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:17.226776+00:00", "epoch": 0, "step": 575, "train_loss": 4.53280782699585, "perplexity": 93.01937706574041, "lr": 0.001305, "grad_norm": 0.277152, "tokens_per_sec": 106267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:17.535246+00:00", "epoch": 0, "step": 576, "train_loss": 4.478015422821045, "perplexity": 88.05973780116243, "lr": 0.001305, "grad_norm": 0.288, "tokens_per_sec": 106227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:17.841893+00:00", "epoch": 0, "step": 577, "train_loss": 4.5214152336120605, "perplexity": 91.96565881472482, "lr": 0.001305, "grad_norm": 0.3029, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:18.149642+00:00", "epoch": 0, "step": 578, "train_loss": 4.399893760681152, "perplexity": 81.44221583980486, "lr": 0.001305, "grad_norm": 0.317842, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:18.456797+00:00", "epoch": 0, "step": 579, "train_loss": 4.3829193115234375, "perplexity": 80.07144603232955, "lr": 0.001305, "grad_norm": 0.337997, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:18.763707+00:00", "epoch": 0, "step": 580, "train_loss": 4.348398685455322, "perplexity": 77.35449481813588, "lr": 0.001305, "grad_norm": 0.340568, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:19.072323+00:00", "epoch": 0, "step": 581, "train_loss": 4.522702693939209, "perplexity": 92.08413720365617, "lr": 0.001305, "grad_norm": 0.362886, "tokens_per_sec": 106119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:19.379917+00:00", "epoch": 0, "step": 582, "train_loss": 4.568612098693848, "perplexity": 96.4102090187994, "lr": 0.001305, "grad_norm": 0.34932, "tokens_per_sec": 106530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:19.687614+00:00", "epoch": 0, "step": 583, "train_loss": 4.447836875915527, "perplexity": 85.44192248174774, "lr": 0.001305, "grad_norm": 0.298579, "tokens_per_sec": 106496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:19.994887+00:00", "epoch": 0, "step": 584, "train_loss": 4.349475860595703, "perplexity": 77.43786405052184, "lr": 0.001305, "grad_norm": 0.359348, "tokens_per_sec": 106641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:20.301352+00:00", "epoch": 0, "step": 585, "train_loss": 4.456638813018799, "perplexity": 86.19729640949566, "lr": 0.001305, "grad_norm": 0.376887, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:20.610047+00:00", "epoch": 0, "step": 586, "train_loss": 4.384339332580566, "perplexity": 80.1852299404229, "lr": 0.001305, "grad_norm": 0.480652, "tokens_per_sec": 106150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:20.917067+00:00", "epoch": 0, "step": 587, "train_loss": 4.448451042175293, "perplexity": 85.49441414535991, "lr": 0.001305, "grad_norm": 0.526434, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:21.225945+00:00", "epoch": 0, "step": 588, "train_loss": 4.462789058685303, "perplexity": 86.7290645343056, "lr": 0.001305, "grad_norm": 0.520237, "tokens_per_sec": 106088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:21.533726+00:00", "epoch": 0, "step": 589, "train_loss": 4.448687553405762, "perplexity": 85.5146369258107, "lr": 0.001305, "grad_norm": 0.496416, "tokens_per_sec": 106465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:21.840767+00:00", "epoch": 0, "step": 590, "train_loss": 4.4593119621276855, "perplexity": 86.42802288132009, "lr": 0.001305, "grad_norm": 0.443034, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:22.159055+00:00", "epoch": 0, "step": 591, "train_loss": 4.584646224975586, "perplexity": 97.96852219434278, "lr": 0.001305, "grad_norm": 0.489963, "tokens_per_sec": 102901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:22.466769+00:00", "epoch": 0, "step": 592, "train_loss": 4.527966499328613, "perplexity": 92.5701281400643, "lr": 0.001305, "grad_norm": 0.504784, "tokens_per_sec": 106488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:22.776536+00:00", "epoch": 0, "step": 593, "train_loss": 4.491466522216797, "perplexity": 89.25224034208449, "lr": 0.001305, "grad_norm": 0.424804, "tokens_per_sec": 105782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:23.085004+00:00", "epoch": 0, "step": 594, "train_loss": 4.471937656402588, "perplexity": 87.52615442401482, "lr": 0.001305, "grad_norm": 0.405551, "tokens_per_sec": 106229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:23.392450+00:00", "epoch": 0, "step": 595, "train_loss": 4.452552795410156, "perplexity": 85.84581131471163, "lr": 0.001305, "grad_norm": 0.409784, "tokens_per_sec": 106581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:23.699416+00:00", "epoch": 0, "step": 596, "train_loss": 4.464187145233154, "perplexity": 86.8504040745636, "lr": 0.001305, "grad_norm": 0.423551, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:24.017902+00:00", "epoch": 0, "step": 597, "train_loss": 4.417353630065918, "perplexity": 82.87667256365341, "lr": 0.001305, "grad_norm": 0.354107, "tokens_per_sec": 102886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:24.326177+00:00", "epoch": 0, "step": 598, "train_loss": 4.44570255279541, "perplexity": 85.25975628108691, "lr": 0.001305, "grad_norm": 0.40751, "tokens_per_sec": 106297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:24.644750+00:00", "epoch": 0, "step": 599, "train_loss": 4.533431529998779, "perplexity": 93.07741162683065, "lr": 0.001305, "grad_norm": 0.407069, "tokens_per_sec": 102860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:24.954093+00:00", "epoch": 0, "step": 600, "train_loss": 4.492940425872803, "perplexity": 89.38388653850424, "lr": 0.001305, "grad_norm": 0.364917, "tokens_per_sec": 105985} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:31:30.515096+00:00", "step": 600, "epoch": 0, "val_loss": 4.454707050323487, "val_ppl": 86.03094441587422, "eval_train_loss": 4.492940425872803, "eval_train_ppl": 89.38388653850424} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:31:31.439103+00:00", "step": 600, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_4p4547_epoch_0000_step_0000600.pt", "val_loss": 4.454707050323487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:32.618164+00:00", "epoch": 0, "step": 601, "train_loss": 4.420554161071777, "perplexity": 83.1423468464617, "lr": 0.001305, "grad_norm": 0.306464, "tokens_per_sec": 4275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:32.923383+00:00", "epoch": 0, "step": 602, "train_loss": 4.357110023498535, "perplexity": 78.03129962933346, "lr": 0.001305, "grad_norm": 0.281825, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:33.230135+00:00", "epoch": 0, "step": 603, "train_loss": 4.478672981262207, "perplexity": 88.11766126702497, "lr": 0.001305, "grad_norm": 0.288693, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:33.536632+00:00", "epoch": 0, "step": 604, "train_loss": 4.370822906494141, "perplexity": 79.10870398794272, "lr": 0.001305, "grad_norm": 0.304667, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:33.843295+00:00", "epoch": 0, "step": 605, "train_loss": 4.491522789001465, "perplexity": 89.25726241995972, "lr": 0.001305, "grad_norm": 0.291588, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:34.151009+00:00", "epoch": 0, "step": 606, "train_loss": 4.426736831665039, "perplexity": 83.65798094472594, "lr": 0.001305, "grad_norm": 0.265328, "tokens_per_sec": 106490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:34.457966+00:00", "epoch": 0, "step": 607, "train_loss": 4.463079452514648, "perplexity": 86.75425377669455, "lr": 0.001305, "grad_norm": 0.255757, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:34.764222+00:00", "epoch": 0, "step": 608, "train_loss": 4.31564998626709, "perplexity": 74.86226707890751, "lr": 0.001305, "grad_norm": 0.258642, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:35.070714+00:00", "epoch": 0, "step": 609, "train_loss": 4.366811752319336, "perplexity": 78.79202233374801, "lr": 0.001305, "grad_norm": 0.270023, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:35.377756+00:00", "epoch": 0, "step": 610, "train_loss": 4.4222540855407715, "perplexity": 83.28380275439092, "lr": 0.001305, "grad_norm": 0.285648, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:35.685547+00:00", "epoch": 0, "step": 611, "train_loss": 4.469620227813721, "perplexity": 87.32355365855386, "lr": 0.001305, "grad_norm": 0.281033, "tokens_per_sec": 106392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:35.992480+00:00", "epoch": 0, "step": 612, "train_loss": 4.469875812530518, "perplexity": 87.34587507667054, "lr": 0.001305, "grad_norm": 0.30636, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:36.298805+00:00", "epoch": 0, "step": 613, "train_loss": 4.500936031341553, "perplexity": 90.10142960346771, "lr": 0.001305, "grad_norm": 0.335513, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:36.605549+00:00", "epoch": 0, "step": 614, "train_loss": 4.365800857543945, "perplexity": 78.71241213557458, "lr": 0.001305, "grad_norm": 0.370941, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:36.912952+00:00", "epoch": 0, "step": 615, "train_loss": 4.41871976852417, "perplexity": 82.98997094635246, "lr": 0.001305, "grad_norm": 0.420076, "tokens_per_sec": 106596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:37.220540+00:00", "epoch": 0, "step": 616, "train_loss": 4.320085525512695, "perplexity": 75.19505911300233, "lr": 0.001305, "grad_norm": 0.361855, "tokens_per_sec": 106531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:37.529188+00:00", "epoch": 0, "step": 617, "train_loss": 4.530252933502197, "perplexity": 92.7820257972062, "lr": 0.001305, "grad_norm": 0.40877, "tokens_per_sec": 106167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:37.835746+00:00", "epoch": 0, "step": 618, "train_loss": 4.353672504425049, "perplexity": 77.76352605060049, "lr": 0.001305, "grad_norm": 0.410134, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:38.142412+00:00", "epoch": 0, "step": 619, "train_loss": 4.369203567504883, "perplexity": 78.98070384496725, "lr": 0.001305, "grad_norm": 0.413727, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:38.449098+00:00", "epoch": 0, "step": 620, "train_loss": 4.459447383880615, "perplexity": 86.43972790822068, "lr": 0.001305, "grad_norm": 0.446471, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:38.757312+00:00", "epoch": 0, "step": 621, "train_loss": 4.426015377044678, "perplexity": 83.59764727446282, "lr": 0.001305, "grad_norm": 0.461919, "tokens_per_sec": 106260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:39.065788+00:00", "epoch": 0, "step": 622, "train_loss": 4.396190643310547, "perplexity": 81.14118347871722, "lr": 0.001305, "grad_norm": 0.340346, "tokens_per_sec": 106225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:39.373116+00:00", "epoch": 0, "step": 623, "train_loss": 4.32664680480957, "perplexity": 75.69005703156562, "lr": 0.001305, "grad_norm": 0.350536, "tokens_per_sec": 106622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:39.680009+00:00", "epoch": 0, "step": 624, "train_loss": 4.362576007843018, "perplexity": 78.45898528821562, "lr": 0.001305, "grad_norm": 0.31172, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:39.987145+00:00", "epoch": 0, "step": 625, "train_loss": 4.443763256072998, "perplexity": 85.09457253709357, "lr": 0.001305, "grad_norm": 0.31601, "tokens_per_sec": 106689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:40.294074+00:00", "epoch": 0, "step": 626, "train_loss": 4.362833499908447, "perplexity": 78.47919045561251, "lr": 0.001305, "grad_norm": 0.304509, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:40.600782+00:00", "epoch": 0, "step": 627, "train_loss": 4.44706392288208, "perplexity": 85.37590540591827, "lr": 0.001305, "grad_norm": 0.30798, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:40.909829+00:00", "epoch": 0, "step": 628, "train_loss": 4.346817970275879, "perplexity": 77.23231598438751, "lr": 0.001305, "grad_norm": 0.326087, "tokens_per_sec": 106029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:41.216758+00:00", "epoch": 0, "step": 629, "train_loss": 4.449141502380371, "perplexity": 85.55346501987724, "lr": 0.001305, "grad_norm": 0.316109, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:41.523696+00:00", "epoch": 0, "step": 630, "train_loss": 4.371119976043701, "perplexity": 79.13220826603253, "lr": 0.001305, "grad_norm": 0.291013, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:41.830710+00:00", "epoch": 0, "step": 631, "train_loss": 4.3087358474731445, "perplexity": 74.34644426408298, "lr": 0.001305, "grad_norm": 0.288123, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:42.138387+00:00", "epoch": 0, "step": 632, "train_loss": 4.414585113525391, "perplexity": 82.64754444368162, "lr": 0.001305, "grad_norm": 0.300045, "tokens_per_sec": 106501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:42.444881+00:00", "epoch": 0, "step": 633, "train_loss": 4.436816215515137, "perplexity": 84.5054657341078, "lr": 0.001305, "grad_norm": 0.287259, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:42.751520+00:00", "epoch": 0, "step": 634, "train_loss": 4.315843105316162, "perplexity": 74.87672580481994, "lr": 0.001305, "grad_norm": 0.344694, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:43.059348+00:00", "epoch": 0, "step": 635, "train_loss": 4.398799896240234, "perplexity": 81.35317780255768, "lr": 0.001305, "grad_norm": 0.373082, "tokens_per_sec": 106450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:43.379245+00:00", "epoch": 0, "step": 636, "train_loss": 4.4150214195251465, "perplexity": 82.68361193084586, "lr": 0.001305, "grad_norm": 0.377288, "tokens_per_sec": 102432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:43.684639+00:00", "epoch": 0, "step": 637, "train_loss": 4.429204940795898, "perplexity": 83.86471298503962, "lr": 0.001305, "grad_norm": 0.368217, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:43.991976+00:00", "epoch": 0, "step": 638, "train_loss": 4.339628219604492, "perplexity": 76.67902627960852, "lr": 0.001305, "grad_norm": 0.432012, "tokens_per_sec": 106623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:44.299919+00:00", "epoch": 0, "step": 639, "train_loss": 4.378809452056885, "perplexity": 79.74303895757423, "lr": 0.001305, "grad_norm": 0.415449, "tokens_per_sec": 106406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:44.607108+00:00", "epoch": 0, "step": 640, "train_loss": 4.374125957489014, "perplexity": 79.37043609063306, "lr": 0.001305, "grad_norm": 0.40386, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:44.924227+00:00", "epoch": 0, "step": 641, "train_loss": 4.4052019119262695, "perplexity": 81.87567284912836, "lr": 0.001305, "grad_norm": 0.396909, "tokens_per_sec": 103272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:45.231304+00:00", "epoch": 0, "step": 642, "train_loss": 4.376051425933838, "perplexity": 79.52340858537207, "lr": 0.001305, "grad_norm": 0.370273, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:45.550507+00:00", "epoch": 0, "step": 643, "train_loss": 4.31930685043335, "perplexity": 75.13652938515872, "lr": 0.001305, "grad_norm": 0.40463, "tokens_per_sec": 102656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:45.857046+00:00", "epoch": 0, "step": 644, "train_loss": 4.396832466125488, "perplexity": 81.19327845758988, "lr": 0.001305, "grad_norm": 0.408973, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:46.179185+00:00", "epoch": 0, "step": 645, "train_loss": 4.317887783050537, "perplexity": 75.02998120445461, "lr": 0.001305, "grad_norm": 0.39723, "tokens_per_sec": 101719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:46.485811+00:00", "epoch": 0, "step": 646, "train_loss": 4.3141679763793945, "perplexity": 74.75140263027727, "lr": 0.001305, "grad_norm": 0.336019, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:46.793924+00:00", "epoch": 0, "step": 647, "train_loss": 4.185210704803467, "perplexity": 65.70734431132105, "lr": 0.001305, "grad_norm": 0.313899, "tokens_per_sec": 106351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:47.112967+00:00", "epoch": 0, "step": 648, "train_loss": 4.352022171020508, "perplexity": 77.63529614605568, "lr": 0.001305, "grad_norm": 0.324906, "tokens_per_sec": 102709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:47.419832+00:00", "epoch": 0, "step": 649, "train_loss": 4.430088043212891, "perplexity": 83.93880682718732, "lr": 0.001305, "grad_norm": 0.360537, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:47.738453+00:00", "epoch": 0, "step": 650, "train_loss": 4.4338603019714355, "perplexity": 84.25604370002074, "lr": 0.001305, "grad_norm": 0.35923, "tokens_per_sec": 102896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:48.059422+00:00", "epoch": 0, "step": 651, "train_loss": 4.423399448394775, "perplexity": 83.3792475774493, "lr": 0.001305, "grad_norm": 0.340755, "tokens_per_sec": 102043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:48.366332+00:00", "epoch": 0, "step": 652, "train_loss": 4.493228912353516, "perplexity": 89.40967630118425, "lr": 0.001305, "grad_norm": 0.357672, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:48.684838+00:00", "epoch": 0, "step": 653, "train_loss": 4.385301113128662, "perplexity": 80.26238763326052, "lr": 0.001305, "grad_norm": 0.376295, "tokens_per_sec": 102880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:49.005386+00:00", "epoch": 0, "step": 654, "train_loss": 4.340756416320801, "perplexity": 76.76558412322328, "lr": 0.001305, "grad_norm": 0.351841, "tokens_per_sec": 102229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:49.313860+00:00", "epoch": 0, "step": 655, "train_loss": 4.352548122406006, "perplexity": 77.67613927743649, "lr": 0.001305, "grad_norm": 0.335395, "tokens_per_sec": 106222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:49.631253+00:00", "epoch": 0, "step": 656, "train_loss": 4.284222602844238, "perplexity": 72.54612763929113, "lr": 0.001305, "grad_norm": 0.347225, "tokens_per_sec": 103244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:49.937850+00:00", "epoch": 0, "step": 657, "train_loss": 4.336111545562744, "perplexity": 76.40984472764653, "lr": 0.001305, "grad_norm": 0.332064, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:50.258582+00:00", "epoch": 0, "step": 658, "train_loss": 4.300267219543457, "perplexity": 73.71949035636969, "lr": 0.001305, "grad_norm": 0.341484, "tokens_per_sec": 102170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:50.578597+00:00", "epoch": 0, "step": 659, "train_loss": 4.42257833480835, "perplexity": 83.31081184503745, "lr": 0.001305, "grad_norm": 0.329282, "tokens_per_sec": 102391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:50.886881+00:00", "epoch": 0, "step": 660, "train_loss": 4.239240646362305, "perplexity": 69.35516673981917, "lr": 0.001305, "grad_norm": 0.291925, "tokens_per_sec": 106347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:51.205504+00:00", "epoch": 0, "step": 661, "train_loss": 4.330893516540527, "perplexity": 76.0121743703154, "lr": 0.001305, "grad_norm": 0.341313, "tokens_per_sec": 102792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:51.514257+00:00", "epoch": 0, "step": 662, "train_loss": 4.394599437713623, "perplexity": 81.01217384107395, "lr": 0.001305, "grad_norm": 0.368135, "tokens_per_sec": 106129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:51.832688+00:00", "epoch": 0, "step": 663, "train_loss": 4.305535316467285, "perplexity": 74.10887653826833, "lr": 0.001305, "grad_norm": 0.332295, "tokens_per_sec": 102908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:52.141877+00:00", "epoch": 0, "step": 664, "train_loss": 4.353917598724365, "perplexity": 77.78258778339584, "lr": 0.001305, "grad_norm": 0.30705, "tokens_per_sec": 105977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:52.462069+00:00", "epoch": 0, "step": 665, "train_loss": 4.5321455001831055, "perplexity": 92.95778823642299, "lr": 0.001305, "grad_norm": 0.324212, "tokens_per_sec": 102340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:52.769060+00:00", "epoch": 0, "step": 666, "train_loss": 4.332763671875, "perplexity": 76.15446195216836, "lr": 0.001305, "grad_norm": 0.288679, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:53.089073+00:00", "epoch": 0, "step": 667, "train_loss": 4.363364219665527, "perplexity": 78.52085196682263, "lr": 0.001305, "grad_norm": 0.297403, "tokens_per_sec": 102397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:53.396245+00:00", "epoch": 0, "step": 668, "train_loss": 4.3155412673950195, "perplexity": 74.85412858008229, "lr": 0.001305, "grad_norm": 0.332366, "tokens_per_sec": 106676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:53.705479+00:00", "epoch": 0, "step": 669, "train_loss": 4.201035022735596, "perplexity": 66.7553886416344, "lr": 0.001305, "grad_norm": 0.36293, "tokens_per_sec": 105964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:54.023502+00:00", "epoch": 0, "step": 670, "train_loss": 4.364161968231201, "perplexity": 78.58351685594401, "lr": 0.001305, "grad_norm": 0.335851, "tokens_per_sec": 103091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:54.330688+00:00", "epoch": 0, "step": 671, "train_loss": 4.285852432250977, "perplexity": 72.66446185756232, "lr": 0.001305, "grad_norm": 0.328445, "tokens_per_sec": 106614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:54.648434+00:00", "epoch": 0, "step": 672, "train_loss": 4.342993259429932, "perplexity": 76.93748888139794, "lr": 0.001305, "grad_norm": 0.2827, "tokens_per_sec": 103127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:54.967320+00:00", "epoch": 0, "step": 673, "train_loss": 4.160429954528809, "perplexity": 64.09907636425454, "lr": 0.001305, "grad_norm": 0.312769, "tokens_per_sec": 102759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:55.274291+00:00", "epoch": 0, "step": 674, "train_loss": 4.294951915740967, "perplexity": 73.32868840460338, "lr": 0.001305, "grad_norm": 0.329094, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:55.595931+00:00", "epoch": 0, "step": 675, "train_loss": 4.314676761627197, "perplexity": 74.78944471799399, "lr": 0.001305, "grad_norm": 0.378442, "tokens_per_sec": 101878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:55.903344+00:00", "epoch": 0, "step": 676, "train_loss": 4.3153533935546875, "perplexity": 74.84006676844588, "lr": 0.001305, "grad_norm": 0.368361, "tokens_per_sec": 106593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:56.212549+00:00", "epoch": 0, "step": 677, "train_loss": 4.361978530883789, "perplexity": 78.41212185356679, "lr": 0.001305, "grad_norm": 0.294773, "tokens_per_sec": 105975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:56.530451+00:00", "epoch": 0, "step": 678, "train_loss": 4.417483806610107, "perplexity": 82.88746186472339, "lr": 0.001305, "grad_norm": 0.258991, "tokens_per_sec": 103076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:56.837330+00:00", "epoch": 0, "step": 679, "train_loss": 4.256101608276367, "perplexity": 70.53447576818742, "lr": 0.001305, "grad_norm": 0.288338, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:57.157572+00:00", "epoch": 0, "step": 680, "train_loss": 4.368643760681152, "perplexity": 78.93650228133461, "lr": 0.001305, "grad_norm": 0.313098, "tokens_per_sec": 102378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:57.464674+00:00", "epoch": 0, "step": 681, "train_loss": 4.336148262023926, "perplexity": 76.412650278249, "lr": 0.001305, "grad_norm": 0.323384, "tokens_per_sec": 106641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:57.787704+00:00", "epoch": 0, "step": 682, "train_loss": 4.33472204208374, "perplexity": 76.30374671141911, "lr": 0.001305, "grad_norm": 0.324471, "tokens_per_sec": 101440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:58.095894+00:00", "epoch": 0, "step": 683, "train_loss": 4.395893096923828, "perplexity": 81.1170438042736, "lr": 0.001305, "grad_norm": 0.298988, "tokens_per_sec": 106324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:58.404471+00:00", "epoch": 0, "step": 684, "train_loss": 4.289640426635742, "perplexity": 72.94023641753802, "lr": 0.001305, "grad_norm": 0.313882, "tokens_per_sec": 106191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:58.711278+00:00", "epoch": 0, "step": 685, "train_loss": 4.29581356048584, "perplexity": 73.3918989122057, "lr": 0.001305, "grad_norm": 0.318896, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:59.029193+00:00", "epoch": 0, "step": 686, "train_loss": 4.250446796417236, "perplexity": 70.13674219226343, "lr": 0.001305, "grad_norm": 0.334077, "tokens_per_sec": 103071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:59.350034+00:00", "epoch": 0, "step": 687, "train_loss": 4.264211177825928, "perplexity": 71.10880564152824, "lr": 0.001305, "grad_norm": 0.352032, "tokens_per_sec": 102132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:59.657580+00:00", "epoch": 0, "step": 688, "train_loss": 4.32258939743042, "perplexity": 75.38357381971804, "lr": 0.001305, "grad_norm": 0.397896, "tokens_per_sec": 106547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:31:59.981678+00:00", "epoch": 0, "step": 689, "train_loss": 4.319307327270508, "perplexity": 75.13656521305641, "lr": 0.001305, "grad_norm": 0.414285, "tokens_per_sec": 101105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:00.289798+00:00", "epoch": 0, "step": 690, "train_loss": 4.203848361968994, "perplexity": 66.94345862382454, "lr": 0.001305, "grad_norm": 0.434284, "tokens_per_sec": 106407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:00.597073+00:00", "epoch": 0, "step": 691, "train_loss": 4.202538013458252, "perplexity": 66.85579680882029, "lr": 0.001305, "grad_norm": 0.382184, "tokens_per_sec": 106583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:00.904955+00:00", "epoch": 0, "step": 692, "train_loss": 4.297893047332764, "perplexity": 73.54467519374462, "lr": 0.001305, "grad_norm": 0.391687, "tokens_per_sec": 106429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:01.224765+00:00", "epoch": 0, "step": 693, "train_loss": 4.323681354522705, "perplexity": 75.46593440672213, "lr": 0.001305, "grad_norm": 0.403453, "tokens_per_sec": 102461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:01.531913+00:00", "epoch": 0, "step": 694, "train_loss": 4.375258445739746, "perplexity": 79.46037309364858, "lr": 0.001305, "grad_norm": 0.362039, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:01.839414+00:00", "epoch": 0, "step": 695, "train_loss": 4.32761287689209, "perplexity": 75.763214414551, "lr": 0.001305, "grad_norm": 0.36796, "tokens_per_sec": 106562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:02.148418+00:00", "epoch": 0, "step": 696, "train_loss": 4.328186988830566, "perplexity": 75.80672346878242, "lr": 0.001305, "grad_norm": 0.349297, "tokens_per_sec": 106044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:02.456692+00:00", "epoch": 0, "step": 697, "train_loss": 4.267044544219971, "perplexity": 71.3105686410258, "lr": 0.001305, "grad_norm": 0.31514, "tokens_per_sec": 106296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:02.775732+00:00", "epoch": 0, "step": 698, "train_loss": 4.3096723556518555, "perplexity": 74.41610293005841, "lr": 0.001305, "grad_norm": 0.305205, "tokens_per_sec": 102708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:03.082312+00:00", "epoch": 0, "step": 699, "train_loss": 4.25549840927124, "perplexity": 70.49194227194785, "lr": 0.001305, "grad_norm": 0.295705, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:03.390184+00:00", "epoch": 0, "step": 700, "train_loss": 4.297580242156982, "perplexity": 73.52167363638353, "lr": 0.001305, "grad_norm": 0.292494, "tokens_per_sec": 106489} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:32:08.978821+00:00", "step": 700, "epoch": 0, "val_loss": 4.295912361145019, "val_ppl": 73.39915043841854, "eval_train_loss": 4.297580242156982, "eval_train_ppl": 73.52167363638353} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:32:09.908376+00:00", "step": 700, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_4p2959_epoch_0000_step_0000700.pt", "val_loss": 4.295912361145019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:10.924605+00:00", "epoch": 0, "step": 701, "train_loss": 4.300752639770508, "perplexity": 73.75528397488692, "lr": 0.001305, "grad_norm": 0.271632, "tokens_per_sec": 4349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:11.230667+00:00", "epoch": 0, "step": 702, "train_loss": 4.177489280700684, "perplexity": 65.20194375625184, "lr": 0.001305, "grad_norm": 0.265091, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:11.536491+00:00", "epoch": 0, "step": 703, "train_loss": 4.324515342712402, "perplexity": 75.52889835668766, "lr": 0.001305, "grad_norm": 0.287945, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:11.843181+00:00", "epoch": 0, "step": 704, "train_loss": 4.298485279083252, "perplexity": 73.58824358549275, "lr": 0.001305, "grad_norm": 0.308931, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:12.150408+00:00", "epoch": 0, "step": 705, "train_loss": 4.2825541496276855, "perplexity": 72.42518873778768, "lr": 0.001305, "grad_norm": 0.332058, "tokens_per_sec": 106657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:12.456873+00:00", "epoch": 0, "step": 706, "train_loss": 4.403700351715088, "perplexity": 81.7528238521794, "lr": 0.001305, "grad_norm": 0.374235, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:12.763309+00:00", "epoch": 0, "step": 707, "train_loss": 4.485370635986328, "perplexity": 88.70982377366452, "lr": 0.001305, "grad_norm": 0.404616, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:13.069744+00:00", "epoch": 0, "step": 708, "train_loss": 4.218634605407715, "perplexity": 67.9406551265169, "lr": 0.001305, "grad_norm": 0.351225, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:13.376052+00:00", "epoch": 0, "step": 709, "train_loss": 4.3060150146484375, "perplexity": 74.14443495952412, "lr": 0.001305, "grad_norm": 0.332871, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:13.799975+00:00", "epoch": 0, "step": 710, "train_loss": 4.235026836395264, "perplexity": 69.0635321248953, "lr": 0.001305, "grad_norm": 0.388548, "tokens_per_sec": 77328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:14.106046+00:00", "epoch": 0, "step": 711, "train_loss": 4.269808292388916, "perplexity": 71.50792569150458, "lr": 0.001305, "grad_norm": 0.493846, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:14.412140+00:00", "epoch": 0, "step": 712, "train_loss": 4.331814765930176, "perplexity": 76.08223280525023, "lr": 0.001305, "grad_norm": 0.494922, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:14.718362+00:00", "epoch": 0, "step": 713, "train_loss": 4.2455220222473145, "perplexity": 69.79218370870505, "lr": 0.001305, "grad_norm": 0.476932, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:15.025814+00:00", "epoch": 0, "step": 714, "train_loss": 4.225194931030273, "perplexity": 68.38783316011452, "lr": 0.001305, "grad_norm": 0.525668, "tokens_per_sec": 106581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:15.332794+00:00", "epoch": 0, "step": 715, "train_loss": 4.2685394287109375, "perplexity": 71.41724942198532, "lr": 0.001305, "grad_norm": 0.563713, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:15.639713+00:00", "epoch": 0, "step": 716, "train_loss": 4.195542812347412, "perplexity": 66.38975897926659, "lr": 0.001305, "grad_norm": 0.458193, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:15.945794+00:00", "epoch": 0, "step": 717, "train_loss": 4.299596786499023, "perplexity": 73.67008293804956, "lr": 0.001305, "grad_norm": 0.466146, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:16.251619+00:00", "epoch": 0, "step": 718, "train_loss": 4.314894676208496, "perplexity": 74.80574420440865, "lr": 0.001305, "grad_norm": 0.473773, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:16.558074+00:00", "epoch": 0, "step": 719, "train_loss": 4.289519786834717, "perplexity": 72.93143745269322, "lr": 0.001305, "grad_norm": 0.41705, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:16.865486+00:00", "epoch": 0, "step": 720, "train_loss": 4.322603225708008, "perplexity": 75.38461625190988, "lr": 0.001305, "grad_norm": 0.386347, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:17.172174+00:00", "epoch": 0, "step": 721, "train_loss": 4.32322359085083, "perplexity": 75.43139674913222, "lr": 0.001305, "grad_norm": 0.336926, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:17.478530+00:00", "epoch": 0, "step": 722, "train_loss": 4.285754680633545, "perplexity": 72.65735913604286, "lr": 0.001305, "grad_norm": 0.29452, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:17.784892+00:00", "epoch": 0, "step": 723, "train_loss": 4.273344993591309, "perplexity": 71.76127560567096, "lr": 0.001305, "grad_norm": 0.305667, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:18.090812+00:00", "epoch": 0, "step": 724, "train_loss": 4.207982540130615, "perplexity": 67.22078767739326, "lr": 0.001305, "grad_norm": 0.303105, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:18.397355+00:00", "epoch": 0, "step": 725, "train_loss": 4.413660049438477, "perplexity": 82.5711255200956, "lr": 0.001305, "grad_norm": 0.307344, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:18.705506+00:00", "epoch": 0, "step": 726, "train_loss": 4.352348327636719, "perplexity": 77.66062154134332, "lr": 0.001305, "grad_norm": 0.308337, "tokens_per_sec": 106338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:19.013254+00:00", "epoch": 0, "step": 727, "train_loss": 4.18131685256958, "perplexity": 65.4519871061491, "lr": 0.001305, "grad_norm": 0.265171, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:19.320268+00:00", "epoch": 0, "step": 728, "train_loss": 4.251908302307129, "perplexity": 70.23932239660243, "lr": 0.001305, "grad_norm": 0.278664, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:19.626375+00:00", "epoch": 0, "step": 729, "train_loss": 4.282997131347656, "perplexity": 72.4572788796123, "lr": 0.001305, "grad_norm": 0.275292, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:19.933073+00:00", "epoch": 0, "step": 730, "train_loss": 4.185741424560547, "perplexity": 65.74222575244858, "lr": 0.001305, "grad_norm": 0.294701, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:20.239782+00:00", "epoch": 0, "step": 731, "train_loss": 4.204827785491943, "perplexity": 67.00905674083387, "lr": 0.001305, "grad_norm": 0.30228, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:20.547898+00:00", "epoch": 0, "step": 732, "train_loss": 4.1995930671691895, "perplexity": 66.65919970414853, "lr": 0.001305, "grad_norm": 0.311536, "tokens_per_sec": 106350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:20.855984+00:00", "epoch": 0, "step": 733, "train_loss": 4.258439540863037, "perplexity": 70.69957353609428, "lr": 0.001305, "grad_norm": 0.257633, "tokens_per_sec": 106360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:21.161997+00:00", "epoch": 0, "step": 734, "train_loss": 4.25071907043457, "perplexity": 70.15584120478727, "lr": 0.001305, "grad_norm": 0.265799, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:21.467295+00:00", "epoch": 0, "step": 735, "train_loss": 4.221487522125244, "perplexity": 68.13476090948944, "lr": 0.001305, "grad_norm": 0.288934, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:21.772791+00:00", "epoch": 0, "step": 736, "train_loss": 4.2542572021484375, "perplexity": 70.40450144841212, "lr": 0.001305, "grad_norm": 0.294891, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:22.080468+00:00", "epoch": 0, "step": 737, "train_loss": 4.187510013580322, "perplexity": 65.85859960947893, "lr": 0.001305, "grad_norm": 0.256921, "tokens_per_sec": 106502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:22.387968+00:00", "epoch": 0, "step": 738, "train_loss": 4.29651403427124, "perplexity": 73.44332602302318, "lr": 0.001305, "grad_norm": 0.243552, "tokens_per_sec": 106562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:22.694473+00:00", "epoch": 0, "step": 739, "train_loss": 4.195700645446777, "perplexity": 66.40023830766307, "lr": 0.001305, "grad_norm": 0.260233, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:23.000584+00:00", "epoch": 0, "step": 740, "train_loss": 4.1703314781188965, "perplexity": 64.73690741951793, "lr": 0.001305, "grad_norm": 0.267948, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:23.306823+00:00", "epoch": 0, "step": 741, "train_loss": 4.226687431335449, "perplexity": 68.48997822878823, "lr": 0.001305, "grad_norm": 0.328739, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:23.613287+00:00", "epoch": 0, "step": 742, "train_loss": 4.245528221130371, "perplexity": 69.79261634363105, "lr": 0.001305, "grad_norm": 0.388465, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:23.920701+00:00", "epoch": 0, "step": 743, "train_loss": 4.257235050201416, "perplexity": 70.61446782479598, "lr": 0.001305, "grad_norm": 0.379867, "tokens_per_sec": 106593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:24.227336+00:00", "epoch": 0, "step": 744, "train_loss": 4.295414447784424, "perplexity": 73.36261311772382, "lr": 0.001305, "grad_norm": 0.310171, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:24.533233+00:00", "epoch": 0, "step": 745, "train_loss": 4.280832290649414, "perplexity": 72.30059007775141, "lr": 0.001305, "grad_norm": 0.317835, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:24.839975+00:00", "epoch": 0, "step": 746, "train_loss": 4.247539520263672, "perplexity": 69.9331314339644, "lr": 0.001305, "grad_norm": 0.284257, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:25.146155+00:00", "epoch": 0, "step": 747, "train_loss": 4.194400310516357, "perplexity": 66.31395187119537, "lr": 0.001305, "grad_norm": 0.302025, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:25.454036+00:00", "epoch": 0, "step": 748, "train_loss": 4.182227611541748, "perplexity": 65.51162524451466, "lr": 0.001305, "grad_norm": 0.32071, "tokens_per_sec": 106431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:25.760820+00:00", "epoch": 0, "step": 749, "train_loss": 4.23219108581543, "perplexity": 68.86796259788227, "lr": 0.001305, "grad_norm": 0.308759, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:26.067824+00:00", "epoch": 0, "step": 750, "train_loss": 4.222700119018555, "perplexity": 68.21743102152547, "lr": 0.001305, "grad_norm": 0.30846, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:32:26.979011+00:00", "step": 750, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0000_step_0000750.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:27.975833+00:00", "epoch": 0, "step": 751, "train_loss": 4.201521396636963, "perplexity": 66.78786461753866, "lr": 0.001305, "grad_norm": 0.323495, "tokens_per_sec": 17172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:28.282242+00:00", "epoch": 0, "step": 752, "train_loss": 4.299644470214844, "perplexity": 73.67359588510335, "lr": 0.001305, "grad_norm": 0.325473, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:28.588360+00:00", "epoch": 0, "step": 753, "train_loss": 4.21038293838501, "perplexity": 67.38233815396411, "lr": 0.001305, "grad_norm": 0.374564, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:28.894350+00:00", "epoch": 0, "step": 754, "train_loss": 4.2287139892578125, "perplexity": 68.6289178738381, "lr": 0.001305, "grad_norm": 0.381567, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:29.199561+00:00", "epoch": 0, "step": 755, "train_loss": 4.2966203689575195, "perplexity": 73.4511360112841, "lr": 0.001305, "grad_norm": 0.355467, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:29.506497+00:00", "epoch": 0, "step": 756, "train_loss": 4.216547966003418, "perplexity": 67.79903528453268, "lr": 0.001305, "grad_norm": 0.346479, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:29.813059+00:00", "epoch": 0, "step": 757, "train_loss": 4.2600507736206055, "perplexity": 70.81357882478578, "lr": 0.001305, "grad_norm": 0.377289, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:30.120103+00:00", "epoch": 0, "step": 758, "train_loss": 4.150954723358154, "perplexity": 63.49459113905231, "lr": 0.001305, "grad_norm": 0.410589, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:30.427846+00:00", "epoch": 0, "step": 759, "train_loss": 4.195404529571533, "perplexity": 66.38057905383242, "lr": 0.001305, "grad_norm": 0.425988, "tokens_per_sec": 106479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:30.734161+00:00", "epoch": 0, "step": 760, "train_loss": 4.214702129364014, "perplexity": 67.67400476954045, "lr": 0.001305, "grad_norm": 0.424327, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:31.040743+00:00", "epoch": 0, "step": 761, "train_loss": 4.314418792724609, "perplexity": 74.77015385534324, "lr": 0.001305, "grad_norm": 0.425713, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:31.348159+00:00", "epoch": 0, "step": 762, "train_loss": 4.233648777008057, "perplexity": 68.96842402349411, "lr": 0.001305, "grad_norm": 0.409582, "tokens_per_sec": 106592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:31.654859+00:00", "epoch": 0, "step": 763, "train_loss": 4.289178371429443, "perplexity": 72.90654178654194, "lr": 0.001305, "grad_norm": 0.369169, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:31.962894+00:00", "epoch": 0, "step": 764, "train_loss": 4.2137675285339355, "perplexity": 67.61078613520387, "lr": 0.001305, "grad_norm": 0.313525, "tokens_per_sec": 106378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:32.269111+00:00", "epoch": 0, "step": 765, "train_loss": 4.181725025177002, "perplexity": 65.47870826743922, "lr": 0.001305, "grad_norm": 0.325129, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:32.575637+00:00", "epoch": 0, "step": 766, "train_loss": 4.312158584594727, "perplexity": 74.60134858511125, "lr": 0.001305, "grad_norm": 0.3129, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:32.880990+00:00", "epoch": 0, "step": 767, "train_loss": 4.286411762237549, "perplexity": 72.70511663869658, "lr": 0.001305, "grad_norm": 0.291915, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:33.187511+00:00", "epoch": 0, "step": 768, "train_loss": 4.214005470275879, "perplexity": 67.6268754775182, "lr": 0.001305, "grad_norm": 0.320254, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:33.495431+00:00", "epoch": 0, "step": 769, "train_loss": 4.2726149559021, "perplexity": 71.70890628796681, "lr": 0.001305, "grad_norm": 0.326558, "tokens_per_sec": 106418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:33.802120+00:00", "epoch": 0, "step": 770, "train_loss": 4.127763748168945, "perplexity": 62.03903279529152, "lr": 0.001305, "grad_norm": 0.280013, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:34.109096+00:00", "epoch": 0, "step": 771, "train_loss": 4.1458892822265625, "perplexity": 63.17377624591458, "lr": 0.001305, "grad_norm": 0.286483, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:34.415216+00:00", "epoch": 0, "step": 772, "train_loss": 4.300243854522705, "perplexity": 73.71776791907016, "lr": 0.001305, "grad_norm": 0.281442, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:34.720960+00:00", "epoch": 0, "step": 773, "train_loss": 4.207580089569092, "perplexity": 67.19374007667255, "lr": 0.001305, "grad_norm": 0.297253, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:35.027517+00:00", "epoch": 0, "step": 774, "train_loss": 4.311919689178467, "perplexity": 74.5835287935055, "lr": 0.001305, "grad_norm": 0.318183, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:35.334243+00:00", "epoch": 0, "step": 775, "train_loss": 4.286248207092285, "perplexity": 72.69322631517163, "lr": 0.001305, "grad_norm": 0.332854, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:35.640651+00:00", "epoch": 0, "step": 776, "train_loss": 4.142999172210693, "perplexity": 62.99146066539145, "lr": 0.001305, "grad_norm": 0.32017, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:35.946786+00:00", "epoch": 0, "step": 777, "train_loss": 4.26713752746582, "perplexity": 71.31719963744243, "lr": 0.001305, "grad_norm": 0.292896, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:36.252975+00:00", "epoch": 0, "step": 778, "train_loss": 4.243360996246338, "perplexity": 69.6415238336315, "lr": 0.001305, "grad_norm": 0.285911, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:36.558992+00:00", "epoch": 0, "step": 779, "train_loss": 4.341030120849609, "perplexity": 76.78659808693178, "lr": 0.001305, "grad_norm": 0.280911, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:36.866728+00:00", "epoch": 0, "step": 780, "train_loss": 4.2098894119262695, "perplexity": 67.34909139198653, "lr": 0.001305, "grad_norm": 0.3031, "tokens_per_sec": 106539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:37.174348+00:00", "epoch": 0, "step": 781, "train_loss": 4.132169723510742, "perplexity": 62.3129782994181, "lr": 0.001305, "grad_norm": 0.326218, "tokens_per_sec": 106463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:37.481605+00:00", "epoch": 0, "step": 782, "train_loss": 4.180303573608398, "perplexity": 65.38569957419931, "lr": 0.001305, "grad_norm": 0.323785, "tokens_per_sec": 106647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:37.788235+00:00", "epoch": 0, "step": 783, "train_loss": 4.20635986328125, "perplexity": 67.11179851244522, "lr": 0.001305, "grad_norm": 0.304241, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:38.094870+00:00", "epoch": 0, "step": 784, "train_loss": 4.157318592071533, "perplexity": 63.89995084098493, "lr": 0.001305, "grad_norm": 0.309572, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:38.401820+00:00", "epoch": 0, "step": 785, "train_loss": 4.031894207000732, "perplexity": 56.36758205548096, "lr": 0.001305, "grad_norm": 0.270025, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:38.709261+00:00", "epoch": 0, "step": 786, "train_loss": 4.191540718078613, "perplexity": 66.12459187115691, "lr": 0.001305, "grad_norm": 0.266705, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:39.015854+00:00", "epoch": 0, "step": 787, "train_loss": 4.255152225494385, "perplexity": 70.46754332863712, "lr": 0.001305, "grad_norm": 0.263614, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:39.322003+00:00", "epoch": 0, "step": 788, "train_loss": 4.118139266967773, "perplexity": 61.44480345196623, "lr": 0.001305, "grad_norm": 0.269693, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:39.627657+00:00", "epoch": 0, "step": 789, "train_loss": 4.1089935302734375, "perplexity": 60.885407402044336, "lr": 0.001305, "grad_norm": 0.281062, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:39.934747+00:00", "epoch": 0, "step": 790, "train_loss": 4.214434623718262, "perplexity": 67.65590401232926, "lr": 0.001305, "grad_norm": 0.291043, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:40.241292+00:00", "epoch": 0, "step": 791, "train_loss": 4.152914047241211, "perplexity": 63.61911956383617, "lr": 0.001305, "grad_norm": 0.288806, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:40.549173+00:00", "epoch": 0, "step": 792, "train_loss": 4.096189498901367, "perplexity": 60.11079837732328, "lr": 0.001305, "grad_norm": 0.277635, "tokens_per_sec": 106431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:40.856882+00:00", "epoch": 0, "step": 793, "train_loss": 4.172957897186279, "perplexity": 64.90715714322192, "lr": 0.001305, "grad_norm": 0.305392, "tokens_per_sec": 106491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:41.163964+00:00", "epoch": 0, "step": 794, "train_loss": 4.153744220733643, "perplexity": 63.6719563993455, "lr": 0.001305, "grad_norm": 0.335026, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:41.469710+00:00", "epoch": 0, "step": 795, "train_loss": 4.159989356994629, "perplexity": 64.07084069000625, "lr": 0.001305, "grad_norm": 0.336553, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:41.776172+00:00", "epoch": 0, "step": 796, "train_loss": 4.191742420196533, "perplexity": 66.13793068657236, "lr": 0.001305, "grad_norm": 0.353379, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:42.082678+00:00", "epoch": 0, "step": 797, "train_loss": 4.050264358520508, "perplexity": 57.412632558071294, "lr": 0.001305, "grad_norm": 0.312948, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:42.390365+00:00", "epoch": 0, "step": 798, "train_loss": 4.287223815917969, "perplexity": 72.76418107474899, "lr": 0.001305, "grad_norm": 0.306547, "tokens_per_sec": 106498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:42.696928+00:00", "epoch": 0, "step": 799, "train_loss": 4.159154891967773, "perplexity": 64.01739811529013, "lr": 0.001305, "grad_norm": 0.353943, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:43.003567+00:00", "epoch": 0, "step": 800, "train_loss": 4.262724876403809, "perplexity": 71.00319502662356, "lr": 0.001305, "grad_norm": 0.362401, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:32:48.599299+00:00", "step": 800, "epoch": 0, "val_loss": 4.196767997741699, "val_ppl": 66.47114859079886, "eval_train_loss": 4.262724876403809, "eval_train_ppl": 71.00319502662356} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:32:49.576906+00:00", "step": 800, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_4p1968_epoch_0000_step_0000800.pt", "val_loss": 4.196767997741699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:50.576426+00:00", "epoch": 0, "step": 801, "train_loss": 4.1869072914123535, "perplexity": 65.81891713149881, "lr": 0.001305, "grad_norm": 0.333209, "tokens_per_sec": 4327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:50.881871+00:00", "epoch": 0, "step": 802, "train_loss": 4.239950656890869, "perplexity": 69.4044271240411, "lr": 0.001305, "grad_norm": 0.306601, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:51.186316+00:00", "epoch": 0, "step": 803, "train_loss": 4.182608127593994, "perplexity": 65.53655821292544, "lr": 0.001305, "grad_norm": 0.357298, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:51.491149+00:00", "epoch": 0, "step": 804, "train_loss": 4.234219074249268, "perplexity": 69.00776774318643, "lr": 0.001305, "grad_norm": 0.344116, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:51.798121+00:00", "epoch": 0, "step": 805, "train_loss": 4.151264667510986, "perplexity": 63.514273966443376, "lr": 0.001305, "grad_norm": 0.319891, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:52.105173+00:00", "epoch": 0, "step": 806, "train_loss": 4.1870574951171875, "perplexity": 65.82880411921262, "lr": 0.001305, "grad_norm": 0.354268, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:52.504645+00:00", "epoch": 0, "step": 807, "train_loss": 4.1486029624938965, "perplexity": 63.345442494234696, "lr": 0.001305, "grad_norm": 0.368237, "tokens_per_sec": 82028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:52.854616+00:00", "epoch": 0, "step": 808, "train_loss": 4.19832706451416, "perplexity": 66.57486237724522, "lr": 0.001305, "grad_norm": 0.326535, "tokens_per_sec": 93632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:53.161114+00:00", "epoch": 0, "step": 809, "train_loss": 4.182496547698975, "perplexity": 65.52924605859239, "lr": 0.001305, "grad_norm": 0.301731, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:53.467336+00:00", "epoch": 0, "step": 810, "train_loss": 4.069387912750244, "perplexity": 58.52113158867422, "lr": 0.001305, "grad_norm": 0.292055, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:53.772976+00:00", "epoch": 0, "step": 811, "train_loss": 4.067751407623291, "perplexity": 58.425439778219875, "lr": 0.001305, "grad_norm": 0.300751, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:54.078089+00:00", "epoch": 0, "step": 812, "train_loss": 4.2109503746032715, "perplexity": 67.42058418316871, "lr": 0.001305, "grad_norm": 0.301459, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:54.383407+00:00", "epoch": 0, "step": 813, "train_loss": 4.155066967010498, "perplexity": 63.75623396924566, "lr": 0.001305, "grad_norm": 0.290999, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:54.689825+00:00", "epoch": 0, "step": 814, "train_loss": 4.155735969543457, "perplexity": 63.79890132195517, "lr": 0.001305, "grad_norm": 0.286564, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:54.996900+00:00", "epoch": 0, "step": 815, "train_loss": 4.168683052062988, "perplexity": 64.63028132131261, "lr": 0.001305, "grad_norm": 0.268784, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:55.303640+00:00", "epoch": 0, "step": 816, "train_loss": 4.126933574676514, "perplexity": 61.987551007133426, "lr": 0.001305, "grad_norm": 0.253778, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:55.608844+00:00", "epoch": 0, "step": 817, "train_loss": 4.194488048553467, "perplexity": 66.31977038241416, "lr": 0.001305, "grad_norm": 0.286925, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:55.914590+00:00", "epoch": 0, "step": 818, "train_loss": 4.092554092407227, "perplexity": 59.89266792719941, "lr": 0.001305, "grad_norm": 0.290273, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:56.221046+00:00", "epoch": 0, "step": 819, "train_loss": 4.1807661056518555, "perplexity": 65.41594955068865, "lr": 0.001305, "grad_norm": 0.308704, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:56.526885+00:00", "epoch": 0, "step": 820, "train_loss": 4.119129657745361, "perplexity": 61.5056879633884, "lr": 0.001305, "grad_norm": 0.292754, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:56.833806+00:00", "epoch": 0, "step": 821, "train_loss": 4.121086120605469, "perplexity": 61.62613934848336, "lr": 0.001305, "grad_norm": 0.328436, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:57.139817+00:00", "epoch": 0, "step": 822, "train_loss": 4.314697742462158, "perplexity": 74.79101387945154, "lr": 0.001305, "grad_norm": 0.317182, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:57.445173+00:00", "epoch": 0, "step": 823, "train_loss": 4.1835174560546875, "perplexity": 65.59617957411633, "lr": 0.001305, "grad_norm": 0.308898, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:57.750926+00:00", "epoch": 0, "step": 824, "train_loss": 4.237821102142334, "perplexity": 69.25678385969397, "lr": 0.001305, "grad_norm": 0.308818, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:58.056820+00:00", "epoch": 0, "step": 825, "train_loss": 4.166666507720947, "perplexity": 64.50008281284288, "lr": 0.001305, "grad_norm": 0.291101, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:58.363455+00:00", "epoch": 0, "step": 826, "train_loss": 4.17518424987793, "perplexity": 65.05182434758969, "lr": 0.001305, "grad_norm": 0.305478, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:58.670010+00:00", "epoch": 0, "step": 827, "train_loss": 4.193276405334473, "perplexity": 66.23946314403025, "lr": 0.001305, "grad_norm": 0.312856, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:58.975989+00:00", "epoch": 0, "step": 828, "train_loss": 4.19303560256958, "perplexity": 66.22351441848583, "lr": 0.001305, "grad_norm": 0.297343, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:59.281931+00:00", "epoch": 0, "step": 829, "train_loss": 4.242959976196289, "perplexity": 69.61360178528307, "lr": 0.001305, "grad_norm": 0.314321, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:59.587910+00:00", "epoch": 0, "step": 830, "train_loss": 4.1892805099487305, "perplexity": 65.97530530393404, "lr": 0.001305, "grad_norm": 0.317796, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:32:59.894449+00:00", "epoch": 0, "step": 831, "train_loss": 4.106512069702148, "perplexity": 60.73450996466096, "lr": 0.001305, "grad_norm": 0.308128, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:00.201081+00:00", "epoch": 0, "step": 832, "train_loss": 4.189323902130127, "perplexity": 65.97816817846221, "lr": 0.001305, "grad_norm": 0.314783, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:00.506957+00:00", "epoch": 0, "step": 833, "train_loss": 4.167547225952148, "perplexity": 64.5569142342511, "lr": 0.001305, "grad_norm": 0.315696, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:00.812293+00:00", "epoch": 0, "step": 834, "train_loss": 4.152634620666504, "perplexity": 63.601345174604575, "lr": 0.001305, "grad_norm": 0.326456, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:01.118051+00:00", "epoch": 0, "step": 835, "train_loss": 4.063875198364258, "perplexity": 58.199408902085295, "lr": 0.001305, "grad_norm": 0.319941, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:01.424740+00:00", "epoch": 0, "step": 836, "train_loss": 4.1249003410339355, "perplexity": 61.86164387568786, "lr": 0.001305, "grad_norm": 0.316537, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:01.731511+00:00", "epoch": 0, "step": 837, "train_loss": 4.254265785217285, "perplexity": 70.40510573768856, "lr": 0.001305, "grad_norm": 0.298814, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:02.039215+00:00", "epoch": 0, "step": 838, "train_loss": 4.130617141723633, "perplexity": 62.216307368392165, "lr": 0.001305, "grad_norm": 0.317493, "tokens_per_sec": 106492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:02.344604+00:00", "epoch": 0, "step": 839, "train_loss": 4.093147277832031, "perplexity": 59.92820592413534, "lr": 0.001305, "grad_norm": 0.295218, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:02.651027+00:00", "epoch": 0, "step": 840, "train_loss": 4.150961399078369, "perplexity": 63.49501501259274, "lr": 0.001305, "grad_norm": 0.322876, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:02.957336+00:00", "epoch": 0, "step": 841, "train_loss": 4.107545375823975, "perplexity": 60.79729974055663, "lr": 0.001305, "grad_norm": 0.336815, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:03.263877+00:00", "epoch": 0, "step": 842, "train_loss": 4.071581840515137, "perplexity": 58.64966366759237, "lr": 0.001305, "grad_norm": 0.354161, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:03.571539+00:00", "epoch": 0, "step": 843, "train_loss": 4.0661821365356445, "perplexity": 58.333826326763884, "lr": 0.001305, "grad_norm": 0.307045, "tokens_per_sec": 106507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:03.877429+00:00", "epoch": 0, "step": 844, "train_loss": 4.176117420196533, "perplexity": 65.11255711184114, "lr": 0.001305, "grad_norm": 0.319408, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:04.183489+00:00", "epoch": 0, "step": 845, "train_loss": 4.154839038848877, "perplexity": 63.74170378402728, "lr": 0.001305, "grad_norm": 0.348909, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:04.490468+00:00", "epoch": 0, "step": 846, "train_loss": 4.143327713012695, "perplexity": 63.01215933038955, "lr": 0.001305, "grad_norm": 0.340042, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:04.796564+00:00", "epoch": 0, "step": 847, "train_loss": 4.212814807891846, "perplexity": 67.5464026182477, "lr": 0.001305, "grad_norm": 0.319942, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:05.103686+00:00", "epoch": 0, "step": 848, "train_loss": 4.082345962524414, "perplexity": 59.284385787720275, "lr": 0.001305, "grad_norm": 0.341023, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:05.410867+00:00", "epoch": 0, "step": 849, "train_loss": 4.096609592437744, "perplexity": 60.13605584006502, "lr": 0.001305, "grad_norm": 0.346434, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:05.716925+00:00", "epoch": 0, "step": 850, "train_loss": 4.0849480628967285, "perplexity": 59.43885058935405, "lr": 0.001305, "grad_norm": 0.331878, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:06.022839+00:00", "epoch": 0, "step": 851, "train_loss": 4.093515872955322, "perplexity": 59.950299240080675, "lr": 0.001305, "grad_norm": 0.298751, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:06.329960+00:00", "epoch": 0, "step": 852, "train_loss": 4.138942718505859, "perplexity": 62.73645627781955, "lr": 0.001305, "grad_norm": 0.266104, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:06.636904+00:00", "epoch": 0, "step": 853, "train_loss": 4.08244514465332, "perplexity": 59.2902660309163, "lr": 0.001305, "grad_norm": 0.260961, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:06.943880+00:00", "epoch": 0, "step": 854, "train_loss": 4.1014814376831055, "perplexity": 60.42974421906276, "lr": 0.001305, "grad_norm": 0.288397, "tokens_per_sec": 106733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:07.250236+00:00", "epoch": 0, "step": 855, "train_loss": 4.191503047943115, "perplexity": 66.12210099573748, "lr": 0.001305, "grad_norm": 0.288999, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:07.556351+00:00", "epoch": 0, "step": 856, "train_loss": 4.167111873626709, "perplexity": 64.52881534841747, "lr": 0.001305, "grad_norm": 0.27871, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:07.862254+00:00", "epoch": 0, "step": 857, "train_loss": 4.135094165802002, "perplexity": 62.49547572998794, "lr": 0.001305, "grad_norm": 0.331724, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:08.168898+00:00", "epoch": 0, "step": 858, "train_loss": 4.1712188720703125, "perplexity": 64.79438005625902, "lr": 0.001305, "grad_norm": 0.348966, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:08.477029+00:00", "epoch": 0, "step": 859, "train_loss": 4.154183864593506, "perplexity": 63.69995553839574, "lr": 0.001305, "grad_norm": 0.319197, "tokens_per_sec": 106344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:08.784087+00:00", "epoch": 0, "step": 860, "train_loss": 4.047682762145996, "perplexity": 57.264607466738916, "lr": 0.001305, "grad_norm": 0.282718, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:09.089756+00:00", "epoch": 0, "step": 861, "train_loss": 4.086157321929932, "perplexity": 59.51077103281125, "lr": 0.001305, "grad_norm": 0.373528, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:09.396033+00:00", "epoch": 0, "step": 862, "train_loss": 4.1862969398498535, "perplexity": 65.77875670982034, "lr": 0.001305, "grad_norm": 0.419796, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:09.702882+00:00", "epoch": 0, "step": 863, "train_loss": 4.143005847930908, "perplexity": 62.991881180162395, "lr": 0.001305, "grad_norm": 0.402271, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:10.009807+00:00", "epoch": 0, "step": 864, "train_loss": 4.1025471687316895, "perplexity": 60.4941804034574, "lr": 0.001305, "grad_norm": 0.393838, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:10.316549+00:00", "epoch": 0, "step": 865, "train_loss": 4.156179904937744, "perplexity": 63.827230199999185, "lr": 0.001305, "grad_norm": 0.352528, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:10.622792+00:00", "epoch": 0, "step": 866, "train_loss": 4.174091339111328, "perplexity": 64.98076734492763, "lr": 0.001305, "grad_norm": 0.358603, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:10.930312+00:00", "epoch": 0, "step": 867, "train_loss": 4.137269973754883, "perplexity": 62.63160192162213, "lr": 0.001305, "grad_norm": 0.295945, "tokens_per_sec": 106556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:11.238243+00:00", "epoch": 0, "step": 868, "train_loss": 4.1820173263549805, "perplexity": 65.49785056852059, "lr": 0.001305, "grad_norm": 0.30056, "tokens_per_sec": 106413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:11.546486+00:00", "epoch": 0, "step": 869, "train_loss": 4.063590049743652, "perplexity": 58.18281578678121, "lr": 0.001305, "grad_norm": 0.287623, "tokens_per_sec": 106307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:11.854704+00:00", "epoch": 0, "step": 870, "train_loss": 4.094218730926514, "perplexity": 59.992450597250304, "lr": 0.001305, "grad_norm": 0.339019, "tokens_per_sec": 106380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:12.160897+00:00", "epoch": 0, "step": 871, "train_loss": 4.175603866577148, "perplexity": 65.07912690730497, "lr": 0.001305, "grad_norm": 0.363491, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:12.467377+00:00", "epoch": 0, "step": 872, "train_loss": 4.012044906616211, "perplexity": 55.25975614484122, "lr": 0.001305, "grad_norm": 0.318527, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:12.773675+00:00", "epoch": 0, "step": 873, "train_loss": 4.091139316558838, "perplexity": 59.807993139166676, "lr": 0.001305, "grad_norm": 0.305222, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:13.081450+00:00", "epoch": 0, "step": 874, "train_loss": 4.1590576171875, "perplexity": 64.01117113982355, "lr": 0.001305, "grad_norm": 0.295923, "tokens_per_sec": 106467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:13.388886+00:00", "epoch": 0, "step": 875, "train_loss": 4.11774206161499, "perplexity": 61.420402093630294, "lr": 0.001305, "grad_norm": 0.267716, "tokens_per_sec": 106585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:13.695107+00:00", "epoch": 0, "step": 876, "train_loss": 4.063603401184082, "perplexity": 58.183592616366106, "lr": 0.001305, "grad_norm": 0.288101, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:14.000773+00:00", "epoch": 0, "step": 877, "train_loss": 4.233785629272461, "perplexity": 68.97786315436261, "lr": 0.001305, "grad_norm": 0.307955, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:14.307827+00:00", "epoch": 0, "step": 878, "train_loss": 4.004549026489258, "perplexity": 54.847084238777775, "lr": 0.001305, "grad_norm": 0.29728, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:14.615277+00:00", "epoch": 0, "step": 879, "train_loss": 4.120721817016602, "perplexity": 61.60369281367553, "lr": 0.001305, "grad_norm": 0.304786, "tokens_per_sec": 106580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:14.923129+00:00", "epoch": 0, "step": 880, "train_loss": 4.163177013397217, "perplexity": 64.27540237852381, "lr": 0.001305, "grad_norm": 0.309382, "tokens_per_sec": 106515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:15.228999+00:00", "epoch": 0, "step": 881, "train_loss": 4.016380786895752, "perplexity": 55.499876021095126, "lr": 0.001305, "grad_norm": 0.313934, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:15.535391+00:00", "epoch": 0, "step": 882, "train_loss": 4.173000812530518, "perplexity": 64.90994271598574, "lr": 0.001305, "grad_norm": 0.305603, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:15.841935+00:00", "epoch": 0, "step": 883, "train_loss": 4.08030891418457, "perplexity": 59.16374354682875, "lr": 0.001305, "grad_norm": 0.294058, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:16.149201+00:00", "epoch": 0, "step": 884, "train_loss": 4.063138961791992, "perplexity": 58.15657613822573, "lr": 0.001305, "grad_norm": 0.293282, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:16.456092+00:00", "epoch": 0, "step": 885, "train_loss": 4.006515979766846, "perplexity": 54.955072059570185, "lr": 0.001305, "grad_norm": 0.28327, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:16.762987+00:00", "epoch": 0, "step": 886, "train_loss": 4.065899848937988, "perplexity": 58.31736173505236, "lr": 0.001305, "grad_norm": 0.252568, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:17.069700+00:00", "epoch": 0, "step": 887, "train_loss": 4.106112480163574, "perplexity": 60.710245937997115, "lr": 0.001305, "grad_norm": 0.264104, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:17.376285+00:00", "epoch": 0, "step": 888, "train_loss": 4.090715408325195, "perplexity": 59.78264541137218, "lr": 0.001305, "grad_norm": 0.284685, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:17.683354+00:00", "epoch": 0, "step": 889, "train_loss": 4.091914176940918, "perplexity": 59.854353942837335, "lr": 0.001305, "grad_norm": 0.328501, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:17.992105+00:00", "epoch": 0, "step": 890, "train_loss": 4.142885208129883, "perplexity": 62.984282310522886, "lr": 0.001305, "grad_norm": 0.337846, "tokens_per_sec": 106184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:18.298707+00:00", "epoch": 0, "step": 891, "train_loss": 4.067446708679199, "perplexity": 58.407640320287086, "lr": 0.001305, "grad_norm": 0.336973, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:18.604806+00:00", "epoch": 0, "step": 892, "train_loss": 4.051459312438965, "perplexity": 57.4812790148233, "lr": 0.001305, "grad_norm": 0.349438, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:18.911004+00:00", "epoch": 0, "step": 893, "train_loss": 4.054959774017334, "perplexity": 57.682842600466714, "lr": 0.001305, "grad_norm": 0.321662, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:19.218087+00:00", "epoch": 0, "step": 894, "train_loss": 4.182229518890381, "perplexity": 65.51175019814266, "lr": 0.001305, "grad_norm": 0.297823, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:19.525047+00:00", "epoch": 0, "step": 895, "train_loss": 4.127501964569092, "perplexity": 62.02279411955717, "lr": 0.001305, "grad_norm": 0.291188, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:19.832178+00:00", "epoch": 0, "step": 896, "train_loss": 4.073283672332764, "perplexity": 58.74956051101795, "lr": 0.001305, "grad_norm": 0.261988, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:20.138333+00:00", "epoch": 0, "step": 897, "train_loss": 3.9979794025421143, "perplexity": 54.487940531998, "lr": 0.001305, "grad_norm": 0.277813, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:20.444353+00:00", "epoch": 0, "step": 898, "train_loss": 4.052916526794434, "perplexity": 57.565102619421275, "lr": 0.001305, "grad_norm": 0.2732, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:20.751292+00:00", "epoch": 0, "step": 899, "train_loss": 4.2001118659973145, "perplexity": 66.69379139112625, "lr": 0.001305, "grad_norm": 0.275187, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:21.059279+00:00", "epoch": 0, "step": 900, "train_loss": 4.073366641998291, "perplexity": 58.75443514462398, "lr": 0.001305, "grad_norm": 0.276763, "tokens_per_sec": 106453} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:33:26.650603+00:00", "step": 900, "epoch": 0, "val_loss": 4.113304162025452, "val_ppl": 61.14842845860683, "eval_train_loss": 4.073366641998291, "eval_train_ppl": 58.75443514462398} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:33:27.571361+00:00", "step": 900, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_4p1133_epoch_0000_step_0000900.pt", "val_loss": 4.113304162025452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:28.537507+00:00", "epoch": 0, "step": 901, "train_loss": 4.054601669311523, "perplexity": 57.66218980123996, "lr": 0.001305, "grad_norm": 0.28167, "tokens_per_sec": 4382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:28.843150+00:00", "epoch": 0, "step": 902, "train_loss": 3.9931728839874268, "perplexity": 54.22667163472499, "lr": 0.001305, "grad_norm": 0.276377, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:29.148036+00:00", "epoch": 0, "step": 903, "train_loss": 4.064308166503906, "perplexity": 58.22461284774527, "lr": 0.001305, "grad_norm": 0.26841, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:29.453981+00:00", "epoch": 0, "step": 904, "train_loss": 4.09212064743042, "perplexity": 59.866713376479986, "lr": 0.001305, "grad_norm": 0.268299, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:29.760534+00:00", "epoch": 0, "step": 905, "train_loss": 4.032789707183838, "perplexity": 56.418081843442806, "lr": 0.001305, "grad_norm": 0.298151, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:30.068131+00:00", "epoch": 0, "step": 906, "train_loss": 4.055247783660889, "perplexity": 57.69945820801605, "lr": 0.001305, "grad_norm": 0.326225, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:30.374416+00:00", "epoch": 0, "step": 907, "train_loss": 4.244104862213135, "perplexity": 69.69334706546678, "lr": 0.001305, "grad_norm": 0.37494, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:30.680169+00:00", "epoch": 0, "step": 908, "train_loss": 4.136751651763916, "perplexity": 62.599146996793905, "lr": 0.001305, "grad_norm": 0.338515, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:30.986087+00:00", "epoch": 0, "step": 909, "train_loss": 4.142854690551758, "perplexity": 62.98236021209587, "lr": 0.001305, "grad_norm": 0.297083, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:31.292686+00:00", "epoch": 0, "step": 910, "train_loss": 4.119897842407227, "perplexity": 61.55295384163636, "lr": 0.001305, "grad_norm": 0.335003, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:31.600424+00:00", "epoch": 0, "step": 911, "train_loss": 4.034756183624268, "perplexity": 56.5291358289378, "lr": 0.001305, "grad_norm": 0.341136, "tokens_per_sec": 106425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:31.907414+00:00", "epoch": 0, "step": 912, "train_loss": 4.091876029968262, "perplexity": 59.852070723983346, "lr": 0.001305, "grad_norm": 0.34578, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:32.212993+00:00", "epoch": 0, "step": 913, "train_loss": 4.155308246612549, "perplexity": 63.771618903966846, "lr": 0.001305, "grad_norm": 0.338092, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:32.519089+00:00", "epoch": 0, "step": 914, "train_loss": 4.042648792266846, "perplexity": 56.977063508639155, "lr": 0.001305, "grad_norm": 0.302554, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:32.828177+00:00", "epoch": 0, "step": 915, "train_loss": 4.1297383308410645, "perplexity": 62.16165501846091, "lr": 0.001305, "grad_norm": 0.322103, "tokens_per_sec": 106015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:33.134749+00:00", "epoch": 0, "step": 916, "train_loss": 4.1645708084106445, "perplexity": 64.36505157563381, "lr": 0.001305, "grad_norm": 0.384248, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:33.441117+00:00", "epoch": 0, "step": 917, "train_loss": 4.1402764320373535, "perplexity": 62.82018456083602, "lr": 0.001305, "grad_norm": 0.363586, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:33.747548+00:00", "epoch": 0, "step": 918, "train_loss": 4.042497634887695, "perplexity": 56.96845165593644, "lr": 0.001305, "grad_norm": 0.329466, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:34.053988+00:00", "epoch": 0, "step": 919, "train_loss": 4.080169200897217, "perplexity": 59.155478163131086, "lr": 0.001305, "grad_norm": 0.41555, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:34.360300+00:00", "epoch": 0, "step": 920, "train_loss": 4.178372383117676, "perplexity": 65.25954918237737, "lr": 0.001305, "grad_norm": 0.439328, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:34.666520+00:00", "epoch": 0, "step": 921, "train_loss": 4.101901054382324, "perplexity": 60.45510686979034, "lr": 0.001305, "grad_norm": 0.366999, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:34.974477+00:00", "epoch": 0, "step": 922, "train_loss": 4.223120212554932, "perplexity": 68.24609474366652, "lr": 0.001305, "grad_norm": 0.374705, "tokens_per_sec": 106405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:35.281911+00:00", "epoch": 0, "step": 923, "train_loss": 4.013494968414307, "perplexity": 55.339944331060856, "lr": 0.001305, "grad_norm": 0.338628, "tokens_per_sec": 106585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:35.588791+00:00", "epoch": 0, "step": 924, "train_loss": 4.036521911621094, "perplexity": 56.629039081736074, "lr": 0.001305, "grad_norm": 0.291063, "tokens_per_sec": 106779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:35.894543+00:00", "epoch": 0, "step": 925, "train_loss": 4.1129326820373535, "perplexity": 61.12571725976989, "lr": 0.001305, "grad_norm": 0.298113, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:36.201107+00:00", "epoch": 0, "step": 926, "train_loss": 4.124670505523682, "perplexity": 61.847427506978306, "lr": 0.001305, "grad_norm": 0.290734, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:36.507484+00:00", "epoch": 0, "step": 927, "train_loss": 4.1506547927856445, "perplexity": 63.47555002562947, "lr": 0.001305, "grad_norm": 0.274076, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:36.814508+00:00", "epoch": 0, "step": 928, "train_loss": 4.195714473724365, "perplexity": 66.4011565149389, "lr": 0.001305, "grad_norm": 0.30145, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:37.121963+00:00", "epoch": 0, "step": 929, "train_loss": 4.060977935791016, "perplexity": 58.03103396387193, "lr": 0.001305, "grad_norm": 0.316595, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:37.428377+00:00", "epoch": 0, "step": 930, "train_loss": 4.11881685256958, "perplexity": 61.48645167461566, "lr": 0.001305, "grad_norm": 0.316228, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:37.734740+00:00", "epoch": 0, "step": 931, "train_loss": 4.049557209014893, "perplexity": 57.37204759485137, "lr": 0.001305, "grad_norm": 0.291053, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:38.040520+00:00", "epoch": 0, "step": 932, "train_loss": 4.043027877807617, "perplexity": 56.99866678405686, "lr": 0.001305, "grad_norm": 0.3054, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:38.346159+00:00", "epoch": 0, "step": 933, "train_loss": 4.057250499725342, "perplexity": 57.81512962972477, "lr": 0.001305, "grad_norm": 0.381264, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:38.654292+00:00", "epoch": 0, "step": 934, "train_loss": 4.053637504577637, "perplexity": 57.606620744522125, "lr": 0.001305, "grad_norm": 0.388119, "tokens_per_sec": 106344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:38.960502+00:00", "epoch": 0, "step": 935, "train_loss": 4.025135517120361, "perplexity": 55.98789558594648, "lr": 0.001305, "grad_norm": 0.354882, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:39.266272+00:00", "epoch": 0, "step": 936, "train_loss": 4.070948123931885, "perplexity": 58.61250817739268, "lr": 0.001305, "grad_norm": 0.262056, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:39.572128+00:00", "epoch": 0, "step": 937, "train_loss": 4.135715961456299, "perplexity": 62.534347229022245, "lr": 0.001305, "grad_norm": 0.298102, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:39.878640+00:00", "epoch": 0, "step": 938, "train_loss": 3.987325429916382, "perplexity": 53.91050893721298, "lr": 0.001305, "grad_norm": 0.283432, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:40.185289+00:00", "epoch": 0, "step": 939, "train_loss": 4.1944661140441895, "perplexity": 66.31831570674929, "lr": 0.001305, "grad_norm": 0.268813, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:40.491999+00:00", "epoch": 0, "step": 940, "train_loss": 4.143035888671875, "perplexity": 62.993773531371616, "lr": 0.001305, "grad_norm": 0.276933, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:40.798159+00:00", "epoch": 0, "step": 941, "train_loss": 4.047226905822754, "perplexity": 57.23850898235874, "lr": 0.001305, "grad_norm": 0.276106, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:41.104229+00:00", "epoch": 0, "step": 942, "train_loss": 4.066096305847168, "perplexity": 58.32881970915258, "lr": 0.001305, "grad_norm": 0.26036, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:41.410316+00:00", "epoch": 0, "step": 943, "train_loss": 4.150547027587891, "perplexity": 63.46870993899668, "lr": 0.001305, "grad_norm": 0.309271, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:41.716800+00:00", "epoch": 0, "step": 944, "train_loss": 4.148221969604492, "perplexity": 63.32131292795669, "lr": 0.001305, "grad_norm": 0.308274, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:42.024481+00:00", "epoch": 0, "step": 945, "train_loss": 4.052793979644775, "perplexity": 57.55804861240854, "lr": 0.001305, "grad_norm": 0.284373, "tokens_per_sec": 106499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:42.332545+00:00", "epoch": 0, "step": 946, "train_loss": 3.9800000190734863, "perplexity": 53.517035248247595, "lr": 0.001305, "grad_norm": 0.258741, "tokens_per_sec": 106369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:42.639741+00:00", "epoch": 0, "step": 947, "train_loss": 4.0116119384765625, "perplexity": 55.23583560981422, "lr": 0.001305, "grad_norm": 0.28923, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:42.946549+00:00", "epoch": 0, "step": 948, "train_loss": 4.060802459716797, "perplexity": 58.02085179923836, "lr": 0.001305, "grad_norm": 0.314938, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:43.252673+00:00", "epoch": 0, "step": 949, "train_loss": 4.055590629577637, "perplexity": 57.71924362314183, "lr": 0.001305, "grad_norm": 0.308714, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:43.559273+00:00", "epoch": 0, "step": 950, "train_loss": 4.16005277633667, "perplexity": 64.07490414941643, "lr": 0.001305, "grad_norm": 0.331705, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:43.867230+00:00", "epoch": 0, "step": 951, "train_loss": 4.011694431304932, "perplexity": 55.24039235806795, "lr": 0.001305, "grad_norm": 0.361136, "tokens_per_sec": 106347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:44.175142+00:00", "epoch": 0, "step": 952, "train_loss": 3.958103656768799, "perplexity": 52.3579431158728, "lr": 0.001305, "grad_norm": 0.332145, "tokens_per_sec": 106421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:44.481788+00:00", "epoch": 0, "step": 953, "train_loss": 4.061156749725342, "perplexity": 58.04141164917774, "lr": 0.001305, "grad_norm": 0.349548, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:44.789192+00:00", "epoch": 0, "step": 954, "train_loss": 4.214336395263672, "perplexity": 67.6492586038237, "lr": 0.001305, "grad_norm": 0.386021, "tokens_per_sec": 106596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:45.094950+00:00", "epoch": 0, "step": 955, "train_loss": 4.026762962341309, "perplexity": 56.079087003439724, "lr": 0.001305, "grad_norm": 0.36628, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:45.402487+00:00", "epoch": 0, "step": 956, "train_loss": 3.997237205505371, "perplexity": 54.44751474780127, "lr": 0.001305, "grad_norm": 0.348699, "tokens_per_sec": 106549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:45.710097+00:00", "epoch": 0, "step": 957, "train_loss": 4.072167873382568, "perplexity": 58.684044371292615, "lr": 0.001305, "grad_norm": 0.330364, "tokens_per_sec": 106525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:46.018055+00:00", "epoch": 0, "step": 958, "train_loss": 3.935795307159424, "perplexity": 51.202855765474425, "lr": 0.001305, "grad_norm": 0.296443, "tokens_per_sec": 106405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:46.325217+00:00", "epoch": 0, "step": 959, "train_loss": 3.996562957763672, "perplexity": 54.41081600735316, "lr": 0.001305, "grad_norm": 0.276792, "tokens_per_sec": 106679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:46.632171+00:00", "epoch": 0, "step": 960, "train_loss": 3.99711537361145, "perplexity": 54.44088170802636, "lr": 0.001305, "grad_norm": 0.2652, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:46.938112+00:00", "epoch": 0, "step": 961, "train_loss": 4.006242752075195, "perplexity": 54.940058863190764, "lr": 0.001305, "grad_norm": 0.294555, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:47.245373+00:00", "epoch": 0, "step": 962, "train_loss": 4.01767110824585, "perplexity": 55.57153491761272, "lr": 0.001305, "grad_norm": 0.27422, "tokens_per_sec": 106646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:47.552257+00:00", "epoch": 0, "step": 963, "train_loss": 4.01902961730957, "perplexity": 55.64708065464882, "lr": 0.001305, "grad_norm": 0.292403, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:47.859348+00:00", "epoch": 0, "step": 964, "train_loss": 3.984926223754883, "perplexity": 53.78132154756082, "lr": 0.001305, "grad_norm": 0.316103, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:48.166378+00:00", "epoch": 0, "step": 965, "train_loss": 4.04016637802124, "perplexity": 56.835798246506286, "lr": 0.001305, "grad_norm": 0.297963, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:48.472807+00:00", "epoch": 0, "step": 966, "train_loss": 4.089048862457275, "perplexity": 59.68309786400547, "lr": 0.001305, "grad_norm": 0.267841, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:48.779236+00:00", "epoch": 0, "step": 967, "train_loss": 4.042877674102783, "perplexity": 56.99010601608092, "lr": 0.001305, "grad_norm": 0.299257, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:49.087106+00:00", "epoch": 0, "step": 968, "train_loss": 4.1125054359436035, "perplexity": 61.09960711396907, "lr": 0.001305, "grad_norm": 0.306497, "tokens_per_sec": 106435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:49.394094+00:00", "epoch": 0, "step": 969, "train_loss": 4.008237361907959, "perplexity": 55.04975220615451, "lr": 0.001305, "grad_norm": 0.292477, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:49.700948+00:00", "epoch": 0, "step": 970, "train_loss": 4.185989856719971, "perplexity": 65.75856026448263, "lr": 0.001305, "grad_norm": 0.312394, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:50.006683+00:00", "epoch": 0, "step": 971, "train_loss": 4.0995192527771, "perplexity": 60.31128614341331, "lr": 0.001305, "grad_norm": 0.289914, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:50.313612+00:00", "epoch": 0, "step": 972, "train_loss": 4.071750640869141, "perplexity": 58.65956458719753, "lr": 0.001305, "grad_norm": 0.273859, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:50.620969+00:00", "epoch": 0, "step": 973, "train_loss": 4.081226348876953, "perplexity": 59.218047323965486, "lr": 0.001305, "grad_norm": 0.26708, "tokens_per_sec": 106614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:50.928025+00:00", "epoch": 0, "step": 974, "train_loss": 4.104762554168701, "perplexity": 60.62834689011418, "lr": 0.001305, "grad_norm": 0.257899, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:51.234845+00:00", "epoch": 0, "step": 975, "train_loss": 4.066319465637207, "perplexity": 58.34183780881627, "lr": 0.001305, "grad_norm": 0.264799, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:51.541906+00:00", "epoch": 0, "step": 976, "train_loss": 4.025588035583496, "perplexity": 56.013236875679496, "lr": 0.001305, "grad_norm": 0.265668, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:51.847875+00:00", "epoch": 0, "step": 977, "train_loss": 4.157103061676025, "perplexity": 63.88617994338405, "lr": 0.001305, "grad_norm": 0.276349, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:52.154680+00:00", "epoch": 0, "step": 978, "train_loss": 4.045084476470947, "perplexity": 57.11601078937897, "lr": 0.001305, "grad_norm": 0.286921, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:52.461893+00:00", "epoch": 0, "step": 979, "train_loss": 4.139382362365723, "perplexity": 62.76404403956234, "lr": 0.001305, "grad_norm": 0.296966, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:52.768978+00:00", "epoch": 0, "step": 980, "train_loss": 4.034256458282471, "perplexity": 56.500893844419615, "lr": 0.001305, "grad_norm": 0.29362, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:53.075734+00:00", "epoch": 0, "step": 981, "train_loss": 3.9701764583587646, "perplexity": 52.99388122798087, "lr": 0.001305, "grad_norm": 0.301704, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:53.381867+00:00", "epoch": 0, "step": 982, "train_loss": 4.011613845825195, "perplexity": 55.235940963910224, "lr": 0.001305, "grad_norm": 0.314486, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:53.688432+00:00", "epoch": 0, "step": 983, "train_loss": 3.9863901138305664, "perplexity": 53.860109144552126, "lr": 0.001305, "grad_norm": 0.307365, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:53.995555+00:00", "epoch": 0, "step": 984, "train_loss": 4.046492099761963, "perplexity": 57.1964652279424, "lr": 0.001305, "grad_norm": 0.302243, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:54.302337+00:00", "epoch": 0, "step": 985, "train_loss": 3.991164207458496, "perplexity": 54.11785711534273, "lr": 0.001305, "grad_norm": 0.348012, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:54.609215+00:00", "epoch": 0, "step": 986, "train_loss": 4.025976657867432, "perplexity": 56.03500909803612, "lr": 0.001305, "grad_norm": 0.34631, "tokens_per_sec": 106779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:54.914919+00:00", "epoch": 0, "step": 987, "train_loss": 4.026087760925293, "perplexity": 56.04123510475198, "lr": 0.001305, "grad_norm": 0.335844, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:55.221418+00:00", "epoch": 0, "step": 988, "train_loss": 4.065795421600342, "perplexity": 58.311272126194154, "lr": 0.001305, "grad_norm": 0.402123, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:55.527468+00:00", "epoch": 0, "step": 989, "train_loss": 4.120201587677002, "perplexity": 61.57165309996832, "lr": 0.001305, "grad_norm": 0.400604, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:55.835704+00:00", "epoch": 0, "step": 990, "train_loss": 3.902393341064453, "perplexity": 49.520827619231184, "lr": 0.001305, "grad_norm": 0.380787, "tokens_per_sec": 106363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:56.143722+00:00", "epoch": 0, "step": 991, "train_loss": 4.151159286499023, "perplexity": 63.507581120634335, "lr": 0.001305, "grad_norm": 0.351852, "tokens_per_sec": 106329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:56.451995+00:00", "epoch": 0, "step": 992, "train_loss": 4.040367603302002, "perplexity": 56.84723619672953, "lr": 0.001305, "grad_norm": 0.329314, "tokens_per_sec": 106296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:56.757778+00:00", "epoch": 0, "step": 993, "train_loss": 3.9833734035491943, "perplexity": 53.69787363134617, "lr": 0.001305, "grad_norm": 0.333619, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:57.063696+00:00", "epoch": 0, "step": 994, "train_loss": 4.127874374389648, "perplexity": 62.04589631866157, "lr": 0.001305, "grad_norm": 0.346476, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:57.370470+00:00", "epoch": 0, "step": 995, "train_loss": 4.084075450897217, "perplexity": 59.387006158418615, "lr": 0.001305, "grad_norm": 0.292502, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:57.676866+00:00", "epoch": 0, "step": 996, "train_loss": 4.08484411239624, "perplexity": 59.432672212215095, "lr": 0.001305, "grad_norm": 0.275431, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:57.984063+00:00", "epoch": 0, "step": 997, "train_loss": 4.127431392669678, "perplexity": 62.0184172076145, "lr": 0.001305, "grad_norm": 0.308626, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:58.290162+00:00", "epoch": 0, "step": 998, "train_loss": 4.0695881843566895, "perplexity": 58.53285288339057, "lr": 0.001305, "grad_norm": 0.280369, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:58.597097+00:00", "epoch": 0, "step": 999, "train_loss": 4.076048851013184, "perplexity": 58.91223835622595, "lr": 0.001305, "grad_norm": 0.300194, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:33:58.904498+00:00", "epoch": 0, "step": 1000, "train_loss": 4.087186813354492, "perplexity": 59.572068408405926, "lr": 0.001305, "grad_norm": 0.289555, "tokens_per_sec": 106656} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:34:04.505949+00:00", "step": 1000, "epoch": 0, "val_loss": 4.058231282234192, "val_ppl": 57.871861513884646, "eval_train_loss": 4.087186813354492, "eval_train_ppl": 59.572068408405926} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:34:05.470066+00:00", "step": 1000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_4p0582_epoch_0000_step_0001000.pt", "val_loss": 4.058231282234192} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:34:07.093156+00:00", "step": 1000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0000_step_0001000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:08.066001+00:00", "epoch": 0, "step": 1001, "train_loss": 4.1441521644592285, "perplexity": 63.064131217503565, "lr": 0.001305, "grad_norm": 0.281629, "tokens_per_sec": 3577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:08.371821+00:00", "epoch": 0, "step": 1002, "train_loss": 4.099072456359863, "perplexity": 60.28434529583079, "lr": 0.001305, "grad_norm": 0.287092, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:08.677022+00:00", "epoch": 0, "step": 1003, "train_loss": 4.036978721618652, "perplexity": 56.654913702383496, "lr": 0.001305, "grad_norm": 0.289321, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:08.982190+00:00", "epoch": 0, "step": 1004, "train_loss": 3.9123849868774414, "perplexity": 50.018102348624296, "lr": 0.001305, "grad_norm": 0.257959, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:09.288046+00:00", "epoch": 0, "step": 1005, "train_loss": 4.107052326202393, "perplexity": 60.76733104356081, "lr": 0.001305, "grad_norm": 0.247853, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:09.593883+00:00", "epoch": 0, "step": 1006, "train_loss": 4.057790279388428, "perplexity": 57.84634548500157, "lr": 0.001305, "grad_norm": 0.244618, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:09.900198+00:00", "epoch": 0, "step": 1007, "train_loss": 4.003311634063721, "perplexity": 54.77925884414689, "lr": 0.001305, "grad_norm": 0.256814, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:10.206832+00:00", "epoch": 0, "step": 1008, "train_loss": 4.027257919311523, "perplexity": 56.10685060876355, "lr": 0.001305, "grad_norm": 0.238865, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:10.512134+00:00", "epoch": 0, "step": 1009, "train_loss": 3.9862329959869385, "perplexity": 53.851647425106734, "lr": 0.001305, "grad_norm": 0.248832, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:10.817982+00:00", "epoch": 0, "step": 1010, "train_loss": 3.944279909133911, "perplexity": 51.63913984771217, "lr": 0.001305, "grad_norm": 0.261569, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:11.123841+00:00", "epoch": 0, "step": 1011, "train_loss": 3.968545913696289, "perplexity": 52.907542746304784, "lr": 0.001305, "grad_norm": 0.254571, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:11.430021+00:00", "epoch": 0, "step": 1012, "train_loss": 4.047692775726318, "perplexity": 57.26518089335643, "lr": 0.001305, "grad_norm": 0.250418, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:11.736524+00:00", "epoch": 0, "step": 1013, "train_loss": 3.959660291671753, "perplexity": 52.43950878508741, "lr": 0.001305, "grad_norm": 0.254042, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:12.042931+00:00", "epoch": 0, "step": 1014, "train_loss": 3.9234907627105713, "perplexity": 50.576688204316426, "lr": 0.001305, "grad_norm": 0.264876, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:12.807753+00:00", "epoch": 0, "step": 1015, "train_loss": 4.01912260055542, "perplexity": 55.6522551413967, "lr": 0.001305, "grad_norm": 0.270596, "tokens_per_sec": 42844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:13.114539+00:00", "epoch": 0, "step": 1016, "train_loss": 4.031113624572754, "perplexity": 56.32359967958894, "lr": 0.001305, "grad_norm": 0.278855, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:13.421576+00:00", "epoch": 0, "step": 1017, "train_loss": 4.087916851043701, "perplexity": 59.61557414204861, "lr": 0.001305, "grad_norm": 0.246094, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:13.727070+00:00", "epoch": 0, "step": 1018, "train_loss": 4.129187107086182, "perplexity": 62.12739947967264, "lr": 0.001305, "grad_norm": 0.257455, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:14.032291+00:00", "epoch": 0, "step": 1019, "train_loss": 4.0073676109313965, "perplexity": 55.00189344603232, "lr": 0.001305, "grad_norm": 0.264769, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:14.338660+00:00", "epoch": 0, "step": 1020, "train_loss": 3.960592269897461, "perplexity": 52.48840404656232, "lr": 0.001305, "grad_norm": 0.270716, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:14.645403+00:00", "epoch": 0, "step": 1021, "train_loss": 4.075910568237305, "perplexity": 58.904092371610474, "lr": 0.001305, "grad_norm": 0.26564, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:14.952801+00:00", "epoch": 0, "step": 1022, "train_loss": 3.975370168685913, "perplexity": 53.26983208028034, "lr": 0.001305, "grad_norm": 0.232384, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:15.258785+00:00", "epoch": 0, "step": 1023, "train_loss": 4.000161170959473, "perplexity": 54.606950378531515, "lr": 0.001305, "grad_norm": 0.244252, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:15.565180+00:00", "epoch": 0, "step": 1024, "train_loss": 4.1064043045043945, "perplexity": 60.727965250836654, "lr": 0.001305, "grad_norm": 0.250227, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:15.871381+00:00", "epoch": 0, "step": 1025, "train_loss": 3.9614641666412354, "perplexity": 52.534188471881656, "lr": 0.001305, "grad_norm": 0.243324, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:16.178111+00:00", "epoch": 0, "step": 1026, "train_loss": 4.093003273010254, "perplexity": 59.91957659486922, "lr": 0.001305, "grad_norm": 0.233572, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:16.484051+00:00", "epoch": 0, "step": 1027, "train_loss": 4.0064849853515625, "perplexity": 54.953368785640976, "lr": 0.001305, "grad_norm": 0.256177, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:16.790493+00:00", "epoch": 0, "step": 1028, "train_loss": 4.009486198425293, "perplexity": 55.11854329244955, "lr": 0.001305, "grad_norm": 0.247404, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:17.096713+00:00", "epoch": 0, "step": 1029, "train_loss": 4.057577610015869, "perplexity": 57.834044647054576, "lr": 0.001305, "grad_norm": 0.24699, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:17.403102+00:00", "epoch": 0, "step": 1030, "train_loss": 4.05721378326416, "perplexity": 57.81300690173178, "lr": 0.001305, "grad_norm": 0.283894, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:17.708882+00:00", "epoch": 0, "step": 1031, "train_loss": 4.019762992858887, "perplexity": 55.68790583125515, "lr": 0.001305, "grad_norm": 0.332941, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:18.016583+00:00", "epoch": 0, "step": 1032, "train_loss": 4.056066036224365, "perplexity": 57.746690258844, "lr": 0.001305, "grad_norm": 0.419328, "tokens_per_sec": 106493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:18.323491+00:00", "epoch": 0, "step": 1033, "train_loss": 4.0663862228393555, "perplexity": 58.34573267668038, "lr": 0.001305, "grad_norm": 0.467357, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:18.630719+00:00", "epoch": 0, "step": 1034, "train_loss": 4.001575469970703, "perplexity": 54.68423557376337, "lr": 0.001305, "grad_norm": 0.505979, "tokens_per_sec": 106657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:18.937635+00:00", "epoch": 0, "step": 1035, "train_loss": 4.0692338943481445, "perplexity": 58.51211895157167, "lr": 0.001305, "grad_norm": 0.546987, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:19.244136+00:00", "epoch": 0, "step": 1036, "train_loss": 3.975301504135132, "perplexity": 53.26617445676632, "lr": 0.001305, "grad_norm": 0.388407, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:19.550830+00:00", "epoch": 0, "step": 1037, "train_loss": 3.910492420196533, "perplexity": 49.923529275828834, "lr": 0.001305, "grad_norm": 0.33878, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:19.857871+00:00", "epoch": 0, "step": 1038, "train_loss": 4.08123254776001, "perplexity": 59.218414410853455, "lr": 0.001305, "grad_norm": 0.306651, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:20.163795+00:00", "epoch": 0, "step": 1039, "train_loss": 3.8767614364624023, "perplexity": 48.26764384385958, "lr": 0.001305, "grad_norm": 0.322246, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:20.470478+00:00", "epoch": 0, "step": 1040, "train_loss": 4.080260276794434, "perplexity": 59.160866046729524, "lr": 0.001305, "grad_norm": 0.336274, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:20.778098+00:00", "epoch": 0, "step": 1041, "train_loss": 4.0124077796936035, "perplexity": 55.279812061265446, "lr": 0.001305, "grad_norm": 0.331216, "tokens_per_sec": 106453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:21.085284+00:00", "epoch": 0, "step": 1042, "train_loss": 3.904141902923584, "perplexity": 49.60749359796967, "lr": 0.001305, "grad_norm": 0.33272, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:21.391859+00:00", "epoch": 0, "step": 1043, "train_loss": 4.057636737823486, "perplexity": 57.83746434841892, "lr": 0.001305, "grad_norm": 0.312375, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:21.699201+00:00", "epoch": 0, "step": 1044, "train_loss": 4.0837860107421875, "perplexity": 59.36981966150055, "lr": 0.001305, "grad_norm": 0.306032, "tokens_per_sec": 106626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:22.004940+00:00", "epoch": 0, "step": 1045, "train_loss": 3.918031692504883, "perplexity": 50.30133877239512, "lr": 0.001305, "grad_norm": 0.258853, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:22.312341+00:00", "epoch": 0, "step": 1046, "train_loss": 3.9896011352539062, "perplexity": 54.0333330729135, "lr": 0.001305, "grad_norm": 0.273516, "tokens_per_sec": 106596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:22.620113+00:00", "epoch": 0, "step": 1047, "train_loss": 4.028600692749023, "perplexity": 56.18224000153231, "lr": 0.001305, "grad_norm": 0.272105, "tokens_per_sec": 106469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:22.926794+00:00", "epoch": 0, "step": 1048, "train_loss": 4.063503265380859, "perplexity": 58.177766647284, "lr": 0.001305, "grad_norm": 0.268121, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:23.233065+00:00", "epoch": 0, "step": 1049, "train_loss": 3.975489854812622, "perplexity": 53.2762081217067, "lr": 0.001305, "grad_norm": 0.254939, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:23.539758+00:00", "epoch": 0, "step": 1050, "train_loss": 4.000992298126221, "perplexity": 54.65235456419681, "lr": 0.001305, "grad_norm": 0.27755, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:23.847791+00:00", "epoch": 0, "step": 1051, "train_loss": 4.00183629989624, "perplexity": 54.698500719163654, "lr": 0.001305, "grad_norm": 0.262474, "tokens_per_sec": 106292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:24.154834+00:00", "epoch": 0, "step": 1052, "train_loss": 4.109462261199951, "perplexity": 60.91395296503851, "lr": 0.001305, "grad_norm": 0.280485, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:24.460792+00:00", "epoch": 0, "step": 1053, "train_loss": 4.161055564880371, "perplexity": 64.13918995638453, "lr": 0.001305, "grad_norm": 0.269253, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:24.766929+00:00", "epoch": 0, "step": 1054, "train_loss": 3.8983330726623535, "perplexity": 49.320167410425604, "lr": 0.001305, "grad_norm": 0.272548, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:25.072825+00:00", "epoch": 0, "step": 1055, "train_loss": 4.119334697723389, "perplexity": 61.518300381277726, "lr": 0.001305, "grad_norm": 0.28357, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:25.379968+00:00", "epoch": 0, "step": 1056, "train_loss": 4.060200214385986, "perplexity": 57.985919532105825, "lr": 0.001305, "grad_norm": 0.287073, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:25.687565+00:00", "epoch": 0, "step": 1057, "train_loss": 3.9954371452331543, "perplexity": 54.349594097559326, "lr": 0.001305, "grad_norm": 0.245509, "tokens_per_sec": 106529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:25.995279+00:00", "epoch": 0, "step": 1058, "train_loss": 4.174349784851074, "perplexity": 64.99756351776952, "lr": 0.001305, "grad_norm": 0.254892, "tokens_per_sec": 106489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:26.302363+00:00", "epoch": 0, "step": 1059, "train_loss": 4.011069297790527, "perplexity": 55.205870528956495, "lr": 0.001305, "grad_norm": 0.246205, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:26.609052+00:00", "epoch": 0, "step": 1060, "train_loss": 4.028796195983887, "perplexity": 56.19322488494961, "lr": 0.001305, "grad_norm": 0.248589, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:26.915530+00:00", "epoch": 0, "step": 1061, "train_loss": 3.897231340408325, "perplexity": 49.265859712979704, "lr": 0.001305, "grad_norm": 0.257786, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:27.222864+00:00", "epoch": 0, "step": 1062, "train_loss": 4.067622184753418, "perplexity": 58.417890363007196, "lr": 0.001305, "grad_norm": 0.261339, "tokens_per_sec": 106621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:27.529792+00:00", "epoch": 0, "step": 1063, "train_loss": 3.9655215740203857, "perplexity": 52.74777408475236, "lr": 0.001305, "grad_norm": 0.282655, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:27.837033+00:00", "epoch": 0, "step": 1064, "train_loss": 3.9176347255706787, "perplexity": 50.281374766943465, "lr": 0.001305, "grad_norm": 0.317933, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:28.143289+00:00", "epoch": 0, "step": 1065, "train_loss": 3.9974193572998047, "perplexity": 54.45743336363444, "lr": 0.001305, "grad_norm": 0.328847, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:28.449523+00:00", "epoch": 0, "step": 1066, "train_loss": 3.9929733276367188, "perplexity": 54.21585143767795, "lr": 0.001305, "grad_norm": 0.362848, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:28.756160+00:00", "epoch": 0, "step": 1067, "train_loss": 4.033279895782471, "perplexity": 56.44574412323237, "lr": 0.001305, "grad_norm": 0.372623, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:29.063241+00:00", "epoch": 0, "step": 1068, "train_loss": 3.954562187194824, "perplexity": 52.17284700295448, "lr": 0.001305, "grad_norm": 0.338033, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:29.369274+00:00", "epoch": 0, "step": 1069, "train_loss": 4.079719543457031, "perplexity": 59.128884441727735, "lr": 0.001305, "grad_norm": 0.281286, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:29.676378+00:00", "epoch": 0, "step": 1070, "train_loss": 4.087978839874268, "perplexity": 59.61926975631544, "lr": 0.001305, "grad_norm": 0.312273, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:29.982811+00:00", "epoch": 0, "step": 1071, "train_loss": 4.041576862335205, "perplexity": 56.91602081143938, "lr": 0.001305, "grad_norm": 0.294004, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:30.289058+00:00", "epoch": 0, "step": 1072, "train_loss": 4.038137435913086, "perplexity": 56.72059860841097, "lr": 0.001305, "grad_norm": 0.277102, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:30.595087+00:00", "epoch": 0, "step": 1073, "train_loss": 4.0248494148254395, "perplexity": 55.971879611744235, "lr": 0.001305, "grad_norm": 0.241272, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:30.902988+00:00", "epoch": 0, "step": 1074, "train_loss": 3.9216384887695312, "perplexity": 50.483093031441115, "lr": 0.001305, "grad_norm": 0.264355, "tokens_per_sec": 106424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:31.210777+00:00", "epoch": 0, "step": 1075, "train_loss": 4.00624942779541, "perplexity": 54.94042562887653, "lr": 0.001305, "grad_norm": 0.259266, "tokens_per_sec": 106463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:31.518665+00:00", "epoch": 0, "step": 1076, "train_loss": 3.9053759574890137, "perplexity": 49.6687497408613, "lr": 0.001305, "grad_norm": 0.266383, "tokens_per_sec": 106432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:31.824807+00:00", "epoch": 0, "step": 1077, "train_loss": 3.9605929851531982, "perplexity": 52.48844158920789, "lr": 0.001305, "grad_norm": 0.25961, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:32.131397+00:00", "epoch": 0, "step": 1078, "train_loss": 3.8421754837036133, "perplexity": 46.62680002735589, "lr": 0.001305, "grad_norm": 0.228675, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:32.438004+00:00", "epoch": 0, "step": 1079, "train_loss": 3.9118919372558594, "perplexity": 49.993447020838545, "lr": 0.001305, "grad_norm": 0.247103, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:32.746668+00:00", "epoch": 0, "step": 1080, "train_loss": 4.134206295013428, "perplexity": 62.44001244842654, "lr": 0.001305, "grad_norm": 0.262212, "tokens_per_sec": 106220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:33.054643+00:00", "epoch": 0, "step": 1081, "train_loss": 3.8887221813201904, "perplexity": 48.84842719363392, "lr": 0.001305, "grad_norm": 0.255895, "tokens_per_sec": 106342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:33.361416+00:00", "epoch": 0, "step": 1082, "train_loss": 3.9691736698150635, "perplexity": 52.94076620701742, "lr": 0.001305, "grad_norm": 0.241838, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:33.668386+00:00", "epoch": 0, "step": 1083, "train_loss": 4.008169651031494, "perplexity": 55.046024865375635, "lr": 0.001305, "grad_norm": 0.246636, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:33.975913+00:00", "epoch": 0, "step": 1084, "train_loss": 4.003704071044922, "perplexity": 54.80076046986071, "lr": 0.001305, "grad_norm": 0.3075, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:34.283918+00:00", "epoch": 0, "step": 1085, "train_loss": 4.023764133453369, "perplexity": 55.91116732435822, "lr": 0.001305, "grad_norm": 0.34941, "tokens_per_sec": 106389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:34.592266+00:00", "epoch": 0, "step": 1086, "train_loss": 3.9978864192962646, "perplexity": 54.482874301968586, "lr": 0.001305, "grad_norm": 0.392677, "tokens_per_sec": 106269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:34.900281+00:00", "epoch": 0, "step": 1087, "train_loss": 4.0809221267700195, "perplexity": 59.20003462490965, "lr": 0.001305, "grad_norm": 0.372312, "tokens_per_sec": 106385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:35.207567+00:00", "epoch": 0, "step": 1088, "train_loss": 4.03130578994751, "perplexity": 56.33442416524132, "lr": 0.001305, "grad_norm": 0.366868, "tokens_per_sec": 106637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:35.525711+00:00", "epoch": 0, "step": 1089, "train_loss": 3.9495019912719727, "perplexity": 51.90950900810051, "lr": 0.001305, "grad_norm": 0.348072, "tokens_per_sec": 102998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:35.832945+00:00", "epoch": 0, "step": 1090, "train_loss": 4.037606239318848, "perplexity": 56.69047682061086, "lr": 0.001305, "grad_norm": 0.345449, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:36.152322+00:00", "epoch": 0, "step": 1091, "train_loss": 3.928891658782959, "perplexity": 50.850586623648354, "lr": 0.001305, "grad_norm": 0.318323, "tokens_per_sec": 102546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:36.460703+00:00", "epoch": 0, "step": 1092, "train_loss": 3.8808681964874268, "perplexity": 48.4662750603019, "lr": 0.001305, "grad_norm": 0.328987, "tokens_per_sec": 106258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:36.768909+00:00", "epoch": 0, "step": 1093, "train_loss": 4.060635566711426, "perplexity": 58.011169332897815, "lr": 0.001305, "grad_norm": 0.311965, "tokens_per_sec": 106319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:37.085546+00:00", "epoch": 0, "step": 1094, "train_loss": 3.918843984603882, "perplexity": 50.342214751819014, "lr": 0.001305, "grad_norm": 0.328005, "tokens_per_sec": 103488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:37.392768+00:00", "epoch": 0, "step": 1095, "train_loss": 4.017186641693115, "perplexity": 55.544618888145486, "lr": 0.001305, "grad_norm": 0.339493, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:37.699413+00:00", "epoch": 0, "step": 1096, "train_loss": 4.047680854797363, "perplexity": 57.26449824327232, "lr": 0.001305, "grad_norm": 0.284325, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:38.018688+00:00", "epoch": 0, "step": 1097, "train_loss": 3.988290309906006, "perplexity": 53.962551211754025, "lr": 0.001305, "grad_norm": 0.267861, "tokens_per_sec": 102633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:38.326311+00:00", "epoch": 0, "step": 1098, "train_loss": 4.025616645812988, "perplexity": 56.01483945016601, "lr": 0.001305, "grad_norm": 0.266161, "tokens_per_sec": 106521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:38.643744+00:00", "epoch": 0, "step": 1099, "train_loss": 4.138766288757324, "perplexity": 62.7253886769702, "lr": 0.001305, "grad_norm": 0.2482, "tokens_per_sec": 103228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:38.951716+00:00", "epoch": 0, "step": 1100, "train_loss": 3.9567761421203613, "perplexity": 52.2884832940906, "lr": 0.001305, "grad_norm": 0.26034, "tokens_per_sec": 106457} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:34:44.552176+00:00", "step": 1100, "epoch": 0, "val_loss": 4.007558274269104, "val_ppl": 55.012381290408904, "eval_train_loss": 3.9567761421203613, "eval_train_ppl": 52.2884832940906} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:34:45.519264+00:00", "step": 1100, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_4p0076_epoch_0000_step_0001100.pt", "val_loss": 4.007558274269104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:46.491997+00:00", "epoch": 0, "step": 1101, "train_loss": 4.0437541007995605, "perplexity": 57.04007556056843, "lr": 0.001305, "grad_norm": 0.259528, "tokens_per_sec": 4346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:46.798053+00:00", "epoch": 0, "step": 1102, "train_loss": 3.905341863632202, "perplexity": 49.667056370486556, "lr": 0.001305, "grad_norm": 0.285822, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:47.103297+00:00", "epoch": 0, "step": 1103, "train_loss": 3.948481559753418, "perplexity": 51.85656592598483, "lr": 0.001305, "grad_norm": 0.271024, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:47.408191+00:00", "epoch": 0, "step": 1104, "train_loss": 3.997697114944458, "perplexity": 54.472561432930576, "lr": 0.001305, "grad_norm": 0.254025, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:47.714499+00:00", "epoch": 0, "step": 1105, "train_loss": 3.958773612976074, "perplexity": 52.393032397692075, "lr": 0.001305, "grad_norm": 0.292637, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:48.021298+00:00", "epoch": 0, "step": 1106, "train_loss": 3.9975576400756836, "perplexity": 54.46496440938216, "lr": 0.001305, "grad_norm": 0.294623, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:48.328996+00:00", "epoch": 0, "step": 1107, "train_loss": 3.98958158493042, "perplexity": 54.03227671409899, "lr": 0.001305, "grad_norm": 0.289244, "tokens_per_sec": 106495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:48.634761+00:00", "epoch": 0, "step": 1108, "train_loss": 3.929331064224243, "perplexity": 50.87293555786534, "lr": 0.001305, "grad_norm": 0.277636, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:48.940601+00:00", "epoch": 0, "step": 1109, "train_loss": 4.049584865570068, "perplexity": 57.37363432999294, "lr": 0.001305, "grad_norm": 0.273111, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:49.247349+00:00", "epoch": 0, "step": 1110, "train_loss": 3.954371213912964, "perplexity": 52.16288433447035, "lr": 0.001305, "grad_norm": 0.278665, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:49.553654+00:00", "epoch": 0, "step": 1111, "train_loss": 3.992633581161499, "perplexity": 54.19743492190127, "lr": 0.001305, "grad_norm": 0.289083, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:49.861640+00:00", "epoch": 0, "step": 1112, "train_loss": 3.948606491088867, "perplexity": 51.863044840719084, "lr": 0.001305, "grad_norm": 0.301504, "tokens_per_sec": 106395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:50.169442+00:00", "epoch": 0, "step": 1113, "train_loss": 3.9377505779266357, "perplexity": 51.30306915277675, "lr": 0.001305, "grad_norm": 0.293345, "tokens_per_sec": 106458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:50.476264+00:00", "epoch": 0, "step": 1114, "train_loss": 3.976619243621826, "perplexity": 53.33641166516108, "lr": 0.001305, "grad_norm": 0.262867, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:50.794037+00:00", "epoch": 0, "step": 1115, "train_loss": 3.946390390396118, "perplexity": 51.748238369455265, "lr": 0.001305, "grad_norm": 0.288172, "tokens_per_sec": 103117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:51.100792+00:00", "epoch": 0, "step": 1116, "train_loss": 3.9497599601745605, "perplexity": 51.922901774557566, "lr": 0.001305, "grad_norm": 0.275824, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:51.407987+00:00", "epoch": 0, "step": 1117, "train_loss": 4.047934055328369, "perplexity": 57.2789994804183, "lr": 0.001305, "grad_norm": 0.266359, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:51.714500+00:00", "epoch": 0, "step": 1118, "train_loss": 3.9337658882141113, "perplexity": 51.0990490891753, "lr": 0.001305, "grad_norm": 0.288395, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:52.021993+00:00", "epoch": 0, "step": 1119, "train_loss": 4.068431854248047, "perplexity": 58.4652087002968, "lr": 0.001305, "grad_norm": 0.27478, "tokens_per_sec": 106566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:52.328605+00:00", "epoch": 0, "step": 1120, "train_loss": 3.95625376701355, "perplexity": 52.26117622493237, "lr": 0.001305, "grad_norm": 0.291643, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:52.634643+00:00", "epoch": 0, "step": 1121, "train_loss": 3.980205774307251, "perplexity": 53.52804779125081, "lr": 0.001305, "grad_norm": 0.24917, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:52.941259+00:00", "epoch": 0, "step": 1122, "train_loss": 3.9107744693756104, "perplexity": 49.93761215221622, "lr": 0.001305, "grad_norm": 0.265916, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:53.248327+00:00", "epoch": 0, "step": 1123, "train_loss": 4.038055896759033, "perplexity": 56.71597384733519, "lr": 0.001305, "grad_norm": 0.288199, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:53.556294+00:00", "epoch": 0, "step": 1124, "train_loss": 3.9168179035186768, "perplexity": 50.2403206004857, "lr": 0.001305, "grad_norm": 0.264001, "tokens_per_sec": 106401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:53.863347+00:00", "epoch": 0, "step": 1125, "train_loss": 3.985640048980713, "perplexity": 53.81972571686163, "lr": 0.001305, "grad_norm": 0.281341, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:54.169696+00:00", "epoch": 0, "step": 1126, "train_loss": 3.9262125492095947, "perplexity": 50.71453466064955, "lr": 0.001305, "grad_norm": 0.300438, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:54.475505+00:00", "epoch": 0, "step": 1127, "train_loss": 4.038968563079834, "perplexity": 56.76776023476825, "lr": 0.001305, "grad_norm": 0.296806, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:54.782144+00:00", "epoch": 0, "step": 1128, "train_loss": 4.083053112030029, "perplexity": 59.32632353820291, "lr": 0.001305, "grad_norm": 0.273968, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:55.089926+00:00", "epoch": 0, "step": 1129, "train_loss": 3.9621028900146484, "perplexity": 52.56775400436587, "lr": 0.001305, "grad_norm": 0.282608, "tokens_per_sec": 106464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:55.397590+00:00", "epoch": 0, "step": 1130, "train_loss": 4.013320446014404, "perplexity": 55.330287113890606, "lr": 0.001305, "grad_norm": 0.339157, "tokens_per_sec": 106562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:55.704176+00:00", "epoch": 0, "step": 1131, "train_loss": 3.8560147285461426, "perplexity": 47.276565490346336, "lr": 0.001305, "grad_norm": 0.307401, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:56.011231+00:00", "epoch": 0, "step": 1132, "train_loss": 3.9546854496002197, "perplexity": 52.1792783499359, "lr": 0.001305, "grad_norm": 0.263309, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:56.317001+00:00", "epoch": 0, "step": 1133, "train_loss": 4.011613845825195, "perplexity": 55.235940963910224, "lr": 0.001305, "grad_norm": 0.250919, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:56.623779+00:00", "epoch": 0, "step": 1134, "train_loss": 3.8618056774139404, "perplexity": 47.5511359080593, "lr": 0.001305, "grad_norm": 0.26475, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:56.930656+00:00", "epoch": 0, "step": 1135, "train_loss": 3.973384380340576, "perplexity": 53.164154430027416, "lr": 0.001305, "grad_norm": 0.257203, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:57.237901+00:00", "epoch": 0, "step": 1136, "train_loss": 3.900155544281006, "perplexity": 49.41013397160921, "lr": 0.001305, "grad_norm": 0.254079, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:57.545624+00:00", "epoch": 0, "step": 1137, "train_loss": 3.93023943901062, "perplexity": 50.919168244959664, "lr": 0.001305, "grad_norm": 0.287197, "tokens_per_sec": 106486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:57.853430+00:00", "epoch": 0, "step": 1138, "train_loss": 3.9526724815368652, "perplexity": 52.07434877442237, "lr": 0.001305, "grad_norm": 0.28811, "tokens_per_sec": 106457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:58.160916+00:00", "epoch": 0, "step": 1139, "train_loss": 3.938880205154419, "perplexity": 51.361055241726895, "lr": 0.001305, "grad_norm": 0.249479, "tokens_per_sec": 106567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:58.468365+00:00", "epoch": 0, "step": 1140, "train_loss": 3.963557720184326, "perplexity": 52.64428681646695, "lr": 0.001305, "grad_norm": 0.260647, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:58.775932+00:00", "epoch": 0, "step": 1141, "train_loss": 3.84696364402771, "perplexity": 46.85059196957247, "lr": 0.001305, "grad_norm": 0.303303, "tokens_per_sec": 106481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:59.082529+00:00", "epoch": 0, "step": 1142, "train_loss": 3.9534683227539062, "perplexity": 52.11580818289243, "lr": 0.001305, "grad_norm": 0.272044, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:59.389082+00:00", "epoch": 0, "step": 1143, "train_loss": 3.926612138748169, "perplexity": 50.73480370753345, "lr": 0.001305, "grad_norm": 0.28343, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:34:59.696465+00:00", "epoch": 0, "step": 1144, "train_loss": 4.022297382354736, "perplexity": 55.82921967135597, "lr": 0.001305, "grad_norm": 0.319704, "tokens_per_sec": 106604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:00.004978+00:00", "epoch": 0, "step": 1145, "train_loss": 3.959508180618286, "perplexity": 52.43153276279879, "lr": 0.001305, "grad_norm": 0.321079, "tokens_per_sec": 106222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:00.312743+00:00", "epoch": 0, "step": 1146, "train_loss": 3.8966991901397705, "perplexity": 49.239649846915015, "lr": 0.001305, "grad_norm": 0.305874, "tokens_per_sec": 106461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:00.619867+00:00", "epoch": 0, "step": 1147, "train_loss": 3.93045711517334, "perplexity": 50.93025334054887, "lr": 0.001305, "grad_norm": 0.256098, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:00.926359+00:00", "epoch": 0, "step": 1148, "train_loss": 3.9824159145355225, "perplexity": 53.6464831141406, "lr": 0.001305, "grad_norm": 0.262072, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:01.234070+00:00", "epoch": 0, "step": 1149, "train_loss": 3.992077112197876, "perplexity": 54.167284121229685, "lr": 0.001305, "grad_norm": 0.252678, "tokens_per_sec": 106495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:01.541824+00:00", "epoch": 0, "step": 1150, "train_loss": 4.017108917236328, "perplexity": 55.540301880585666, "lr": 0.001305, "grad_norm": 0.249111, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:01.848195+00:00", "epoch": 0, "step": 1151, "train_loss": 3.9411964416503906, "perplexity": 51.480157473569335, "lr": 0.001305, "grad_norm": 0.245959, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:02.155582+00:00", "epoch": 0, "step": 1152, "train_loss": 3.889112710952759, "perplexity": 48.86750767746206, "lr": 0.001305, "grad_norm": 0.245246, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:02.462452+00:00", "epoch": 0, "step": 1153, "train_loss": 3.861926317214966, "perplexity": 47.5568728136764, "lr": 0.001305, "grad_norm": 0.27797, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:02.769365+00:00", "epoch": 0, "step": 1154, "train_loss": 4.054347991943359, "perplexity": 57.647564063876324, "lr": 0.001305, "grad_norm": 0.3, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:03.075937+00:00", "epoch": 0, "step": 1155, "train_loss": 3.9244844913482666, "perplexity": 50.6269726882127, "lr": 0.001305, "grad_norm": 0.294671, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:03.382741+00:00", "epoch": 0, "step": 1156, "train_loss": 3.8137779235839844, "perplexity": 45.32133638976703, "lr": 0.001305, "grad_norm": 0.266611, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:03.690785+00:00", "epoch": 0, "step": 1157, "train_loss": 4.026779651641846, "perplexity": 56.08002293198655, "lr": 0.001305, "grad_norm": 0.305321, "tokens_per_sec": 106374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:03.997961+00:00", "epoch": 0, "step": 1158, "train_loss": 4.051217079162598, "perplexity": 57.46735682255496, "lr": 0.001305, "grad_norm": 0.311704, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:04.306222+00:00", "epoch": 0, "step": 1159, "train_loss": 3.883340358734131, "perplexity": 48.586239780772914, "lr": 0.001305, "grad_norm": 0.322674, "tokens_per_sec": 106301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:04.613462+00:00", "epoch": 0, "step": 1160, "train_loss": 3.9510436058044434, "perplexity": 51.9895951767006, "lr": 0.001305, "grad_norm": 0.289663, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:04.931615+00:00", "epoch": 0, "step": 1161, "train_loss": 3.9731099605560303, "perplexity": 53.149567135835774, "lr": 0.001305, "grad_norm": 0.247981, "tokens_per_sec": 102944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:05.237762+00:00", "epoch": 0, "step": 1162, "train_loss": 3.9886553287506104, "perplexity": 53.98225215523832, "lr": 0.001305, "grad_norm": 0.278915, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:05.544189+00:00", "epoch": 0, "step": 1163, "train_loss": 3.968203544616699, "perplexity": 52.889431940058316, "lr": 0.001305, "grad_norm": 0.288212, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:05.853725+00:00", "epoch": 0, "step": 1164, "train_loss": 3.9216268062591553, "perplexity": 50.48250326562794, "lr": 0.001305, "grad_norm": 0.313209, "tokens_per_sec": 105863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:06.161151+00:00", "epoch": 0, "step": 1165, "train_loss": 4.000921249389648, "perplexity": 54.648471721391495, "lr": 0.001305, "grad_norm": 0.312001, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:06.468128+00:00", "epoch": 0, "step": 1166, "train_loss": 3.9721996784210205, "perplexity": 53.10120804793466, "lr": 0.001305, "grad_norm": 0.311726, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:06.786673+00:00", "epoch": 0, "step": 1167, "train_loss": 3.994682550430298, "perplexity": 54.30859764611587, "lr": 0.001305, "grad_norm": 0.289754, "tokens_per_sec": 102867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:07.094316+00:00", "epoch": 0, "step": 1168, "train_loss": 4.009594917297363, "perplexity": 55.12453604406316, "lr": 0.001305, "grad_norm": 0.320232, "tokens_per_sec": 106523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:07.401152+00:00", "epoch": 0, "step": 1169, "train_loss": 3.868854522705078, "perplexity": 47.88750060713261, "lr": 0.001305, "grad_norm": 0.277913, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:07.709578+00:00", "epoch": 0, "step": 1170, "train_loss": 4.0261430740356445, "perplexity": 56.04433500550536, "lr": 0.001305, "grad_norm": 0.228723, "tokens_per_sec": 106299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:08.029682+00:00", "epoch": 0, "step": 1171, "train_loss": 3.955810070037842, "perplexity": 52.23799324258433, "lr": 0.001305, "grad_norm": 0.26525, "tokens_per_sec": 102315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:08.336271+00:00", "epoch": 0, "step": 1172, "train_loss": 3.994131326675415, "perplexity": 54.27866970625245, "lr": 0.001305, "grad_norm": 0.263991, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:08.643207+00:00", "epoch": 0, "step": 1173, "train_loss": 3.9874703884124756, "perplexity": 53.918324289949375, "lr": 0.001305, "grad_norm": 0.292876, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:08.963130+00:00", "epoch": 0, "step": 1174, "train_loss": 3.9815008640289307, "perplexity": 53.597416325308835, "lr": 0.001305, "grad_norm": 0.294682, "tokens_per_sec": 102425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:09.270168+00:00", "epoch": 0, "step": 1175, "train_loss": 3.977630376815796, "perplexity": 53.39036915595082, "lr": 0.001305, "grad_norm": 0.296406, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:09.577355+00:00", "epoch": 0, "step": 1176, "train_loss": 3.886021852493286, "perplexity": 48.71669831322534, "lr": 0.001305, "grad_norm": 0.285906, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:09.884649+00:00", "epoch": 0, "step": 1177, "train_loss": 3.954960584640503, "perplexity": 52.19363667293497, "lr": 0.001305, "grad_norm": 0.3466, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:10.202729+00:00", "epoch": 0, "step": 1178, "train_loss": 3.9845616817474365, "perplexity": 53.76171956973001, "lr": 0.001305, "grad_norm": 0.415435, "tokens_per_sec": 103018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:10.510104+00:00", "epoch": 0, "step": 1179, "train_loss": 3.9733965396881104, "perplexity": 53.16480087538766, "lr": 0.001305, "grad_norm": 0.371715, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:10.826949+00:00", "epoch": 0, "step": 1180, "train_loss": 3.910010814666748, "perplexity": 49.899491616862555, "lr": 0.001305, "grad_norm": 0.30249, "tokens_per_sec": 103474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:11.132929+00:00", "epoch": 0, "step": 1181, "train_loss": 3.9624171257019043, "perplexity": 52.58427526432171, "lr": 0.001305, "grad_norm": 0.300063, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:11.453430+00:00", "epoch": 0, "step": 1182, "train_loss": 3.946866512298584, "perplexity": 51.77288270554553, "lr": 0.001305, "grad_norm": 0.302539, "tokens_per_sec": 102241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:11.760530+00:00", "epoch": 0, "step": 1183, "train_loss": 4.012011528015137, "perplexity": 55.257911682268364, "lr": 0.001305, "grad_norm": 0.26435, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:12.067904+00:00", "epoch": 0, "step": 1184, "train_loss": 3.910120725631714, "perplexity": 49.90497641955193, "lr": 0.001305, "grad_norm": 0.287549, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:12.384301+00:00", "epoch": 0, "step": 1185, "train_loss": 3.951448440551758, "perplexity": 52.01064663221976, "lr": 0.001305, "grad_norm": 0.269035, "tokens_per_sec": 103567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:12.691275+00:00", "epoch": 0, "step": 1186, "train_loss": 3.946991443634033, "perplexity": 51.77935116497021, "lr": 0.001305, "grad_norm": 0.254325, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:13.010513+00:00", "epoch": 0, "step": 1187, "train_loss": 3.971404790878296, "perplexity": 53.059015330608034, "lr": 0.001305, "grad_norm": 0.269635, "tokens_per_sec": 102645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:13.317274+00:00", "epoch": 0, "step": 1188, "train_loss": 3.9926812648773193, "perplexity": 54.2000193186026, "lr": 0.001305, "grad_norm": 0.249069, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:13.623751+00:00", "epoch": 0, "step": 1189, "train_loss": 3.8838231563568115, "perplexity": 48.60970276531489, "lr": 0.001305, "grad_norm": 0.251713, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:13.931970+00:00", "epoch": 0, "step": 1190, "train_loss": 3.9441452026367188, "perplexity": 51.632184188562015, "lr": 0.001305, "grad_norm": 0.256404, "tokens_per_sec": 106372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:14.253227+00:00", "epoch": 0, "step": 1191, "train_loss": 3.8942604064941406, "perplexity": 49.119711305674436, "lr": 0.001305, "grad_norm": 0.256649, "tokens_per_sec": 101947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:14.560192+00:00", "epoch": 0, "step": 1192, "train_loss": 3.9508676528930664, "perplexity": 51.980448260804785, "lr": 0.001305, "grad_norm": 0.251236, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:14.866685+00:00", "epoch": 0, "step": 1193, "train_loss": 3.984304189682007, "perplexity": 53.747878135623154, "lr": 0.001305, "grad_norm": 0.27014, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:15.185721+00:00", "epoch": 0, "step": 1194, "train_loss": 3.905275344848633, "perplexity": 49.66375268819301, "lr": 0.001305, "grad_norm": 0.263826, "tokens_per_sec": 102709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:15.507500+00:00", "epoch": 0, "step": 1195, "train_loss": 3.9202582836151123, "perplexity": 50.413464068408345, "lr": 0.001305, "grad_norm": 0.255222, "tokens_per_sec": 101834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:15.814047+00:00", "epoch": 0, "step": 1196, "train_loss": 3.8778324127197266, "perplexity": 48.319365035553844, "lr": 0.001305, "grad_norm": 0.236555, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:16.121279+00:00", "epoch": 0, "step": 1197, "train_loss": 4.000663757324219, "perplexity": 54.63440198503585, "lr": 0.001305, "grad_norm": 0.262885, "tokens_per_sec": 106656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:16.440404+00:00", "epoch": 0, "step": 1198, "train_loss": 4.017489433288574, "perplexity": 55.561439878416515, "lr": 0.001305, "grad_norm": 0.275607, "tokens_per_sec": 102681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:16.747761+00:00", "epoch": 0, "step": 1199, "train_loss": 4.043247222900391, "perplexity": 57.01117053317844, "lr": 0.001305, "grad_norm": 0.2903, "tokens_per_sec": 106611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:17.065469+00:00", "epoch": 0, "step": 1200, "train_loss": 3.995788335800171, "perplexity": 54.3686845143185, "lr": 0.001305, "grad_norm": 0.306091, "tokens_per_sec": 103192} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:35:22.665152+00:00", "step": 1200, "epoch": 0, "val_loss": 3.968159782886505, "val_ppl": 52.887117457650874, "eval_train_loss": 3.995788335800171, "eval_train_ppl": 54.3686845143185} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:35:23.728025+00:00", "step": 1200, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p9682_epoch_0000_step_0001200.pt", "val_loss": 3.968159782886505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:24.725180+00:00", "epoch": 0, "step": 1201, "train_loss": 4.022805213928223, "perplexity": 55.85757871202685, "lr": 0.001305, "grad_norm": 0.272265, "tokens_per_sec": 4278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:25.032445+00:00", "epoch": 0, "step": 1202, "train_loss": 3.9841270446777344, "perplexity": 53.73835781078507, "lr": 0.001305, "grad_norm": 0.253213, "tokens_per_sec": 106643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:25.339987+00:00", "epoch": 0, "step": 1203, "train_loss": 3.911747694015503, "perplexity": 49.98623632410331, "lr": 0.001305, "grad_norm": 0.304211, "tokens_per_sec": 106548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:25.646978+00:00", "epoch": 0, "step": 1204, "train_loss": 3.876687526702881, "perplexity": 48.26407652574185, "lr": 0.001305, "grad_norm": 0.257711, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:25.953238+00:00", "epoch": 0, "step": 1205, "train_loss": 3.884369134902954, "perplexity": 48.636249866573536, "lr": 0.001305, "grad_norm": 0.27257, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:26.259812+00:00", "epoch": 0, "step": 1206, "train_loss": 3.943179130554199, "perplexity": 51.58232786313756, "lr": 0.001305, "grad_norm": 0.248879, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:26.566180+00:00", "epoch": 0, "step": 1207, "train_loss": 3.9500935077667236, "perplexity": 51.94022342205891, "lr": 0.001305, "grad_norm": 0.250364, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:26.873808+00:00", "epoch": 0, "step": 1208, "train_loss": 3.8846583366394043, "perplexity": 48.650317588596145, "lr": 0.001305, "grad_norm": 0.259924, "tokens_per_sec": 106519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:27.179619+00:00", "epoch": 0, "step": 1209, "train_loss": 3.9056973457336426, "perplexity": 49.684715258580866, "lr": 0.001305, "grad_norm": 0.252203, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:27.486030+00:00", "epoch": 0, "step": 1210, "train_loss": 3.947193145751953, "perplexity": 51.78979622312446, "lr": 0.001305, "grad_norm": 0.265717, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:27.792038+00:00", "epoch": 0, "step": 1211, "train_loss": 3.9649486541748047, "perplexity": 52.71756249340549, "lr": 0.001305, "grad_norm": 0.281883, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:28.098670+00:00", "epoch": 0, "step": 1212, "train_loss": 3.801523447036743, "perplexity": 44.76933627999372, "lr": 0.001305, "grad_norm": 0.316135, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:28.405500+00:00", "epoch": 0, "step": 1213, "train_loss": 3.9602572917938232, "perplexity": 52.47082452505361, "lr": 0.001305, "grad_norm": 0.298087, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:28.711152+00:00", "epoch": 0, "step": 1214, "train_loss": 3.964252471923828, "perplexity": 52.680874234422, "lr": 0.001305, "grad_norm": 0.303426, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:29.045140+00:00", "epoch": 0, "step": 1215, "train_loss": 4.082171440124512, "perplexity": 59.274040237227275, "lr": 0.001305, "grad_norm": 0.313099, "tokens_per_sec": 98112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:29.351247+00:00", "epoch": 0, "step": 1216, "train_loss": 3.940539598464966, "perplexity": 51.44635418589315, "lr": 0.001305, "grad_norm": 0.254809, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:29.657346+00:00", "epoch": 0, "step": 1217, "train_loss": 3.978085517883301, "perplexity": 53.41467483640003, "lr": 0.001305, "grad_norm": 0.265828, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:29.964082+00:00", "epoch": 0, "step": 1218, "train_loss": 3.9698245525360107, "perplexity": 52.97523565354194, "lr": 0.001305, "grad_norm": 0.267583, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:30.270088+00:00", "epoch": 0, "step": 1219, "train_loss": 3.883507490158081, "perplexity": 48.59436074682727, "lr": 0.001305, "grad_norm": 0.284327, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:30.577244+00:00", "epoch": 0, "step": 1220, "train_loss": 3.9451451301574707, "perplexity": 51.6838384514443, "lr": 0.001305, "grad_norm": 0.28917, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:30.884311+00:00", "epoch": 0, "step": 1221, "train_loss": 3.9090566635131836, "perplexity": 49.85190266651042, "lr": 0.001305, "grad_norm": 0.275057, "tokens_per_sec": 106664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:31.192786+00:00", "epoch": 0, "step": 1222, "train_loss": 3.9352197647094727, "perplexity": 51.17339482724619, "lr": 0.001305, "grad_norm": 0.250585, "tokens_per_sec": 106216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:31.518253+00:00", "epoch": 0, "step": 1223, "train_loss": 3.946906089782715, "perplexity": 51.77493178653768, "lr": 0.001305, "grad_norm": 0.265385, "tokens_per_sec": 100681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:31.825191+00:00", "epoch": 0, "step": 1224, "train_loss": 3.8001115322113037, "perplexity": 44.706170393294315, "lr": 0.001305, "grad_norm": 0.272686, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:32.131404+00:00", "epoch": 0, "step": 1225, "train_loss": 3.8900797367095947, "perplexity": 48.91478667237709, "lr": 0.001305, "grad_norm": 0.25093, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:32.437614+00:00", "epoch": 0, "step": 1226, "train_loss": 4.001016139984131, "perplexity": 54.653657593402045, "lr": 0.001305, "grad_norm": 0.263127, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:32.744899+00:00", "epoch": 0, "step": 1227, "train_loss": 3.9090282917022705, "perplexity": 49.8504882978185, "lr": 0.001305, "grad_norm": 0.265649, "tokens_per_sec": 106638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:33.051802+00:00", "epoch": 0, "step": 1228, "train_loss": 3.959136486053467, "perplexity": 52.4120478684847, "lr": 0.001305, "grad_norm": 0.262274, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:33.358740+00:00", "epoch": 0, "step": 1229, "train_loss": 3.9304120540618896, "perplexity": 50.927958418433164, "lr": 0.001305, "grad_norm": 0.258187, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:33.664969+00:00", "epoch": 0, "step": 1230, "train_loss": 3.91316556930542, "perplexity": 50.05716084260227, "lr": 0.001305, "grad_norm": 0.248139, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:33.971984+00:00", "epoch": 0, "step": 1231, "train_loss": 3.9571750164031982, "perplexity": 52.309343985485235, "lr": 0.001305, "grad_norm": 0.267714, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:34.277972+00:00", "epoch": 0, "step": 1232, "train_loss": 3.9622137546539307, "perplexity": 52.57358223251767, "lr": 0.001305, "grad_norm": 0.339376, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:34.584918+00:00", "epoch": 0, "step": 1233, "train_loss": 3.964684009552002, "perplexity": 52.703612919885565, "lr": 0.001305, "grad_norm": 0.359568, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:34.891245+00:00", "epoch": 0, "step": 1234, "train_loss": 3.831623077392578, "perplexity": 46.1373620046261, "lr": 0.001305, "grad_norm": 0.308097, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:35.198993+00:00", "epoch": 0, "step": 1235, "train_loss": 3.940694808959961, "perplexity": 51.454339819703094, "lr": 0.001305, "grad_norm": 0.305408, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:35.506557+00:00", "epoch": 0, "step": 1236, "train_loss": 3.9757330417633057, "perplexity": 53.28916577580614, "lr": 0.001305, "grad_norm": 0.34442, "tokens_per_sec": 106540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:35.813355+00:00", "epoch": 0, "step": 1237, "train_loss": 3.9291276931762695, "perplexity": 50.86259052762292, "lr": 0.001305, "grad_norm": 0.350573, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:36.119817+00:00", "epoch": 0, "step": 1238, "train_loss": 3.936479091644287, "perplexity": 51.23787945679126, "lr": 0.001305, "grad_norm": 0.33082, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:36.426379+00:00", "epoch": 0, "step": 1239, "train_loss": 3.927149772644043, "perplexity": 50.7620877914741, "lr": 0.001305, "grad_norm": 0.312187, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:36.734143+00:00", "epoch": 0, "step": 1240, "train_loss": 3.8676552772521973, "perplexity": 47.83010616167222, "lr": 0.001305, "grad_norm": 0.343997, "tokens_per_sec": 106527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:37.041200+00:00", "epoch": 0, "step": 1241, "train_loss": 3.868863105773926, "perplexity": 47.887911630611185, "lr": 0.001305, "grad_norm": 0.340413, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:37.348049+00:00", "epoch": 0, "step": 1242, "train_loss": 4.062589168548584, "perplexity": 58.12461083356463, "lr": 0.001305, "grad_norm": 0.320093, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:37.654150+00:00", "epoch": 0, "step": 1243, "train_loss": 3.9786770343780518, "perplexity": 53.446279844147696, "lr": 0.001305, "grad_norm": 0.318905, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:37.960838+00:00", "epoch": 0, "step": 1244, "train_loss": 3.9835684299468994, "perplexity": 53.708347155478556, "lr": 0.001305, "grad_norm": 0.339854, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:38.268419+00:00", "epoch": 0, "step": 1245, "train_loss": 3.9796323776245117, "perplexity": 53.49736378410846, "lr": 0.001305, "grad_norm": 0.300042, "tokens_per_sec": 106536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:38.577085+00:00", "epoch": 0, "step": 1246, "train_loss": 3.9034619331359863, "perplexity": 49.57377346671909, "lr": 0.001305, "grad_norm": 0.30116, "tokens_per_sec": 106163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:38.884295+00:00", "epoch": 0, "step": 1247, "train_loss": 3.9066343307495117, "perplexity": 49.73129090923171, "lr": 0.001305, "grad_norm": 0.304913, "tokens_per_sec": 106660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:39.191773+00:00", "epoch": 0, "step": 1248, "train_loss": 3.924976110458374, "perplexity": 50.65186799447586, "lr": 0.001305, "grad_norm": 0.255979, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:39.498301+00:00", "epoch": 0, "step": 1249, "train_loss": 3.8604137897491455, "perplexity": 47.484996108805255, "lr": 0.001305, "grad_norm": 0.27647, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:39.806538+00:00", "epoch": 0, "step": 1250, "train_loss": 3.887937545776367, "perplexity": 48.810114014304716, "lr": 0.001305, "grad_norm": 0.257133, "tokens_per_sec": 106384} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:35:40.772314+00:00", "step": 1250, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0000_step_0001250.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:41.779335+00:00", "epoch": 0, "step": 1251, "train_loss": 3.8693270683288574, "perplexity": 47.91013498344479, "lr": 0.001305, "grad_norm": 0.273762, "tokens_per_sec": 16608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:42.085633+00:00", "epoch": 0, "step": 1252, "train_loss": 3.947859764099121, "perplexity": 51.82433176121564, "lr": 0.001305, "grad_norm": 0.249174, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:42.391633+00:00", "epoch": 0, "step": 1253, "train_loss": 3.860623598098755, "perplexity": 47.494959902677074, "lr": 0.001305, "grad_norm": 0.244152, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:42.697925+00:00", "epoch": 0, "step": 1254, "train_loss": 3.7706830501556396, "perplexity": 43.409705717841504, "lr": 0.001305, "grad_norm": 0.259546, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:43.004268+00:00", "epoch": 0, "step": 1255, "train_loss": 3.9449281692504883, "perplexity": 51.67262629532131, "lr": 0.001305, "grad_norm": 0.24397, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:43.311546+00:00", "epoch": 0, "step": 1256, "train_loss": 3.853567361831665, "perplexity": 47.16100386610745, "lr": 0.001305, "grad_norm": 0.238756, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:43.618205+00:00", "epoch": 0, "step": 1257, "train_loss": 3.90476655960083, "perplexity": 49.63849093042452, "lr": 0.001305, "grad_norm": 0.235329, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:43.924203+00:00", "epoch": 0, "step": 1258, "train_loss": 3.880124807357788, "perplexity": 48.43025914684589, "lr": 0.001305, "grad_norm": 0.249431, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:44.229708+00:00", "epoch": 0, "step": 1259, "train_loss": 3.8795266151428223, "perplexity": 48.40129720612284, "lr": 0.001305, "grad_norm": 0.245535, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:44.536389+00:00", "epoch": 0, "step": 1260, "train_loss": 3.728729009628296, "perplexity": 41.62616806915372, "lr": 0.001305, "grad_norm": 0.241762, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:44.842834+00:00", "epoch": 0, "step": 1261, "train_loss": 4.010851860046387, "perplexity": 55.19386799395471, "lr": 0.001305, "grad_norm": 0.249507, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:45.150748+00:00", "epoch": 0, "step": 1262, "train_loss": 4.138768672943115, "perplexity": 62.72553822612889, "lr": 0.001305, "grad_norm": 0.238626, "tokens_per_sec": 106419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:45.457788+00:00", "epoch": 0, "step": 1263, "train_loss": 3.8462209701538086, "perplexity": 46.81581017630313, "lr": 0.001305, "grad_norm": 0.244, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:45.763981+00:00", "epoch": 0, "step": 1264, "train_loss": 3.9243483543395996, "perplexity": 50.620080952713835, "lr": 0.001305, "grad_norm": 0.252417, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:46.091760+00:00", "epoch": 0, "step": 1265, "train_loss": 3.823911666870117, "perplexity": 45.782946145180645, "lr": 0.001305, "grad_norm": 0.254337, "tokens_per_sec": 99969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:46.397761+00:00", "epoch": 0, "step": 1266, "train_loss": 3.8670263290405273, "perplexity": 47.80003296017236, "lr": 0.001305, "grad_norm": 0.265109, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:46.704899+00:00", "epoch": 0, "step": 1267, "train_loss": 3.9439356327056885, "perplexity": 51.62136476903478, "lr": 0.001305, "grad_norm": 0.273634, "tokens_per_sec": 106688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:47.012910+00:00", "epoch": 0, "step": 1268, "train_loss": 3.9602766036987305, "perplexity": 52.471837846411795, "lr": 0.001305, "grad_norm": 0.285719, "tokens_per_sec": 106387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:47.320553+00:00", "epoch": 0, "step": 1269, "train_loss": 3.8965816497802734, "perplexity": 49.23386254089818, "lr": 0.001305, "grad_norm": 0.297772, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:47.627752+00:00", "epoch": 0, "step": 1270, "train_loss": 3.896611213684082, "perplexity": 49.23531810759047, "lr": 0.001305, "grad_norm": 0.316833, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:47.933932+00:00", "epoch": 0, "step": 1271, "train_loss": 3.9925131797790527, "perplexity": 54.190909868632225, "lr": 0.001305, "grad_norm": 0.308222, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:48.240605+00:00", "epoch": 0, "step": 1272, "train_loss": 3.9757606983184814, "perplexity": 53.29063959093992, "lr": 0.001305, "grad_norm": 0.26926, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:48.547577+00:00", "epoch": 0, "step": 1273, "train_loss": 3.91402006149292, "perplexity": 50.0999525754677, "lr": 0.001305, "grad_norm": 0.291698, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:48.854998+00:00", "epoch": 0, "step": 1274, "train_loss": 4.003490924835205, "perplexity": 54.7890811402236, "lr": 0.001305, "grad_norm": 0.282807, "tokens_per_sec": 106587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:49.162489+00:00", "epoch": 0, "step": 1275, "train_loss": 3.876361846923828, "perplexity": 48.248360451305494, "lr": 0.001305, "grad_norm": 0.256088, "tokens_per_sec": 106565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:49.469703+00:00", "epoch": 0, "step": 1276, "train_loss": 3.8643784523010254, "perplexity": 47.67363178596673, "lr": 0.001305, "grad_norm": 0.250636, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:49.775720+00:00", "epoch": 0, "step": 1277, "train_loss": 3.997687578201294, "perplexity": 54.47204194457982, "lr": 0.001305, "grad_norm": 0.245939, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:50.082549+00:00", "epoch": 0, "step": 1278, "train_loss": 3.981865167617798, "perplexity": 53.61694561350905, "lr": 0.001305, "grad_norm": 0.238029, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:50.389453+00:00", "epoch": 0, "step": 1279, "train_loss": 3.9471583366394043, "perplexity": 51.787993497654675, "lr": 0.001305, "grad_norm": 0.254934, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:50.696734+00:00", "epoch": 0, "step": 1280, "train_loss": 3.858436107635498, "perplexity": 47.39117868243572, "lr": 0.001305, "grad_norm": 0.234388, "tokens_per_sec": 106697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:51.003802+00:00", "epoch": 0, "step": 1281, "train_loss": 3.855609178543091, "perplexity": 47.257396366349326, "lr": 0.001305, "grad_norm": 0.272182, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:51.310546+00:00", "epoch": 0, "step": 1282, "train_loss": 3.855994939804077, "perplexity": 47.27562995584266, "lr": 0.001305, "grad_norm": 0.305904, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:51.616272+00:00", "epoch": 0, "step": 1283, "train_loss": 4.066611289978027, "perplexity": 58.358865861656376, "lr": 0.001305, "grad_norm": 0.286783, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:51.922953+00:00", "epoch": 0, "step": 1284, "train_loss": 3.7771947383880615, "perplexity": 43.69329851961589, "lr": 0.001305, "grad_norm": 0.25296, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:52.243194+00:00", "epoch": 0, "step": 1285, "train_loss": 3.9525017738342285, "perplexity": 52.06546004068603, "lr": 0.001305, "grad_norm": 0.271934, "tokens_per_sec": 102323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:52.551077+00:00", "epoch": 0, "step": 1286, "train_loss": 3.9340591430664062, "perplexity": 51.1140363307016, "lr": 0.001305, "grad_norm": 0.268311, "tokens_per_sec": 106433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:52.858172+00:00", "epoch": 0, "step": 1287, "train_loss": 3.890934944152832, "perplexity": 48.956636854766714, "lr": 0.001305, "grad_norm": 0.286173, "tokens_per_sec": 106700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:53.164344+00:00", "epoch": 0, "step": 1288, "train_loss": 3.845081090927124, "perplexity": 46.762476209717896, "lr": 0.001305, "grad_norm": 0.313818, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:53.482922+00:00", "epoch": 0, "step": 1289, "train_loss": 3.947904586791992, "perplexity": 51.82665471938166, "lr": 0.001305, "grad_norm": 0.277549, "tokens_per_sec": 102857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:53.789658+00:00", "epoch": 0, "step": 1290, "train_loss": 3.9525132179260254, "perplexity": 52.066055885999624, "lr": 0.001305, "grad_norm": 0.268737, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:54.110450+00:00", "epoch": 0, "step": 1291, "train_loss": 3.923760414123535, "perplexity": 50.590328118682194, "lr": 0.001305, "grad_norm": 0.293354, "tokens_per_sec": 102096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:54.417905+00:00", "epoch": 0, "step": 1292, "train_loss": 3.9508461952209473, "perplexity": 51.97933289335604, "lr": 0.001305, "grad_norm": 0.274475, "tokens_per_sec": 106575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:54.738733+00:00", "epoch": 0, "step": 1293, "train_loss": 3.970095157623291, "perplexity": 52.989572961596416, "lr": 0.001305, "grad_norm": 0.248647, "tokens_per_sec": 102134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:55.047094+00:00", "epoch": 0, "step": 1294, "train_loss": 3.8198108673095703, "perplexity": 45.595583889840725, "lr": 0.001305, "grad_norm": 0.256292, "tokens_per_sec": 106269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:55.363578+00:00", "epoch": 0, "step": 1295, "train_loss": 3.8761651515960693, "perplexity": 48.238871157513174, "lr": 0.001305, "grad_norm": 0.326071, "tokens_per_sec": 103534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:55.670147+00:00", "epoch": 0, "step": 1296, "train_loss": 3.9741311073303223, "perplexity": 53.20386836491549, "lr": 0.001305, "grad_norm": 0.334326, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:55.991312+00:00", "epoch": 0, "step": 1297, "train_loss": 3.942168951034546, "perplexity": 51.53024676201394, "lr": 0.001305, "grad_norm": 0.302681, "tokens_per_sec": 102028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:56.312271+00:00", "epoch": 0, "step": 1298, "train_loss": 3.8802759647369385, "perplexity": 48.437580291198664, "lr": 0.001305, "grad_norm": 0.325617, "tokens_per_sec": 102095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:56.618832+00:00", "epoch": 0, "step": 1299, "train_loss": 3.8119704723358154, "perplexity": 45.23949426884317, "lr": 0.001305, "grad_norm": 0.302453, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:35:56.941916+00:00", "epoch": 0, "step": 1300, "train_loss": 3.8433730602264404, "perplexity": 46.68267263758869, "lr": 0.001305, "grad_norm": 0.280696, "tokens_per_sec": 101485} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:36:02.579602+00:00", "step": 1300, "epoch": 0, "val_loss": 3.9362255096435548, "val_ppl": 51.22488810006206, "eval_train_loss": 3.8433730602264404, "eval_train_ppl": 46.68267263758869} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:36:03.669376+00:00", "step": 1300, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p9362_epoch_0000_step_0001300.pt", "val_loss": 3.9362255096435548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:04.881158+00:00", "epoch": 0, "step": 1301, "train_loss": 3.72271728515625, "perplexity": 41.376673712968916, "lr": 0.001305, "grad_norm": 0.290055, "tokens_per_sec": 4127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:05.186851+00:00", "epoch": 0, "step": 1302, "train_loss": 3.93682599067688, "perplexity": 51.25565691091948, "lr": 0.001305, "grad_norm": 0.30292, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:05.493151+00:00", "epoch": 0, "step": 1303, "train_loss": 3.9307126998901367, "perplexity": 50.943271998539444, "lr": 0.001305, "grad_norm": 0.276923, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:05.800279+00:00", "epoch": 0, "step": 1304, "train_loss": 3.960129737854004, "perplexity": 52.464132091492075, "lr": 0.001305, "grad_norm": 0.304098, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:06.107281+00:00", "epoch": 0, "step": 1305, "train_loss": 4.041742324829102, "perplexity": 56.92543905734725, "lr": 0.001305, "grad_norm": 0.317609, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:06.413943+00:00", "epoch": 0, "step": 1306, "train_loss": 3.9265432357788086, "perplexity": 50.73130804934008, "lr": 0.001305, "grad_norm": 0.32909, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:06.719471+00:00", "epoch": 0, "step": 1307, "train_loss": 3.9364211559295654, "perplexity": 51.23491103961363, "lr": 0.001305, "grad_norm": 0.306942, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:07.025483+00:00", "epoch": 0, "step": 1308, "train_loss": 3.8743958473205566, "perplexity": 48.15359737640638, "lr": 0.001305, "grad_norm": 0.301868, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:07.332918+00:00", "epoch": 0, "step": 1309, "train_loss": 3.908189296722412, "perplexity": 49.80868152868022, "lr": 0.001305, "grad_norm": 0.303077, "tokens_per_sec": 106586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:07.639685+00:00", "epoch": 0, "step": 1310, "train_loss": 3.865480899810791, "perplexity": 47.7262184442998, "lr": 0.001305, "grad_norm": 0.279521, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:07.945852+00:00", "epoch": 0, "step": 1311, "train_loss": 3.9215550422668457, "perplexity": 50.47888056964294, "lr": 0.001305, "grad_norm": 0.29308, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:08.252325+00:00", "epoch": 0, "step": 1312, "train_loss": 3.8238871097564697, "perplexity": 45.781821861973675, "lr": 0.001305, "grad_norm": 0.260343, "tokens_per_sec": 106921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:08.557491+00:00", "epoch": 0, "step": 1313, "train_loss": 3.8962275981903076, "perplexity": 49.216434299015845, "lr": 0.001305, "grad_norm": 0.265965, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:08.863760+00:00", "epoch": 0, "step": 1314, "train_loss": 3.9230146408081055, "perplexity": 50.55261326706543, "lr": 0.001305, "grad_norm": 0.255958, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:09.170541+00:00", "epoch": 0, "step": 1315, "train_loss": 3.852940797805786, "perplexity": 47.13146373301938, "lr": 0.001305, "grad_norm": 0.272944, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:09.477941+00:00", "epoch": 0, "step": 1316, "train_loss": 3.9763665199279785, "perplexity": 53.32293399332372, "lr": 0.001305, "grad_norm": 0.252854, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:09.784285+00:00", "epoch": 0, "step": 1317, "train_loss": 3.88714861869812, "perplexity": 48.771621579525046, "lr": 0.001305, "grad_norm": 0.236704, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:10.089833+00:00", "epoch": 0, "step": 1318, "train_loss": 3.9265339374542236, "perplexity": 50.7308363353643, "lr": 0.001305, "grad_norm": 0.240416, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:10.395839+00:00", "epoch": 0, "step": 1319, "train_loss": 3.821422815322876, "perplexity": 45.66914086979393, "lr": 0.001305, "grad_norm": 0.235821, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:10.703086+00:00", "epoch": 0, "step": 1320, "train_loss": 3.9084270000457764, "perplexity": 49.820522625090206, "lr": 0.001305, "grad_norm": 0.248167, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:11.009594+00:00", "epoch": 0, "step": 1321, "train_loss": 3.8572452068328857, "perplexity": 47.33477408251402, "lr": 0.001305, "grad_norm": 0.233625, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:11.315794+00:00", "epoch": 0, "step": 1322, "train_loss": 3.947585105895996, "perplexity": 51.81009973793639, "lr": 0.001305, "grad_norm": 0.242321, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:11.621686+00:00", "epoch": 0, "step": 1323, "train_loss": 3.9489662647247314, "perplexity": 51.88170715383124, "lr": 0.001305, "grad_norm": 0.276199, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:11.927630+00:00", "epoch": 0, "step": 1324, "train_loss": 3.8679492473602295, "perplexity": 47.844168850051496, "lr": 0.001305, "grad_norm": 0.239262, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:12.233812+00:00", "epoch": 0, "step": 1325, "train_loss": 3.905439615249634, "perplexity": 49.67191164288134, "lr": 0.001305, "grad_norm": 0.260064, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:12.541319+00:00", "epoch": 0, "step": 1326, "train_loss": 3.949070930480957, "perplexity": 51.88713767613468, "lr": 0.001305, "grad_norm": 0.243782, "tokens_per_sec": 106561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:12.848856+00:00", "epoch": 0, "step": 1327, "train_loss": 3.921198606491089, "perplexity": 50.46089129688841, "lr": 0.001305, "grad_norm": 0.257756, "tokens_per_sec": 106549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:13.155300+00:00", "epoch": 0, "step": 1328, "train_loss": 3.9521586894989014, "perplexity": 52.047600260815, "lr": 0.001305, "grad_norm": 0.270546, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:13.460670+00:00", "epoch": 0, "step": 1329, "train_loss": 3.8012239933013916, "perplexity": 44.755931942105164, "lr": 0.001305, "grad_norm": 0.271361, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:13.777108+00:00", "epoch": 0, "step": 1330, "train_loss": 3.8099639415740967, "perplexity": 45.14881084191653, "lr": 0.001305, "grad_norm": 0.26678, "tokens_per_sec": 103609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:14.083743+00:00", "epoch": 0, "step": 1331, "train_loss": 3.942830801010132, "perplexity": 51.56436334335919, "lr": 0.001305, "grad_norm": 0.301685, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:14.390383+00:00", "epoch": 0, "step": 1332, "train_loss": 3.9535109996795654, "perplexity": 52.11803237282438, "lr": 0.001305, "grad_norm": 0.265581, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:14.708330+00:00", "epoch": 0, "step": 1333, "train_loss": 3.880441188812256, "perplexity": 48.44558400679794, "lr": 0.001305, "grad_norm": 0.254186, "tokens_per_sec": 103061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:15.014845+00:00", "epoch": 0, "step": 1334, "train_loss": 4.015727996826172, "perplexity": 55.46365807580929, "lr": 0.001305, "grad_norm": 0.267101, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:15.320003+00:00", "epoch": 0, "step": 1335, "train_loss": 3.7460129261016846, "perplexity": 42.3518848288452, "lr": 0.001305, "grad_norm": 0.270026, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:15.625697+00:00", "epoch": 0, "step": 1336, "train_loss": 3.920698881149292, "perplexity": 50.43568101037186, "lr": 0.001305, "grad_norm": 0.295778, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:15.945472+00:00", "epoch": 0, "step": 1337, "train_loss": 3.986670732498169, "perplexity": 53.87522541747218, "lr": 0.001305, "grad_norm": 0.324857, "tokens_per_sec": 102471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:16.253048+00:00", "epoch": 0, "step": 1338, "train_loss": 3.9575607776641846, "perplexity": 52.32952679660549, "lr": 0.001305, "grad_norm": 0.311671, "tokens_per_sec": 106537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:16.559205+00:00", "epoch": 0, "step": 1339, "train_loss": 3.931994915008545, "perplexity": 51.00863412727695, "lr": 0.001305, "grad_norm": 0.268852, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:16.875842+00:00", "epoch": 0, "step": 1340, "train_loss": 3.94567608833313, "perplexity": 51.71128769457471, "lr": 0.001305, "grad_norm": 0.266576, "tokens_per_sec": 103544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:17.181413+00:00", "epoch": 0, "step": 1341, "train_loss": 3.8698103427886963, "perplexity": 47.93329432375781, "lr": 0.001305, "grad_norm": 0.25728, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:17.487671+00:00", "epoch": 0, "step": 1342, "train_loss": 3.87119197845459, "perplexity": 47.99956644419959, "lr": 0.001305, "grad_norm": 0.284246, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:17.809320+00:00", "epoch": 0, "step": 1343, "train_loss": 3.8634915351867676, "perplexity": 47.63136797105778, "lr": 0.001305, "grad_norm": 0.287245, "tokens_per_sec": 101875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:18.115873+00:00", "epoch": 0, "step": 1344, "train_loss": 3.917487621307373, "perplexity": 50.27397870635974, "lr": 0.001305, "grad_norm": 0.283743, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:18.423448+00:00", "epoch": 0, "step": 1345, "train_loss": 3.903381109237671, "perplexity": 49.569766883009336, "lr": 0.001305, "grad_norm": 0.25315, "tokens_per_sec": 106537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:18.745324+00:00", "epoch": 0, "step": 1346, "train_loss": 3.9375290870666504, "perplexity": 51.29170725019545, "lr": 0.001305, "grad_norm": 0.269188, "tokens_per_sec": 101803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:19.051958+00:00", "epoch": 0, "step": 1347, "train_loss": 3.7931365966796875, "perplexity": 44.395432684453176, "lr": 0.001305, "grad_norm": 0.235802, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:19.358254+00:00", "epoch": 0, "step": 1348, "train_loss": 3.8726704120635986, "perplexity": 48.07058310021859, "lr": 0.001305, "grad_norm": 0.238587, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:19.663634+00:00", "epoch": 0, "step": 1349, "train_loss": 3.9228856563568115, "perplexity": 50.54609318648525, "lr": 0.001305, "grad_norm": 0.254084, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:19.982552+00:00", "epoch": 0, "step": 1350, "train_loss": 3.818701982498169, "perplexity": 45.545051661786374, "lr": 0.001305, "grad_norm": 0.276037, "tokens_per_sec": 102798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:20.288622+00:00", "epoch": 0, "step": 1351, "train_loss": 3.7926506996154785, "perplexity": 44.37386631398987, "lr": 0.001305, "grad_norm": 0.296133, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:20.605102+00:00", "epoch": 0, "step": 1352, "train_loss": 3.9336354732513428, "perplexity": 51.09238544311985, "lr": 0.001305, "grad_norm": 0.258371, "tokens_per_sec": 103539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:20.912089+00:00", "epoch": 0, "step": 1353, "train_loss": 3.9176342487335205, "perplexity": 50.28135079092133, "lr": 0.001305, "grad_norm": 0.284786, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:21.228999+00:00", "epoch": 0, "step": 1354, "train_loss": 3.9048118591308594, "perplexity": 49.64073958166607, "lr": 0.001305, "grad_norm": 0.254384, "tokens_per_sec": 103399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:21.534208+00:00", "epoch": 0, "step": 1355, "train_loss": 3.843350410461426, "perplexity": 46.68161529799748, "lr": 0.001305, "grad_norm": 0.241488, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:21.853668+00:00", "epoch": 0, "step": 1356, "train_loss": 3.9504306316375732, "perplexity": 51.95773666313168, "lr": 0.001305, "grad_norm": 0.24996, "tokens_per_sec": 102573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:22.160631+00:00", "epoch": 0, "step": 1357, "train_loss": 3.9211957454681396, "perplexity": 50.460746927326895, "lr": 0.001305, "grad_norm": 0.260236, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:22.480008+00:00", "epoch": 0, "step": 1358, "train_loss": 3.910900115966797, "perplexity": 49.94388703715583, "lr": 0.001305, "grad_norm": 0.261949, "tokens_per_sec": 102599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:22.787164+00:00", "epoch": 0, "step": 1359, "train_loss": 3.9895009994506836, "perplexity": 54.027922672597356, "lr": 0.001305, "grad_norm": 0.240837, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:23.103412+00:00", "epoch": 0, "step": 1360, "train_loss": 3.892160654067993, "perplexity": 49.01668028036634, "lr": 0.001305, "grad_norm": 0.316342, "tokens_per_sec": 103665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:23.409709+00:00", "epoch": 0, "step": 1361, "train_loss": 3.8704640865325928, "perplexity": 47.96464066016558, "lr": 0.001305, "grad_norm": 0.320143, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:23.727198+00:00", "epoch": 0, "step": 1362, "train_loss": 3.8146893978118896, "perplexity": 45.362664451730986, "lr": 0.001305, "grad_norm": 0.326118, "tokens_per_sec": 103210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:24.033461+00:00", "epoch": 0, "step": 1363, "train_loss": 3.8392229080200195, "perplexity": 46.489333910813556, "lr": 0.001305, "grad_norm": 0.336435, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:24.340493+00:00", "epoch": 0, "step": 1364, "train_loss": 3.730738639831543, "perplexity": 41.70990538608967, "lr": 0.001305, "grad_norm": 0.308475, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:24.660189+00:00", "epoch": 0, "step": 1365, "train_loss": 3.8870012760162354, "perplexity": 48.76443596738834, "lr": 0.001305, "grad_norm": 0.314695, "tokens_per_sec": 102498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:24.966668+00:00", "epoch": 0, "step": 1366, "train_loss": 3.932180166244507, "perplexity": 51.018084415105484, "lr": 0.001305, "grad_norm": 0.311149, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:25.272609+00:00", "epoch": 0, "step": 1367, "train_loss": 3.840104818344116, "perplexity": 46.53035141857316, "lr": 0.001305, "grad_norm": 0.28654, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:25.588455+00:00", "epoch": 0, "step": 1368, "train_loss": 3.8693671226501465, "perplexity": 47.91205402981721, "lr": 0.001305, "grad_norm": 0.285683, "tokens_per_sec": 103747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:25.894042+00:00", "epoch": 0, "step": 1369, "train_loss": 3.897132158279419, "perplexity": 49.260973662439426, "lr": 0.001305, "grad_norm": 0.294089, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:26.214188+00:00", "epoch": 0, "step": 1370, "train_loss": 3.975036144256592, "perplexity": 53.25204162640819, "lr": 0.001305, "grad_norm": 0.307412, "tokens_per_sec": 102403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:26.520475+00:00", "epoch": 0, "step": 1371, "train_loss": 3.86896014213562, "perplexity": 47.892558724789865, "lr": 0.001305, "grad_norm": 0.303884, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:26.842941+00:00", "epoch": 0, "step": 1372, "train_loss": 3.8902127742767334, "perplexity": 48.92129460948355, "lr": 0.001305, "grad_norm": 0.284138, "tokens_per_sec": 101617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:27.150497+00:00", "epoch": 0, "step": 1373, "train_loss": 3.8505747318267822, "perplexity": 47.0200794034397, "lr": 0.001305, "grad_norm": 0.278907, "tokens_per_sec": 106544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:27.458077+00:00", "epoch": 0, "step": 1374, "train_loss": 3.849719762802124, "perplexity": 46.97989587229776, "lr": 0.001305, "grad_norm": 0.261526, "tokens_per_sec": 106535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:27.775129+00:00", "epoch": 0, "step": 1375, "train_loss": 3.9140429496765137, "perplexity": 50.10109928550329, "lr": 0.001305, "grad_norm": 0.25772, "tokens_per_sec": 103352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:28.081253+00:00", "epoch": 0, "step": 1376, "train_loss": 3.8965327739715576, "perplexity": 49.23145625485535, "lr": 0.001305, "grad_norm": 0.241725, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:28.399344+00:00", "epoch": 0, "step": 1377, "train_loss": 3.910717248916626, "perplexity": 49.93475478087911, "lr": 0.001305, "grad_norm": 0.244756, "tokens_per_sec": 103015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:28.706143+00:00", "epoch": 0, "step": 1378, "train_loss": 3.875877618789673, "perplexity": 48.22500289339786, "lr": 0.001305, "grad_norm": 0.243354, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:29.025238+00:00", "epoch": 0, "step": 1379, "train_loss": 3.9118196964263916, "perplexity": 49.989835583206, "lr": 0.001305, "grad_norm": 0.244395, "tokens_per_sec": 102690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:29.332354+00:00", "epoch": 0, "step": 1380, "train_loss": 3.7765755653381348, "perplexity": 43.66625318044675, "lr": 0.001305, "grad_norm": 0.238239, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:29.649392+00:00", "epoch": 0, "step": 1381, "train_loss": 3.8093149662017822, "perplexity": 45.11951988117496, "lr": 0.001305, "grad_norm": 0.250095, "tokens_per_sec": 103302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:29.956162+00:00", "epoch": 0, "step": 1382, "train_loss": 3.916593074798584, "perplexity": 50.229026403185586, "lr": 0.001305, "grad_norm": 0.238031, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:30.276024+00:00", "epoch": 0, "step": 1383, "train_loss": 3.989866018295288, "perplexity": 54.047647482252295, "lr": 0.001305, "grad_norm": 0.245414, "tokens_per_sec": 102443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:30.582285+00:00", "epoch": 0, "step": 1384, "train_loss": 3.8971521854400635, "perplexity": 49.26196022975151, "lr": 0.001305, "grad_norm": 0.251156, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:30.900253+00:00", "epoch": 0, "step": 1385, "train_loss": 3.7945568561553955, "perplexity": 44.4585305154412, "lr": 0.001305, "grad_norm": 0.259202, "tokens_per_sec": 103055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:31.206814+00:00", "epoch": 0, "step": 1386, "train_loss": 3.899379253387451, "perplexity": 49.37179221866005, "lr": 0.001305, "grad_norm": 0.285402, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:31.523104+00:00", "epoch": 0, "step": 1387, "train_loss": 3.9235751628875732, "perplexity": 50.58095706589685, "lr": 0.001305, "grad_norm": 0.280231, "tokens_per_sec": 103602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:31.841656+00:00", "epoch": 0, "step": 1388, "train_loss": 3.8550665378570557, "perplexity": 47.23175953678733, "lr": 0.001305, "grad_norm": 0.2589, "tokens_per_sec": 102865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:32.161715+00:00", "epoch": 0, "step": 1389, "train_loss": 3.860431432723999, "perplexity": 47.485833892788, "lr": 0.001305, "grad_norm": 0.294649, "tokens_per_sec": 102381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:32.468223+00:00", "epoch": 0, "step": 1390, "train_loss": 4.023011207580566, "perplexity": 55.86908620387126, "lr": 0.001305, "grad_norm": 0.286809, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:32.791197+00:00", "epoch": 0, "step": 1391, "train_loss": 3.9114863872528076, "perplexity": 49.973176288922204, "lr": 0.001305, "grad_norm": 0.287066, "tokens_per_sec": 101399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:33.098440+00:00", "epoch": 0, "step": 1392, "train_loss": 3.908923625946045, "perplexity": 49.845270931807136, "lr": 0.001305, "grad_norm": 0.278924, "tokens_per_sec": 106651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:33.405013+00:00", "epoch": 0, "step": 1393, "train_loss": 3.8529069423675537, "perplexity": 47.12986810367068, "lr": 0.001305, "grad_norm": 0.284324, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:33.721873+00:00", "epoch": 0, "step": 1394, "train_loss": 3.942661762237549, "perplexity": 51.55564770333191, "lr": 0.001305, "grad_norm": 0.276876, "tokens_per_sec": 103415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:34.042569+00:00", "epoch": 0, "step": 1395, "train_loss": 3.768404483795166, "perplexity": 43.31090642579246, "lr": 0.001305, "grad_norm": 0.256817, "tokens_per_sec": 102177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:34.348415+00:00", "epoch": 0, "step": 1396, "train_loss": 3.863560438156128, "perplexity": 47.634650026816075, "lr": 0.001305, "grad_norm": 0.236151, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:34.669087+00:00", "epoch": 0, "step": 1397, "train_loss": 3.918509006500244, "perplexity": 50.32535403633144, "lr": 0.001305, "grad_norm": 0.24508, "tokens_per_sec": 102186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:34.975800+00:00", "epoch": 0, "step": 1398, "train_loss": 3.828533887863159, "perplexity": 45.99505486903383, "lr": 0.001305, "grad_norm": 0.253335, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:35.282951+00:00", "epoch": 0, "step": 1399, "train_loss": 3.8210573196411133, "perplexity": 45.652452046048495, "lr": 0.001305, "grad_norm": 0.268626, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:35.603574+00:00", "epoch": 0, "step": 1400, "train_loss": 3.755174160003662, "perplexity": 42.74166305056964, "lr": 0.001305, "grad_norm": 0.287748, "tokens_per_sec": 102256} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:36:41.189000+00:00", "step": 1400, "epoch": 0, "val_loss": 3.9075106263160704, "val_ppl": 49.77488931873127, "eval_train_loss": 3.755174160003662, "eval_train_ppl": 42.74166305056964} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:36:42.170729+00:00", "step": 1400, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p9075_epoch_0000_step_0001400.pt", "val_loss": 3.9075106263160704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:43.435599+00:00", "epoch": 0, "step": 1401, "train_loss": 3.9519712924957275, "perplexity": 52.03784761034102, "lr": 0.001305, "grad_norm": 0.295152, "tokens_per_sec": 4184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:43.740975+00:00", "epoch": 0, "step": 1402, "train_loss": 3.921196699142456, "perplexity": 50.46079505046817, "lr": 0.001305, "grad_norm": 0.300381, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:44.046405+00:00", "epoch": 0, "step": 1403, "train_loss": 3.8224472999572754, "perplexity": 45.71595217751968, "lr": 0.001305, "grad_norm": 0.274682, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:44.350757+00:00", "epoch": 0, "step": 1404, "train_loss": 4.020685195922852, "perplexity": 55.739285076045675, "lr": 0.001305, "grad_norm": 0.263116, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:44.655471+00:00", "epoch": 0, "step": 1405, "train_loss": 3.8881404399871826, "perplexity": 48.82001830859547, "lr": 0.001305, "grad_norm": 0.295405, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:44.961394+00:00", "epoch": 0, "step": 1406, "train_loss": 3.923032522201538, "perplexity": 50.553517226314305, "lr": 0.001305, "grad_norm": 0.302336, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:45.267451+00:00", "epoch": 0, "step": 1407, "train_loss": 3.792914867401123, "perplexity": 44.385590008437546, "lr": 0.001305, "grad_norm": 0.298661, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:45.574048+00:00", "epoch": 0, "step": 1408, "train_loss": 3.9364917278289795, "perplexity": 51.238526912190004, "lr": 0.001305, "grad_norm": 0.270878, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:45.880121+00:00", "epoch": 0, "step": 1409, "train_loss": 3.8927345275878906, "perplexity": 49.04481772810839, "lr": 0.001305, "grad_norm": 0.275579, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:46.185952+00:00", "epoch": 0, "step": 1410, "train_loss": 3.7789816856384277, "perplexity": 43.77144594113005, "lr": 0.001305, "grad_norm": 0.259714, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:46.491640+00:00", "epoch": 0, "step": 1411, "train_loss": 3.8069393634796143, "perplexity": 45.01246104190609, "lr": 0.001305, "grad_norm": 0.257133, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:46.798121+00:00", "epoch": 0, "step": 1412, "train_loss": 3.794919013977051, "perplexity": 44.47463443591103, "lr": 0.001305, "grad_norm": 0.277342, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:47.104684+00:00", "epoch": 0, "step": 1413, "train_loss": 3.86970591545105, "perplexity": 47.928289038795796, "lr": 0.001305, "grad_norm": 0.266377, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:47.410827+00:00", "epoch": 0, "step": 1414, "train_loss": 3.8012588024139404, "perplexity": 44.75748988349248, "lr": 0.001305, "grad_norm": 0.265874, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:47.716849+00:00", "epoch": 0, "step": 1415, "train_loss": 3.9639079570770264, "perplexity": 52.66272801710586, "lr": 0.001305, "grad_norm": 0.263606, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:48.022494+00:00", "epoch": 0, "step": 1416, "train_loss": 3.935537576675415, "perplexity": 51.18966092910418, "lr": 0.001305, "grad_norm": 0.258644, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:48.328009+00:00", "epoch": 0, "step": 1417, "train_loss": 3.8950681686401367, "perplexity": 49.159404378211896, "lr": 0.001305, "grad_norm": 0.273603, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:48.634947+00:00", "epoch": 0, "step": 1418, "train_loss": 3.9085586071014404, "perplexity": 49.827079788859585, "lr": 0.001305, "grad_norm": 0.277799, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:48.941200+00:00", "epoch": 0, "step": 1419, "train_loss": 3.789355993270874, "perplexity": 44.227908031981926, "lr": 0.001305, "grad_norm": 0.270356, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:49.247366+00:00", "epoch": 0, "step": 1420, "train_loss": 3.9174532890319824, "perplexity": 50.27225271590657, "lr": 0.001305, "grad_norm": 0.281074, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:49.552895+00:00", "epoch": 0, "step": 1421, "train_loss": 3.8003222942352295, "perplexity": 44.71559374925628, "lr": 0.001305, "grad_norm": 0.281036, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:49.859263+00:00", "epoch": 0, "step": 1422, "train_loss": 3.9198241233825684, "perplexity": 50.39158129778318, "lr": 0.001305, "grad_norm": 0.274135, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:50.166826+00:00", "epoch": 0, "step": 1423, "train_loss": 3.83779239654541, "perplexity": 46.42287792956182, "lr": 0.001305, "grad_norm": 0.305841, "tokens_per_sec": 106541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:50.473909+00:00", "epoch": 0, "step": 1424, "train_loss": 3.8609137535095215, "perplexity": 47.50874282177456, "lr": 0.001305, "grad_norm": 0.292954, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:50.780224+00:00", "epoch": 0, "step": 1425, "train_loss": 3.8374063968658447, "perplexity": 46.40496217151738, "lr": 0.001305, "grad_norm": 0.258395, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:51.086180+00:00", "epoch": 0, "step": 1426, "train_loss": 3.835191011428833, "perplexity": 46.30227108628496, "lr": 0.001305, "grad_norm": 0.268291, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:51.391913+00:00", "epoch": 0, "step": 1427, "train_loss": 3.9089531898498535, "perplexity": 49.84674457438548, "lr": 0.001305, "grad_norm": 0.285534, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:51.698010+00:00", "epoch": 0, "step": 1428, "train_loss": 3.8875577449798584, "perplexity": 48.79157941407516, "lr": 0.001305, "grad_norm": 0.257202, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:52.006084+00:00", "epoch": 0, "step": 1429, "train_loss": 3.792630910873413, "perplexity": 44.37298821968316, "lr": 0.001305, "grad_norm": 0.279764, "tokens_per_sec": 106363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:52.313922+00:00", "epoch": 0, "step": 1430, "train_loss": 3.865891218185425, "perplexity": 47.7458054068498, "lr": 0.001305, "grad_norm": 0.287319, "tokens_per_sec": 106498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:52.619670+00:00", "epoch": 0, "step": 1431, "train_loss": 3.84405779838562, "perplexity": 46.71464899138157, "lr": 0.001305, "grad_norm": 0.278676, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:52.924830+00:00", "epoch": 0, "step": 1432, "train_loss": 3.7966361045837402, "perplexity": 44.55106701498918, "lr": 0.001305, "grad_norm": 0.262125, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:53.230201+00:00", "epoch": 0, "step": 1433, "train_loss": 3.9011433124542236, "perplexity": 49.45896384170637, "lr": 0.001305, "grad_norm": 0.286641, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:53.536454+00:00", "epoch": 0, "step": 1434, "train_loss": 3.805388927459717, "perplexity": 44.94272617465686, "lr": 0.001305, "grad_norm": 0.303767, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:53.842714+00:00", "epoch": 0, "step": 1435, "train_loss": 3.9136245250701904, "perplexity": 50.080140137975704, "lr": 0.001305, "grad_norm": 0.303093, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:54.149102+00:00", "epoch": 0, "step": 1436, "train_loss": 3.889503240585327, "perplexity": 48.88659561423974, "lr": 0.001305, "grad_norm": 0.30382, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:54.454589+00:00", "epoch": 0, "step": 1437, "train_loss": 3.874319076538086, "perplexity": 48.14990072895609, "lr": 0.001305, "grad_norm": 0.325577, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:54.760072+00:00", "epoch": 0, "step": 1438, "train_loss": 3.820613384246826, "perplexity": 45.632189804645094, "lr": 0.001305, "grad_norm": 0.332472, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:55.066308+00:00", "epoch": 0, "step": 1439, "train_loss": 3.856217384338379, "perplexity": 47.286147331054956, "lr": 0.001305, "grad_norm": 0.326118, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:55.372735+00:00", "epoch": 0, "step": 1440, "train_loss": 3.8926053047180176, "perplexity": 49.0384804254802, "lr": 0.001305, "grad_norm": 0.299729, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:55.678531+00:00", "epoch": 0, "step": 1441, "train_loss": 3.9043796062469482, "perplexity": 49.61928686565562, "lr": 0.001305, "grad_norm": 0.268886, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:55.984388+00:00", "epoch": 0, "step": 1442, "train_loss": 3.8085596561431885, "perplexity": 45.08545352092162, "lr": 0.001305, "grad_norm": 0.300211, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:56.289762+00:00", "epoch": 0, "step": 1443, "train_loss": 3.8513870239257812, "perplexity": 47.05828895899252, "lr": 0.001305, "grad_norm": 0.266369, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:56.595776+00:00", "epoch": 0, "step": 1444, "train_loss": 3.7386152744293213, "perplexity": 42.039736343627574, "lr": 0.001305, "grad_norm": 0.263446, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:56.901508+00:00", "epoch": 0, "step": 1445, "train_loss": 3.8102409839630127, "perplexity": 45.16132070913063, "lr": 0.001305, "grad_norm": 0.275632, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:57.208617+00:00", "epoch": 0, "step": 1446, "train_loss": 3.899648666381836, "perplexity": 49.38509541298606, "lr": 0.001305, "grad_norm": 0.29098, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:57.515583+00:00", "epoch": 0, "step": 1447, "train_loss": 3.8758749961853027, "perplexity": 48.22487641846036, "lr": 0.001305, "grad_norm": 0.257635, "tokens_per_sec": 106747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:57.821514+00:00", "epoch": 0, "step": 1448, "train_loss": 3.8465347290039062, "perplexity": 46.83050135569369, "lr": 0.001305, "grad_norm": 0.236329, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:58.126537+00:00", "epoch": 0, "step": 1449, "train_loss": 3.850790023803711, "perplexity": 47.0302035390733, "lr": 0.001305, "grad_norm": 0.26012, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:58.433484+00:00", "epoch": 0, "step": 1450, "train_loss": 3.8396799564361572, "perplexity": 46.51058664363877, "lr": 0.001305, "grad_norm": 0.262535, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:58.739696+00:00", "epoch": 0, "step": 1451, "train_loss": 3.8319029808044434, "perplexity": 46.15027781717133, "lr": 0.001305, "grad_norm": 0.248892, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:59.046143+00:00", "epoch": 0, "step": 1452, "train_loss": 3.816296100616455, "perplexity": 45.43560735501998, "lr": 0.001305, "grad_norm": 0.243052, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:59.352352+00:00", "epoch": 0, "step": 1453, "train_loss": 3.8199901580810547, "perplexity": 45.60375949013556, "lr": 0.001305, "grad_norm": 0.25861, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:59.657699+00:00", "epoch": 0, "step": 1454, "train_loss": 3.809490442276001, "perplexity": 45.127437972091776, "lr": 0.001305, "grad_norm": 0.237989, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:36:59.963028+00:00", "epoch": 0, "step": 1455, "train_loss": 3.799485206604004, "perplexity": 44.67817854089473, "lr": 0.001305, "grad_norm": 0.24674, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:00.268982+00:00", "epoch": 0, "step": 1456, "train_loss": 3.9094502925872803, "perplexity": 49.871529687428605, "lr": 0.001305, "grad_norm": 0.255902, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:00.576159+00:00", "epoch": 0, "step": 1457, "train_loss": 3.899186372756958, "perplexity": 49.36227027457714, "lr": 0.001305, "grad_norm": 0.283606, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:00.882710+00:00", "epoch": 0, "step": 1458, "train_loss": 3.9313700199127197, "perplexity": 50.97676903917164, "lr": 0.001305, "grad_norm": 0.308119, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:01.188458+00:00", "epoch": 0, "step": 1459, "train_loss": 3.7984509468078613, "perplexity": 44.63199358480361, "lr": 0.001305, "grad_norm": 0.290015, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:01.494215+00:00", "epoch": 0, "step": 1460, "train_loss": 3.960259199142456, "perplexity": 52.470924605304475, "lr": 0.001305, "grad_norm": 0.27152, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:01.799900+00:00", "epoch": 0, "step": 1461, "train_loss": 3.7242441177368164, "perplexity": 41.439897220042255, "lr": 0.001305, "grad_norm": 0.274516, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:02.105983+00:00", "epoch": 0, "step": 1462, "train_loss": 3.874985694885254, "perplexity": 48.182009037000924, "lr": 0.001305, "grad_norm": 0.275909, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:02.413224+00:00", "epoch": 0, "step": 1463, "train_loss": 3.8222780227661133, "perplexity": 45.70821416449681, "lr": 0.001305, "grad_norm": 0.260966, "tokens_per_sec": 106652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:02.720048+00:00", "epoch": 0, "step": 1464, "train_loss": 3.8435583114624023, "perplexity": 46.69132146147073, "lr": 0.001305, "grad_norm": 0.265407, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:03.026585+00:00", "epoch": 0, "step": 1465, "train_loss": 3.857287645339966, "perplexity": 47.33678294228526, "lr": 0.001305, "grad_norm": 0.252712, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:03.333613+00:00", "epoch": 0, "step": 1466, "train_loss": 3.8030929565429688, "perplexity": 44.8396573492304, "lr": 0.001305, "grad_norm": 0.276147, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:03.639616+00:00", "epoch": 0, "step": 1467, "train_loss": 3.880586862564087, "perplexity": 48.452641770832855, "lr": 0.001305, "grad_norm": 0.233771, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:03.947022+00:00", "epoch": 0, "step": 1468, "train_loss": 3.921076774597168, "perplexity": 50.45474392541333, "lr": 0.001305, "grad_norm": 0.240654, "tokens_per_sec": 106595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:04.253592+00:00", "epoch": 0, "step": 1469, "train_loss": 3.7418057918548584, "perplexity": 42.174079052412914, "lr": 0.001305, "grad_norm": 0.235104, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:04.559596+00:00", "epoch": 0, "step": 1470, "train_loss": 3.8936965465545654, "perplexity": 49.0920224752704, "lr": 0.001305, "grad_norm": 0.251728, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:04.865112+00:00", "epoch": 0, "step": 1471, "train_loss": 3.903254508972168, "perplexity": 49.56349173458714, "lr": 0.001305, "grad_norm": 0.262825, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:05.171145+00:00", "epoch": 0, "step": 1472, "train_loss": 3.903853178024292, "perplexity": 49.5931727468691, "lr": 0.001305, "grad_norm": 0.272289, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:05.477656+00:00", "epoch": 0, "step": 1473, "train_loss": 3.9530982971191406, "perplexity": 52.09652756526961, "lr": 0.001305, "grad_norm": 0.247691, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:05.784399+00:00", "epoch": 0, "step": 1474, "train_loss": 3.8738462924957275, "perplexity": 48.12714160474961, "lr": 0.001305, "grad_norm": 0.254546, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:06.091937+00:00", "epoch": 0, "step": 1475, "train_loss": 3.8979406356811523, "perplexity": 49.300816150138154, "lr": 0.001305, "grad_norm": 0.263114, "tokens_per_sec": 106549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:06.399045+00:00", "epoch": 0, "step": 1476, "train_loss": 3.8326616287231445, "perplexity": 46.18530291356316, "lr": 0.001305, "grad_norm": 0.286492, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:06.705673+00:00", "epoch": 0, "step": 1477, "train_loss": 3.855062484741211, "perplexity": 47.23156810138233, "lr": 0.001305, "grad_norm": 0.289059, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:07.011596+00:00", "epoch": 0, "step": 1478, "train_loss": 3.891948699951172, "perplexity": 49.006292094136306, "lr": 0.001305, "grad_norm": 0.296016, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:07.318305+00:00", "epoch": 0, "step": 1479, "train_loss": 3.745298385620117, "perplexity": 42.32163350185031, "lr": 0.001305, "grad_norm": 0.269746, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:07.625292+00:00", "epoch": 0, "step": 1480, "train_loss": 3.8238565921783447, "perplexity": 45.78042473296691, "lr": 0.001305, "grad_norm": 0.272098, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:07.931588+00:00", "epoch": 0, "step": 1481, "train_loss": 3.7870800495147705, "perplexity": 44.127362262510715, "lr": 0.001305, "grad_norm": 0.299621, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:08.238865+00:00", "epoch": 0, "step": 1482, "train_loss": 3.97159743309021, "perplexity": 53.06923772128367, "lr": 0.001305, "grad_norm": 0.284692, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:08.544604+00:00", "epoch": 0, "step": 1483, "train_loss": 3.801283836364746, "perplexity": 44.758610354317256, "lr": 0.001305, "grad_norm": 0.313454, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:08.879748+00:00", "epoch": 0, "step": 1484, "train_loss": 3.837292194366455, "perplexity": 46.39966291145345, "lr": 0.001305, "grad_norm": 0.320878, "tokens_per_sec": 97772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:09.186112+00:00", "epoch": 0, "step": 1485, "train_loss": 3.7526936531066895, "perplexity": 42.63577344482744, "lr": 0.001305, "grad_norm": 0.298504, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:09.492740+00:00", "epoch": 0, "step": 1486, "train_loss": 3.8527262210845947, "perplexity": 47.12135150302997, "lr": 0.001305, "grad_norm": 0.301955, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:09.799008+00:00", "epoch": 0, "step": 1487, "train_loss": 3.9674453735351562, "perplexity": 52.84934789944244, "lr": 0.001305, "grad_norm": 0.273274, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:10.106369+00:00", "epoch": 0, "step": 1488, "train_loss": 3.809323787689209, "perplexity": 45.11991790420787, "lr": 0.001305, "grad_norm": 0.261232, "tokens_per_sec": 106612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:10.412400+00:00", "epoch": 0, "step": 1489, "train_loss": 3.9318501949310303, "perplexity": 51.001252687926254, "lr": 0.001305, "grad_norm": 0.279634, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:10.718765+00:00", "epoch": 0, "step": 1490, "train_loss": 3.811836004257202, "perplexity": 45.23341140995574, "lr": 0.001305, "grad_norm": 0.244659, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:11.025833+00:00", "epoch": 0, "step": 1491, "train_loss": 3.8166680335998535, "perplexity": 45.45250949905364, "lr": 0.001305, "grad_norm": 0.249394, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:11.333232+00:00", "epoch": 0, "step": 1492, "train_loss": 3.8710920810699463, "perplexity": 47.994771652545346, "lr": 0.001305, "grad_norm": 0.260017, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:11.640197+00:00", "epoch": 0, "step": 1493, "train_loss": 3.750248908996582, "perplexity": 42.53166719723775, "lr": 0.001305, "grad_norm": 0.246135, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:11.946867+00:00", "epoch": 0, "step": 1494, "train_loss": 3.781581401824951, "perplexity": 43.88538732110269, "lr": 0.001305, "grad_norm": 0.240765, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:12.252877+00:00", "epoch": 0, "step": 1495, "train_loss": 3.7975523471832275, "perplexity": 44.59190530647512, "lr": 0.001305, "grad_norm": 0.239643, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:12.559011+00:00", "epoch": 0, "step": 1496, "train_loss": 3.8335766792297363, "perplexity": 46.227584140169995, "lr": 0.001305, "grad_norm": 0.288019, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:12.865311+00:00", "epoch": 0, "step": 1497, "train_loss": 3.9055726528167725, "perplexity": 49.67852031275237, "lr": 0.001305, "grad_norm": 0.28424, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:13.172816+00:00", "epoch": 0, "step": 1498, "train_loss": 3.8812029361724854, "perplexity": 48.482501361593535, "lr": 0.001305, "grad_norm": 0.292403, "tokens_per_sec": 106561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:13.479241+00:00", "epoch": 0, "step": 1499, "train_loss": 3.780472755432129, "perplexity": 43.83676090448279, "lr": 0.001305, "grad_norm": 0.302859, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:13.786513+00:00", "epoch": 0, "step": 1500, "train_loss": 3.819892168045044, "perplexity": 45.599290995038466, "lr": 0.001305, "grad_norm": 0.314054, "tokens_per_sec": 106697} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:37:19.366314+00:00", "step": 1500, "epoch": 0, "val_loss": 3.8826794266700744, "val_ppl": 48.55413818668487, "eval_train_loss": 3.819892168045044, "eval_train_ppl": 45.599290995038466} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:37:20.412284+00:00", "step": 1500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p8827_epoch_0000_step_0001500.pt", "val_loss": 3.8826794266700744} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:37:22.151701+00:00", "step": 1500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0000_step_0001500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:23.178857+00:00", "epoch": 0, "step": 1501, "train_loss": 3.9225528240203857, "perplexity": 50.529272611563705, "lr": 0.001305, "grad_norm": 0.292887, "tokens_per_sec": 3489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:23.482783+00:00", "epoch": 0, "step": 1502, "train_loss": 3.9986300468444824, "perplexity": 54.523404335962816, "lr": 0.001305, "grad_norm": 0.25713, "tokens_per_sec": 107816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:23.788221+00:00", "epoch": 0, "step": 1503, "train_loss": 3.8680877685546875, "perplexity": 47.85079674050951, "lr": 0.001305, "grad_norm": 0.250275, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:24.094597+00:00", "epoch": 0, "step": 1504, "train_loss": 3.8982746601104736, "perplexity": 49.31728657772723, "lr": 0.001305, "grad_norm": 0.248045, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:24.400541+00:00", "epoch": 0, "step": 1505, "train_loss": 3.953166961669922, "perplexity": 52.10010487274784, "lr": 0.001305, "grad_norm": 0.23346, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:24.706039+00:00", "epoch": 0, "step": 1506, "train_loss": 3.72807240486145, "perplexity": 41.59884509995155, "lr": 0.001305, "grad_norm": 0.256604, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:25.011379+00:00", "epoch": 0, "step": 1507, "train_loss": 3.791940212249756, "perplexity": 44.3423504397465, "lr": 0.001305, "grad_norm": 0.251613, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:25.316963+00:00", "epoch": 0, "step": 1508, "train_loss": 3.79945707321167, "perplexity": 44.67692160985001, "lr": 0.001305, "grad_norm": 0.248438, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:25.622841+00:00", "epoch": 0, "step": 1509, "train_loss": 3.8825714588165283, "perplexity": 48.54889618359308, "lr": 0.001305, "grad_norm": 0.27895, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:25.928564+00:00", "epoch": 0, "step": 1510, "train_loss": 3.915978193283081, "perplexity": 50.19815099663972, "lr": 0.001305, "grad_norm": 0.301788, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:26.234631+00:00", "epoch": 0, "step": 1511, "train_loss": 3.9737086296081543, "perplexity": 53.18139566324033, "lr": 0.001305, "grad_norm": 0.300276, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:26.540518+00:00", "epoch": 0, "step": 1512, "train_loss": 3.9628632068634033, "perplexity": 52.607737351516704, "lr": 0.001305, "grad_norm": 0.265623, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:26.845606+00:00", "epoch": 0, "step": 1513, "train_loss": 3.814350128173828, "perplexity": 45.34727688739523, "lr": 0.001305, "grad_norm": 0.259871, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:27.152245+00:00", "epoch": 0, "step": 1514, "train_loss": 3.854104995727539, "perplexity": 47.18636603751992, "lr": 0.001305, "grad_norm": 0.288637, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:27.722819+00:00", "epoch": 0, "step": 1515, "train_loss": 3.89608097076416, "perplexity": 49.20921834897144, "lr": 0.001305, "grad_norm": 0.317815, "tokens_per_sec": 57430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:28.027726+00:00", "epoch": 0, "step": 1516, "train_loss": 3.9804205894470215, "perplexity": 53.53954766144762, "lr": 0.001305, "grad_norm": 0.295401, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:28.332818+00:00", "epoch": 0, "step": 1517, "train_loss": 3.784517765045166, "perplexity": 44.01444013858372, "lr": 0.001305, "grad_norm": 0.254238, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:28.639357+00:00", "epoch": 0, "step": 1518, "train_loss": 3.8798248767852783, "perplexity": 48.41573560962848, "lr": 0.001305, "grad_norm": 0.250509, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:28.946031+00:00", "epoch": 0, "step": 1519, "train_loss": 3.870722532272339, "perplexity": 47.97703851922057, "lr": 0.001305, "grad_norm": 0.283753, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:29.251868+00:00", "epoch": 0, "step": 1520, "train_loss": 3.7643492221832275, "perplexity": 43.13562501578298, "lr": 0.001305, "grad_norm": 0.279821, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:29.557077+00:00", "epoch": 0, "step": 1521, "train_loss": 3.8695554733276367, "perplexity": 47.92107914757052, "lr": 0.001305, "grad_norm": 0.274799, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:29.861950+00:00", "epoch": 0, "step": 1522, "train_loss": 3.903351068496704, "perplexity": 49.56827779284942, "lr": 0.001305, "grad_norm": 0.240751, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:30.168212+00:00", "epoch": 0, "step": 1523, "train_loss": 3.8131890296936035, "perplexity": 45.29465478875021, "lr": 0.001305, "grad_norm": 0.246832, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:30.474942+00:00", "epoch": 0, "step": 1524, "train_loss": 3.806549072265625, "perplexity": 44.994896501707224, "lr": 0.001305, "grad_norm": 0.284671, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:30.781323+00:00", "epoch": 0, "step": 1525, "train_loss": 3.8174474239349365, "perplexity": 45.48794855429709, "lr": 0.001305, "grad_norm": 0.255853, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:31.086787+00:00", "epoch": 0, "step": 1526, "train_loss": 3.788783550262451, "perplexity": 44.20259732041296, "lr": 0.001305, "grad_norm": 0.253043, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:31.392950+00:00", "epoch": 0, "step": 1527, "train_loss": 3.8035593032836914, "perplexity": 44.86057305389662, "lr": 0.001305, "grad_norm": 0.260374, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:31.700167+00:00", "epoch": 0, "step": 1528, "train_loss": 3.897447347640991, "perplexity": 49.27650264443492, "lr": 0.001305, "grad_norm": 0.249192, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:32.006021+00:00", "epoch": 0, "step": 1529, "train_loss": 3.8413989543914795, "perplexity": 46.59060700469384, "lr": 0.001305, "grad_norm": 0.265318, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:32.312919+00:00", "epoch": 0, "step": 1530, "train_loss": 3.852513313293457, "perplexity": 47.11132006808926, "lr": 0.001305, "grad_norm": 0.275587, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:32.633253+00:00", "epoch": 0, "step": 1531, "train_loss": 3.761915445327759, "perplexity": 43.03077017836858, "lr": 0.001305, "grad_norm": 0.251703, "tokens_per_sec": 102230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:32.939239+00:00", "epoch": 0, "step": 1532, "train_loss": 3.8064115047454834, "perplexity": 44.988707091117156, "lr": 0.001305, "grad_norm": 0.259527, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:33.245194+00:00", "epoch": 0, "step": 1533, "train_loss": 3.8383350372314453, "perplexity": 46.4480757079408, "lr": 0.001305, "grad_norm": 0.269492, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:33.551224+00:00", "epoch": 0, "step": 1534, "train_loss": 3.7802326679229736, "perplexity": 43.82623750906624, "lr": 0.001305, "grad_norm": 0.229604, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:33.858247+00:00", "epoch": 0, "step": 1535, "train_loss": 3.9201555252075195, "perplexity": 50.40828392727553, "lr": 0.001305, "grad_norm": 0.262711, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:34.164298+00:00", "epoch": 0, "step": 1536, "train_loss": 3.8439390659332275, "perplexity": 46.70910277580857, "lr": 0.001305, "grad_norm": 0.269251, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:34.469877+00:00", "epoch": 0, "step": 1537, "train_loss": 3.861704111099243, "perplexity": 47.54630655967916, "lr": 0.001305, "grad_norm": 0.284161, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:34.775616+00:00", "epoch": 0, "step": 1538, "train_loss": 3.8028526306152344, "perplexity": 44.82888251176692, "lr": 0.001305, "grad_norm": 0.298976, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:35.080889+00:00", "epoch": 0, "step": 1539, "train_loss": 3.8565056324005127, "perplexity": 47.299779436008436, "lr": 0.001305, "grad_norm": 0.328588, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:35.387263+00:00", "epoch": 0, "step": 1540, "train_loss": 3.8289175033569336, "perplexity": 46.01270266948702, "lr": 0.001305, "grad_norm": 0.341094, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:35.693598+00:00", "epoch": 0, "step": 1541, "train_loss": 3.7672858238220215, "perplexity": 43.26248333790814, "lr": 0.001305, "grad_norm": 0.361023, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:36.000076+00:00", "epoch": 0, "step": 1542, "train_loss": 3.891491174697876, "perplexity": 48.9838756063785, "lr": 0.001305, "grad_norm": 0.334186, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:36.305988+00:00", "epoch": 0, "step": 1543, "train_loss": 3.908413887023926, "perplexity": 49.81986933177175, "lr": 0.001305, "grad_norm": 0.332321, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:36.611951+00:00", "epoch": 0, "step": 1544, "train_loss": 3.90309739112854, "perplexity": 49.55570503737369, "lr": 0.001305, "grad_norm": 0.30469, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:36.918564+00:00", "epoch": 0, "step": 1545, "train_loss": 3.84837007522583, "perplexity": 46.916530461876114, "lr": 0.001305, "grad_norm": 0.26843, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:37.225401+00:00", "epoch": 0, "step": 1546, "train_loss": 3.8138787746429443, "perplexity": 45.32590732502338, "lr": 0.001305, "grad_norm": 0.249292, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:37.532156+00:00", "epoch": 0, "step": 1547, "train_loss": 3.866570234298706, "perplexity": 47.778236587463674, "lr": 0.001305, "grad_norm": 0.238887, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:37.840060+00:00", "epoch": 0, "step": 1548, "train_loss": 3.8689393997192383, "perplexity": 47.89156532769797, "lr": 0.001305, "grad_norm": 0.240226, "tokens_per_sec": 106423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:38.145969+00:00", "epoch": 0, "step": 1549, "train_loss": 3.836578607559204, "perplexity": 46.36656453482588, "lr": 0.001305, "grad_norm": 0.250331, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:38.452866+00:00", "epoch": 0, "step": 1550, "train_loss": 3.837677478790283, "perplexity": 46.41754342316424, "lr": 0.001305, "grad_norm": 0.229699, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:38.760247+00:00", "epoch": 0, "step": 1551, "train_loss": 3.8079166412353516, "perplexity": 45.056472220883634, "lr": 0.001305, "grad_norm": 0.231667, "tokens_per_sec": 106547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:39.067687+00:00", "epoch": 0, "step": 1552, "train_loss": 3.7283003330230713, "perplexity": 41.608327728878784, "lr": 0.001305, "grad_norm": 0.229794, "tokens_per_sec": 106583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:39.374844+00:00", "epoch": 0, "step": 1553, "train_loss": 3.870522975921631, "perplexity": 47.96746535172085, "lr": 0.001305, "grad_norm": 0.247616, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:39.681361+00:00", "epoch": 0, "step": 1554, "train_loss": 3.774120807647705, "perplexity": 43.55919456487954, "lr": 0.001305, "grad_norm": 0.255496, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:39.988250+00:00", "epoch": 0, "step": 1555, "train_loss": 3.7340362071990967, "perplexity": 41.847673634193804, "lr": 0.001305, "grad_norm": 0.265491, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:40.295523+00:00", "epoch": 0, "step": 1556, "train_loss": 3.9031245708465576, "perplexity": 49.55705196576725, "lr": 0.001305, "grad_norm": 0.252704, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:40.602845+00:00", "epoch": 0, "step": 1557, "train_loss": 3.765174150466919, "perplexity": 43.17122349397021, "lr": 0.001305, "grad_norm": 0.262543, "tokens_per_sec": 106624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:40.910194+00:00", "epoch": 0, "step": 1558, "train_loss": 3.9025111198425293, "perplexity": 49.52666046528349, "lr": 0.001305, "grad_norm": 0.256543, "tokens_per_sec": 106615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:41.216280+00:00", "epoch": 0, "step": 1559, "train_loss": 3.6602184772491455, "perplexity": 38.869834118159446, "lr": 0.001305, "grad_norm": 0.259233, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:41.524212+00:00", "epoch": 0, "step": 1560, "train_loss": 3.8660764694213867, "perplexity": 47.75465119563477, "lr": 0.001305, "grad_norm": 0.254807, "tokens_per_sec": 106470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:41.830685+00:00", "epoch": 0, "step": 1561, "train_loss": 3.753880262374878, "perplexity": 42.686395477099964, "lr": 0.001305, "grad_norm": 0.270023, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:42.137534+00:00", "epoch": 0, "step": 1562, "train_loss": 3.7837042808532715, "perplexity": 43.97864964678926, "lr": 0.001305, "grad_norm": 0.289813, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:42.443641+00:00", "epoch": 0, "step": 1563, "train_loss": 3.9812138080596924, "perplexity": 53.582033075049445, "lr": 0.001305, "grad_norm": 0.294232, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:42.750523+00:00", "epoch": 0, "step": 1564, "train_loss": 3.914047956466675, "perplexity": 50.101350131822215, "lr": 0.001305, "grad_norm": 0.277989, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:43.056723+00:00", "epoch": 0, "step": 1565, "train_loss": 3.8180017471313477, "perplexity": 45.51317056926603, "lr": 0.001305, "grad_norm": 0.261407, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:43.362842+00:00", "epoch": 0, "step": 1566, "train_loss": 3.915666103363037, "perplexity": 50.182487104107445, "lr": 0.001305, "grad_norm": 0.244534, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:43.669773+00:00", "epoch": 0, "step": 1567, "train_loss": 3.8955576419830322, "perplexity": 49.183472486075736, "lr": 0.001305, "grad_norm": 0.231446, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:43.976456+00:00", "epoch": 0, "step": 1568, "train_loss": 3.8137121200561523, "perplexity": 45.31835418406743, "lr": 0.001305, "grad_norm": 0.259199, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:44.282382+00:00", "epoch": 0, "step": 1569, "train_loss": 3.754809856414795, "perplexity": 42.72609494525667, "lr": 0.001305, "grad_norm": 0.237588, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:44.589829+00:00", "epoch": 0, "step": 1570, "train_loss": 3.8610215187072754, "perplexity": 47.513862886717234, "lr": 0.001305, "grad_norm": 0.244905, "tokens_per_sec": 106638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:44.896632+00:00", "epoch": 0, "step": 1571, "train_loss": 3.8623390197753906, "perplexity": 47.57650370743382, "lr": 0.001305, "grad_norm": 0.259656, "tokens_per_sec": 106747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:45.203542+00:00", "epoch": 0, "step": 1572, "train_loss": 3.834909200668335, "perplexity": 46.289224446485434, "lr": 0.001305, "grad_norm": 0.231104, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:45.509654+00:00", "epoch": 0, "step": 1573, "train_loss": 3.7981224060058594, "perplexity": 44.61733256234022, "lr": 0.001305, "grad_norm": 0.245912, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:45.815880+00:00", "epoch": 0, "step": 1574, "train_loss": 3.8733294010162354, "perplexity": 48.102271523442084, "lr": 0.001305, "grad_norm": 0.264993, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:46.136042+00:00", "epoch": 0, "step": 1575, "train_loss": 3.9310367107391357, "perplexity": 50.95978084572885, "lr": 0.001305, "grad_norm": 0.313436, "tokens_per_sec": 102348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:46.443962+00:00", "epoch": 0, "step": 1576, "train_loss": 3.907914161682129, "perplexity": 49.79497930014941, "lr": 0.001305, "grad_norm": 0.308295, "tokens_per_sec": 106417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:46.752506+00:00", "epoch": 0, "step": 1577, "train_loss": 3.9198198318481445, "perplexity": 50.39136504104141, "lr": 0.001305, "grad_norm": 0.278929, "tokens_per_sec": 106202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:47.059282+00:00", "epoch": 0, "step": 1578, "train_loss": 3.775815725326538, "perplexity": 43.63308641643667, "lr": 0.001305, "grad_norm": 0.311777, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:47.366008+00:00", "epoch": 0, "step": 1579, "train_loss": 3.7991695404052734, "perplexity": 44.664077375856195, "lr": 0.001305, "grad_norm": 0.30263, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:47.673756+00:00", "epoch": 0, "step": 1580, "train_loss": 3.805145740509033, "perplexity": 44.931798018969324, "lr": 0.001305, "grad_norm": 0.297531, "tokens_per_sec": 106529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:47.994575+00:00", "epoch": 0, "step": 1581, "train_loss": 3.851527690887451, "perplexity": 47.064908971119344, "lr": 0.001305, "grad_norm": 0.307347, "tokens_per_sec": 102090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:48.301495+00:00", "epoch": 0, "step": 1582, "train_loss": 3.822828531265259, "perplexity": 45.73338385230299, "lr": 0.001305, "grad_norm": 0.281351, "tokens_per_sec": 106763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:48.609473+00:00", "epoch": 0, "step": 1583, "train_loss": 3.8636162281036377, "perplexity": 47.637307635573954, "lr": 0.001305, "grad_norm": 0.244425, "tokens_per_sec": 106398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:48.915545+00:00", "epoch": 0, "step": 1584, "train_loss": 3.7968451976776123, "perplexity": 44.56038330937911, "lr": 0.001305, "grad_norm": 0.261454, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:49.231660+00:00", "epoch": 0, "step": 1585, "train_loss": 3.798492908477783, "perplexity": 44.63386645708052, "lr": 0.001305, "grad_norm": 0.26106, "tokens_per_sec": 103658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:49.538325+00:00", "epoch": 0, "step": 1586, "train_loss": 3.8394875526428223, "perplexity": 46.5016386911759, "lr": 0.001305, "grad_norm": 0.242011, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:49.853966+00:00", "epoch": 0, "step": 1587, "train_loss": 3.826571464538574, "perplexity": 45.904881608536996, "lr": 0.001305, "grad_norm": 0.255922, "tokens_per_sec": 103815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:50.161643+00:00", "epoch": 0, "step": 1588, "train_loss": 3.8447213172912598, "perplexity": 46.74565532966593, "lr": 0.001305, "grad_norm": 0.278855, "tokens_per_sec": 106501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:50.481955+00:00", "epoch": 0, "step": 1589, "train_loss": 3.7685253620147705, "perplexity": 43.31614208748302, "lr": 0.001305, "grad_norm": 0.266759, "tokens_per_sec": 102300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:50.791066+00:00", "epoch": 0, "step": 1590, "train_loss": 3.7951712608337402, "perplexity": 44.485854437695494, "lr": 0.001305, "grad_norm": 0.243474, "tokens_per_sec": 106063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:51.108698+00:00", "epoch": 0, "step": 1591, "train_loss": 3.7954976558685303, "perplexity": 44.50037676958193, "lr": 0.001305, "grad_norm": 0.252701, "tokens_per_sec": 103111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:51.415784+00:00", "epoch": 0, "step": 1592, "train_loss": 3.78096866607666, "perplexity": 43.85850541205784, "lr": 0.001305, "grad_norm": 0.256922, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:51.731426+00:00", "epoch": 0, "step": 1593, "train_loss": 3.933546781539917, "perplexity": 51.0878541729601, "lr": 0.001305, "grad_norm": 0.30535, "tokens_per_sec": 103811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:52.049188+00:00", "epoch": 0, "step": 1594, "train_loss": 3.922367572784424, "perplexity": 50.51991286833892, "lr": 0.001305, "grad_norm": 0.265977, "tokens_per_sec": 103121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:52.355476+00:00", "epoch": 0, "step": 1595, "train_loss": 3.774972438812256, "perplexity": 43.59630673317022, "lr": 0.001305, "grad_norm": 0.28197, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:52.675378+00:00", "epoch": 0, "step": 1596, "train_loss": 3.7629079818725586, "perplexity": 43.073500992760984, "lr": 0.001305, "grad_norm": 0.265887, "tokens_per_sec": 102431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:52.984575+00:00", "epoch": 0, "step": 1597, "train_loss": 3.771592855453491, "perplexity": 43.44921806962975, "lr": 0.001305, "grad_norm": 0.243316, "tokens_per_sec": 105978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:53.302049+00:00", "epoch": 0, "step": 1598, "train_loss": 3.828084945678711, "perplexity": 45.97441038306417, "lr": 0.001305, "grad_norm": 0.225543, "tokens_per_sec": 103216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:53.609263+00:00", "epoch": 0, "step": 1599, "train_loss": 3.789337635040283, "perplexity": 44.227096093300624, "lr": 0.001305, "grad_norm": 0.236079, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:37:53.927114+00:00", "epoch": 0, "step": 1600, "train_loss": 3.910210609436035, "perplexity": 49.909462270286724, "lr": 0.001305, "grad_norm": 0.236641, "tokens_per_sec": 103143} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:37:59.481394+00:00", "step": 1600, "epoch": 0, "val_loss": 3.85900616645813, "val_ppl": 47.418202143708015, "eval_train_loss": 3.910210609436035, "eval_train_ppl": 49.909462270286724} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:38:00.446813+00:00", "step": 1600, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p8590_epoch_0000_step_0001600.pt", "val_loss": 3.85900616645813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:01.534215+00:00", "epoch": 0, "step": 1601, "train_loss": 3.7604055404663086, "perplexity": 42.965846835654126, "lr": 0.001305, "grad_norm": 0.22671, "tokens_per_sec": 4307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:01.839864+00:00", "epoch": 0, "step": 1602, "train_loss": 3.7921035289764404, "perplexity": 44.34959287866343, "lr": 0.001305, "grad_norm": 0.241602, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:02.145093+00:00", "epoch": 0, "step": 1603, "train_loss": 3.7436652183532715, "perplexity": 42.25257160550026, "lr": 0.001305, "grad_norm": 0.254084, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:02.450386+00:00", "epoch": 0, "step": 1604, "train_loss": 3.871577262878418, "perplexity": 48.01806349258454, "lr": 0.001305, "grad_norm": 0.223686, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:02.755187+00:00", "epoch": 0, "step": 1605, "train_loss": 3.818708896636963, "perplexity": 45.54536656768359, "lr": 0.001305, "grad_norm": 0.232563, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:03.061663+00:00", "epoch": 0, "step": 1606, "train_loss": 3.834897994995117, "perplexity": 46.288705747468974, "lr": 0.001305, "grad_norm": 0.245691, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:03.368181+00:00", "epoch": 0, "step": 1607, "train_loss": 3.8109703063964844, "perplexity": 45.194269887275894, "lr": 0.001305, "grad_norm": 0.26607, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:03.674541+00:00", "epoch": 0, "step": 1608, "train_loss": 3.8670198917388916, "perplexity": 47.79972525793238, "lr": 0.001305, "grad_norm": 0.287122, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:03.980226+00:00", "epoch": 0, "step": 1609, "train_loss": 3.7412328720092773, "perplexity": 42.14992360578278, "lr": 0.001305, "grad_norm": 0.301344, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:04.286800+00:00", "epoch": 0, "step": 1610, "train_loss": 3.7410287857055664, "perplexity": 42.14132226141052, "lr": 0.001305, "grad_norm": 0.319027, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:04.594473+00:00", "epoch": 0, "step": 1611, "train_loss": 3.8372466564178467, "perplexity": 46.397550014097206, "lr": 0.001305, "grad_norm": 0.346747, "tokens_per_sec": 106442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:04.902235+00:00", "epoch": 0, "step": 1612, "train_loss": 3.7582945823669434, "perplexity": 42.875243396956854, "lr": 0.001305, "grad_norm": 0.364894, "tokens_per_sec": 106470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:05.208457+00:00", "epoch": 0, "step": 1613, "train_loss": 3.6887810230255127, "perplexity": 39.99606295023028, "lr": 0.001305, "grad_norm": 0.39176, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:05.514822+00:00", "epoch": 0, "step": 1614, "train_loss": 3.837271213531494, "perplexity": 46.398689417996046, "lr": 0.001305, "grad_norm": 0.338104, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:05.874692+00:00", "epoch": 0, "step": 1615, "train_loss": 3.902553081512451, "perplexity": 49.5287387302657, "lr": 0.001305, "grad_norm": 0.301039, "tokens_per_sec": 91058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:06.181071+00:00", "epoch": 0, "step": 1616, "train_loss": 3.885957956314087, "perplexity": 48.71358560178617, "lr": 0.001305, "grad_norm": 0.264476, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:06.486912+00:00", "epoch": 0, "step": 1617, "train_loss": 3.977670192718506, "perplexity": 53.392494984015386, "lr": 0.001305, "grad_norm": 0.261879, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:06.794637+00:00", "epoch": 0, "step": 1618, "train_loss": 3.8787598609924316, "perplexity": 48.36419953481999, "lr": 0.001305, "grad_norm": 0.25215, "tokens_per_sec": 106489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:07.102143+00:00", "epoch": 0, "step": 1619, "train_loss": 3.8547091484069824, "perplexity": 47.21488242025182, "lr": 0.001305, "grad_norm": 0.252653, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:07.410848+00:00", "epoch": 0, "step": 1620, "train_loss": 3.774754285812378, "perplexity": 43.58679710538741, "lr": 0.001305, "grad_norm": 0.250901, "tokens_per_sec": 106204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:07.715950+00:00", "epoch": 0, "step": 1621, "train_loss": 3.769333600997925, "perplexity": 43.35116603407085, "lr": 0.001305, "grad_norm": 0.266346, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:08.022064+00:00", "epoch": 0, "step": 1622, "train_loss": 3.883943557739258, "perplexity": 48.615555793077384, "lr": 0.001305, "grad_norm": 0.250024, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:08.328469+00:00", "epoch": 0, "step": 1623, "train_loss": 3.849527597427368, "perplexity": 46.9708688303717, "lr": 0.001305, "grad_norm": 0.242013, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:08.635213+00:00", "epoch": 0, "step": 1624, "train_loss": 3.8370206356048584, "perplexity": 46.38706438715196, "lr": 0.001305, "grad_norm": 0.285869, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:08.943004+00:00", "epoch": 0, "step": 1625, "train_loss": 3.783452272415161, "perplexity": 43.96756805236787, "lr": 0.001305, "grad_norm": 0.283394, "tokens_per_sec": 106462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:09.248998+00:00", "epoch": 0, "step": 1626, "train_loss": 3.7670552730560303, "perplexity": 43.25251028892711, "lr": 0.001305, "grad_norm": 0.266106, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:09.555237+00:00", "epoch": 0, "step": 1627, "train_loss": 3.8490004539489746, "perplexity": 46.946114968185015, "lr": 0.001305, "grad_norm": 0.275802, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:09.861534+00:00", "epoch": 0, "step": 1628, "train_loss": 3.979837417602539, "perplexity": 53.50833400703197, "lr": 0.001305, "grad_norm": 0.328015, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:10.167333+00:00", "epoch": 0, "step": 1629, "train_loss": 3.8149118423461914, "perplexity": 45.37275625089081, "lr": 0.001305, "grad_norm": 0.290953, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:10.474772+00:00", "epoch": 0, "step": 1630, "train_loss": 3.8265998363494873, "perplexity": 45.90618403163394, "lr": 0.001305, "grad_norm": 0.250306, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:10.782217+00:00", "epoch": 0, "step": 1631, "train_loss": 3.838858127593994, "perplexity": 46.47237860445389, "lr": 0.001305, "grad_norm": 0.302113, "tokens_per_sec": 106526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:11.089735+00:00", "epoch": 0, "step": 1632, "train_loss": 3.826756000518799, "perplexity": 45.91335349252139, "lr": 0.001305, "grad_norm": 0.277754, "tokens_per_sec": 106556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:11.396771+00:00", "epoch": 0, "step": 1633, "train_loss": 3.8490395545959473, "perplexity": 46.94795062754062, "lr": 0.001305, "grad_norm": 0.268174, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:11.703587+00:00", "epoch": 0, "step": 1634, "train_loss": 3.7490901947021484, "perplexity": 42.48241368737391, "lr": 0.001305, "grad_norm": 0.261358, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:12.011103+00:00", "epoch": 0, "step": 1635, "train_loss": 3.7699124813079834, "perplexity": 43.37626843544801, "lr": 0.001305, "grad_norm": 0.273689, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:12.318432+00:00", "epoch": 0, "step": 1636, "train_loss": 3.8894243240356445, "perplexity": 48.88273780501263, "lr": 0.001305, "grad_norm": 0.284149, "tokens_per_sec": 106622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:12.625458+00:00", "epoch": 0, "step": 1637, "train_loss": 3.7678163051605225, "perplexity": 43.2854393663101, "lr": 0.001305, "grad_norm": 0.23005, "tokens_per_sec": 106727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:12.933015+00:00", "epoch": 0, "step": 1638, "train_loss": 3.6897218227386475, "perplexity": 40.03370894067216, "lr": 0.001305, "grad_norm": 0.256194, "tokens_per_sec": 106543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:13.239362+00:00", "epoch": 0, "step": 1639, "train_loss": 3.8289780616760254, "perplexity": 46.01548920579068, "lr": 0.001305, "grad_norm": 0.265012, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:13.545886+00:00", "epoch": 0, "step": 1640, "train_loss": 3.7935895919799805, "perplexity": 44.415548162578254, "lr": 0.001305, "grad_norm": 0.282182, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:13.852950+00:00", "epoch": 0, "step": 1641, "train_loss": 3.839848518371582, "perplexity": 46.51842721893387, "lr": 0.001305, "grad_norm": 0.292629, "tokens_per_sec": 106658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:14.160509+00:00", "epoch": 0, "step": 1642, "train_loss": 3.792560338973999, "perplexity": 44.369856844116676, "lr": 0.001305, "grad_norm": 0.255477, "tokens_per_sec": 106542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:14.467156+00:00", "epoch": 0, "step": 1643, "train_loss": 3.7714829444885254, "perplexity": 43.4444427865775, "lr": 0.001305, "grad_norm": 0.256392, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:14.773804+00:00", "epoch": 0, "step": 1644, "train_loss": 3.7741615772247314, "perplexity": 43.56097049101919, "lr": 0.001305, "grad_norm": 0.248813, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:15.079944+00:00", "epoch": 0, "step": 1645, "train_loss": 3.7564988136291504, "perplexity": 42.79831846560497, "lr": 0.001305, "grad_norm": 0.265689, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:15.387869+00:00", "epoch": 0, "step": 1646, "train_loss": 3.7539496421813965, "perplexity": 42.689357153698225, "lr": 0.001305, "grad_norm": 0.270213, "tokens_per_sec": 106415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:15.695349+00:00", "epoch": 0, "step": 1647, "train_loss": 3.9175846576690674, "perplexity": 50.278857347040244, "lr": 0.001305, "grad_norm": 0.264848, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:16.002866+00:00", "epoch": 0, "step": 1648, "train_loss": 3.866405725479126, "perplexity": 47.77037729264044, "lr": 0.001305, "grad_norm": 0.275288, "tokens_per_sec": 106556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:16.309358+00:00", "epoch": 0, "step": 1649, "train_loss": 3.8221030235290527, "perplexity": 45.70021596175072, "lr": 0.001305, "grad_norm": 0.287628, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:16.615906+00:00", "epoch": 0, "step": 1650, "train_loss": 3.7313575744628906, "perplexity": 41.73572908176876, "lr": 0.001305, "grad_norm": 0.286101, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:16.922844+00:00", "epoch": 0, "step": 1651, "train_loss": 3.766407012939453, "perplexity": 43.22448049784438, "lr": 0.001305, "grad_norm": 0.264625, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:17.229193+00:00", "epoch": 0, "step": 1652, "train_loss": 3.8964157104492188, "perplexity": 49.22569338449382, "lr": 0.001305, "grad_norm": 0.221449, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:17.536667+00:00", "epoch": 0, "step": 1653, "train_loss": 3.7685599327087402, "perplexity": 43.31763958245965, "lr": 0.001305, "grad_norm": 0.235673, "tokens_per_sec": 106572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:17.843808+00:00", "epoch": 0, "step": 1654, "train_loss": 3.8188419342041016, "perplexity": 45.55142621551766, "lr": 0.001305, "grad_norm": 0.236451, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:18.150752+00:00", "epoch": 0, "step": 1655, "train_loss": 3.8248867988586426, "perplexity": 45.82761233467148, "lr": 0.001305, "grad_norm": 0.246731, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:18.456259+00:00", "epoch": 0, "step": 1656, "train_loss": 3.7820968627929688, "perplexity": 43.90801435650884, "lr": 0.001305, "grad_norm": 0.258558, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:18.770927+00:00", "epoch": 0, "step": 1657, "train_loss": 3.862098455429077, "perplexity": 47.56505987346397, "lr": 0.001305, "grad_norm": 0.286106, "tokens_per_sec": 104135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:19.078571+00:00", "epoch": 0, "step": 1658, "train_loss": 3.8333306312561035, "perplexity": 46.216211335951456, "lr": 0.001305, "grad_norm": 0.281537, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:19.386825+00:00", "epoch": 0, "step": 1659, "train_loss": 3.7865192890167236, "perplexity": 44.1026243175503, "lr": 0.001305, "grad_norm": 0.235149, "tokens_per_sec": 106303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:19.695360+00:00", "epoch": 0, "step": 1660, "train_loss": 3.840320587158203, "perplexity": 46.54039230053344, "lr": 0.001305, "grad_norm": 0.294149, "tokens_per_sec": 106262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:20.001839+00:00", "epoch": 0, "step": 1661, "train_loss": 3.7311720848083496, "perplexity": 41.72798825374332, "lr": 0.001305, "grad_norm": 0.26751, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:20.308609+00:00", "epoch": 0, "step": 1662, "train_loss": 3.9193365573883057, "perplexity": 50.367018064930384, "lr": 0.001305, "grad_norm": 0.277588, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:20.615000+00:00", "epoch": 0, "step": 1663, "train_loss": 3.8966357707977295, "perplexity": 49.23652719973855, "lr": 0.001305, "grad_norm": 0.274579, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:20.933209+00:00", "epoch": 0, "step": 1664, "train_loss": 3.8564600944519043, "perplexity": 47.29762555012544, "lr": 0.001305, "grad_norm": 0.282042, "tokens_per_sec": 102977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:21.239544+00:00", "epoch": 0, "step": 1665, "train_loss": 3.758399724960327, "perplexity": 42.87975164824004, "lr": 0.001305, "grad_norm": 0.294726, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:21.547333+00:00", "epoch": 0, "step": 1666, "train_loss": 3.737133026123047, "perplexity": 41.97746917473921, "lr": 0.001305, "grad_norm": 0.259662, "tokens_per_sec": 106462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:21.868174+00:00", "epoch": 0, "step": 1667, "train_loss": 3.6998417377471924, "perplexity": 40.440903585073265, "lr": 0.001305, "grad_norm": 0.258577, "tokens_per_sec": 102131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:22.175132+00:00", "epoch": 0, "step": 1668, "train_loss": 3.7683229446411133, "perplexity": 43.307375035096506, "lr": 0.001305, "grad_norm": 0.298903, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:22.482524+00:00", "epoch": 0, "step": 1669, "train_loss": 3.754612445831299, "perplexity": 42.71766119440624, "lr": 0.001305, "grad_norm": 0.299281, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:22.789558+00:00", "epoch": 0, "step": 1670, "train_loss": 3.9258670806884766, "perplexity": 50.69701741136446, "lr": 0.001305, "grad_norm": 0.241405, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:23.105502+00:00", "epoch": 0, "step": 1671, "train_loss": 3.918154001235962, "perplexity": 50.30749144156687, "lr": 0.001305, "grad_norm": 0.253883, "tokens_per_sec": 103658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:23.413830+00:00", "epoch": 0, "step": 1672, "train_loss": 3.9069666862487793, "perplexity": 49.747822124219, "lr": 0.001305, "grad_norm": 0.258547, "tokens_per_sec": 106277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:23.721787+00:00", "epoch": 0, "step": 1673, "train_loss": 3.843372106552124, "perplexity": 46.682628117544006, "lr": 0.001305, "grad_norm": 0.243776, "tokens_per_sec": 106404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:24.028348+00:00", "epoch": 0, "step": 1674, "train_loss": 3.8217825889587402, "perplexity": 45.6855743786458, "lr": 0.001305, "grad_norm": 0.256883, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:24.334785+00:00", "epoch": 0, "step": 1675, "train_loss": 3.7095794677734375, "perplexity": 40.836629796705374, "lr": 0.001305, "grad_norm": 0.260527, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:24.651757+00:00", "epoch": 0, "step": 1676, "train_loss": 3.7548279762268066, "perplexity": 42.72686914107919, "lr": 0.001305, "grad_norm": 0.220072, "tokens_per_sec": 103379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:24.969000+00:00", "epoch": 0, "step": 1677, "train_loss": 3.857635021209717, "perplexity": 47.35322945482646, "lr": 0.001305, "grad_norm": 0.239105, "tokens_per_sec": 103289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:25.275631+00:00", "epoch": 0, "step": 1678, "train_loss": 3.9048919677734375, "perplexity": 49.64471639321689, "lr": 0.001305, "grad_norm": 0.229301, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:25.595309+00:00", "epoch": 0, "step": 1679, "train_loss": 3.748147487640381, "perplexity": 42.44238408704786, "lr": 0.001305, "grad_norm": 0.224025, "tokens_per_sec": 102503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:25.901629+00:00", "epoch": 0, "step": 1680, "train_loss": 3.835843086242676, "perplexity": 46.33247347712307, "lr": 0.001305, "grad_norm": 0.249236, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:26.207532+00:00", "epoch": 0, "step": 1681, "train_loss": 3.7275679111480713, "perplexity": 41.577864036966986, "lr": 0.001305, "grad_norm": 0.245269, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:26.524948+00:00", "epoch": 0, "step": 1682, "train_loss": 3.661193609237671, "perplexity": 38.9077558231256, "lr": 0.001305, "grad_norm": 0.281496, "tokens_per_sec": 103234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:26.831435+00:00", "epoch": 0, "step": 1683, "train_loss": 3.756483554840088, "perplexity": 42.797665420073635, "lr": 0.001305, "grad_norm": 0.2995, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:27.151439+00:00", "epoch": 0, "step": 1684, "train_loss": 3.77846097946167, "perplexity": 43.74865981181686, "lr": 0.001305, "grad_norm": 0.277943, "tokens_per_sec": 102400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:27.458346+00:00", "epoch": 0, "step": 1685, "train_loss": 3.7509803771972656, "perplexity": 42.562789140282796, "lr": 0.001305, "grad_norm": 0.25445, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:27.776262+00:00", "epoch": 0, "step": 1686, "train_loss": 3.774264335632324, "perplexity": 43.565446976974386, "lr": 0.001305, "grad_norm": 0.296043, "tokens_per_sec": 103072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:28.084546+00:00", "epoch": 0, "step": 1687, "train_loss": 3.8737118244171143, "perplexity": 48.12067047557877, "lr": 0.001305, "grad_norm": 0.341866, "tokens_per_sec": 106291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:28.406062+00:00", "epoch": 0, "step": 1688, "train_loss": 3.7971737384796143, "perplexity": 44.57502561861539, "lr": 0.001305, "grad_norm": 0.343767, "tokens_per_sec": 101917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:28.712318+00:00", "epoch": 0, "step": 1689, "train_loss": 3.810051918029785, "perplexity": 45.15278304900074, "lr": 0.001305, "grad_norm": 0.301097, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:29.018539+00:00", "epoch": 0, "step": 1690, "train_loss": 3.7905919551849365, "perplexity": 44.28260583708702, "lr": 0.001305, "grad_norm": 0.254308, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:29.337623+00:00", "epoch": 0, "step": 1691, "train_loss": 3.787558078765869, "perplexity": 44.14846147506462, "lr": 0.001305, "grad_norm": 0.299397, "tokens_per_sec": 102646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:29.645461+00:00", "epoch": 0, "step": 1692, "train_loss": 3.6867504119873047, "perplexity": 39.9149289070394, "lr": 0.001305, "grad_norm": 0.283544, "tokens_per_sec": 106446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:29.965472+00:00", "epoch": 0, "step": 1693, "train_loss": 3.721576452255249, "perplexity": 41.32949675789117, "lr": 0.001305, "grad_norm": 0.291637, "tokens_per_sec": 102396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:30.272870+00:00", "epoch": 0, "step": 1694, "train_loss": 3.7421536445617676, "perplexity": 42.18875197183211, "lr": 0.001305, "grad_norm": 0.299412, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:30.579228+00:00", "epoch": 0, "step": 1695, "train_loss": 3.775137186050415, "perplexity": 43.60348969596885, "lr": 0.001305, "grad_norm": 0.283512, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:30.897236+00:00", "epoch": 0, "step": 1696, "train_loss": 3.8475353717803955, "perplexity": 46.87738541177863, "lr": 0.001305, "grad_norm": 0.256037, "tokens_per_sec": 103041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:31.213500+00:00", "epoch": 0, "step": 1697, "train_loss": 3.7861995697021484, "perplexity": 44.08852611058715, "lr": 0.001305, "grad_norm": 0.29404, "tokens_per_sec": 103610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:31.519679+00:00", "epoch": 0, "step": 1698, "train_loss": 3.801178455352783, "perplexity": 44.75389389518105, "lr": 0.001305, "grad_norm": 0.266701, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:31.837196+00:00", "epoch": 0, "step": 1699, "train_loss": 3.7530455589294434, "perplexity": 42.65077986202822, "lr": 0.001305, "grad_norm": 0.275934, "tokens_per_sec": 103201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:32.157777+00:00", "epoch": 0, "step": 1700, "train_loss": 3.832484006881714, "perplexity": 46.177100123531524, "lr": 0.001305, "grad_norm": 0.276686, "tokens_per_sec": 102274} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:38:37.735416+00:00", "step": 1700, "epoch": 0, "val_loss": 3.84590859413147, "val_ppl": 46.80118832361468, "eval_train_loss": 3.832484006881714, "eval_train_ppl": 46.177100123531524} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:38:38.777391+00:00", "step": 1700, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p8459_epoch_0000_step_0001700.pt", "val_loss": 3.84590859413147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:39.851533+00:00", "epoch": 0, "step": 1701, "train_loss": 3.83774471282959, "perplexity": 46.42066436701893, "lr": 0.001305, "grad_norm": 0.300421, "tokens_per_sec": 4259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:40.156417+00:00", "epoch": 0, "step": 1702, "train_loss": 3.854249954223633, "perplexity": 47.19320659796354, "lr": 0.001305, "grad_norm": 0.274406, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:40.462210+00:00", "epoch": 0, "step": 1703, "train_loss": 3.780782699584961, "perplexity": 43.85034995801928, "lr": 0.001305, "grad_norm": 0.263057, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:40.768295+00:00", "epoch": 0, "step": 1704, "train_loss": 3.833446502685547, "perplexity": 46.22156678468817, "lr": 0.001305, "grad_norm": 0.286268, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:41.074528+00:00", "epoch": 0, "step": 1705, "train_loss": 3.8392786979675293, "perplexity": 46.491927620663006, "lr": 0.001305, "grad_norm": 0.302909, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:41.380923+00:00", "epoch": 0, "step": 1706, "train_loss": 3.834937572479248, "perplexity": 46.290537774239404, "lr": 0.001305, "grad_norm": 0.32168, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:41.687233+00:00", "epoch": 0, "step": 1707, "train_loss": 3.8115122318267822, "perplexity": 45.218768449028794, "lr": 0.001305, "grad_norm": 0.323912, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:41.993118+00:00", "epoch": 0, "step": 1708, "train_loss": 3.8256025314331055, "perplexity": 45.86042439055715, "lr": 0.001305, "grad_norm": 0.27318, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:42.298562+00:00", "epoch": 0, "step": 1709, "train_loss": 3.885016679763794, "perplexity": 48.66775421936518, "lr": 0.001305, "grad_norm": 0.275973, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:42.607113+00:00", "epoch": 0, "step": 1710, "train_loss": 3.8247900009155273, "perplexity": 45.82317653075141, "lr": 0.001305, "grad_norm": 0.272036, "tokens_per_sec": 106255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:42.914637+00:00", "epoch": 0, "step": 1711, "train_loss": 3.855078935623169, "perplexity": 47.23234510872507, "lr": 0.001305, "grad_norm": 0.260185, "tokens_per_sec": 106503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:43.326679+00:00", "epoch": 0, "step": 1712, "train_loss": 3.793567419052124, "perplexity": 44.41456335075126, "lr": 0.001305, "grad_norm": 0.253903, "tokens_per_sec": 79525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:43.633918+00:00", "epoch": 0, "step": 1713, "train_loss": 3.7744483947753906, "perplexity": 43.57346633380761, "lr": 0.001305, "grad_norm": 0.25167, "tokens_per_sec": 106652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:43.942609+00:00", "epoch": 0, "step": 1714, "train_loss": 3.8502206802368164, "perplexity": 47.003434816261745, "lr": 0.001305, "grad_norm": 0.242787, "tokens_per_sec": 106152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:44.249511+00:00", "epoch": 0, "step": 1715, "train_loss": 3.860398530960083, "perplexity": 47.48427155079395, "lr": 0.001305, "grad_norm": 0.251102, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:44.556453+00:00", "epoch": 0, "step": 1716, "train_loss": 3.819728136062622, "perplexity": 45.591811866364445, "lr": 0.001305, "grad_norm": 0.260394, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:44.862629+00:00", "epoch": 0, "step": 1717, "train_loss": 3.7547991275787354, "perplexity": 42.72563654644758, "lr": 0.001305, "grad_norm": 0.254382, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:45.168992+00:00", "epoch": 0, "step": 1718, "train_loss": 3.8386502265930176, "perplexity": 46.462717954688436, "lr": 0.001305, "grad_norm": 0.234238, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:45.476697+00:00", "epoch": 0, "step": 1719, "train_loss": 3.7710721492767334, "perplexity": 43.426599682683545, "lr": 0.001305, "grad_norm": 0.225353, "tokens_per_sec": 106493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:45.784434+00:00", "epoch": 0, "step": 1720, "train_loss": 3.7626991271972656, "perplexity": 43.064505830071, "lr": 0.001305, "grad_norm": 0.234234, "tokens_per_sec": 106537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:46.091916+00:00", "epoch": 0, "step": 1721, "train_loss": 3.8114116191864014, "perplexity": 45.214219098205305, "lr": 0.001305, "grad_norm": 0.244138, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:46.398690+00:00", "epoch": 0, "step": 1722, "train_loss": 3.7409563064575195, "perplexity": 42.138268000747914, "lr": 0.001305, "grad_norm": 0.254581, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:46.705496+00:00", "epoch": 0, "step": 1723, "train_loss": 3.7553482055664062, "perplexity": 42.74910269476765, "lr": 0.001305, "grad_norm": 0.243253, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:47.012734+00:00", "epoch": 0, "step": 1724, "train_loss": 3.768052816390991, "perplexity": 43.29567806957244, "lr": 0.001305, "grad_norm": 0.272796, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:47.319938+00:00", "epoch": 0, "step": 1725, "train_loss": 3.840934991836548, "perplexity": 46.56899572143289, "lr": 0.001305, "grad_norm": 0.278485, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:47.626926+00:00", "epoch": 0, "step": 1726, "train_loss": 3.7834692001342773, "perplexity": 43.96831232930952, "lr": 0.001305, "grad_norm": 0.259797, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:47.933694+00:00", "epoch": 0, "step": 1727, "train_loss": 3.7727270126342773, "perplexity": 43.49852426750689, "lr": 0.001305, "grad_norm": 0.224375, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:48.240657+00:00", "epoch": 0, "step": 1728, "train_loss": 3.706272602081299, "perplexity": 40.70181158236904, "lr": 0.001305, "grad_norm": 0.235505, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:48.547907+00:00", "epoch": 0, "step": 1729, "train_loss": 3.874521255493164, "perplexity": 48.15963660973453, "lr": 0.001305, "grad_norm": 0.241783, "tokens_per_sec": 106650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:48.855679+00:00", "epoch": 0, "step": 1730, "train_loss": 3.773869514465332, "perplexity": 43.548249811484276, "lr": 0.001305, "grad_norm": 0.221806, "tokens_per_sec": 106522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:49.162479+00:00", "epoch": 0, "step": 1731, "train_loss": 3.8507680892944336, "perplexity": 47.02917196595102, "lr": 0.001305, "grad_norm": 0.239841, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:49.468645+00:00", "epoch": 0, "step": 1732, "train_loss": 3.87872576713562, "perplexity": 48.36255064083498, "lr": 0.001305, "grad_norm": 0.248019, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:49.775951+00:00", "epoch": 0, "step": 1733, "train_loss": 3.8613293170928955, "perplexity": 47.52848982796731, "lr": 0.001305, "grad_norm": 0.264413, "tokens_per_sec": 106630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:50.083079+00:00", "epoch": 0, "step": 1734, "train_loss": 3.824700117111206, "perplexity": 45.81905795441817, "lr": 0.001305, "grad_norm": 0.290215, "tokens_per_sec": 106693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:50.390871+00:00", "epoch": 0, "step": 1735, "train_loss": 3.7979378700256348, "perplexity": 44.609099818782894, "lr": 0.001305, "grad_norm": 0.294973, "tokens_per_sec": 106461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:50.698037+00:00", "epoch": 0, "step": 1736, "train_loss": 3.7425880432128906, "perplexity": 42.20708268991259, "lr": 0.001305, "grad_norm": 0.269039, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:51.015406+00:00", "epoch": 0, "step": 1737, "train_loss": 3.8360819816589355, "perplexity": 46.343543414886774, "lr": 0.001305, "grad_norm": 0.272074, "tokens_per_sec": 103249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:51.323079+00:00", "epoch": 0, "step": 1738, "train_loss": 3.75844144821167, "perplexity": 42.88154076821926, "lr": 0.001305, "grad_norm": 0.244055, "tokens_per_sec": 106504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:51.642743+00:00", "epoch": 0, "step": 1739, "train_loss": 3.8742051124572754, "perplexity": 48.14441368244744, "lr": 0.001305, "grad_norm": 0.246324, "tokens_per_sec": 102508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:51.950451+00:00", "epoch": 0, "step": 1740, "train_loss": 3.725034713745117, "perplexity": 41.47267239162157, "lr": 0.001305, "grad_norm": 0.254463, "tokens_per_sec": 106550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:52.257999+00:00", "epoch": 0, "step": 1741, "train_loss": 3.719045639038086, "perplexity": 41.22503176768489, "lr": 0.001305, "grad_norm": 0.292067, "tokens_per_sec": 106487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:52.573333+00:00", "epoch": 0, "step": 1742, "train_loss": 3.753962278366089, "perplexity": 42.68989658770781, "lr": 0.001305, "grad_norm": 0.256041, "tokens_per_sec": 103916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:52.879454+00:00", "epoch": 0, "step": 1743, "train_loss": 3.767939329147339, "perplexity": 43.290764841205906, "lr": 0.001305, "grad_norm": 0.254569, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:53.197997+00:00", "epoch": 0, "step": 1744, "train_loss": 3.8925397396087646, "perplexity": 49.03526531755411, "lr": 0.001305, "grad_norm": 0.275704, "tokens_per_sec": 102869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:53.517757+00:00", "epoch": 0, "step": 1745, "train_loss": 3.7883152961730957, "perplexity": 44.18190411867381, "lr": 0.001305, "grad_norm": 0.262147, "tokens_per_sec": 102477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:53.824480+00:00", "epoch": 0, "step": 1746, "train_loss": 3.768278121948242, "perplexity": 43.30543392542946, "lr": 0.001305, "grad_norm": 0.270625, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:54.141172+00:00", "epoch": 0, "step": 1747, "train_loss": 3.7749855518341064, "perplexity": 43.59687841624126, "lr": 0.001305, "grad_norm": 0.228677, "tokens_per_sec": 103470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:54.449078+00:00", "epoch": 0, "step": 1748, "train_loss": 3.755479574203491, "perplexity": 42.75471895501742, "lr": 0.001305, "grad_norm": 0.234243, "tokens_per_sec": 106421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:54.766259+00:00", "epoch": 0, "step": 1749, "train_loss": 3.813730478286743, "perplexity": 45.31918615650028, "lr": 0.001305, "grad_norm": 0.231186, "tokens_per_sec": 103310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:55.072556+00:00", "epoch": 0, "step": 1750, "train_loss": 3.7427947521209717, "perplexity": 42.21580817167504, "lr": 0.001305, "grad_norm": 0.228736, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:38:55.957621+00:00", "step": 1750, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0000_step_0001750.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:57.024939+00:00", "epoch": 0, "step": 1751, "train_loss": 3.8254897594451904, "perplexity": 45.8552529109365, "lr": 0.001305, "grad_norm": 0.231868, "tokens_per_sec": 16782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:57.331691+00:00", "epoch": 0, "step": 1752, "train_loss": 3.801027774810791, "perplexity": 44.74715086222735, "lr": 0.001305, "grad_norm": 0.228757, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:57.637082+00:00", "epoch": 0, "step": 1753, "train_loss": 3.828233003616333, "perplexity": 45.98121776337978, "lr": 0.001305, "grad_norm": 0.236642, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:57.943475+00:00", "epoch": 0, "step": 1754, "train_loss": 3.7625560760498047, "perplexity": 43.05834584370423, "lr": 0.001305, "grad_norm": 0.232218, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:58.249389+00:00", "epoch": 0, "step": 1755, "train_loss": 3.804924488067627, "perplexity": 44.92185784864488, "lr": 0.001305, "grad_norm": 0.23989, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:58.556394+00:00", "epoch": 0, "step": 1756, "train_loss": 3.8817336559295654, "perplexity": 48.5082388120214, "lr": 0.001305, "grad_norm": 0.240306, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:58.863175+00:00", "epoch": 0, "step": 1757, "train_loss": 3.725132942199707, "perplexity": 41.47674638822623, "lr": 0.001305, "grad_norm": 0.274704, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:59.168915+00:00", "epoch": 0, "step": 1758, "train_loss": 3.759143829345703, "perplexity": 42.91167053350381, "lr": 0.001305, "grad_norm": 0.290722, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:59.474794+00:00", "epoch": 0, "step": 1759, "train_loss": 3.794999599456787, "perplexity": 44.478218590076594, "lr": 0.001305, "grad_norm": 0.267438, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:38:59.780783+00:00", "epoch": 0, "step": 1760, "train_loss": 3.9441306591033936, "perplexity": 51.63143327963106, "lr": 0.001305, "grad_norm": 0.241903, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:00.087525+00:00", "epoch": 0, "step": 1761, "train_loss": 3.793490409851074, "perplexity": 44.4111431524077, "lr": 0.001305, "grad_norm": 0.243881, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:00.395761+00:00", "epoch": 0, "step": 1762, "train_loss": 3.76893949508667, "perplexity": 43.33408444947485, "lr": 0.001305, "grad_norm": 0.245074, "tokens_per_sec": 106308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:00.703718+00:00", "epoch": 0, "step": 1763, "train_loss": 3.8909456729888916, "perplexity": 48.95716210531522, "lr": 0.001305, "grad_norm": 0.261122, "tokens_per_sec": 106406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:01.010251+00:00", "epoch": 0, "step": 1764, "train_loss": 3.860126495361328, "perplexity": 47.47135589538987, "lr": 0.001305, "grad_norm": 0.256808, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:01.317470+00:00", "epoch": 0, "step": 1765, "train_loss": 3.787278413772583, "perplexity": 44.13611642220263, "lr": 0.001305, "grad_norm": 0.256065, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:01.624498+00:00", "epoch": 0, "step": 1766, "train_loss": 3.7450919151306152, "perplexity": 42.312896235489525, "lr": 0.001305, "grad_norm": 0.258559, "tokens_per_sec": 106727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:01.931491+00:00", "epoch": 0, "step": 1767, "train_loss": 3.8931076526641846, "perplexity": 49.063120993956396, "lr": 0.001305, "grad_norm": 0.273861, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:02.238181+00:00", "epoch": 0, "step": 1768, "train_loss": 3.820779800415039, "perplexity": 45.63978437073131, "lr": 0.001305, "grad_norm": 0.256409, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:02.545347+00:00", "epoch": 0, "step": 1769, "train_loss": 3.8392891883850098, "perplexity": 46.49241534295142, "lr": 0.001305, "grad_norm": 0.283976, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:02.851578+00:00", "epoch": 0, "step": 1770, "train_loss": 3.8131120204925537, "perplexity": 45.291166817877745, "lr": 0.001305, "grad_norm": 0.30439, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:03.157647+00:00", "epoch": 0, "step": 1771, "train_loss": 3.7801883220672607, "perplexity": 43.82429404015394, "lr": 0.001305, "grad_norm": 0.283493, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:03.463930+00:00", "epoch": 0, "step": 1772, "train_loss": 3.7792980670928955, "perplexity": 43.78529660579224, "lr": 0.001305, "grad_norm": 0.280929, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:03.770817+00:00", "epoch": 0, "step": 1773, "train_loss": 3.7674460411071777, "perplexity": 43.269415290832164, "lr": 0.001305, "grad_norm": 0.294472, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:04.077402+00:00", "epoch": 0, "step": 1774, "train_loss": 3.7784264087677, "perplexity": 43.74714741642935, "lr": 0.001305, "grad_norm": 0.261539, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:04.384353+00:00", "epoch": 0, "step": 1775, "train_loss": 3.9224040508270264, "perplexity": 50.52175576948532, "lr": 0.001305, "grad_norm": 0.260631, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:04.690842+00:00", "epoch": 0, "step": 1776, "train_loss": 3.749499797821045, "perplexity": 42.49981818074256, "lr": 0.001305, "grad_norm": 0.230884, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:04.997724+00:00", "epoch": 0, "step": 1777, "train_loss": 3.896920680999756, "perplexity": 49.25055718719848, "lr": 0.001305, "grad_norm": 0.265593, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:05.304352+00:00", "epoch": 0, "step": 1778, "train_loss": 3.894129753112793, "perplexity": 49.11329406852752, "lr": 0.001305, "grad_norm": 0.260273, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:05.611019+00:00", "epoch": 0, "step": 1779, "train_loss": 3.934412717819214, "perplexity": 51.13211215885215, "lr": 0.001305, "grad_norm": 0.247886, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:05.918239+00:00", "epoch": 0, "step": 1780, "train_loss": 3.7073986530303955, "perplexity": 40.74766971037997, "lr": 0.001305, "grad_norm": 0.237017, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:06.224794+00:00", "epoch": 0, "step": 1781, "train_loss": 3.808772087097168, "perplexity": 45.09503208417931, "lr": 0.001305, "grad_norm": 0.270186, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:06.531270+00:00", "epoch": 0, "step": 1782, "train_loss": 3.709892749786377, "perplexity": 40.8494251824672, "lr": 0.001305, "grad_norm": 0.295414, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:06.837044+00:00", "epoch": 0, "step": 1783, "train_loss": 3.868144989013672, "perplexity": 47.85353486339935, "lr": 0.001305, "grad_norm": 0.329089, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:07.144610+00:00", "epoch": 0, "step": 1784, "train_loss": 3.8473663330078125, "perplexity": 46.869461985788696, "lr": 0.001305, "grad_norm": 0.301252, "tokens_per_sec": 106541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:07.451651+00:00", "epoch": 0, "step": 1785, "train_loss": 3.7947194576263428, "perplexity": 44.465760125655855, "lr": 0.001305, "grad_norm": 0.263752, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:07.758521+00:00", "epoch": 0, "step": 1786, "train_loss": 3.7595579624176025, "perplexity": 42.929445355759015, "lr": 0.001305, "grad_norm": 0.29871, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:08.064231+00:00", "epoch": 0, "step": 1787, "train_loss": 3.8203492164611816, "perplexity": 45.62013684218258, "lr": 0.001305, "grad_norm": 0.321045, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:08.380272+00:00", "epoch": 0, "step": 1788, "train_loss": 3.814171075820923, "perplexity": 45.33915807763838, "lr": 0.001305, "grad_norm": 0.317575, "tokens_per_sec": 103683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:08.686438+00:00", "epoch": 0, "step": 1789, "train_loss": 3.7684953212738037, "perplexity": 43.31484085802394, "lr": 0.001305, "grad_norm": 0.324162, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:08.994932+00:00", "epoch": 0, "step": 1790, "train_loss": 3.7950525283813477, "perplexity": 44.48057283665626, "lr": 0.001305, "grad_norm": 0.271353, "tokens_per_sec": 106282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:09.302568+00:00", "epoch": 0, "step": 1791, "train_loss": 3.69059681892395, "perplexity": 40.06875361302018, "lr": 0.001305, "grad_norm": 0.305192, "tokens_per_sec": 106454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:09.610483+00:00", "epoch": 0, "step": 1792, "train_loss": 3.9258577823638916, "perplexity": 50.69654601623267, "lr": 0.001305, "grad_norm": 0.288239, "tokens_per_sec": 106420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:09.917313+00:00", "epoch": 0, "step": 1793, "train_loss": 3.8279800415039062, "perplexity": 45.96958772844352, "lr": 0.001305, "grad_norm": 0.275371, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:10.224038+00:00", "epoch": 0, "step": 1794, "train_loss": 3.8485145568847656, "perplexity": 46.92330952974244, "lr": 0.001305, "grad_norm": 0.258472, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:10.530545+00:00", "epoch": 0, "step": 1795, "train_loss": 3.7713534832000732, "perplexity": 43.438818777091875, "lr": 0.001305, "grad_norm": 0.277973, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:10.836788+00:00", "epoch": 0, "step": 1796, "train_loss": 3.6656298637390137, "perplexity": 39.08074395601915, "lr": 0.001305, "grad_norm": 0.243139, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:11.143487+00:00", "epoch": 0, "step": 1797, "train_loss": 3.7620370388031006, "perplexity": 43.03600275737849, "lr": 0.001305, "grad_norm": 0.250467, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:11.450714+00:00", "epoch": 0, "step": 1798, "train_loss": 3.7648470401763916, "perplexity": 43.15710405194397, "lr": 0.001305, "grad_norm": 0.230578, "tokens_per_sec": 106657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:11.756289+00:00", "epoch": 0, "step": 1799, "train_loss": 3.803454637527466, "perplexity": 44.85587793380666, "lr": 0.001305, "grad_norm": 0.240406, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:12.063074+00:00", "epoch": 0, "step": 1800, "train_loss": 3.739260196685791, "perplexity": 42.06685744981417, "lr": 0.001305, "grad_norm": 0.244393, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:39:17.610713+00:00", "step": 1800, "epoch": 0, "val_loss": 3.8228560209274294, "val_ppl": 45.734641064855104, "eval_train_loss": 3.739260196685791, "eval_train_ppl": 42.06685744981417} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:39:18.473525+00:00", "step": 1800, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p8229_epoch_0000_step_0001800.pt", "val_loss": 3.8228560209274294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:19.535003+00:00", "epoch": 0, "step": 1801, "train_loss": 3.651341438293457, "perplexity": 38.52631207162991, "lr": 0.001305, "grad_norm": 0.229478, "tokens_per_sec": 4385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:19.840241+00:00", "epoch": 0, "step": 1802, "train_loss": 3.7593438625335693, "perplexity": 42.9202551503328, "lr": 0.001305, "grad_norm": 0.255018, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:20.145879+00:00", "epoch": 0, "step": 1803, "train_loss": 3.9269070625305176, "perplexity": 50.74976881441408, "lr": 0.001305, "grad_norm": 0.26556, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:20.451841+00:00", "epoch": 0, "step": 1804, "train_loss": 3.748736619949341, "perplexity": 42.46739563361451, "lr": 0.001305, "grad_norm": 0.267553, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:20.757581+00:00", "epoch": 0, "step": 1805, "train_loss": 3.8110737800598145, "perplexity": 45.19894654589399, "lr": 0.001305, "grad_norm": 0.297853, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:21.063911+00:00", "epoch": 0, "step": 1806, "train_loss": 3.7404699325561523, "perplexity": 42.117778030240906, "lr": 0.001305, "grad_norm": 0.27186, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:21.369497+00:00", "epoch": 0, "step": 1807, "train_loss": 3.774958372116089, "perplexity": 43.59569348148263, "lr": 0.001305, "grad_norm": 0.247394, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:21.674961+00:00", "epoch": 0, "step": 1808, "train_loss": 3.7518231868743896, "perplexity": 42.59867659187704, "lr": 0.001305, "grad_norm": 0.252076, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:21.980483+00:00", "epoch": 0, "step": 1809, "train_loss": 3.8013808727264404, "perplexity": 44.76295377775213, "lr": 0.001305, "grad_norm": 0.282127, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:22.285378+00:00", "epoch": 0, "step": 1810, "train_loss": 3.8659303188323975, "perplexity": 47.74767233523025, "lr": 0.001305, "grad_norm": 0.340008, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:22.591174+00:00", "epoch": 0, "step": 1811, "train_loss": 3.7630491256713867, "perplexity": 43.07958097938604, "lr": 0.001305, "grad_norm": 0.300857, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:22.897804+00:00", "epoch": 0, "step": 1812, "train_loss": 3.7487289905548096, "perplexity": 42.46707163433447, "lr": 0.001305, "grad_norm": 0.285603, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:23.204165+00:00", "epoch": 0, "step": 1813, "train_loss": 3.611818313598633, "perplexity": 37.03332983817039, "lr": 0.001305, "grad_norm": 0.309462, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:23.509960+00:00", "epoch": 0, "step": 1814, "train_loss": 3.7543418407440186, "perplexity": 42.706103141874756, "lr": 0.001305, "grad_norm": 0.256255, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:23.870150+00:00", "epoch": 0, "step": 1815, "train_loss": 3.83351469039917, "perplexity": 46.22471863510481, "lr": 0.001305, "grad_norm": 0.251839, "tokens_per_sec": 90974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:24.176083+00:00", "epoch": 0, "step": 1816, "train_loss": 3.677241563796997, "perplexity": 39.53718271936834, "lr": 0.001305, "grad_norm": 0.235651, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:24.481141+00:00", "epoch": 0, "step": 1817, "train_loss": 3.741346597671509, "perplexity": 42.15471740634176, "lr": 0.001305, "grad_norm": 0.23283, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:24.786093+00:00", "epoch": 0, "step": 1818, "train_loss": 3.6522669792175293, "perplexity": 38.56198625651819, "lr": 0.001305, "grad_norm": 0.257101, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:25.091453+00:00", "epoch": 0, "step": 1819, "train_loss": 3.7661044597625732, "perplexity": 43.2114047721017, "lr": 0.001305, "grad_norm": 0.291119, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:25.398339+00:00", "epoch": 0, "step": 1820, "train_loss": 3.7745203971862793, "perplexity": 43.57660384138712, "lr": 0.001305, "grad_norm": 0.296857, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:25.703398+00:00", "epoch": 0, "step": 1821, "train_loss": 3.9547324180603027, "perplexity": 52.18172918784383, "lr": 0.001305, "grad_norm": 0.274924, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:26.008909+00:00", "epoch": 0, "step": 1822, "train_loss": 3.816911458969116, "perplexity": 45.463575139736555, "lr": 0.001305, "grad_norm": 0.252466, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:26.314312+00:00", "epoch": 0, "step": 1823, "train_loss": 3.8088748455047607, "perplexity": 45.099666215960546, "lr": 0.001305, "grad_norm": 0.266285, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:26.620468+00:00", "epoch": 0, "step": 1824, "train_loss": 3.7986512184143066, "perplexity": 44.64093300098344, "lr": 0.001305, "grad_norm": 0.2689, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:26.925445+00:00", "epoch": 0, "step": 1825, "train_loss": 3.776418685913086, "perplexity": 43.65940338106411, "lr": 0.001305, "grad_norm": 0.257596, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:27.231590+00:00", "epoch": 0, "step": 1826, "train_loss": 3.7393462657928467, "perplexity": 42.07047826248933, "lr": 0.001305, "grad_norm": 0.307613, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:27.537219+00:00", "epoch": 0, "step": 1827, "train_loss": 3.827486038208008, "perplexity": 45.94688420886304, "lr": 0.001305, "grad_norm": 0.291394, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:27.843111+00:00", "epoch": 0, "step": 1828, "train_loss": 3.7315561771392822, "perplexity": 41.74401873241171, "lr": 0.001305, "grad_norm": 0.246646, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:28.148220+00:00", "epoch": 0, "step": 1829, "train_loss": 3.8884074687957764, "perplexity": 48.833056400615554, "lr": 0.001305, "grad_norm": 0.263118, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:28.453933+00:00", "epoch": 0, "step": 1830, "train_loss": 3.7994937896728516, "perplexity": 44.67856201842283, "lr": 0.001305, "grad_norm": 0.268953, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:28.760506+00:00", "epoch": 0, "step": 1831, "train_loss": 3.7437899112701416, "perplexity": 42.25784053039099, "lr": 0.001305, "grad_norm": 0.251387, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:29.066919+00:00", "epoch": 0, "step": 1832, "train_loss": 3.8548905849456787, "perplexity": 47.2234497022786, "lr": 0.001305, "grad_norm": 0.246547, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:29.371713+00:00", "epoch": 0, "step": 1833, "train_loss": 3.841033697128296, "perplexity": 46.57359255460414, "lr": 0.001305, "grad_norm": 0.249028, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:29.677841+00:00", "epoch": 0, "step": 1834, "train_loss": 3.7282917499542236, "perplexity": 41.60797060326987, "lr": 0.001305, "grad_norm": 0.243774, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:29.983991+00:00", "epoch": 0, "step": 1835, "train_loss": 3.687469720840454, "perplexity": 39.9436503973478, "lr": 0.001305, "grad_norm": 0.252329, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:30.290990+00:00", "epoch": 0, "step": 1836, "train_loss": 3.656277894973755, "perplexity": 38.71696573182692, "lr": 0.001305, "grad_norm": 0.250083, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:30.597434+00:00", "epoch": 0, "step": 1837, "train_loss": 3.726168155670166, "perplexity": 41.51970590709792, "lr": 0.001305, "grad_norm": 0.282326, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:30.903100+00:00", "epoch": 0, "step": 1838, "train_loss": 3.723125457763672, "perplexity": 41.39356598501162, "lr": 0.001305, "grad_norm": 0.23663, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:31.208603+00:00", "epoch": 0, "step": 1839, "train_loss": 3.7909255027770996, "perplexity": 44.29737865722102, "lr": 0.001305, "grad_norm": 0.242818, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:31.514842+00:00", "epoch": 0, "step": 1840, "train_loss": 3.727097511291504, "perplexity": 41.558310415058656, "lr": 0.001305, "grad_norm": 0.252815, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:31.821100+00:00", "epoch": 0, "step": 1841, "train_loss": 3.796509265899658, "perplexity": 44.545416574629336, "lr": 0.001305, "grad_norm": 0.271685, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:32.127628+00:00", "epoch": 0, "step": 1842, "train_loss": 3.7989768981933594, "perplexity": 44.655474017908965, "lr": 0.001305, "grad_norm": 0.252021, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:32.432788+00:00", "epoch": 0, "step": 1843, "train_loss": 3.6713569164276123, "perplexity": 39.305203568271445, "lr": 0.001305, "grad_norm": 0.238354, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:32.738121+00:00", "epoch": 0, "step": 1844, "train_loss": 3.6556813716888428, "perplexity": 38.693877047400335, "lr": 0.001305, "grad_norm": 0.247876, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:33.043394+00:00", "epoch": 0, "step": 1845, "train_loss": 3.740126132965088, "perplexity": 42.10330044421424, "lr": 0.001305, "grad_norm": 0.252517, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:33.349890+00:00", "epoch": 0, "step": 1846, "train_loss": 3.813875436782837, "perplexity": 45.32575603373798, "lr": 0.001305, "grad_norm": 0.279005, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:33.655390+00:00", "epoch": 0, "step": 1847, "train_loss": 3.8484771251678467, "perplexity": 46.92155314257572, "lr": 0.001305, "grad_norm": 0.268928, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:33.960967+00:00", "epoch": 0, "step": 1848, "train_loss": 3.7452685832977295, "perplexity": 42.3203722376791, "lr": 0.001305, "grad_norm": 0.260572, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:34.265493+00:00", "epoch": 0, "step": 1849, "train_loss": 3.79158878326416, "perplexity": 44.32676999038364, "lr": 0.001305, "grad_norm": 0.252759, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:34.570837+00:00", "epoch": 0, "step": 1850, "train_loss": 3.7231504917144775, "perplexity": 41.394602242476914, "lr": 0.001305, "grad_norm": 0.25253, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:34.876483+00:00", "epoch": 0, "step": 1851, "train_loss": 3.811356782913208, "perplexity": 45.21173978691333, "lr": 0.001305, "grad_norm": 0.235798, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:35.181740+00:00", "epoch": 0, "step": 1852, "train_loss": 3.7975730895996094, "perplexity": 44.5928302599351, "lr": 0.001305, "grad_norm": 0.247573, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:35.487451+00:00", "epoch": 0, "step": 1853, "train_loss": 3.7626471519470215, "perplexity": 43.062267599770635, "lr": 0.001305, "grad_norm": 0.239765, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:35.793651+00:00", "epoch": 0, "step": 1854, "train_loss": 3.713846445083618, "perplexity": 41.011251057004536, "lr": 0.001305, "grad_norm": 0.254978, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:36.100157+00:00", "epoch": 0, "step": 1855, "train_loss": 3.6929209232330322, "perplexity": 40.16198587473648, "lr": 0.001305, "grad_norm": 0.270131, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:36.405919+00:00", "epoch": 0, "step": 1856, "train_loss": 3.7085916996002197, "perplexity": 40.79631258879619, "lr": 0.001305, "grad_norm": 0.27747, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:36.711867+00:00", "epoch": 0, "step": 1857, "train_loss": 3.896883487701416, "perplexity": 49.248725430596366, "lr": 0.001305, "grad_norm": 0.301648, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:37.017357+00:00", "epoch": 0, "step": 1858, "train_loss": 3.8241240978240967, "perplexity": 45.792672893200866, "lr": 0.001305, "grad_norm": 0.285776, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:37.323322+00:00", "epoch": 0, "step": 1859, "train_loss": 3.8050718307495117, "perplexity": 44.928477243303256, "lr": 0.001305, "grad_norm": 0.298415, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:37.628614+00:00", "epoch": 0, "step": 1860, "train_loss": 3.6941494941711426, "perplexity": 40.211358045796096, "lr": 0.001305, "grad_norm": 0.296181, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:37.933875+00:00", "epoch": 0, "step": 1861, "train_loss": 3.760139226913452, "perplexity": 42.954405971828, "lr": 0.001305, "grad_norm": 0.263444, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:38.239903+00:00", "epoch": 0, "step": 1862, "train_loss": 3.695066452026367, "perplexity": 40.24824707668404, "lr": 0.001305, "grad_norm": 0.28962, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:38.546213+00:00", "epoch": 0, "step": 1863, "train_loss": 3.8103744983673096, "perplexity": 45.1673507985053, "lr": 0.001305, "grad_norm": 0.305935, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:38.852516+00:00", "epoch": 0, "step": 1864, "train_loss": 3.7581825256347656, "perplexity": 42.870439206466436, "lr": 0.001305, "grad_norm": 0.321865, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:39.158158+00:00", "epoch": 0, "step": 1865, "train_loss": 3.7714948654174805, "perplexity": 43.44496068778037, "lr": 0.001305, "grad_norm": 0.280165, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:39.463577+00:00", "epoch": 0, "step": 1866, "train_loss": 3.788025140762329, "perplexity": 44.169086359796786, "lr": 0.001305, "grad_norm": 0.249144, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:39.769269+00:00", "epoch": 0, "step": 1867, "train_loss": 3.9140050411224365, "perplexity": 50.099200061270345, "lr": 0.001305, "grad_norm": 0.260587, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:40.075175+00:00", "epoch": 0, "step": 1868, "train_loss": 3.7331430912017822, "perplexity": 41.81031549248096, "lr": 0.001305, "grad_norm": 0.236811, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:40.381515+00:00", "epoch": 0, "step": 1869, "train_loss": 3.7490170001983643, "perplexity": 42.47930432198012, "lr": 0.001305, "grad_norm": 0.239678, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:40.688310+00:00", "epoch": 0, "step": 1870, "train_loss": 3.905289649963379, "perplexity": 49.66446313895546, "lr": 0.001305, "grad_norm": 0.232191, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:40.993725+00:00", "epoch": 0, "step": 1871, "train_loss": 3.787771224975586, "perplexity": 44.15787255522526, "lr": 0.001305, "grad_norm": 0.228981, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:41.300405+00:00", "epoch": 0, "step": 1872, "train_loss": 3.7177038192749023, "perplexity": 41.16975230115562, "lr": 0.001305, "grad_norm": 0.260347, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:41.605753+00:00", "epoch": 0, "step": 1873, "train_loss": 3.706839084625244, "perplexity": 40.72487498002673, "lr": 0.001305, "grad_norm": 0.238372, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:41.912299+00:00", "epoch": 0, "step": 1874, "train_loss": 3.8067681789398193, "perplexity": 45.004756263965845, "lr": 0.001305, "grad_norm": 0.252753, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:42.218937+00:00", "epoch": 0, "step": 1875, "train_loss": 3.632856607437134, "perplexity": 37.820701348404974, "lr": 0.001305, "grad_norm": 0.250155, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:42.524400+00:00", "epoch": 0, "step": 1876, "train_loss": 3.7332797050476074, "perplexity": 41.81602775065345, "lr": 0.001305, "grad_norm": 0.263493, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:42.830187+00:00", "epoch": 0, "step": 1877, "train_loss": 3.64597749710083, "perplexity": 38.32021244659879, "lr": 0.001305, "grad_norm": 0.258988, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:43.135331+00:00", "epoch": 0, "step": 1878, "train_loss": 3.730966567993164, "perplexity": 41.71941333166883, "lr": 0.001305, "grad_norm": 0.263791, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:43.440422+00:00", "epoch": 0, "step": 1879, "train_loss": 3.809051513671875, "perplexity": 45.107634595187136, "lr": 0.001305, "grad_norm": 0.246014, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:43.746090+00:00", "epoch": 0, "step": 1880, "train_loss": 3.835766553878784, "perplexity": 46.3289276790888, "lr": 0.001305, "grad_norm": 0.226768, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:44.052040+00:00", "epoch": 0, "step": 1881, "train_loss": 3.6925320625305176, "perplexity": 40.14637149280137, "lr": 0.001305, "grad_norm": 0.258991, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:44.358076+00:00", "epoch": 0, "step": 1882, "train_loss": 3.7508583068847656, "perplexity": 42.55759380441616, "lr": 0.001305, "grad_norm": 0.247891, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:44.662939+00:00", "epoch": 0, "step": 1883, "train_loss": 3.777815818786621, "perplexity": 43.72044399976506, "lr": 0.001305, "grad_norm": 0.225767, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:44.968903+00:00", "epoch": 0, "step": 1884, "train_loss": 3.7033140659332275, "perplexity": 40.58157175615105, "lr": 0.001305, "grad_norm": 0.246002, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:45.275548+00:00", "epoch": 0, "step": 1885, "train_loss": 3.726896286010742, "perplexity": 41.549948673702495, "lr": 0.001305, "grad_norm": 0.251602, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:45.583299+00:00", "epoch": 0, "step": 1886, "train_loss": 3.7481658458709717, "perplexity": 42.44316326127386, "lr": 0.001305, "grad_norm": 0.222787, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:45.890130+00:00", "epoch": 0, "step": 1887, "train_loss": 3.8158693313598633, "perplexity": 45.416220971696724, "lr": 0.001305, "grad_norm": 0.251298, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:46.196180+00:00", "epoch": 0, "step": 1888, "train_loss": 3.735454559326172, "perplexity": 41.907070483972625, "lr": 0.001305, "grad_norm": 0.222007, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:46.501248+00:00", "epoch": 0, "step": 1889, "train_loss": 3.8061306476593018, "perplexity": 44.976073468136775, "lr": 0.001305, "grad_norm": 0.245596, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:46.808983+00:00", "epoch": 0, "step": 1890, "train_loss": 3.7931530475616455, "perplexity": 44.39616303448318, "lr": 0.001305, "grad_norm": 0.266993, "tokens_per_sec": 106539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:47.116147+00:00", "epoch": 0, "step": 1891, "train_loss": 3.730884313583374, "perplexity": 41.71598186707693, "lr": 0.001305, "grad_norm": 0.267294, "tokens_per_sec": 106623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:47.423798+00:00", "epoch": 0, "step": 1892, "train_loss": 3.7295565605163574, "perplexity": 41.66063009908093, "lr": 0.001305, "grad_norm": 0.291198, "tokens_per_sec": 106509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:47.730614+00:00", "epoch": 0, "step": 1893, "train_loss": 3.8852875232696533, "perplexity": 48.68093734974229, "lr": 0.001305, "grad_norm": 0.289483, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:48.036718+00:00", "epoch": 0, "step": 1894, "train_loss": 3.7422163486480713, "perplexity": 42.191397461917454, "lr": 0.001305, "grad_norm": 0.307213, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:48.341919+00:00", "epoch": 0, "step": 1895, "train_loss": 3.751683473587036, "perplexity": 42.592725406472994, "lr": 0.001305, "grad_norm": 0.307352, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:48.648266+00:00", "epoch": 0, "step": 1896, "train_loss": 3.7386434078216553, "perplexity": 42.04091908066087, "lr": 0.001305, "grad_norm": 0.298734, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:48.955048+00:00", "epoch": 0, "step": 1897, "train_loss": 3.839092493057251, "perplexity": 46.48327140139201, "lr": 0.001305, "grad_norm": 0.345863, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:49.261451+00:00", "epoch": 0, "step": 1898, "train_loss": 3.6018753051757812, "perplexity": 36.66693169617856, "lr": 0.001305, "grad_norm": 0.364085, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:49.567657+00:00", "epoch": 0, "step": 1899, "train_loss": 3.7167019844055176, "perplexity": 41.12852766132522, "lr": 0.001305, "grad_norm": 0.380367, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:49.874541+00:00", "epoch": 0, "step": 1900, "train_loss": 3.7915637493133545, "perplexity": 44.325660330093974, "lr": 0.001305, "grad_norm": 0.377616, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:39:55.458338+00:00", "step": 1900, "epoch": 0, "val_loss": 3.8154653549194335, "val_ppl": 45.39787759380653, "eval_train_loss": 3.7915637493133545, "eval_train_ppl": 44.325660330093974} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:39:56.355848+00:00", "step": 1900, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p8155_epoch_0000_step_0001900.pt", "val_loss": 3.8154653549194335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:57.409906+00:00", "epoch": 0, "step": 1901, "train_loss": 3.8528034687042236, "perplexity": 47.12499165586204, "lr": 0.001305, "grad_norm": 0.324244, "tokens_per_sec": 4348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:57.715412+00:00", "epoch": 0, "step": 1902, "train_loss": 3.7723875045776367, "perplexity": 43.48375867472678, "lr": 0.001305, "grad_norm": 0.276807, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:58.019889+00:00", "epoch": 0, "step": 1903, "train_loss": 3.623765707015991, "perplexity": 37.478435229388545, "lr": 0.001305, "grad_norm": 0.283661, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:58.325945+00:00", "epoch": 0, "step": 1904, "train_loss": 3.774810314178467, "perplexity": 43.58923927082691, "lr": 0.001305, "grad_norm": 0.263707, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:58.630851+00:00", "epoch": 0, "step": 1905, "train_loss": 3.857881784439087, "perplexity": 47.364915932485026, "lr": 0.001305, "grad_norm": 0.230306, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:58.937267+00:00", "epoch": 0, "step": 1906, "train_loss": 3.8404946327209473, "perplexity": 46.54849315424009, "lr": 0.001305, "grad_norm": 0.240002, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:59.243159+00:00", "epoch": 0, "step": 1907, "train_loss": 3.7581841945648193, "perplexity": 42.87051075429055, "lr": 0.001305, "grad_norm": 0.234691, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:59.549229+00:00", "epoch": 0, "step": 1908, "train_loss": 3.765151023864746, "perplexity": 43.1702251018039, "lr": 0.001305, "grad_norm": 0.250324, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:39:59.854552+00:00", "epoch": 0, "step": 1909, "train_loss": 3.7531752586364746, "perplexity": 42.6563120144325, "lr": 0.001305, "grad_norm": 0.235329, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:00.160839+00:00", "epoch": 0, "step": 1910, "train_loss": 3.7546985149383545, "perplexity": 42.72133802358918, "lr": 0.001305, "grad_norm": 0.240537, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:00.465605+00:00", "epoch": 0, "step": 1911, "train_loss": 3.7757699489593506, "perplexity": 43.63108909796669, "lr": 0.001305, "grad_norm": 0.237033, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:00.771580+00:00", "epoch": 0, "step": 1912, "train_loss": 3.6412529945373535, "perplexity": 38.139595503229145, "lr": 0.001305, "grad_norm": 0.254891, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:01.075724+00:00", "epoch": 0, "step": 1913, "train_loss": 3.742748975753784, "perplexity": 42.21387572956949, "lr": 0.001305, "grad_norm": 0.260162, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:01.381644+00:00", "epoch": 0, "step": 1914, "train_loss": 3.8183300495147705, "perplexity": 45.528115104670114, "lr": 0.001305, "grad_norm": 0.236434, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:01.687144+00:00", "epoch": 0, "step": 1915, "train_loss": 3.803983211517334, "perplexity": 44.87959385143161, "lr": 0.001305, "grad_norm": 0.241421, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:01.992568+00:00", "epoch": 0, "step": 1916, "train_loss": 3.770831823348999, "perplexity": 43.41616439881132, "lr": 0.001305, "grad_norm": 0.25456, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:02.298764+00:00", "epoch": 0, "step": 1917, "train_loss": 3.8037333488464355, "perplexity": 44.86838151707329, "lr": 0.001305, "grad_norm": 0.243219, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:02.603765+00:00", "epoch": 0, "step": 1918, "train_loss": 3.7636828422546387, "perplexity": 43.10688987638874, "lr": 0.001305, "grad_norm": 0.23199, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:02.908979+00:00", "epoch": 0, "step": 1919, "train_loss": 3.715524911880493, "perplexity": 41.08014488202666, "lr": 0.001305, "grad_norm": 0.231718, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:03.214507+00:00", "epoch": 0, "step": 1920, "train_loss": 3.8596296310424805, "perplexity": 47.447774931233106, "lr": 0.001305, "grad_norm": 0.256051, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:03.520842+00:00", "epoch": 0, "step": 1921, "train_loss": 3.6847496032714844, "perplexity": 39.835146610563534, "lr": 0.001305, "grad_norm": 0.212215, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:03.827440+00:00", "epoch": 0, "step": 1922, "train_loss": 3.7821133136749268, "perplexity": 43.90873668801151, "lr": 0.001305, "grad_norm": 0.228737, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:04.133368+00:00", "epoch": 0, "step": 1923, "train_loss": 3.789562940597534, "perplexity": 44.237061826456895, "lr": 0.001305, "grad_norm": 0.251533, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:04.438884+00:00", "epoch": 0, "step": 1924, "train_loss": 3.639517068862915, "perplexity": 38.073445432596976, "lr": 0.001305, "grad_norm": 0.285519, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:04.744342+00:00", "epoch": 0, "step": 1925, "train_loss": 3.8381526470184326, "perplexity": 46.43960480604669, "lr": 0.001305, "grad_norm": 0.275073, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:05.051614+00:00", "epoch": 0, "step": 1926, "train_loss": 3.8622071743011475, "perplexity": 47.57023137423815, "lr": 0.001305, "grad_norm": 0.234115, "tokens_per_sec": 106642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:05.357882+00:00", "epoch": 0, "step": 1927, "train_loss": 3.804337501525879, "perplexity": 44.89549706012835, "lr": 0.001305, "grad_norm": 0.243375, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:05.663842+00:00", "epoch": 0, "step": 1928, "train_loss": 3.710956573486328, "perplexity": 40.89290489237805, "lr": 0.001305, "grad_norm": 0.241481, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:05.968769+00:00", "epoch": 0, "step": 1929, "train_loss": 3.835723876953125, "perplexity": 46.326950545075675, "lr": 0.001305, "grad_norm": 0.255955, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:06.274795+00:00", "epoch": 0, "step": 1930, "train_loss": 3.8044240474700928, "perplexity": 44.899382751455356, "lr": 0.001305, "grad_norm": 0.28491, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:06.579534+00:00", "epoch": 0, "step": 1931, "train_loss": 3.7150213718414307, "perplexity": 41.05946459138221, "lr": 0.001305, "grad_norm": 0.331539, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:06.885942+00:00", "epoch": 0, "step": 1932, "train_loss": 3.7534968852996826, "perplexity": 42.670033628230605, "lr": 0.001305, "grad_norm": 0.321624, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:07.191919+00:00", "epoch": 0, "step": 1933, "train_loss": 3.785659074783325, "perplexity": 44.064702924981326, "lr": 0.001305, "grad_norm": 0.245776, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:07.498495+00:00", "epoch": 0, "step": 1934, "train_loss": 3.7972095012664795, "perplexity": 44.57661977426166, "lr": 0.001305, "grad_norm": 0.272024, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:07.803995+00:00", "epoch": 0, "step": 1935, "train_loss": 3.7923223972320557, "perplexity": 44.35930065901775, "lr": 0.001305, "grad_norm": 0.2772, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:08.109027+00:00", "epoch": 0, "step": 1936, "train_loss": 3.7054288387298584, "perplexity": 40.66748336989903, "lr": 0.001305, "grad_norm": 0.253782, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:08.415327+00:00", "epoch": 0, "step": 1937, "train_loss": 3.7503390312194824, "perplexity": 42.53550041835584, "lr": 0.001305, "grad_norm": 0.249312, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:08.721761+00:00", "epoch": 0, "step": 1938, "train_loss": 3.6821322441101074, "perplexity": 39.73102005238192, "lr": 0.001305, "grad_norm": 0.24427, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:09.026670+00:00", "epoch": 0, "step": 1939, "train_loss": 3.696392774581909, "perplexity": 40.30166465123915, "lr": 0.001305, "grad_norm": 0.237495, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:40:14.599408+00:00", "step": 1939, "epoch": 1, "val_loss": 3.80453999042511, "val_ppl": 44.904588820367685, "eval_train_loss": 3.696392774581909, "eval_train_ppl": 40.30166465123915} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:40:15.545311+00:00", "step": 1939, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p8045_epoch_0001_step_0001939.pt", "val_loss": 3.80453999042511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:16.625995+00:00", "epoch": 1, "step": 1940, "train_loss": 3.7313144207000732, "perplexity": 41.73392806687552, "lr": 0.001305, "grad_norm": 0.248956, "tokens_per_sec": 4312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:16.930857+00:00", "epoch": 1, "step": 1941, "train_loss": 3.6760435104370117, "perplexity": 39.48984342792793, "lr": 0.001305, "grad_norm": 0.244714, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:17.235268+00:00", "epoch": 1, "step": 1942, "train_loss": 3.688011407852173, "perplexity": 39.965293215256146, "lr": 0.001305, "grad_norm": 0.257134, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:17.540360+00:00", "epoch": 1, "step": 1943, "train_loss": 3.742142677307129, "perplexity": 42.188289279583586, "lr": 0.001305, "grad_norm": 0.256459, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:17.845884+00:00", "epoch": 1, "step": 1944, "train_loss": 3.6706528663635254, "perplexity": 39.27754047642373, "lr": 0.001305, "grad_norm": 0.26013, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:18.151986+00:00", "epoch": 1, "step": 1945, "train_loss": 3.7844293117523193, "perplexity": 44.01054708859974, "lr": 0.001305, "grad_norm": 0.290077, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:18.825318+00:00", "epoch": 1, "step": 1946, "train_loss": 3.6957671642303467, "perplexity": 40.276459397803954, "lr": 0.001305, "grad_norm": 0.251646, "tokens_per_sec": 48665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:19.133025+00:00", "epoch": 1, "step": 1947, "train_loss": 3.769188642501831, "perplexity": 43.34488236968491, "lr": 0.001305, "grad_norm": 0.263262, "tokens_per_sec": 106491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:19.440190+00:00", "epoch": 1, "step": 1948, "train_loss": 3.781249761581421, "perplexity": 43.87083557367001, "lr": 0.001305, "grad_norm": 0.290949, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:19.745989+00:00", "epoch": 1, "step": 1949, "train_loss": 3.6842610836029053, "perplexity": 39.81569111052768, "lr": 0.001305, "grad_norm": 0.261775, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:20.051510+00:00", "epoch": 1, "step": 1950, "train_loss": 3.626142740249634, "perplexity": 37.56762868137614, "lr": 0.001305, "grad_norm": 0.27315, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:20.357390+00:00", "epoch": 1, "step": 1951, "train_loss": 3.9113810062408447, "perplexity": 49.96791034250413, "lr": 0.001305, "grad_norm": 0.267259, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:20.664757+00:00", "epoch": 1, "step": 1952, "train_loss": 3.7262685298919678, "perplexity": 41.523873624429974, "lr": 0.001305, "grad_norm": 0.250251, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:20.971354+00:00", "epoch": 1, "step": 1953, "train_loss": 3.7725276947021484, "perplexity": 43.48985509558864, "lr": 0.001305, "grad_norm": 0.258346, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:21.277140+00:00", "epoch": 1, "step": 1954, "train_loss": 3.70065975189209, "perplexity": 40.473998350385926, "lr": 0.001305, "grad_norm": 0.290285, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:21.582566+00:00", "epoch": 1, "step": 1955, "train_loss": 3.7523765563964844, "perplexity": 42.622255924624895, "lr": 0.001305, "grad_norm": 0.287014, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:21.888601+00:00", "epoch": 1, "step": 1956, "train_loss": 3.799975872039795, "perplexity": 44.70010595791173, "lr": 0.001305, "grad_norm": 0.274543, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:22.194984+00:00", "epoch": 1, "step": 1957, "train_loss": 3.7744944095611572, "perplexity": 43.57547140365714, "lr": 0.001305, "grad_norm": 0.276568, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:22.500698+00:00", "epoch": 1, "step": 1958, "train_loss": 3.7949581146240234, "perplexity": 44.47637345688954, "lr": 0.001305, "grad_norm": 0.275366, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:22.806917+00:00", "epoch": 1, "step": 1959, "train_loss": 3.7610280513763428, "perplexity": 42.99260187085483, "lr": 0.001305, "grad_norm": 0.252058, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:23.113572+00:00", "epoch": 1, "step": 1960, "train_loss": 3.81954288482666, "perplexity": 45.58336670912851, "lr": 0.001305, "grad_norm": 0.248118, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:23.418404+00:00", "epoch": 1, "step": 1961, "train_loss": 3.6966285705566406, "perplexity": 40.31116874200803, "lr": 0.001305, "grad_norm": 0.244934, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:23.724590+00:00", "epoch": 1, "step": 1962, "train_loss": 3.839250087738037, "perplexity": 46.49059749497194, "lr": 0.001305, "grad_norm": 0.265508, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:24.031589+00:00", "epoch": 1, "step": 1963, "train_loss": 3.7936642169952393, "perplexity": 44.4188627972134, "lr": 0.001305, "grad_norm": 0.27099, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:24.337972+00:00", "epoch": 1, "step": 1964, "train_loss": 3.7184414863586426, "perplexity": 41.20013307634519, "lr": 0.001305, "grad_norm": 0.29037, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:24.644346+00:00", "epoch": 1, "step": 1965, "train_loss": 3.787642002105713, "perplexity": 44.152166716876536, "lr": 0.001305, "grad_norm": 0.286664, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:24.950225+00:00", "epoch": 1, "step": 1966, "train_loss": 3.767354965209961, "perplexity": 43.26547466946304, "lr": 0.001305, "grad_norm": 0.240339, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:25.256479+00:00", "epoch": 1, "step": 1967, "train_loss": 3.697120428085327, "perplexity": 40.331000970759625, "lr": 0.001305, "grad_norm": 0.245518, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:25.562982+00:00", "epoch": 1, "step": 1968, "train_loss": 3.7538669109344482, "perplexity": 42.68582555603824, "lr": 0.001305, "grad_norm": 0.22465, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:25.869707+00:00", "epoch": 1, "step": 1969, "train_loss": 3.5963594913482666, "perplexity": 36.46524048426163, "lr": 0.001305, "grad_norm": 0.237612, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:26.176368+00:00", "epoch": 1, "step": 1970, "train_loss": 3.6476595401763916, "perplexity": 38.38472293409824, "lr": 0.001305, "grad_norm": 0.266546, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:26.482478+00:00", "epoch": 1, "step": 1971, "train_loss": 3.6915178298950195, "perplexity": 40.10567437429807, "lr": 0.001305, "grad_norm": 0.272422, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:26.788695+00:00", "epoch": 1, "step": 1972, "train_loss": 3.7424864768981934, "perplexity": 42.202796089760426, "lr": 0.001305, "grad_norm": 0.259945, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:27.095465+00:00", "epoch": 1, "step": 1973, "train_loss": 3.7696280479431152, "perplexity": 43.363932531915935, "lr": 0.001305, "grad_norm": 0.246929, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:27.402476+00:00", "epoch": 1, "step": 1974, "train_loss": 3.7115235328674316, "perplexity": 40.916096082037406, "lr": 0.001305, "grad_norm": 0.241608, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:27.708544+00:00", "epoch": 1, "step": 1975, "train_loss": 3.649299383163452, "perplexity": 38.44771949092602, "lr": 0.001305, "grad_norm": 0.21843, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:28.015362+00:00", "epoch": 1, "step": 1976, "train_loss": 3.800935983657837, "perplexity": 44.74304365816368, "lr": 0.001305, "grad_norm": 0.226254, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:28.321293+00:00", "epoch": 1, "step": 1977, "train_loss": 3.7812933921813965, "perplexity": 43.872749726305024, "lr": 0.001305, "grad_norm": 0.225263, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:28.627645+00:00", "epoch": 1, "step": 1978, "train_loss": 3.770951509475708, "perplexity": 43.42136102234044, "lr": 0.001305, "grad_norm": 0.231806, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:28.934438+00:00", "epoch": 1, "step": 1979, "train_loss": 3.773184299468994, "perplexity": 43.51842011869273, "lr": 0.001305, "grad_norm": 0.230063, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:29.242264+00:00", "epoch": 1, "step": 1980, "train_loss": 3.7148778438568115, "perplexity": 41.05357183207792, "lr": 0.001305, "grad_norm": 0.228713, "tokens_per_sec": 106511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:29.548566+00:00", "epoch": 1, "step": 1981, "train_loss": 3.608673095703125, "perplexity": 36.917034928717904, "lr": 0.001305, "grad_norm": 0.24214, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:29.854496+00:00", "epoch": 1, "step": 1982, "train_loss": 3.6274619102478027, "perplexity": 37.61721947218514, "lr": 0.001305, "grad_norm": 0.24578, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:30.161142+00:00", "epoch": 1, "step": 1983, "train_loss": 3.7199132442474365, "perplexity": 41.26081434033082, "lr": 0.001305, "grad_norm": 0.247316, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:30.467307+00:00", "epoch": 1, "step": 1984, "train_loss": 3.6790807247161865, "perplexity": 39.60996486920791, "lr": 0.001305, "grad_norm": 0.25517, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:30.773633+00:00", "epoch": 1, "step": 1985, "train_loss": 3.7275402545928955, "perplexity": 41.57671415237716, "lr": 0.001305, "grad_norm": 0.261055, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:31.080271+00:00", "epoch": 1, "step": 1986, "train_loss": 3.7468011379241943, "perplexity": 42.38528024477471, "lr": 0.001305, "grad_norm": 0.28316, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:31.386854+00:00", "epoch": 1, "step": 1987, "train_loss": 3.731637716293335, "perplexity": 41.74742264316003, "lr": 0.001305, "grad_norm": 0.296262, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:31.693495+00:00", "epoch": 1, "step": 1988, "train_loss": 3.7483510971069336, "perplexity": 42.45102663805379, "lr": 0.001305, "grad_norm": 0.314982, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:31.999393+00:00", "epoch": 1, "step": 1989, "train_loss": 3.652630090713501, "perplexity": 38.575991099541305, "lr": 0.001305, "grad_norm": 0.298048, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:32.306663+00:00", "epoch": 1, "step": 1990, "train_loss": 3.7834041118621826, "perplexity": 43.96545060096657, "lr": 0.001305, "grad_norm": 0.291046, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:32.613385+00:00", "epoch": 1, "step": 1991, "train_loss": 3.797666072845459, "perplexity": 44.59697683881248, "lr": 0.001305, "grad_norm": 0.320051, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:32.919717+00:00", "epoch": 1, "step": 1992, "train_loss": 3.788881301879883, "perplexity": 44.20691840698887, "lr": 0.001305, "grad_norm": 0.358427, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:33.225757+00:00", "epoch": 1, "step": 1993, "train_loss": 3.8281185626983643, "perplexity": 45.97595593169979, "lr": 0.001305, "grad_norm": 0.332462, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:33.531585+00:00", "epoch": 1, "step": 1994, "train_loss": 3.8034610748291016, "perplexity": 44.856166685552445, "lr": 0.001305, "grad_norm": 0.273647, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:33.837697+00:00", "epoch": 1, "step": 1995, "train_loss": 3.7081620693206787, "perplexity": 40.77878902221137, "lr": 0.001305, "grad_norm": 0.29307, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:34.156471+00:00", "epoch": 1, "step": 1996, "train_loss": 3.7649805545806885, "perplexity": 43.1628665316611, "lr": 0.001305, "grad_norm": 0.279577, "tokens_per_sec": 102795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:34.463332+00:00", "epoch": 1, "step": 1997, "train_loss": 3.713360548019409, "perplexity": 40.99132865102765, "lr": 0.001305, "grad_norm": 0.244537, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:34.769975+00:00", "epoch": 1, "step": 1998, "train_loss": 3.689727306365967, "perplexity": 40.03392847121412, "lr": 0.001305, "grad_norm": 0.249441, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:35.076302+00:00", "epoch": 1, "step": 1999, "train_loss": 3.7284350395202637, "perplexity": 41.61393301848716, "lr": 0.001305, "grad_norm": 0.245315, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:35.383315+00:00", "epoch": 1, "step": 2000, "train_loss": 3.67490291595459, "perplexity": 39.44482720790876, "lr": 0.001305, "grad_norm": 0.25341, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:40:40.987101+00:00", "step": 2000, "epoch": 1, "val_loss": 3.7983508110046387, "val_ppl": 44.627524548035595, "eval_train_loss": 3.67490291595459, "eval_train_ppl": 39.44482720790876} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:40:41.902074+00:00", "step": 2000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p7984_epoch_0001_step_0002000.pt", "val_loss": 3.7983508110046387} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:40:43.799959+00:00", "step": 2000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0001_step_0002000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:44.826058+00:00", "epoch": 1, "step": 2001, "train_loss": 3.756481885910034, "perplexity": 42.797593993823185, "lr": 0.001305, "grad_norm": 0.270522, "tokens_per_sec": 3470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:45.131674+00:00", "epoch": 1, "step": 2002, "train_loss": 3.7744572162628174, "perplexity": 43.57385071828843, "lr": 0.001305, "grad_norm": 0.279874, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:45.436496+00:00", "epoch": 1, "step": 2003, "train_loss": 3.770136594772339, "perplexity": 43.38599073064583, "lr": 0.001305, "grad_norm": 0.27726, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:45.742419+00:00", "epoch": 1, "step": 2004, "train_loss": 3.753962755203247, "perplexity": 42.68991694384163, "lr": 0.001305, "grad_norm": 0.281963, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:46.048273+00:00", "epoch": 1, "step": 2005, "train_loss": 3.691538095474243, "perplexity": 40.10648714725505, "lr": 0.001305, "grad_norm": 0.282571, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:46.354023+00:00", "epoch": 1, "step": 2006, "train_loss": 3.6786704063415527, "perplexity": 39.59371550673745, "lr": 0.001305, "grad_norm": 0.279982, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:46.658547+00:00", "epoch": 1, "step": 2007, "train_loss": 3.7145700454711914, "perplexity": 41.0409375534491, "lr": 0.001305, "grad_norm": 0.248107, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:46.963510+00:00", "epoch": 1, "step": 2008, "train_loss": 3.7138190269470215, "perplexity": 41.0101266203361, "lr": 0.001305, "grad_norm": 0.237215, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:47.269101+00:00", "epoch": 1, "step": 2009, "train_loss": 3.8376216888427734, "perplexity": 46.414953863089515, "lr": 0.001305, "grad_norm": 0.257449, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:47.576063+00:00", "epoch": 1, "step": 2010, "train_loss": 3.793842077255249, "perplexity": 44.426763850320704, "lr": 0.001305, "grad_norm": 0.242306, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:47.882515+00:00", "epoch": 1, "step": 2011, "train_loss": 3.7558889389038086, "perplexity": 42.77222481062479, "lr": 0.001305, "grad_norm": 0.241624, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:48.187806+00:00", "epoch": 1, "step": 2012, "train_loss": 3.6829802989959717, "perplexity": 39.764728429314346, "lr": 0.001305, "grad_norm": 0.271698, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:48.492823+00:00", "epoch": 1, "step": 2013, "train_loss": 3.6334335803985596, "perplexity": 37.842529166890166, "lr": 0.001305, "grad_norm": 0.23703, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:48.797996+00:00", "epoch": 1, "step": 2014, "train_loss": 3.748098850250244, "perplexity": 42.440319850454635, "lr": 0.001305, "grad_norm": 0.245483, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:49.242272+00:00", "epoch": 1, "step": 2015, "train_loss": 3.7082560062408447, "perplexity": 40.782619835984825, "lr": 0.001305, "grad_norm": 0.260362, "tokens_per_sec": 73757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:49.548979+00:00", "epoch": 1, "step": 2016, "train_loss": 3.7575082778930664, "perplexity": 42.841543652121636, "lr": 0.001305, "grad_norm": 0.252043, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:49.854975+00:00", "epoch": 1, "step": 2017, "train_loss": 3.762197494506836, "perplexity": 43.042908683519684, "lr": 0.001305, "grad_norm": 0.249994, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:50.159626+00:00", "epoch": 1, "step": 2018, "train_loss": 3.655749797821045, "perplexity": 38.69652481033365, "lr": 0.001305, "grad_norm": 0.254198, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:50.465409+00:00", "epoch": 1, "step": 2019, "train_loss": 3.821552276611328, "perplexity": 45.67505363834249, "lr": 0.001305, "grad_norm": 0.26991, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:50.773236+00:00", "epoch": 1, "step": 2020, "train_loss": 3.749980926513672, "perplexity": 42.52027098252109, "lr": 0.001305, "grad_norm": 0.271771, "tokens_per_sec": 106514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:51.079891+00:00", "epoch": 1, "step": 2021, "train_loss": 3.7755565643310547, "perplexity": 43.6217798874939, "lr": 0.001305, "grad_norm": 0.256709, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:51.386216+00:00", "epoch": 1, "step": 2022, "train_loss": 3.7946338653564453, "perplexity": 44.461954363188234, "lr": 0.001305, "grad_norm": 0.273902, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:51.691860+00:00", "epoch": 1, "step": 2023, "train_loss": 3.6462368965148926, "perplexity": 38.33015397661197, "lr": 0.001305, "grad_norm": 0.296027, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:51.997351+00:00", "epoch": 1, "step": 2024, "train_loss": 3.760202407836914, "perplexity": 42.95711995659919, "lr": 0.001305, "grad_norm": 0.294536, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:52.304170+00:00", "epoch": 1, "step": 2025, "train_loss": 3.78267765045166, "perplexity": 43.93352299619927, "lr": 0.001305, "grad_norm": 0.287024, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:52.611004+00:00", "epoch": 1, "step": 2026, "train_loss": 3.722808837890625, "perplexity": 41.380462033999606, "lr": 0.001305, "grad_norm": 0.310547, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:52.918819+00:00", "epoch": 1, "step": 2027, "train_loss": 3.825047492980957, "perplexity": 45.834977154339136, "lr": 0.001305, "grad_norm": 0.285461, "tokens_per_sec": 106454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:53.226790+00:00", "epoch": 1, "step": 2028, "train_loss": 3.7228217124938965, "perplexity": 41.38099479446102, "lr": 0.001305, "grad_norm": 0.273062, "tokens_per_sec": 106397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:53.533053+00:00", "epoch": 1, "step": 2029, "train_loss": 3.802534818649292, "perplexity": 44.81463762020325, "lr": 0.001305, "grad_norm": 0.298642, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:53.839219+00:00", "epoch": 1, "step": 2030, "train_loss": 3.8121206760406494, "perplexity": 45.24628991884029, "lr": 0.001305, "grad_norm": 0.274708, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:54.145584+00:00", "epoch": 1, "step": 2031, "train_loss": 3.7402122020721436, "perplexity": 42.106924393640384, "lr": 0.001305, "grad_norm": 0.266132, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:54.451995+00:00", "epoch": 1, "step": 2032, "train_loss": 3.660943031311035, "perplexity": 38.89800761973469, "lr": 0.001305, "grad_norm": 0.284561, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:54.758001+00:00", "epoch": 1, "step": 2033, "train_loss": 3.7371089458465576, "perplexity": 41.97645835784559, "lr": 0.001305, "grad_norm": 0.281025, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:55.063826+00:00", "epoch": 1, "step": 2034, "train_loss": 3.739445209503174, "perplexity": 42.07464107764265, "lr": 0.001305, "grad_norm": 0.230532, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:55.369827+00:00", "epoch": 1, "step": 2035, "train_loss": 3.728389024734497, "perplexity": 41.61201820632958, "lr": 0.001305, "grad_norm": 0.256742, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:55.677302+00:00", "epoch": 1, "step": 2036, "train_loss": 3.623816967010498, "perplexity": 37.48035642301229, "lr": 0.001305, "grad_norm": 0.255059, "tokens_per_sec": 106573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:55.985221+00:00", "epoch": 1, "step": 2037, "train_loss": 3.8045289516448975, "perplexity": 44.90409313121707, "lr": 0.001305, "grad_norm": 0.24976, "tokens_per_sec": 106418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:56.291933+00:00", "epoch": 1, "step": 2038, "train_loss": 3.7793054580688477, "perplexity": 43.78562022306243, "lr": 0.001305, "grad_norm": 0.256709, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:56.597873+00:00", "epoch": 1, "step": 2039, "train_loss": 3.7290384769439697, "perplexity": 41.63905200112297, "lr": 0.001305, "grad_norm": 0.266986, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:56.904312+00:00", "epoch": 1, "step": 2040, "train_loss": 3.752171516418457, "perplexity": 42.6135175540949, "lr": 0.001305, "grad_norm": 0.25848, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:57.210540+00:00", "epoch": 1, "step": 2041, "train_loss": 3.750169277191162, "perplexity": 42.52828045863916, "lr": 0.001305, "grad_norm": 0.264761, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:57.517497+00:00", "epoch": 1, "step": 2042, "train_loss": 3.6152374744415283, "perplexity": 37.16016946841372, "lr": 0.001305, "grad_norm": 0.212812, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:57.825733+00:00", "epoch": 1, "step": 2043, "train_loss": 3.8085579872131348, "perplexity": 45.08537827651604, "lr": 0.001305, "grad_norm": 0.234185, "tokens_per_sec": 106309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:58.131279+00:00", "epoch": 1, "step": 2044, "train_loss": 3.7774930000305176, "perplexity": 43.70633249826799, "lr": 0.001305, "grad_norm": 0.249113, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:58.437317+00:00", "epoch": 1, "step": 2045, "train_loss": 3.740164041519165, "perplexity": 42.1048965497088, "lr": 0.001305, "grad_norm": 0.266402, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:58.743709+00:00", "epoch": 1, "step": 2046, "train_loss": 3.7269723415374756, "perplexity": 41.55310889710932, "lr": 0.001305, "grad_norm": 0.263272, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:59.051219+00:00", "epoch": 1, "step": 2047, "train_loss": 3.666811227798462, "perplexity": 39.12693982404172, "lr": 0.001305, "grad_norm": 0.25228, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:59.359438+00:00", "epoch": 1, "step": 2048, "train_loss": 3.741880178451538, "perplexity": 42.17721635530692, "lr": 0.001305, "grad_norm": 0.261253, "tokens_per_sec": 106315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:59.665071+00:00", "epoch": 1, "step": 2049, "train_loss": 3.7843220233917236, "perplexity": 44.00582552244276, "lr": 0.001305, "grad_norm": 0.271972, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:40:59.971602+00:00", "epoch": 1, "step": 2050, "train_loss": 3.6132986545562744, "perplexity": 37.088192390750635, "lr": 0.001305, "grad_norm": 0.308876, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:00.278071+00:00", "epoch": 1, "step": 2051, "train_loss": 3.790915012359619, "perplexity": 44.296913961663044, "lr": 0.001305, "grad_norm": 0.278106, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:00.585069+00:00", "epoch": 1, "step": 2052, "train_loss": 3.7530009746551514, "perplexity": 42.648878350349165, "lr": 0.001305, "grad_norm": 0.268324, "tokens_per_sec": 106737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:00.891630+00:00", "epoch": 1, "step": 2053, "train_loss": 3.7144744396209717, "perplexity": 41.03701398728141, "lr": 0.001305, "grad_norm": 0.293233, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:01.199628+00:00", "epoch": 1, "step": 2054, "train_loss": 3.7757651805877686, "perplexity": 43.630881049217365, "lr": 0.001305, "grad_norm": 0.275276, "tokens_per_sec": 106391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:01.506298+00:00", "epoch": 1, "step": 2055, "train_loss": 3.7412166595458984, "perplexity": 42.149240257229295, "lr": 0.001305, "grad_norm": 0.269879, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:01.812659+00:00", "epoch": 1, "step": 2056, "train_loss": 3.7163219451904297, "perplexity": 41.11290017767149, "lr": 0.001305, "grad_norm": 0.267089, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:02.118773+00:00", "epoch": 1, "step": 2057, "train_loss": 3.7566921710968018, "perplexity": 42.80659464018751, "lr": 0.001305, "grad_norm": 0.276162, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:02.425370+00:00", "epoch": 1, "step": 2058, "train_loss": 3.7704918384552, "perplexity": 43.40140606771601, "lr": 0.001305, "grad_norm": 0.243385, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:02.732500+00:00", "epoch": 1, "step": 2059, "train_loss": 3.771090030670166, "perplexity": 43.42737621774066, "lr": 0.001305, "grad_norm": 0.228751, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:03.039719+00:00", "epoch": 1, "step": 2060, "train_loss": 3.684084892272949, "perplexity": 39.80867654892841, "lr": 0.001305, "grad_norm": 0.24346, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:03.345749+00:00", "epoch": 1, "step": 2061, "train_loss": 3.6915712356567383, "perplexity": 40.1078163055825, "lr": 0.001305, "grad_norm": 0.239361, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:03.652124+00:00", "epoch": 1, "step": 2062, "train_loss": 3.773545026779175, "perplexity": 43.53412123306558, "lr": 0.001305, "grad_norm": 0.252044, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:03.959596+00:00", "epoch": 1, "step": 2063, "train_loss": 3.600585699081421, "perplexity": 36.61967627459065, "lr": 0.001305, "grad_norm": 0.238329, "tokens_per_sec": 106574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:04.266535+00:00", "epoch": 1, "step": 2064, "train_loss": 3.7767906188964844, "perplexity": 43.67564477338455, "lr": 0.001305, "grad_norm": 0.227576, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:04.573281+00:00", "epoch": 1, "step": 2065, "train_loss": 3.665440320968628, "perplexity": 39.07333718551307, "lr": 0.001305, "grad_norm": 0.226638, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:04.879457+00:00", "epoch": 1, "step": 2066, "train_loss": 3.6797990798950195, "perplexity": 39.63842911510028, "lr": 0.001305, "grad_norm": 0.226199, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:05.185161+00:00", "epoch": 1, "step": 2067, "train_loss": 3.8568637371063232, "perplexity": 47.31672074280807, "lr": 0.001305, "grad_norm": 0.269195, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:05.493287+00:00", "epoch": 1, "step": 2068, "train_loss": 3.688080072402954, "perplexity": 39.968037508378366, "lr": 0.001305, "grad_norm": 0.284122, "tokens_per_sec": 106346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:05.799478+00:00", "epoch": 1, "step": 2069, "train_loss": 3.6945362091064453, "perplexity": 40.226911385681795, "lr": 0.001305, "grad_norm": 0.2908, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:06.107203+00:00", "epoch": 1, "step": 2070, "train_loss": 3.687967300415039, "perplexity": 39.963530487473115, "lr": 0.001305, "grad_norm": 0.303469, "tokens_per_sec": 106540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:06.412975+00:00", "epoch": 1, "step": 2071, "train_loss": 3.7683496475219727, "perplexity": 43.308531482212565, "lr": 0.001305, "grad_norm": 0.261358, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:06.719895+00:00", "epoch": 1, "step": 2072, "train_loss": 3.8150713443756104, "perplexity": 45.37999387478566, "lr": 0.001305, "grad_norm": 0.302173, "tokens_per_sec": 106763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:07.026394+00:00", "epoch": 1, "step": 2073, "train_loss": 3.7334067821502686, "perplexity": 41.82134194795404, "lr": 0.001305, "grad_norm": 0.30847, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:07.332314+00:00", "epoch": 1, "step": 2074, "train_loss": 3.670609474182129, "perplexity": 39.27583617523951, "lr": 0.001305, "grad_norm": 0.301013, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:07.638506+00:00", "epoch": 1, "step": 2075, "train_loss": 3.752445697784424, "perplexity": 42.6252029884375, "lr": 0.001305, "grad_norm": 0.279128, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:07.945272+00:00", "epoch": 1, "step": 2076, "train_loss": 3.864917039871216, "perplexity": 47.69931512721582, "lr": 0.001305, "grad_norm": 0.299534, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:08.253421+00:00", "epoch": 1, "step": 2077, "train_loss": 3.669797897338867, "perplexity": 39.24397374725317, "lr": 0.001305, "grad_norm": 0.301152, "tokens_per_sec": 106339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:08.574045+00:00", "epoch": 1, "step": 2078, "train_loss": 3.719522714614868, "perplexity": 41.24470391567088, "lr": 0.001305, "grad_norm": 0.273837, "tokens_per_sec": 102200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:08.879648+00:00", "epoch": 1, "step": 2079, "train_loss": 3.7392759323120117, "perplexity": 42.06751940336739, "lr": 0.001305, "grad_norm": 0.275424, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:09.186061+00:00", "epoch": 1, "step": 2080, "train_loss": 3.69492506980896, "perplexity": 40.242557092506345, "lr": 0.001305, "grad_norm": 0.247736, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:09.491900+00:00", "epoch": 1, "step": 2081, "train_loss": 3.7711520195007324, "perplexity": 43.43006831344603, "lr": 0.001305, "grad_norm": 0.236905, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:09.798459+00:00", "epoch": 1, "step": 2082, "train_loss": 3.646648645401001, "perplexity": 38.34593962445464, "lr": 0.001305, "grad_norm": 0.265755, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:10.106515+00:00", "epoch": 1, "step": 2083, "train_loss": 3.5736184120178223, "perplexity": 35.64533960750997, "lr": 0.001305, "grad_norm": 0.275581, "tokens_per_sec": 106371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:10.414802+00:00", "epoch": 1, "step": 2084, "train_loss": 3.5975821018218994, "perplexity": 36.50985053400141, "lr": 0.001305, "grad_norm": 0.239321, "tokens_per_sec": 106290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:10.721367+00:00", "epoch": 1, "step": 2085, "train_loss": 3.7684009075164795, "perplexity": 43.310751534197884, "lr": 0.001305, "grad_norm": 0.263655, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:11.044660+00:00", "epoch": 1, "step": 2086, "train_loss": 3.8063392639160156, "perplexity": 44.98545718698951, "lr": 0.001305, "grad_norm": 0.284394, "tokens_per_sec": 101358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:11.350911+00:00", "epoch": 1, "step": 2087, "train_loss": 3.694143772125244, "perplexity": 40.21112795521802, "lr": 0.001305, "grad_norm": 0.269105, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:11.658163+00:00", "epoch": 1, "step": 2088, "train_loss": 3.6857268810272217, "perplexity": 39.87409564215373, "lr": 0.001305, "grad_norm": 0.259273, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:11.974736+00:00", "epoch": 1, "step": 2089, "train_loss": 3.6168811321258545, "perplexity": 37.22129829017797, "lr": 0.001305, "grad_norm": 0.237059, "tokens_per_sec": 103509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:12.283088+00:00", "epoch": 1, "step": 2090, "train_loss": 3.794074058532715, "perplexity": 44.43707122326492, "lr": 0.001305, "grad_norm": 0.247556, "tokens_per_sec": 106326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:12.601840+00:00", "epoch": 1, "step": 2091, "train_loss": 3.6924691200256348, "perplexity": 40.14384465914111, "lr": 0.001305, "grad_norm": 0.253419, "tokens_per_sec": 102747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:12.907704+00:00", "epoch": 1, "step": 2092, "train_loss": 3.6842758655548096, "perplexity": 39.81627966850872, "lr": 0.001305, "grad_norm": 0.229085, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:13.230496+00:00", "epoch": 1, "step": 2093, "train_loss": 3.706521987915039, "perplexity": 40.71196330338034, "lr": 0.001305, "grad_norm": 0.248901, "tokens_per_sec": 101515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:13.536775+00:00", "epoch": 1, "step": 2094, "train_loss": 3.6405327320098877, "perplexity": 38.112134872394314, "lr": 0.001305, "grad_norm": 0.232573, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:13.855764+00:00", "epoch": 1, "step": 2095, "train_loss": 3.796868324279785, "perplexity": 44.561413851553006, "lr": 0.001305, "grad_norm": 0.234302, "tokens_per_sec": 102724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:14.162404+00:00", "epoch": 1, "step": 2096, "train_loss": 3.6664321422576904, "perplexity": 39.11211017792959, "lr": 0.001305, "grad_norm": 0.251747, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:14.479810+00:00", "epoch": 1, "step": 2097, "train_loss": 3.6890509128570557, "perplexity": 40.006858937720395, "lr": 0.001305, "grad_norm": 0.280615, "tokens_per_sec": 103238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:14.787176+00:00", "epoch": 1, "step": 2098, "train_loss": 3.7374744415283203, "perplexity": 41.99180337620922, "lr": 0.001305, "grad_norm": 0.307688, "tokens_per_sec": 106609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:15.093007+00:00", "epoch": 1, "step": 2099, "train_loss": 3.818167209625244, "perplexity": 45.520701915034, "lr": 0.001305, "grad_norm": 0.254376, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:15.412601+00:00", "epoch": 1, "step": 2100, "train_loss": 3.723968267440796, "perplexity": 41.42846758863313, "lr": 0.001305, "grad_norm": 0.239684, "tokens_per_sec": 102585} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:41:21.050416+00:00", "step": 2100, "epoch": 1, "val_loss": 3.785566258430481, "val_ppl": 44.06061318976675, "eval_train_loss": 3.723968267440796, "eval_train_ppl": 41.42846758863313} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:41:21.969213+00:00", "step": 2100, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p7856_epoch_0001_step_0002100.pt", "val_loss": 3.785566258430481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:22.980023+00:00", "epoch": 1, "step": 2101, "train_loss": 3.653323173522949, "perplexity": 38.60273672322519, "lr": 0.001305, "grad_norm": 0.262734, "tokens_per_sec": 4330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:23.286219+00:00", "epoch": 1, "step": 2102, "train_loss": 3.6246230602264404, "perplexity": 37.51058126443916, "lr": 0.001305, "grad_norm": 0.27213, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:23.592343+00:00", "epoch": 1, "step": 2103, "train_loss": 3.679736852645874, "perplexity": 39.63596260143897, "lr": 0.001305, "grad_norm": 0.288053, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:23.897739+00:00", "epoch": 1, "step": 2104, "train_loss": 3.6649680137634277, "perplexity": 39.05488692426778, "lr": 0.001305, "grad_norm": 0.247047, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:24.204581+00:00", "epoch": 1, "step": 2105, "train_loss": 3.7227587699890137, "perplexity": 41.37839025296316, "lr": 0.001305, "grad_norm": 0.231164, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:24.511952+00:00", "epoch": 1, "step": 2106, "train_loss": 3.6680047512054443, "perplexity": 39.17366662178187, "lr": 0.001305, "grad_norm": 0.250756, "tokens_per_sec": 106608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:24.816942+00:00", "epoch": 1, "step": 2107, "train_loss": 3.726227283477783, "perplexity": 41.52216094886102, "lr": 0.001305, "grad_norm": 0.241471, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:25.123072+00:00", "epoch": 1, "step": 2108, "train_loss": 3.8394525051116943, "perplexity": 46.5000089521057, "lr": 0.001305, "grad_norm": 0.256375, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:25.428656+00:00", "epoch": 1, "step": 2109, "train_loss": 3.7907590866088867, "perplexity": 44.290007470562344, "lr": 0.001305, "grad_norm": 0.288507, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:25.737048+00:00", "epoch": 1, "step": 2110, "train_loss": 3.6191585063934326, "perplexity": 37.306161713319725, "lr": 0.001305, "grad_norm": 0.263576, "tokens_per_sec": 106307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:26.043593+00:00", "epoch": 1, "step": 2111, "train_loss": 3.7218017578125, "perplexity": 41.33880957226413, "lr": 0.001305, "grad_norm": 0.251661, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:26.349792+00:00", "epoch": 1, "step": 2112, "train_loss": 3.8313891887664795, "perplexity": 46.12657226226322, "lr": 0.001305, "grad_norm": 0.272668, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:26.655860+00:00", "epoch": 1, "step": 2113, "train_loss": 3.7204535007476807, "perplexity": 41.28311178611121, "lr": 0.001305, "grad_norm": 0.297119, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:26.962936+00:00", "epoch": 1, "step": 2114, "train_loss": 3.551157236099243, "perplexity": 34.8536280352449, "lr": 0.001305, "grad_norm": 0.255207, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:27.269462+00:00", "epoch": 1, "step": 2115, "train_loss": 3.668606758117676, "perplexity": 39.197256539800044, "lr": 0.001305, "grad_norm": 0.232197, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:27.577044+00:00", "epoch": 1, "step": 2116, "train_loss": 3.729844093322754, "perplexity": 41.672610619283326, "lr": 0.001305, "grad_norm": 0.242376, "tokens_per_sec": 106536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:27.884547+00:00", "epoch": 1, "step": 2117, "train_loss": 3.721937894821167, "perplexity": 41.34443769723051, "lr": 0.001305, "grad_norm": 0.258183, "tokens_per_sec": 106561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:28.191351+00:00", "epoch": 1, "step": 2118, "train_loss": 3.626695156097412, "perplexity": 37.58838736800902, "lr": 0.001305, "grad_norm": 0.258329, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:28.497215+00:00", "epoch": 1, "step": 2119, "train_loss": 3.64937162399292, "perplexity": 38.450497086399885, "lr": 0.001305, "grad_norm": 0.267389, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:28.804374+00:00", "epoch": 1, "step": 2120, "train_loss": 3.7706949710845947, "perplexity": 43.41022320494378, "lr": 0.001305, "grad_norm": 0.238813, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:29.110410+00:00", "epoch": 1, "step": 2121, "train_loss": 3.711215019226074, "perplexity": 40.90347485525543, "lr": 0.001305, "grad_norm": 0.26024, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:29.418307+00:00", "epoch": 1, "step": 2122, "train_loss": 3.789919137954712, "perplexity": 44.25282175762507, "lr": 0.001305, "grad_norm": 0.267602, "tokens_per_sec": 106425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:29.725641+00:00", "epoch": 1, "step": 2123, "train_loss": 3.7146317958831787, "perplexity": 41.043471926499855, "lr": 0.001305, "grad_norm": 0.256514, "tokens_per_sec": 106620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:30.032267+00:00", "epoch": 1, "step": 2124, "train_loss": 3.691748857498169, "perplexity": 40.11494096249907, "lr": 0.001305, "grad_norm": 0.274095, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:30.338117+00:00", "epoch": 1, "step": 2125, "train_loss": 3.7537338733673096, "perplexity": 42.6801471153863, "lr": 0.001305, "grad_norm": 0.29341, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:30.644497+00:00", "epoch": 1, "step": 2126, "train_loss": 3.6596736907958984, "perplexity": 38.84866412617792, "lr": 0.001305, "grad_norm": 0.287957, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:30.951658+00:00", "epoch": 1, "step": 2127, "train_loss": 3.6749672889709473, "perplexity": 39.447366472145, "lr": 0.001305, "grad_norm": 0.268374, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:31.258478+00:00", "epoch": 1, "step": 2128, "train_loss": 3.8644895553588867, "perplexity": 47.678928766487424, "lr": 0.001305, "grad_norm": 0.263996, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:31.565466+00:00", "epoch": 1, "step": 2129, "train_loss": 3.7238054275512695, "perplexity": 41.421721930793744, "lr": 0.001305, "grad_norm": 0.315958, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:31.872409+00:00", "epoch": 1, "step": 2130, "train_loss": 3.8115880489349365, "perplexity": 45.22219693525421, "lr": 0.001305, "grad_norm": 0.288842, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:32.178951+00:00", "epoch": 1, "step": 2131, "train_loss": 3.772792100906372, "perplexity": 43.501355603432515, "lr": 0.001305, "grad_norm": 0.260986, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:32.485789+00:00", "epoch": 1, "step": 2132, "train_loss": 3.735719919204712, "perplexity": 41.918192414698055, "lr": 0.001305, "grad_norm": 0.234287, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:32.793485+00:00", "epoch": 1, "step": 2133, "train_loss": 3.62874174118042, "perplexity": 37.66539397429799, "lr": 0.001305, "grad_norm": 0.234427, "tokens_per_sec": 106495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:33.101248+00:00", "epoch": 1, "step": 2134, "train_loss": 3.6805272102355957, "perplexity": 39.667301568170686, "lr": 0.001305, "grad_norm": 0.236624, "tokens_per_sec": 106470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:33.407523+00:00", "epoch": 1, "step": 2135, "train_loss": 3.8314692974090576, "perplexity": 46.130267547364085, "lr": 0.001305, "grad_norm": 0.249708, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:33.714164+00:00", "epoch": 1, "step": 2136, "train_loss": 3.6478893756866455, "perplexity": 38.39354612038168, "lr": 0.001305, "grad_norm": 0.235592, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:34.021412+00:00", "epoch": 1, "step": 2137, "train_loss": 3.6811299324035645, "perplexity": 39.69121713666698, "lr": 0.001305, "grad_norm": 0.2492, "tokens_per_sec": 106651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:34.328331+00:00", "epoch": 1, "step": 2138, "train_loss": 3.6990718841552734, "perplexity": 40.409781991260324, "lr": 0.001305, "grad_norm": 0.247212, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:34.635290+00:00", "epoch": 1, "step": 2139, "train_loss": 3.669769763946533, "perplexity": 39.24286969667342, "lr": 0.001305, "grad_norm": 0.225457, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:34.943059+00:00", "epoch": 1, "step": 2140, "train_loss": 3.8087167739868164, "perplexity": 45.092537806677086, "lr": 0.001305, "grad_norm": 0.236747, "tokens_per_sec": 106525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:35.261611+00:00", "epoch": 1, "step": 2141, "train_loss": 3.72102952003479, "perplexity": 41.3068985048786, "lr": 0.001305, "grad_norm": 0.237257, "tokens_per_sec": 102814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:35.568923+00:00", "epoch": 1, "step": 2142, "train_loss": 3.7769925594329834, "perplexity": 43.68446554712794, "lr": 0.001305, "grad_norm": 0.231102, "tokens_per_sec": 106627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:35.891048+00:00", "epoch": 1, "step": 2143, "train_loss": 3.7084732055664062, "perplexity": 40.79147875554869, "lr": 0.001305, "grad_norm": 0.233526, "tokens_per_sec": 101724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:36.197219+00:00", "epoch": 1, "step": 2144, "train_loss": 3.8139142990112305, "perplexity": 45.32751752784864, "lr": 0.001305, "grad_norm": 0.253412, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:36.503862+00:00", "epoch": 1, "step": 2145, "train_loss": 3.7682416439056396, "perplexity": 43.30385425677759, "lr": 0.001305, "grad_norm": 0.239419, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:36.811397+00:00", "epoch": 1, "step": 2146, "train_loss": 3.6802241802215576, "perplexity": 39.65528300630408, "lr": 0.001305, "grad_norm": 0.222107, "tokens_per_sec": 106551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:37.128762+00:00", "epoch": 1, "step": 2147, "train_loss": 3.7801971435546875, "perplexity": 43.824680637317975, "lr": 0.001305, "grad_norm": 0.23011, "tokens_per_sec": 103250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:37.434754+00:00", "epoch": 1, "step": 2148, "train_loss": 3.6700820922851562, "perplexity": 39.255128271219036, "lr": 0.001305, "grad_norm": 0.252654, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:37.756982+00:00", "epoch": 1, "step": 2149, "train_loss": 3.729111909866333, "perplexity": 41.642109790665685, "lr": 0.001305, "grad_norm": 0.244465, "tokens_per_sec": 101693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:38.063913+00:00", "epoch": 1, "step": 2150, "train_loss": 3.6403400897979736, "perplexity": 38.10479357357646, "lr": 0.001305, "grad_norm": 0.212809, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:38.370040+00:00", "epoch": 1, "step": 2151, "train_loss": 3.701566219329834, "perplexity": 40.51070334539393, "lr": 0.001305, "grad_norm": 0.223406, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:38.678084+00:00", "epoch": 1, "step": 2152, "train_loss": 3.690530300140381, "perplexity": 40.066088376915815, "lr": 0.001305, "grad_norm": 0.248084, "tokens_per_sec": 106374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:38.995250+00:00", "epoch": 1, "step": 2153, "train_loss": 3.6951956748962402, "perplexity": 40.2534484067368, "lr": 0.001305, "grad_norm": 0.222119, "tokens_per_sec": 103316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:39.303705+00:00", "epoch": 1, "step": 2154, "train_loss": 3.749096393585205, "perplexity": 42.48267703170454, "lr": 0.001305, "grad_norm": 0.21696, "tokens_per_sec": 106232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:39.622674+00:00", "epoch": 1, "step": 2155, "train_loss": 3.8424229621887207, "perplexity": 46.638340585153045, "lr": 0.001305, "grad_norm": 0.240385, "tokens_per_sec": 102732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:39.928868+00:00", "epoch": 1, "step": 2156, "train_loss": 3.6555676460266113, "perplexity": 38.68947681082347, "lr": 0.001305, "grad_norm": 0.274743, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:40.247911+00:00", "epoch": 1, "step": 2157, "train_loss": 3.6669559478759766, "perplexity": 39.1326026875611, "lr": 0.001305, "grad_norm": 0.335716, "tokens_per_sec": 102704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:40.555540+00:00", "epoch": 1, "step": 2158, "train_loss": 3.7020347118377686, "perplexity": 40.52968675284727, "lr": 0.001305, "grad_norm": 0.318998, "tokens_per_sec": 106517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:40.863608+00:00", "epoch": 1, "step": 2159, "train_loss": 3.6285879611968994, "perplexity": 37.659602235971455, "lr": 0.001305, "grad_norm": 0.292573, "tokens_per_sec": 106367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:41.181471+00:00", "epoch": 1, "step": 2160, "train_loss": 3.713242769241333, "perplexity": 40.98650102672884, "lr": 0.001305, "grad_norm": 0.341219, "tokens_per_sec": 103152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:41.487838+00:00", "epoch": 1, "step": 2161, "train_loss": 3.737575054168701, "perplexity": 41.99602849496787, "lr": 0.001305, "grad_norm": 0.293894, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:41.807714+00:00", "epoch": 1, "step": 2162, "train_loss": 3.733097791671753, "perplexity": 41.8084215477365, "lr": 0.001305, "grad_norm": 0.289441, "tokens_per_sec": 102440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:42.114061+00:00", "epoch": 1, "step": 2163, "train_loss": 3.6924262046813965, "perplexity": 40.142121909194984, "lr": 0.001305, "grad_norm": 0.302531, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:42.433753+00:00", "epoch": 1, "step": 2164, "train_loss": 3.693964958190918, "perplexity": 40.203938288050146, "lr": 0.001305, "grad_norm": 0.303035, "tokens_per_sec": 102499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:42.740107+00:00", "epoch": 1, "step": 2165, "train_loss": 3.716261386871338, "perplexity": 41.110410524929094, "lr": 0.001305, "grad_norm": 0.271062, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:43.048788+00:00", "epoch": 1, "step": 2166, "train_loss": 3.6899781227111816, "perplexity": 40.043970894187126, "lr": 0.001305, "grad_norm": 0.257134, "tokens_per_sec": 106155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:43.366341+00:00", "epoch": 1, "step": 2167, "train_loss": 3.687958002090454, "perplexity": 39.96315889532267, "lr": 0.001305, "grad_norm": 0.26028, "tokens_per_sec": 103189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:43.684514+00:00", "epoch": 1, "step": 2168, "train_loss": 3.632551670074463, "perplexity": 37.80917016171561, "lr": 0.001305, "grad_norm": 0.267365, "tokens_per_sec": 102989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:43.991206+00:00", "epoch": 1, "step": 2169, "train_loss": 3.7123422622680664, "perplexity": 40.94960900999665, "lr": 0.001305, "grad_norm": 0.26897, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:44.311569+00:00", "epoch": 1, "step": 2170, "train_loss": 3.729875087738037, "perplexity": 41.67390225749948, "lr": 0.001305, "grad_norm": 0.278668, "tokens_per_sec": 102337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:44.617734+00:00", "epoch": 1, "step": 2171, "train_loss": 3.7641048431396484, "perplexity": 43.125084860946366, "lr": 0.001305, "grad_norm": 0.231195, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:44.940736+00:00", "epoch": 1, "step": 2172, "train_loss": 3.6248345375061035, "perplexity": 37.51851473896878, "lr": 0.001305, "grad_norm": 0.241185, "tokens_per_sec": 101449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:45.248012+00:00", "epoch": 1, "step": 2173, "train_loss": 3.7727630138397217, "perplexity": 43.50009029500484, "lr": 0.001305, "grad_norm": 0.267198, "tokens_per_sec": 106641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:45.555695+00:00", "epoch": 1, "step": 2174, "train_loss": 3.6822006702423096, "perplexity": 39.73373878542768, "lr": 0.001305, "grad_norm": 0.276929, "tokens_per_sec": 106499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:45.862846+00:00", "epoch": 1, "step": 2175, "train_loss": 3.742140531539917, "perplexity": 42.18819875343284, "lr": 0.001305, "grad_norm": 0.274248, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:46.182978+00:00", "epoch": 1, "step": 2176, "train_loss": 3.7064857482910156, "perplexity": 40.71048794387036, "lr": 0.001305, "grad_norm": 0.270184, "tokens_per_sec": 102358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:46.489441+00:00", "epoch": 1, "step": 2177, "train_loss": 3.7129251956939697, "perplexity": 40.97348686478993, "lr": 0.001305, "grad_norm": 0.272919, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:46.810293+00:00", "epoch": 1, "step": 2178, "train_loss": 3.7055888175964355, "perplexity": 40.67398982822913, "lr": 0.001305, "grad_norm": 0.25336, "tokens_per_sec": 102127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:47.116845+00:00", "epoch": 1, "step": 2179, "train_loss": 3.6834888458251953, "perplexity": 39.784955798718364, "lr": 0.001305, "grad_norm": 0.245894, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:47.432113+00:00", "epoch": 1, "step": 2180, "train_loss": 3.7728161811828613, "perplexity": 43.50240314071554, "lr": 0.001305, "grad_norm": 0.281445, "tokens_per_sec": 103988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:47.740331+00:00", "epoch": 1, "step": 2181, "train_loss": 3.7269692420959473, "perplexity": 41.552980105877566, "lr": 0.001305, "grad_norm": 0.321026, "tokens_per_sec": 106262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:48.058371+00:00", "epoch": 1, "step": 2182, "train_loss": 3.7180514335632324, "perplexity": 41.18406598297854, "lr": 0.001305, "grad_norm": 0.277446, "tokens_per_sec": 103031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:48.378412+00:00", "epoch": 1, "step": 2183, "train_loss": 3.7655699253082275, "perplexity": 43.18831295966485, "lr": 0.001305, "grad_norm": 0.280002, "tokens_per_sec": 102386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:48.684579+00:00", "epoch": 1, "step": 2184, "train_loss": 3.6051418781280518, "perplexity": 36.78690274391061, "lr": 0.001305, "grad_norm": 0.300762, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:49.005121+00:00", "epoch": 1, "step": 2185, "train_loss": 3.7048065662384033, "perplexity": 40.642184985761844, "lr": 0.001305, "grad_norm": 0.306511, "tokens_per_sec": 102228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:49.312895+00:00", "epoch": 1, "step": 2186, "train_loss": 3.7155520915985107, "perplexity": 41.081261443954524, "lr": 0.001305, "grad_norm": 0.273553, "tokens_per_sec": 106468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:49.619584+00:00", "epoch": 1, "step": 2187, "train_loss": 3.65971040725708, "perplexity": 38.85009053783251, "lr": 0.001305, "grad_norm": 0.267597, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:49.926765+00:00", "epoch": 1, "step": 2188, "train_loss": 3.6553843021392822, "perplexity": 38.68238398197952, "lr": 0.001305, "grad_norm": 0.276378, "tokens_per_sec": 106673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:50.246085+00:00", "epoch": 1, "step": 2189, "train_loss": 3.766874313354492, "perplexity": 43.24468403571404, "lr": 0.001305, "grad_norm": 0.243403, "tokens_per_sec": 102619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:50.566464+00:00", "epoch": 1, "step": 2190, "train_loss": 3.71213436126709, "perplexity": 40.941096430211466, "lr": 0.001305, "grad_norm": 0.245924, "tokens_per_sec": 102328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:50.872903+00:00", "epoch": 1, "step": 2191, "train_loss": 3.7114992141723633, "perplexity": 40.91510106807217, "lr": 0.001305, "grad_norm": 0.262567, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:51.179305+00:00", "epoch": 1, "step": 2192, "train_loss": 3.765737533569336, "perplexity": 43.19555228436859, "lr": 0.001305, "grad_norm": 0.262161, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:51.499557+00:00", "epoch": 1, "step": 2193, "train_loss": 3.7730066776275635, "perplexity": 43.51069098322705, "lr": 0.001305, "grad_norm": 0.261395, "tokens_per_sec": 102319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:51.806235+00:00", "epoch": 1, "step": 2194, "train_loss": 3.6539807319641113, "perplexity": 38.6281286260246, "lr": 0.001305, "grad_norm": 0.249338, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:52.123578+00:00", "epoch": 1, "step": 2195, "train_loss": 3.6498496532440186, "perplexity": 38.4688819426259, "lr": 0.001305, "grad_norm": 0.246993, "tokens_per_sec": 103259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:52.431739+00:00", "epoch": 1, "step": 2196, "train_loss": 3.656275987625122, "perplexity": 38.71689188514569, "lr": 0.001305, "grad_norm": 0.271753, "tokens_per_sec": 106333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:52.750786+00:00", "epoch": 1, "step": 2197, "train_loss": 3.834843635559082, "perplexity": 46.28618958791888, "lr": 0.001305, "grad_norm": 0.304148, "tokens_per_sec": 102706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:53.057950+00:00", "epoch": 1, "step": 2198, "train_loss": 3.8961591720581055, "perplexity": 49.21306672399236, "lr": 0.001305, "grad_norm": 0.286949, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:53.380209+00:00", "epoch": 1, "step": 2199, "train_loss": 3.775237798690796, "perplexity": 43.60787697890134, "lr": 0.001305, "grad_norm": 0.244347, "tokens_per_sec": 101683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:41:53.687556+00:00", "epoch": 1, "step": 2200, "train_loss": 3.712090015411377, "perplexity": 40.93928090251231, "lr": 0.001305, "grad_norm": 0.21794, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:41:59.258086+00:00", "step": 2200, "epoch": 1, "val_loss": 3.7709935665130616, "val_ppl": 43.42318723454517, "eval_train_loss": 3.712090015411377, "eval_train_ppl": 40.93928090251231} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:42:00.160658+00:00", "step": 2200, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p7710_epoch_0001_step_0002200.pt", "val_loss": 3.7709935665130616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:01.136690+00:00", "epoch": 1, "step": 2201, "train_loss": 3.8596975803375244, "perplexity": 47.450999083629284, "lr": 0.001305, "grad_norm": 0.233385, "tokens_per_sec": 4399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:01.444000+00:00", "epoch": 1, "step": 2202, "train_loss": 3.6974127292633057, "perplexity": 40.342791492960245, "lr": 0.001305, "grad_norm": 0.25866, "tokens_per_sec": 106628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:01.749426+00:00", "epoch": 1, "step": 2203, "train_loss": 3.7759132385253906, "perplexity": 43.63734142572535, "lr": 0.001305, "grad_norm": 0.241679, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:02.056248+00:00", "epoch": 1, "step": 2204, "train_loss": 3.6859889030456543, "perplexity": 39.88454490208537, "lr": 0.001305, "grad_norm": 0.247345, "tokens_per_sec": 106797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:02.362398+00:00", "epoch": 1, "step": 2205, "train_loss": 3.626948356628418, "perplexity": 37.59790597265728, "lr": 0.001305, "grad_norm": 0.23055, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:02.668765+00:00", "epoch": 1, "step": 2206, "train_loss": 3.853963851928711, "perplexity": 47.17970644455566, "lr": 0.001305, "grad_norm": 0.261913, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:02.974789+00:00", "epoch": 1, "step": 2207, "train_loss": 3.7412304878234863, "perplexity": 42.14982311265363, "lr": 0.001305, "grad_norm": 0.223929, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:03.280692+00:00", "epoch": 1, "step": 2208, "train_loss": 3.6644086837768555, "perplexity": 39.03304846291134, "lr": 0.001305, "grad_norm": 0.23431, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:03.586403+00:00", "epoch": 1, "step": 2209, "train_loss": 3.727224826812744, "perplexity": 41.56360176983966, "lr": 0.001305, "grad_norm": 0.229848, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:03.892678+00:00", "epoch": 1, "step": 2210, "train_loss": 3.7015154361724854, "perplexity": 40.508646136207865, "lr": 0.001305, "grad_norm": 0.241528, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:04.198699+00:00", "epoch": 1, "step": 2211, "train_loss": 3.792295455932617, "perplexity": 44.35810557791441, "lr": 0.001305, "grad_norm": 0.249227, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:04.503715+00:00", "epoch": 1, "step": 2212, "train_loss": 3.7958621978759766, "perplexity": 44.51660198346797, "lr": 0.001305, "grad_norm": 0.223608, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:04.809000+00:00", "epoch": 1, "step": 2213, "train_loss": 3.738145112991333, "perplexity": 42.01997552648672, "lr": 0.001305, "grad_norm": 0.235374, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:05.115061+00:00", "epoch": 1, "step": 2214, "train_loss": 3.7181756496429443, "perplexity": 41.18918202394224, "lr": 0.001305, "grad_norm": 0.232813, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:05.486374+00:00", "epoch": 1, "step": 2215, "train_loss": 3.726511001586914, "perplexity": 41.53394320919374, "lr": 0.001305, "grad_norm": 0.265591, "tokens_per_sec": 88249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:05.792501+00:00", "epoch": 1, "step": 2216, "train_loss": 3.6948513984680176, "perplexity": 40.239592478567275, "lr": 0.001305, "grad_norm": 0.25061, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:06.098668+00:00", "epoch": 1, "step": 2217, "train_loss": 3.6325647830963135, "perplexity": 37.80966595744078, "lr": 0.001305, "grad_norm": 0.234578, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:06.406679+00:00", "epoch": 1, "step": 2218, "train_loss": 3.679657459259033, "perplexity": 39.63281589304292, "lr": 0.001305, "grad_norm": 0.216512, "tokens_per_sec": 106384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:06.712144+00:00", "epoch": 1, "step": 2219, "train_loss": 3.704909563064575, "perplexity": 40.64637121740465, "lr": 0.001305, "grad_norm": 0.2276, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:07.018590+00:00", "epoch": 1, "step": 2220, "train_loss": 3.742814064025879, "perplexity": 42.21662344722027, "lr": 0.001305, "grad_norm": 0.248108, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:07.325230+00:00", "epoch": 1, "step": 2221, "train_loss": 3.7703285217285156, "perplexity": 43.39431847092083, "lr": 0.001305, "grad_norm": 0.22835, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:07.632957+00:00", "epoch": 1, "step": 2222, "train_loss": 3.5975143909454346, "perplexity": 36.50737850371476, "lr": 0.001305, "grad_norm": 0.241478, "tokens_per_sec": 106484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:07.940562+00:00", "epoch": 1, "step": 2223, "train_loss": 3.669184923171997, "perplexity": 39.21992557634753, "lr": 0.001305, "grad_norm": 0.268804, "tokens_per_sec": 106527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:08.247133+00:00", "epoch": 1, "step": 2224, "train_loss": 3.7102627754211426, "perplexity": 40.864543313825735, "lr": 0.001305, "grad_norm": 0.251707, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:08.553419+00:00", "epoch": 1, "step": 2225, "train_loss": 3.5760154724121094, "perplexity": 35.730886128450855, "lr": 0.001305, "grad_norm": 0.237912, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:08.860130+00:00", "epoch": 1, "step": 2226, "train_loss": 3.866305351257324, "perplexity": 47.76558261882936, "lr": 0.001305, "grad_norm": 0.283322, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:09.166514+00:00", "epoch": 1, "step": 2227, "train_loss": 3.6612138748168945, "perplexity": 38.90854431932328, "lr": 0.001305, "grad_norm": 0.289553, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:09.473712+00:00", "epoch": 1, "step": 2228, "train_loss": 3.7107555866241455, "perplexity": 40.88468678163201, "lr": 0.001305, "grad_norm": 0.244696, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:09.780446+00:00", "epoch": 1, "step": 2229, "train_loss": 3.747666120529175, "perplexity": 42.421958635691844, "lr": 0.001305, "grad_norm": 0.254978, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:10.087557+00:00", "epoch": 1, "step": 2230, "train_loss": 3.7161197662353516, "perplexity": 41.10458885468895, "lr": 0.001305, "grad_norm": 0.221153, "tokens_per_sec": 106751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:10.394399+00:00", "epoch": 1, "step": 2231, "train_loss": 3.6074960231781006, "perplexity": 36.87360646543759, "lr": 0.001305, "grad_norm": 0.238948, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:10.701885+00:00", "epoch": 1, "step": 2232, "train_loss": 3.6330416202545166, "perplexity": 37.827699310253195, "lr": 0.001305, "grad_norm": 0.259162, "tokens_per_sec": 106567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:11.009123+00:00", "epoch": 1, "step": 2233, "train_loss": 3.632357120513916, "perplexity": 37.8018151197593, "lr": 0.001305, "grad_norm": 0.265582, "tokens_per_sec": 106655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:11.316975+00:00", "epoch": 1, "step": 2234, "train_loss": 3.838179111480713, "perplexity": 46.44083382147893, "lr": 0.001305, "grad_norm": 0.259818, "tokens_per_sec": 106440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:11.624140+00:00", "epoch": 1, "step": 2235, "train_loss": 3.6212313175201416, "perplexity": 37.38357053963627, "lr": 0.001305, "grad_norm": 0.255668, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:11.930814+00:00", "epoch": 1, "step": 2236, "train_loss": 3.636716842651367, "perplexity": 37.966980305520856, "lr": 0.001305, "grad_norm": 0.232661, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:12.236766+00:00", "epoch": 1, "step": 2237, "train_loss": 3.6814115047454834, "perplexity": 39.702394659196514, "lr": 0.001305, "grad_norm": 0.23169, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:12.544990+00:00", "epoch": 1, "step": 2238, "train_loss": 3.688802480697632, "perplexity": 39.996921181842914, "lr": 0.001305, "grad_norm": 0.218784, "tokens_per_sec": 106312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:12.851093+00:00", "epoch": 1, "step": 2239, "train_loss": 3.772063970565796, "perplexity": 43.46969247540247, "lr": 0.001305, "grad_norm": 0.245743, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:13.158055+00:00", "epoch": 1, "step": 2240, "train_loss": 3.738168239593506, "perplexity": 42.02094731698109, "lr": 0.001305, "grad_norm": 0.247555, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:13.464163+00:00", "epoch": 1, "step": 2241, "train_loss": 3.6693267822265625, "perplexity": 39.2254896725593, "lr": 0.001305, "grad_norm": 0.243428, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:13.771860+00:00", "epoch": 1, "step": 2242, "train_loss": 3.6988866329193115, "perplexity": 40.40229672255054, "lr": 0.001305, "grad_norm": 0.241977, "tokens_per_sec": 106495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:14.078489+00:00", "epoch": 1, "step": 2243, "train_loss": 3.7912065982818604, "perplexity": 44.3098322014699, "lr": 0.001305, "grad_norm": 0.217151, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:14.384881+00:00", "epoch": 1, "step": 2244, "train_loss": 3.681899309158325, "perplexity": 39.72176638693466, "lr": 0.001305, "grad_norm": 0.229342, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:14.690736+00:00", "epoch": 1, "step": 2245, "train_loss": 3.754734992980957, "perplexity": 42.72289644280151, "lr": 0.001305, "grad_norm": 0.23502, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:14.997075+00:00", "epoch": 1, "step": 2246, "train_loss": 3.819969415664673, "perplexity": 45.60281356777803, "lr": 0.001305, "grad_norm": 0.252246, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:15.303402+00:00", "epoch": 1, "step": 2247, "train_loss": 3.7109835147857666, "perplexity": 40.89400661521452, "lr": 0.001305, "grad_norm": 0.259905, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:15.610030+00:00", "epoch": 1, "step": 2248, "train_loss": 3.7519304752349854, "perplexity": 42.60324717923315, "lr": 0.001305, "grad_norm": 0.271679, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:15.916014+00:00", "epoch": 1, "step": 2249, "train_loss": 3.724248170852661, "perplexity": 41.440065181086666, "lr": 0.001305, "grad_norm": 0.26259, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:16.221640+00:00", "epoch": 1, "step": 2250, "train_loss": 3.876439332962036, "perplexity": 48.2520991704543, "lr": 0.001305, "grad_norm": 0.259478, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:42:17.086605+00:00", "step": 2250, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0001_step_0002250.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:18.063121+00:00", "epoch": 1, "step": 2251, "train_loss": 3.6793692111968994, "perplexity": 39.62139345699146, "lr": 0.001305, "grad_norm": 0.3235, "tokens_per_sec": 17793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:18.369384+00:00", "epoch": 1, "step": 2252, "train_loss": 3.6953744888305664, "perplexity": 40.260646927795364, "lr": 0.001305, "grad_norm": 0.365966, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:18.675999+00:00", "epoch": 1, "step": 2253, "train_loss": 3.6604690551757812, "perplexity": 38.87957526100816, "lr": 0.001305, "grad_norm": 0.305495, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:18.982811+00:00", "epoch": 1, "step": 2254, "train_loss": 3.6929023265838623, "perplexity": 40.16123900331988, "lr": 0.001305, "grad_norm": 0.271867, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:19.289005+00:00", "epoch": 1, "step": 2255, "train_loss": 3.7326858043670654, "perplexity": 41.79120055648827, "lr": 0.001305, "grad_norm": 0.287854, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:19.594897+00:00", "epoch": 1, "step": 2256, "train_loss": 3.661794662475586, "perplexity": 38.93114848515648, "lr": 0.001305, "grad_norm": 0.271951, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:19.900778+00:00", "epoch": 1, "step": 2257, "train_loss": 3.738433361053467, "perplexity": 42.03208944882658, "lr": 0.001305, "grad_norm": 0.288978, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:20.207557+00:00", "epoch": 1, "step": 2258, "train_loss": 3.6393983364105225, "perplexity": 38.06892514740731, "lr": 0.001305, "grad_norm": 0.26767, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:20.514020+00:00", "epoch": 1, "step": 2259, "train_loss": 3.805417060852051, "perplexity": 44.94399058379087, "lr": 0.001305, "grad_norm": 0.244461, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:20.819781+00:00", "epoch": 1, "step": 2260, "train_loss": 3.7766683101654053, "perplexity": 43.67030318736121, "lr": 0.001305, "grad_norm": 0.234916, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:21.126380+00:00", "epoch": 1, "step": 2261, "train_loss": 3.6704869270324707, "perplexity": 39.27102332837215, "lr": 0.001305, "grad_norm": 0.253392, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:21.433342+00:00", "epoch": 1, "step": 2262, "train_loss": 3.766458511352539, "perplexity": 43.226706547315004, "lr": 0.001305, "grad_norm": 0.278228, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:21.739328+00:00", "epoch": 1, "step": 2263, "train_loss": 3.6133291721343994, "perplexity": 37.08932424983015, "lr": 0.001305, "grad_norm": 0.250871, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:22.046700+00:00", "epoch": 1, "step": 2264, "train_loss": 3.696077346801758, "perplexity": 40.28895439131173, "lr": 0.001305, "grad_norm": 0.256516, "tokens_per_sec": 106615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:22.354084+00:00", "epoch": 1, "step": 2265, "train_loss": 3.603132963180542, "perplexity": 36.71307516660664, "lr": 0.001305, "grad_norm": 0.297207, "tokens_per_sec": 106597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:22.660642+00:00", "epoch": 1, "step": 2266, "train_loss": 3.736504077911377, "perplexity": 41.95107582146755, "lr": 0.001305, "grad_norm": 0.286045, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:22.966578+00:00", "epoch": 1, "step": 2267, "train_loss": 3.6268937587738037, "perplexity": 37.595853263690444, "lr": 0.001305, "grad_norm": 0.240105, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:23.272652+00:00", "epoch": 1, "step": 2268, "train_loss": 3.7066712379455566, "perplexity": 40.718040018609486, "lr": 0.001305, "grad_norm": 0.251909, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:23.579132+00:00", "epoch": 1, "step": 2269, "train_loss": 3.607491970062256, "perplexity": 36.873457012741845, "lr": 0.001305, "grad_norm": 0.301839, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:23.885983+00:00", "epoch": 1, "step": 2270, "train_loss": 3.688899517059326, "perplexity": 40.000802525866085, "lr": 0.001305, "grad_norm": 0.308666, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:24.191218+00:00", "epoch": 1, "step": 2271, "train_loss": 3.697983980178833, "perplexity": 40.36584393327246, "lr": 0.001305, "grad_norm": 0.249785, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:24.497325+00:00", "epoch": 1, "step": 2272, "train_loss": 3.718319892883301, "perplexity": 41.195123713538976, "lr": 0.001305, "grad_norm": 0.250126, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:24.802842+00:00", "epoch": 1, "step": 2273, "train_loss": 3.665797472000122, "perplexity": 39.08729476052558, "lr": 0.001305, "grad_norm": 0.263699, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:25.110041+00:00", "epoch": 1, "step": 2274, "train_loss": 3.650965929031372, "perplexity": 38.51184780051689, "lr": 0.001305, "grad_norm": 0.236559, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:25.416229+00:00", "epoch": 1, "step": 2275, "train_loss": 3.6341805458068848, "perplexity": 37.8708067870286, "lr": 0.001305, "grad_norm": 0.230117, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:25.723262+00:00", "epoch": 1, "step": 2276, "train_loss": 3.7001121044158936, "perplexity": 40.45183893566537, "lr": 0.001305, "grad_norm": 0.245533, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:26.029123+00:00", "epoch": 1, "step": 2277, "train_loss": 3.7659316062927246, "perplexity": 43.20393617635475, "lr": 0.001305, "grad_norm": 0.234223, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:26.335708+00:00", "epoch": 1, "step": 2278, "train_loss": 3.6179275512695312, "perplexity": 37.26026775490232, "lr": 0.001305, "grad_norm": 0.227676, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:26.641221+00:00", "epoch": 1, "step": 2279, "train_loss": 3.6155624389648438, "perplexity": 37.17224716747692, "lr": 0.001305, "grad_norm": 0.226661, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:26.948004+00:00", "epoch": 1, "step": 2280, "train_loss": 3.594592571258545, "perplexity": 36.40086620712329, "lr": 0.001305, "grad_norm": 0.260729, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:27.255547+00:00", "epoch": 1, "step": 2281, "train_loss": 3.7295982837677, "perplexity": 41.66236835228419, "lr": 0.001305, "grad_norm": 0.236443, "tokens_per_sec": 106488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:27.563134+00:00", "epoch": 1, "step": 2282, "train_loss": 3.68060040473938, "perplexity": 39.67020510288552, "lr": 0.001305, "grad_norm": 0.252975, "tokens_per_sec": 106532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:27.869863+00:00", "epoch": 1, "step": 2283, "train_loss": 3.7046666145324707, "perplexity": 40.63649744064036, "lr": 0.001305, "grad_norm": 0.242007, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:28.176854+00:00", "epoch": 1, "step": 2284, "train_loss": 3.742880344390869, "perplexity": 42.2194216731637, "lr": 0.001305, "grad_norm": 0.246105, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:28.483758+00:00", "epoch": 1, "step": 2285, "train_loss": 3.6179311275482178, "perplexity": 37.26040100824203, "lr": 0.001305, "grad_norm": 0.246731, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:28.791418+00:00", "epoch": 1, "step": 2286, "train_loss": 3.665614604949951, "perplexity": 39.080147635740296, "lr": 0.001305, "grad_norm": 0.23699, "tokens_per_sec": 106506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:29.098227+00:00", "epoch": 1, "step": 2287, "train_loss": 3.703293561935425, "perplexity": 40.580739680223395, "lr": 0.001305, "grad_norm": 0.236813, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:29.405051+00:00", "epoch": 1, "step": 2288, "train_loss": 3.6223909854888916, "perplexity": 37.426948215938175, "lr": 0.001305, "grad_norm": 0.234485, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:29.710936+00:00", "epoch": 1, "step": 2289, "train_loss": 3.723453998565674, "perplexity": 41.407167694613996, "lr": 0.001305, "grad_norm": 0.238563, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:30.017628+00:00", "epoch": 1, "step": 2290, "train_loss": 3.7707748413085938, "perplexity": 43.41369052766109, "lr": 0.001305, "grad_norm": 0.212172, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:30.323773+00:00", "epoch": 1, "step": 2291, "train_loss": 3.7842228412628174, "perplexity": 44.001461147420756, "lr": 0.001305, "grad_norm": 0.239822, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:30.630880+00:00", "epoch": 1, "step": 2292, "train_loss": 3.6830482482910156, "perplexity": 39.76743050637981, "lr": 0.001305, "grad_norm": 0.234927, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:30.937607+00:00", "epoch": 1, "step": 2293, "train_loss": 3.7185044288635254, "perplexity": 41.20272639753673, "lr": 0.001305, "grad_norm": 0.248559, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:31.244594+00:00", "epoch": 1, "step": 2294, "train_loss": 3.7634575366973877, "perplexity": 43.0971787485704, "lr": 0.001305, "grad_norm": 0.25407, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:31.550288+00:00", "epoch": 1, "step": 2295, "train_loss": 3.7021074295043945, "perplexity": 40.53263408425725, "lr": 0.001305, "grad_norm": 0.276778, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:31.856175+00:00", "epoch": 1, "step": 2296, "train_loss": 3.769038200378418, "perplexity": 43.33836196402626, "lr": 0.001305, "grad_norm": 0.250105, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:32.164275+00:00", "epoch": 1, "step": 2297, "train_loss": 3.7175137996673584, "perplexity": 41.161929984201, "lr": 0.001305, "grad_norm": 0.226642, "tokens_per_sec": 106356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:32.472063+00:00", "epoch": 1, "step": 2298, "train_loss": 3.767429828643799, "perplexity": 43.26871379270786, "lr": 0.001305, "grad_norm": 0.259844, "tokens_per_sec": 106463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:32.779901+00:00", "epoch": 1, "step": 2299, "train_loss": 3.7069153785705566, "perplexity": 40.727982159939316, "lr": 0.001305, "grad_norm": 0.285202, "tokens_per_sec": 106445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:33.087169+00:00", "epoch": 1, "step": 2300, "train_loss": 3.669889450073242, "perplexity": 39.24756680483209, "lr": 0.001305, "grad_norm": 0.293903, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:42:38.847720+00:00", "step": 2300, "epoch": 1, "val_loss": 3.764524984359741, "val_ppl": 43.14320729344031, "eval_train_loss": 3.669889450073242, "eval_train_ppl": 39.24756680483209} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:42:39.838838+00:00", "step": 2300, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p7645_epoch_0001_step_0002300.pt", "val_loss": 3.764524984359741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:40.812588+00:00", "epoch": 1, "step": 2301, "train_loss": 3.7253682613372803, "perplexity": 41.486507808895155, "lr": 0.001305, "grad_norm": 0.274034, "tokens_per_sec": 4241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:41.117687+00:00", "epoch": 1, "step": 2302, "train_loss": 3.7088921070098877, "perplexity": 40.808569944393035, "lr": 0.001305, "grad_norm": 0.248436, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:41.422814+00:00", "epoch": 1, "step": 2303, "train_loss": 3.6664175987243652, "perplexity": 39.11154135378816, "lr": 0.001305, "grad_norm": 0.257076, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:41.729988+00:00", "epoch": 1, "step": 2304, "train_loss": 3.705159902572632, "perplexity": 40.656547883737176, "lr": 0.001305, "grad_norm": 0.248489, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:42.037360+00:00", "epoch": 1, "step": 2305, "train_loss": 3.644298553466797, "perplexity": 38.255928949141655, "lr": 0.001305, "grad_norm": 0.247288, "tokens_per_sec": 106609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:42.343479+00:00", "epoch": 1, "step": 2306, "train_loss": 3.7647593021392822, "perplexity": 43.153317698453165, "lr": 0.001305, "grad_norm": 0.252739, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:42.648805+00:00", "epoch": 1, "step": 2307, "train_loss": 3.6834282875061035, "perplexity": 39.782546561620464, "lr": 0.001305, "grad_norm": 0.243397, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:42.953881+00:00", "epoch": 1, "step": 2308, "train_loss": 3.770897626876831, "perplexity": 43.41902142959402, "lr": 0.001305, "grad_norm": 0.239229, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:43.259449+00:00", "epoch": 1, "step": 2309, "train_loss": 3.6514229774475098, "perplexity": 38.52945360260215, "lr": 0.001305, "grad_norm": 0.241877, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:43.567390+00:00", "epoch": 1, "step": 2310, "train_loss": 3.6025824546813965, "perplexity": 36.69286986880203, "lr": 0.001305, "grad_norm": 0.23361, "tokens_per_sec": 106464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:43.874913+00:00", "epoch": 1, "step": 2311, "train_loss": 3.652792453765869, "perplexity": 38.5822549236974, "lr": 0.001305, "grad_norm": 0.221512, "tokens_per_sec": 106502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:44.180977+00:00", "epoch": 1, "step": 2312, "train_loss": 3.60577392578125, "perplexity": 36.81016116890052, "lr": 0.001305, "grad_norm": 0.260913, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:44.486212+00:00", "epoch": 1, "step": 2313, "train_loss": 3.6916682720184326, "perplexity": 40.11170841098712, "lr": 0.001305, "grad_norm": 0.246509, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:44.792016+00:00", "epoch": 1, "step": 2314, "train_loss": 3.6855931282043457, "perplexity": 39.86876272595622, "lr": 0.001305, "grad_norm": 0.275824, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:45.187388+00:00", "epoch": 1, "step": 2315, "train_loss": 3.7527973651885986, "perplexity": 42.64019551896251, "lr": 0.001305, "grad_norm": 0.301538, "tokens_per_sec": 82879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:45.492704+00:00", "epoch": 1, "step": 2316, "train_loss": 3.6385409832000732, "perplexity": 38.03630061958407, "lr": 0.001305, "grad_norm": 0.313144, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:45.799272+00:00", "epoch": 1, "step": 2317, "train_loss": 3.6453187465667725, "perplexity": 38.29497729893046, "lr": 0.001305, "grad_norm": 0.294499, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:46.106421+00:00", "epoch": 1, "step": 2318, "train_loss": 3.63641357421875, "perplexity": 37.95546786468057, "lr": 0.001305, "grad_norm": 0.261672, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:46.413965+00:00", "epoch": 1, "step": 2319, "train_loss": 3.719942092895508, "perplexity": 41.26200467621256, "lr": 0.001305, "grad_norm": 0.243155, "tokens_per_sec": 106548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:46.719851+00:00", "epoch": 1, "step": 2320, "train_loss": 3.724937677383423, "perplexity": 41.46864822963113, "lr": 0.001305, "grad_norm": 0.275458, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:47.024728+00:00", "epoch": 1, "step": 2321, "train_loss": 3.774625539779663, "perplexity": 43.58118583940322, "lr": 0.001305, "grad_norm": 0.240364, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:47.330688+00:00", "epoch": 1, "step": 2322, "train_loss": 3.7097907066345215, "perplexity": 40.84525699104141, "lr": 0.001305, "grad_norm": 0.257378, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:47.636783+00:00", "epoch": 1, "step": 2323, "train_loss": 3.6340107917785645, "perplexity": 37.864378610640614, "lr": 0.001305, "grad_norm": 0.241243, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:47.943152+00:00", "epoch": 1, "step": 2324, "train_loss": 3.653834104537964, "perplexity": 38.62246509817167, "lr": 0.001305, "grad_norm": 0.252785, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:48.249731+00:00", "epoch": 1, "step": 2325, "train_loss": 3.845663547515869, "perplexity": 46.789721255850345, "lr": 0.001305, "grad_norm": 0.258831, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:48.556824+00:00", "epoch": 1, "step": 2326, "train_loss": 3.738083839416504, "perplexity": 42.01740089125134, "lr": 0.001305, "grad_norm": 0.244563, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:48.862315+00:00", "epoch": 1, "step": 2327, "train_loss": 3.615576982498169, "perplexity": 37.172787787223626, "lr": 0.001305, "grad_norm": 0.226404, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:49.168568+00:00", "epoch": 1, "step": 2328, "train_loss": 3.654953718185425, "perplexity": 38.66573155353145, "lr": 0.001305, "grad_norm": 0.239648, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:49.474984+00:00", "epoch": 1, "step": 2329, "train_loss": 3.7242348194122314, "perplexity": 41.439511900218555, "lr": 0.001305, "grad_norm": 0.251311, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:49.782242+00:00", "epoch": 1, "step": 2330, "train_loss": 3.7662086486816406, "perplexity": 43.21590715620149, "lr": 0.001305, "grad_norm": 0.284334, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:50.089144+00:00", "epoch": 1, "step": 2331, "train_loss": 3.7201786041259766, "perplexity": 41.271764757849105, "lr": 0.001305, "grad_norm": 0.323889, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:50.395273+00:00", "epoch": 1, "step": 2332, "train_loss": 3.703627347946167, "perplexity": 40.59428722429879, "lr": 0.001305, "grad_norm": 0.288217, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:50.701338+00:00", "epoch": 1, "step": 2333, "train_loss": 3.666839361190796, "perplexity": 39.12804061307501, "lr": 0.001305, "grad_norm": 0.259301, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:51.008087+00:00", "epoch": 1, "step": 2334, "train_loss": 3.7574543952941895, "perplexity": 42.83923530060031, "lr": 0.001305, "grad_norm": 0.25247, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:51.314908+00:00", "epoch": 1, "step": 2335, "train_loss": 3.7957327365875244, "perplexity": 44.51083917985572, "lr": 0.001305, "grad_norm": 0.252709, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:51.622005+00:00", "epoch": 1, "step": 2336, "train_loss": 3.7579338550567627, "perplexity": 42.85977991495074, "lr": 0.001305, "grad_norm": 0.261559, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:51.928415+00:00", "epoch": 1, "step": 2337, "train_loss": 3.7321231365203857, "perplexity": 41.76769260586518, "lr": 0.001305, "grad_norm": 0.266497, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:52.235476+00:00", "epoch": 1, "step": 2338, "train_loss": 3.741159439086914, "perplexity": 42.14682852735674, "lr": 0.001305, "grad_norm": 0.253703, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:52.541227+00:00", "epoch": 1, "step": 2339, "train_loss": 3.6460955142974854, "perplexity": 38.32473515752055, "lr": 0.001305, "grad_norm": 0.275098, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:52.847925+00:00", "epoch": 1, "step": 2340, "train_loss": 3.688870429992676, "perplexity": 39.99963903677827, "lr": 0.001305, "grad_norm": 0.282704, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:53.154762+00:00", "epoch": 1, "step": 2341, "train_loss": 3.6740856170654297, "perplexity": 39.412602164988314, "lr": 0.001305, "grad_norm": 0.317878, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:53.462847+00:00", "epoch": 1, "step": 2342, "train_loss": 3.609179973602295, "perplexity": 36.93575210108592, "lr": 0.001305, "grad_norm": 0.289333, "tokens_per_sec": 106359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:53.769543+00:00", "epoch": 1, "step": 2343, "train_loss": 3.7941465377807617, "perplexity": 44.44029210549474, "lr": 0.001305, "grad_norm": 0.259431, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:54.076905+00:00", "epoch": 1, "step": 2344, "train_loss": 3.735614061355591, "perplexity": 41.913755279866905, "lr": 0.001305, "grad_norm": 0.254385, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:54.383675+00:00", "epoch": 1, "step": 2345, "train_loss": 3.5804433822631836, "perplexity": 35.88945006546394, "lr": 0.001305, "grad_norm": 0.273615, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:54.690798+00:00", "epoch": 1, "step": 2346, "train_loss": 3.817884683609009, "perplexity": 45.50784294904746, "lr": 0.001305, "grad_norm": 0.260115, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:54.998066+00:00", "epoch": 1, "step": 2347, "train_loss": 3.789112091064453, "perplexity": 44.21712206304183, "lr": 0.001305, "grad_norm": 0.287218, "tokens_per_sec": 106643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:55.305888+00:00", "epoch": 1, "step": 2348, "train_loss": 3.5680766105651855, "perplexity": 35.44834656502129, "lr": 0.001305, "grad_norm": 0.284086, "tokens_per_sec": 106452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:55.612851+00:00", "epoch": 1, "step": 2349, "train_loss": 3.695594310760498, "perplexity": 40.26949807370568, "lr": 0.001305, "grad_norm": 0.271739, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:55.931372+00:00", "epoch": 1, "step": 2350, "train_loss": 3.7208499908447266, "perplexity": 41.2994833764819, "lr": 0.001305, "grad_norm": 0.255971, "tokens_per_sec": 102927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:56.237723+00:00", "epoch": 1, "step": 2351, "train_loss": 3.565847635269165, "perplexity": 35.36942107040927, "lr": 0.001305, "grad_norm": 0.255287, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:56.544893+00:00", "epoch": 1, "step": 2352, "train_loss": 3.746248483657837, "perplexity": 42.36186231042338, "lr": 0.001305, "grad_norm": 0.266384, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:56.852237+00:00", "epoch": 1, "step": 2353, "train_loss": 3.6754419803619385, "perplexity": 39.46609624248527, "lr": 0.001305, "grad_norm": 0.275678, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:57.173586+00:00", "epoch": 1, "step": 2354, "train_loss": 3.6572299003601074, "perplexity": 38.753842042184765, "lr": 0.001305, "grad_norm": 0.226795, "tokens_per_sec": 101971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:57.480837+00:00", "epoch": 1, "step": 2355, "train_loss": 3.7457797527313232, "perplexity": 42.34201064836122, "lr": 0.001305, "grad_norm": 0.238702, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:57.788283+00:00", "epoch": 1, "step": 2356, "train_loss": 3.710669994354248, "perplexity": 40.88118751824341, "lr": 0.001305, "grad_norm": 0.252683, "tokens_per_sec": 106582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:58.094562+00:00", "epoch": 1, "step": 2357, "train_loss": 3.7021102905273438, "perplexity": 40.53275004921945, "lr": 0.001305, "grad_norm": 0.271023, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:58.416024+00:00", "epoch": 1, "step": 2358, "train_loss": 3.69464111328125, "perplexity": 40.231131577979674, "lr": 0.001305, "grad_norm": 0.23157, "tokens_per_sec": 101934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:58.722224+00:00", "epoch": 1, "step": 2359, "train_loss": 3.559748888015747, "perplexity": 35.15436835345292, "lr": 0.001305, "grad_norm": 0.226069, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:59.029748+00:00", "epoch": 1, "step": 2360, "train_loss": 3.6365931034088135, "perplexity": 37.96228259078762, "lr": 0.001305, "grad_norm": 0.236268, "tokens_per_sec": 106609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:59.336373+00:00", "epoch": 1, "step": 2361, "train_loss": 3.6182422637939453, "perplexity": 37.27199587322388, "lr": 0.001305, "grad_norm": 0.274712, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:59.643491+00:00", "epoch": 1, "step": 2362, "train_loss": 3.608473539352417, "perplexity": 36.90966863496836, "lr": 0.001305, "grad_norm": 0.292565, "tokens_per_sec": 106697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:42:59.960402+00:00", "epoch": 1, "step": 2363, "train_loss": 3.657834529876709, "perplexity": 38.77728084414687, "lr": 0.001305, "grad_norm": 0.26991, "tokens_per_sec": 103399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:00.280203+00:00", "epoch": 1, "step": 2364, "train_loss": 3.6865806579589844, "perplexity": 39.90815376213765, "lr": 0.001305, "grad_norm": 0.261352, "tokens_per_sec": 102463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:00.586164+00:00", "epoch": 1, "step": 2365, "train_loss": 3.718078136444092, "perplexity": 41.18516573086894, "lr": 0.001305, "grad_norm": 0.252872, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:00.894308+00:00", "epoch": 1, "step": 2366, "train_loss": 3.8123793601989746, "perplexity": 45.25799593127999, "lr": 0.001305, "grad_norm": 0.299405, "tokens_per_sec": 106340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:01.201529+00:00", "epoch": 1, "step": 2367, "train_loss": 3.6858556270599365, "perplexity": 39.87922960425729, "lr": 0.001305, "grad_norm": 0.306917, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:01.508196+00:00", "epoch": 1, "step": 2368, "train_loss": 3.636484146118164, "perplexity": 37.95814654865973, "lr": 0.001305, "grad_norm": 0.321569, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:01.827583+00:00", "epoch": 1, "step": 2369, "train_loss": 3.791863203048706, "perplexity": 44.338935802247406, "lr": 0.001305, "grad_norm": 0.293371, "tokens_per_sec": 102597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:02.133483+00:00", "epoch": 1, "step": 2370, "train_loss": 3.669825315475464, "perplexity": 39.24504975863702, "lr": 0.001305, "grad_norm": 0.255532, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:02.439387+00:00", "epoch": 1, "step": 2371, "train_loss": 3.6864068508148193, "perplexity": 39.901218042659615, "lr": 0.001305, "grad_norm": 0.290123, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:02.759624+00:00", "epoch": 1, "step": 2372, "train_loss": 3.678133964538574, "perplexity": 39.5724814785238, "lr": 0.001305, "grad_norm": 0.264125, "tokens_per_sec": 102325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:03.066953+00:00", "epoch": 1, "step": 2373, "train_loss": 3.7218995094299316, "perplexity": 41.342850705272944, "lr": 0.001305, "grad_norm": 0.246127, "tokens_per_sec": 106623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:03.375251+00:00", "epoch": 1, "step": 2374, "train_loss": 3.7642228603363037, "perplexity": 43.13017466290335, "lr": 0.001305, "grad_norm": 0.262199, "tokens_per_sec": 106287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:03.689742+00:00", "epoch": 1, "step": 2375, "train_loss": 3.6389060020446777, "perplexity": 38.050187120352305, "lr": 0.001305, "grad_norm": 0.242291, "tokens_per_sec": 104194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:03.995441+00:00", "epoch": 1, "step": 2376, "train_loss": 3.7053565979003906, "perplexity": 40.664545623281924, "lr": 0.001305, "grad_norm": 0.21948, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:04.314422+00:00", "epoch": 1, "step": 2377, "train_loss": 3.666356325149536, "perplexity": 39.109144923252025, "lr": 0.001305, "grad_norm": 0.243611, "tokens_per_sec": 102729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:04.620714+00:00", "epoch": 1, "step": 2378, "train_loss": 3.5510122776031494, "perplexity": 34.84857607191285, "lr": 0.001305, "grad_norm": 0.246877, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:04.943980+00:00", "epoch": 1, "step": 2379, "train_loss": 3.699244737625122, "perplexity": 40.41676756601636, "lr": 0.001305, "grad_norm": 0.239488, "tokens_per_sec": 101367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:05.251311+00:00", "epoch": 1, "step": 2380, "train_loss": 3.5933303833007812, "perplexity": 36.35495045539472, "lr": 0.001305, "grad_norm": 0.234323, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:05.556759+00:00", "epoch": 1, "step": 2381, "train_loss": 3.7169320583343506, "perplexity": 41.13799135190387, "lr": 0.001305, "grad_norm": 0.216742, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:05.862475+00:00", "epoch": 1, "step": 2382, "train_loss": 3.6130211353302, "perplexity": 37.07790113237895, "lr": 0.001305, "grad_norm": 0.224162, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:06.180797+00:00", "epoch": 1, "step": 2383, "train_loss": 3.7199172973632812, "perplexity": 41.260981575530096, "lr": 0.001305, "grad_norm": 0.238065, "tokens_per_sec": 102940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:06.486987+00:00", "epoch": 1, "step": 2384, "train_loss": 3.769200563430786, "perplexity": 43.34539908402806, "lr": 0.001305, "grad_norm": 0.246195, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:06.808724+00:00", "epoch": 1, "step": 2385, "train_loss": 3.6332263946533203, "perplexity": 37.83468954643983, "lr": 0.001305, "grad_norm": 0.253175, "tokens_per_sec": 101848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:07.116824+00:00", "epoch": 1, "step": 2386, "train_loss": 3.7321324348449707, "perplexity": 41.768080977233794, "lr": 0.001305, "grad_norm": 0.262698, "tokens_per_sec": 106355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:07.424342+00:00", "epoch": 1, "step": 2387, "train_loss": 3.6890382766723633, "perplexity": 40.00635340685589, "lr": 0.001305, "grad_norm": 0.254128, "tokens_per_sec": 106556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:07.731345+00:00", "epoch": 1, "step": 2388, "train_loss": 3.646766424179077, "perplexity": 38.350456228342615, "lr": 0.001305, "grad_norm": 0.269443, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:08.037834+00:00", "epoch": 1, "step": 2389, "train_loss": 3.642822027206421, "perplexity": 38.199484746361634, "lr": 0.001305, "grad_norm": 0.274447, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:08.356096+00:00", "epoch": 1, "step": 2390, "train_loss": 3.8069286346435547, "perplexity": 45.01197811318158, "lr": 0.001305, "grad_norm": 0.247663, "tokens_per_sec": 103014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:08.662032+00:00", "epoch": 1, "step": 2391, "train_loss": 3.7031211853027344, "perplexity": 40.57374511183263, "lr": 0.001305, "grad_norm": 0.222481, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:08.983780+00:00", "epoch": 1, "step": 2392, "train_loss": 3.6735587120056152, "perplexity": 39.39184093556587, "lr": 0.001305, "grad_norm": 0.258941, "tokens_per_sec": 101843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:09.291385+00:00", "epoch": 1, "step": 2393, "train_loss": 3.7974162101745605, "perplexity": 44.58583511107443, "lr": 0.001305, "grad_norm": 0.276873, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:09.599132+00:00", "epoch": 1, "step": 2394, "train_loss": 3.6249613761901855, "perplexity": 37.52327383981965, "lr": 0.001305, "grad_norm": 0.264706, "tokens_per_sec": 106476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:09.916500+00:00", "epoch": 1, "step": 2395, "train_loss": 3.7057361602783203, "perplexity": 40.679983284508495, "lr": 0.001305, "grad_norm": 0.242092, "tokens_per_sec": 103249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:10.222527+00:00", "epoch": 1, "step": 2396, "train_loss": 3.7494635581970215, "perplexity": 42.498278031218014, "lr": 0.001305, "grad_norm": 0.238047, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:10.540263+00:00", "epoch": 1, "step": 2397, "train_loss": 3.849126100540161, "perplexity": 46.95201395808644, "lr": 0.001305, "grad_norm": 0.218983, "tokens_per_sec": 103131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:10.847117+00:00", "epoch": 1, "step": 2398, "train_loss": 3.663275957107544, "perplexity": 38.988859719541686, "lr": 0.001305, "grad_norm": 0.225085, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:11.154667+00:00", "epoch": 1, "step": 2399, "train_loss": 3.67745304107666, "perplexity": 39.54554481938129, "lr": 0.001305, "grad_norm": 0.225887, "tokens_per_sec": 106546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:11.461842+00:00", "epoch": 1, "step": 2400, "train_loss": 3.707273483276367, "perplexity": 40.74256965377816, "lr": 0.001305, "grad_norm": 0.23145, "tokens_per_sec": 106733} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:43:17.050354+00:00", "step": 2400, "epoch": 1, "val_loss": 3.7508641958236693, "val_ppl": 42.557844424223894, "eval_train_loss": 3.707273483276367, "eval_train_ppl": 40.74256965377816} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:43:17.976079+00:00", "step": 2400, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p7509_epoch_0001_step_0002400.pt", "val_loss": 3.7508641958236693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:18.968568+00:00", "epoch": 1, "step": 2401, "train_loss": 3.7915713787078857, "perplexity": 44.32599850933454, "lr": 0.001305, "grad_norm": 0.242479, "tokens_per_sec": 4365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:19.274000+00:00", "epoch": 1, "step": 2402, "train_loss": 3.7532918453216553, "perplexity": 42.66128546236557, "lr": 0.001305, "grad_norm": 0.246029, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:19.579767+00:00", "epoch": 1, "step": 2403, "train_loss": 3.710251808166504, "perplexity": 40.86409514443112, "lr": 0.001305, "grad_norm": 0.224039, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:19.884909+00:00", "epoch": 1, "step": 2404, "train_loss": 3.6083579063415527, "perplexity": 36.905400905604054, "lr": 0.001305, "grad_norm": 0.223051, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:20.190525+00:00", "epoch": 1, "step": 2405, "train_loss": 3.6368372440338135, "perplexity": 37.97155185764203, "lr": 0.001305, "grad_norm": 0.248417, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:20.496356+00:00", "epoch": 1, "step": 2406, "train_loss": 3.824563980102539, "perplexity": 45.812820709497885, "lr": 0.001305, "grad_norm": 0.241686, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:20.803076+00:00", "epoch": 1, "step": 2407, "train_loss": 3.66178560256958, "perplexity": 38.93079577420827, "lr": 0.001305, "grad_norm": 0.262025, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:21.108214+00:00", "epoch": 1, "step": 2408, "train_loss": 3.7722926139831543, "perplexity": 43.47963267077838, "lr": 0.001305, "grad_norm": 0.268386, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:21.414576+00:00", "epoch": 1, "step": 2409, "train_loss": 3.6174521446228027, "perplexity": 37.242558185909715, "lr": 0.001305, "grad_norm": 0.264115, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:21.720661+00:00", "epoch": 1, "step": 2410, "train_loss": 3.7404375076293945, "perplexity": 42.11641238651364, "lr": 0.001305, "grad_norm": 0.263709, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:22.026164+00:00", "epoch": 1, "step": 2411, "train_loss": 3.760857343673706, "perplexity": 42.98526332894971, "lr": 0.001305, "grad_norm": 0.271079, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:22.332892+00:00", "epoch": 1, "step": 2412, "train_loss": 3.6947524547576904, "perplexity": 40.23561122094884, "lr": 0.001305, "grad_norm": 0.320346, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:22.638078+00:00", "epoch": 1, "step": 2413, "train_loss": 3.667842149734497, "perplexity": 39.16729744379966, "lr": 0.001305, "grad_norm": 0.281012, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:22.943846+00:00", "epoch": 1, "step": 2414, "train_loss": 3.5617988109588623, "perplexity": 35.22650601275151, "lr": 0.001305, "grad_norm": 0.237283, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:23.301387+00:00", "epoch": 1, "step": 2415, "train_loss": 3.5639700889587402, "perplexity": 35.30307564716893, "lr": 0.001305, "grad_norm": 0.245751, "tokens_per_sec": 91649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:23.606718+00:00", "epoch": 1, "step": 2416, "train_loss": 3.705965995788574, "perplexity": 40.68933406375321, "lr": 0.001305, "grad_norm": 0.2411, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:23.911642+00:00", "epoch": 1, "step": 2417, "train_loss": 3.729360342025757, "perplexity": 41.652456315075426, "lr": 0.001305, "grad_norm": 0.244383, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:24.218265+00:00", "epoch": 1, "step": 2418, "train_loss": 3.6501986980438232, "perplexity": 38.48231164947059, "lr": 0.001305, "grad_norm": 0.249442, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:24.524696+00:00", "epoch": 1, "step": 2419, "train_loss": 3.668269395828247, "perplexity": 39.18403509392848, "lr": 0.001305, "grad_norm": 0.243026, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:24.831033+00:00", "epoch": 1, "step": 2420, "train_loss": 3.677652359008789, "perplexity": 39.55342774117734, "lr": 0.001305, "grad_norm": 0.257181, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:25.136902+00:00", "epoch": 1, "step": 2421, "train_loss": 3.785700559616089, "perplexity": 44.06653097973096, "lr": 0.001305, "grad_norm": 0.272482, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:25.443235+00:00", "epoch": 1, "step": 2422, "train_loss": 3.5508527755737305, "perplexity": 34.843018096572976, "lr": 0.001305, "grad_norm": 0.271959, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:25.749281+00:00", "epoch": 1, "step": 2423, "train_loss": 3.7174015045166016, "perplexity": 41.157307958588376, "lr": 0.001305, "grad_norm": 0.265844, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:26.056585+00:00", "epoch": 1, "step": 2424, "train_loss": 3.75091814994812, "perplexity": 42.56014065740339, "lr": 0.001305, "grad_norm": 0.287745, "tokens_per_sec": 106630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:26.363823+00:00", "epoch": 1, "step": 2425, "train_loss": 3.676992416381836, "perplexity": 39.52733335951328, "lr": 0.001305, "grad_norm": 0.307654, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:26.670729+00:00", "epoch": 1, "step": 2426, "train_loss": 3.651163339614868, "perplexity": 38.51945119733351, "lr": 0.001305, "grad_norm": 0.34124, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:26.976640+00:00", "epoch": 1, "step": 2427, "train_loss": 3.6430747509002686, "perplexity": 38.209139881239096, "lr": 0.001305, "grad_norm": 0.27476, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:27.282671+00:00", "epoch": 1, "step": 2428, "train_loss": 3.7226884365081787, "perplexity": 41.3754800690882, "lr": 0.001305, "grad_norm": 0.224919, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:27.589630+00:00", "epoch": 1, "step": 2429, "train_loss": 3.682973861694336, "perplexity": 39.76447245258688, "lr": 0.001305, "grad_norm": 0.276884, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:27.897290+00:00", "epoch": 1, "step": 2430, "train_loss": 3.777508497238159, "perplexity": 43.70700982962632, "lr": 0.001305, "grad_norm": 0.269257, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:28.202934+00:00", "epoch": 1, "step": 2431, "train_loss": 3.7481648921966553, "perplexity": 42.443122784338456, "lr": 0.001305, "grad_norm": 0.240118, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:28.508667+00:00", "epoch": 1, "step": 2432, "train_loss": 3.7031593322753906, "perplexity": 40.57529290689963, "lr": 0.001305, "grad_norm": 0.241923, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:28.815051+00:00", "epoch": 1, "step": 2433, "train_loss": 3.6152548789978027, "perplexity": 37.160816230302686, "lr": 0.001305, "grad_norm": 0.244035, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:29.121761+00:00", "epoch": 1, "step": 2434, "train_loss": 3.695377826690674, "perplexity": 40.26078131242692, "lr": 0.001305, "grad_norm": 0.241085, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:29.428521+00:00", "epoch": 1, "step": 2435, "train_loss": 3.530471086502075, "perplexity": 34.14004674232755, "lr": 0.001305, "grad_norm": 0.239632, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:29.736106+00:00", "epoch": 1, "step": 2436, "train_loss": 3.6307644844055176, "perplexity": 37.741658500554074, "lr": 0.001305, "grad_norm": 0.239977, "tokens_per_sec": 106532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:30.042546+00:00", "epoch": 1, "step": 2437, "train_loss": 3.4807214736938477, "perplexity": 32.48314936152009, "lr": 0.001305, "grad_norm": 0.226577, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:30.348747+00:00", "epoch": 1, "step": 2438, "train_loss": 3.6896042823791504, "perplexity": 40.02900364066802, "lr": 0.001305, "grad_norm": 0.242816, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:30.655074+00:00", "epoch": 1, "step": 2439, "train_loss": 3.6431379318237305, "perplexity": 38.21155404624527, "lr": 0.001305, "grad_norm": 0.234801, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:30.962899+00:00", "epoch": 1, "step": 2440, "train_loss": 3.6974353790283203, "perplexity": 40.34370525805584, "lr": 0.001305, "grad_norm": 0.211915, "tokens_per_sec": 106508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:31.269543+00:00", "epoch": 1, "step": 2441, "train_loss": 3.735856771469116, "perplexity": 41.923929406800966, "lr": 0.001305, "grad_norm": 0.266881, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:31.576970+00:00", "epoch": 1, "step": 2442, "train_loss": 3.639751672744751, "perplexity": 38.0823786585341, "lr": 0.001305, "grad_norm": 0.2386, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:31.883300+00:00", "epoch": 1, "step": 2443, "train_loss": 3.6677000522613525, "perplexity": 39.16173226521126, "lr": 0.001305, "grad_norm": 0.255935, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:32.189603+00:00", "epoch": 1, "step": 2444, "train_loss": 3.7223594188690186, "perplexity": 41.361869045572945, "lr": 0.001305, "grad_norm": 0.258942, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:32.496118+00:00", "epoch": 1, "step": 2445, "train_loss": 3.680096387863159, "perplexity": 39.650215687954926, "lr": 0.001305, "grad_norm": 0.252563, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:32.802354+00:00", "epoch": 1, "step": 2446, "train_loss": 3.6960716247558594, "perplexity": 40.28872385672507, "lr": 0.001305, "grad_norm": 0.299283, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:33.109736+00:00", "epoch": 1, "step": 2447, "train_loss": 3.6747052669525146, "perplexity": 39.43703174758194, "lr": 0.001305, "grad_norm": 0.291463, "tokens_per_sec": 106603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:33.416634+00:00", "epoch": 1, "step": 2448, "train_loss": 3.6641077995300293, "perplexity": 39.021305800202946, "lr": 0.001305, "grad_norm": 0.234355, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:33.722376+00:00", "epoch": 1, "step": 2449, "train_loss": 3.7015318870544434, "perplexity": 40.50931254464522, "lr": 0.001305, "grad_norm": 0.273657, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:34.029073+00:00", "epoch": 1, "step": 2450, "train_loss": 3.664855718612671, "perplexity": 39.05050149608859, "lr": 0.001305, "grad_norm": 0.262666, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:34.335822+00:00", "epoch": 1, "step": 2451, "train_loss": 3.627958059310913, "perplexity": 37.63588785114949, "lr": 0.001305, "grad_norm": 0.23438, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:34.642377+00:00", "epoch": 1, "step": 2452, "train_loss": 3.737725019454956, "perplexity": 42.00232691366296, "lr": 0.001305, "grad_norm": 0.239225, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:34.949044+00:00", "epoch": 1, "step": 2453, "train_loss": 3.660778284072876, "perplexity": 38.89159980825856, "lr": 0.001305, "grad_norm": 0.222569, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:35.255571+00:00", "epoch": 1, "step": 2454, "train_loss": 3.768953800201416, "perplexity": 43.33470435295921, "lr": 0.001305, "grad_norm": 0.234211, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:35.561246+00:00", "epoch": 1, "step": 2455, "train_loss": 3.694838285446167, "perplexity": 40.23906481937146, "lr": 0.001305, "grad_norm": 0.230318, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:35.867186+00:00", "epoch": 1, "step": 2456, "train_loss": 3.806124448776245, "perplexity": 44.97579466758113, "lr": 0.001305, "grad_norm": 0.242474, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:36.173386+00:00", "epoch": 1, "step": 2457, "train_loss": 3.7771639823913574, "perplexity": 43.691954709335846, "lr": 0.001305, "grad_norm": 0.243043, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:36.480292+00:00", "epoch": 1, "step": 2458, "train_loss": 3.605405807495117, "perplexity": 36.79661316924495, "lr": 0.001305, "grad_norm": 0.228152, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:36.786425+00:00", "epoch": 1, "step": 2459, "train_loss": 3.7642319202423096, "perplexity": 43.13056542000192, "lr": 0.001305, "grad_norm": 0.226856, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:37.092397+00:00", "epoch": 1, "step": 2460, "train_loss": 3.753588914871216, "perplexity": 42.673960713843954, "lr": 0.001305, "grad_norm": 0.234401, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:37.399049+00:00", "epoch": 1, "step": 2461, "train_loss": 3.7166059017181396, "perplexity": 41.124576111700414, "lr": 0.001305, "grad_norm": 0.212796, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:37.705332+00:00", "epoch": 1, "step": 2462, "train_loss": 3.6709399223327637, "perplexity": 39.28881694728619, "lr": 0.001305, "grad_norm": 0.21229, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:38.012529+00:00", "epoch": 1, "step": 2463, "train_loss": 3.640718698501587, "perplexity": 38.11922311147475, "lr": 0.001305, "grad_norm": 0.224212, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:38.318829+00:00", "epoch": 1, "step": 2464, "train_loss": 3.6566824913024902, "perplexity": 38.732633643397804, "lr": 0.001305, "grad_norm": 0.234967, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:38.625024+00:00", "epoch": 1, "step": 2465, "train_loss": 3.6713645458221436, "perplexity": 39.30550344432053, "lr": 0.001305, "grad_norm": 0.217054, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:38.931105+00:00", "epoch": 1, "step": 2466, "train_loss": 3.6708011627197266, "perplexity": 39.28336562447037, "lr": 0.001305, "grad_norm": 0.224288, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:39.236395+00:00", "epoch": 1, "step": 2467, "train_loss": 3.6528561115264893, "perplexity": 38.58471106182081, "lr": 0.001305, "grad_norm": 0.236686, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:39.543535+00:00", "epoch": 1, "step": 2468, "train_loss": 3.6688928604125977, "perplexity": 39.2084725692401, "lr": 0.001305, "grad_norm": 0.272489, "tokens_per_sec": 106688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:39.850776+00:00", "epoch": 1, "step": 2469, "train_loss": 3.694105386734009, "perplexity": 40.20958446496337, "lr": 0.001305, "grad_norm": 0.279378, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:40.158463+00:00", "epoch": 1, "step": 2470, "train_loss": 3.7454373836517334, "perplexity": 42.32751653445735, "lr": 0.001305, "grad_norm": 0.30347, "tokens_per_sec": 106550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:40.464046+00:00", "epoch": 1, "step": 2471, "train_loss": 3.630326509475708, "perplexity": 37.725132219633934, "lr": 0.001305, "grad_norm": 0.300923, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:40.770110+00:00", "epoch": 1, "step": 2472, "train_loss": 3.745316982269287, "perplexity": 42.32242054973904, "lr": 0.001305, "grad_norm": 0.277783, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:41.077083+00:00", "epoch": 1, "step": 2473, "train_loss": 3.725675344467163, "perplexity": 41.49924957185112, "lr": 0.001305, "grad_norm": 0.299516, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:41.384819+00:00", "epoch": 1, "step": 2474, "train_loss": 3.609807252883911, "perplexity": 36.95892840137639, "lr": 0.001305, "grad_norm": 0.270007, "tokens_per_sec": 106481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:41.691084+00:00", "epoch": 1, "step": 2475, "train_loss": 3.7242343425750732, "perplexity": 41.43949214032418, "lr": 0.001305, "grad_norm": 0.272089, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:41.997663+00:00", "epoch": 1, "step": 2476, "train_loss": 3.6407864093780518, "perplexity": 38.12180428486756, "lr": 0.001305, "grad_norm": 0.328026, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:42.303833+00:00", "epoch": 1, "step": 2477, "train_loss": 3.711517572402954, "perplexity": 40.915852203826965, "lr": 0.001305, "grad_norm": 0.319145, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:42.609878+00:00", "epoch": 1, "step": 2478, "train_loss": 3.760044813156128, "perplexity": 42.950350676407425, "lr": 0.001305, "grad_norm": 0.263534, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:42.916356+00:00", "epoch": 1, "step": 2479, "train_loss": 3.727393865585327, "perplexity": 41.570628223921815, "lr": 0.001305, "grad_norm": 0.290825, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:43.223240+00:00", "epoch": 1, "step": 2480, "train_loss": 3.6178207397460938, "perplexity": 37.2562881414778, "lr": 0.001305, "grad_norm": 0.284452, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:43.530683+00:00", "epoch": 1, "step": 2481, "train_loss": 3.6607375144958496, "perplexity": 38.89001424650606, "lr": 0.001305, "grad_norm": 0.244554, "tokens_per_sec": 106529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:43.838081+00:00", "epoch": 1, "step": 2482, "train_loss": 3.6909372806549072, "perplexity": 40.08239781276469, "lr": 0.001305, "grad_norm": 0.290786, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:44.144473+00:00", "epoch": 1, "step": 2483, "train_loss": 3.6600565910339355, "perplexity": 38.86354213713445, "lr": 0.001305, "grad_norm": 0.271499, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:44.450395+00:00", "epoch": 1, "step": 2484, "train_loss": 3.7172160148620605, "perplexity": 41.14967441174717, "lr": 0.001305, "grad_norm": 0.271773, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:44.756473+00:00", "epoch": 1, "step": 2485, "train_loss": 3.7108166217803955, "perplexity": 40.887182261033175, "lr": 0.001305, "grad_norm": 0.265975, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:45.062815+00:00", "epoch": 1, "step": 2486, "train_loss": 3.602358341217041, "perplexity": 36.684647424033294, "lr": 0.001305, "grad_norm": 0.261642, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:45.370496+00:00", "epoch": 1, "step": 2487, "train_loss": 3.69704532623291, "perplexity": 40.32797215161328, "lr": 0.001305, "grad_norm": 0.229316, "tokens_per_sec": 106501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:45.676383+00:00", "epoch": 1, "step": 2488, "train_loss": 3.7626752853393555, "perplexity": 43.063479104481594, "lr": 0.001305, "grad_norm": 0.242589, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:45.982490+00:00", "epoch": 1, "step": 2489, "train_loss": 3.7274036407470703, "perplexity": 41.5710345855226, "lr": 0.001305, "grad_norm": 0.254949, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:46.290245+00:00", "epoch": 1, "step": 2490, "train_loss": 3.736300468444824, "perplexity": 41.9425350548182, "lr": 0.001305, "grad_norm": 0.259626, "tokens_per_sec": 106526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:46.598105+00:00", "epoch": 1, "step": 2491, "train_loss": 3.6572632789611816, "perplexity": 38.75513561280705, "lr": 0.001305, "grad_norm": 0.234767, "tokens_per_sec": 106382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:46.906072+00:00", "epoch": 1, "step": 2492, "train_loss": 3.7406833171844482, "perplexity": 42.12676627559313, "lr": 0.001305, "grad_norm": 0.212234, "tokens_per_sec": 106406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:47.213147+00:00", "epoch": 1, "step": 2493, "train_loss": 3.623339891433716, "perplexity": 37.46247972496086, "lr": 0.001305, "grad_norm": 0.251677, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:47.519807+00:00", "epoch": 1, "step": 2494, "train_loss": 3.6954333782196045, "perplexity": 40.26301792250775, "lr": 0.001305, "grad_norm": 0.246283, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:47.827300+00:00", "epoch": 1, "step": 2495, "train_loss": 3.7014083862304688, "perplexity": 40.50430992008779, "lr": 0.001305, "grad_norm": 0.259488, "tokens_per_sec": 106567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:48.134763+00:00", "epoch": 1, "step": 2496, "train_loss": 3.692002534866333, "perplexity": 40.12511850599814, "lr": 0.001305, "grad_norm": 0.242766, "tokens_per_sec": 106574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:48.441250+00:00", "epoch": 1, "step": 2497, "train_loss": 3.694274663925171, "perplexity": 40.216391606610046, "lr": 0.001305, "grad_norm": 0.26076, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:48.748666+00:00", "epoch": 1, "step": 2498, "train_loss": 3.692512273788452, "perplexity": 40.14557705447153, "lr": 0.001305, "grad_norm": 0.257473, "tokens_per_sec": 106593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:49.055797+00:00", "epoch": 1, "step": 2499, "train_loss": 3.661900520324707, "perplexity": 38.9352698709356, "lr": 0.001305, "grad_norm": 0.257627, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:49.363538+00:00", "epoch": 1, "step": 2500, "train_loss": 3.7079129219055176, "perplexity": 40.768630357888114, "lr": 0.001305, "grad_norm": 0.267045, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:43:54.927961+00:00", "step": 2500, "epoch": 1, "val_loss": 3.744227170944214, "val_ppl": 42.2763222203227, "eval_train_loss": 3.7079129219055176, "eval_train_ppl": 40.768630357888114} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:43:55.815789+00:00", "step": 2500, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p7442_epoch_0001_step_0002500.pt", "val_loss": 3.744227170944214} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:43:57.471571+00:00", "step": 2500, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0001_step_0002500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:58.477001+00:00", "epoch": 1, "step": 2501, "train_loss": 3.6928443908691406, "perplexity": 40.158912300634356, "lr": 0.001305, "grad_norm": 0.245317, "tokens_per_sec": 3595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:58.782237+00:00", "epoch": 1, "step": 2502, "train_loss": 3.6421213150024414, "perplexity": 38.17272727695163, "lr": 0.001305, "grad_norm": 0.250021, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:59.087585+00:00", "epoch": 1, "step": 2503, "train_loss": 3.72227144241333, "perplexity": 41.35823033499644, "lr": 0.001305, "grad_norm": 0.264298, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:59.392133+00:00", "epoch": 1, "step": 2504, "train_loss": 3.6520650386810303, "perplexity": 38.55419981455067, "lr": 0.001305, "grad_norm": 0.266403, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:43:59.697415+00:00", "epoch": 1, "step": 2505, "train_loss": 3.7524044513702393, "perplexity": 42.62344488791826, "lr": 0.001305, "grad_norm": 0.286127, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:00.004328+00:00", "epoch": 1, "step": 2506, "train_loss": 3.6846749782562256, "perplexity": 39.83217402305596, "lr": 0.001305, "grad_norm": 0.247352, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:00.310172+00:00", "epoch": 1, "step": 2507, "train_loss": 3.645069122314453, "perplexity": 38.28541913687879, "lr": 0.0013028489010989011, "grad_norm": 0.258512, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:00.616330+00:00", "epoch": 1, "step": 2508, "train_loss": 3.663907766342163, "perplexity": 39.013501024642125, "lr": 0.001300697802197802, "grad_norm": 0.293485, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:00.921190+00:00", "epoch": 1, "step": 2509, "train_loss": 3.745832920074463, "perplexity": 42.344261920417125, "lr": 0.0012985467032967032, "grad_norm": 0.282485, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:01.226389+00:00", "epoch": 1, "step": 2510, "train_loss": 3.775576114654541, "perplexity": 43.622632715738256, "lr": 0.0012963956043956044, "grad_norm": 0.261254, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:01.532293+00:00", "epoch": 1, "step": 2511, "train_loss": 3.6905932426452637, "perplexity": 40.06861031624687, "lr": 0.0012942445054945054, "grad_norm": 0.281414, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:01.839502+00:00", "epoch": 1, "step": 2512, "train_loss": 3.592512607574463, "perplexity": 36.325232412384075, "lr": 0.0012920934065934065, "grad_norm": 0.286119, "tokens_per_sec": 106665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:02.148105+00:00", "epoch": 1, "step": 2513, "train_loss": 3.5541749000549316, "perplexity": 34.95896342576286, "lr": 0.0012899423076923077, "grad_norm": 0.267355, "tokens_per_sec": 106183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:02.456376+00:00", "epoch": 1, "step": 2514, "train_loss": 3.717710494995117, "perplexity": 41.17002713982067, "lr": 0.0012877912087912087, "grad_norm": 0.23112, "tokens_per_sec": 106297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:03.164696+00:00", "epoch": 1, "step": 2515, "train_loss": 3.6413514614105225, "perplexity": 38.143351174843865, "lr": 0.0012856401098901098, "grad_norm": 0.237272, "tokens_per_sec": 46261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:03.470728+00:00", "epoch": 1, "step": 2516, "train_loss": 3.6321582794189453, "perplexity": 37.79429931269942, "lr": 0.001283489010989011, "grad_norm": 0.259444, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:03.777265+00:00", "epoch": 1, "step": 2517, "train_loss": 3.6179182529449463, "perplexity": 37.259921298449356, "lr": 0.001281337912087912, "grad_norm": 0.238384, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:04.083511+00:00", "epoch": 1, "step": 2518, "train_loss": 3.578852653503418, "perplexity": 35.832405068660115, "lr": 0.001279186813186813, "grad_norm": 0.278427, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:04.390454+00:00", "epoch": 1, "step": 2519, "train_loss": 3.723464250564575, "perplexity": 41.40759220302774, "lr": 0.0012770357142857143, "grad_norm": 0.295294, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:04.696913+00:00", "epoch": 1, "step": 2520, "train_loss": 3.710385799407959, "perplexity": 40.86957094211673, "lr": 0.0012748846153846154, "grad_norm": 0.251997, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:05.003372+00:00", "epoch": 1, "step": 2521, "train_loss": 3.7653441429138184, "perplexity": 43.178562899691684, "lr": 0.0012727335164835164, "grad_norm": 0.244965, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:05.309112+00:00", "epoch": 1, "step": 2522, "train_loss": 3.6313095092773438, "perplexity": 37.76223424977905, "lr": 0.0012705824175824176, "grad_norm": 0.243873, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:05.614675+00:00", "epoch": 1, "step": 2523, "train_loss": 3.686478853225708, "perplexity": 39.904091129989446, "lr": 0.0012684313186813187, "grad_norm": 0.243165, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:05.922088+00:00", "epoch": 1, "step": 2524, "train_loss": 3.6778481006622314, "perplexity": 39.56117075231289, "lr": 0.0012662802197802197, "grad_norm": 0.28423, "tokens_per_sec": 106593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:06.228440+00:00", "epoch": 1, "step": 2525, "train_loss": 3.65230655670166, "perplexity": 38.56351247311902, "lr": 0.0012641291208791209, "grad_norm": 0.241286, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:06.534292+00:00", "epoch": 1, "step": 2526, "train_loss": 3.87038254737854, "perplexity": 47.960729823387034, "lr": 0.001261978021978022, "grad_norm": 0.232442, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:06.840126+00:00", "epoch": 1, "step": 2527, "train_loss": 3.6120262145996094, "perplexity": 37.0410299049113, "lr": 0.001259826923076923, "grad_norm": 0.282703, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:07.145641+00:00", "epoch": 1, "step": 2528, "train_loss": 3.7105331420898438, "perplexity": 40.87559321796506, "lr": 0.0012576758241758241, "grad_norm": 0.255381, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:07.451248+00:00", "epoch": 1, "step": 2529, "train_loss": 3.650197744369507, "perplexity": 38.48227494989583, "lr": 0.0012555247252747253, "grad_norm": 0.24573, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:07.757764+00:00", "epoch": 1, "step": 2530, "train_loss": 3.7595250606536865, "perplexity": 42.92803292451874, "lr": 0.0012533736263736263, "grad_norm": 0.258455, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:08.065236+00:00", "epoch": 1, "step": 2531, "train_loss": 3.5863845348358154, "perplexity": 36.10330942052512, "lr": 0.0012512225274725274, "grad_norm": 0.246781, "tokens_per_sec": 106515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:08.371975+00:00", "epoch": 1, "step": 2532, "train_loss": 3.604328155517578, "perplexity": 36.756980585198214, "lr": 0.0012490714285714286, "grad_norm": 0.256428, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:08.677786+00:00", "epoch": 1, "step": 2533, "train_loss": 3.6893250942230225, "perplexity": 40.01782957685566, "lr": 0.0012469203296703295, "grad_norm": 0.238088, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:08.983199+00:00", "epoch": 1, "step": 2534, "train_loss": 3.728296995162964, "perplexity": 41.608188846333306, "lr": 0.0012447692307692307, "grad_norm": 0.240399, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:09.289750+00:00", "epoch": 1, "step": 2535, "train_loss": 3.7371859550476074, "perplexity": 41.97969105583877, "lr": 0.0012426181318681319, "grad_norm": 0.251741, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:09.595905+00:00", "epoch": 1, "step": 2536, "train_loss": 3.6301541328430176, "perplexity": 37.71862984881861, "lr": 0.0012404670329670328, "grad_norm": 0.271055, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:09.901854+00:00", "epoch": 1, "step": 2537, "train_loss": 3.7205886840820312, "perplexity": 41.28869295204664, "lr": 0.001238315934065934, "grad_norm": 0.230553, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:10.207693+00:00", "epoch": 1, "step": 2538, "train_loss": 3.6413676738739014, "perplexity": 38.14396957754084, "lr": 0.0012361648351648352, "grad_norm": 0.233002, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:10.512859+00:00", "epoch": 1, "step": 2539, "train_loss": 3.577089309692383, "perplexity": 35.76927589452477, "lr": 0.0012340137362637361, "grad_norm": 0.230565, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:10.820187+00:00", "epoch": 1, "step": 2540, "train_loss": 3.559516191482544, "perplexity": 35.146189005499984, "lr": 0.0012318626373626373, "grad_norm": 0.230824, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:11.126561+00:00", "epoch": 1, "step": 2541, "train_loss": 3.6269736289978027, "perplexity": 37.598856172831965, "lr": 0.0012297115384615385, "grad_norm": 0.232912, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:11.433136+00:00", "epoch": 1, "step": 2542, "train_loss": 3.6283915042877197, "perplexity": 37.6522044736098, "lr": 0.0012275604395604394, "grad_norm": 0.235014, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:11.738662+00:00", "epoch": 1, "step": 2543, "train_loss": 3.6643738746643066, "perplexity": 39.03168978078167, "lr": 0.0012254093406593406, "grad_norm": 0.227685, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:12.044157+00:00", "epoch": 1, "step": 2544, "train_loss": 3.650392532348633, "perplexity": 38.489771564567, "lr": 0.0012232582417582418, "grad_norm": 0.228181, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:12.349732+00:00", "epoch": 1, "step": 2545, "train_loss": 3.6404290199279785, "perplexity": 38.10818238850441, "lr": 0.0012211071428571427, "grad_norm": 0.233851, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:12.656185+00:00", "epoch": 1, "step": 2546, "train_loss": 3.661607265472412, "perplexity": 38.923853588142514, "lr": 0.0012189560439560439, "grad_norm": 0.245041, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:12.964323+00:00", "epoch": 1, "step": 2547, "train_loss": 3.6778745651245117, "perplexity": 39.56221773127783, "lr": 0.001216804945054945, "grad_norm": 0.241605, "tokens_per_sec": 106342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:13.270481+00:00", "epoch": 1, "step": 2548, "train_loss": 3.7737889289855957, "perplexity": 43.54474059627933, "lr": 0.0012146538461538462, "grad_norm": 0.229019, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:13.575563+00:00", "epoch": 1, "step": 2549, "train_loss": 3.7854771614074707, "perplexity": 44.05668769517708, "lr": 0.0012125027472527472, "grad_norm": 0.221774, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:13.881953+00:00", "epoch": 1, "step": 2550, "train_loss": 3.663705587387085, "perplexity": 39.005614113081656, "lr": 0.0012103516483516483, "grad_norm": 0.220736, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:14.188065+00:00", "epoch": 1, "step": 2551, "train_loss": 3.726181745529175, "perplexity": 41.52027015788132, "lr": 0.0012082005494505495, "grad_norm": 0.247746, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:14.494792+00:00", "epoch": 1, "step": 2552, "train_loss": 3.6295669078826904, "perplexity": 37.696487029946304, "lr": 0.0012060494505494504, "grad_norm": 0.268094, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:14.802102+00:00", "epoch": 1, "step": 2553, "train_loss": 3.591951847076416, "perplexity": 36.30486836717453, "lr": 0.0012038983516483516, "grad_norm": 0.244321, "tokens_per_sec": 106628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:15.108242+00:00", "epoch": 1, "step": 2554, "train_loss": 3.6301825046539307, "perplexity": 37.71970000983371, "lr": 0.0012017472527472528, "grad_norm": 0.247894, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:15.414140+00:00", "epoch": 1, "step": 2555, "train_loss": 3.5959060192108154, "perplexity": 36.44870826245084, "lr": 0.0011995961538461537, "grad_norm": 0.212877, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:15.720458+00:00", "epoch": 1, "step": 2556, "train_loss": 3.6138365268707275, "perplexity": 37.10814646852266, "lr": 0.001197445054945055, "grad_norm": 0.204124, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:16.026516+00:00", "epoch": 1, "step": 2557, "train_loss": 3.678703784942627, "perplexity": 39.59503711162893, "lr": 0.001195293956043956, "grad_norm": 0.249321, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:16.334348+00:00", "epoch": 1, "step": 2558, "train_loss": 3.590320110321045, "perplexity": 36.24567668481623, "lr": 0.001193142857142857, "grad_norm": 0.202925, "tokens_per_sec": 106447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:16.640690+00:00", "epoch": 1, "step": 2559, "train_loss": 3.627319812774658, "perplexity": 37.61187454011109, "lr": 0.0011909917582417582, "grad_norm": 0.229401, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:16.946101+00:00", "epoch": 1, "step": 2560, "train_loss": 3.679381847381592, "perplexity": 39.6218941234002, "lr": 0.0011888406593406594, "grad_norm": 0.210956, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:17.252031+00:00", "epoch": 1, "step": 2561, "train_loss": 3.789841890335083, "perplexity": 44.24940346451164, "lr": 0.0011866895604395603, "grad_norm": 0.215388, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:17.558693+00:00", "epoch": 1, "step": 2562, "train_loss": 3.66654896736145, "perplexity": 39.11667972117287, "lr": 0.0011845384615384615, "grad_norm": 0.215715, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:17.865140+00:00", "epoch": 1, "step": 2563, "train_loss": 3.7906301021575928, "perplexity": 44.28429511666128, "lr": 0.0011823873626373627, "grad_norm": 0.205346, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:18.173471+00:00", "epoch": 1, "step": 2564, "train_loss": 3.6628897190093994, "perplexity": 38.97380364431689, "lr": 0.0011802362637362636, "grad_norm": 0.209755, "tokens_per_sec": 106275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:18.479256+00:00", "epoch": 1, "step": 2565, "train_loss": 3.721926689147949, "perplexity": 41.34397440756804, "lr": 0.0011780851648351648, "grad_norm": 0.238018, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:18.785437+00:00", "epoch": 1, "step": 2566, "train_loss": 3.610684633255005, "perplexity": 36.991369669277155, "lr": 0.001175934065934066, "grad_norm": 0.236066, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:19.091073+00:00", "epoch": 1, "step": 2567, "train_loss": 3.6970059871673584, "perplexity": 40.32638571807785, "lr": 0.0011737829670329669, "grad_norm": 0.233946, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:19.397513+00:00", "epoch": 1, "step": 2568, "train_loss": 3.6661698818206787, "perplexity": 39.101853963780265, "lr": 0.001171631868131868, "grad_norm": 0.219081, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:19.703679+00:00", "epoch": 1, "step": 2569, "train_loss": 3.647373676300049, "perplexity": 38.373751696622904, "lr": 0.0011694807692307692, "grad_norm": 0.214544, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:20.009717+00:00", "epoch": 1, "step": 2570, "train_loss": 3.687159299850464, "perplexity": 39.931252974157374, "lr": 0.0011673296703296702, "grad_norm": 0.231828, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:20.315221+00:00", "epoch": 1, "step": 2571, "train_loss": 3.6706767082214355, "perplexity": 39.278476937126264, "lr": 0.0011651785714285713, "grad_norm": 0.207225, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:20.621272+00:00", "epoch": 1, "step": 2572, "train_loss": 3.664142370223999, "perplexity": 39.02265481714216, "lr": 0.0011630274725274725, "grad_norm": 0.249426, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:20.927726+00:00", "epoch": 1, "step": 2573, "train_loss": 3.578425168991089, "perplexity": 35.81709054404802, "lr": 0.0011608763736263737, "grad_norm": 0.226233, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:21.234460+00:00", "epoch": 1, "step": 2574, "train_loss": 3.6776387691497803, "perplexity": 39.55289021932345, "lr": 0.0011587252747252746, "grad_norm": 0.254384, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:21.541204+00:00", "epoch": 1, "step": 2575, "train_loss": 3.636096239089966, "perplexity": 37.9434251722834, "lr": 0.0011565741758241758, "grad_norm": 0.250324, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:21.847944+00:00", "epoch": 1, "step": 2576, "train_loss": 3.5939319133758545, "perplexity": 36.376825630099574, "lr": 0.001154423076923077, "grad_norm": 0.242441, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:22.154551+00:00", "epoch": 1, "step": 2577, "train_loss": 3.544309139251709, "perplexity": 34.61576240798438, "lr": 0.001152271978021978, "grad_norm": 0.25745, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:22.462195+00:00", "epoch": 1, "step": 2578, "train_loss": 3.61232852935791, "perplexity": 37.05222964775269, "lr": 0.001150120879120879, "grad_norm": 0.259551, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:22.769612+00:00", "epoch": 1, "step": 2579, "train_loss": 3.703552722930908, "perplexity": 40.59125798802506, "lr": 0.0011479697802197803, "grad_norm": 0.222476, "tokens_per_sec": 106592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:23.076576+00:00", "epoch": 1, "step": 2580, "train_loss": 3.6742568016052246, "perplexity": 39.41934957066134, "lr": 0.0011458186813186812, "grad_norm": 0.206226, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:23.382835+00:00", "epoch": 1, "step": 2581, "train_loss": 3.621964931488037, "perplexity": 37.411005711336095, "lr": 0.0011436675824175824, "grad_norm": 0.206031, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:23.689058+00:00", "epoch": 1, "step": 2582, "train_loss": 3.687073230743408, "perplexity": 39.92781627476922, "lr": 0.0011415164835164836, "grad_norm": 0.219949, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:23.996167+00:00", "epoch": 1, "step": 2583, "train_loss": 3.7052853107452393, "perplexity": 40.6616468668322, "lr": 0.0011393653846153845, "grad_norm": 0.22493, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:24.304346+00:00", "epoch": 1, "step": 2584, "train_loss": 3.5904934406280518, "perplexity": 36.25195970358674, "lr": 0.0011372142857142857, "grad_norm": 0.247872, "tokens_per_sec": 106328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:24.611653+00:00", "epoch": 1, "step": 2585, "train_loss": 3.6340174674987793, "perplexity": 37.86463138348205, "lr": 0.0011350631868131868, "grad_norm": 0.234223, "tokens_per_sec": 106630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:24.917851+00:00", "epoch": 1, "step": 2586, "train_loss": 3.7191925048828125, "perplexity": 41.231086761424336, "lr": 0.0011329120879120878, "grad_norm": 0.228612, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:25.223985+00:00", "epoch": 1, "step": 2587, "train_loss": 3.586872100830078, "perplexity": 36.12091645842659, "lr": 0.001130760989010989, "grad_norm": 0.215476, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:25.530958+00:00", "epoch": 1, "step": 2588, "train_loss": 3.7105751037597656, "perplexity": 40.87730846210254, "lr": 0.0011286098901098901, "grad_norm": 0.211972, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:25.838517+00:00", "epoch": 1, "step": 2589, "train_loss": 3.6055831909179688, "perplexity": 36.803140857372966, "lr": 0.001126458791208791, "grad_norm": 0.218622, "tokens_per_sec": 106543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:26.147063+00:00", "epoch": 1, "step": 2590, "train_loss": 3.5645790100097656, "perplexity": 35.32457897934783, "lr": 0.0011243076923076922, "grad_norm": 0.212, "tokens_per_sec": 106256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:26.452748+00:00", "epoch": 1, "step": 2591, "train_loss": 3.669896364212036, "perplexity": 39.247838168894425, "lr": 0.0011221565934065934, "grad_norm": 0.212675, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:26.758680+00:00", "epoch": 1, "step": 2592, "train_loss": 3.594102144241333, "perplexity": 36.38301861571364, "lr": 0.0011200054945054946, "grad_norm": 0.22874, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:27.065902+00:00", "epoch": 1, "step": 2593, "train_loss": 3.6167685985565186, "perplexity": 37.217109880298835, "lr": 0.0011178543956043955, "grad_norm": 0.225242, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:27.373652+00:00", "epoch": 1, "step": 2594, "train_loss": 3.6589624881744385, "perplexity": 38.82104467708765, "lr": 0.0011157032967032967, "grad_norm": 0.215742, "tokens_per_sec": 106482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:27.680814+00:00", "epoch": 1, "step": 2595, "train_loss": 3.6618878841400146, "perplexity": 38.93477788078291, "lr": 0.0011135521978021977, "grad_norm": 0.21335, "tokens_per_sec": 106673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:27.988010+00:00", "epoch": 1, "step": 2596, "train_loss": 3.606605291366577, "perplexity": 36.840776594628984, "lr": 0.0011114010989010988, "grad_norm": 0.207236, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:28.305955+00:00", "epoch": 1, "step": 2597, "train_loss": 3.734557628631592, "perplexity": 41.86949959789486, "lr": 0.00110925, "grad_norm": 0.202264, "tokens_per_sec": 103062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:28.612932+00:00", "epoch": 1, "step": 2598, "train_loss": 3.640669584274292, "perplexity": 38.11735096126153, "lr": 0.0011070989010989012, "grad_norm": 0.199986, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:28.920545+00:00", "epoch": 1, "step": 2599, "train_loss": 3.6792633533477783, "perplexity": 39.61719944348946, "lr": 0.0011049478021978021, "grad_norm": 0.210676, "tokens_per_sec": 106523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:29.227712+00:00", "epoch": 1, "step": 2600, "train_loss": 3.692474603652954, "perplexity": 40.144064793627955, "lr": 0.0011027967032967033, "grad_norm": 0.194217, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:44:34.936400+00:00", "step": 2600, "epoch": 1, "val_loss": 3.7191783785820007, "val_ppl": 41.23050432280381, "eval_train_loss": 3.692474603652954, "eval_train_ppl": 40.144064793627955} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:44:35.844122+00:00", "step": 2600, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p7192_epoch_0001_step_0002600.pt", "val_loss": 3.7191783785820007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:36.895647+00:00", "epoch": 1, "step": 2601, "train_loss": 3.5921740531921387, "perplexity": 36.312936427309175, "lr": 0.0011006456043956044, "grad_norm": 0.207035, "tokens_per_sec": 4273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:37.201925+00:00", "epoch": 1, "step": 2602, "train_loss": 3.680739402770996, "perplexity": 39.67571956654955, "lr": 0.0010984945054945054, "grad_norm": 0.211614, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:37.508277+00:00", "epoch": 1, "step": 2603, "train_loss": 3.602731227874756, "perplexity": 36.69832919031621, "lr": 0.0010963434065934066, "grad_norm": 0.198434, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:37.815623+00:00", "epoch": 1, "step": 2604, "train_loss": 3.6195549964904785, "perplexity": 37.320956169731765, "lr": 0.0010941923076923077, "grad_norm": 0.213651, "tokens_per_sec": 106617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:38.121621+00:00", "epoch": 1, "step": 2605, "train_loss": 3.5632431507110596, "perplexity": 35.27742181672522, "lr": 0.0010920412087912087, "grad_norm": 0.223167, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:38.427912+00:00", "epoch": 1, "step": 2606, "train_loss": 3.6757256984710693, "perplexity": 39.47729507726696, "lr": 0.0010898901098901099, "grad_norm": 0.213911, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:38.732941+00:00", "epoch": 1, "step": 2607, "train_loss": 3.732059955596924, "perplexity": 41.76505376783846, "lr": 0.001087739010989011, "grad_norm": 0.203819, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:39.039309+00:00", "epoch": 1, "step": 2608, "train_loss": 3.6398637294769287, "perplexity": 38.086646284543775, "lr": 0.001085587912087912, "grad_norm": 0.203455, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:39.345296+00:00", "epoch": 1, "step": 2609, "train_loss": 3.5581867694854736, "perplexity": 35.0994959329987, "lr": 0.0010834368131868131, "grad_norm": 0.211274, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:39.652379+00:00", "epoch": 1, "step": 2610, "train_loss": 3.6839523315429688, "perplexity": 39.80339983145602, "lr": 0.0010812857142857143, "grad_norm": 0.22275, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:39.957668+00:00", "epoch": 1, "step": 2611, "train_loss": 3.6650705337524414, "perplexity": 39.05889103609343, "lr": 0.0010791346153846153, "grad_norm": 0.218813, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:40.262743+00:00", "epoch": 1, "step": 2612, "train_loss": 3.664262294769287, "perplexity": 39.02733487189815, "lr": 0.0010769835164835164, "grad_norm": 0.237751, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:40.568254+00:00", "epoch": 1, "step": 2613, "train_loss": 3.619067907333374, "perplexity": 37.30278196323751, "lr": 0.0010748324175824176, "grad_norm": 0.217129, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:40.874575+00:00", "epoch": 1, "step": 2614, "train_loss": 3.654431104660034, "perplexity": 38.645529598619696, "lr": 0.0010726813186813186, "grad_norm": 0.204786, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:41.180660+00:00", "epoch": 1, "step": 2615, "train_loss": 3.8304502964019775, "perplexity": 46.083284700126214, "lr": 0.0010705302197802197, "grad_norm": 0.258124, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:41.486895+00:00", "epoch": 1, "step": 2616, "train_loss": 3.654646635055542, "perplexity": 38.65385978257037, "lr": 0.001068379120879121, "grad_norm": 0.257175, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:41.792779+00:00", "epoch": 1, "step": 2617, "train_loss": 3.652284860610962, "perplexity": 38.56267580473101, "lr": 0.0010662280219780218, "grad_norm": 0.224596, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:42.098004+00:00", "epoch": 1, "step": 2618, "train_loss": 3.627408981323242, "perplexity": 37.61522848590437, "lr": 0.001064076923076923, "grad_norm": 0.208475, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:42.403564+00:00", "epoch": 1, "step": 2619, "train_loss": 3.5493416786193848, "perplexity": 34.79040667853495, "lr": 0.0010619258241758242, "grad_norm": 0.229246, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:42.709675+00:00", "epoch": 1, "step": 2620, "train_loss": 3.5813260078430176, "perplexity": 35.921140995693385, "lr": 0.0010597747252747251, "grad_norm": 0.219873, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:43.016183+00:00", "epoch": 1, "step": 2621, "train_loss": 3.5759353637695312, "perplexity": 35.728023890311526, "lr": 0.0010576236263736263, "grad_norm": 0.216271, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:43.322216+00:00", "epoch": 1, "step": 2622, "train_loss": 3.693906307220459, "perplexity": 40.20158035720142, "lr": 0.0010554725274725275, "grad_norm": 0.217148, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:43.627238+00:00", "epoch": 1, "step": 2623, "train_loss": 3.684295654296875, "perplexity": 39.81706759039305, "lr": 0.0010533214285714286, "grad_norm": 0.229928, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:43.932874+00:00", "epoch": 1, "step": 2624, "train_loss": 3.625026226043701, "perplexity": 37.525707297535426, "lr": 0.0010511703296703296, "grad_norm": 0.212675, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:44.239729+00:00", "epoch": 1, "step": 2625, "train_loss": 3.559373378753662, "perplexity": 35.141170040732874, "lr": 0.0010490192307692308, "grad_norm": 0.217637, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:44.546078+00:00", "epoch": 1, "step": 2626, "train_loss": 3.661221742630005, "perplexity": 38.90885044568265, "lr": 0.0010468681318681317, "grad_norm": 0.215259, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:44.851884+00:00", "epoch": 1, "step": 2627, "train_loss": 3.668325662612915, "perplexity": 39.18623991562206, "lr": 0.0010447170329670329, "grad_norm": 0.212633, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:45.156690+00:00", "epoch": 1, "step": 2628, "train_loss": 3.6763789653778076, "perplexity": 39.503092713161934, "lr": 0.001042565934065934, "grad_norm": 0.204256, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:45.461525+00:00", "epoch": 1, "step": 2629, "train_loss": 3.6864311695098877, "perplexity": 39.90218840001291, "lr": 0.0010404148351648352, "grad_norm": 0.205395, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:45.767420+00:00", "epoch": 1, "step": 2630, "train_loss": 3.6373445987701416, "perplexity": 37.99082179225567, "lr": 0.0010382637362637362, "grad_norm": 0.220755, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:46.074304+00:00", "epoch": 1, "step": 2631, "train_loss": 3.592463970184326, "perplexity": 36.32346569084814, "lr": 0.0010361126373626373, "grad_norm": 0.203678, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:46.380566+00:00", "epoch": 1, "step": 2632, "train_loss": 3.5330708026885986, "perplexity": 34.228916642658994, "lr": 0.0010339615384615383, "grad_norm": 0.205318, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:46.685702+00:00", "epoch": 1, "step": 2633, "train_loss": 3.688206434249878, "perplexity": 39.973088262520434, "lr": 0.0010318104395604395, "grad_norm": 0.229136, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:46.990751+00:00", "epoch": 1, "step": 2634, "train_loss": 3.6418330669403076, "perplexity": 38.16172564796298, "lr": 0.0010296593406593406, "grad_norm": 0.250693, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:47.295746+00:00", "epoch": 1, "step": 2635, "train_loss": 3.597669839859009, "perplexity": 36.51305397715227, "lr": 0.0010275082417582418, "grad_norm": 0.252877, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:47.602555+00:00", "epoch": 1, "step": 2636, "train_loss": 3.6789684295654297, "perplexity": 39.605517111966904, "lr": 0.001025357142857143, "grad_norm": 0.254298, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:47.909111+00:00", "epoch": 1, "step": 2637, "train_loss": 3.582516670227051, "perplexity": 35.963936419463316, "lr": 0.001023206043956044, "grad_norm": 0.24719, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:48.215466+00:00", "epoch": 1, "step": 2638, "train_loss": 3.7650439739227295, "perplexity": 43.16560397905978, "lr": 0.001021054945054945, "grad_norm": 0.241578, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:48.520582+00:00", "epoch": 1, "step": 2639, "train_loss": 3.585139751434326, "perplexity": 36.05839657939852, "lr": 0.001018903846153846, "grad_norm": 0.236165, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:48.826549+00:00", "epoch": 1, "step": 2640, "train_loss": 3.6256978511810303, "perplexity": 37.55091897130172, "lr": 0.0010167527472527472, "grad_norm": 0.207676, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:49.132374+00:00", "epoch": 1, "step": 2641, "train_loss": 3.6277594566345215, "perplexity": 37.62841400528134, "lr": 0.0010146016483516484, "grad_norm": 0.22034, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:49.438953+00:00", "epoch": 1, "step": 2642, "train_loss": 3.6140732765197754, "perplexity": 37.116932849221136, "lr": 0.0010124505494505495, "grad_norm": 0.197804, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:49.745581+00:00", "epoch": 1, "step": 2643, "train_loss": 3.6223721504211426, "perplexity": 37.426243283471635, "lr": 0.0010102994505494505, "grad_norm": 0.200843, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:50.051162+00:00", "epoch": 1, "step": 2644, "train_loss": 3.6112022399902344, "perplexity": 37.01052160752186, "lr": 0.0010081483516483517, "grad_norm": 0.20972, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:50.356851+00:00", "epoch": 1, "step": 2645, "train_loss": 3.6493289470672607, "perplexity": 38.448856172409, "lr": 0.0010059972527472526, "grad_norm": 0.197109, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:50.662706+00:00", "epoch": 1, "step": 2646, "train_loss": 3.737293004989624, "perplexity": 41.98418521987789, "lr": 0.0010038461538461538, "grad_norm": 0.210681, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:50.968755+00:00", "epoch": 1, "step": 2647, "train_loss": 3.677067995071411, "perplexity": 39.530320896466634, "lr": 0.001001695054945055, "grad_norm": 0.208102, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:51.275171+00:00", "epoch": 1, "step": 2648, "train_loss": 3.567856788635254, "perplexity": 35.44055509746557, "lr": 0.0009995439560439561, "grad_norm": 0.205603, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:51.581612+00:00", "epoch": 1, "step": 2649, "train_loss": 3.5794677734375, "perplexity": 35.85445307569307, "lr": 0.000997392857142857, "grad_norm": 0.21515, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:51.887675+00:00", "epoch": 1, "step": 2650, "train_loss": 3.617781162261963, "perplexity": 37.25481366050343, "lr": 0.0009952417582417582, "grad_norm": 0.206265, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:52.193199+00:00", "epoch": 1, "step": 2651, "train_loss": 3.55989933013916, "perplexity": 35.15965744911667, "lr": 0.0009930906593406592, "grad_norm": 0.209159, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:52.499420+00:00", "epoch": 1, "step": 2652, "train_loss": 3.653049945831299, "perplexity": 38.592190827359914, "lr": 0.0009909395604395604, "grad_norm": 0.194787, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:52.806575+00:00", "epoch": 1, "step": 2653, "train_loss": 3.619171142578125, "perplexity": 37.3066331238477, "lr": 0.0009887884615384615, "grad_norm": 0.215527, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:53.112719+00:00", "epoch": 1, "step": 2654, "train_loss": 3.6483213901519775, "perplexity": 38.410136271015574, "lr": 0.0009866373626373627, "grad_norm": 0.198809, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:53.419920+00:00", "epoch": 1, "step": 2655, "train_loss": 3.671199083328247, "perplexity": 39.29900039571693, "lr": 0.0009844862637362639, "grad_norm": 0.207857, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:53.726311+00:00", "epoch": 1, "step": 2656, "train_loss": 3.649864912033081, "perplexity": 38.46946893565932, "lr": 0.0009823351648351648, "grad_norm": 0.208426, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:54.031704+00:00", "epoch": 1, "step": 2657, "train_loss": 3.662522315979004, "perplexity": 38.959487180869, "lr": 0.0009801840659340658, "grad_norm": 0.2102, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:54.337763+00:00", "epoch": 1, "step": 2658, "train_loss": 3.5670790672302246, "perplexity": 35.413002934502344, "lr": 0.000978032967032967, "grad_norm": 0.197395, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:54.644085+00:00", "epoch": 1, "step": 2659, "train_loss": 3.4887044429779053, "perplexity": 32.743499144658436, "lr": 0.0009758818681318681, "grad_norm": 0.220356, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:54.951369+00:00", "epoch": 1, "step": 2660, "train_loss": 3.699174165725708, "perplexity": 40.41391537860438, "lr": 0.0009737307692307693, "grad_norm": 0.214628, "tokens_per_sec": 106693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:55.258466+00:00", "epoch": 1, "step": 2661, "train_loss": 3.616102457046509, "perplexity": 37.19232627413585, "lr": 0.0009715796703296702, "grad_norm": 0.215617, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:55.565429+00:00", "epoch": 1, "step": 2662, "train_loss": 3.5596959590911865, "perplexity": 35.152507719783465, "lr": 0.0009694285714285714, "grad_norm": 0.220721, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:55.871201+00:00", "epoch": 1, "step": 2663, "train_loss": 3.712178945541382, "perplexity": 40.942921799975615, "lr": 0.0009672774725274726, "grad_norm": 0.212898, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:56.177075+00:00", "epoch": 1, "step": 2664, "train_loss": 3.7324697971343994, "perplexity": 41.782174329806494, "lr": 0.0009651263736263735, "grad_norm": 0.206871, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:56.483943+00:00", "epoch": 1, "step": 2665, "train_loss": 3.653137445449829, "perplexity": 38.59556777707431, "lr": 0.0009629752747252747, "grad_norm": 0.197351, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:56.790827+00:00", "epoch": 1, "step": 2666, "train_loss": 3.7048511505126953, "perplexity": 40.64399702847908, "lr": 0.0009608241758241758, "grad_norm": 0.203172, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:57.097244+00:00", "epoch": 1, "step": 2667, "train_loss": 3.7391741275787354, "perplexity": 42.063236948765706, "lr": 0.0009586730769230768, "grad_norm": 0.213525, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:57.403368+00:00", "epoch": 1, "step": 2668, "train_loss": 3.736654043197632, "perplexity": 41.95736749831658, "lr": 0.000956521978021978, "grad_norm": 0.192832, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:57.711098+00:00", "epoch": 1, "step": 2669, "train_loss": 3.628734827041626, "perplexity": 37.665133551436625, "lr": 0.0009543708791208791, "grad_norm": 0.20334, "tokens_per_sec": 106484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:58.017644+00:00", "epoch": 1, "step": 2670, "train_loss": 3.6155080795288086, "perplexity": 37.17022656000478, "lr": 0.0009522197802197802, "grad_norm": 0.195257, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:58.324009+00:00", "epoch": 1, "step": 2671, "train_loss": 3.6715970039367676, "perplexity": 39.31464138959916, "lr": 0.0009500686813186813, "grad_norm": 0.204725, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:58.632776+00:00", "epoch": 1, "step": 2672, "train_loss": 3.7303237915039062, "perplexity": 41.69260569021742, "lr": 0.0009479175824175824, "grad_norm": 0.187667, "tokens_per_sec": 106126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:58.938988+00:00", "epoch": 1, "step": 2673, "train_loss": 3.561356544494629, "perplexity": 35.2109299551278, "lr": 0.0009457664835164834, "grad_norm": 0.206401, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:59.244845+00:00", "epoch": 1, "step": 2674, "train_loss": 3.5182340145111084, "perplexity": 33.72481830426251, "lr": 0.0009436153846153845, "grad_norm": 0.194108, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:59.550758+00:00", "epoch": 1, "step": 2675, "train_loss": 3.654120445251465, "perplexity": 38.633525865883534, "lr": 0.0009414642857142857, "grad_norm": 0.199078, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:44:59.856731+00:00", "epoch": 1, "step": 2676, "train_loss": 3.5986135005950928, "perplexity": 36.54752617501671, "lr": 0.0009393131868131868, "grad_norm": 0.212451, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:00.162826+00:00", "epoch": 1, "step": 2677, "train_loss": 3.6478986740112305, "perplexity": 38.39390311769521, "lr": 0.0009371620879120878, "grad_norm": 0.224352, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:00.469322+00:00", "epoch": 1, "step": 2678, "train_loss": 3.6276957988739014, "perplexity": 37.62601874094947, "lr": 0.000935010989010989, "grad_norm": 0.213751, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:00.775128+00:00", "epoch": 1, "step": 2679, "train_loss": 3.6802690029144287, "perplexity": 39.657060502710785, "lr": 0.0009328598901098902, "grad_norm": 0.203774, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:01.081274+00:00", "epoch": 1, "step": 2680, "train_loss": 3.6776068210601807, "perplexity": 39.55162660022802, "lr": 0.0009307087912087911, "grad_norm": 0.237057, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:01.387457+00:00", "epoch": 1, "step": 2681, "train_loss": 3.4692938327789307, "perplexity": 32.11405654161005, "lr": 0.0009285576923076923, "grad_norm": 0.204637, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:01.694345+00:00", "epoch": 1, "step": 2682, "train_loss": 3.6699411869049072, "perplexity": 39.24959740211702, "lr": 0.0009264065934065933, "grad_norm": 0.218922, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:02.001007+00:00", "epoch": 1, "step": 2683, "train_loss": 3.7242729663848877, "perplexity": 41.4410927222975, "lr": 0.0009242554945054944, "grad_norm": 0.216744, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:02.306878+00:00", "epoch": 1, "step": 2684, "train_loss": 3.7415974140167236, "perplexity": 42.16529182455821, "lr": 0.0009221043956043956, "grad_norm": 0.201113, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:02.612941+00:00", "epoch": 1, "step": 2685, "train_loss": 3.6023497581481934, "perplexity": 36.68433255853006, "lr": 0.0009199532967032966, "grad_norm": 0.21228, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:02.918780+00:00", "epoch": 1, "step": 2686, "train_loss": 3.605999231338501, "perplexity": 36.81845563713498, "lr": 0.0009178021978021977, "grad_norm": 0.205193, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:03.224792+00:00", "epoch": 1, "step": 2687, "train_loss": 3.6726582050323486, "perplexity": 39.35638427499465, "lr": 0.0009156510989010989, "grad_norm": 0.204577, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:03.531234+00:00", "epoch": 1, "step": 2688, "train_loss": 3.6282660961151123, "perplexity": 37.6474828755218, "lr": 0.0009135, "grad_norm": 0.213523, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:03.837821+00:00", "epoch": 1, "step": 2689, "train_loss": 3.5917303562164307, "perplexity": 36.296828061118255, "lr": 0.0009113489010989011, "grad_norm": 0.224137, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:04.143498+00:00", "epoch": 1, "step": 2690, "train_loss": 3.5047481060028076, "perplexity": 33.273061512254344, "lr": 0.0009091978021978022, "grad_norm": 0.245632, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:04.448886+00:00", "epoch": 1, "step": 2691, "train_loss": 3.6565258502960205, "perplexity": 38.72656699983566, "lr": 0.0009070467032967032, "grad_norm": 0.202299, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:04.754669+00:00", "epoch": 1, "step": 2692, "train_loss": 3.5994067192077637, "perplexity": 36.576527853840894, "lr": 0.0009048956043956043, "grad_norm": 0.219144, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:05.061452+00:00", "epoch": 1, "step": 2693, "train_loss": 3.667813777923584, "perplexity": 39.16618621240651, "lr": 0.0009027445054945054, "grad_norm": 0.226401, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:05.368881+00:00", "epoch": 1, "step": 2694, "train_loss": 3.662705659866333, "perplexity": 38.96663081954832, "lr": 0.0009005934065934066, "grad_norm": 0.212628, "tokens_per_sec": 106587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:05.675622+00:00", "epoch": 1, "step": 2695, "train_loss": 3.649837017059326, "perplexity": 38.46839584579997, "lr": 0.0008984423076923077, "grad_norm": 0.220507, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:05.981789+00:00", "epoch": 1, "step": 2696, "train_loss": 3.6269147396087646, "perplexity": 37.5966420643578, "lr": 0.0008962912087912087, "grad_norm": 0.198471, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:06.287235+00:00", "epoch": 1, "step": 2697, "train_loss": 3.5932650566101074, "perplexity": 36.35257558436393, "lr": 0.0008941401098901099, "grad_norm": 0.218903, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:06.593433+00:00", "epoch": 1, "step": 2698, "train_loss": 3.567095994949341, "perplexity": 35.413602400942864, "lr": 0.0008919890109890108, "grad_norm": 0.204669, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:06.900733+00:00", "epoch": 1, "step": 2699, "train_loss": 3.6010024547576904, "perplexity": 36.63494091313269, "lr": 0.000889837912087912, "grad_norm": 0.2194, "tokens_per_sec": 106633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:07.209544+00:00", "epoch": 1, "step": 2700, "train_loss": 3.625822067260742, "perplexity": 37.55558368895639, "lr": 0.0008876868131868132, "grad_norm": 0.215434, "tokens_per_sec": 106168} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:45:12.756774+00:00", "step": 2700, "epoch": 1, "val_loss": 3.692422604560852, "val_ppl": 40.141977392977346, "eval_train_loss": 3.625822067260742, "eval_train_ppl": 37.55558368895639} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:45:13.657301+00:00", "step": 2700, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p6924_epoch_0001_step_0002700.pt", "val_loss": 3.692422604560852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:14.878934+00:00", "epoch": 1, "step": 2701, "train_loss": 3.6094963550567627, "perplexity": 36.947439736835626, "lr": 0.0008855357142857142, "grad_norm": 0.217547, "tokens_per_sec": 4272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:15.183217+00:00", "epoch": 1, "step": 2702, "train_loss": 3.622248888015747, "perplexity": 37.42163031900798, "lr": 0.0008833846153846153, "grad_norm": 0.220849, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:15.488041+00:00", "epoch": 1, "step": 2703, "train_loss": 3.6751480102539062, "perplexity": 39.454496095041094, "lr": 0.0008812335164835165, "grad_norm": 0.233006, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:15.793324+00:00", "epoch": 1, "step": 2704, "train_loss": 3.574629545211792, "perplexity": 35.6814000214659, "lr": 0.0008790824175824175, "grad_norm": 0.213458, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:16.098971+00:00", "epoch": 1, "step": 2705, "train_loss": 3.5856878757476807, "perplexity": 36.07816648095115, "lr": 0.0008769313186813186, "grad_norm": 0.221103, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:16.404264+00:00", "epoch": 1, "step": 2706, "train_loss": 3.6252570152282715, "perplexity": 37.53436882437796, "lr": 0.0008747802197802198, "grad_norm": 0.226929, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:16.709949+00:00", "epoch": 1, "step": 2707, "train_loss": 3.7060813903808594, "perplexity": 40.69402966378606, "lr": 0.0008726291208791208, "grad_norm": 0.198455, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:17.016040+00:00", "epoch": 1, "step": 2708, "train_loss": 3.5795936584472656, "perplexity": 35.858966897973964, "lr": 0.0008704780219780219, "grad_norm": 0.234684, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:17.322079+00:00", "epoch": 1, "step": 2709, "train_loss": 3.760575294494629, "perplexity": 42.97314108033087, "lr": 0.000868326923076923, "grad_norm": 0.228379, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:17.629093+00:00", "epoch": 1, "step": 2710, "train_loss": 3.6489460468292236, "perplexity": 38.43413691441186, "lr": 0.0008661758241758241, "grad_norm": 0.23511, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:17.935694+00:00", "epoch": 1, "step": 2711, "train_loss": 3.5729825496673584, "perplexity": 35.62268128263535, "lr": 0.0008640247252747252, "grad_norm": 0.218642, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:18.241186+00:00", "epoch": 1, "step": 2712, "train_loss": 3.73724627494812, "perplexity": 41.9822233429997, "lr": 0.0008618736263736263, "grad_norm": 0.211284, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:18.546723+00:00", "epoch": 1, "step": 2713, "train_loss": 3.658553123474121, "perplexity": 38.805155964128396, "lr": 0.0008597225274725275, "grad_norm": 0.231898, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:18.853490+00:00", "epoch": 1, "step": 2714, "train_loss": 3.6884548664093018, "perplexity": 39.98302009679877, "lr": 0.0008575714285714286, "grad_norm": 0.213746, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:19.216295+00:00", "epoch": 1, "step": 2715, "train_loss": 3.6734259128570557, "perplexity": 39.3866100799637, "lr": 0.0008554203296703296, "grad_norm": 0.20591, "tokens_per_sec": 90318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:19.521745+00:00", "epoch": 1, "step": 2716, "train_loss": 3.5780372619628906, "perplexity": 35.803199537280136, "lr": 0.0008532692307692307, "grad_norm": 0.22418, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:19.828807+00:00", "epoch": 1, "step": 2717, "train_loss": 3.5694544315338135, "perplexity": 35.49722170308479, "lr": 0.0008511181318681317, "grad_norm": 0.196165, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:20.136561+00:00", "epoch": 1, "step": 2718, "train_loss": 3.690031051635742, "perplexity": 40.046090434593694, "lr": 0.0008489670329670329, "grad_norm": 0.203495, "tokens_per_sec": 106474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:20.442439+00:00", "epoch": 1, "step": 2719, "train_loss": 3.7207727432250977, "perplexity": 41.29629321291702, "lr": 0.0008468159340659341, "grad_norm": 0.19862, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:20.748131+00:00", "epoch": 1, "step": 2720, "train_loss": 3.5516560077667236, "perplexity": 34.87101637346253, "lr": 0.0008446648351648351, "grad_norm": 0.221749, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:21.053934+00:00", "epoch": 1, "step": 2721, "train_loss": 3.5647706985473633, "perplexity": 35.33135094526701, "lr": 0.0008425137362637362, "grad_norm": 0.191659, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:21.360230+00:00", "epoch": 1, "step": 2722, "train_loss": 3.66355037689209, "perplexity": 38.9995605022117, "lr": 0.0008403626373626374, "grad_norm": 0.199147, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:21.666648+00:00", "epoch": 1, "step": 2723, "train_loss": 3.6379315853118896, "perplexity": 38.01312843956807, "lr": 0.0008382115384615384, "grad_norm": 0.216944, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:21.974396+00:00", "epoch": 1, "step": 2724, "train_loss": 3.544459581375122, "perplexity": 34.62097046853062, "lr": 0.0008360604395604395, "grad_norm": 0.20331, "tokens_per_sec": 106474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:22.279982+00:00", "epoch": 1, "step": 2725, "train_loss": 3.5707435607910156, "perplexity": 35.54301171841449, "lr": 0.0008339093406593407, "grad_norm": 0.194675, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:22.585975+00:00", "epoch": 1, "step": 2726, "train_loss": 3.7382137775421143, "perplexity": 42.022860908290674, "lr": 0.0008317582417582417, "grad_norm": 0.207006, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:22.892715+00:00", "epoch": 1, "step": 2727, "train_loss": 3.67134952545166, "perplexity": 39.30491306553063, "lr": 0.0008296071428571428, "grad_norm": 0.193526, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:23.200237+00:00", "epoch": 1, "step": 2728, "train_loss": 3.60429048538208, "perplexity": 36.75559597083854, "lr": 0.000827456043956044, "grad_norm": 0.203295, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:23.507054+00:00", "epoch": 1, "step": 2729, "train_loss": 3.5989041328430176, "perplexity": 36.55814960838644, "lr": 0.000825304945054945, "grad_norm": 0.194787, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:23.813434+00:00", "epoch": 1, "step": 2730, "train_loss": 3.657431125640869, "perplexity": 38.76164107958573, "lr": 0.0008231538461538461, "grad_norm": 0.199574, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:24.118878+00:00", "epoch": 1, "step": 2731, "train_loss": 3.66530704498291, "perplexity": 39.068129994988816, "lr": 0.0008210027472527472, "grad_norm": 0.193902, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:24.425083+00:00", "epoch": 1, "step": 2732, "train_loss": 3.717879295349121, "perplexity": 41.17697724154953, "lr": 0.0008188516483516484, "grad_norm": 0.196754, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:24.731840+00:00", "epoch": 1, "step": 2733, "train_loss": 3.663503408432007, "perplexity": 38.99772879592754, "lr": 0.0008167005494505494, "grad_norm": 0.208192, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:25.039740+00:00", "epoch": 1, "step": 2734, "train_loss": 3.6135315895080566, "perplexity": 37.096832533313396, "lr": 0.0008145494505494505, "grad_norm": 0.216466, "tokens_per_sec": 106419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:25.346360+00:00", "epoch": 1, "step": 2735, "train_loss": 3.714486837387085, "perplexity": 41.03752275773662, "lr": 0.0008123983516483516, "grad_norm": 0.199508, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:25.652887+00:00", "epoch": 1, "step": 2736, "train_loss": 3.677039384841919, "perplexity": 39.52918994109241, "lr": 0.0008102472527472528, "grad_norm": 0.214241, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:25.958595+00:00", "epoch": 1, "step": 2737, "train_loss": 3.6140730381011963, "perplexity": 37.1169239998558, "lr": 0.0008080961538461538, "grad_norm": 0.203806, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:26.265096+00:00", "epoch": 1, "step": 2738, "train_loss": 3.5723636150360107, "perplexity": 35.60063999329034, "lr": 0.0008059450549450549, "grad_norm": 0.213197, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:26.571538+00:00", "epoch": 1, "step": 2739, "train_loss": 3.59676194190979, "perplexity": 36.47991889423965, "lr": 0.000803793956043956, "grad_norm": 0.207118, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:26.877977+00:00", "epoch": 1, "step": 2740, "train_loss": 3.6018478870391846, "perplexity": 36.66592637101887, "lr": 0.0008016428571428571, "grad_norm": 0.204755, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:27.183769+00:00", "epoch": 1, "step": 2741, "train_loss": 3.5481998920440674, "perplexity": 34.75070612833025, "lr": 0.0007994917582417582, "grad_norm": 0.202613, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:27.489575+00:00", "epoch": 1, "step": 2742, "train_loss": 3.517242193222046, "perplexity": 33.69138589372749, "lr": 0.0007973406593406593, "grad_norm": 0.210444, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:27.796101+00:00", "epoch": 1, "step": 2743, "train_loss": 3.672032117843628, "perplexity": 39.33175145895562, "lr": 0.0007951895604395604, "grad_norm": 0.226984, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:28.104431+00:00", "epoch": 1, "step": 2744, "train_loss": 3.5441908836364746, "perplexity": 34.611669141734644, "lr": 0.0007930384615384615, "grad_norm": 0.199089, "tokens_per_sec": 106277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:28.411811+00:00", "epoch": 1, "step": 2745, "train_loss": 3.4662535190582275, "perplexity": 32.01656800783723, "lr": 0.0007908873626373626, "grad_norm": 0.191825, "tokens_per_sec": 106605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:28.717804+00:00", "epoch": 1, "step": 2746, "train_loss": 3.664057970046997, "perplexity": 39.01936143715139, "lr": 0.0007887362637362637, "grad_norm": 0.198672, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:29.024168+00:00", "epoch": 1, "step": 2747, "train_loss": 3.6324431896209717, "perplexity": 37.80506882825162, "lr": 0.0007865851648351649, "grad_norm": 0.208485, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:29.330736+00:00", "epoch": 1, "step": 2748, "train_loss": 3.660764694213867, "perplexity": 38.89107128049186, "lr": 0.0007844340659340659, "grad_norm": 0.200101, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:29.637348+00:00", "epoch": 1, "step": 2749, "train_loss": 3.6526129245758057, "perplexity": 38.57532890445004, "lr": 0.0007822829670329671, "grad_norm": 0.207872, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:29.944843+00:00", "epoch": 1, "step": 2750, "train_loss": 3.600515127182007, "perplexity": 36.61709204566807, "lr": 0.000780131868131868, "grad_norm": 0.21413, "tokens_per_sec": 106625} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:45:30.816165+00:00", "step": 2750, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0001_step_0002750.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:31.821545+00:00", "epoch": 1, "step": 2751, "train_loss": 3.5968003273010254, "perplexity": 36.48131921707444, "lr": 0.0007779807692307692, "grad_norm": 0.18629, "tokens_per_sec": 17459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:32.129328+00:00", "epoch": 1, "step": 2752, "train_loss": 3.516235589981079, "perplexity": 33.65748909866824, "lr": 0.0007758296703296703, "grad_norm": 0.196574, "tokens_per_sec": 106465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:32.435171+00:00", "epoch": 1, "step": 2753, "train_loss": 3.727964162826538, "perplexity": 41.59434259999222, "lr": 0.0007736785714285714, "grad_norm": 0.189016, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:32.741039+00:00", "epoch": 1, "step": 2754, "train_loss": 3.5777676105499268, "perplexity": 35.79354645547834, "lr": 0.0007715274725274725, "grad_norm": 0.206266, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:33.047490+00:00", "epoch": 1, "step": 2755, "train_loss": 3.7082324028015137, "perplexity": 40.78165723725213, "lr": 0.0007693763736263735, "grad_norm": 0.193251, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:33.353902+00:00", "epoch": 1, "step": 2756, "train_loss": 3.722364664077759, "perplexity": 41.362085997778955, "lr": 0.0007672252747252747, "grad_norm": 0.216668, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:33.660705+00:00", "epoch": 1, "step": 2757, "train_loss": 3.616271495819092, "perplexity": 37.198613750717435, "lr": 0.0007650741758241758, "grad_norm": 0.184701, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:33.967420+00:00", "epoch": 1, "step": 2758, "train_loss": 3.605945587158203, "perplexity": 36.816480594237746, "lr": 0.0007629230769230768, "grad_norm": 0.205392, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:34.273583+00:00", "epoch": 1, "step": 2759, "train_loss": 3.609848737716675, "perplexity": 36.96046166814369, "lr": 0.000760771978021978, "grad_norm": 0.193204, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:34.580690+00:00", "epoch": 1, "step": 2760, "train_loss": 3.6176326274871826, "perplexity": 37.24928043609518, "lr": 0.0007586208791208791, "grad_norm": 0.211116, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:34.887398+00:00", "epoch": 1, "step": 2761, "train_loss": 3.6302127838134766, "perplexity": 37.720842147939734, "lr": 0.0007564697802197801, "grad_norm": 0.203591, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:35.193802+00:00", "epoch": 1, "step": 2762, "train_loss": 3.5732288360595703, "perplexity": 35.63145574476032, "lr": 0.0007543186813186813, "grad_norm": 0.207995, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:35.500641+00:00", "epoch": 1, "step": 2763, "train_loss": 3.556373119354248, "perplexity": 35.0358954195672, "lr": 0.0007521675824175824, "grad_norm": 0.201053, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:35.808192+00:00", "epoch": 1, "step": 2764, "train_loss": 3.5614399909973145, "perplexity": 35.21386830668474, "lr": 0.0007500164835164835, "grad_norm": 0.193748, "tokens_per_sec": 106547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:36.168703+00:00", "epoch": 1, "step": 2765, "train_loss": 3.618976354598999, "perplexity": 37.29936694787836, "lr": 0.0007478653846153846, "grad_norm": 0.190577, "tokens_per_sec": 90893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:36.475886+00:00", "epoch": 1, "step": 2766, "train_loss": 3.549025297164917, "perplexity": 34.77940138009644, "lr": 0.0007457142857142857, "grad_norm": 0.202633, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:36.782488+00:00", "epoch": 1, "step": 2767, "train_loss": 3.705125093460083, "perplexity": 40.655132690017005, "lr": 0.0007435631868131867, "grad_norm": 0.193825, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:37.105621+00:00", "epoch": 1, "step": 2768, "train_loss": 3.6204373836517334, "perplexity": 37.35390223575686, "lr": 0.0007414120879120879, "grad_norm": 0.210873, "tokens_per_sec": 101412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:37.411733+00:00", "epoch": 1, "step": 2769, "train_loss": 3.6446526050567627, "perplexity": 38.26947591963343, "lr": 0.0007392609890109889, "grad_norm": 0.212176, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:37.718022+00:00", "epoch": 1, "step": 2770, "train_loss": 3.5393316745758057, "perplexity": 34.4438917682316, "lr": 0.0007371098901098901, "grad_norm": 0.198236, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:38.024722+00:00", "epoch": 1, "step": 2771, "train_loss": 3.7248427867889404, "perplexity": 41.46471343163885, "lr": 0.0007349587912087912, "grad_norm": 0.200725, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:38.343421+00:00", "epoch": 1, "step": 2772, "train_loss": 3.6435375213623047, "perplexity": 38.22682603455498, "lr": 0.0007328076923076923, "grad_norm": 0.18081, "tokens_per_sec": 102819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:38.663105+00:00", "epoch": 1, "step": 2773, "train_loss": 3.5730297565460205, "perplexity": 35.624362957921285, "lr": 0.0007306565934065934, "grad_norm": 0.192105, "tokens_per_sec": 102503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:38.970392+00:00", "epoch": 1, "step": 2774, "train_loss": 3.6336731910705566, "perplexity": 37.85159772715244, "lr": 0.0007285054945054944, "grad_norm": 0.176501, "tokens_per_sec": 106635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:39.288903+00:00", "epoch": 1, "step": 2775, "train_loss": 3.6741888523101807, "perplexity": 39.41667114464653, "lr": 0.0007263543956043955, "grad_norm": 0.187286, "tokens_per_sec": 102878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:39.595563+00:00", "epoch": 1, "step": 2776, "train_loss": 3.6490020751953125, "perplexity": 38.43629037663213, "lr": 0.0007242032967032967, "grad_norm": 0.192813, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:39.912310+00:00", "epoch": 1, "step": 2777, "train_loss": 3.568157196044922, "perplexity": 35.451203302139305, "lr": 0.0007220521978021977, "grad_norm": 0.196834, "tokens_per_sec": 103452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:40.218951+00:00", "epoch": 1, "step": 2778, "train_loss": 3.5863168239593506, "perplexity": 36.10086491656167, "lr": 0.0007199010989010989, "grad_norm": 0.191499, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:40.537214+00:00", "epoch": 1, "step": 2779, "train_loss": 3.63716721534729, "perplexity": 37.98408344790218, "lr": 0.00071775, "grad_norm": 0.192826, "tokens_per_sec": 102960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:40.857277+00:00", "epoch": 1, "step": 2780, "train_loss": 3.61838436126709, "perplexity": 37.27729250596667, "lr": 0.000715598901098901, "grad_norm": 0.187099, "tokens_per_sec": 102430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:41.177326+00:00", "epoch": 1, "step": 2781, "train_loss": 3.461926221847534, "perplexity": 31.878322133620756, "lr": 0.0007134478021978022, "grad_norm": 0.186218, "tokens_per_sec": 102334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:41.485282+00:00", "epoch": 1, "step": 2782, "train_loss": 3.630642890930176, "perplexity": 37.73706964012523, "lr": 0.0007112967032967033, "grad_norm": 0.189615, "tokens_per_sec": 106404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:41.805544+00:00", "epoch": 1, "step": 2783, "train_loss": 3.6540069580078125, "perplexity": 38.62914170229847, "lr": 0.0007091456043956044, "grad_norm": 0.180457, "tokens_per_sec": 102318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:42.126324+00:00", "epoch": 1, "step": 2784, "train_loss": 3.5495259761810303, "perplexity": 34.796819056529124, "lr": 0.0007069945054945055, "grad_norm": 0.182432, "tokens_per_sec": 102150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:42.432182+00:00", "epoch": 1, "step": 2785, "train_loss": 3.5948402881622314, "perplexity": 36.40988443392913, "lr": 0.0007048434065934066, "grad_norm": 0.181122, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:42.749419+00:00", "epoch": 1, "step": 2786, "train_loss": 3.5780575275421143, "perplexity": 35.80392511720894, "lr": 0.0007026923076923077, "grad_norm": 0.184249, "tokens_per_sec": 103291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:43.067018+00:00", "epoch": 1, "step": 2787, "train_loss": 3.5447235107421875, "perplexity": 34.630109165285745, "lr": 0.0007005412087912088, "grad_norm": 0.18898, "tokens_per_sec": 103175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:43.385259+00:00", "epoch": 1, "step": 2788, "train_loss": 3.725069761276245, "perplexity": 41.47412593186953, "lr": 0.0006983901098901098, "grad_norm": 0.194657, "tokens_per_sec": 102965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:43.705340+00:00", "epoch": 1, "step": 2789, "train_loss": 3.524531364440918, "perplexity": 33.93786539843537, "lr": 0.0006962390109890109, "grad_norm": 0.198914, "tokens_per_sec": 102376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:44.026293+00:00", "epoch": 1, "step": 2790, "train_loss": 3.741553544998169, "perplexity": 42.163442115161565, "lr": 0.0006940879120879121, "grad_norm": 0.190505, "tokens_per_sec": 102147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:44.345308+00:00", "epoch": 1, "step": 2791, "train_loss": 3.667846918106079, "perplexity": 39.16748420847301, "lr": 0.0006919368131868131, "grad_norm": 0.184181, "tokens_per_sec": 102664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:44.666061+00:00", "epoch": 1, "step": 2792, "train_loss": 3.703411102294922, "perplexity": 40.58550983529145, "lr": 0.0006897857142857142, "grad_norm": 0.176147, "tokens_per_sec": 102160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:44.972630+00:00", "epoch": 1, "step": 2793, "train_loss": 3.6316065788269043, "perplexity": 37.773453926127615, "lr": 0.0006876346153846153, "grad_norm": 0.189449, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:45.290174+00:00", "epoch": 1, "step": 2794, "train_loss": 3.641674518585205, "perplexity": 38.15567564875501, "lr": 0.0006854835164835164, "grad_norm": 0.192704, "tokens_per_sec": 103193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:45.610036+00:00", "epoch": 1, "step": 2795, "train_loss": 3.652576446533203, "perplexity": 38.573921777623625, "lr": 0.0006833324175824176, "grad_norm": 0.18403, "tokens_per_sec": 102445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:45.929742+00:00", "epoch": 1, "step": 2796, "train_loss": 3.5718533992767334, "perplexity": 35.58248061871889, "lr": 0.0006811813186813186, "grad_norm": 0.187679, "tokens_per_sec": 102494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:46.250101+00:00", "epoch": 1, "step": 2797, "train_loss": 3.6318318843841553, "perplexity": 37.781965454025, "lr": 0.0006790302197802198, "grad_norm": 0.188153, "tokens_per_sec": 102286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:46.568782+00:00", "epoch": 1, "step": 2798, "train_loss": 3.6448771953582764, "perplexity": 38.278071838012806, "lr": 0.0006768791208791209, "grad_norm": 0.172916, "tokens_per_sec": 102824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:46.886365+00:00", "epoch": 1, "step": 2799, "train_loss": 3.602071762084961, "perplexity": 36.674135875880964, "lr": 0.0006747280219780219, "grad_norm": 0.185073, "tokens_per_sec": 103180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:47.205959+00:00", "epoch": 1, "step": 2800, "train_loss": 3.5800724029541016, "perplexity": 35.876138291424354, "lr": 0.0006725769230769231, "grad_norm": 0.196943, "tokens_per_sec": 102585} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:45:52.914896+00:00", "step": 2800, "epoch": 1, "val_loss": 3.6690022587776183, "val_ppl": 39.21276214666623, "eval_train_loss": 3.5800724029541016, "eval_train_ppl": 35.876138291424354} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:45:53.786491+00:00", "step": 2800, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p6690_epoch_0001_step_0002800.pt", "val_loss": 3.6690022587776183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:54.801038+00:00", "epoch": 1, "step": 2801, "train_loss": 3.646250009536743, "perplexity": 38.33065660405408, "lr": 0.0006704258241758242, "grad_norm": 0.190897, "tokens_per_sec": 4314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:55.107322+00:00", "epoch": 1, "step": 2802, "train_loss": 3.5026302337646484, "perplexity": 33.20266798755665, "lr": 0.0006682747252747253, "grad_norm": 0.191427, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:55.413357+00:00", "epoch": 1, "step": 2803, "train_loss": 3.6291136741638184, "perplexity": 37.679405582177765, "lr": 0.0006661236263736264, "grad_norm": 0.179387, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:55.719980+00:00", "epoch": 1, "step": 2804, "train_loss": 3.634657621383667, "perplexity": 37.888878334423914, "lr": 0.0006639725274725275, "grad_norm": 0.191042, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:56.025816+00:00", "epoch": 1, "step": 2805, "train_loss": 3.5926148891448975, "perplexity": 36.328948004216656, "lr": 0.0006618214285714285, "grad_norm": 0.18693, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:56.332778+00:00", "epoch": 1, "step": 2806, "train_loss": 3.5459470748901367, "perplexity": 34.67250725840375, "lr": 0.0006596703296703296, "grad_norm": 0.17733, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:56.639240+00:00", "epoch": 1, "step": 2807, "train_loss": 3.5228404998779297, "perplexity": 33.88052955172576, "lr": 0.0006575192307692307, "grad_norm": 0.195083, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:56.946673+00:00", "epoch": 1, "step": 2808, "train_loss": 3.6021642684936523, "perplexity": 36.67752862540582, "lr": 0.0006553681318681318, "grad_norm": 0.176933, "tokens_per_sec": 106586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:57.252742+00:00", "epoch": 1, "step": 2809, "train_loss": 3.6429102420806885, "perplexity": 38.20285465774159, "lr": 0.0006532170329670328, "grad_norm": 0.190332, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:57.559236+00:00", "epoch": 1, "step": 2810, "train_loss": 3.5644638538360596, "perplexity": 35.320511370204436, "lr": 0.000651065934065934, "grad_norm": 0.181875, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:57.865263+00:00", "epoch": 1, "step": 2811, "train_loss": 3.6833019256591797, "perplexity": 39.77751988315848, "lr": 0.0006489148351648351, "grad_norm": 0.196207, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:58.172128+00:00", "epoch": 1, "step": 2812, "train_loss": 3.662297487258911, "perplexity": 38.950728953818036, "lr": 0.0006467637362637362, "grad_norm": 0.183949, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:58.479669+00:00", "epoch": 1, "step": 2813, "train_loss": 3.602975606918335, "perplexity": 36.70729858882657, "lr": 0.0006446126373626373, "grad_norm": 0.187971, "tokens_per_sec": 106548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:58.787980+00:00", "epoch": 1, "step": 2814, "train_loss": 3.5454440116882324, "perplexity": 34.6550691824792, "lr": 0.0006424615384615385, "grad_norm": 0.182025, "tokens_per_sec": 106284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:59.095700+00:00", "epoch": 1, "step": 2815, "train_loss": 3.7611095905303955, "perplexity": 42.996107594166816, "lr": 0.0006403104395604395, "grad_norm": 0.188966, "tokens_per_sec": 106486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:59.401508+00:00", "epoch": 1, "step": 2816, "train_loss": 3.5940158367156982, "perplexity": 36.37987862290625, "lr": 0.0006381593406593406, "grad_norm": 0.194462, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:45:59.707549+00:00", "epoch": 1, "step": 2817, "train_loss": 3.6703286170959473, "perplexity": 39.264806827241756, "lr": 0.0006360082417582418, "grad_norm": 0.184896, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:00.013854+00:00", "epoch": 1, "step": 2818, "train_loss": 3.60577654838562, "perplexity": 36.810257707516655, "lr": 0.0006338571428571428, "grad_norm": 0.19088, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:00.321427+00:00", "epoch": 1, "step": 2819, "train_loss": 3.597537040710449, "perplexity": 36.5082053966236, "lr": 0.000631706043956044, "grad_norm": 0.189839, "tokens_per_sec": 106537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:00.628125+00:00", "epoch": 1, "step": 2820, "train_loss": 3.5655226707458496, "perplexity": 35.35792913068867, "lr": 0.000629554945054945, "grad_norm": 0.187141, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:00.934301+00:00", "epoch": 1, "step": 2821, "train_loss": 3.693532705307007, "perplexity": 40.18656377514273, "lr": 0.0006274038461538462, "grad_norm": 0.177518, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:01.240710+00:00", "epoch": 1, "step": 2822, "train_loss": 3.574479818344116, "perplexity": 35.67605795714166, "lr": 0.0006252527472527472, "grad_norm": 0.19113, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:01.546912+00:00", "epoch": 1, "step": 2823, "train_loss": 3.659964084625244, "perplexity": 38.85994717670332, "lr": 0.0006231016483516482, "grad_norm": 0.174598, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:01.854062+00:00", "epoch": 1, "step": 2824, "train_loss": 3.5913212299346924, "perplexity": 36.28198111216032, "lr": 0.0006209505494505494, "grad_norm": 0.19292, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:02.161551+00:00", "epoch": 1, "step": 2825, "train_loss": 3.6768736839294434, "perplexity": 39.522640460892156, "lr": 0.0006187994505494505, "grad_norm": 0.177427, "tokens_per_sec": 106566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:02.468804+00:00", "epoch": 1, "step": 2826, "train_loss": 3.487969160079956, "perplexity": 32.71943225878744, "lr": 0.0006166483516483516, "grad_norm": 0.202741, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:02.774971+00:00", "epoch": 1, "step": 2827, "train_loss": 3.6389095783233643, "perplexity": 38.05032319866885, "lr": 0.0006144972527472527, "grad_norm": 0.185416, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:03.081733+00:00", "epoch": 1, "step": 2828, "train_loss": 3.5416293144226074, "perplexity": 34.523122413324984, "lr": 0.0006123461538461537, "grad_norm": 0.188795, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:03.388169+00:00", "epoch": 1, "step": 2829, "train_loss": 3.5284945964813232, "perplexity": 34.07263592098627, "lr": 0.0006101950549450549, "grad_norm": 0.179483, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:03.696296+00:00", "epoch": 1, "step": 2830, "train_loss": 3.5574185848236084, "perplexity": 35.0725433921734, "lr": 0.000608043956043956, "grad_norm": 0.182266, "tokens_per_sec": 106402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:04.001809+00:00", "epoch": 1, "step": 2831, "train_loss": 3.561772108078003, "perplexity": 35.22556537611726, "lr": 0.0006058928571428571, "grad_norm": 0.178337, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:04.308519+00:00", "epoch": 1, "step": 2832, "train_loss": 3.565892219543457, "perplexity": 35.37099802553326, "lr": 0.0006037417582417582, "grad_norm": 0.193636, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:04.614452+00:00", "epoch": 1, "step": 2833, "train_loss": 3.5553197860717773, "perplexity": 34.99901037437383, "lr": 0.0006015906593406594, "grad_norm": 0.182143, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:04.921903+00:00", "epoch": 1, "step": 2834, "train_loss": 3.617053985595703, "perplexity": 37.227732676826705, "lr": 0.0005994395604395604, "grad_norm": 0.203998, "tokens_per_sec": 106579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:05.228769+00:00", "epoch": 1, "step": 2835, "train_loss": 3.5376205444335938, "perplexity": 34.38500418334105, "lr": 0.0005972884615384615, "grad_norm": 0.206927, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:05.535242+00:00", "epoch": 1, "step": 2836, "train_loss": 3.5182266235351562, "perplexity": 33.72456904586257, "lr": 0.0005951373626373627, "grad_norm": 0.20136, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:05.842427+00:00", "epoch": 1, "step": 2837, "train_loss": 3.6422078609466553, "perplexity": 38.17603111464185, "lr": 0.0005929862637362637, "grad_norm": 0.20094, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:06.148717+00:00", "epoch": 1, "step": 2838, "train_loss": 3.6418306827545166, "perplexity": 38.16163466342739, "lr": 0.0005908351648351649, "grad_norm": 0.199037, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:06.454538+00:00", "epoch": 1, "step": 2839, "train_loss": 3.4257309436798096, "perplexity": 30.745109583811, "lr": 0.0005886840659340658, "grad_norm": 0.200799, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:06.761555+00:00", "epoch": 1, "step": 2840, "train_loss": 3.665391206741333, "perplexity": 39.07141817587509, "lr": 0.0005865329670329669, "grad_norm": 0.203657, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:07.069732+00:00", "epoch": 1, "step": 2841, "train_loss": 3.572125196456909, "perplexity": 35.59215315103867, "lr": 0.0005843818681318681, "grad_norm": 0.198415, "tokens_per_sec": 106271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:07.377000+00:00", "epoch": 1, "step": 2842, "train_loss": 3.60821533203125, "perplexity": 36.90013951860172, "lr": 0.0005822307692307691, "grad_norm": 0.189883, "tokens_per_sec": 106643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:07.683232+00:00", "epoch": 1, "step": 2843, "train_loss": 3.6110966205596924, "perplexity": 37.00661278373311, "lr": 0.0005800796703296703, "grad_norm": 0.191777, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:07.989597+00:00", "epoch": 1, "step": 2844, "train_loss": 3.5585057735443115, "perplexity": 35.11069460077976, "lr": 0.0005779285714285714, "grad_norm": 0.184692, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:08.295571+00:00", "epoch": 1, "step": 2845, "train_loss": 3.5552361011505127, "perplexity": 34.99608160749484, "lr": 0.0005757774725274725, "grad_norm": 0.19109, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:08.602117+00:00", "epoch": 1, "step": 2846, "train_loss": 3.6744937896728516, "perplexity": 39.4286925931919, "lr": 0.0005736263736263736, "grad_norm": 0.190226, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:08.909459+00:00", "epoch": 1, "step": 2847, "train_loss": 3.473764419555664, "perplexity": 32.25794661501137, "lr": 0.0005714752747252746, "grad_norm": 0.196063, "tokens_per_sec": 106617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:09.216960+00:00", "epoch": 1, "step": 2848, "train_loss": 3.549798011779785, "perplexity": 34.806286317693576, "lr": 0.0005693241758241758, "grad_norm": 0.204668, "tokens_per_sec": 106563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:09.522949+00:00", "epoch": 1, "step": 2849, "train_loss": 3.6016249656677246, "perplexity": 36.65775366339583, "lr": 0.0005671730769230769, "grad_norm": 0.185507, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:09.827954+00:00", "epoch": 1, "step": 2850, "train_loss": 3.5173544883728027, "perplexity": 33.695169485421154, "lr": 0.000565021978021978, "grad_norm": 0.184385, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:10.133682+00:00", "epoch": 1, "step": 2851, "train_loss": 3.519099473953247, "perplexity": 33.754018400625284, "lr": 0.0005628708791208791, "grad_norm": 0.189088, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:10.440490+00:00", "epoch": 1, "step": 2852, "train_loss": 3.639718532562256, "perplexity": 38.081116622467675, "lr": 0.0005607197802197803, "grad_norm": 0.194483, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:10.747103+00:00", "epoch": 1, "step": 2853, "train_loss": 3.598266363143921, "perplexity": 36.53484136174794, "lr": 0.0005585686813186813, "grad_norm": 0.183746, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:11.053761+00:00", "epoch": 1, "step": 2854, "train_loss": 3.5532209873199463, "perplexity": 34.92563152573862, "lr": 0.0005564175824175824, "grad_norm": 0.18582, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:11.360076+00:00", "epoch": 1, "step": 2855, "train_loss": 3.6427764892578125, "perplexity": 38.19774525979501, "lr": 0.0005542664835164836, "grad_norm": 0.18824, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:11.665985+00:00", "epoch": 1, "step": 2856, "train_loss": 3.6552743911743164, "perplexity": 38.67813259747005, "lr": 0.0005521153846153845, "grad_norm": 0.195432, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:11.972238+00:00", "epoch": 1, "step": 2857, "train_loss": 3.60099720954895, "perplexity": 36.63474875572437, "lr": 0.0005499642857142856, "grad_norm": 0.188774, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:12.280072+00:00", "epoch": 1, "step": 2858, "train_loss": 3.6385550498962402, "perplexity": 38.03683566843137, "lr": 0.0005478131868131867, "grad_norm": 0.186852, "tokens_per_sec": 106447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:12.587128+00:00", "epoch": 1, "step": 2859, "train_loss": 3.5957190990448, "perplexity": 36.44189590055511, "lr": 0.0005456620879120878, "grad_norm": 0.189372, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:12.894399+00:00", "epoch": 1, "step": 2860, "train_loss": 3.5785083770751953, "perplexity": 35.82007093952524, "lr": 0.000543510989010989, "grad_norm": 0.192558, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:13.200961+00:00", "epoch": 1, "step": 2861, "train_loss": 3.6143012046813965, "perplexity": 37.1253938076992, "lr": 0.00054135989010989, "grad_norm": 0.192991, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:13.507238+00:00", "epoch": 1, "step": 2862, "train_loss": 3.628173828125, "perplexity": 37.64400937819286, "lr": 0.0005392087912087912, "grad_norm": 0.193048, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:13.814294+00:00", "epoch": 1, "step": 2863, "train_loss": 3.5157158374786377, "perplexity": 33.64000007985767, "lr": 0.0005370576923076923, "grad_norm": 0.195836, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:14.121242+00:00", "epoch": 1, "step": 2864, "train_loss": 3.658564567565918, "perplexity": 38.80560005643655, "lr": 0.0005349065934065933, "grad_norm": 0.191312, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:14.427511+00:00", "epoch": 1, "step": 2865, "train_loss": 3.569855213165283, "perplexity": 35.51145118877937, "lr": 0.0005327554945054945, "grad_norm": 0.195355, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:14.733308+00:00", "epoch": 1, "step": 2866, "train_loss": 3.6398732662200928, "perplexity": 38.08700950883936, "lr": 0.0005306043956043955, "grad_norm": 0.191253, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:15.039057+00:00", "epoch": 1, "step": 2867, "train_loss": 3.56447172164917, "perplexity": 35.32078926648007, "lr": 0.0005284532967032967, "grad_norm": 0.188222, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:15.345563+00:00", "epoch": 1, "step": 2868, "train_loss": 3.541358232498169, "perplexity": 34.513765087221834, "lr": 0.0005263021978021978, "grad_norm": 0.205893, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:15.651800+00:00", "epoch": 1, "step": 2869, "train_loss": 3.5722761154174805, "perplexity": 35.59752508715003, "lr": 0.0005241510989010989, "grad_norm": 0.182525, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:15.958796+00:00", "epoch": 1, "step": 2870, "train_loss": 3.796687602996826, "perplexity": 44.5533613833194, "lr": 0.000522, "grad_norm": 0.191299, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:16.264780+00:00", "epoch": 1, "step": 2871, "train_loss": 3.609300374984741, "perplexity": 36.940199484430764, "lr": 0.0005198489010989011, "grad_norm": 0.192275, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:16.571058+00:00", "epoch": 1, "step": 2872, "train_loss": 3.6012613773345947, "perplexity": 36.64442775456436, "lr": 0.0005176978021978022, "grad_norm": 0.193845, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:16.876392+00:00", "epoch": 1, "step": 2873, "train_loss": 3.511180877685547, "perplexity": 33.487789427512006, "lr": 0.0005155467032967032, "grad_norm": 0.190813, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:17.182801+00:00", "epoch": 1, "step": 2874, "train_loss": 3.6134557723999023, "perplexity": 37.094020065366976, "lr": 0.0005133956043956044, "grad_norm": 0.195861, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:17.490318+00:00", "epoch": 1, "step": 2875, "train_loss": 3.584444761276245, "perplexity": 36.033345054949464, "lr": 0.0005112445054945054, "grad_norm": 0.18498, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:17.798677+00:00", "epoch": 1, "step": 2876, "train_loss": 3.5282771587371826, "perplexity": 34.06522804929933, "lr": 0.0005090934065934065, "grad_norm": 0.188762, "tokens_per_sec": 106265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:18.104741+00:00", "epoch": 1, "step": 2877, "train_loss": 3.6474289894104004, "perplexity": 38.3758743268892, "lr": 0.0005069423076923076, "grad_norm": 0.176321, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:18.412053+00:00", "epoch": 1, "step": 2878, "train_loss": 3.5815181732177734, "perplexity": 35.928044458496494, "lr": 0.0005047912087912087, "grad_norm": 0.182133, "tokens_per_sec": 106628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:18.718553+00:00", "epoch": 1, "step": 2879, "train_loss": 3.5823638439178467, "perplexity": 35.958440603759186, "lr": 0.0005026401098901099, "grad_norm": 0.181814, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:19.026568+00:00", "epoch": 1, "step": 2880, "train_loss": 3.531381845474243, "perplexity": 34.17115425978341, "lr": 0.0005004890109890109, "grad_norm": 0.186398, "tokens_per_sec": 106444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:19.332876+00:00", "epoch": 1, "step": 2881, "train_loss": 3.6556272506713867, "perplexity": 38.691782952073005, "lr": 0.0004983379120879121, "grad_norm": 0.186611, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:19.639748+00:00", "epoch": 1, "step": 2882, "train_loss": 3.5851664543151855, "perplexity": 36.05935945532208, "lr": 0.0004961868131868132, "grad_norm": 0.191551, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:19.945843+00:00", "epoch": 1, "step": 2883, "train_loss": 3.5193326473236084, "perplexity": 33.76188985653028, "lr": 0.0004940357142857142, "grad_norm": 0.189369, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:20.252152+00:00", "epoch": 1, "step": 2884, "train_loss": 3.599754810333252, "perplexity": 36.5892620347871, "lr": 0.0004918846153846154, "grad_norm": 0.179414, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:20.557643+00:00", "epoch": 1, "step": 2885, "train_loss": 3.652312994003296, "perplexity": 38.563760718879955, "lr": 0.0004897335164835164, "grad_norm": 0.193013, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:20.864536+00:00", "epoch": 1, "step": 2886, "train_loss": 3.645737409591675, "perplexity": 38.31101334658087, "lr": 0.0004875824175824176, "grad_norm": 0.190376, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:21.171648+00:00", "epoch": 1, "step": 2887, "train_loss": 3.49981951713562, "perplexity": 33.10947572638756, "lr": 0.0004854313186813187, "grad_norm": 0.18683, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:21.477512+00:00", "epoch": 1, "step": 2888, "train_loss": 3.5867300033569336, "perplexity": 36.11578413212412, "lr": 0.00048328021978021984, "grad_norm": 0.195493, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:21.783834+00:00", "epoch": 1, "step": 2889, "train_loss": 3.660012722015381, "perplexity": 38.861837269079054, "lr": 0.0004811291208791209, "grad_norm": 0.200237, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:22.090668+00:00", "epoch": 1, "step": 2890, "train_loss": 3.5295162200927734, "perplexity": 34.10746311746148, "lr": 0.00047897802197802185, "grad_norm": 0.190239, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:22.396165+00:00", "epoch": 1, "step": 2891, "train_loss": 3.6207120418548584, "perplexity": 37.3641632004892, "lr": 0.000476826923076923, "grad_norm": 0.18386, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:22.703243+00:00", "epoch": 1, "step": 2892, "train_loss": 3.6017656326293945, "perplexity": 36.66291056091936, "lr": 0.0004746758241758241, "grad_norm": 0.185127, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:23.009346+00:00", "epoch": 1, "step": 2893, "train_loss": 3.535658121109009, "perplexity": 34.31759241596219, "lr": 0.00047252472527472525, "grad_norm": 0.200293, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:23.315967+00:00", "epoch": 1, "step": 2894, "train_loss": 3.5880208015441895, "perplexity": 36.16243242109965, "lr": 0.0004703736263736263, "grad_norm": 0.18233, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:23.622801+00:00", "epoch": 1, "step": 2895, "train_loss": 3.463815689086914, "perplexity": 31.938612118978025, "lr": 0.00046822252747252737, "grad_norm": 0.192714, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:23.928354+00:00", "epoch": 1, "step": 2896, "train_loss": 3.5533130168914795, "perplexity": 34.92884586454831, "lr": 0.00046607142857142854, "grad_norm": 0.193393, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:24.234544+00:00", "epoch": 1, "step": 2897, "train_loss": 3.597527027130127, "perplexity": 36.507839820606804, "lr": 0.0004639203296703296, "grad_norm": 0.186891, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:24.541078+00:00", "epoch": 1, "step": 2898, "train_loss": 3.6621134281158447, "perplexity": 38.94356037576639, "lr": 0.00046176923076923077, "grad_norm": 0.185404, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:24.848954+00:00", "epoch": 1, "step": 2899, "train_loss": 3.5553317070007324, "perplexity": 34.99942759757684, "lr": 0.00045961813186813183, "grad_norm": 0.192043, "tokens_per_sec": 106433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:25.156425+00:00", "epoch": 1, "step": 2900, "train_loss": 3.53690767288208, "perplexity": 34.36050082697059, "lr": 0.000457467032967033, "grad_norm": 0.181783, "tokens_per_sec": 106624} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:46:30.719785+00:00", "step": 2900, "epoch": 1, "val_loss": 3.6457739591598513, "val_ppl": 38.31241362316469, "eval_train_loss": 3.53690767288208, "eval_train_ppl": 34.36050082697059} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:46:31.652588+00:00", "step": 2900, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p6458_epoch_0001_step_0002900.pt", "val_loss": 3.6457739591598513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:32.730715+00:00", "epoch": 1, "step": 2901, "train_loss": 3.594458818435669, "perplexity": 36.39599781410087, "lr": 0.00045531593406593406, "grad_norm": 0.185197, "tokens_per_sec": 4326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:33.034779+00:00", "epoch": 1, "step": 2902, "train_loss": 3.629800319671631, "perplexity": 37.70528686137298, "lr": 0.0004531648351648351, "grad_norm": 0.175703, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:33.339801+00:00", "epoch": 1, "step": 2903, "train_loss": 3.618182897567749, "perplexity": 37.26978324116454, "lr": 0.0004510137362637363, "grad_norm": 0.188475, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:33.644752+00:00", "epoch": 1, "step": 2904, "train_loss": 3.600672960281372, "perplexity": 36.622871890908826, "lr": 0.00044886263736263734, "grad_norm": 0.174047, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:33.950747+00:00", "epoch": 1, "step": 2905, "train_loss": 3.581220865249634, "perplexity": 35.91736435231831, "lr": 0.0004467115384615385, "grad_norm": 0.18768, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:34.256342+00:00", "epoch": 1, "step": 2906, "train_loss": 3.6032521724700928, "perplexity": 36.71745196708694, "lr": 0.0004445604395604396, "grad_norm": 0.171129, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:34.562108+00:00", "epoch": 1, "step": 2907, "train_loss": 3.5417416095733643, "perplexity": 34.52699941024088, "lr": 0.0004424093406593405, "grad_norm": 0.199443, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:34.866555+00:00", "epoch": 1, "step": 2908, "train_loss": 3.5441009998321533, "perplexity": 34.60855825304963, "lr": 0.0004402582417582417, "grad_norm": 0.165775, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:35.171939+00:00", "epoch": 1, "step": 2909, "train_loss": 3.5676686763763428, "perplexity": 35.433888921603476, "lr": 0.00043810714285714275, "grad_norm": 0.191608, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:35.478861+00:00", "epoch": 1, "step": 2910, "train_loss": 3.4897520542144775, "perplexity": 32.77781957638318, "lr": 0.0004359560439560439, "grad_norm": 0.175773, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:35.785202+00:00", "epoch": 1, "step": 2911, "train_loss": 3.6285970211029053, "perplexity": 37.659943429973524, "lr": 0.000433804945054945, "grad_norm": 0.175211, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:36.090676+00:00", "epoch": 1, "step": 2912, "train_loss": 3.5587046146392822, "perplexity": 35.11767674388484, "lr": 0.00043165384615384604, "grad_norm": 0.186792, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:36.395637+00:00", "epoch": 1, "step": 2913, "train_loss": 3.570969820022583, "perplexity": 35.55105456278307, "lr": 0.0004295027472527472, "grad_norm": 0.182802, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:36.702095+00:00", "epoch": 1, "step": 2914, "train_loss": 3.514338970184326, "perplexity": 33.593714136080436, "lr": 0.00042735164835164827, "grad_norm": 0.185584, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:37.075879+00:00", "epoch": 1, "step": 2915, "train_loss": 3.533437728881836, "perplexity": 34.24147843322538, "lr": 0.00042520054945054944, "grad_norm": 0.16868, "tokens_per_sec": 87665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:37.381339+00:00", "epoch": 1, "step": 2916, "train_loss": 3.53667950630188, "perplexity": 34.352661803338556, "lr": 0.0004230494505494505, "grad_norm": 0.17545, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:37.687429+00:00", "epoch": 1, "step": 2917, "train_loss": 3.5814242362976074, "perplexity": 35.92466964716465, "lr": 0.00042089835164835167, "grad_norm": 0.173958, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:37.993297+00:00", "epoch": 1, "step": 2918, "train_loss": 3.5673608779907227, "perplexity": 35.42298410612556, "lr": 0.0004187472527472527, "grad_norm": 0.165459, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:38.299770+00:00", "epoch": 1, "step": 2919, "train_loss": 3.5832290649414062, "perplexity": 35.98956606580431, "lr": 0.0004165961538461539, "grad_norm": 0.171302, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:38.605221+00:00", "epoch": 1, "step": 2920, "train_loss": 3.5773093700408936, "perplexity": 35.777148159999065, "lr": 0.00041444505494505496, "grad_norm": 0.174662, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:38.910711+00:00", "epoch": 1, "step": 2921, "train_loss": 3.6075806617736816, "perplexity": 36.87672752778214, "lr": 0.000412293956043956, "grad_norm": 0.175353, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:39.216224+00:00", "epoch": 1, "step": 2922, "train_loss": 3.4774696826934814, "perplexity": 32.37769250351996, "lr": 0.0004101428571428572, "grad_norm": 0.176058, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:39.522111+00:00", "epoch": 1, "step": 2923, "train_loss": 3.5310916900634766, "perplexity": 34.1612407527812, "lr": 0.00040799175824175824, "grad_norm": 0.175593, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:39.827956+00:00", "epoch": 1, "step": 2924, "train_loss": 3.587768077850342, "perplexity": 36.1532944723364, "lr": 0.0004058406593406594, "grad_norm": 0.177682, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:40.133423+00:00", "epoch": 1, "step": 2925, "train_loss": 3.656404972076416, "perplexity": 38.721886084281394, "lr": 0.00040368956043956036, "grad_norm": 0.180628, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:40.438674+00:00", "epoch": 1, "step": 2926, "train_loss": 3.534156084060669, "perplexity": 34.26608481359103, "lr": 0.0004015384615384614, "grad_norm": 0.176191, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:40.744337+00:00", "epoch": 1, "step": 2927, "train_loss": 3.5110697746276855, "perplexity": 33.48406903838273, "lr": 0.0003993873626373626, "grad_norm": 0.175609, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:41.050346+00:00", "epoch": 1, "step": 2928, "train_loss": 3.4923763275146484, "perplexity": 32.863950499245284, "lr": 0.00039723626373626365, "grad_norm": 0.167434, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:41.355588+00:00", "epoch": 1, "step": 2929, "train_loss": 3.623760223388672, "perplexity": 37.47822971218072, "lr": 0.0003950851648351648, "grad_norm": 0.166466, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:41.660663+00:00", "epoch": 1, "step": 2930, "train_loss": 3.5990633964538574, "perplexity": 36.56397245497017, "lr": 0.0003929340659340659, "grad_norm": 0.167276, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:41.966039+00:00", "epoch": 1, "step": 2931, "train_loss": 3.494467258453369, "perplexity": 32.93273864067755, "lr": 0.00039078296703296694, "grad_norm": 0.179784, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:42.272737+00:00", "epoch": 1, "step": 2932, "train_loss": 3.508929967880249, "perplexity": 33.41249620484847, "lr": 0.0003886318681318681, "grad_norm": 0.167466, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:42.578905+00:00", "epoch": 1, "step": 2933, "train_loss": 3.6221015453338623, "perplexity": 37.41611692182562, "lr": 0.00038648076923076917, "grad_norm": 0.165913, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:42.884737+00:00", "epoch": 1, "step": 2934, "train_loss": 3.5211405754089355, "perplexity": 33.82298413580622, "lr": 0.00038432967032967034, "grad_norm": 0.168765, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:43.190772+00:00", "epoch": 1, "step": 2935, "train_loss": 3.6579225063323975, "perplexity": 38.78069248194648, "lr": 0.0003821785714285714, "grad_norm": 0.174501, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:43.496292+00:00", "epoch": 1, "step": 2936, "train_loss": 3.5332369804382324, "perplexity": 34.234605199641784, "lr": 0.00038002747252747257, "grad_norm": 0.177381, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:43.804088+00:00", "epoch": 1, "step": 2937, "train_loss": 3.5227720737457275, "perplexity": 33.87821131744638, "lr": 0.0003778763736263736, "grad_norm": 0.176629, "tokens_per_sec": 106462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:44.110355+00:00", "epoch": 1, "step": 2938, "train_loss": 3.4420008659362793, "perplexity": 31.24942156016195, "lr": 0.0003757252747252747, "grad_norm": 0.190569, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:44.416339+00:00", "epoch": 1, "step": 2939, "train_loss": 3.583388090133667, "perplexity": 35.99528976856171, "lr": 0.00037357417582417585, "grad_norm": 0.174143, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:44.722102+00:00", "epoch": 1, "step": 2940, "train_loss": 3.6129238605499268, "perplexity": 37.07429456311025, "lr": 0.0003714230769230769, "grad_norm": 0.182593, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:45.029571+00:00", "epoch": 1, "step": 2941, "train_loss": 3.4995994567871094, "perplexity": 33.10219044525036, "lr": 0.0003692719780219781, "grad_norm": 0.185795, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:45.336524+00:00", "epoch": 1, "step": 2942, "train_loss": 3.5937082767486572, "perplexity": 36.368691349102996, "lr": 0.00036712087912087903, "grad_norm": 0.179654, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:45.641797+00:00", "epoch": 1, "step": 2943, "train_loss": 3.529309034347534, "perplexity": 34.10039726929405, "lr": 0.0003649697802197801, "grad_norm": 0.175229, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:45.947060+00:00", "epoch": 1, "step": 2944, "train_loss": 3.601017713546753, "perplexity": 36.635499922233286, "lr": 0.00036281868131868126, "grad_norm": 0.167464, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:46.251982+00:00", "epoch": 1, "step": 2945, "train_loss": 3.524994134902954, "perplexity": 33.953574474654204, "lr": 0.0003606675824175823, "grad_norm": 0.176395, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:46.557124+00:00", "epoch": 1, "step": 2946, "train_loss": 3.602653741836548, "perplexity": 36.695485692345514, "lr": 0.0003585164835164835, "grad_norm": 0.178036, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:46.862388+00:00", "epoch": 1, "step": 2947, "train_loss": 3.6396517753601074, "perplexity": 38.078574518520085, "lr": 0.00035636538461538455, "grad_norm": 0.171783, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:47.168095+00:00", "epoch": 1, "step": 2948, "train_loss": 3.6609373092651367, "perplexity": 38.89778504418653, "lr": 0.0003542142857142857, "grad_norm": 0.173877, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:47.472701+00:00", "epoch": 1, "step": 2949, "train_loss": 3.6732823848724365, "perplexity": 39.38095740486618, "lr": 0.0003520631868131868, "grad_norm": 0.176856, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:47.778839+00:00", "epoch": 1, "step": 2950, "train_loss": 3.570910930633545, "perplexity": 35.548961044543816, "lr": 0.00034991208791208784, "grad_norm": 0.175389, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:48.083934+00:00", "epoch": 1, "step": 2951, "train_loss": 3.6664905548095703, "perplexity": 39.114394882821564, "lr": 0.000347760989010989, "grad_norm": 0.165208, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:48.389770+00:00", "epoch": 1, "step": 2952, "train_loss": 3.5682809352874756, "perplexity": 35.45559027859851, "lr": 0.00034560989010989007, "grad_norm": 0.173482, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:48.695937+00:00", "epoch": 1, "step": 2953, "train_loss": 3.604097366333008, "perplexity": 36.74849845045182, "lr": 0.00034345879120879124, "grad_norm": 0.179307, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:49.001362+00:00", "epoch": 1, "step": 2954, "train_loss": 3.653158187866211, "perplexity": 38.59636835071452, "lr": 0.0003413076923076923, "grad_norm": 0.172362, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:49.306396+00:00", "epoch": 1, "step": 2955, "train_loss": 3.5413520336151123, "perplexity": 34.513551141091334, "lr": 0.00033915659340659347, "grad_norm": 0.163081, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:49.612266+00:00", "epoch": 1, "step": 2956, "train_loss": 3.617077112197876, "perplexity": 37.22859363774563, "lr": 0.0003370054945054945, "grad_norm": 0.18347, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:49.918681+00:00", "epoch": 1, "step": 2957, "train_loss": 3.6457314491271973, "perplexity": 38.31078499582726, "lr": 0.0003348543956043956, "grad_norm": 0.16296, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:50.223773+00:00", "epoch": 1, "step": 2958, "train_loss": 3.639831066131592, "perplexity": 38.08540226758045, "lr": 0.00033270329670329675, "grad_norm": 0.174218, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:50.529026+00:00", "epoch": 1, "step": 2959, "train_loss": 3.536503791809082, "perplexity": 34.34662607309169, "lr": 0.0003305521978021977, "grad_norm": 0.17382, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:50.834491+00:00", "epoch": 1, "step": 2960, "train_loss": 3.5649497509002686, "perplexity": 35.33767767317594, "lr": 0.00032840109890109877, "grad_norm": 0.189079, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:51.139723+00:00", "epoch": 1, "step": 2961, "train_loss": 3.6181833744049072, "perplexity": 37.2698010127863, "lr": 0.00032624999999999993, "grad_norm": 0.169621, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:51.445279+00:00", "epoch": 1, "step": 2962, "train_loss": 3.6457948684692383, "perplexity": 38.313214717649636, "lr": 0.0003240989010989011, "grad_norm": 0.178196, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:51.751467+00:00", "epoch": 1, "step": 2963, "train_loss": 3.528671979904175, "perplexity": 34.07868037784888, "lr": 0.00032194780219780205, "grad_norm": 0.167938, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:52.058262+00:00", "epoch": 1, "step": 2964, "train_loss": 3.50282883644104, "perplexity": 33.20926278113247, "lr": 0.0003197967032967032, "grad_norm": 0.186581, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:52.364396+00:00", "epoch": 1, "step": 2965, "train_loss": 3.5409040451049805, "perplexity": 34.498092929545386, "lr": 0.0003176456043956044, "grad_norm": 0.17331, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:52.670446+00:00", "epoch": 1, "step": 2966, "train_loss": 3.5909664630889893, "perplexity": 36.26911175111159, "lr": 0.00031549450549450556, "grad_norm": 0.172745, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:52.976002+00:00", "epoch": 1, "step": 2967, "train_loss": 3.5596160888671875, "perplexity": 35.14970019323814, "lr": 0.0003133434065934065, "grad_norm": 0.177492, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:53.283184+00:00", "epoch": 1, "step": 2968, "train_loss": 3.5329301357269287, "perplexity": 34.22410210358485, "lr": 0.0003111923076923077, "grad_norm": 0.168505, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:53.590504+00:00", "epoch": 1, "step": 2969, "train_loss": 3.4786882400512695, "perplexity": 32.41717062724236, "lr": 0.00030904120879120885, "grad_norm": 0.178133, "tokens_per_sec": 106626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:53.897033+00:00", "epoch": 1, "step": 2970, "train_loss": 3.5020864009857178, "perplexity": 33.18461619737958, "lr": 0.0003068901098901098, "grad_norm": 0.18283, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:54.202723+00:00", "epoch": 1, "step": 2971, "train_loss": 3.6403424739837646, "perplexity": 38.104884422592164, "lr": 0.00030473901098901097, "grad_norm": 0.167204, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:54.508050+00:00", "epoch": 1, "step": 2972, "train_loss": 3.5761096477508545, "perplexity": 35.73425125520915, "lr": 0.00030258791208791214, "grad_norm": 0.168752, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:54.813922+00:00", "epoch": 1, "step": 2973, "train_loss": 3.59798002243042, "perplexity": 36.52438144682599, "lr": 0.0003004368131868133, "grad_norm": 0.184042, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:55.120918+00:00", "epoch": 1, "step": 2974, "train_loss": 3.532217502593994, "perplexity": 34.19972156270331, "lr": 0.00029828571428571426, "grad_norm": 0.172054, "tokens_per_sec": 106737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:55.426902+00:00", "epoch": 1, "step": 2975, "train_loss": 3.6271464824676514, "perplexity": 37.60535582731151, "lr": 0.0002961346153846154, "grad_norm": 0.164618, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:55.732660+00:00", "epoch": 1, "step": 2976, "train_loss": 3.5200815200805664, "perplexity": 33.78718268544048, "lr": 0.0002939835164835164, "grad_norm": 0.172687, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:56.038959+00:00", "epoch": 1, "step": 2977, "train_loss": 3.6409106254577637, "perplexity": 38.1265399200623, "lr": 0.00029183241758241754, "grad_norm": 0.16343, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:56.345108+00:00", "epoch": 1, "step": 2978, "train_loss": 3.6860129833221436, "perplexity": 39.88550534451808, "lr": 0.0002896813186813187, "grad_norm": 0.17409, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:56.651673+00:00", "epoch": 1, "step": 2979, "train_loss": 3.6105804443359375, "perplexity": 36.98751577922629, "lr": 0.00028753021978021966, "grad_norm": 0.159355, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:56.958646+00:00", "epoch": 1, "step": 2980, "train_loss": 3.572284460067749, "perplexity": 35.5978221372867, "lr": 0.00028537912087912083, "grad_norm": 0.176496, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:57.265061+00:00", "epoch": 1, "step": 2981, "train_loss": 3.5124704837799072, "perplexity": 33.531003343319945, "lr": 0.000283228021978022, "grad_norm": 0.167269, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:57.571389+00:00", "epoch": 1, "step": 2982, "train_loss": 3.5571298599243164, "perplexity": 35.06241853733377, "lr": 0.00028107692307692295, "grad_norm": 0.160409, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:57.875774+00:00", "epoch": 1, "step": 2983, "train_loss": 3.6404669284820557, "perplexity": 38.10962704197945, "lr": 0.0002789258241758241, "grad_norm": 0.162251, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:58.181812+00:00", "epoch": 1, "step": 2984, "train_loss": 3.5826752185821533, "perplexity": 35.9696388944727, "lr": 0.0002767747252747253, "grad_norm": 0.159279, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:58.488065+00:00", "epoch": 1, "step": 2985, "train_loss": 3.501572608947754, "perplexity": 33.16757058511939, "lr": 0.00027462362637362646, "grad_norm": 0.162663, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:58.794805+00:00", "epoch": 1, "step": 2986, "train_loss": 3.585526943206787, "perplexity": 36.07236079712284, "lr": 0.0002724725274725274, "grad_norm": 0.169504, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:59.100748+00:00", "epoch": 1, "step": 2987, "train_loss": 3.6722354888916016, "perplexity": 39.339751211899966, "lr": 0.0002703214285714286, "grad_norm": 0.159561, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:59.405947+00:00", "epoch": 1, "step": 2988, "train_loss": 3.661431312561035, "perplexity": 38.917005425276436, "lr": 0.00026817032967032975, "grad_norm": 0.167488, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:46:59.714325+00:00", "epoch": 1, "step": 2989, "train_loss": 3.5903589725494385, "perplexity": 36.24708529995262, "lr": 0.0002660192307692307, "grad_norm": 0.167561, "tokens_per_sec": 106267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:00.020642+00:00", "epoch": 1, "step": 2990, "train_loss": 3.5628461837768555, "perplexity": 35.26342062592877, "lr": 0.00026386813186813187, "grad_norm": 0.158339, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:00.326923+00:00", "epoch": 1, "step": 2991, "train_loss": 3.564527988433838, "perplexity": 35.322776709637026, "lr": 0.00026171703296703304, "grad_norm": 0.167164, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:00.634350+00:00", "epoch": 1, "step": 2992, "train_loss": 3.5103697776794434, "perplexity": 33.4606384938527, "lr": 0.0002595659340659342, "grad_norm": 0.175708, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:00.940352+00:00", "epoch": 1, "step": 2993, "train_loss": 3.63668155670166, "perplexity": 37.965640628199296, "lr": 0.00025741483516483516, "grad_norm": 0.16173, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:01.246066+00:00", "epoch": 1, "step": 2994, "train_loss": 3.6263160705566406, "perplexity": 37.574140854351164, "lr": 0.0002552637362637361, "grad_norm": 0.163431, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:01.551764+00:00", "epoch": 1, "step": 2995, "train_loss": 3.6013941764831543, "perplexity": 36.64929442650742, "lr": 0.0002531126373626373, "grad_norm": 0.168652, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:01.857841+00:00", "epoch": 1, "step": 2996, "train_loss": 3.5970234870910645, "perplexity": 36.48946128906902, "lr": 0.00025096153846153844, "grad_norm": 0.162057, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:02.165445+00:00", "epoch": 1, "step": 2997, "train_loss": 3.5496745109558105, "perplexity": 34.80198797808355, "lr": 0.0002488104395604394, "grad_norm": 0.16048, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:02.471489+00:00", "epoch": 1, "step": 2998, "train_loss": 3.545358180999756, "perplexity": 34.65209484167875, "lr": 0.00024665934065934056, "grad_norm": 0.160902, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:02.776338+00:00", "epoch": 1, "step": 2999, "train_loss": 3.44191837310791, "perplexity": 31.246843813316826, "lr": 0.00024450824175824173, "grad_norm": 0.162869, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:03.082864+00:00", "epoch": 1, "step": 3000, "train_loss": 3.5244863033294678, "perplexity": 33.936336154955235, "lr": 0.0002423571428571429, "grad_norm": 0.161458, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T15:47:08.670269+00:00", "step": 3000, "epoch": 1, "val_loss": 3.623665285110474, "val_ppl": 37.47467176247742, "eval_train_loss": 3.5244863033294678, "eval_train_ppl": 33.936336154955235} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T15:47:09.591550+00:00", "step": 3000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p6237_epoch_0001_step_0003000.pt", "val_loss": 3.623665285110474} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:47:11.291021+00:00", "step": 3000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0001_step_0003000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:12.403293+00:00", "epoch": 1, "step": 3001, "train_loss": 3.5351200103759766, "perplexity": 34.29913071881551, "lr": 0.00024020604395604385, "grad_norm": 0.162521, "tokens_per_sec": 3516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:12.708254+00:00", "epoch": 1, "step": 3002, "train_loss": 3.474059820175171, "perplexity": 32.26747704000289, "lr": 0.00023805494505494502, "grad_norm": 0.156224, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:13.012412+00:00", "epoch": 1, "step": 3003, "train_loss": 3.5584187507629395, "perplexity": 35.107639303421855, "lr": 0.0002359038461538462, "grad_norm": 0.152752, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:13.316993+00:00", "epoch": 1, "step": 3004, "train_loss": 3.4992997646331787, "perplexity": 33.09227146489053, "lr": 0.00023375274725274714, "grad_norm": 0.166401, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:13.622002+00:00", "epoch": 1, "step": 3005, "train_loss": 3.477428436279297, "perplexity": 32.376357067345786, "lr": 0.0002316016483516483, "grad_norm": 0.166597, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:13.927587+00:00", "epoch": 1, "step": 3006, "train_loss": 3.4898924827575684, "perplexity": 32.78242284103929, "lr": 0.00022945054945054948, "grad_norm": 0.156815, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:14.232430+00:00", "epoch": 1, "step": 3007, "train_loss": 3.4792394638061523, "perplexity": 32.43504466760343, "lr": 0.00022729945054945065, "grad_norm": 0.162527, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:14.537688+00:00", "epoch": 1, "step": 3008, "train_loss": 3.5203845500946045, "perplexity": 33.79742276733167, "lr": 0.0002251483516483516, "grad_norm": 0.160014, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:14.842172+00:00", "epoch": 1, "step": 3009, "train_loss": 3.5454981327056885, "perplexity": 34.6569448008381, "lr": 0.00022299725274725277, "grad_norm": 0.164384, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:15.147886+00:00", "epoch": 1, "step": 3010, "train_loss": 3.5011227130889893, "perplexity": 33.152651988626886, "lr": 0.00022084615384615372, "grad_norm": 0.166024, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:15.453926+00:00", "epoch": 1, "step": 3011, "train_loss": 3.5291285514831543, "perplexity": 34.094243287279255, "lr": 0.0002186950549450549, "grad_norm": 0.15976, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:15.761485+00:00", "epoch": 1, "step": 3012, "train_loss": 3.652003288269043, "perplexity": 38.55181915033254, "lr": 0.00021654395604395606, "grad_norm": 0.167667, "tokens_per_sec": 106542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:16.067762+00:00", "epoch": 1, "step": 3013, "train_loss": 3.4810564517974854, "perplexity": 32.494032327969364, "lr": 0.000214392857142857, "grad_norm": 0.166118, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:16.372740+00:00", "epoch": 1, "step": 3014, "train_loss": 3.622633695602417, "perplexity": 37.436033217254824, "lr": 0.00021224175824175818, "grad_norm": 0.163314, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:16.677943+00:00", "epoch": 1, "step": 3015, "train_loss": 3.5219175815582275, "perplexity": 33.8492750152319, "lr": 0.00021009065934065934, "grad_norm": 0.153748, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:16.983083+00:00", "epoch": 1, "step": 3016, "train_loss": 3.6271328926086426, "perplexity": 37.604844779300386, "lr": 0.0002079395604395603, "grad_norm": 0.172744, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:17.288576+00:00", "epoch": 1, "step": 3017, "train_loss": 3.571262836456299, "perplexity": 35.5614731323374, "lr": 0.00020578846153846146, "grad_norm": 0.161786, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:17.595650+00:00", "epoch": 1, "step": 3018, "train_loss": 3.459625482559204, "perplexity": 31.805062733148294, "lr": 0.00020363736263736263, "grad_norm": 0.155484, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:17.900828+00:00", "epoch": 1, "step": 3019, "train_loss": 3.6178529262542725, "perplexity": 37.25748731059921, "lr": 0.0002014862637362638, "grad_norm": 0.166514, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:18.206388+00:00", "epoch": 1, "step": 3020, "train_loss": 3.5878653526306152, "perplexity": 36.156811447166056, "lr": 0.00019933516483516475, "grad_norm": 0.165577, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:18.511966+00:00", "epoch": 1, "step": 3021, "train_loss": 3.6158876419067383, "perplexity": 37.18433765743751, "lr": 0.00019718406593406592, "grad_norm": 0.162251, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:18.816332+00:00", "epoch": 1, "step": 3022, "train_loss": 3.637413501739502, "perplexity": 37.99343956287065, "lr": 0.0001950329670329671, "grad_norm": 0.159068, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:19.121895+00:00", "epoch": 1, "step": 3023, "train_loss": 3.5626604557037354, "perplexity": 35.256871826931025, "lr": 0.00019288186813186804, "grad_norm": 0.165828, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:19.427844+00:00", "epoch": 1, "step": 3024, "train_loss": 3.455983877182007, "perplexity": 31.68945187799545, "lr": 0.0001907307692307692, "grad_norm": 0.161528, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:19.733259+00:00", "epoch": 1, "step": 3025, "train_loss": 3.5155529975891113, "perplexity": 33.63452259194987, "lr": 0.00018857967032967038, "grad_norm": 0.158294, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:20.038796+00:00", "epoch": 1, "step": 3026, "train_loss": 3.5643606185913086, "perplexity": 35.31686523677638, "lr": 0.00018642857142857155, "grad_norm": 0.159714, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:20.343865+00:00", "epoch": 1, "step": 3027, "train_loss": 3.497058391571045, "perplexity": 33.01818240069774, "lr": 0.0001842774725274725, "grad_norm": 0.15797, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:20.649172+00:00", "epoch": 1, "step": 3028, "train_loss": 3.563558340072632, "perplexity": 35.28854263727565, "lr": 0.00018212637362637345, "grad_norm": 0.15847, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:20.955512+00:00", "epoch": 1, "step": 3029, "train_loss": 3.553812026977539, "perplexity": 34.94628006048717, "lr": 0.00017997527472527462, "grad_norm": 0.163749, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:21.260713+00:00", "epoch": 1, "step": 3030, "train_loss": 3.551964282989502, "perplexity": 34.881767900933184, "lr": 0.00017782417582417579, "grad_norm": 0.159229, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:21.565916+00:00", "epoch": 1, "step": 3031, "train_loss": 3.5710861682891846, "perplexity": 35.55519110699252, "lr": 0.00017567307692307695, "grad_norm": 0.154126, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:21.871320+00:00", "epoch": 1, "step": 3032, "train_loss": 3.5318942070007324, "perplexity": 34.188666730516864, "lr": 0.0001735219780219779, "grad_norm": 0.153429, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:22.176740+00:00", "epoch": 1, "step": 3033, "train_loss": 3.5238144397735596, "perplexity": 33.91354322519848, "lr": 0.00017137087912087907, "grad_norm": 0.150826, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:22.482613+00:00", "epoch": 1, "step": 3034, "train_loss": 3.6172492504119873, "perplexity": 37.2350026529707, "lr": 0.00016921978021978024, "grad_norm": 0.155783, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:22.787735+00:00", "epoch": 1, "step": 3035, "train_loss": 3.5478339195251465, "perplexity": 34.7379906517736, "lr": 0.0001670686813186812, "grad_norm": 0.162829, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:23.093517+00:00", "epoch": 1, "step": 3036, "train_loss": 3.62914776802063, "perplexity": 37.68069024033578, "lr": 0.00016491758241758236, "grad_norm": 0.157259, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:23.399361+00:00", "epoch": 1, "step": 3037, "train_loss": 3.6164112091064453, "perplexity": 37.203811254401245, "lr": 0.00016276648351648353, "grad_norm": 0.151647, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:23.705580+00:00", "epoch": 1, "step": 3038, "train_loss": 3.483081817626953, "perplexity": 32.559911322720254, "lr": 0.0001606153846153847, "grad_norm": 0.147978, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:24.011889+00:00", "epoch": 1, "step": 3039, "train_loss": 3.432919502258301, "perplexity": 30.96691889449547, "lr": 0.00015846428571428565, "grad_norm": 0.159344, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:24.318600+00:00", "epoch": 1, "step": 3040, "train_loss": 3.49619722366333, "perplexity": 32.98976044143928, "lr": 0.00015631318681318682, "grad_norm": 0.15541, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:24.623348+00:00", "epoch": 1, "step": 3041, "train_loss": 3.577932834625244, "perplexity": 35.79946089968461, "lr": 0.000154162087912088, "grad_norm": 0.14842, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:24.929089+00:00", "epoch": 1, "step": 3042, "train_loss": 3.698037624359131, "perplexity": 40.36800938396369, "lr": 0.00015201098901098894, "grad_norm": 0.158718, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:25.234582+00:00", "epoch": 1, "step": 3043, "train_loss": 3.533156394958496, "perplexity": 34.2318464987153, "lr": 0.0001498598901098901, "grad_norm": 0.163513, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:25.540366+00:00", "epoch": 1, "step": 3044, "train_loss": 3.504533290863037, "perplexity": 33.265914722541794, "lr": 0.00014770879120879106, "grad_norm": 0.15122, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:25.846889+00:00", "epoch": 1, "step": 3045, "train_loss": 3.586376190185547, "perplexity": 36.103008152291466, "lr": 0.00014555769230769223, "grad_norm": 0.152611, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:26.153185+00:00", "epoch": 1, "step": 3046, "train_loss": 3.5146360397338867, "perplexity": 33.60369528808167, "lr": 0.0001434065934065934, "grad_norm": 0.164144, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:26.458976+00:00", "epoch": 1, "step": 3047, "train_loss": 3.485006332397461, "perplexity": 32.6226336886881, "lr": 0.00014125549450549435, "grad_norm": 0.149335, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:26.765661+00:00", "epoch": 1, "step": 3048, "train_loss": 3.5203559398651123, "perplexity": 33.79645582914228, "lr": 0.00013910439560439552, "grad_norm": 0.15067, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:27.071912+00:00", "epoch": 1, "step": 3049, "train_loss": 3.575617551803589, "perplexity": 35.716670900953204, "lr": 0.00013695329670329669, "grad_norm": 0.154962, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:27.378867+00:00", "epoch": 1, "step": 3050, "train_loss": 3.5037314891815186, "perplexity": 33.23925274641704, "lr": 0.00013480219780219785, "grad_norm": 0.155569, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:27.686319+00:00", "epoch": 1, "step": 3051, "train_loss": 3.6347267627716064, "perplexity": 37.891498114626, "lr": 0.0001326510989010988, "grad_norm": 0.152947, "tokens_per_sec": 106522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T15:47:27.992944+00:00", "epoch": 1, "step": 3052, "train_loss": 3.5226492881774902, "perplexity": 33.87405181738742, "lr": 0.00013049999999999997, "grad_norm": 0.149647, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T15:47:28.926034+00:00", "step": 3052, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/last_epoch_0001_step_0003052.pt", "category": "last"} +{"stage": "pretraining", "type": "training_complete", "timestamp": "2026-04-08T15:47:28.928899+00:00", "epochs": 2, "total_steps": 3052, "total_time_hours": 0.48, "best_val_loss": 3.623665285110474, "best_checkpoint_path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints/best_loss_3p6237_epoch_0001_step_0003000.pt", "checkpoint_dir": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_151821/checkpoints", "stopped_early": false, "stop_reason": null, "run_dir": "runs/big_run/exp_c/run_20260408_151821"}