diff --git "a/runs/big_run/exp_c/run_20260408_135803/metrics.jsonl" "b/runs/big_run/exp_c/run_20260408_135803/metrics.jsonl" new file mode 100644--- /dev/null +++ "b/runs/big_run/exp_c/run_20260408_135803/metrics.jsonl" @@ -0,0 +1,3132 @@ +{"stage": "pretraining", "type": "config", "timestamp": "2026-04-08T13:58:04.147943+00:00", "device": "auto", "train_bin": "data/exp_c/train.bin", "val_bin": "data/exp_a/val.bin", "num_workers": 4, "pin_memory": true, "batch_size": 32, "gradient_accumulation_steps": 1, "train_samples_per_epoch": null, "val_sequence_stride": null, "learning_rate": 0.001305, "min_lr": 0.0001305, "weight_decay": 0.063, "beta1": 0.9, "beta2": 0.9978, "grad_clip": 1.786, "warmup_steps": 305, "max_steps": 3052, "lr_schedule": "wsd", "lr_decay_ratio": 0.179, "z_loss_coeff": 0.000275, "save_every": 250, "eval_every": 100, "checkpoint_dir": "checkpoints", "keep_last_checkpoints": 10, "keep_best_checkpoints": 10, "early_stopping_patience": 0, "early_stopping_min_delta": 0.001, "runs_dir": "runs/big_run/exp_c", "log_every": 10, "hf_upload": {"repo_id": "ParrotLabs/Preprocessed", "repo_type": "dataset", "path_in_repo": "", "private": null}, "compile": true} +{"stage": "pretraining", "type": "checkpoint_policy", "timestamp": "2026-04-08T13:58:04.148036+00:00", "checkpoint_dir": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints", "keep_last": 10, "keep_best": 10} +{"stage": "pretraining", "type": "model_architecture", "timestamp": "2026-04-08T13:58:05.860742+00:00", "vocab_size": 50258, "context_length": 1024, "n_layers": 14, "n_heads": 6, "d_model": 384, "d_ff": 768, "dropout": 0.005, "bias": false, "total_params": 39966592, "total_params_non_embedding": 39966592, "trainable_params": 39966592, "non_trainable_params": 0, "params_size_mb": 159.87} +{"stage": "pretraining", "type": "initial_validation", "timestamp": "2026-04-08T13:58:14.734412+00:00", "val_loss": 10.892181491851806, "val_ppl": 53754.4370232034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:54.380532+00:00", "epoch": 0, "step": 1, "train_loss": 10.925138473510742, "perplexity": 55555.53741348256, "lr": 4.278688524590164e-06, "grad_norm": 7.191318, "tokens_per_sec": 830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:54.747333+00:00", "epoch": 0, "step": 2, "train_loss": 10.88792610168457, "perplexity": 53526.176932772694, "lr": 8.557377049180328e-06, "grad_norm": 7.041293, "tokens_per_sec": 61667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:55.055916+00:00", "epoch": 0, "step": 3, "train_loss": 10.813188552856445, "perplexity": 49671.596655645204, "lr": 1.2836065573770493e-05, "grad_norm": 6.803502, "tokens_per_sec": 106183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:55.359592+00:00", "epoch": 0, "step": 4, "train_loss": 10.707443237304688, "perplexity": 44687.238036758565, "lr": 1.7114754098360656e-05, "grad_norm": 6.585616, "tokens_per_sec": 107904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:55.663983+00:00", "epoch": 0, "step": 5, "train_loss": 10.601115226745605, "perplexity": 40179.62184275455, "lr": 2.1393442622950817e-05, "grad_norm": 5.849903, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:55.967496+00:00", "epoch": 0, "step": 6, "train_loss": 10.502140045166016, "perplexity": 36393.302708297146, "lr": 2.5672131147540985e-05, "grad_norm": 4.206442, "tokens_per_sec": 107962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:56.272491+00:00", "epoch": 0, "step": 7, "train_loss": 10.420207977294922, "perplexity": 33530.40698406568, "lr": 2.9950819672131143e-05, "grad_norm": 2.950845, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:56.578815+00:00", "epoch": 0, "step": 8, "train_loss": 10.350739479064941, "perplexity": 31280.165296646857, "lr": 3.422950819672131e-05, "grad_norm": 2.3349, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:56.883658+00:00", "epoch": 0, "step": 9, "train_loss": 10.308854103088379, "perplexity": 29997.043419021684, "lr": 3.850819672131147e-05, "grad_norm": 1.99535, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:57.188233+00:00", "epoch": 0, "step": 10, "train_loss": 10.240365028381348, "perplexity": 28011.348997646375, "lr": 4.2786885245901634e-05, "grad_norm": 1.825579, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:57.492160+00:00", "epoch": 0, "step": 11, "train_loss": 10.214095115661621, "perplexity": 27285.07465298869, "lr": 4.70655737704918e-05, "grad_norm": 1.765255, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:57.796517+00:00", "epoch": 0, "step": 12, "train_loss": 10.215256690979004, "perplexity": 27316.78673665973, "lr": 5.134426229508197e-05, "grad_norm": 1.740022, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:58.102441+00:00", "epoch": 0, "step": 13, "train_loss": 10.191204071044922, "perplexity": 26667.585244007605, "lr": 5.562295081967213e-05, "grad_norm": 1.730684, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:58.407694+00:00", "epoch": 0, "step": 14, "train_loss": 10.177032470703125, "perplexity": 26292.328160904388, "lr": 5.990163934426229e-05, "grad_norm": 1.70856, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:58.710660+00:00", "epoch": 0, "step": 15, "train_loss": 10.106959342956543, "perplexity": 24513.011641317797, "lr": 6.418032786885245e-05, "grad_norm": 1.71246, "tokens_per_sec": 108162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:59.015171+00:00", "epoch": 0, "step": 16, "train_loss": 10.095610618591309, "perplexity": 24236.392833030146, "lr": 6.845901639344262e-05, "grad_norm": 1.689028, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:59.319050+00:00", "epoch": 0, "step": 17, "train_loss": 10.040719985961914, "perplexity": 22941.894813940256, "lr": 7.273770491803279e-05, "grad_norm": 1.700618, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:59.624136+00:00", "epoch": 0, "step": 18, "train_loss": 9.999996185302734, "perplexity": 22026.38177066814, "lr": 7.701639344262295e-05, "grad_norm": 1.706825, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:58:59.928425+00:00", "epoch": 0, "step": 19, "train_loss": 9.951403617858887, "perplexity": 20981.651944395217, "lr": 8.12950819672131e-05, "grad_norm": 1.681398, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:00.233044+00:00", "epoch": 0, "step": 20, "train_loss": 9.883415222167969, "perplexity": 19602.55555479656, "lr": 8.557377049180327e-05, "grad_norm": 1.708033, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:00.537164+00:00", "epoch": 0, "step": 21, "train_loss": 9.81299114227295, "perplexity": 18269.5522607858, "lr": 8.985245901639344e-05, "grad_norm": 1.708457, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:00.841453+00:00", "epoch": 0, "step": 22, "train_loss": 9.799005508422852, "perplexity": 18015.819435246416, "lr": 9.41311475409836e-05, "grad_norm": 1.760293, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:01.146956+00:00", "epoch": 0, "step": 23, "train_loss": 9.712841987609863, "perplexity": 16528.509008265064, "lr": 9.840983606557377e-05, "grad_norm": 1.686511, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:01.452184+00:00", "epoch": 0, "step": 24, "train_loss": 9.667057037353516, "perplexity": 15788.814690830517, "lr": 0.00010268852459016394, "grad_norm": 1.68156, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:01.756655+00:00", "epoch": 0, "step": 25, "train_loss": 9.62434196472168, "perplexity": 15128.595364826899, "lr": 0.0001069672131147541, "grad_norm": 1.650393, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:02.060383+00:00", "epoch": 0, "step": 26, "train_loss": 9.52663516998291, "perplexity": 13720.346690985823, "lr": 0.00011124590163934426, "grad_norm": 1.652482, "tokens_per_sec": 107886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:02.364292+00:00", "epoch": 0, "step": 27, "train_loss": 9.529010772705078, "perplexity": 13752.979529917597, "lr": 0.0001155245901639344, "grad_norm": 1.631608, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:02.668556+00:00", "epoch": 0, "step": 28, "train_loss": 9.461267471313477, "perplexity": 12852.163877895762, "lr": 0.00011980327868852457, "grad_norm": 1.625257, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:02.973050+00:00", "epoch": 0, "step": 29, "train_loss": 9.362518310546875, "perplexity": 11643.674038716323, "lr": 0.00012408196721311476, "grad_norm": 1.625832, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:03.277646+00:00", "epoch": 0, "step": 30, "train_loss": 9.331758499145508, "perplexity": 11290.969197798668, "lr": 0.0001283606557377049, "grad_norm": 1.601424, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:03.581609+00:00", "epoch": 0, "step": 31, "train_loss": 9.210369110107422, "perplexity": 10000.287385441832, "lr": 0.00013263934426229506, "grad_norm": 1.588692, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:03.886048+00:00", "epoch": 0, "step": 32, "train_loss": 9.185859680175781, "perplexity": 9758.165299968457, "lr": 0.00013691803278688525, "grad_norm": 1.589675, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:04.190536+00:00", "epoch": 0, "step": 33, "train_loss": 9.115782737731934, "perplexity": 9097.752989032091, "lr": 0.0001411967213114754, "grad_norm": 1.576652, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:04.495702+00:00", "epoch": 0, "step": 34, "train_loss": 9.075517654418945, "perplexity": 8738.708199099872, "lr": 0.00014547540983606558, "grad_norm": 1.673376, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:04.800794+00:00", "epoch": 0, "step": 35, "train_loss": 8.962420463562012, "perplexity": 7804.224460463483, "lr": 0.00014975409836065574, "grad_norm": 1.816523, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:05.104827+00:00", "epoch": 0, "step": 36, "train_loss": 8.934361457824707, "perplexity": 7588.28931144448, "lr": 0.0001540327868852459, "grad_norm": 2.14214, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:05.409086+00:00", "epoch": 0, "step": 37, "train_loss": 8.86970329284668, "perplexity": 7113.1701315735445, "lr": 0.00015831147540983607, "grad_norm": 1.61454, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:05.713410+00:00", "epoch": 0, "step": 38, "train_loss": 8.758869171142578, "perplexity": 6366.907622666542, "lr": 0.0001625901639344262, "grad_norm": 1.771541, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:06.019720+00:00", "epoch": 0, "step": 39, "train_loss": 8.75025749206543, "perplexity": 6312.313269428354, "lr": 0.00016686885245901638, "grad_norm": 1.672411, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:06.324787+00:00", "epoch": 0, "step": 40, "train_loss": 8.669316291809082, "perplexity": 5821.517769177923, "lr": 0.00017114754098360654, "grad_norm": 1.593671, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:06.629440+00:00", "epoch": 0, "step": 41, "train_loss": 8.62218952178955, "perplexity": 5553.5326723494345, "lr": 0.00017542622950819672, "grad_norm": 1.564365, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:06.933776+00:00", "epoch": 0, "step": 42, "train_loss": 8.546272277832031, "perplexity": 5147.530055980369, "lr": 0.00017970491803278687, "grad_norm": 1.560771, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:07.238554+00:00", "epoch": 0, "step": 43, "train_loss": 8.490973472595215, "perplexity": 4870.605166842665, "lr": 0.00018398360655737706, "grad_norm": 1.529922, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:07.543165+00:00", "epoch": 0, "step": 44, "train_loss": 8.437338829040527, "perplexity": 4616.253983242687, "lr": 0.0001882622950819672, "grad_norm": 1.482227, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:07.849370+00:00", "epoch": 0, "step": 45, "train_loss": 8.358017921447754, "perplexity": 4264.234337126251, "lr": 0.00019254098360655736, "grad_norm": 1.447832, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:08.155469+00:00", "epoch": 0, "step": 46, "train_loss": 8.276278495788574, "perplexity": 3929.5433248717113, "lr": 0.00019681967213114755, "grad_norm": 1.45705, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:08.459192+00:00", "epoch": 0, "step": 47, "train_loss": 8.257515907287598, "perplexity": 3856.502282715257, "lr": 0.0002010983606557377, "grad_norm": 1.458459, "tokens_per_sec": 107887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:08.763071+00:00", "epoch": 0, "step": 48, "train_loss": 8.11115550994873, "perplexity": 3331.425301765008, "lr": 0.00020537704918032788, "grad_norm": 1.431449, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:09.068025+00:00", "epoch": 0, "step": 49, "train_loss": 8.102071762084961, "perplexity": 3301.300504472354, "lr": 0.00020965573770491804, "grad_norm": 1.373614, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:09.372842+00:00", "epoch": 0, "step": 50, "train_loss": 8.043302536010742, "perplexity": 3112.876620461769, "lr": 0.0002139344262295082, "grad_norm": 1.346588, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:09.677700+00:00", "epoch": 0, "step": 51, "train_loss": 7.9968719482421875, "perplexity": 2971.6479648773147, "lr": 0.00021821311475409837, "grad_norm": 1.31938, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:09.982754+00:00", "epoch": 0, "step": 52, "train_loss": 7.882281303405762, "perplexity": 2649.910920224801, "lr": 0.00022249180327868853, "grad_norm": 1.29988, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:10.286909+00:00", "epoch": 0, "step": 53, "train_loss": 7.93437385559082, "perplexity": 2791.610240857729, "lr": 0.0002267704918032787, "grad_norm": 1.247676, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:10.591146+00:00", "epoch": 0, "step": 54, "train_loss": 7.791971683502197, "perplexity": 2421.0864955568654, "lr": 0.0002310491803278688, "grad_norm": 1.21947, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:10.897682+00:00", "epoch": 0, "step": 55, "train_loss": 7.756300449371338, "perplexity": 2336.2455391721655, "lr": 0.000235327868852459, "grad_norm": 1.199373, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:11.202725+00:00", "epoch": 0, "step": 56, "train_loss": 7.799519062042236, "perplexity": 2439.428481705651, "lr": 0.00023960655737704915, "grad_norm": 1.154524, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:11.507498+00:00", "epoch": 0, "step": 57, "train_loss": 7.701381206512451, "perplexity": 2211.400283952955, "lr": 0.00024388524590163933, "grad_norm": 1.096203, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:11.811627+00:00", "epoch": 0, "step": 58, "train_loss": 7.7130560874938965, "perplexity": 2237.369417369794, "lr": 0.0002481639344262295, "grad_norm": 1.057032, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:12.115349+00:00", "epoch": 0, "step": 59, "train_loss": 7.633842945098877, "perplexity": 2066.978059773607, "lr": 0.00025244262295081964, "grad_norm": 0.999513, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:12.420345+00:00", "epoch": 0, "step": 60, "train_loss": 7.593359470367432, "perplexity": 1984.9707755487223, "lr": 0.0002567213114754098, "grad_norm": 0.987262, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:12.725935+00:00", "epoch": 0, "step": 61, "train_loss": 7.509757041931152, "perplexity": 1825.769903396014, "lr": 0.000261, "grad_norm": 0.949903, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:13.030380+00:00", "epoch": 0, "step": 62, "train_loss": 7.510984420776367, "perplexity": 1828.0121905380809, "lr": 0.00026527868852459013, "grad_norm": 0.917726, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:13.334505+00:00", "epoch": 0, "step": 63, "train_loss": 7.476302623748779, "perplexity": 1765.7002338493708, "lr": 0.0002695573770491803, "grad_norm": 0.924764, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:13.639070+00:00", "epoch": 0, "step": 64, "train_loss": 7.46420955657959, "perplexity": 1744.4760934534002, "lr": 0.0002738360655737705, "grad_norm": 0.826455, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:13.944451+00:00", "epoch": 0, "step": 65, "train_loss": 7.363767623901367, "perplexity": 1577.7698220872721, "lr": 0.0002781147540983607, "grad_norm": 0.832717, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:14.250077+00:00", "epoch": 0, "step": 66, "train_loss": 7.422722816467285, "perplexity": 1673.5841721646511, "lr": 0.0002823934426229508, "grad_norm": 0.76535, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:14.555343+00:00", "epoch": 0, "step": 67, "train_loss": 7.35529088973999, "perplexity": 1564.4520122502597, "lr": 0.000286672131147541, "grad_norm": 0.893571, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:14.859575+00:00", "epoch": 0, "step": 68, "train_loss": 7.309744358062744, "perplexity": 1494.795008051111, "lr": 0.00029095081967213117, "grad_norm": 1.128903, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:15.165820+00:00", "epoch": 0, "step": 69, "train_loss": 7.275897979736328, "perplexity": 1445.0482337342517, "lr": 0.0002952295081967213, "grad_norm": 0.903697, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:15.471145+00:00", "epoch": 0, "step": 70, "train_loss": 7.325888156890869, "perplexity": 1519.1225188150192, "lr": 0.0002995081967213115, "grad_norm": 0.677426, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:15.777322+00:00", "epoch": 0, "step": 71, "train_loss": 7.296803951263428, "perplexity": 1475.5763692533799, "lr": 0.00030378688524590166, "grad_norm": 1.214379, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:16.083072+00:00", "epoch": 0, "step": 72, "train_loss": 7.343814373016357, "perplexity": 1546.6001868933552, "lr": 0.0003080655737704918, "grad_norm": 1.629923, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:16.389495+00:00", "epoch": 0, "step": 73, "train_loss": 7.197364330291748, "perplexity": 1335.9051155662964, "lr": 0.00031234426229508197, "grad_norm": 1.361096, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:16.694474+00:00", "epoch": 0, "step": 74, "train_loss": 7.245260238647461, "perplexity": 1401.446559080639, "lr": 0.00031662295081967215, "grad_norm": 0.911053, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:17.000229+00:00", "epoch": 0, "step": 75, "train_loss": 7.273672580718994, "perplexity": 1441.8360003910507, "lr": 0.0003209016393442623, "grad_norm": 1.643601, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:17.306966+00:00", "epoch": 0, "step": 76, "train_loss": 7.176508903503418, "perplexity": 1308.3327602246086, "lr": 0.0003251803278688524, "grad_norm": 2.407869, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:17.615036+00:00", "epoch": 0, "step": 77, "train_loss": 7.0620341300964355, "perplexity": 1166.816209418219, "lr": 0.0003294590163934426, "grad_norm": 0.980315, "tokens_per_sec": 106366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:17.921252+00:00", "epoch": 0, "step": 78, "train_loss": 7.091020584106445, "perplexity": 1201.1330325611227, "lr": 0.00033373770491803277, "grad_norm": 0.730347, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:18.226929+00:00", "epoch": 0, "step": 79, "train_loss": 7.113033771514893, "perplexity": 1227.8669691342097, "lr": 0.00033801639344262295, "grad_norm": 1.965192, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:18.532451+00:00", "epoch": 0, "step": 80, "train_loss": 7.106605052947998, "perplexity": 1219.9986765699666, "lr": 0.0003422950819672131, "grad_norm": 1.36936, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:18.838827+00:00", "epoch": 0, "step": 81, "train_loss": 7.0668559074401855, "perplexity": 1172.4559231764008, "lr": 0.00034657377049180326, "grad_norm": 1.079535, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:19.145922+00:00", "epoch": 0, "step": 82, "train_loss": 7.130200386047363, "perplexity": 1249.1272495080307, "lr": 0.00035085245901639344, "grad_norm": 1.130158, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:19.452512+00:00", "epoch": 0, "step": 83, "train_loss": 7.118668556213379, "perplexity": 1234.805264686444, "lr": 0.00035513114754098357, "grad_norm": 0.722526, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:19.759681+00:00", "epoch": 0, "step": 84, "train_loss": 7.046634197235107, "perplexity": 1148.984970473216, "lr": 0.00035940983606557375, "grad_norm": 1.336739, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:20.066808+00:00", "epoch": 0, "step": 85, "train_loss": 6.970933437347412, "perplexity": 1065.2165995751513, "lr": 0.00036368852459016393, "grad_norm": 0.597529, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:20.372404+00:00", "epoch": 0, "step": 86, "train_loss": 7.047848701477051, "perplexity": 1150.381265325165, "lr": 0.0003679672131147541, "grad_norm": 0.634005, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:20.678045+00:00", "epoch": 0, "step": 87, "train_loss": 7.057628631591797, "perplexity": 1161.6871087706963, "lr": 0.00037224590163934424, "grad_norm": 0.617516, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:20.983623+00:00", "epoch": 0, "step": 88, "train_loss": 6.988141059875488, "perplexity": 1083.705059749176, "lr": 0.0003765245901639344, "grad_norm": 0.512915, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:21.291182+00:00", "epoch": 0, "step": 89, "train_loss": 6.946172714233398, "perplexity": 1039.164926424877, "lr": 0.0003808032786885246, "grad_norm": 0.528798, "tokens_per_sec": 106543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:21.598451+00:00", "epoch": 0, "step": 90, "train_loss": 6.876177787780762, "perplexity": 968.9158710896673, "lr": 0.00038508196721311473, "grad_norm": 0.506407, "tokens_per_sec": 106700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:21.904478+00:00", "epoch": 0, "step": 91, "train_loss": 6.934144496917725, "perplexity": 1026.7404965292067, "lr": 0.0003893606557377049, "grad_norm": 0.532158, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:22.210599+00:00", "epoch": 0, "step": 92, "train_loss": 6.881779193878174, "perplexity": 974.3583910097465, "lr": 0.0003936393442622951, "grad_norm": 0.52553, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:22.518057+00:00", "epoch": 0, "step": 93, "train_loss": 6.889942169189453, "perplexity": 982.34460579076, "lr": 0.0003979180327868853, "grad_norm": 0.365925, "tokens_per_sec": 106577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:22.826943+00:00", "epoch": 0, "step": 94, "train_loss": 6.811304569244385, "perplexity": 908.0546547487628, "lr": 0.0004021967213114754, "grad_norm": 0.626427, "tokens_per_sec": 106084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:23.133688+00:00", "epoch": 0, "step": 95, "train_loss": 6.889155387878418, "perplexity": 981.5720193820556, "lr": 0.0004064754098360656, "grad_norm": 0.727004, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:23.438383+00:00", "epoch": 0, "step": 96, "train_loss": 6.833338737487793, "perplexity": 928.284944100429, "lr": 0.00041075409836065577, "grad_norm": 0.988601, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:23.744598+00:00", "epoch": 0, "step": 97, "train_loss": 6.8595428466796875, "perplexity": 952.9313321773151, "lr": 0.0004150327868852459, "grad_norm": 0.907062, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:24.051069+00:00", "epoch": 0, "step": 98, "train_loss": 6.769334316253662, "perplexity": 870.7320688364543, "lr": 0.0004193114754098361, "grad_norm": 0.493316, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:24.358050+00:00", "epoch": 0, "step": 99, "train_loss": 6.785677433013916, "perplexity": 885.0794659527894, "lr": 0.00042359016393442626, "grad_norm": 0.491036, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:59:24.665019+00:00", "epoch": 0, "step": 100, "train_loss": 6.736417770385742, "perplexity": 842.5371619437138, "lr": 0.0004278688524590164, "grad_norm": 0.528769, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:00:00.220147+00:00", "step": 100, "epoch": 0, "val_loss": 6.7636210203170775, "val_ppl": 865.7715029296917, "eval_train_loss": 6.736417770385742, "eval_train_ppl": 842.5371619437138} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:00:01.366045+00:00", "step": 100, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_6p7636_epoch_0000_step_0000100.pt", "val_loss": 6.7636210203170775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:03.065243+00:00", "epoch": 0, "step": 101, "train_loss": 6.7008748054504395, "perplexity": 813.1168331628596, "lr": 0.00043214754098360657, "grad_norm": 0.447665, "tokens_per_sec": 853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:03.369185+00:00", "epoch": 0, "step": 102, "train_loss": 6.825942039489746, "perplexity": 921.4440319827478, "lr": 0.00043642622950819675, "grad_norm": 0.504999, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:03.674321+00:00", "epoch": 0, "step": 103, "train_loss": 6.651531219482422, "perplexity": 773.9685343605538, "lr": 0.00044070491803278693, "grad_norm": 0.366627, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:03.979725+00:00", "epoch": 0, "step": 104, "train_loss": 6.641878604888916, "perplexity": 766.533655130142, "lr": 0.00044498360655737706, "grad_norm": 0.473638, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:04.285738+00:00", "epoch": 0, "step": 105, "train_loss": 6.685661792755127, "perplexity": 800.840493278854, "lr": 0.00044926229508196724, "grad_norm": 0.408439, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:04.591670+00:00", "epoch": 0, "step": 106, "train_loss": 6.788459777832031, "perplexity": 887.5454912940355, "lr": 0.0004535409836065574, "grad_norm": 0.633951, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:04.897444+00:00", "epoch": 0, "step": 107, "train_loss": 6.813784599304199, "perplexity": 910.3094524158038, "lr": 0.0004578196721311475, "grad_norm": 0.891409, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:05.203656+00:00", "epoch": 0, "step": 108, "train_loss": 6.710447788238525, "perplexity": 820.9381636122109, "lr": 0.0004620983606557376, "grad_norm": 1.379622, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:05.509806+00:00", "epoch": 0, "step": 109, "train_loss": 6.827373504638672, "perplexity": 922.7639915138568, "lr": 0.0004663770491803278, "grad_norm": 2.229, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:05.815050+00:00", "epoch": 0, "step": 110, "train_loss": 6.683359622955322, "perplexity": 798.9989430747481, "lr": 0.000470655737704918, "grad_norm": 1.528188, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:06.120789+00:00", "epoch": 0, "step": 111, "train_loss": 6.7339863777160645, "perplexity": 840.4911116480426, "lr": 0.00047493442622950817, "grad_norm": 2.027878, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:06.426394+00:00", "epoch": 0, "step": 112, "train_loss": 6.874141693115234, "perplexity": 966.945073699082, "lr": 0.0004792131147540983, "grad_norm": 2.061504, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:06.732234+00:00", "epoch": 0, "step": 113, "train_loss": 6.564482688903809, "perplexity": 709.4447975870972, "lr": 0.0004834918032786885, "grad_norm": 0.93322, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:07.037815+00:00", "epoch": 0, "step": 114, "train_loss": 6.575873851776123, "perplexity": 717.5724023749223, "lr": 0.00048777049180327866, "grad_norm": 1.444847, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:07.343339+00:00", "epoch": 0, "step": 115, "train_loss": 6.5840630531311035, "perplexity": 723.4728743661083, "lr": 0.0004920491803278688, "grad_norm": 0.871756, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:07.648315+00:00", "epoch": 0, "step": 116, "train_loss": 6.69766902923584, "perplexity": 810.5143362995087, "lr": 0.000496327868852459, "grad_norm": 0.72726, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:07.953679+00:00", "epoch": 0, "step": 117, "train_loss": 6.574490547180176, "perplexity": 716.5804674050684, "lr": 0.0005006065573770491, "grad_norm": 0.738156, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:08.258987+00:00", "epoch": 0, "step": 118, "train_loss": 6.591068267822266, "perplexity": 728.5587502020745, "lr": 0.0005048852459016393, "grad_norm": 0.599755, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:08.565579+00:00", "epoch": 0, "step": 119, "train_loss": 6.541970252990723, "perplexity": 693.6519023127539, "lr": 0.0005091639344262295, "grad_norm": 0.522352, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:08.871088+00:00", "epoch": 0, "step": 120, "train_loss": 6.533705711364746, "perplexity": 687.9428112977881, "lr": 0.0005134426229508196, "grad_norm": 0.493476, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:09.177966+00:00", "epoch": 0, "step": 121, "train_loss": 6.6364545822143555, "perplexity": 762.3872145634215, "lr": 0.0005177213114754098, "grad_norm": 0.550117, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:09.484306+00:00", "epoch": 0, "step": 122, "train_loss": 6.5721235275268555, "perplexity": 714.8863131946889, "lr": 0.000522, "grad_norm": 0.431684, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:09.791272+00:00", "epoch": 0, "step": 123, "train_loss": 6.510031223297119, "perplexity": 671.8473945544263, "lr": 0.0005262786885245901, "grad_norm": 0.503934, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:10.097175+00:00", "epoch": 0, "step": 124, "train_loss": 6.498055458068848, "perplexity": 663.8494939652197, "lr": 0.0005305573770491803, "grad_norm": 0.486991, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:10.403338+00:00", "epoch": 0, "step": 125, "train_loss": 6.37823486328125, "perplexity": 588.8873230809103, "lr": 0.0005348360655737705, "grad_norm": 0.618751, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:10.709992+00:00", "epoch": 0, "step": 126, "train_loss": 6.570863723754883, "perplexity": 713.9862637826562, "lr": 0.0005391147540983606, "grad_norm": 0.584842, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:11.018353+00:00", "epoch": 0, "step": 127, "train_loss": 6.38267183303833, "perplexity": 591.5060035306309, "lr": 0.0005433934426229507, "grad_norm": 0.628301, "tokens_per_sec": 106264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:11.326435+00:00", "epoch": 0, "step": 128, "train_loss": 6.475228309631348, "perplexity": 648.867353367207, "lr": 0.000547672131147541, "grad_norm": 0.528016, "tokens_per_sec": 106361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:11.633079+00:00", "epoch": 0, "step": 129, "train_loss": 6.46022891998291, "perplexity": 639.2073671617715, "lr": 0.0005519508196721311, "grad_norm": 0.458335, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:11.938804+00:00", "epoch": 0, "step": 130, "train_loss": 6.382452487945557, "perplexity": 591.3762738197182, "lr": 0.0005562295081967213, "grad_norm": 0.36243, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:12.243847+00:00", "epoch": 0, "step": 131, "train_loss": 6.513660430908203, "perplexity": 674.2900980902618, "lr": 0.0005605081967213115, "grad_norm": 0.466239, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:12.550769+00:00", "epoch": 0, "step": 132, "train_loss": 6.447922229766846, "perplexity": 631.3890476051538, "lr": 0.0005647868852459016, "grad_norm": 0.563351, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:12.857921+00:00", "epoch": 0, "step": 133, "train_loss": 6.401353359222412, "perplexity": 602.6601018172981, "lr": 0.0005690655737704918, "grad_norm": 0.997855, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:13.165263+00:00", "epoch": 0, "step": 134, "train_loss": 6.391906261444092, "perplexity": 596.9935214218921, "lr": 0.000573344262295082, "grad_norm": 1.474248, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:13.471899+00:00", "epoch": 0, "step": 135, "train_loss": 6.477176189422607, "perplexity": 650.1325007495265, "lr": 0.0005776229508196721, "grad_norm": 1.829093, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:13.778171+00:00", "epoch": 0, "step": 136, "train_loss": 6.47504186630249, "perplexity": 648.7463876548338, "lr": 0.0005819016393442623, "grad_norm": 1.48348, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:14.084551+00:00", "epoch": 0, "step": 137, "train_loss": 6.345588207244873, "perplexity": 569.9725531181161, "lr": 0.0005861803278688525, "grad_norm": 1.440356, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:14.390957+00:00", "epoch": 0, "step": 138, "train_loss": 6.438258647918701, "perplexity": 625.3169542047508, "lr": 0.0005904590163934426, "grad_norm": 1.136676, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:14.699030+00:00", "epoch": 0, "step": 139, "train_loss": 6.353419303894043, "perplexity": 574.4535860697009, "lr": 0.0005947377049180328, "grad_norm": 0.802681, "tokens_per_sec": 106364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:15.006179+00:00", "epoch": 0, "step": 140, "train_loss": 6.417585372924805, "perplexity": 612.522314091332, "lr": 0.000599016393442623, "grad_norm": 0.788093, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:15.312773+00:00", "epoch": 0, "step": 141, "train_loss": 6.29037618637085, "perplexity": 539.3561893730882, "lr": 0.0006032950819672131, "grad_norm": 0.774035, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:15.618377+00:00", "epoch": 0, "step": 142, "train_loss": 6.369144916534424, "perplexity": 583.5586241601421, "lr": 0.0006075737704918033, "grad_norm": 0.784441, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:15.924586+00:00", "epoch": 0, "step": 143, "train_loss": 6.3818864822387695, "perplexity": 591.0416461833872, "lr": 0.0006118524590163934, "grad_norm": 0.553014, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:16.232043+00:00", "epoch": 0, "step": 144, "train_loss": 6.288963794708252, "perplexity": 538.5949449022665, "lr": 0.0006161311475409836, "grad_norm": 0.790794, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:16.538800+00:00", "epoch": 0, "step": 145, "train_loss": 6.498678684234619, "perplexity": 664.2633512899542, "lr": 0.0006204098360655738, "grad_norm": 0.780672, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:16.844805+00:00", "epoch": 0, "step": 146, "train_loss": 6.354985237121582, "perplexity": 575.3538467178734, "lr": 0.0006246885245901639, "grad_norm": 0.506654, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:17.150736+00:00", "epoch": 0, "step": 147, "train_loss": 6.334907054901123, "perplexity": 563.9169872290768, "lr": 0.0006289672131147542, "grad_norm": 0.434589, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:17.457175+00:00", "epoch": 0, "step": 148, "train_loss": 6.208045959472656, "perplexity": 496.7296724326574, "lr": 0.0006332459016393443, "grad_norm": 0.450263, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:17.765052+00:00", "epoch": 0, "step": 149, "train_loss": 6.281805038452148, "perplexity": 534.7530429979895, "lr": 0.0006375245901639344, "grad_norm": 0.471826, "tokens_per_sec": 106432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:18.073323+00:00", "epoch": 0, "step": 150, "train_loss": 6.348578929901123, "perplexity": 571.6797345267466, "lr": 0.0006418032786885245, "grad_norm": 0.355838, "tokens_per_sec": 106352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:18.381007+00:00", "epoch": 0, "step": 151, "train_loss": 6.398825645446777, "perplexity": 601.1386732539547, "lr": 0.0006460819672131147, "grad_norm": 0.436587, "tokens_per_sec": 106444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:18.687022+00:00", "epoch": 0, "step": 152, "train_loss": 6.204312324523926, "perplexity": 494.87852307598075, "lr": 0.0006503606557377048, "grad_norm": 0.449853, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:18.992888+00:00", "epoch": 0, "step": 153, "train_loss": 6.2759599685668945, "perplexity": 531.6364911948957, "lr": 0.000654639344262295, "grad_norm": 0.477354, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:19.299885+00:00", "epoch": 0, "step": 154, "train_loss": 6.123437881469727, "perplexity": 456.4311565165103, "lr": 0.0006589180327868852, "grad_norm": 0.445684, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:19.607135+00:00", "epoch": 0, "step": 155, "train_loss": 6.310477256774902, "perplexity": 550.3075241433045, "lr": 0.0006631967213114754, "grad_norm": 0.598103, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:19.913464+00:00", "epoch": 0, "step": 156, "train_loss": 6.308550834655762, "perplexity": 549.2484200248024, "lr": 0.0006674754098360655, "grad_norm": 0.764661, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:20.219918+00:00", "epoch": 0, "step": 157, "train_loss": 6.163554668426514, "perplexity": 475.11394925177706, "lr": 0.0006717540983606557, "grad_norm": 0.886103, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:20.526352+00:00", "epoch": 0, "step": 158, "train_loss": 6.225561141967773, "perplexity": 505.5066238663114, "lr": 0.0006760327868852459, "grad_norm": 0.678047, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:20.833280+00:00", "epoch": 0, "step": 159, "train_loss": 6.295027256011963, "perplexity": 541.8706154237982, "lr": 0.000680311475409836, "grad_norm": 0.576371, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:21.140135+00:00", "epoch": 0, "step": 160, "train_loss": 6.115100860595703, "perplexity": 452.6416987803206, "lr": 0.0006845901639344261, "grad_norm": 0.67423, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:21.446946+00:00", "epoch": 0, "step": 161, "train_loss": 6.346736907958984, "perplexity": 570.6276571841262, "lr": 0.0006888688524590164, "grad_norm": 0.470855, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:21.754048+00:00", "epoch": 0, "step": 162, "train_loss": 6.195205211639404, "perplexity": 490.39206883346975, "lr": 0.0006931475409836065, "grad_norm": 0.495202, "tokens_per_sec": 106700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:22.060734+00:00", "epoch": 0, "step": 163, "train_loss": 6.1501240730285645, "perplexity": 468.77554557602997, "lr": 0.0006974262295081966, "grad_norm": 0.473236, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:22.367608+00:00", "epoch": 0, "step": 164, "train_loss": 6.192189693450928, "perplexity": 488.91551004426566, "lr": 0.0007017049180327869, "grad_norm": 0.488709, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:22.675114+00:00", "epoch": 0, "step": 165, "train_loss": 6.1884074211120605, "perplexity": 487.0697911406993, "lr": 0.000705983606557377, "grad_norm": 0.507606, "tokens_per_sec": 106560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:22.982433+00:00", "epoch": 0, "step": 166, "train_loss": 6.224830150604248, "perplexity": 505.1372379154786, "lr": 0.0007102622950819671, "grad_norm": 0.387085, "tokens_per_sec": 106625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:23.289021+00:00", "epoch": 0, "step": 167, "train_loss": 6.1226115226745605, "perplexity": 456.0541364143284, "lr": 0.0007145409836065574, "grad_norm": 0.441355, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:23.595728+00:00", "epoch": 0, "step": 168, "train_loss": 6.3125200271606445, "perplexity": 551.4328250310247, "lr": 0.0007188196721311475, "grad_norm": 0.419652, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:23.903529+00:00", "epoch": 0, "step": 169, "train_loss": 6.171010494232178, "perplexity": 478.6695546112057, "lr": 0.0007230983606557376, "grad_norm": 0.356037, "tokens_per_sec": 106459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:24.210444+00:00", "epoch": 0, "step": 170, "train_loss": 6.136927127838135, "perplexity": 462.62978224032304, "lr": 0.0007273770491803279, "grad_norm": 0.442195, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:24.518107+00:00", "epoch": 0, "step": 171, "train_loss": 6.024311065673828, "perplexity": 413.3567682380369, "lr": 0.000731655737704918, "grad_norm": 0.685158, "tokens_per_sec": 106452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:24.826765+00:00", "epoch": 0, "step": 172, "train_loss": 6.250603199005127, "perplexity": 518.3253837470106, "lr": 0.0007359344262295082, "grad_norm": 1.278025, "tokens_per_sec": 106161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:25.134047+00:00", "epoch": 0, "step": 173, "train_loss": 6.23824405670166, "perplexity": 511.9587506727392, "lr": 0.0007402131147540983, "grad_norm": 2.048225, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:25.440908+00:00", "epoch": 0, "step": 174, "train_loss": 6.222705364227295, "perplexity": 504.0650686625536, "lr": 0.0007444918032786885, "grad_norm": 1.562256, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:25.748239+00:00", "epoch": 0, "step": 175, "train_loss": 6.335722923278809, "perplexity": 564.3772570008415, "lr": 0.0007487704918032787, "grad_norm": 2.048106, "tokens_per_sec": 106622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:26.055595+00:00", "epoch": 0, "step": 176, "train_loss": 6.3766913414001465, "perplexity": 587.9790637516289, "lr": 0.0007530491803278688, "grad_norm": 2.49816, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:26.363561+00:00", "epoch": 0, "step": 177, "train_loss": 6.102909564971924, "perplexity": 447.1569112841788, "lr": 0.000757327868852459, "grad_norm": 0.939938, "tokens_per_sec": 106401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:26.671704+00:00", "epoch": 0, "step": 178, "train_loss": 6.160481929779053, "perplexity": 473.65628890949955, "lr": 0.0007616065573770492, "grad_norm": 2.147427, "tokens_per_sec": 106341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:26.979026+00:00", "epoch": 0, "step": 179, "train_loss": 6.175258159637451, "perplexity": 480.70710707587165, "lr": 0.0007658852459016393, "grad_norm": 1.298857, "tokens_per_sec": 106625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:27.286519+00:00", "epoch": 0, "step": 180, "train_loss": 6.123754501342773, "perplexity": 456.5756945719441, "lr": 0.0007701639344262295, "grad_norm": 1.489507, "tokens_per_sec": 106624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:27.593609+00:00", "epoch": 0, "step": 181, "train_loss": 6.250177383422852, "perplexity": 518.1047197063499, "lr": 0.0007744426229508197, "grad_norm": 1.856845, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:27.902361+00:00", "epoch": 0, "step": 182, "train_loss": 6.124340057373047, "perplexity": 456.84312351285024, "lr": 0.0007787213114754098, "grad_norm": 1.709776, "tokens_per_sec": 106132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:28.210590+00:00", "epoch": 0, "step": 183, "train_loss": 6.142582893371582, "perplexity": 465.2537210142305, "lr": 0.000783, "grad_norm": 1.192055, "tokens_per_sec": 106310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:28.518760+00:00", "epoch": 0, "step": 184, "train_loss": 6.226996898651123, "perplexity": 506.2329296544326, "lr": 0.0007872786885245902, "grad_norm": 1.059669, "tokens_per_sec": 106331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:28.827004+00:00", "epoch": 0, "step": 185, "train_loss": 6.246550559997559, "perplexity": 516.2290487917934, "lr": 0.0007915573770491803, "grad_norm": 0.719653, "tokens_per_sec": 106305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:29.134541+00:00", "epoch": 0, "step": 186, "train_loss": 6.16975212097168, "perplexity": 478.06758847155, "lr": 0.0007958360655737705, "grad_norm": 0.850066, "tokens_per_sec": 106550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:29.441936+00:00", "epoch": 0, "step": 187, "train_loss": 6.1798996925354, "perplexity": 482.94351108397353, "lr": 0.0008001147540983607, "grad_norm": 0.659281, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:29.749102+00:00", "epoch": 0, "step": 188, "train_loss": 6.093594074249268, "perplexity": 443.01076689327033, "lr": 0.0008043934426229508, "grad_norm": 0.599429, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:30.057168+00:00", "epoch": 0, "step": 189, "train_loss": 6.073849678039551, "perplexity": 434.3495731790681, "lr": 0.000808672131147541, "grad_norm": 0.492632, "tokens_per_sec": 106368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:30.365459+00:00", "epoch": 0, "step": 190, "train_loss": 6.095026016235352, "perplexity": 443.6455870149923, "lr": 0.0008129508196721312, "grad_norm": 0.463723, "tokens_per_sec": 106352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:30.672762+00:00", "epoch": 0, "step": 191, "train_loss": 6.0404276847839355, "perplexity": 420.0726551562665, "lr": 0.0008172295081967213, "grad_norm": 0.510622, "tokens_per_sec": 106568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:30.979859+00:00", "epoch": 0, "step": 192, "train_loss": 6.146116733551025, "perplexity": 466.90076178258465, "lr": 0.0008215081967213115, "grad_norm": 0.406616, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:31.289728+00:00", "epoch": 0, "step": 193, "train_loss": 6.122403144836426, "perplexity": 455.95911473985933, "lr": 0.0008257868852459017, "grad_norm": 0.399615, "tokens_per_sec": 105747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:31.597516+00:00", "epoch": 0, "step": 194, "train_loss": 6.148472785949707, "perplexity": 468.0021013397218, "lr": 0.0008300655737704918, "grad_norm": 0.416895, "tokens_per_sec": 106463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:31.905041+00:00", "epoch": 0, "step": 195, "train_loss": 6.055781364440918, "perplexity": 426.57208352846385, "lr": 0.000834344262295082, "grad_norm": 0.498441, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:32.211889+00:00", "epoch": 0, "step": 196, "train_loss": 6.0180206298828125, "perplexity": 410.76473508484645, "lr": 0.0008386229508196722, "grad_norm": 0.377901, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:32.519008+00:00", "epoch": 0, "step": 197, "train_loss": 5.930573463439941, "perplexity": 376.3702865491578, "lr": 0.0008429016393442623, "grad_norm": 0.262911, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:32.826370+00:00", "epoch": 0, "step": 198, "train_loss": 6.000246047973633, "perplexity": 403.52806854259074, "lr": 0.0008471803278688525, "grad_norm": 0.371591, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:33.134526+00:00", "epoch": 0, "step": 199, "train_loss": 6.009607791900635, "perplexity": 407.3235333522445, "lr": 0.0008514590163934426, "grad_norm": 0.349525, "tokens_per_sec": 106336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:33.442846+00:00", "epoch": 0, "step": 200, "train_loss": 6.012701511383057, "perplexity": 408.58562938192597, "lr": 0.0008557377049180328, "grad_norm": 0.275546, "tokens_per_sec": 106336} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:00:39.155330+00:00", "step": 200, "epoch": 0, "val_loss": 5.9880232334136965, "val_ppl": 398.6258403356609, "eval_train_loss": 6.012701511383057, "eval_train_ppl": 408.58562938192597} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:00:40.139963+00:00", "step": 200, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_5p9880_epoch_0000_step_0000200.pt", "val_loss": 5.9880232334136965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:41.085663+00:00", "epoch": 0, "step": 201, "train_loss": 5.961450576782227, "perplexity": 388.1727904811908, "lr": 0.000860016393442623, "grad_norm": 0.292132, "tokens_per_sec": 4287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:41.392432+00:00", "epoch": 0, "step": 202, "train_loss": 5.847477912902832, "perplexity": 346.35972856582083, "lr": 0.0008642950819672131, "grad_norm": 0.332264, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:41.698464+00:00", "epoch": 0, "step": 203, "train_loss": 5.962233543395996, "perplexity": 388.47683582965726, "lr": 0.0008685737704918034, "grad_norm": 0.361177, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:42.004498+00:00", "epoch": 0, "step": 204, "train_loss": 5.950173854827881, "perplexity": 383.8200622317276, "lr": 0.0008728524590163935, "grad_norm": 0.264279, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:42.311445+00:00", "epoch": 0, "step": 205, "train_loss": 5.9767022132873535, "perplexity": 394.1384380873338, "lr": 0.0008771311475409836, "grad_norm": 0.281593, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:42.619598+00:00", "epoch": 0, "step": 206, "train_loss": 5.9850850105285645, "perplexity": 397.45630778403176, "lr": 0.0008814098360655739, "grad_norm": 0.391732, "tokens_per_sec": 106336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:42.929004+00:00", "epoch": 0, "step": 207, "train_loss": 5.886225700378418, "perplexity": 360.04380347888156, "lr": 0.000885688524590164, "grad_norm": 0.493298, "tokens_per_sec": 105906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:43.236618+00:00", "epoch": 0, "step": 208, "train_loss": 5.987180709838867, "perplexity": 398.2901301093905, "lr": 0.0008899672131147541, "grad_norm": 0.687134, "tokens_per_sec": 106523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:43.543612+00:00", "epoch": 0, "step": 209, "train_loss": 5.968730926513672, "perplexity": 391.0091364193105, "lr": 0.0008942459016393444, "grad_norm": 0.694923, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:43.850839+00:00", "epoch": 0, "step": 210, "train_loss": 5.962837219238281, "perplexity": 388.71142071022365, "lr": 0.0008985245901639345, "grad_norm": 0.618451, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:44.158575+00:00", "epoch": 0, "step": 211, "train_loss": 5.825278282165527, "perplexity": 338.7553895457588, "lr": 0.0009028032786885246, "grad_norm": 0.991402, "tokens_per_sec": 106417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:44.467057+00:00", "epoch": 0, "step": 212, "train_loss": 5.932689189910889, "perplexity": 377.1674260945536, "lr": 0.0009070819672131148, "grad_norm": 1.072554, "tokens_per_sec": 106224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:44.774193+00:00", "epoch": 0, "step": 213, "train_loss": 5.990201950073242, "perplexity": 399.49527988221615, "lr": 0.000911360655737705, "grad_norm": 0.649383, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:45.081214+00:00", "epoch": 0, "step": 214, "train_loss": 5.776695251464844, "perplexity": 322.69101370683944, "lr": 0.000915639344262295, "grad_norm": 0.668172, "tokens_per_sec": 106727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:45.388293+00:00", "epoch": 0, "step": 215, "train_loss": 5.851956844329834, "perplexity": 347.91452936300385, "lr": 0.0009199180327868851, "grad_norm": 0.613148, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:45.696210+00:00", "epoch": 0, "step": 216, "train_loss": 5.758347988128662, "perplexity": 316.82449851728313, "lr": 0.0009241967213114752, "grad_norm": 0.43088, "tokens_per_sec": 106418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:46.004292+00:00", "epoch": 0, "step": 217, "train_loss": 5.983169078826904, "perplexity": 396.6955376684338, "lr": 0.0009284754098360655, "grad_norm": 0.515445, "tokens_per_sec": 106362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:46.312534+00:00", "epoch": 0, "step": 218, "train_loss": 5.778013229370117, "perplexity": 323.1165937240601, "lr": 0.0009327540983606556, "grad_norm": 0.413547, "tokens_per_sec": 106306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:46.621617+00:00", "epoch": 0, "step": 219, "train_loss": 5.840498924255371, "perplexity": 343.9509033113437, "lr": 0.0009370327868852458, "grad_norm": 0.430166, "tokens_per_sec": 106017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:46.928780+00:00", "epoch": 0, "step": 220, "train_loss": 5.8730010986328125, "perplexity": 355.31371328338633, "lr": 0.000941311475409836, "grad_norm": 0.418847, "tokens_per_sec": 106737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:47.236436+00:00", "epoch": 0, "step": 221, "train_loss": 5.856788158416748, "perplexity": 349.5994807169102, "lr": 0.0009455901639344261, "grad_norm": 0.456606, "tokens_per_sec": 106453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:47.543917+00:00", "epoch": 0, "step": 222, "train_loss": 5.874536037445068, "perplexity": 355.85951687282727, "lr": 0.0009498688524590163, "grad_norm": 0.429806, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:47.850681+00:00", "epoch": 0, "step": 223, "train_loss": 5.869314670562744, "perplexity": 354.00628618566833, "lr": 0.0009541475409836065, "grad_norm": 0.375675, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:48.158705+00:00", "epoch": 0, "step": 224, "train_loss": 5.955068588256836, "perplexity": 385.7033644922259, "lr": 0.0009584262295081966, "grad_norm": 0.350385, "tokens_per_sec": 106382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:48.467185+00:00", "epoch": 0, "step": 225, "train_loss": 5.77197790145874, "perplexity": 321.1723520968174, "lr": 0.0009627049180327868, "grad_norm": 0.387328, "tokens_per_sec": 106224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:48.775323+00:00", "epoch": 0, "step": 226, "train_loss": 5.892688751220703, "perplexity": 362.3783208120638, "lr": 0.000966983606557377, "grad_norm": 0.506808, "tokens_per_sec": 106342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:49.083177+00:00", "epoch": 0, "step": 227, "train_loss": 5.727747440338135, "perplexity": 307.2763299619301, "lr": 0.0009712622950819671, "grad_norm": 0.739258, "tokens_per_sec": 106441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:49.390004+00:00", "epoch": 0, "step": 228, "train_loss": 5.796812057495117, "perplexity": 329.24826053461095, "lr": 0.0009755409836065573, "grad_norm": 0.827508, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:49.697748+00:00", "epoch": 0, "step": 229, "train_loss": 5.801435947418213, "perplexity": 330.7741934027516, "lr": 0.0009798196721311474, "grad_norm": 0.834975, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:50.006084+00:00", "epoch": 0, "step": 230, "train_loss": 5.840269088745117, "perplexity": 343.8718602637764, "lr": 0.0009840983606557376, "grad_norm": 0.867641, "tokens_per_sec": 106332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:50.314371+00:00", "epoch": 0, "step": 231, "train_loss": 5.76706600189209, "perplexity": 319.5986538476255, "lr": 0.0009883770491803277, "grad_norm": 0.579703, "tokens_per_sec": 106233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:50.621259+00:00", "epoch": 0, "step": 232, "train_loss": 5.881410121917725, "perplexity": 358.3141522720109, "lr": 0.000992655737704918, "grad_norm": 0.951759, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:50.927542+00:00", "epoch": 0, "step": 233, "train_loss": 5.875965595245361, "perplexity": 356.3686024178885, "lr": 0.0009969344262295082, "grad_norm": 0.937999, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:51.235254+00:00", "epoch": 0, "step": 234, "train_loss": 5.743649005889893, "perplexity": 312.2015603197318, "lr": 0.0010012131147540983, "grad_norm": 0.683544, "tokens_per_sec": 106490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:51.544167+00:00", "epoch": 0, "step": 235, "train_loss": 5.889297008514404, "perplexity": 361.1513088163092, "lr": 0.0010054918032786884, "grad_norm": 0.573685, "tokens_per_sec": 106075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:51.852827+00:00", "epoch": 0, "step": 236, "train_loss": 5.678746700286865, "perplexity": 292.58250646662566, "lr": 0.0010097704918032786, "grad_norm": 0.530284, "tokens_per_sec": 106163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:52.161474+00:00", "epoch": 0, "step": 237, "train_loss": 5.7315287590026855, "perplexity": 308.4404392305705, "lr": 0.0010140491803278687, "grad_norm": 0.482581, "tokens_per_sec": 106167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:52.469005+00:00", "epoch": 0, "step": 238, "train_loss": 5.83985710144043, "perplexity": 343.73021860219507, "lr": 0.001018327868852459, "grad_norm": 0.479033, "tokens_per_sec": 106552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:52.776458+00:00", "epoch": 0, "step": 239, "train_loss": 5.745815277099609, "perplexity": 312.87860663969, "lr": 0.0010226065573770492, "grad_norm": 0.526028, "tokens_per_sec": 106580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:53.083460+00:00", "epoch": 0, "step": 240, "train_loss": 5.8025689125061035, "perplexity": 331.1491613885612, "lr": 0.0010268852459016393, "grad_norm": 0.54173, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:53.392782+00:00", "epoch": 0, "step": 241, "train_loss": 5.744084358215332, "perplexity": 312.33750758535473, "lr": 0.0010311639344262294, "grad_norm": 0.526991, "tokens_per_sec": 105879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:53.700855+00:00", "epoch": 0, "step": 242, "train_loss": 5.686390399932861, "perplexity": 294.8274883207724, "lr": 0.0010354426229508195, "grad_norm": 0.510096, "tokens_per_sec": 106365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:54.009364+00:00", "epoch": 0, "step": 243, "train_loss": 5.791051387786865, "perplexity": 327.3570226820662, "lr": 0.0010397213114754099, "grad_norm": 0.498116, "tokens_per_sec": 106214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:54.316668+00:00", "epoch": 0, "step": 244, "train_loss": 5.6584272384643555, "perplexity": 286.69738117232225, "lr": 0.001044, "grad_norm": 0.665662, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:54.624521+00:00", "epoch": 0, "step": 245, "train_loss": 5.7926764488220215, "perplexity": 327.8894303044864, "lr": 0.0010482786885245901, "grad_norm": 0.853437, "tokens_per_sec": 106440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:54.931251+00:00", "epoch": 0, "step": 246, "train_loss": 5.7810468673706055, "perplexity": 324.09830082046216, "lr": 0.0010525573770491803, "grad_norm": 0.698642, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:55.239782+00:00", "epoch": 0, "step": 247, "train_loss": 5.695779800415039, "perplexity": 297.6087785724082, "lr": 0.0010568360655737704, "grad_norm": 0.591746, "tokens_per_sec": 106207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:55.548251+00:00", "epoch": 0, "step": 248, "train_loss": 5.673626899719238, "perplexity": 291.0883704867931, "lr": 0.0010611147540983605, "grad_norm": 0.616794, "tokens_per_sec": 106227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:55.856429+00:00", "epoch": 0, "step": 249, "train_loss": 5.704098701477051, "perplexity": 300.0948830477465, "lr": 0.0010653934426229509, "grad_norm": 0.693812, "tokens_per_sec": 106329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:56.162652+00:00", "epoch": 0, "step": 250, "train_loss": 5.763881683349609, "perplexity": 318.5825685562259, "lr": 0.001069672131147541, "grad_norm": 0.596376, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:00:57.073517+00:00", "step": 250, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0000_step_0000250.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:58.030506+00:00", "epoch": 0, "step": 251, "train_loss": 5.685887813568115, "perplexity": 294.6793492746406, "lr": 0.0010739508196721311, "grad_norm": 0.613539, "tokens_per_sec": 17542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:58.337268+00:00", "epoch": 0, "step": 252, "train_loss": 5.679304599761963, "perplexity": 292.7457836352945, "lr": 0.0010782295081967212, "grad_norm": 0.539202, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:58.644760+00:00", "epoch": 0, "step": 253, "train_loss": 5.755523681640625, "perplexity": 315.9309514499051, "lr": 0.0010825081967213114, "grad_norm": 0.524855, "tokens_per_sec": 106565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:58.952785+00:00", "epoch": 0, "step": 254, "train_loss": 5.674073696136475, "perplexity": 291.2184567867081, "lr": 0.0010867868852459015, "grad_norm": 0.337207, "tokens_per_sec": 106382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:59.261020+00:00", "epoch": 0, "step": 255, "train_loss": 5.776025295257568, "perplexity": 322.4748972615351, "lr": 0.0010910655737704918, "grad_norm": 0.404562, "tokens_per_sec": 106308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:59.570015+00:00", "epoch": 0, "step": 256, "train_loss": 5.648301124572754, "perplexity": 283.80890006407674, "lr": 0.001095344262295082, "grad_norm": 0.347379, "tokens_per_sec": 106048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:00:59.878419+00:00", "epoch": 0, "step": 257, "train_loss": 5.568302154541016, "perplexity": 261.9889047392782, "lr": 0.001099622950819672, "grad_norm": 0.396247, "tokens_per_sec": 106250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:00.185401+00:00", "epoch": 0, "step": 258, "train_loss": 5.67234992980957, "perplexity": 290.71689662757615, "lr": 0.0011039016393442622, "grad_norm": 0.425567, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:00.493137+00:00", "epoch": 0, "step": 259, "train_loss": 5.692895412445068, "perplexity": 296.75159620927553, "lr": 0.0011081803278688524, "grad_norm": 0.534469, "tokens_per_sec": 106480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:00.801157+00:00", "epoch": 0, "step": 260, "train_loss": 5.642047882080078, "perplexity": 282.0397115378819, "lr": 0.0011124590163934427, "grad_norm": 0.676309, "tokens_per_sec": 106444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:01.110920+00:00", "epoch": 0, "step": 261, "train_loss": 5.599400520324707, "perplexity": 270.2643408738938, "lr": 0.0011167377049180328, "grad_norm": 0.500466, "tokens_per_sec": 105724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:01.420049+00:00", "epoch": 0, "step": 262, "train_loss": 5.5633368492126465, "perplexity": 260.6912740668576, "lr": 0.001121016393442623, "grad_norm": 0.500103, "tokens_per_sec": 106003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:01.728059+00:00", "epoch": 0, "step": 263, "train_loss": 5.651444435119629, "perplexity": 284.702403116048, "lr": 0.001125295081967213, "grad_norm": 0.575465, "tokens_per_sec": 106385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:02.036020+00:00", "epoch": 0, "step": 264, "train_loss": 5.598125457763672, "perplexity": 269.9199565337099, "lr": 0.0011295737704918032, "grad_norm": 0.548313, "tokens_per_sec": 106404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:02.416432+00:00", "epoch": 0, "step": 265, "train_loss": 5.566240310668945, "perplexity": 261.44928102258854, "lr": 0.0011338524590163933, "grad_norm": 0.541777, "tokens_per_sec": 86139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:02.723089+00:00", "epoch": 0, "step": 266, "train_loss": 5.472217082977295, "perplexity": 237.98724575654265, "lr": 0.0011381311475409837, "grad_norm": 0.585346, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:03.030962+00:00", "epoch": 0, "step": 267, "train_loss": 5.606888294219971, "perplexity": 272.29561451862713, "lr": 0.0011424098360655738, "grad_norm": 0.582527, "tokens_per_sec": 106434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:03.339051+00:00", "epoch": 0, "step": 268, "train_loss": 5.49722957611084, "perplexity": 244.01497005871164, "lr": 0.001146688524590164, "grad_norm": 0.61808, "tokens_per_sec": 106360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:03.648156+00:00", "epoch": 0, "step": 269, "train_loss": 5.460811614990234, "perplexity": 235.28831041152884, "lr": 0.001150967213114754, "grad_norm": 0.810759, "tokens_per_sec": 106008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:03.956198+00:00", "epoch": 0, "step": 270, "train_loss": 5.615996837615967, "perplexity": 274.7871608931253, "lr": 0.0011552459016393442, "grad_norm": 1.077888, "tokens_per_sec": 106437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:04.264498+00:00", "epoch": 0, "step": 271, "train_loss": 5.733131408691406, "perplexity": 308.93515752866574, "lr": 0.0011595245901639343, "grad_norm": 1.264861, "tokens_per_sec": 106225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:04.571492+00:00", "epoch": 0, "step": 272, "train_loss": 5.771587371826172, "perplexity": 321.04694926451674, "lr": 0.0011638032786885247, "grad_norm": 1.329842, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:04.879435+00:00", "epoch": 0, "step": 273, "train_loss": 5.697127819061279, "perplexity": 298.0102312772826, "lr": 0.0011680819672131148, "grad_norm": 1.139105, "tokens_per_sec": 106410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:05.187442+00:00", "epoch": 0, "step": 274, "train_loss": 5.612246513366699, "perplexity": 273.75854995811835, "lr": 0.001172360655737705, "grad_norm": 0.972852, "tokens_per_sec": 106387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:05.494795+00:00", "epoch": 0, "step": 275, "train_loss": 5.6590399742126465, "perplexity": 286.87310473726507, "lr": 0.001176639344262295, "grad_norm": 0.736843, "tokens_per_sec": 106614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:05.803382+00:00", "epoch": 0, "step": 276, "train_loss": 5.810926914215088, "perplexity": 333.9285053466379, "lr": 0.0011809180327868852, "grad_norm": 1.012516, "tokens_per_sec": 106189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:06.112201+00:00", "epoch": 0, "step": 277, "train_loss": 5.6035308837890625, "perplexity": 271.3829393520355, "lr": 0.0011851967213114755, "grad_norm": 0.978494, "tokens_per_sec": 106108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:06.419785+00:00", "epoch": 0, "step": 278, "train_loss": 5.687301158905029, "perplexity": 295.0961274151657, "lr": 0.0011894754098360656, "grad_norm": 0.695275, "tokens_per_sec": 106532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:06.726983+00:00", "epoch": 0, "step": 279, "train_loss": 5.605428218841553, "perplexity": 271.89833249764365, "lr": 0.0011937540983606558, "grad_norm": 0.687576, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:07.034192+00:00", "epoch": 0, "step": 280, "train_loss": 5.663059234619141, "perplexity": 288.0284426955588, "lr": 0.001198032786885246, "grad_norm": 0.836709, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:07.341370+00:00", "epoch": 0, "step": 281, "train_loss": 5.632937431335449, "perplexity": 279.4818718658257, "lr": 0.001202311475409836, "grad_norm": 0.684744, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:07.649593+00:00", "epoch": 0, "step": 282, "train_loss": 5.613656997680664, "perplexity": 274.14495454338964, "lr": 0.0012065901639344262, "grad_norm": 0.450102, "tokens_per_sec": 106313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:07.956806+00:00", "epoch": 0, "step": 283, "train_loss": 5.5709919929504395, "perplexity": 262.694561183527, "lr": 0.0012108688524590165, "grad_norm": 0.556894, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:08.264200+00:00", "epoch": 0, "step": 284, "train_loss": 5.535525798797607, "perplexity": 253.54106418238612, "lr": 0.0012151475409836066, "grad_norm": 0.489879, "tokens_per_sec": 106600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:08.571858+00:00", "epoch": 0, "step": 285, "train_loss": 5.481876850128174, "perplexity": 240.2972864001177, "lr": 0.0012194262295081968, "grad_norm": 0.421385, "tokens_per_sec": 106508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:08.879420+00:00", "epoch": 0, "step": 286, "train_loss": 5.497481346130371, "perplexity": 244.0764134469662, "lr": 0.0012237049180327869, "grad_norm": 0.408225, "tokens_per_sec": 106542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:09.187744+00:00", "epoch": 0, "step": 287, "train_loss": 5.5209455490112305, "perplexity": 249.8711909821748, "lr": 0.001227983606557377, "grad_norm": 0.384657, "tokens_per_sec": 106277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:09.497076+00:00", "epoch": 0, "step": 288, "train_loss": 5.539152145385742, "perplexity": 254.462161054287, "lr": 0.0012322622950819671, "grad_norm": 0.367619, "tokens_per_sec": 105933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:09.804725+00:00", "epoch": 0, "step": 289, "train_loss": 5.609882831573486, "perplexity": 273.1122359990274, "lr": 0.0012365409836065575, "grad_norm": 0.369676, "tokens_per_sec": 106511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:10.112089+00:00", "epoch": 0, "step": 290, "train_loss": 5.513192176818848, "perplexity": 247.94133774127832, "lr": 0.0012408196721311476, "grad_norm": 0.328525, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:10.418966+00:00", "epoch": 0, "step": 291, "train_loss": 5.59016227722168, "perplexity": 267.7790706316108, "lr": 0.0012450983606557377, "grad_norm": 0.333991, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:10.726920+00:00", "epoch": 0, "step": 292, "train_loss": 5.526872634887695, "perplexity": 251.35659670425312, "lr": 0.0012493770491803279, "grad_norm": 0.400788, "tokens_per_sec": 106407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:11.036295+00:00", "epoch": 0, "step": 293, "train_loss": 5.470344543457031, "perplexity": 237.54202221297962, "lr": 0.001253655737704918, "grad_norm": 0.427473, "tokens_per_sec": 105916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:11.343114+00:00", "epoch": 0, "step": 294, "train_loss": 5.519253730773926, "perplexity": 249.44881173949324, "lr": 0.0012579344262295083, "grad_norm": 0.410927, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:11.650330+00:00", "epoch": 0, "step": 295, "train_loss": 5.452943801879883, "perplexity": 233.44436936197707, "lr": 0.0012622131147540985, "grad_norm": 0.351796, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:11.957569+00:00", "epoch": 0, "step": 296, "train_loss": 5.551644802093506, "perplexity": 257.66100892899993, "lr": 0.0012664918032786886, "grad_norm": 0.328019, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:12.264930+00:00", "epoch": 0, "step": 297, "train_loss": 5.40281343460083, "perplexity": 222.0302057604797, "lr": 0.0012707704918032787, "grad_norm": 0.343907, "tokens_per_sec": 106611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:12.573346+00:00", "epoch": 0, "step": 298, "train_loss": 5.4967265129089355, "perplexity": 243.89224597813396, "lr": 0.0012750491803278688, "grad_norm": 0.428794, "tokens_per_sec": 106246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:12.882559+00:00", "epoch": 0, "step": 299, "train_loss": 5.5067009925842285, "perplexity": 246.33711711731223, "lr": 0.001279327868852459, "grad_norm": 0.477079, "tokens_per_sec": 105979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:13.191634+00:00", "epoch": 0, "step": 300, "train_loss": 5.459721565246582, "perplexity": 235.0319741840816, "lr": 0.001283606557377049, "grad_norm": 0.368563, "tokens_per_sec": 106075} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:01:18.942698+00:00", "step": 300, "epoch": 0, "val_loss": 5.46511480808258, "val_ppl": 236.30298304396914, "eval_train_loss": 5.459721565246582, "eval_train_ppl": 235.0319741840816} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:01:20.053213+00:00", "step": 300, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_5p4651_epoch_0000_step_0000300.pt", "val_loss": 5.46511480808258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:21.020083+00:00", "epoch": 0, "step": 301, "train_loss": 5.436943531036377, "perplexity": 229.73891937889582, "lr": 0.0012878852459016392, "grad_norm": 0.466919, "tokens_per_sec": 4186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:21.326482+00:00", "epoch": 0, "step": 302, "train_loss": 5.482722759246826, "perplexity": 240.50064206395004, "lr": 0.0012921639344262294, "grad_norm": 0.570851, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:21.632813+00:00", "epoch": 0, "step": 303, "train_loss": 5.460334300994873, "perplexity": 235.17603080646964, "lr": 0.0012964426229508195, "grad_norm": 0.667666, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:21.940429+00:00", "epoch": 0, "step": 304, "train_loss": 5.460587978363037, "perplexity": 235.2356972107081, "lr": 0.0013007213114754096, "grad_norm": 0.682361, "tokens_per_sec": 106522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:22.247602+00:00", "epoch": 0, "step": 305, "train_loss": 5.406628608703613, "perplexity": 222.87890759471873, "lr": 0.001305, "grad_norm": 0.820245, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:22.553582+00:00", "epoch": 0, "step": 306, "train_loss": 5.406545162200928, "perplexity": 222.8603099053244, "lr": 0.001305, "grad_norm": 0.80075, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:22.860023+00:00", "epoch": 0, "step": 307, "train_loss": 5.516578674316406, "perplexity": 248.78241380826472, "lr": 0.001305, "grad_norm": 0.805395, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:23.166250+00:00", "epoch": 0, "step": 308, "train_loss": 5.638689994812012, "perplexity": 281.094242259425, "lr": 0.001305, "grad_norm": 0.728669, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:23.472843+00:00", "epoch": 0, "step": 309, "train_loss": 5.235008239746094, "perplexity": 187.73065319074243, "lr": 0.001305, "grad_norm": 0.578751, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:23.780815+00:00", "epoch": 0, "step": 310, "train_loss": 5.445008277893066, "perplexity": 231.59919685757538, "lr": 0.001305, "grad_norm": 0.601564, "tokens_per_sec": 106456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:24.089118+00:00", "epoch": 0, "step": 311, "train_loss": 5.544313907623291, "perplexity": 255.77902998640133, "lr": 0.001305, "grad_norm": 0.676637, "tokens_per_sec": 106233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:24.394964+00:00", "epoch": 0, "step": 312, "train_loss": 5.358590126037598, "perplexity": 212.42524240850065, "lr": 0.001305, "grad_norm": 0.731232, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:24.700186+00:00", "epoch": 0, "step": 313, "train_loss": 5.41255521774292, "perplexity": 224.20374576420005, "lr": 0.001305, "grad_norm": 0.410842, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:25.006892+00:00", "epoch": 0, "step": 314, "train_loss": 5.517688274383545, "perplexity": 249.0586159999759, "lr": 0.001305, "grad_norm": 0.431568, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:25.378329+00:00", "epoch": 0, "step": 315, "train_loss": 5.502549171447754, "perplexity": 245.31648966517895, "lr": 0.001305, "grad_norm": 0.365318, "tokens_per_sec": 88219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:25.684511+00:00", "epoch": 0, "step": 316, "train_loss": 5.455609321594238, "perplexity": 234.06744997863498, "lr": 0.001305, "grad_norm": 0.329419, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:25.991084+00:00", "epoch": 0, "step": 317, "train_loss": 5.338178634643555, "perplexity": 208.13327810940396, "lr": 0.001305, "grad_norm": 0.329334, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:26.298166+00:00", "epoch": 0, "step": 318, "train_loss": 5.376036643981934, "perplexity": 216.16384114762096, "lr": 0.001305, "grad_norm": 0.284618, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:26.605622+00:00", "epoch": 0, "step": 319, "train_loss": 5.401314735412598, "perplexity": 221.6976984977628, "lr": 0.001305, "grad_norm": 0.342597, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:26.912543+00:00", "epoch": 0, "step": 320, "train_loss": 5.461167335510254, "perplexity": 235.37202217977708, "lr": 0.001305, "grad_norm": 0.262643, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:27.217987+00:00", "epoch": 0, "step": 321, "train_loss": 5.245477676391602, "perplexity": 189.70641186414323, "lr": 0.001305, "grad_norm": 0.327174, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:27.523691+00:00", "epoch": 0, "step": 322, "train_loss": 5.375264644622803, "perplexity": 215.99702719920322, "lr": 0.001305, "grad_norm": 0.311463, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:27.830304+00:00", "epoch": 0, "step": 323, "train_loss": 5.379399299621582, "perplexity": 216.89194920939698, "lr": 0.001305, "grad_norm": 0.352623, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:28.137565+00:00", "epoch": 0, "step": 324, "train_loss": 5.32429838180542, "perplexity": 205.26429276793087, "lr": 0.001305, "grad_norm": 0.384594, "tokens_per_sec": 106646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:28.443794+00:00", "epoch": 0, "step": 325, "train_loss": 5.3028974533081055, "perplexity": 200.91811827773617, "lr": 0.001305, "grad_norm": 0.470038, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:28.749165+00:00", "epoch": 0, "step": 326, "train_loss": 5.339245319366455, "perplexity": 208.35540914832126, "lr": 0.001305, "grad_norm": 0.433014, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:29.055065+00:00", "epoch": 0, "step": 327, "train_loss": 5.203151226043701, "perplexity": 181.8443726680268, "lr": 0.001305, "grad_norm": 0.382551, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:29.361507+00:00", "epoch": 0, "step": 328, "train_loss": 5.31166410446167, "perplexity": 202.6872406400518, "lr": 0.001305, "grad_norm": 0.447435, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:29.668328+00:00", "epoch": 0, "step": 329, "train_loss": 5.349128723144531, "perplexity": 210.42487964680154, "lr": 0.001305, "grad_norm": 0.349571, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:29.976864+00:00", "epoch": 0, "step": 330, "train_loss": 5.3209004402160645, "perplexity": 204.56800034117234, "lr": 0.001305, "grad_norm": 0.329855, "tokens_per_sec": 106262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:30.284149+00:00", "epoch": 0, "step": 331, "train_loss": 5.326422691345215, "perplexity": 205.70080113853268, "lr": 0.001305, "grad_norm": 0.385401, "tokens_per_sec": 106580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:30.589599+00:00", "epoch": 0, "step": 332, "train_loss": 5.223235130310059, "perplexity": 185.53343906628277, "lr": 0.001305, "grad_norm": 0.398314, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:30.896335+00:00", "epoch": 0, "step": 333, "train_loss": 5.302866458892822, "perplexity": 200.91189103464575, "lr": 0.001305, "grad_norm": 0.380776, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:31.203957+00:00", "epoch": 0, "step": 334, "train_loss": 5.311126708984375, "perplexity": 202.5783466958037, "lr": 0.001305, "grad_norm": 0.657441, "tokens_per_sec": 106521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:31.511756+00:00", "epoch": 0, "step": 335, "train_loss": 5.3660502433776855, "perplexity": 214.0158854577169, "lr": 0.001305, "grad_norm": 0.836551, "tokens_per_sec": 106459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:31.817844+00:00", "epoch": 0, "step": 336, "train_loss": 5.2125244140625, "perplexity": 183.5568472988784, "lr": 0.001305, "grad_norm": 0.696309, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:32.123894+00:00", "epoch": 0, "step": 337, "train_loss": 5.232586860656738, "perplexity": 187.27663600835925, "lr": 0.001305, "grad_norm": 0.538974, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:32.430407+00:00", "epoch": 0, "step": 338, "train_loss": 5.346421241760254, "perplexity": 209.8559287618895, "lr": 0.001305, "grad_norm": 0.476622, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:32.738318+00:00", "epoch": 0, "step": 339, "train_loss": 5.163821697235107, "perplexity": 174.83133291346797, "lr": 0.001305, "grad_norm": 0.482362, "tokens_per_sec": 106420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:33.046588+00:00", "epoch": 0, "step": 340, "train_loss": 5.192441940307617, "perplexity": 179.90733995299988, "lr": 0.001305, "grad_norm": 0.58713, "tokens_per_sec": 106354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:33.353340+00:00", "epoch": 0, "step": 341, "train_loss": 5.337008476257324, "perplexity": 207.88987164840051, "lr": 0.001305, "grad_norm": 0.488389, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:33.659936+00:00", "epoch": 0, "step": 342, "train_loss": 5.351051330566406, "perplexity": 210.82983324081047, "lr": 0.001305, "grad_norm": 0.401231, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:33.967026+00:00", "epoch": 0, "step": 343, "train_loss": 5.341485977172852, "perplexity": 208.82278574230077, "lr": 0.001305, "grad_norm": 0.413318, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:34.273761+00:00", "epoch": 0, "step": 344, "train_loss": 5.2823710441589355, "perplexity": 196.8360294152895, "lr": 0.001305, "grad_norm": 0.351662, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:34.581184+00:00", "epoch": 0, "step": 345, "train_loss": 5.309459686279297, "perplexity": 202.24092531499628, "lr": 0.001305, "grad_norm": 0.398872, "tokens_per_sec": 106589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:34.889769+00:00", "epoch": 0, "step": 346, "train_loss": 5.4063944816589355, "perplexity": 222.82673172289225, "lr": 0.001305, "grad_norm": 0.444245, "tokens_per_sec": 106188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:35.196604+00:00", "epoch": 0, "step": 347, "train_loss": 5.210809230804443, "perplexity": 183.24228351187696, "lr": 0.001305, "grad_norm": 0.469042, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:35.501941+00:00", "epoch": 0, "step": 348, "train_loss": 5.301784515380859, "perplexity": 200.694633269173, "lr": 0.001305, "grad_norm": 0.51145, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:35.808311+00:00", "epoch": 0, "step": 349, "train_loss": 5.289669990539551, "perplexity": 198.27798100180894, "lr": 0.001305, "grad_norm": 0.461933, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:36.114416+00:00", "epoch": 0, "step": 350, "train_loss": 5.262245178222656, "perplexity": 192.91413204582526, "lr": 0.001305, "grad_norm": 0.411537, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:36.422418+00:00", "epoch": 0, "step": 351, "train_loss": 5.274269104003906, "perplexity": 195.2477185725516, "lr": 0.001305, "grad_norm": 0.39611, "tokens_per_sec": 106331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:36.729205+00:00", "epoch": 0, "step": 352, "train_loss": 5.3104753494262695, "perplexity": 202.44643831797325, "lr": 0.001305, "grad_norm": 0.355544, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:37.036150+00:00", "epoch": 0, "step": 353, "train_loss": 5.370416641235352, "perplexity": 214.95240708657107, "lr": 0.001305, "grad_norm": 0.338525, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:37.342914+00:00", "epoch": 0, "step": 354, "train_loss": 5.215454578399658, "perplexity": 184.09548779386822, "lr": 0.001305, "grad_norm": 0.367756, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:37.649146+00:00", "epoch": 0, "step": 355, "train_loss": 5.253167152404785, "perplexity": 191.17077765511547, "lr": 0.001305, "grad_norm": 0.401091, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:37.956413+00:00", "epoch": 0, "step": 356, "train_loss": 5.174892902374268, "perplexity": 176.77768079268947, "lr": 0.001305, "grad_norm": 0.404829, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:38.266152+00:00", "epoch": 0, "step": 357, "train_loss": 5.204789161682129, "perplexity": 182.14246610895796, "lr": 0.001305, "grad_norm": 0.425978, "tokens_per_sec": 105792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:38.573562+00:00", "epoch": 0, "step": 358, "train_loss": 5.18817663192749, "perplexity": 179.14161385811173, "lr": 0.001305, "grad_norm": 0.424156, "tokens_per_sec": 106595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:38.880148+00:00", "epoch": 0, "step": 359, "train_loss": 5.198083400726318, "perplexity": 180.92514835387973, "lr": 0.001305, "grad_norm": 0.392291, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:39.187217+00:00", "epoch": 0, "step": 360, "train_loss": 5.261115550994873, "perplexity": 192.69633402808785, "lr": 0.001305, "grad_norm": 0.349198, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:39.494306+00:00", "epoch": 0, "step": 361, "train_loss": 5.207937717437744, "perplexity": 182.71685559311263, "lr": 0.001305, "grad_norm": 0.333668, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:39.801321+00:00", "epoch": 0, "step": 362, "train_loss": 5.186452388763428, "perplexity": 178.83299629733784, "lr": 0.001305, "grad_norm": 0.325014, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:40.108324+00:00", "epoch": 0, "step": 363, "train_loss": 5.154169082641602, "perplexity": 173.1518720721077, "lr": 0.001305, "grad_norm": 0.2751, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:40.414850+00:00", "epoch": 0, "step": 364, "train_loss": 5.148106098175049, "perplexity": 172.10523105169364, "lr": 0.001305, "grad_norm": 0.421844, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:40.722687+00:00", "epoch": 0, "step": 365, "train_loss": 5.090207099914551, "perplexity": 162.4234964627658, "lr": 0.001305, "grad_norm": 0.627227, "tokens_per_sec": 106446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:41.029848+00:00", "epoch": 0, "step": 366, "train_loss": 5.158966064453125, "perplexity": 173.9844738442624, "lr": 0.001305, "grad_norm": 0.70277, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:41.337914+00:00", "epoch": 0, "step": 367, "train_loss": 5.169605255126953, "perplexity": 175.8454097079065, "lr": 0.001305, "grad_norm": 0.740966, "tokens_per_sec": 106367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:41.645660+00:00", "epoch": 0, "step": 368, "train_loss": 5.229701995849609, "perplexity": 186.73714678283724, "lr": 0.001305, "grad_norm": 0.77568, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:41.951829+00:00", "epoch": 0, "step": 369, "train_loss": 5.201250076293945, "perplexity": 181.49898770273373, "lr": 0.001305, "grad_norm": 0.5777, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:42.260219+00:00", "epoch": 0, "step": 370, "train_loss": 5.389519691467285, "perplexity": 219.09812557812884, "lr": 0.001305, "grad_norm": 0.654417, "tokens_per_sec": 106313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:42.566977+00:00", "epoch": 0, "step": 371, "train_loss": 5.289001941680908, "perplexity": 198.14556585772428, "lr": 0.001305, "grad_norm": 0.614225, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:42.874613+00:00", "epoch": 0, "step": 372, "train_loss": 5.187654495239258, "perplexity": 179.0481018642731, "lr": 0.001305, "grad_norm": 0.585289, "tokens_per_sec": 106515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:43.181640+00:00", "epoch": 0, "step": 373, "train_loss": 5.138167858123779, "perplexity": 170.40327916289885, "lr": 0.001305, "grad_norm": 0.504867, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:43.488121+00:00", "epoch": 0, "step": 374, "train_loss": 5.179093837738037, "perplexity": 177.5218744637601, "lr": 0.001305, "grad_norm": 0.528951, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:43.794657+00:00", "epoch": 0, "step": 375, "train_loss": 5.2932658195495605, "perplexity": 198.99223812156626, "lr": 0.001305, "grad_norm": 0.538333, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:44.103288+00:00", "epoch": 0, "step": 376, "train_loss": 5.235806465148926, "perplexity": 187.8805643905136, "lr": 0.001305, "grad_norm": 0.429179, "tokens_per_sec": 106168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:44.410382+00:00", "epoch": 0, "step": 377, "train_loss": 5.253096103668213, "perplexity": 191.1571956953898, "lr": 0.001305, "grad_norm": 0.440591, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:44.716534+00:00", "epoch": 0, "step": 378, "train_loss": 5.156478404998779, "perplexity": 173.55219762393418, "lr": 0.001305, "grad_norm": 0.397711, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:45.023025+00:00", "epoch": 0, "step": 379, "train_loss": 5.012808799743652, "perplexity": 150.3263804134569, "lr": 0.001305, "grad_norm": 0.435042, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:45.329817+00:00", "epoch": 0, "step": 380, "train_loss": 5.128236770629883, "perplexity": 168.71936467918, "lr": 0.001305, "grad_norm": 0.495445, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:45.637228+00:00", "epoch": 0, "step": 381, "train_loss": 5.16624116897583, "perplexity": 175.25484451342143, "lr": 0.001305, "grad_norm": 0.582349, "tokens_per_sec": 106533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:45.943973+00:00", "epoch": 0, "step": 382, "train_loss": 5.1115264892578125, "perplexity": 165.92344200903443, "lr": 0.001305, "grad_norm": 0.503283, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:46.252316+00:00", "epoch": 0, "step": 383, "train_loss": 5.149632453918457, "perplexity": 172.36812544372245, "lr": 0.001305, "grad_norm": 0.478828, "tokens_per_sec": 106272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:46.558930+00:00", "epoch": 0, "step": 384, "train_loss": 5.265932083129883, "perplexity": 193.62670088553622, "lr": 0.001305, "grad_norm": 0.485509, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:46.865743+00:00", "epoch": 0, "step": 385, "train_loss": 5.179430961608887, "perplexity": 177.58173141427633, "lr": 0.001305, "grad_norm": 0.40399, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:47.173336+00:00", "epoch": 0, "step": 386, "train_loss": 5.082340717315674, "perplexity": 161.1508233261919, "lr": 0.001305, "grad_norm": 0.380629, "tokens_per_sec": 106532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:47.483102+00:00", "epoch": 0, "step": 387, "train_loss": 5.033019065856934, "perplexity": 153.39542520832504, "lr": 0.001305, "grad_norm": 0.356498, "tokens_per_sec": 105784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:47.790809+00:00", "epoch": 0, "step": 388, "train_loss": 5.065300464630127, "perplexity": 158.4280369445582, "lr": 0.001305, "grad_norm": 0.344162, "tokens_per_sec": 106490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:48.098343+00:00", "epoch": 0, "step": 389, "train_loss": 5.145480632781982, "perplexity": 171.6539673714055, "lr": 0.001305, "grad_norm": 0.386864, "tokens_per_sec": 106551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:48.405445+00:00", "epoch": 0, "step": 390, "train_loss": 5.070240020751953, "perplexity": 159.21253707041794, "lr": 0.001305, "grad_norm": 0.403107, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:48.712666+00:00", "epoch": 0, "step": 391, "train_loss": 5.080151557922363, "perplexity": 160.7984243570613, "lr": 0.001305, "grad_norm": 0.384463, "tokens_per_sec": 106599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:49.019767+00:00", "epoch": 0, "step": 392, "train_loss": 5.084306240081787, "perplexity": 161.46788042761958, "lr": 0.001305, "grad_norm": 0.37144, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:49.328026+00:00", "epoch": 0, "step": 393, "train_loss": 5.084423542022705, "perplexity": 161.48682203431153, "lr": 0.001305, "grad_norm": 0.420949, "tokens_per_sec": 106299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:49.636574+00:00", "epoch": 0, "step": 394, "train_loss": 5.148723125457764, "perplexity": 172.21145744367124, "lr": 0.001305, "grad_norm": 0.499408, "tokens_per_sec": 106201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:49.943823+00:00", "epoch": 0, "step": 395, "train_loss": 5.196231365203857, "perplexity": 180.59037865060233, "lr": 0.001305, "grad_norm": 0.448374, "tokens_per_sec": 106651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:50.250306+00:00", "epoch": 0, "step": 396, "train_loss": 5.270313262939453, "perplexity": 194.47687530163714, "lr": 0.001305, "grad_norm": 0.421454, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:50.557061+00:00", "epoch": 0, "step": 397, "train_loss": 5.1516642570495605, "perplexity": 172.71869956890004, "lr": 0.001305, "grad_norm": 0.389805, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:50.863552+00:00", "epoch": 0, "step": 398, "train_loss": 4.973910808563232, "perplexity": 144.59125179669786, "lr": 0.001305, "grad_norm": 0.362372, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:51.172280+00:00", "epoch": 0, "step": 399, "train_loss": 5.039333343505859, "perplexity": 154.36707090521975, "lr": 0.001305, "grad_norm": 0.319937, "tokens_per_sec": 106139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:51.479777+00:00", "epoch": 0, "step": 400, "train_loss": 5.085159778594971, "perplexity": 161.60575831587988, "lr": 0.001305, "grad_norm": 0.353946, "tokens_per_sec": 106621} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:01:57.222428+00:00", "step": 400, "epoch": 0, "val_loss": 5.069458365440369, "val_ppl": 159.08813637071378, "eval_train_loss": 5.085159778594971, "eval_train_ppl": 161.60575831587988} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:01:58.118097+00:00", "step": 400, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_5p0695_epoch_0000_step_0000400.pt", "val_loss": 5.069458365440369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:59.116193+00:00", "epoch": 0, "step": 401, "train_loss": 5.126349449157715, "perplexity": 168.40123729819928, "lr": 0.001305, "grad_norm": 0.403871, "tokens_per_sec": 4291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:59.420981+00:00", "epoch": 0, "step": 402, "train_loss": 5.044564723968506, "perplexity": 155.1767397804573, "lr": 0.001305, "grad_norm": 0.460248, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:01:59.728802+00:00", "epoch": 0, "step": 403, "train_loss": 5.116114616394043, "perplexity": 166.68646894962487, "lr": 0.001305, "grad_norm": 0.415262, "tokens_per_sec": 106451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:00.034464+00:00", "epoch": 0, "step": 404, "train_loss": 5.035850524902344, "perplexity": 153.83037355117798, "lr": 0.001305, "grad_norm": 0.374442, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:00.339734+00:00", "epoch": 0, "step": 405, "train_loss": 4.99982213973999, "perplexity": 148.3867646468417, "lr": 0.001305, "grad_norm": 0.432189, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:00.645267+00:00", "epoch": 0, "step": 406, "train_loss": 5.127835750579834, "perplexity": 168.6517183957767, "lr": 0.001305, "grad_norm": 0.336661, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:00.951137+00:00", "epoch": 0, "step": 407, "train_loss": 4.961904525756836, "perplexity": 142.86562822771134, "lr": 0.001305, "grad_norm": 0.348219, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:01.256693+00:00", "epoch": 0, "step": 408, "train_loss": 5.058901309967041, "perplexity": 157.41746827428446, "lr": 0.001305, "grad_norm": 0.40981, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:01.563201+00:00", "epoch": 0, "step": 409, "train_loss": 5.042388916015625, "perplexity": 154.83947204390864, "lr": 0.001305, "grad_norm": 0.402626, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:01.869451+00:00", "epoch": 0, "step": 410, "train_loss": 5.01756477355957, "perplexity": 151.0430315787894, "lr": 0.001305, "grad_norm": 0.393416, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:02.175334+00:00", "epoch": 0, "step": 411, "train_loss": 5.085864543914795, "perplexity": 161.71969259357277, "lr": 0.001305, "grad_norm": 0.370646, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:02.481862+00:00", "epoch": 0, "step": 412, "train_loss": 5.001256465911865, "perplexity": 148.5997523773578, "lr": 0.001305, "grad_norm": 0.346699, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:02.788710+00:00", "epoch": 0, "step": 413, "train_loss": 5.015449523925781, "perplexity": 150.7238755279153, "lr": 0.001305, "grad_norm": 0.379348, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:03.096535+00:00", "epoch": 0, "step": 414, "train_loss": 5.007296085357666, "perplexity": 149.49995403278234, "lr": 0.001305, "grad_norm": 0.379816, "tokens_per_sec": 106451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:03.448197+00:00", "epoch": 0, "step": 415, "train_loss": 5.054950714111328, "perplexity": 156.79680228398615, "lr": 0.001305, "grad_norm": 0.473683, "tokens_per_sec": 93180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:03.755178+00:00", "epoch": 0, "step": 416, "train_loss": 5.075819492340088, "perplexity": 160.10334169609516, "lr": 0.001305, "grad_norm": 0.540859, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:04.060946+00:00", "epoch": 0, "step": 417, "train_loss": 5.070751667022705, "perplexity": 159.29401841430231, "lr": 0.001305, "grad_norm": 0.532716, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:04.367097+00:00", "epoch": 0, "step": 418, "train_loss": 4.966510772705078, "perplexity": 143.52522054805843, "lr": 0.001305, "grad_norm": 0.467185, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:04.673115+00:00", "epoch": 0, "step": 419, "train_loss": 5.1865668296813965, "perplexity": 178.85346328070506, "lr": 0.001305, "grad_norm": 0.470843, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:04.978713+00:00", "epoch": 0, "step": 420, "train_loss": 5.16265344619751, "perplexity": 174.62720528675936, "lr": 0.001305, "grad_norm": 0.53698, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:05.285444+00:00", "epoch": 0, "step": 421, "train_loss": 4.96150016784668, "perplexity": 142.80787105891903, "lr": 0.001305, "grad_norm": 0.567706, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:05.592197+00:00", "epoch": 0, "step": 422, "train_loss": 4.935569763183594, "perplexity": 139.15240386647082, "lr": 0.001305, "grad_norm": 0.485286, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:05.898997+00:00", "epoch": 0, "step": 423, "train_loss": 4.95796537399292, "perplexity": 142.3039657993202, "lr": 0.001305, "grad_norm": 0.528333, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:06.204654+00:00", "epoch": 0, "step": 424, "train_loss": 4.9590559005737305, "perplexity": 142.45923670471592, "lr": 0.001305, "grad_norm": 0.51208, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:06.510865+00:00", "epoch": 0, "step": 425, "train_loss": 5.078585147857666, "perplexity": 160.5467452545614, "lr": 0.001305, "grad_norm": 0.456141, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:06.817884+00:00", "epoch": 0, "step": 426, "train_loss": 5.035870552062988, "perplexity": 153.83345436763102, "lr": 0.001305, "grad_norm": 0.439227, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:07.125462+00:00", "epoch": 0, "step": 427, "train_loss": 4.976996421813965, "perplexity": 145.0380935150133, "lr": 0.001305, "grad_norm": 0.49817, "tokens_per_sec": 106535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:07.432352+00:00", "epoch": 0, "step": 428, "train_loss": 4.9762091636657715, "perplexity": 144.92395602782352, "lr": 0.001305, "grad_norm": 0.551622, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:07.738835+00:00", "epoch": 0, "step": 429, "train_loss": 4.966821670532227, "perplexity": 143.56984916437827, "lr": 0.001305, "grad_norm": 0.488895, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:08.045228+00:00", "epoch": 0, "step": 430, "train_loss": 4.995368480682373, "perplexity": 147.72737003775677, "lr": 0.001305, "grad_norm": 0.488256, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:08.351428+00:00", "epoch": 0, "step": 431, "train_loss": 4.970871448516846, "perplexity": 144.15245409282514, "lr": 0.001305, "grad_norm": 0.495313, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:08.658980+00:00", "epoch": 0, "step": 432, "train_loss": 4.891875743865967, "perplexity": 133.2031949415055, "lr": 0.001305, "grad_norm": 0.593536, "tokens_per_sec": 106544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:08.967169+00:00", "epoch": 0, "step": 433, "train_loss": 4.994375705718994, "perplexity": 147.58078277949716, "lr": 0.001305, "grad_norm": 0.50741, "tokens_per_sec": 106325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:09.273844+00:00", "epoch": 0, "step": 434, "train_loss": 4.953868389129639, "perplexity": 141.72214128241552, "lr": 0.001305, "grad_norm": 0.494659, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:09.579654+00:00", "epoch": 0, "step": 435, "train_loss": 4.9883646965026855, "perplexity": 146.6963342113968, "lr": 0.001305, "grad_norm": 0.48677, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:09.885977+00:00", "epoch": 0, "step": 436, "train_loss": 5.038753986358643, "perplexity": 154.27766314141022, "lr": 0.001305, "grad_norm": 0.488549, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:10.192110+00:00", "epoch": 0, "step": 437, "train_loss": 4.941274166107178, "perplexity": 139.94845358143672, "lr": 0.001305, "grad_norm": 0.538194, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:10.500328+00:00", "epoch": 0, "step": 438, "train_loss": 4.912137508392334, "perplexity": 135.9296548803091, "lr": 0.001305, "grad_norm": 0.512834, "tokens_per_sec": 106314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:10.808711+00:00", "epoch": 0, "step": 439, "train_loss": 5.023162364959717, "perplexity": 151.89087949270015, "lr": 0.001305, "grad_norm": 0.415845, "tokens_per_sec": 106258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:11.114991+00:00", "epoch": 0, "step": 440, "train_loss": 4.995075225830078, "perplexity": 147.68405462121487, "lr": 0.001305, "grad_norm": 0.454784, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:11.422371+00:00", "epoch": 0, "step": 441, "train_loss": 4.853776931762695, "perplexity": 128.2237688519526, "lr": 0.001305, "grad_norm": 0.33669, "tokens_per_sec": 106546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:11.729421+00:00", "epoch": 0, "step": 442, "train_loss": 4.83447790145874, "perplexity": 125.77290021413573, "lr": 0.001305, "grad_norm": 0.337782, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:12.035911+00:00", "epoch": 0, "step": 443, "train_loss": 4.959031581878662, "perplexity": 142.4557723241036, "lr": 0.001305, "grad_norm": 0.336163, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:12.343473+00:00", "epoch": 0, "step": 444, "train_loss": 4.875465393066406, "perplexity": 131.03512185969413, "lr": 0.001305, "grad_norm": 0.293618, "tokens_per_sec": 106542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:12.651763+00:00", "epoch": 0, "step": 445, "train_loss": 5.0319108963012695, "perplexity": 153.22553122118953, "lr": 0.001305, "grad_norm": 0.325383, "tokens_per_sec": 106289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:12.958272+00:00", "epoch": 0, "step": 446, "train_loss": 4.9003987312316895, "perplexity": 134.34333589074896, "lr": 0.001305, "grad_norm": 0.284229, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:13.264985+00:00", "epoch": 0, "step": 447, "train_loss": 4.871699333190918, "perplexity": 130.54256382866802, "lr": 0.001305, "grad_norm": 0.317014, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:13.570828+00:00", "epoch": 0, "step": 448, "train_loss": 4.955869674682617, "perplexity": 142.00605175450576, "lr": 0.001305, "grad_norm": 0.340891, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:13.877398+00:00", "epoch": 0, "step": 449, "train_loss": 4.906075954437256, "perplexity": 135.108202098516, "lr": 0.001305, "grad_norm": 0.366081, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:14.185255+00:00", "epoch": 0, "step": 450, "train_loss": 4.909561634063721, "perplexity": 135.5799677402208, "lr": 0.001305, "grad_norm": 0.36106, "tokens_per_sec": 106500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:14.492873+00:00", "epoch": 0, "step": 451, "train_loss": 4.867860317230225, "perplexity": 130.04236958374835, "lr": 0.001305, "grad_norm": 0.320691, "tokens_per_sec": 106461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:14.800296+00:00", "epoch": 0, "step": 452, "train_loss": 4.898624420166016, "perplexity": 134.10518036674895, "lr": 0.001305, "grad_norm": 0.343883, "tokens_per_sec": 106596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:15.106497+00:00", "epoch": 0, "step": 453, "train_loss": 4.840361595153809, "perplexity": 126.51509070380295, "lr": 0.001305, "grad_norm": 0.517921, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:15.412345+00:00", "epoch": 0, "step": 454, "train_loss": 4.8929243087768555, "perplexity": 133.3429403910228, "lr": 0.001305, "grad_norm": 0.630284, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:15.719559+00:00", "epoch": 0, "step": 455, "train_loss": 4.866737365722656, "perplexity": 129.89642027111393, "lr": 0.001305, "grad_norm": 0.47788, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:16.026957+00:00", "epoch": 0, "step": 456, "train_loss": 4.994022369384766, "perplexity": 147.52864633809924, "lr": 0.001305, "grad_norm": 0.42935, "tokens_per_sec": 106599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:16.335082+00:00", "epoch": 0, "step": 457, "train_loss": 4.784280300140381, "perplexity": 119.61524502294658, "lr": 0.001305, "grad_norm": 0.380782, "tokens_per_sec": 106347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:16.641598+00:00", "epoch": 0, "step": 458, "train_loss": 4.713804721832275, "perplexity": 111.47548729382105, "lr": 0.001305, "grad_norm": 0.357806, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:16.948933+00:00", "epoch": 0, "step": 459, "train_loss": 4.906489849090576, "perplexity": 135.16413423519197, "lr": 0.001305, "grad_norm": 0.339286, "tokens_per_sec": 106620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:17.255878+00:00", "epoch": 0, "step": 460, "train_loss": 4.873895168304443, "perplexity": 130.82952872262322, "lr": 0.001305, "grad_norm": 0.337833, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:17.562948+00:00", "epoch": 0, "step": 461, "train_loss": 4.880515098571777, "perplexity": 131.69848411835284, "lr": 0.001305, "grad_norm": 0.335822, "tokens_per_sec": 106632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:17.870521+00:00", "epoch": 0, "step": 462, "train_loss": 4.873682498931885, "perplexity": 130.801708247224, "lr": 0.001305, "grad_norm": 0.407252, "tokens_per_sec": 106538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:18.177743+00:00", "epoch": 0, "step": 463, "train_loss": 4.780245304107666, "perplexity": 119.13357041491261, "lr": 0.001305, "grad_norm": 0.53039, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:18.484857+00:00", "epoch": 0, "step": 464, "train_loss": 4.790956497192383, "perplexity": 120.41649163320339, "lr": 0.001305, "grad_norm": 0.538647, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:18.791717+00:00", "epoch": 0, "step": 465, "train_loss": 4.867049217224121, "perplexity": 129.93693498176873, "lr": 0.001305, "grad_norm": 0.44773, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:19.099035+00:00", "epoch": 0, "step": 466, "train_loss": 4.923289775848389, "perplexity": 137.45406325210115, "lr": 0.001305, "grad_norm": 0.444648, "tokens_per_sec": 106626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:19.408812+00:00", "epoch": 0, "step": 467, "train_loss": 4.82546854019165, "perplexity": 124.64485582833285, "lr": 0.001305, "grad_norm": 0.392688, "tokens_per_sec": 105781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:19.716534+00:00", "epoch": 0, "step": 468, "train_loss": 4.797218322753906, "perplexity": 121.17288442686531, "lr": 0.001305, "grad_norm": 0.356207, "tokens_per_sec": 106486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:20.024600+00:00", "epoch": 0, "step": 469, "train_loss": 4.854193210601807, "perplexity": 128.27715680496337, "lr": 0.001305, "grad_norm": 0.382308, "tokens_per_sec": 106376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:20.332752+00:00", "epoch": 0, "step": 470, "train_loss": 4.802742004394531, "perplexity": 121.84405682873833, "lr": 0.001305, "grad_norm": 0.360598, "tokens_per_sec": 106398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:20.640911+00:00", "epoch": 0, "step": 471, "train_loss": 4.848554611206055, "perplexity": 127.55588868808515, "lr": 0.001305, "grad_norm": 0.335671, "tokens_per_sec": 106266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:20.948229+00:00", "epoch": 0, "step": 472, "train_loss": 4.825811862945557, "perplexity": 124.68765659031341, "lr": 0.001305, "grad_norm": 0.386639, "tokens_per_sec": 106624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:21.256180+00:00", "epoch": 0, "step": 473, "train_loss": 4.915319919586182, "perplexity": 136.36292799651045, "lr": 0.001305, "grad_norm": 0.37716, "tokens_per_sec": 106406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:21.563900+00:00", "epoch": 0, "step": 474, "train_loss": 4.764525413513184, "perplexity": 117.27544673635144, "lr": 0.001305, "grad_norm": 0.3208, "tokens_per_sec": 106487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:21.871037+00:00", "epoch": 0, "step": 475, "train_loss": 4.8538665771484375, "perplexity": 128.23526403641006, "lr": 0.001305, "grad_norm": 0.313457, "tokens_per_sec": 106688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:22.178147+00:00", "epoch": 0, "step": 476, "train_loss": 4.8154706954956055, "perplexity": 123.40488476527955, "lr": 0.001305, "grad_norm": 0.293366, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:22.485287+00:00", "epoch": 0, "step": 477, "train_loss": 4.759406566619873, "perplexity": 116.67666552230993, "lr": 0.001305, "grad_norm": 0.317555, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:22.792392+00:00", "epoch": 0, "step": 478, "train_loss": 4.846783638000488, "perplexity": 127.33019053868286, "lr": 0.001305, "grad_norm": 0.361, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:23.101416+00:00", "epoch": 0, "step": 479, "train_loss": 4.874781131744385, "perplexity": 130.94549026319376, "lr": 0.001305, "grad_norm": 0.40763, "tokens_per_sec": 106037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:23.409382+00:00", "epoch": 0, "step": 480, "train_loss": 4.8119096755981445, "perplexity": 122.96621902745177, "lr": 0.001305, "grad_norm": 0.404564, "tokens_per_sec": 106457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:23.716208+00:00", "epoch": 0, "step": 481, "train_loss": 4.763080596923828, "perplexity": 117.10612757240558, "lr": 0.001305, "grad_norm": 0.39306, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:24.023195+00:00", "epoch": 0, "step": 482, "train_loss": 4.838909149169922, "perplexity": 126.33146775189958, "lr": 0.001305, "grad_norm": 0.376854, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:24.330721+00:00", "epoch": 0, "step": 483, "train_loss": 4.739465713500977, "perplexity": 114.37307736157759, "lr": 0.001305, "grad_norm": 0.449985, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:24.638862+00:00", "epoch": 0, "step": 484, "train_loss": 4.816110610961914, "perplexity": 123.48387873169368, "lr": 0.001305, "grad_norm": 0.436106, "tokens_per_sec": 106341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:24.947928+00:00", "epoch": 0, "step": 485, "train_loss": 4.712368965148926, "perplexity": 111.31555046057235, "lr": 0.001305, "grad_norm": 0.473212, "tokens_per_sec": 106022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:25.254889+00:00", "epoch": 0, "step": 486, "train_loss": 4.663883209228516, "perplexity": 106.0470866575178, "lr": 0.001305, "grad_norm": 0.569215, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:25.562126+00:00", "epoch": 0, "step": 487, "train_loss": 4.860933303833008, "perplexity": 129.14467710089738, "lr": 0.001305, "grad_norm": 0.591731, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:25.869130+00:00", "epoch": 0, "step": 488, "train_loss": 4.895453929901123, "perplexity": 133.6806744993791, "lr": 0.001305, "grad_norm": 0.599824, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:26.176221+00:00", "epoch": 0, "step": 489, "train_loss": 4.857701778411865, "perplexity": 128.72801638128476, "lr": 0.001305, "grad_norm": 0.674797, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:26.483998+00:00", "epoch": 0, "step": 490, "train_loss": 4.941675662994385, "perplexity": 140.00465373125675, "lr": 0.001305, "grad_norm": 0.64475, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:26.792535+00:00", "epoch": 0, "step": 491, "train_loss": 4.826272487640381, "perplexity": 124.74510403391851, "lr": 0.001305, "grad_norm": 0.465801, "tokens_per_sec": 106144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:27.101686+00:00", "epoch": 0, "step": 492, "train_loss": 4.792550563812256, "perplexity": 120.6085966163897, "lr": 0.001305, "grad_norm": 0.434411, "tokens_per_sec": 105994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:27.409027+00:00", "epoch": 0, "step": 493, "train_loss": 4.686575889587402, "perplexity": 108.48109189957917, "lr": 0.001305, "grad_norm": 0.400645, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:27.716264+00:00", "epoch": 0, "step": 494, "train_loss": 4.776798248291016, "perplexity": 118.72361731958489, "lr": 0.001305, "grad_norm": 0.393868, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:28.023320+00:00", "epoch": 0, "step": 495, "train_loss": 4.788784503936768, "perplexity": 120.15523165567481, "lr": 0.001305, "grad_norm": 0.432604, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:28.331553+00:00", "epoch": 0, "step": 496, "train_loss": 4.780459880828857, "perplexity": 119.15913644867582, "lr": 0.001305, "grad_norm": 0.428081, "tokens_per_sec": 106310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:28.639365+00:00", "epoch": 0, "step": 497, "train_loss": 4.735681056976318, "perplexity": 113.94103263424773, "lr": 0.001305, "grad_norm": 0.39115, "tokens_per_sec": 106454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:28.946786+00:00", "epoch": 0, "step": 498, "train_loss": 4.725703716278076, "perplexity": 112.80985658036647, "lr": 0.001305, "grad_norm": 0.326351, "tokens_per_sec": 106590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:29.253667+00:00", "epoch": 0, "step": 499, "train_loss": 4.806803226470947, "perplexity": 122.33989878284156, "lr": 0.001305, "grad_norm": 0.321897, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:29.561028+00:00", "epoch": 0, "step": 500, "train_loss": 4.743276119232178, "perplexity": 114.80971654897844, "lr": 0.001305, "grad_norm": 0.34309, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:02:35.436276+00:00", "step": 500, "epoch": 0, "val_loss": 4.760609078407287, "val_ppl": 116.81705498098881, "eval_train_loss": 4.743276119232178, "eval_train_ppl": 114.80971654897844} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:02:36.376451+00:00", "step": 500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_4p7606_epoch_0000_step_0000500.pt", "val_loss": 4.760609078407287} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:02:38.040014+00:00", "step": 500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0000_step_0000500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:39.164632+00:00", "epoch": 0, "step": 501, "train_loss": 4.879329204559326, "perplexity": 131.54239624471813, "lr": 0.001305, "grad_norm": 0.332441, "tokens_per_sec": 3412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:39.470714+00:00", "epoch": 0, "step": 502, "train_loss": 4.842297077178955, "perplexity": 126.76019550922713, "lr": 0.001305, "grad_norm": 0.297305, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:39.776801+00:00", "epoch": 0, "step": 503, "train_loss": 4.771894931793213, "perplexity": 118.14290272531429, "lr": 0.001305, "grad_norm": 0.298591, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:40.083390+00:00", "epoch": 0, "step": 504, "train_loss": 4.817488193511963, "perplexity": 123.65410519182696, "lr": 0.001305, "grad_norm": 0.310856, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:40.389807+00:00", "epoch": 0, "step": 505, "train_loss": 4.7398481369018555, "perplexity": 114.41682466725474, "lr": 0.001305, "grad_norm": 0.355654, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:40.695933+00:00", "epoch": 0, "step": 506, "train_loss": 4.789778232574463, "perplexity": 120.274692696438, "lr": 0.001305, "grad_norm": 0.359939, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:41.001903+00:00", "epoch": 0, "step": 507, "train_loss": 4.761443614959717, "perplexity": 116.91458377332833, "lr": 0.001305, "grad_norm": 0.305779, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:41.307522+00:00", "epoch": 0, "step": 508, "train_loss": 4.778563976287842, "perplexity": 118.93343612152265, "lr": 0.001305, "grad_norm": 0.356642, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:41.614115+00:00", "epoch": 0, "step": 509, "train_loss": 4.654097557067871, "perplexity": 105.01440771398286, "lr": 0.001305, "grad_norm": 0.439629, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:41.921457+00:00", "epoch": 0, "step": 510, "train_loss": 4.682651996612549, "perplexity": 108.05625775212852, "lr": 0.001305, "grad_norm": 0.470938, "tokens_per_sec": 106676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:42.229301+00:00", "epoch": 0, "step": 511, "train_loss": 4.716402530670166, "perplexity": 111.76545577819311, "lr": 0.001305, "grad_norm": 0.42895, "tokens_per_sec": 106386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:42.536100+00:00", "epoch": 0, "step": 512, "train_loss": 4.6942057609558105, "perplexity": 109.31195433270912, "lr": 0.001305, "grad_norm": 0.44769, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:42.842648+00:00", "epoch": 0, "step": 513, "train_loss": 4.740741729736328, "perplexity": 114.51911241691411, "lr": 0.001305, "grad_norm": 0.582143, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:43.729961+00:00", "epoch": 0, "step": 514, "train_loss": 4.759441375732422, "perplexity": 116.6807270041798, "lr": 0.001305, "grad_norm": 0.678826, "tokens_per_sec": 36929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:44.036066+00:00", "epoch": 0, "step": 515, "train_loss": 4.657100677490234, "perplexity": 105.3302526492421, "lr": 0.001305, "grad_norm": 0.536716, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:44.342186+00:00", "epoch": 0, "step": 516, "train_loss": 4.621334552764893, "perplexity": 101.62957169297708, "lr": 0.001305, "grad_norm": 0.47628, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:44.649906+00:00", "epoch": 0, "step": 517, "train_loss": 4.710996627807617, "perplexity": 111.16289274690592, "lr": 0.001305, "grad_norm": 0.392422, "tokens_per_sec": 106487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:44.958457+00:00", "epoch": 0, "step": 518, "train_loss": 4.652792930603027, "perplexity": 104.8774924695347, "lr": 0.001305, "grad_norm": 0.349404, "tokens_per_sec": 106201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:45.266404+00:00", "epoch": 0, "step": 519, "train_loss": 4.788848876953125, "perplexity": 120.1629666593277, "lr": 0.001305, "grad_norm": 0.387117, "tokens_per_sec": 106407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:45.572934+00:00", "epoch": 0, "step": 520, "train_loss": 4.827866077423096, "perplexity": 124.94405503819102, "lr": 0.001305, "grad_norm": 0.42721, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:45.879948+00:00", "epoch": 0, "step": 521, "train_loss": 4.732010364532471, "perplexity": 113.5235568281116, "lr": 0.001305, "grad_norm": 0.452573, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:46.187296+00:00", "epoch": 0, "step": 522, "train_loss": 4.73272705078125, "perplexity": 113.60494676224236, "lr": 0.001305, "grad_norm": 0.437246, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:46.495663+00:00", "epoch": 0, "step": 523, "train_loss": 4.62640905380249, "perplexity": 102.14660178536977, "lr": 0.001305, "grad_norm": 0.413089, "tokens_per_sec": 106263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:46.803091+00:00", "epoch": 0, "step": 524, "train_loss": 4.608458042144775, "perplexity": 100.32932670842295, "lr": 0.001305, "grad_norm": 0.35168, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:47.109930+00:00", "epoch": 0, "step": 525, "train_loss": 4.716588020324707, "perplexity": 111.78618903681809, "lr": 0.001305, "grad_norm": 0.319472, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:47.417085+00:00", "epoch": 0, "step": 526, "train_loss": 4.592983245849609, "perplexity": 98.78870198629032, "lr": 0.001305, "grad_norm": 0.320357, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:47.724459+00:00", "epoch": 0, "step": 527, "train_loss": 4.68156099319458, "perplexity": 107.93843229126549, "lr": 0.001305, "grad_norm": 0.284507, "tokens_per_sec": 106605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:48.032259+00:00", "epoch": 0, "step": 528, "train_loss": 4.586392879486084, "perplexity": 98.13978888384649, "lr": 0.001305, "grad_norm": 0.28615, "tokens_per_sec": 106459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:48.340450+00:00", "epoch": 0, "step": 529, "train_loss": 4.566601753234863, "perplexity": 96.2165858828053, "lr": 0.001305, "grad_norm": 0.319482, "tokens_per_sec": 106324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:48.647349+00:00", "epoch": 0, "step": 530, "train_loss": 4.708601951599121, "perplexity": 110.89701208847796, "lr": 0.001305, "grad_norm": 0.357987, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:48.954661+00:00", "epoch": 0, "step": 531, "train_loss": 4.710909843444824, "perplexity": 111.15324596469364, "lr": 0.001305, "grad_norm": 0.425457, "tokens_per_sec": 106576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:49.262823+00:00", "epoch": 0, "step": 532, "train_loss": 4.6045637130737305, "perplexity": 99.93937109531649, "lr": 0.001305, "grad_norm": 0.474656, "tokens_per_sec": 106330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:49.570892+00:00", "epoch": 0, "step": 533, "train_loss": 4.633995056152344, "perplexity": 102.92443273005269, "lr": 0.001305, "grad_norm": 0.44915, "tokens_per_sec": 106366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:49.878300+00:00", "epoch": 0, "step": 534, "train_loss": 4.671853065490723, "perplexity": 106.89564364221133, "lr": 0.001305, "grad_norm": 0.428065, "tokens_per_sec": 106594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:50.185815+00:00", "epoch": 0, "step": 535, "train_loss": 4.705015182495117, "perplexity": 110.49996259987905, "lr": 0.001305, "grad_norm": 0.383459, "tokens_per_sec": 106558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:50.492711+00:00", "epoch": 0, "step": 536, "train_loss": 4.804757595062256, "perplexity": 122.08989224115389, "lr": 0.001305, "grad_norm": 0.354233, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:50.800838+00:00", "epoch": 0, "step": 537, "train_loss": 4.6634602546691895, "perplexity": 106.0022430427869, "lr": 0.001305, "grad_norm": 0.311817, "tokens_per_sec": 106345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:51.109399+00:00", "epoch": 0, "step": 538, "train_loss": 4.500720977783203, "perplexity": 90.08205505377668, "lr": 0.001305, "grad_norm": 0.335701, "tokens_per_sec": 106198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:51.418715+00:00", "epoch": 0, "step": 539, "train_loss": 4.58114767074585, "perplexity": 97.62637286961123, "lr": 0.001305, "grad_norm": 0.352254, "tokens_per_sec": 105937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:51.725606+00:00", "epoch": 0, "step": 540, "train_loss": 4.726729393005371, "perplexity": 112.92562238386674, "lr": 0.001305, "grad_norm": 0.367709, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:52.033397+00:00", "epoch": 0, "step": 541, "train_loss": 4.542352676391602, "perplexity": 93.91148374981698, "lr": 0.001305, "grad_norm": 0.407883, "tokens_per_sec": 106404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:52.340634+00:00", "epoch": 0, "step": 542, "train_loss": 4.608990669250488, "perplexity": 100.38277906114809, "lr": 0.001305, "grad_norm": 0.435287, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:52.649066+00:00", "epoch": 0, "step": 543, "train_loss": 4.687536239624023, "perplexity": 108.58532176072242, "lr": 0.001305, "grad_norm": 0.413221, "tokens_per_sec": 106241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:52.957201+00:00", "epoch": 0, "step": 544, "train_loss": 4.622804641723633, "perplexity": 101.77908607699993, "lr": 0.001305, "grad_norm": 0.483084, "tokens_per_sec": 106343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:53.265852+00:00", "epoch": 0, "step": 545, "train_loss": 4.768008708953857, "perplexity": 117.68466406393229, "lr": 0.001305, "grad_norm": 0.49192, "tokens_per_sec": 106165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:53.572518+00:00", "epoch": 0, "step": 546, "train_loss": 4.573798179626465, "perplexity": 96.91149890693308, "lr": 0.001305, "grad_norm": 0.461777, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:53.878863+00:00", "epoch": 0, "step": 547, "train_loss": 4.723690509796143, "perplexity": 112.58297550173103, "lr": 0.001305, "grad_norm": 0.521043, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:54.185304+00:00", "epoch": 0, "step": 548, "train_loss": 4.532689571380615, "perplexity": 93.0083776524608, "lr": 0.001305, "grad_norm": 0.564101, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:54.493388+00:00", "epoch": 0, "step": 549, "train_loss": 4.512356758117676, "perplexity": 91.13635194678805, "lr": 0.001305, "grad_norm": 0.553411, "tokens_per_sec": 106360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:54.802415+00:00", "epoch": 0, "step": 550, "train_loss": 4.625749111175537, "perplexity": 102.07921312742295, "lr": 0.001305, "grad_norm": 0.523094, "tokens_per_sec": 106092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:55.110695+00:00", "epoch": 0, "step": 551, "train_loss": 4.629052639007568, "perplexity": 102.41699227327568, "lr": 0.001305, "grad_norm": 0.463198, "tokens_per_sec": 106237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:55.418266+00:00", "epoch": 0, "step": 552, "train_loss": 4.697062969207764, "perplexity": 109.62472796759006, "lr": 0.001305, "grad_norm": 0.485206, "tokens_per_sec": 106539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:55.725267+00:00", "epoch": 0, "step": 553, "train_loss": 4.581087112426758, "perplexity": 97.6204609595807, "lr": 0.001305, "grad_norm": 0.489964, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:56.031458+00:00", "epoch": 0, "step": 554, "train_loss": 4.600635528564453, "perplexity": 99.54756086118232, "lr": 0.001305, "grad_norm": 0.392467, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:56.338873+00:00", "epoch": 0, "step": 555, "train_loss": 4.594694137573242, "perplexity": 98.95786342609735, "lr": 0.001305, "grad_norm": 0.382386, "tokens_per_sec": 106590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:56.646635+00:00", "epoch": 0, "step": 556, "train_loss": 4.559581756591797, "perplexity": 95.54351102862074, "lr": 0.001305, "grad_norm": 0.372319, "tokens_per_sec": 106475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:56.954268+00:00", "epoch": 0, "step": 557, "train_loss": 4.541298866271973, "perplexity": 93.81257100467164, "lr": 0.001305, "grad_norm": 0.334109, "tokens_per_sec": 106514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:57.260183+00:00", "epoch": 0, "step": 558, "train_loss": 4.483412742614746, "perplexity": 88.53630931453698, "lr": 0.001305, "grad_norm": 0.352534, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:57.568157+00:00", "epoch": 0, "step": 559, "train_loss": 4.552435398101807, "perplexity": 94.86315677062858, "lr": 0.001305, "grad_norm": 0.372825, "tokens_per_sec": 106397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:57.875205+00:00", "epoch": 0, "step": 560, "train_loss": 4.537672519683838, "perplexity": 93.47299019886786, "lr": 0.001305, "grad_norm": 0.323864, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:58.183852+00:00", "epoch": 0, "step": 561, "train_loss": 4.660463809967041, "perplexity": 105.68508858840015, "lr": 0.001305, "grad_norm": 0.306822, "tokens_per_sec": 106111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:58.491920+00:00", "epoch": 0, "step": 562, "train_loss": 4.475126266479492, "perplexity": 87.80568662476091, "lr": 0.001305, "grad_norm": 0.323159, "tokens_per_sec": 106366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:58.798884+00:00", "epoch": 0, "step": 563, "train_loss": 4.497251987457275, "perplexity": 89.77010266899136, "lr": 0.001305, "grad_norm": 0.330266, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:59.105799+00:00", "epoch": 0, "step": 564, "train_loss": 4.560925483703613, "perplexity": 95.67198173019985, "lr": 0.001305, "grad_norm": 0.275772, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:59.413568+00:00", "epoch": 0, "step": 565, "train_loss": 4.538995265960693, "perplexity": 93.59671305757594, "lr": 0.001305, "grad_norm": 0.260993, "tokens_per_sec": 106470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:02:59.721188+00:00", "epoch": 0, "step": 566, "train_loss": 4.611878871917725, "perplexity": 100.67312395699723, "lr": 0.001305, "grad_norm": 0.280867, "tokens_per_sec": 106521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:00.029877+00:00", "epoch": 0, "step": 567, "train_loss": 4.607213497161865, "perplexity": 100.2045400156622, "lr": 0.001305, "grad_norm": 0.25536, "tokens_per_sec": 106152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:00.338652+00:00", "epoch": 0, "step": 568, "train_loss": 4.552511692047119, "perplexity": 94.87039453121855, "lr": 0.001305, "grad_norm": 0.257174, "tokens_per_sec": 106125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:00.647794+00:00", "epoch": 0, "step": 569, "train_loss": 4.592157363891602, "perplexity": 98.70714786134059, "lr": 0.001305, "grad_norm": 0.279496, "tokens_per_sec": 105995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:00.956101+00:00", "epoch": 0, "step": 570, "train_loss": 4.58549690246582, "perplexity": 98.0518972685466, "lr": 0.001305, "grad_norm": 0.307116, "tokens_per_sec": 106343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:01.263249+00:00", "epoch": 0, "step": 571, "train_loss": 4.638215065002441, "perplexity": 103.35969250116536, "lr": 0.001305, "grad_norm": 0.334051, "tokens_per_sec": 106625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:01.570487+00:00", "epoch": 0, "step": 572, "train_loss": 4.538227558135986, "perplexity": 93.52488570333348, "lr": 0.001305, "grad_norm": 0.375148, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:01.879318+00:00", "epoch": 0, "step": 573, "train_loss": 4.505863666534424, "perplexity": 90.54651228083294, "lr": 0.001305, "grad_norm": 0.3543, "tokens_per_sec": 106103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:02.187788+00:00", "epoch": 0, "step": 574, "train_loss": 4.470066547393799, "perplexity": 87.3625365691247, "lr": 0.001305, "grad_norm": 0.375825, "tokens_per_sec": 106229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:02.495828+00:00", "epoch": 0, "step": 575, "train_loss": 4.584909915924072, "perplexity": 97.99435901319994, "lr": 0.001305, "grad_norm": 0.375599, "tokens_per_sec": 106375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:02.803236+00:00", "epoch": 0, "step": 576, "train_loss": 4.531612396240234, "perplexity": 92.90824527993963, "lr": 0.001305, "grad_norm": 0.388535, "tokens_per_sec": 106600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:03.110756+00:00", "epoch": 0, "step": 577, "train_loss": 4.576507091522217, "perplexity": 97.17437951860934, "lr": 0.001305, "grad_norm": 0.309992, "tokens_per_sec": 106550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:03.418993+00:00", "epoch": 0, "step": 578, "train_loss": 4.452530384063721, "perplexity": 85.84388741605287, "lr": 0.001305, "grad_norm": 0.311425, "tokens_per_sec": 106308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:03.727866+00:00", "epoch": 0, "step": 579, "train_loss": 4.432133674621582, "perplexity": 84.11069043215278, "lr": 0.001305, "grad_norm": 0.322622, "tokens_per_sec": 106089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:04.036448+00:00", "epoch": 0, "step": 580, "train_loss": 4.397405624389648, "perplexity": 81.23982839509846, "lr": 0.001305, "grad_norm": 0.345669, "tokens_per_sec": 106253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:04.343229+00:00", "epoch": 0, "step": 581, "train_loss": 4.580271244049072, "perplexity": 97.54084799373547, "lr": 0.001305, "grad_norm": 0.37401, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:04.649552+00:00", "epoch": 0, "step": 582, "train_loss": 4.615288257598877, "perplexity": 101.01694323751539, "lr": 0.001305, "grad_norm": 0.326169, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:04.957545+00:00", "epoch": 0, "step": 583, "train_loss": 4.497165679931641, "perplexity": 89.76235516789265, "lr": 0.001305, "grad_norm": 0.29975, "tokens_per_sec": 106392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:05.265817+00:00", "epoch": 0, "step": 584, "train_loss": 4.400746822357178, "perplexity": 81.51172071470596, "lr": 0.001305, "grad_norm": 0.323021, "tokens_per_sec": 106296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:05.574825+00:00", "epoch": 0, "step": 585, "train_loss": 4.511096954345703, "perplexity": 91.02161031798563, "lr": 0.001305, "grad_norm": 0.348179, "tokens_per_sec": 106043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:05.882805+00:00", "epoch": 0, "step": 586, "train_loss": 4.426255226135254, "perplexity": 83.61770049891359, "lr": 0.001305, "grad_norm": 0.364806, "tokens_per_sec": 106396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:06.190958+00:00", "epoch": 0, "step": 587, "train_loss": 4.476345062255859, "perplexity": 87.91276906731896, "lr": 0.001305, "grad_norm": 0.368658, "tokens_per_sec": 106337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:06.498789+00:00", "epoch": 0, "step": 588, "train_loss": 4.501930236816406, "perplexity": 90.19105348296509, "lr": 0.001305, "grad_norm": 0.506649, "tokens_per_sec": 106451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:06.806457+00:00", "epoch": 0, "step": 589, "train_loss": 4.506706714630127, "perplexity": 90.62287953169351, "lr": 0.001305, "grad_norm": 0.556211, "tokens_per_sec": 106501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:07.114680+00:00", "epoch": 0, "step": 590, "train_loss": 4.516800880432129, "perplexity": 91.5422743584685, "lr": 0.001305, "grad_norm": 0.535885, "tokens_per_sec": 106379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:07.423288+00:00", "epoch": 0, "step": 591, "train_loss": 4.647029876708984, "perplexity": 104.27481612433571, "lr": 0.001305, "grad_norm": 0.555716, "tokens_per_sec": 106115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:07.731215+00:00", "epoch": 0, "step": 592, "train_loss": 4.5775909423828125, "perplexity": 97.27975915106504, "lr": 0.001305, "grad_norm": 0.52318, "tokens_per_sec": 106414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:08.038564+00:00", "epoch": 0, "step": 593, "train_loss": 4.5483503341674805, "perplexity": 94.47642515969306, "lr": 0.001305, "grad_norm": 0.499108, "tokens_per_sec": 106615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:08.346427+00:00", "epoch": 0, "step": 594, "train_loss": 4.513164520263672, "perplexity": 91.20999818233109, "lr": 0.001305, "grad_norm": 0.45312, "tokens_per_sec": 106437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:08.654038+00:00", "epoch": 0, "step": 595, "train_loss": 4.494059085845947, "perplexity": 89.48393266298328, "lr": 0.001305, "grad_norm": 0.4275, "tokens_per_sec": 106524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:08.962712+00:00", "epoch": 0, "step": 596, "train_loss": 4.5149335861206055, "perplexity": 91.37149748525886, "lr": 0.001305, "grad_norm": 0.438228, "tokens_per_sec": 106158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:09.271297+00:00", "epoch": 0, "step": 597, "train_loss": 4.4745941162109375, "perplexity": 87.75897323541662, "lr": 0.001305, "grad_norm": 0.376527, "tokens_per_sec": 106188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:09.579285+00:00", "epoch": 0, "step": 598, "train_loss": 4.4932732582092285, "perplexity": 89.41364133770466, "lr": 0.001305, "grad_norm": 0.350414, "tokens_per_sec": 106394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:09.885989+00:00", "epoch": 0, "step": 599, "train_loss": 4.580368995666504, "perplexity": 97.55038323542756, "lr": 0.001305, "grad_norm": 0.3514, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:10.193761+00:00", "epoch": 0, "step": 600, "train_loss": 4.539957523345947, "perplexity": 93.68682053228974, "lr": 0.001305, "grad_norm": 0.334339, "tokens_per_sec": 106525} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:03:16.151267+00:00", "step": 600, "epoch": 0, "val_loss": 4.5067647933959964, "val_ppl": 90.62814294954119, "eval_train_loss": 4.539957523345947, "eval_train_ppl": 93.68682053228974} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:03:17.133131+00:00", "step": 600, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_4p5068_epoch_0000_step_0000600.pt", "val_loss": 4.5067647933959964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:18.119674+00:00", "epoch": 0, "step": 601, "train_loss": 4.477025985717773, "perplexity": 87.97265131968028, "lr": 0.001305, "grad_norm": 0.349656, "tokens_per_sec": 4134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:18.425438+00:00", "epoch": 0, "step": 602, "train_loss": 4.407693862915039, "perplexity": 82.07995744092197, "lr": 0.001305, "grad_norm": 0.348679, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:18.732204+00:00", "epoch": 0, "step": 603, "train_loss": 4.530516147613525, "perplexity": 92.80645055000242, "lr": 0.001305, "grad_norm": 0.346711, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:19.038416+00:00", "epoch": 0, "step": 604, "train_loss": 4.424951076507568, "perplexity": 83.5087215837918, "lr": 0.001305, "grad_norm": 0.353064, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:19.344497+00:00", "epoch": 0, "step": 605, "train_loss": 4.552748680114746, "perplexity": 94.8928803470233, "lr": 0.001305, "grad_norm": 0.381567, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:19.652216+00:00", "epoch": 0, "step": 606, "train_loss": 4.4845428466796875, "perplexity": 88.63642111531006, "lr": 0.001305, "grad_norm": 0.350293, "tokens_per_sec": 106486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:19.958992+00:00", "epoch": 0, "step": 607, "train_loss": 4.521514892578125, "perplexity": 91.9748244739082, "lr": 0.001305, "grad_norm": 0.306382, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:20.265064+00:00", "epoch": 0, "step": 608, "train_loss": 4.372864723205566, "perplexity": 79.27039447676144, "lr": 0.001305, "grad_norm": 0.286305, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:20.571836+00:00", "epoch": 0, "step": 609, "train_loss": 4.426074504852295, "perplexity": 83.60259036620376, "lr": 0.001305, "grad_norm": 0.278057, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:20.878816+00:00", "epoch": 0, "step": 610, "train_loss": 4.475263595581055, "perplexity": 87.81774572883117, "lr": 0.001305, "grad_norm": 0.28554, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:21.186799+00:00", "epoch": 0, "step": 611, "train_loss": 4.516977310180664, "perplexity": 91.55842656373677, "lr": 0.001305, "grad_norm": 0.2754, "tokens_per_sec": 106330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:21.494892+00:00", "epoch": 0, "step": 612, "train_loss": 4.517736434936523, "perplexity": 91.62795721983096, "lr": 0.001305, "grad_norm": 0.303258, "tokens_per_sec": 106361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:21.801629+00:00", "epoch": 0, "step": 613, "train_loss": 4.555908679962158, "perplexity": 95.19321611501155, "lr": 0.001305, "grad_norm": 0.399728, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:22.107865+00:00", "epoch": 0, "step": 614, "train_loss": 4.426319599151611, "perplexity": 83.62308339577037, "lr": 0.001305, "grad_norm": 0.472581, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:22.469009+00:00", "epoch": 0, "step": 615, "train_loss": 4.477321624755859, "perplexity": 87.99866331458541, "lr": 0.001305, "grad_norm": 0.420356, "tokens_per_sec": 90733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:22.774474+00:00", "epoch": 0, "step": 616, "train_loss": 4.368898391723633, "perplexity": 78.95660452441922, "lr": 0.001305, "grad_norm": 0.337521, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:23.080600+00:00", "epoch": 0, "step": 617, "train_loss": 4.580416679382324, "perplexity": 97.55503491108364, "lr": 0.001305, "grad_norm": 0.348357, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:23.387177+00:00", "epoch": 0, "step": 618, "train_loss": 4.4076666831970215, "perplexity": 82.07772656114132, "lr": 0.001305, "grad_norm": 0.307587, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:23.695506+00:00", "epoch": 0, "step": 619, "train_loss": 4.417884349822998, "perplexity": 82.92066852491426, "lr": 0.001305, "grad_norm": 0.336788, "tokens_per_sec": 106277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:24.002393+00:00", "epoch": 0, "step": 620, "train_loss": 4.4946818351745605, "perplexity": 89.53967607725622, "lr": 0.001305, "grad_norm": 0.30247, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:24.309451+00:00", "epoch": 0, "step": 621, "train_loss": 4.458831310272217, "perplexity": 86.38649107372746, "lr": 0.001305, "grad_norm": 0.314908, "tokens_per_sec": 106647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:24.617090+00:00", "epoch": 0, "step": 622, "train_loss": 4.432209014892578, "perplexity": 84.11702759308233, "lr": 0.001305, "grad_norm": 0.37316, "tokens_per_sec": 106515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:24.923984+00:00", "epoch": 0, "step": 623, "train_loss": 4.377243518829346, "perplexity": 79.61826440301323, "lr": 0.001305, "grad_norm": 0.412697, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:25.231125+00:00", "epoch": 0, "step": 624, "train_loss": 4.425055027008057, "perplexity": 83.51740280839657, "lr": 0.001305, "grad_norm": 0.519105, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:25.538225+00:00", "epoch": 0, "step": 625, "train_loss": 4.510697841644287, "perplexity": 90.98528968569936, "lr": 0.001305, "grad_norm": 0.555188, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:25.846230+00:00", "epoch": 0, "step": 626, "train_loss": 4.428751468658447, "perplexity": 83.82669129592544, "lr": 0.001305, "grad_norm": 0.516229, "tokens_per_sec": 106387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:26.153009+00:00", "epoch": 0, "step": 627, "train_loss": 4.511730194091797, "perplexity": 91.07926707274015, "lr": 0.001305, "grad_norm": 0.439472, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:26.459548+00:00", "epoch": 0, "step": 628, "train_loss": 4.414236545562744, "perplexity": 82.61874117773665, "lr": 0.001305, "grad_norm": 0.457289, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:26.765795+00:00", "epoch": 0, "step": 629, "train_loss": 4.521154403686523, "perplexity": 91.94167454682758, "lr": 0.001305, "grad_norm": 0.426598, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:27.073524+00:00", "epoch": 0, "step": 630, "train_loss": 4.439376354217529, "perplexity": 84.72208862149854, "lr": 0.001305, "grad_norm": 0.439814, "tokens_per_sec": 106540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:27.380310+00:00", "epoch": 0, "step": 631, "train_loss": 4.375247001647949, "perplexity": 79.45946374704802, "lr": 0.001305, "grad_norm": 0.342246, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:27.686644+00:00", "epoch": 0, "step": 632, "train_loss": 4.470848083496094, "perplexity": 87.43084023285445, "lr": 0.001305, "grad_norm": 0.339328, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:27.992821+00:00", "epoch": 0, "step": 633, "train_loss": 4.489034175872803, "perplexity": 89.03541178956567, "lr": 0.001305, "grad_norm": 0.366226, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:28.299109+00:00", "epoch": 0, "step": 634, "train_loss": 4.375931739807129, "perplexity": 79.5138913061703, "lr": 0.001305, "grad_norm": 0.444647, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:28.604948+00:00", "epoch": 0, "step": 635, "train_loss": 4.456464767456055, "perplexity": 86.18229545799744, "lr": 0.001305, "grad_norm": 0.413054, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:28.911532+00:00", "epoch": 0, "step": 636, "train_loss": 4.471707820892334, "perplexity": 87.50604011723175, "lr": 0.001305, "grad_norm": 0.309885, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:29.218441+00:00", "epoch": 0, "step": 637, "train_loss": 4.479085445404053, "perplexity": 88.15401413917364, "lr": 0.001305, "grad_norm": 0.285256, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:29.525054+00:00", "epoch": 0, "step": 638, "train_loss": 4.37959098815918, "perplexity": 79.80538538124334, "lr": 0.001305, "grad_norm": 0.273799, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:29.831366+00:00", "epoch": 0, "step": 639, "train_loss": 4.422491073608398, "perplexity": 83.30354236080352, "lr": 0.001305, "grad_norm": 0.279834, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:30.138255+00:00", "epoch": 0, "step": 640, "train_loss": 4.408660888671875, "perplexity": 82.15936926433824, "lr": 0.001305, "grad_norm": 0.278376, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:30.466548+00:00", "epoch": 0, "step": 641, "train_loss": 4.4477338790893555, "perplexity": 85.43312268809329, "lr": 0.001305, "grad_norm": 0.301417, "tokens_per_sec": 99764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:30.773966+00:00", "epoch": 0, "step": 642, "train_loss": 4.430140018463135, "perplexity": 83.94316968105657, "lr": 0.001305, "grad_norm": 0.280391, "tokens_per_sec": 106591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:31.081167+00:00", "epoch": 0, "step": 643, "train_loss": 4.359439373016357, "perplexity": 78.21327365777721, "lr": 0.001305, "grad_norm": 0.274685, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:31.387104+00:00", "epoch": 0, "step": 644, "train_loss": 4.4255266189575195, "perplexity": 83.55679823175319, "lr": 0.001305, "grad_norm": 0.283123, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:31.693926+00:00", "epoch": 0, "step": 645, "train_loss": 4.352344512939453, "perplexity": 77.66032529014772, "lr": 0.001305, "grad_norm": 0.298208, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:32.000483+00:00", "epoch": 0, "step": 646, "train_loss": 4.360368251800537, "perplexity": 78.2859580605929, "lr": 0.001305, "grad_norm": 0.304259, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:32.307531+00:00", "epoch": 0, "step": 647, "train_loss": 4.228517055511475, "perplexity": 68.61540385466098, "lr": 0.001305, "grad_norm": 0.279988, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:32.614790+00:00", "epoch": 0, "step": 648, "train_loss": 4.3933424949646, "perplexity": 80.91041014554946, "lr": 0.001305, "grad_norm": 0.280929, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:32.922024+00:00", "epoch": 0, "step": 649, "train_loss": 4.474569797515869, "perplexity": 87.75683907765708, "lr": 0.001305, "grad_norm": 0.272238, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:33.229438+00:00", "epoch": 0, "step": 650, "train_loss": 4.470474720001221, "perplexity": 87.39820284196972, "lr": 0.001305, "grad_norm": 0.288173, "tokens_per_sec": 106651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:33.536687+00:00", "epoch": 0, "step": 651, "train_loss": 4.4669508934021, "perplexity": 87.09076872047366, "lr": 0.001305, "grad_norm": 0.316349, "tokens_per_sec": 106592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:33.844005+00:00", "epoch": 0, "step": 652, "train_loss": 4.535062789916992, "perplexity": 93.22936898498567, "lr": 0.001305, "grad_norm": 0.352605, "tokens_per_sec": 106627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:34.152812+00:00", "epoch": 0, "step": 653, "train_loss": 4.425300121307373, "perplexity": 83.53787495641953, "lr": 0.001305, "grad_norm": 0.401005, "tokens_per_sec": 106111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:34.461762+00:00", "epoch": 0, "step": 654, "train_loss": 4.385068893432617, "perplexity": 80.24375128994721, "lr": 0.001305, "grad_norm": 0.411675, "tokens_per_sec": 106064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:34.769418+00:00", "epoch": 0, "step": 655, "train_loss": 4.39603853225708, "perplexity": 81.12884194648427, "lr": 0.001305, "grad_norm": 0.367338, "tokens_per_sec": 106507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:35.076320+00:00", "epoch": 0, "step": 656, "train_loss": 4.33121395111084, "perplexity": 76.03653520156608, "lr": 0.001305, "grad_norm": 0.397326, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:35.383763+00:00", "epoch": 0, "step": 657, "train_loss": 4.393113613128662, "perplexity": 80.89189334148979, "lr": 0.001305, "grad_norm": 0.441763, "tokens_per_sec": 106582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:35.691885+00:00", "epoch": 0, "step": 658, "train_loss": 4.351939678192139, "perplexity": 77.62889205504473, "lr": 0.001305, "grad_norm": 0.466119, "tokens_per_sec": 106348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:36.000507+00:00", "epoch": 0, "step": 659, "train_loss": 4.477776050567627, "perplexity": 88.03866126595906, "lr": 0.001305, "grad_norm": 0.453271, "tokens_per_sec": 106175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:36.308824+00:00", "epoch": 0, "step": 660, "train_loss": 4.30007791519165, "perplexity": 73.70553625685982, "lr": 0.001305, "grad_norm": 0.411576, "tokens_per_sec": 106338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:36.617478+00:00", "epoch": 0, "step": 661, "train_loss": 4.378195285797119, "perplexity": 79.69407851006177, "lr": 0.001305, "grad_norm": 0.357698, "tokens_per_sec": 106107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:36.925542+00:00", "epoch": 0, "step": 662, "train_loss": 4.443151950836182, "perplexity": 85.04256967570717, "lr": 0.001305, "grad_norm": 0.374816, "tokens_per_sec": 106367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:37.234094+00:00", "epoch": 0, "step": 663, "train_loss": 4.359729766845703, "perplexity": 78.23598960794659, "lr": 0.001305, "grad_norm": 0.298879, "tokens_per_sec": 106199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:37.541225+00:00", "epoch": 0, "step": 664, "train_loss": 4.402142524719238, "perplexity": 81.6255662445998, "lr": 0.001305, "grad_norm": 0.315363, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:37.848530+00:00", "epoch": 0, "step": 665, "train_loss": 4.5774335861206055, "perplexity": 97.26445277608518, "lr": 0.001305, "grad_norm": 0.282773, "tokens_per_sec": 106630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:38.155546+00:00", "epoch": 0, "step": 666, "train_loss": 4.3916521072387695, "perplexity": 80.77375571337478, "lr": 0.001305, "grad_norm": 0.27618, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:38.463298+00:00", "epoch": 0, "step": 667, "train_loss": 4.40626335144043, "perplexity": 81.96262506263275, "lr": 0.001305, "grad_norm": 0.290388, "tokens_per_sec": 106475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:38.770069+00:00", "epoch": 0, "step": 668, "train_loss": 4.356802940368652, "perplexity": 78.00734121241557, "lr": 0.001305, "grad_norm": 0.319254, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:39.077400+00:00", "epoch": 0, "step": 669, "train_loss": 4.244930267333984, "perplexity": 69.75089605837319, "lr": 0.001305, "grad_norm": 0.340679, "tokens_per_sec": 106621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:39.386166+00:00", "epoch": 0, "step": 670, "train_loss": 4.409328937530518, "perplexity": 82.21427407470807, "lr": 0.001305, "grad_norm": 0.322659, "tokens_per_sec": 106180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:39.693610+00:00", "epoch": 0, "step": 671, "train_loss": 4.333000183105469, "perplexity": 76.17247546778576, "lr": 0.001305, "grad_norm": 0.326602, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:40.001888+00:00", "epoch": 0, "step": 672, "train_loss": 4.391048908233643, "perplexity": 80.72504775606066, "lr": 0.001305, "grad_norm": 0.320675, "tokens_per_sec": 106294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:40.310022+00:00", "epoch": 0, "step": 673, "train_loss": 4.208573341369629, "perplexity": 67.26051353592824, "lr": 0.001305, "grad_norm": 0.300009, "tokens_per_sec": 106343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:40.616646+00:00", "epoch": 0, "step": 674, "train_loss": 4.344246864318848, "perplexity": 77.03399857347254, "lr": 0.001305, "grad_norm": 0.321867, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:40.923721+00:00", "epoch": 0, "step": 675, "train_loss": 4.3579325675964355, "perplexity": 78.09551021869542, "lr": 0.001305, "grad_norm": 0.341413, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:41.231384+00:00", "epoch": 0, "step": 676, "train_loss": 4.3596110343933105, "perplexity": 78.22670100847509, "lr": 0.001305, "grad_norm": 0.373741, "tokens_per_sec": 106506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:41.539681+00:00", "epoch": 0, "step": 677, "train_loss": 4.415673732757568, "perplexity": 82.73756514029586, "lr": 0.001305, "grad_norm": 0.380647, "tokens_per_sec": 106288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:41.847859+00:00", "epoch": 0, "step": 678, "train_loss": 4.478085517883301, "perplexity": 88.06591057030363, "lr": 0.001305, "grad_norm": 0.371103, "tokens_per_sec": 106328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:42.155555+00:00", "epoch": 0, "step": 679, "train_loss": 4.312809467315674, "perplexity": 74.6499211196616, "lr": 0.001305, "grad_norm": 0.352714, "tokens_per_sec": 106494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:42.462208+00:00", "epoch": 0, "step": 680, "train_loss": 4.418978214263916, "perplexity": 83.01142212264836, "lr": 0.001305, "grad_norm": 0.347577, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:42.768926+00:00", "epoch": 0, "step": 681, "train_loss": 4.385111331939697, "perplexity": 80.24715678721607, "lr": 0.001305, "grad_norm": 0.364052, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:43.076001+00:00", "epoch": 0, "step": 682, "train_loss": 4.378910541534424, "perplexity": 79.75110054718391, "lr": 0.001305, "grad_norm": 0.369769, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:43.384125+00:00", "epoch": 0, "step": 683, "train_loss": 4.438323497772217, "perplexity": 84.63293536545876, "lr": 0.001305, "grad_norm": 0.312706, "tokens_per_sec": 106347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:43.691980+00:00", "epoch": 0, "step": 684, "train_loss": 4.3348069190979, "perplexity": 76.31022342046718, "lr": 0.001305, "grad_norm": 0.33691, "tokens_per_sec": 106441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:44.000441+00:00", "epoch": 0, "step": 685, "train_loss": 4.343926429748535, "perplexity": 77.00931817167834, "lr": 0.001305, "grad_norm": 0.305899, "tokens_per_sec": 106230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:44.307588+00:00", "epoch": 0, "step": 686, "train_loss": 4.301133155822754, "perplexity": 73.78335438465939, "lr": 0.001305, "grad_norm": 0.309956, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:44.614865+00:00", "epoch": 0, "step": 687, "train_loss": 4.304057598114014, "perplexity": 73.99944536540697, "lr": 0.001305, "grad_norm": 0.349527, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:44.923331+00:00", "epoch": 0, "step": 688, "train_loss": 4.36072301864624, "perplexity": 78.3137362500959, "lr": 0.001305, "grad_norm": 0.354529, "tokens_per_sec": 106230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:45.232114+00:00", "epoch": 0, "step": 689, "train_loss": 4.362985134124756, "perplexity": 78.49109148843287, "lr": 0.001305, "grad_norm": 0.369968, "tokens_per_sec": 106120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:45.539732+00:00", "epoch": 0, "step": 690, "train_loss": 4.244160175323486, "perplexity": 69.69720212788056, "lr": 0.001305, "grad_norm": 0.378293, "tokens_per_sec": 106577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:45.847222+00:00", "epoch": 0, "step": 691, "train_loss": 4.251186847686768, "perplexity": 70.18866618820336, "lr": 0.001305, "grad_norm": 0.393113, "tokens_per_sec": 106510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:46.153685+00:00", "epoch": 0, "step": 692, "train_loss": 4.344082832336426, "perplexity": 77.0213635702733, "lr": 0.001305, "grad_norm": 0.451306, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:46.460831+00:00", "epoch": 0, "step": 693, "train_loss": 4.36367130279541, "perplexity": 78.54496809844474, "lr": 0.001305, "grad_norm": 0.429785, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:46.769338+00:00", "epoch": 0, "step": 694, "train_loss": 4.412436008453369, "perplexity": 82.4701169102233, "lr": 0.001305, "grad_norm": 0.35474, "tokens_per_sec": 106215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:47.078272+00:00", "epoch": 0, "step": 695, "train_loss": 4.379289150238037, "perplexity": 79.78130072463807, "lr": 0.001305, "grad_norm": 0.394549, "tokens_per_sec": 106068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:47.384674+00:00", "epoch": 0, "step": 696, "train_loss": 4.369477272033691, "perplexity": 79.00232417995494, "lr": 0.001305, "grad_norm": 0.342508, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:47.692689+00:00", "epoch": 0, "step": 697, "train_loss": 4.3146185874938965, "perplexity": 74.78509403341735, "lr": 0.001305, "grad_norm": 0.328094, "tokens_per_sec": 106384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:47.999821+00:00", "epoch": 0, "step": 698, "train_loss": 4.361987590789795, "perplexity": 78.41283226323861, "lr": 0.001305, "grad_norm": 0.373243, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:48.306807+00:00", "epoch": 0, "step": 699, "train_loss": 4.2992119789123535, "perplexity": 73.6417395849428, "lr": 0.001305, "grad_norm": 0.456942, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:48.616785+00:00", "epoch": 0, "step": 700, "train_loss": 4.361639499664307, "perplexity": 78.38554220219297, "lr": 0.001305, "grad_norm": 0.536771, "tokens_per_sec": 105767} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:03:54.507494+00:00", "step": 700, "epoch": 0, "val_loss": 4.352842521667481, "val_ppl": 77.69901044193577, "eval_train_loss": 4.361639499664307, "eval_train_ppl": 78.38554220219297} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:03:55.484836+00:00", "step": 700, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_4p3528_epoch_0000_step_0000700.pt", "val_loss": 4.352842521667481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:56.751146+00:00", "epoch": 0, "step": 701, "train_loss": 4.358786582946777, "perplexity": 78.16223347050025, "lr": 0.001305, "grad_norm": 0.446155, "tokens_per_sec": 4028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:57.058468+00:00", "epoch": 0, "step": 702, "train_loss": 4.234419345855713, "perplexity": 69.02158942368845, "lr": 0.001305, "grad_norm": 0.410171, "tokens_per_sec": 106624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:57.364872+00:00", "epoch": 0, "step": 703, "train_loss": 4.379135608673096, "perplexity": 79.7690519192462, "lr": 0.001305, "grad_norm": 0.355491, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:57.671037+00:00", "epoch": 0, "step": 704, "train_loss": 4.358318328857422, "perplexity": 78.12564225270681, "lr": 0.001305, "grad_norm": 0.363521, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:57.977316+00:00", "epoch": 0, "step": 705, "train_loss": 4.330996990203857, "perplexity": 76.02004003539275, "lr": 0.001305, "grad_norm": 0.321928, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:58.284112+00:00", "epoch": 0, "step": 706, "train_loss": 4.456718921661377, "perplexity": 86.20420183449335, "lr": 0.001305, "grad_norm": 0.304976, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:58.590789+00:00", "epoch": 0, "step": 707, "train_loss": 4.533015727996826, "perplexity": 93.0387178977621, "lr": 0.001305, "grad_norm": 0.318566, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:58.897890+00:00", "epoch": 0, "step": 708, "train_loss": 4.266273021697998, "perplexity": 71.25557214950852, "lr": 0.001305, "grad_norm": 0.29009, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:59.204700+00:00", "epoch": 0, "step": 709, "train_loss": 4.346820831298828, "perplexity": 77.23253694813206, "lr": 0.001305, "grad_norm": 0.284498, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:59.510472+00:00", "epoch": 0, "step": 710, "train_loss": 4.272371768951416, "perplexity": 71.69146973796643, "lr": 0.001305, "grad_norm": 0.260157, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:03:59.816578+00:00", "epoch": 0, "step": 711, "train_loss": 4.290309429168701, "perplexity": 72.98904994682275, "lr": 0.001305, "grad_norm": 0.279338, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:00.123296+00:00", "epoch": 0, "step": 712, "train_loss": 4.359388828277588, "perplexity": 78.20932048819866, "lr": 0.001305, "grad_norm": 0.316125, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:00.430141+00:00", "epoch": 0, "step": 713, "train_loss": 4.275683879852295, "perplexity": 71.92931350141603, "lr": 0.001305, "grad_norm": 0.335721, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:00.736726+00:00", "epoch": 0, "step": 714, "train_loss": 4.236496925354004, "perplexity": 69.16513632623239, "lr": 0.001305, "grad_norm": 0.296188, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:01.042344+00:00", "epoch": 0, "step": 715, "train_loss": 4.27640962600708, "perplexity": 71.98153487154319, "lr": 0.001305, "grad_norm": 0.326037, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:01.347834+00:00", "epoch": 0, "step": 716, "train_loss": 4.210241317749023, "perplexity": 67.37279610007232, "lr": 0.001305, "grad_norm": 0.337058, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:01.655098+00:00", "epoch": 0, "step": 717, "train_loss": 4.317867279052734, "perplexity": 75.02844280565658, "lr": 0.001305, "grad_norm": 0.330634, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:01.961527+00:00", "epoch": 0, "step": 718, "train_loss": 4.323474884033203, "perplexity": 75.45035452675252, "lr": 0.001305, "grad_norm": 0.359847, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:02.269588+00:00", "epoch": 0, "step": 719, "train_loss": 4.315308570861816, "perplexity": 74.83671231029714, "lr": 0.001305, "grad_norm": 0.355567, "tokens_per_sec": 106370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:02.576043+00:00", "epoch": 0, "step": 720, "train_loss": 4.349807262420654, "perplexity": 77.463531352849, "lr": 0.001305, "grad_norm": 0.3388, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:02.882717+00:00", "epoch": 0, "step": 721, "train_loss": 4.351908206939697, "perplexity": 77.62644901502908, "lr": 0.001305, "grad_norm": 0.340878, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:03.189401+00:00", "epoch": 0, "step": 722, "train_loss": 4.322673797607422, "perplexity": 75.3899364751923, "lr": 0.001305, "grad_norm": 0.394895, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:03.496107+00:00", "epoch": 0, "step": 723, "train_loss": 4.316041469573975, "perplexity": 74.89158014419795, "lr": 0.001305, "grad_norm": 0.466916, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:03.802491+00:00", "epoch": 0, "step": 724, "train_loss": 4.25467586517334, "perplexity": 70.43398338102216, "lr": 0.001305, "grad_norm": 0.441714, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:04.108790+00:00", "epoch": 0, "step": 725, "train_loss": 4.45191764831543, "perplexity": 85.79130390900323, "lr": 0.001305, "grad_norm": 0.366991, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:04.414931+00:00", "epoch": 0, "step": 726, "train_loss": 4.3931803703308105, "perplexity": 80.8972936382181, "lr": 0.001305, "grad_norm": 0.354924, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:04.720725+00:00", "epoch": 0, "step": 727, "train_loss": 4.218683242797852, "perplexity": 67.9439596630278, "lr": 0.001305, "grad_norm": 0.369532, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:05.026829+00:00", "epoch": 0, "step": 728, "train_loss": 4.2903361320495605, "perplexity": 72.99099899074994, "lr": 0.001305, "grad_norm": 0.358323, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:05.334559+00:00", "epoch": 0, "step": 729, "train_loss": 4.324497699737549, "perplexity": 75.52756581398828, "lr": 0.001305, "grad_norm": 0.331423, "tokens_per_sec": 106483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:05.642064+00:00", "epoch": 0, "step": 730, "train_loss": 4.215602397918701, "perplexity": 67.73495698056696, "lr": 0.001305, "grad_norm": 0.332905, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:05.948920+00:00", "epoch": 0, "step": 731, "train_loss": 4.239758014678955, "perplexity": 69.39105818943521, "lr": 0.001305, "grad_norm": 0.351607, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:06.254857+00:00", "epoch": 0, "step": 732, "train_loss": 4.2364501953125, "perplexity": 69.16190431205791, "lr": 0.001305, "grad_norm": 0.334009, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:06.561961+00:00", "epoch": 0, "step": 733, "train_loss": 4.293955326080322, "perplexity": 73.25564619451183, "lr": 0.001305, "grad_norm": 0.267498, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:06.868212+00:00", "epoch": 0, "step": 734, "train_loss": 4.288909435272217, "perplexity": 72.88693721766406, "lr": 0.001305, "grad_norm": 0.307034, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:07.176247+00:00", "epoch": 0, "step": 735, "train_loss": 4.2628068923950195, "perplexity": 71.0090186628552, "lr": 0.001305, "grad_norm": 0.344129, "tokens_per_sec": 106377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:07.483607+00:00", "epoch": 0, "step": 736, "train_loss": 4.296646595001221, "perplexity": 73.4530623692473, "lr": 0.001305, "grad_norm": 0.363421, "tokens_per_sec": 106612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:07.790538+00:00", "epoch": 0, "step": 737, "train_loss": 4.233499050140381, "perplexity": 68.95809837042968, "lr": 0.001305, "grad_norm": 0.366946, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:08.097147+00:00", "epoch": 0, "step": 738, "train_loss": 4.343224048614502, "perplexity": 76.95524727086305, "lr": 0.001305, "grad_norm": 0.298635, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:08.403284+00:00", "epoch": 0, "step": 739, "train_loss": 4.238488674163818, "perplexity": 69.3030331865574, "lr": 0.001305, "grad_norm": 0.296072, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:08.710934+00:00", "epoch": 0, "step": 740, "train_loss": 4.205368518829346, "perplexity": 67.04530056996174, "lr": 0.001305, "grad_norm": 0.287867, "tokens_per_sec": 106567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:09.017657+00:00", "epoch": 0, "step": 741, "train_loss": 4.260894775390625, "perplexity": 70.87337083938904, "lr": 0.001305, "grad_norm": 0.273786, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:09.323180+00:00", "epoch": 0, "step": 742, "train_loss": 4.280643463134766, "perplexity": 72.28693902590706, "lr": 0.001305, "grad_norm": 0.286402, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:09.630554+00:00", "epoch": 0, "step": 743, "train_loss": 4.2881178855896, "perplexity": 72.82926641330705, "lr": 0.001305, "grad_norm": 0.282719, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:09.937543+00:00", "epoch": 0, "step": 744, "train_loss": 4.3231635093688965, "perplexity": 75.42686485517403, "lr": 0.001305, "grad_norm": 0.279492, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:10.244705+00:00", "epoch": 0, "step": 745, "train_loss": 4.321744441986084, "perplexity": 75.31990496106796, "lr": 0.001305, "grad_norm": 0.275534, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:10.552371+00:00", "epoch": 0, "step": 746, "train_loss": 4.277316093444824, "perplexity": 72.0468133709703, "lr": 0.001305, "grad_norm": 0.251602, "tokens_per_sec": 106504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:10.860052+00:00", "epoch": 0, "step": 747, "train_loss": 4.226263523101807, "perplexity": 68.46095091598816, "lr": 0.001305, "grad_norm": 0.294138, "tokens_per_sec": 106506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:11.166129+00:00", "epoch": 0, "step": 748, "train_loss": 4.218167304992676, "perplexity": 67.90891384711239, "lr": 0.001305, "grad_norm": 0.328095, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:11.472986+00:00", "epoch": 0, "step": 749, "train_loss": 4.274721145629883, "perplexity": 71.86009801311934, "lr": 0.001305, "grad_norm": 0.323384, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:11.779698+00:00", "epoch": 0, "step": 750, "train_loss": 4.253791332244873, "perplexity": 70.37170974903783, "lr": 0.001305, "grad_norm": 0.362452, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:04:12.856601+00:00", "step": 750, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0000_step_0000750.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:13.839419+00:00", "epoch": 0, "step": 751, "train_loss": 4.246506690979004, "perplexity": 69.86093973511868, "lr": 0.001305, "grad_norm": 0.439883, "tokens_per_sec": 15908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:14.144579+00:00", "epoch": 0, "step": 752, "train_loss": 4.348300457000732, "perplexity": 77.34689677883226, "lr": 0.001305, "grad_norm": 0.43029, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:14.450657+00:00", "epoch": 0, "step": 753, "train_loss": 4.249729156494141, "perplexity": 70.08642732212952, "lr": 0.001305, "grad_norm": 0.445224, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:14.758257+00:00", "epoch": 0, "step": 754, "train_loss": 4.279902458190918, "perplexity": 72.23339388776938, "lr": 0.001305, "grad_norm": 0.499144, "tokens_per_sec": 106529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:15.064757+00:00", "epoch": 0, "step": 755, "train_loss": 4.341965198516846, "perplexity": 76.8584331003701, "lr": 0.001305, "grad_norm": 0.40351, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:15.371534+00:00", "epoch": 0, "step": 756, "train_loss": 4.256539344787598, "perplexity": 70.56535804218797, "lr": 0.001305, "grad_norm": 0.359704, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:15.678074+00:00", "epoch": 0, "step": 757, "train_loss": 4.30161190032959, "perplexity": 73.81868621705249, "lr": 0.001305, "grad_norm": 0.332029, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:15.983599+00:00", "epoch": 0, "step": 758, "train_loss": 4.185184001922607, "perplexity": 65.70558975936022, "lr": 0.001305, "grad_norm": 0.293624, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:16.290844+00:00", "epoch": 0, "step": 759, "train_loss": 4.224372863769531, "perplexity": 68.3316368631732, "lr": 0.001305, "grad_norm": 0.283829, "tokens_per_sec": 106651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:16.597528+00:00", "epoch": 0, "step": 760, "train_loss": 4.247179985046387, "perplexity": 69.90799252979035, "lr": 0.001305, "grad_norm": 0.291714, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:16.905266+00:00", "epoch": 0, "step": 761, "train_loss": 4.348376750946045, "perplexity": 77.35279810386001, "lr": 0.001305, "grad_norm": 0.310505, "tokens_per_sec": 106419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:17.211927+00:00", "epoch": 0, "step": 762, "train_loss": 4.266749858856201, "perplexity": 71.28955755614685, "lr": 0.001305, "grad_norm": 0.267192, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:17.518466+00:00", "epoch": 0, "step": 763, "train_loss": 4.323240280151367, "perplexity": 75.43265565688762, "lr": 0.001305, "grad_norm": 0.300738, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:17.824698+00:00", "epoch": 0, "step": 764, "train_loss": 4.2502851486206055, "perplexity": 70.1254056587127, "lr": 0.001305, "grad_norm": 0.305572, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:18.219800+00:00", "epoch": 0, "step": 765, "train_loss": 4.2259931564331055, "perplexity": 68.44244385871134, "lr": 0.001305, "grad_norm": 0.326924, "tokens_per_sec": 82934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:18.525693+00:00", "epoch": 0, "step": 766, "train_loss": 4.3489508628845215, "perplexity": 77.3972200190818, "lr": 0.001305, "grad_norm": 0.340496, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:18.831362+00:00", "epoch": 0, "step": 767, "train_loss": 4.335426330566406, "perplexity": 76.35750549003441, "lr": 0.001305, "grad_norm": 0.356278, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:19.138752+00:00", "epoch": 0, "step": 768, "train_loss": 4.253495693206787, "perplexity": 70.35090819948573, "lr": 0.001305, "grad_norm": 0.391257, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:19.445605+00:00", "epoch": 0, "step": 769, "train_loss": 4.310342311859131, "perplexity": 74.46597516437873, "lr": 0.001305, "grad_norm": 0.362856, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:19.753179+00:00", "epoch": 0, "step": 770, "train_loss": 4.172597408294678, "perplexity": 64.88376305099489, "lr": 0.001305, "grad_norm": 0.338569, "tokens_per_sec": 106592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:20.059616+00:00", "epoch": 0, "step": 771, "train_loss": 4.187657356262207, "perplexity": 65.86830410709636, "lr": 0.001305, "grad_norm": 0.348401, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:20.366070+00:00", "epoch": 0, "step": 772, "train_loss": 4.33850622177124, "perplexity": 76.59304082503424, "lr": 0.001305, "grad_norm": 0.35236, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:20.672187+00:00", "epoch": 0, "step": 773, "train_loss": 4.2370100021362305, "perplexity": 69.20063245716254, "lr": 0.001305, "grad_norm": 0.358104, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:20.979404+00:00", "epoch": 0, "step": 774, "train_loss": 4.355381011962891, "perplexity": 77.89649918147963, "lr": 0.001305, "grad_norm": 0.364117, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:21.286037+00:00", "epoch": 0, "step": 775, "train_loss": 4.335440158843994, "perplexity": 76.35856139011688, "lr": 0.001305, "grad_norm": 0.352894, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:21.593546+00:00", "epoch": 0, "step": 776, "train_loss": 4.18675422668457, "perplexity": 65.80884334785523, "lr": 0.001305, "grad_norm": 0.353398, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:21.900396+00:00", "epoch": 0, "step": 777, "train_loss": 4.308739185333252, "perplexity": 74.34669242252758, "lr": 0.001305, "grad_norm": 0.371489, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:22.205891+00:00", "epoch": 0, "step": 778, "train_loss": 4.291060924530029, "perplexity": 73.04392149455995, "lr": 0.001305, "grad_norm": 0.373914, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:22.511420+00:00", "epoch": 0, "step": 779, "train_loss": 4.383452892303467, "perplexity": 80.11418201749777, "lr": 0.001305, "grad_norm": 0.349391, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:22.818897+00:00", "epoch": 0, "step": 780, "train_loss": 4.242405414581299, "perplexity": 69.57500745630753, "lr": 0.001305, "grad_norm": 0.295823, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:23.125986+00:00", "epoch": 0, "step": 781, "train_loss": 4.167120456695557, "perplexity": 64.52936920605916, "lr": 0.001305, "grad_norm": 0.310867, "tokens_per_sec": 106627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:23.432247+00:00", "epoch": 0, "step": 782, "train_loss": 4.209375858306885, "perplexity": 67.3145129020542, "lr": 0.001305, "grad_norm": 0.304728, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:23.737497+00:00", "epoch": 0, "step": 783, "train_loss": 4.248857021331787, "perplexity": 70.02532913127116, "lr": 0.001305, "grad_norm": 0.279057, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:24.043176+00:00", "epoch": 0, "step": 784, "train_loss": 4.189178943634033, "perplexity": 65.96860477559304, "lr": 0.001305, "grad_norm": 0.283314, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:24.349559+00:00", "epoch": 0, "step": 785, "train_loss": 4.06602144241333, "perplexity": 58.32445317686628, "lr": 0.001305, "grad_norm": 0.29614, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:24.656987+00:00", "epoch": 0, "step": 786, "train_loss": 4.233307361602783, "perplexity": 68.94488116023186, "lr": 0.001305, "grad_norm": 0.298654, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:24.964157+00:00", "epoch": 0, "step": 787, "train_loss": 4.2869553565979, "perplexity": 72.74464947400976, "lr": 0.001305, "grad_norm": 0.31974, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:25.269684+00:00", "epoch": 0, "step": 788, "train_loss": 4.154521942138672, "perplexity": 63.72149470374027, "lr": 0.001305, "grad_norm": 0.287962, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:25.576596+00:00", "epoch": 0, "step": 789, "train_loss": 4.145174980163574, "perplexity": 63.12866719987567, "lr": 0.001305, "grad_norm": 0.312995, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:25.883331+00:00", "epoch": 0, "step": 790, "train_loss": 4.246142864227295, "perplexity": 69.8355270795258, "lr": 0.001305, "grad_norm": 0.329897, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:26.190322+00:00", "epoch": 0, "step": 791, "train_loss": 4.193298816680908, "perplexity": 66.2409476762216, "lr": 0.001305, "grad_norm": 0.320669, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:26.497784+00:00", "epoch": 0, "step": 792, "train_loss": 4.130484104156494, "perplexity": 62.20803081278214, "lr": 0.001305, "grad_norm": 0.299414, "tokens_per_sec": 106580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:26.804969+00:00", "epoch": 0, "step": 793, "train_loss": 4.202103614807129, "perplexity": 66.82676104787107, "lr": 0.001305, "grad_norm": 0.29944, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:27.110245+00:00", "epoch": 0, "step": 794, "train_loss": 4.188201427459717, "perplexity": 65.9041509049107, "lr": 0.001305, "grad_norm": 0.236166, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:27.417057+00:00", "epoch": 0, "step": 795, "train_loss": 4.1849493980407715, "perplexity": 65.69017678098733, "lr": 0.001305, "grad_norm": 0.293242, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:27.722847+00:00", "epoch": 0, "step": 796, "train_loss": 4.226222515106201, "perplexity": 68.45814352717694, "lr": 0.001305, "grad_norm": 0.281782, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:28.030723+00:00", "epoch": 0, "step": 797, "train_loss": 4.0887322425842285, "perplexity": 59.66420400038092, "lr": 0.001305, "grad_norm": 0.299462, "tokens_per_sec": 106433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:28.339097+00:00", "epoch": 0, "step": 798, "train_loss": 4.320217609405518, "perplexity": 75.20499182509218, "lr": 0.001305, "grad_norm": 0.31094, "tokens_per_sec": 106261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:28.646265+00:00", "epoch": 0, "step": 799, "train_loss": 4.188753604888916, "perplexity": 65.94055173846539, "lr": 0.001305, "grad_norm": 0.321177, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:28.953637+00:00", "epoch": 0, "step": 800, "train_loss": 4.298616409301758, "perplexity": 73.59789386066109, "lr": 0.001305, "grad_norm": 0.332597, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:04:34.871831+00:00", "step": 800, "epoch": 0, "val_loss": 4.228866338729858, "val_ppl": 68.63937424973471, "eval_train_loss": 4.298616409301758, "eval_train_ppl": 73.59789386066109} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:04:35.857207+00:00", "step": 800, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_4p2289_epoch_0000_step_0000800.pt", "val_loss": 4.228866338729858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:36.856565+00:00", "epoch": 0, "step": 801, "train_loss": 4.214120388031006, "perplexity": 67.63464745279474, "lr": 0.001305, "grad_norm": 0.324778, "tokens_per_sec": 4146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:37.162688+00:00", "epoch": 0, "step": 802, "train_loss": 4.272418022155762, "perplexity": 71.69478577485414, "lr": 0.001305, "grad_norm": 0.354924, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:37.467900+00:00", "epoch": 0, "step": 803, "train_loss": 4.22130012512207, "perplexity": 68.1219938557742, "lr": 0.001305, "grad_norm": 0.312428, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:37.773024+00:00", "epoch": 0, "step": 804, "train_loss": 4.26834774017334, "perplexity": 71.40356086589593, "lr": 0.001305, "grad_norm": 0.339081, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:38.078865+00:00", "epoch": 0, "step": 805, "train_loss": 4.187559604644775, "perplexity": 65.86186568852045, "lr": 0.001305, "grad_norm": 0.300065, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:38.385022+00:00", "epoch": 0, "step": 806, "train_loss": 4.216752529144287, "perplexity": 67.81290588679698, "lr": 0.001305, "grad_norm": 0.282111, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:38.691118+00:00", "epoch": 0, "step": 807, "train_loss": 4.178972244262695, "perplexity": 65.29870759390394, "lr": 0.001305, "grad_norm": 0.305268, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:38.996795+00:00", "epoch": 0, "step": 808, "train_loss": 4.228003978729248, "perplexity": 68.58020791392218, "lr": 0.001305, "grad_norm": 0.307586, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:39.303273+00:00", "epoch": 0, "step": 809, "train_loss": 4.214118957519531, "perplexity": 67.63455070072469, "lr": 0.001305, "grad_norm": 0.296835, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:39.608955+00:00", "epoch": 0, "step": 810, "train_loss": 4.109990119934082, "perplexity": 60.94611541500981, "lr": 0.001305, "grad_norm": 0.32967, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:39.913941+00:00", "epoch": 0, "step": 811, "train_loss": 4.102031707763672, "perplexity": 60.46300604993868, "lr": 0.001305, "grad_norm": 0.364188, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:40.220805+00:00", "epoch": 0, "step": 812, "train_loss": 4.250276565551758, "perplexity": 70.12480377011099, "lr": 0.001305, "grad_norm": 0.416468, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:40.527120+00:00", "epoch": 0, "step": 813, "train_loss": 4.198798179626465, "perplexity": 66.60623419029812, "lr": 0.001305, "grad_norm": 0.427585, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:40.831669+00:00", "epoch": 0, "step": 814, "train_loss": 4.19912576675415, "perplexity": 66.62805710950103, "lr": 0.001305, "grad_norm": 0.388862, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:41.138694+00:00", "epoch": 0, "step": 815, "train_loss": 4.213666915893555, "perplexity": 67.60398397768982, "lr": 0.001305, "grad_norm": 0.326008, "tokens_per_sec": 106727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:41.443941+00:00", "epoch": 0, "step": 816, "train_loss": 4.164638519287109, "perplexity": 64.36940993724228, "lr": 0.001305, "grad_norm": 0.295088, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:41.750471+00:00", "epoch": 0, "step": 817, "train_loss": 4.231623649597168, "perplexity": 68.82889550671581, "lr": 0.001305, "grad_norm": 0.290121, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:42.055474+00:00", "epoch": 0, "step": 818, "train_loss": 4.1414642333984375, "perplexity": 62.894846794734214, "lr": 0.001305, "grad_norm": 0.355958, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:42.360398+00:00", "epoch": 0, "step": 819, "train_loss": 4.2293701171875, "perplexity": 68.67396199937781, "lr": 0.001305, "grad_norm": 0.425492, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:42.666297+00:00", "epoch": 0, "step": 820, "train_loss": 4.168404579162598, "perplexity": 64.61228604513467, "lr": 0.001305, "grad_norm": 0.495137, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:42.972750+00:00", "epoch": 0, "step": 821, "train_loss": 4.167293071746826, "perplexity": 64.54050890784549, "lr": 0.001305, "grad_norm": 0.431139, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:43.279193+00:00", "epoch": 0, "step": 822, "train_loss": 4.357179641723633, "perplexity": 78.03673221901715, "lr": 0.001305, "grad_norm": 0.373544, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:43.585694+00:00", "epoch": 0, "step": 823, "train_loss": 4.22939920425415, "perplexity": 68.67595955253897, "lr": 0.001305, "grad_norm": 0.372102, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:43.891781+00:00", "epoch": 0, "step": 824, "train_loss": 4.273785591125488, "perplexity": 71.7929004131469, "lr": 0.001305, "grad_norm": 0.34373, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:44.197965+00:00", "epoch": 0, "step": 825, "train_loss": 4.214961051940918, "perplexity": 67.69152936590373, "lr": 0.001305, "grad_norm": 0.346895, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:44.506344+00:00", "epoch": 0, "step": 826, "train_loss": 4.209615707397461, "perplexity": 67.33066016313187, "lr": 0.001305, "grad_norm": 0.32507, "tokens_per_sec": 106260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:44.813191+00:00", "epoch": 0, "step": 827, "train_loss": 4.2309250831604, "perplexity": 68.78083074060585, "lr": 0.001305, "grad_norm": 0.293632, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:45.120853+00:00", "epoch": 0, "step": 828, "train_loss": 4.23520565032959, "perplexity": 69.07588275099212, "lr": 0.001305, "grad_norm": 0.273677, "tokens_per_sec": 106508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:45.427567+00:00", "epoch": 0, "step": 829, "train_loss": 4.282643795013428, "perplexity": 72.43168161279333, "lr": 0.001305, "grad_norm": 0.302493, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:45.733806+00:00", "epoch": 0, "step": 830, "train_loss": 4.230467796325684, "perplexity": 68.74938536255395, "lr": 0.001305, "grad_norm": 0.299927, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:46.040025+00:00", "epoch": 0, "step": 831, "train_loss": 4.141464710235596, "perplexity": 62.894876785341374, "lr": 0.001305, "grad_norm": 0.286816, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:46.346641+00:00", "epoch": 0, "step": 832, "train_loss": 4.224900722503662, "perplexity": 68.36771583597161, "lr": 0.001305, "grad_norm": 0.317435, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:46.653087+00:00", "epoch": 0, "step": 833, "train_loss": 4.205165863037109, "perplexity": 67.03171482812014, "lr": 0.001305, "grad_norm": 0.304632, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:46.959758+00:00", "epoch": 0, "step": 834, "train_loss": 4.185776710510254, "perplexity": 65.74454557024835, "lr": 0.001305, "grad_norm": 0.296834, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:47.266005+00:00", "epoch": 0, "step": 835, "train_loss": 4.09935188293457, "perplexity": 60.301192697641525, "lr": 0.001305, "grad_norm": 0.313195, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:47.571922+00:00", "epoch": 0, "step": 836, "train_loss": 4.156573295593262, "perplexity": 63.85234417540518, "lr": 0.001305, "grad_norm": 0.330908, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:47.878443+00:00", "epoch": 0, "step": 837, "train_loss": 4.297108173370361, "perplexity": 73.48697453994421, "lr": 0.001305, "grad_norm": 0.363317, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:48.185132+00:00", "epoch": 0, "step": 838, "train_loss": 4.169047832489014, "perplexity": 64.65386148339645, "lr": 0.001305, "grad_norm": 0.332896, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:48.492143+00:00", "epoch": 0, "step": 839, "train_loss": 4.123307228088379, "perplexity": 61.76316975102483, "lr": 0.001305, "grad_norm": 0.288278, "tokens_per_sec": 106733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:48.797678+00:00", "epoch": 0, "step": 840, "train_loss": 4.180887222290039, "perplexity": 65.42387299040233, "lr": 0.001305, "grad_norm": 0.278905, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:49.103586+00:00", "epoch": 0, "step": 841, "train_loss": 4.134549617767334, "perplexity": 62.46145320579361, "lr": 0.001305, "grad_norm": 0.282469, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:49.409267+00:00", "epoch": 0, "step": 842, "train_loss": 4.096519470214844, "perplexity": 60.13063648924189, "lr": 0.001305, "grad_norm": 0.262462, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:49.716222+00:00", "epoch": 0, "step": 843, "train_loss": 4.086573123931885, "perplexity": 59.53552087570499, "lr": 0.001305, "grad_norm": 0.250633, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:50.022763+00:00", "epoch": 0, "step": 844, "train_loss": 4.209360599517822, "perplexity": 67.31348577193738, "lr": 0.001305, "grad_norm": 0.284697, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:50.330709+00:00", "epoch": 0, "step": 845, "train_loss": 4.187005043029785, "perplexity": 65.82535135157859, "lr": 0.001305, "grad_norm": 0.279693, "tokens_per_sec": 106407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:50.637019+00:00", "epoch": 0, "step": 846, "train_loss": 4.165198802947998, "perplexity": 64.4054851711295, "lr": 0.001305, "grad_norm": 0.267443, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:50.943714+00:00", "epoch": 0, "step": 847, "train_loss": 4.242526531219482, "perplexity": 69.58343465763909, "lr": 0.001305, "grad_norm": 0.287229, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:51.250082+00:00", "epoch": 0, "step": 848, "train_loss": 4.108734607696533, "perplexity": 60.86964483619408, "lr": 0.001305, "grad_norm": 0.296738, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:51.557639+00:00", "epoch": 0, "step": 849, "train_loss": 4.124941825866699, "perplexity": 61.86421024887094, "lr": 0.001305, "grad_norm": 0.321191, "tokens_per_sec": 106543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:51.864552+00:00", "epoch": 0, "step": 850, "train_loss": 4.116701602935791, "perplexity": 61.35652993712396, "lr": 0.001305, "grad_norm": 0.352183, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:52.171582+00:00", "epoch": 0, "step": 851, "train_loss": 4.122675895690918, "perplexity": 61.724188967201734, "lr": 0.001305, "grad_norm": 0.34026, "tokens_per_sec": 106670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:52.477978+00:00", "epoch": 0, "step": 852, "train_loss": 4.183395862579346, "perplexity": 65.588203991572, "lr": 0.001305, "grad_norm": 0.335789, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:52.783972+00:00", "epoch": 0, "step": 853, "train_loss": 4.1229248046875, "perplexity": 61.739554585383374, "lr": 0.001305, "grad_norm": 0.348691, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:53.090401+00:00", "epoch": 0, "step": 854, "train_loss": 4.137887001037598, "perplexity": 62.67025925385576, "lr": 0.001305, "grad_norm": 0.330102, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:53.398299+00:00", "epoch": 0, "step": 855, "train_loss": 4.224945068359375, "perplexity": 68.37074772805892, "lr": 0.001305, "grad_norm": 0.308331, "tokens_per_sec": 106424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:53.705429+00:00", "epoch": 0, "step": 856, "train_loss": 4.195294380187988, "perplexity": 66.37326767664952, "lr": 0.001305, "grad_norm": 0.323652, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:54.011964+00:00", "epoch": 0, "step": 857, "train_loss": 4.171878337860107, "perplexity": 64.83712382575486, "lr": 0.001305, "grad_norm": 0.357452, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:54.317518+00:00", "epoch": 0, "step": 858, "train_loss": 4.203812122344971, "perplexity": 66.94103266201142, "lr": 0.001305, "grad_norm": 0.37437, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:54.623939+00:00", "epoch": 0, "step": 859, "train_loss": 4.192063808441162, "perplexity": 66.15919005609183, "lr": 0.001305, "grad_norm": 0.357313, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:54.931502+00:00", "epoch": 0, "step": 860, "train_loss": 4.087775707244873, "perplexity": 59.60716036723492, "lr": 0.001305, "grad_norm": 0.321002, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:55.237786+00:00", "epoch": 0, "step": 861, "train_loss": 4.1132988929748535, "perplexity": 61.148106265292085, "lr": 0.001305, "grad_norm": 0.302514, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:55.542977+00:00", "epoch": 0, "step": 862, "train_loss": 4.215283393859863, "perplexity": 67.71335270047469, "lr": 0.001305, "grad_norm": 0.313194, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:55.848645+00:00", "epoch": 0, "step": 863, "train_loss": 4.162082195281982, "perplexity": 64.2050710106887, "lr": 0.001305, "grad_norm": 0.294619, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:56.154882+00:00", "epoch": 0, "step": 864, "train_loss": 4.1236371994018555, "perplexity": 61.78355318806186, "lr": 0.001305, "grad_norm": 0.281003, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:56.460938+00:00", "epoch": 0, "step": 865, "train_loss": 4.171465873718262, "perplexity": 64.81038635162004, "lr": 0.001305, "grad_norm": 0.316014, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:56.767898+00:00", "epoch": 0, "step": 866, "train_loss": 4.197991371154785, "perplexity": 66.55251738878584, "lr": 0.001305, "grad_norm": 0.31174, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:57.074905+00:00", "epoch": 0, "step": 867, "train_loss": 4.15715217590332, "perplexity": 63.88931774080141, "lr": 0.001305, "grad_norm": 0.281565, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:57.380984+00:00", "epoch": 0, "step": 868, "train_loss": 4.2127156257629395, "perplexity": 67.53970355445526, "lr": 0.001305, "grad_norm": 0.283697, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:57.686259+00:00", "epoch": 0, "step": 869, "train_loss": 4.095662593841553, "perplexity": 60.07913403629211, "lr": 0.001305, "grad_norm": 0.324846, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:57.993124+00:00", "epoch": 0, "step": 870, "train_loss": 4.1254425048828125, "perplexity": 61.89519211612917, "lr": 0.001305, "grad_norm": 0.378131, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:58.300202+00:00", "epoch": 0, "step": 871, "train_loss": 4.205267429351807, "perplexity": 67.03852333811486, "lr": 0.001305, "grad_norm": 0.394958, "tokens_per_sec": 106651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:58.608035+00:00", "epoch": 0, "step": 872, "train_loss": 4.046669006347656, "perplexity": 57.20658455438096, "lr": 0.001305, "grad_norm": 0.392184, "tokens_per_sec": 106447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:58.914849+00:00", "epoch": 0, "step": 873, "train_loss": 4.130817413330078, "perplexity": 62.228768776007335, "lr": 0.001305, "grad_norm": 0.38144, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:59.220513+00:00", "epoch": 0, "step": 874, "train_loss": 4.199638843536377, "perplexity": 66.66225118999304, "lr": 0.001305, "grad_norm": 0.347524, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:59.528169+00:00", "epoch": 0, "step": 875, "train_loss": 4.152890682220459, "perplexity": 63.61763311915281, "lr": 0.001305, "grad_norm": 0.30987, "tokens_per_sec": 106508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:04:59.835193+00:00", "epoch": 0, "step": 876, "train_loss": 4.091093063354492, "perplexity": 59.8052268918129, "lr": 0.001305, "grad_norm": 0.302302, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:00.144155+00:00", "epoch": 0, "step": 877, "train_loss": 4.259714603424072, "perplexity": 70.78977741096216, "lr": 0.001305, "grad_norm": 0.29568, "tokens_per_sec": 106058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:00.451262+00:00", "epoch": 0, "step": 878, "train_loss": 4.039159774780273, "perplexity": 56.77861593256802, "lr": 0.001305, "grad_norm": 0.326201, "tokens_per_sec": 106700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:00.758015+00:00", "epoch": 0, "step": 879, "train_loss": 4.149225234985352, "perplexity": 63.38487288751008, "lr": 0.001305, "grad_norm": 0.342781, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:01.064840+00:00", "epoch": 0, "step": 880, "train_loss": 4.207443714141846, "perplexity": 67.18457712647617, "lr": 0.001305, "grad_norm": 0.371499, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:01.372432+00:00", "epoch": 0, "step": 881, "train_loss": 4.059367656707764, "perplexity": 57.93766300053094, "lr": 0.001305, "grad_norm": 0.399107, "tokens_per_sec": 106472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:01.680546+00:00", "epoch": 0, "step": 882, "train_loss": 4.212888717651367, "perplexity": 67.55139514111755, "lr": 0.001305, "grad_norm": 0.349353, "tokens_per_sec": 106350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:01.987603+00:00", "epoch": 0, "step": 883, "train_loss": 4.1210737228393555, "perplexity": 61.62537532675735, "lr": 0.001305, "grad_norm": 0.32582, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:02.294945+00:00", "epoch": 0, "step": 884, "train_loss": 4.103935241699219, "perplexity": 60.578209045419655, "lr": 0.001305, "grad_norm": 0.30456, "tokens_per_sec": 106619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:02.602052+00:00", "epoch": 0, "step": 885, "train_loss": 4.043346405029297, "perplexity": 57.016825302864774, "lr": 0.001305, "grad_norm": 0.294409, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:02.909049+00:00", "epoch": 0, "step": 886, "train_loss": 4.102554798126221, "perplexity": 60.494641939187154, "lr": 0.001305, "grad_norm": 0.29631, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:03.215589+00:00", "epoch": 0, "step": 887, "train_loss": 4.1389641761779785, "perplexity": 62.73780247057131, "lr": 0.001305, "grad_norm": 0.292104, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:03.521503+00:00", "epoch": 0, "step": 888, "train_loss": 4.12463903427124, "perplexity": 61.845481121602134, "lr": 0.001305, "grad_norm": 0.287905, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:03.829299+00:00", "epoch": 0, "step": 889, "train_loss": 4.125784873962402, "perplexity": 61.91638674407061, "lr": 0.001305, "grad_norm": 0.328867, "tokens_per_sec": 106460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:04.136639+00:00", "epoch": 0, "step": 890, "train_loss": 4.169133186340332, "perplexity": 64.6593801749941, "lr": 0.001305, "grad_norm": 0.318635, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:04.443510+00:00", "epoch": 0, "step": 891, "train_loss": 4.102991104125977, "perplexity": 60.52104187322431, "lr": 0.001305, "grad_norm": 0.31879, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:04.749358+00:00", "epoch": 0, "step": 892, "train_loss": 4.0818023681640625, "perplexity": 59.25216788747695, "lr": 0.001305, "grad_norm": 0.333163, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:05.056301+00:00", "epoch": 0, "step": 893, "train_loss": 4.075868606567383, "perplexity": 58.901620709387224, "lr": 0.001305, "grad_norm": 0.314287, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:05.362841+00:00", "epoch": 0, "step": 894, "train_loss": 4.20474910736084, "perplexity": 67.00378480087839, "lr": 0.001305, "grad_norm": 0.261858, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:05.669491+00:00", "epoch": 0, "step": 895, "train_loss": 4.159369468688965, "perplexity": 64.0311362325643, "lr": 0.001305, "grad_norm": 0.26576, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:05.977728+00:00", "epoch": 0, "step": 896, "train_loss": 4.1039323806762695, "perplexity": 60.578035730021284, "lr": 0.001305, "grad_norm": 0.262321, "tokens_per_sec": 106308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:06.285107+00:00", "epoch": 0, "step": 897, "train_loss": 4.044837951660156, "perplexity": 57.10193201108136, "lr": 0.001305, "grad_norm": 0.275287, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:06.590689+00:00", "epoch": 0, "step": 898, "train_loss": 4.086853504180908, "perplexity": 59.55221580022804, "lr": 0.001305, "grad_norm": 0.263334, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:06.897159+00:00", "epoch": 0, "step": 899, "train_loss": 4.23227071762085, "perplexity": 68.87344689643892, "lr": 0.001305, "grad_norm": 0.273731, "tokens_per_sec": 106921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:07.204172+00:00", "epoch": 0, "step": 900, "train_loss": 4.1067328453063965, "perplexity": 60.74792014306277, "lr": 0.001305, "grad_norm": 0.304288, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:05:13.091907+00:00", "step": 900, "epoch": 0, "val_loss": 4.145457231998444, "val_ppl": 63.14648789686962, "eval_train_loss": 4.1067328453063965, "eval_train_ppl": 60.74792014306277} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:05:14.070154+00:00", "step": 900, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_4p1455_epoch_0000_step_0000900.pt", "val_loss": 4.145457231998444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:15.027070+00:00", "epoch": 0, "step": 901, "train_loss": 4.089560031890869, "perplexity": 59.71361383808786, "lr": 0.001305, "grad_norm": 0.306993, "tokens_per_sec": 4189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:15.332349+00:00", "epoch": 0, "step": 902, "train_loss": 4.029262542724609, "perplexity": 56.21943652360648, "lr": 0.001305, "grad_norm": 0.316114, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:15.638226+00:00", "epoch": 0, "step": 903, "train_loss": 4.100391387939453, "perplexity": 60.36390868039268, "lr": 0.001305, "grad_norm": 0.312342, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:15.943109+00:00", "epoch": 0, "step": 904, "train_loss": 4.125463008880615, "perplexity": 61.896461228023206, "lr": 0.001305, "grad_norm": 0.300039, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:16.249483+00:00", "epoch": 0, "step": 905, "train_loss": 4.067574501037598, "perplexity": 58.4151048473367, "lr": 0.001305, "grad_norm": 0.339263, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:16.556519+00:00", "epoch": 0, "step": 906, "train_loss": 4.086500644683838, "perplexity": 59.53120594229329, "lr": 0.001305, "grad_norm": 0.356638, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:16.862675+00:00", "epoch": 0, "step": 907, "train_loss": 4.270926475524902, "perplexity": 71.58792936914782, "lr": 0.001305, "grad_norm": 0.355965, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:17.168845+00:00", "epoch": 0, "step": 908, "train_loss": 4.1674933433532715, "perplexity": 64.55343583365016, "lr": 0.001305, "grad_norm": 0.307264, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:17.474757+00:00", "epoch": 0, "step": 909, "train_loss": 4.173568248748779, "perplexity": 64.94678542036193, "lr": 0.001305, "grad_norm": 0.284535, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:17.780542+00:00", "epoch": 0, "step": 910, "train_loss": 4.151553630828857, "perplexity": 63.53262991375089, "lr": 0.001305, "grad_norm": 0.330322, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:18.087594+00:00", "epoch": 0, "step": 911, "train_loss": 4.074827194213867, "perplexity": 58.84031176342623, "lr": 0.001305, "grad_norm": 0.381305, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:18.395779+00:00", "epoch": 0, "step": 912, "train_loss": 4.128440856933594, "perplexity": 62.08105419307041, "lr": 0.001305, "grad_norm": 0.41792, "tokens_per_sec": 106327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:18.703120+00:00", "epoch": 0, "step": 913, "train_loss": 4.18454647064209, "perplexity": 65.66371374063806, "lr": 0.001305, "grad_norm": 0.405946, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:19.010197+00:00", "epoch": 0, "step": 914, "train_loss": 4.078949451446533, "perplexity": 59.08336728867691, "lr": 0.001305, "grad_norm": 0.410509, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:19.360992+00:00", "epoch": 0, "step": 915, "train_loss": 4.166635036468506, "perplexity": 64.49805294639557, "lr": 0.001305, "grad_norm": 0.40566, "tokens_per_sec": 93410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:19.666689+00:00", "epoch": 0, "step": 916, "train_loss": 4.204157829284668, "perplexity": 66.96417864218564, "lr": 0.001305, "grad_norm": 0.386166, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:19.972697+00:00", "epoch": 0, "step": 917, "train_loss": 4.17168664932251, "perplexity": 64.82469648343425, "lr": 0.001305, "grad_norm": 0.352643, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:20.279242+00:00", "epoch": 0, "step": 918, "train_loss": 4.069337368011475, "perplexity": 58.51817372811835, "lr": 0.001305, "grad_norm": 0.317779, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:20.586643+00:00", "epoch": 0, "step": 919, "train_loss": 4.114680290222168, "perplexity": 61.23263446103211, "lr": 0.001305, "grad_norm": 0.320718, "tokens_per_sec": 106599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:20.894188+00:00", "epoch": 0, "step": 920, "train_loss": 4.198845863342285, "perplexity": 66.60941029876484, "lr": 0.001305, "grad_norm": 0.282815, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:21.199614+00:00", "epoch": 0, "step": 921, "train_loss": 4.122365474700928, "perplexity": 61.705031456956675, "lr": 0.001305, "grad_norm": 0.264712, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:21.505421+00:00", "epoch": 0, "step": 922, "train_loss": 4.241522312164307, "perplexity": 69.51359272080259, "lr": 0.001305, "grad_norm": 0.29656, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:21.812190+00:00", "epoch": 0, "step": 923, "train_loss": 4.043828964233398, "perplexity": 57.04434593633677, "lr": 0.001305, "grad_norm": 0.316281, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:22.118626+00:00", "epoch": 0, "step": 924, "train_loss": 4.065263271331787, "perplexity": 58.28025002201505, "lr": 0.001305, "grad_norm": 0.334802, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:22.425550+00:00", "epoch": 0, "step": 925, "train_loss": 4.1370720863342285, "perplexity": 62.619209141694476, "lr": 0.001305, "grad_norm": 0.290375, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:22.731478+00:00", "epoch": 0, "step": 926, "train_loss": 4.147136211395264, "perplexity": 63.252598602947714, "lr": 0.001305, "grad_norm": 0.301366, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:23.037545+00:00", "epoch": 0, "step": 927, "train_loss": 4.179834365844727, "perplexity": 65.355027292725, "lr": 0.001305, "grad_norm": 0.311102, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:23.343007+00:00", "epoch": 0, "step": 928, "train_loss": 4.220506191253662, "perplexity": 68.06793096168973, "lr": 0.001305, "grad_norm": 0.297358, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:23.649582+00:00", "epoch": 0, "step": 929, "train_loss": 4.088734149932861, "perplexity": 59.66431780092737, "lr": 0.001305, "grad_norm": 0.28075, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:23.956398+00:00", "epoch": 0, "step": 930, "train_loss": 4.139472007751465, "perplexity": 62.769670798703736, "lr": 0.001305, "grad_norm": 0.268309, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:24.262342+00:00", "epoch": 0, "step": 931, "train_loss": 4.073808193206787, "perplexity": 58.78038396491148, "lr": 0.001305, "grad_norm": 0.267021, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:24.567989+00:00", "epoch": 0, "step": 932, "train_loss": 4.072880744934082, "perplexity": 58.72589347179702, "lr": 0.001305, "grad_norm": 0.271197, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:24.874591+00:00", "epoch": 0, "step": 933, "train_loss": 4.090003490447998, "perplexity": 59.74010022349452, "lr": 0.001305, "grad_norm": 0.290728, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:25.181260+00:00", "epoch": 0, "step": 934, "train_loss": 4.068161487579346, "perplexity": 58.44940379324189, "lr": 0.001305, "grad_norm": 0.314038, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:25.488394+00:00", "epoch": 0, "step": 935, "train_loss": 4.054525375366211, "perplexity": 57.65779069309946, "lr": 0.001305, "grad_norm": 0.310268, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:25.794816+00:00", "epoch": 0, "step": 936, "train_loss": 4.095520973205566, "perplexity": 60.070626193577645, "lr": 0.001305, "grad_norm": 0.251911, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:26.100573+00:00", "epoch": 0, "step": 937, "train_loss": 4.154970645904541, "perplexity": 63.75009319402597, "lr": 0.001305, "grad_norm": 0.290326, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:26.406303+00:00", "epoch": 0, "step": 938, "train_loss": 4.0220208168029785, "perplexity": 55.81378136736329, "lr": 0.001305, "grad_norm": 0.299292, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:26.713003+00:00", "epoch": 0, "step": 939, "train_loss": 4.21890926361084, "perplexity": 67.95931814763168, "lr": 0.001305, "grad_norm": 0.297729, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:27.019206+00:00", "epoch": 0, "step": 940, "train_loss": 4.170492649078369, "perplexity": 64.7473419698507, "lr": 0.001305, "grad_norm": 0.28021, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:27.325848+00:00", "epoch": 0, "step": 941, "train_loss": 4.077008247375488, "perplexity": 58.96878566470379, "lr": 0.001305, "grad_norm": 0.30937, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:27.632978+00:00", "epoch": 0, "step": 942, "train_loss": 4.097285747528076, "perplexity": 60.17673089010517, "lr": 0.001305, "grad_norm": 0.304651, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:27.938439+00:00", "epoch": 0, "step": 943, "train_loss": 4.177403926849365, "perplexity": 65.19637875673925, "lr": 0.001305, "grad_norm": 0.298454, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:28.244173+00:00", "epoch": 0, "step": 944, "train_loss": 4.174360275268555, "perplexity": 64.9982453729225, "lr": 0.001305, "grad_norm": 0.310726, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:28.550849+00:00", "epoch": 0, "step": 945, "train_loss": 4.078607082366943, "perplexity": 59.06314243297939, "lr": 0.001305, "grad_norm": 0.315249, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:28.857626+00:00", "epoch": 0, "step": 946, "train_loss": 4.021395206451416, "perplexity": 55.778874608133485, "lr": 0.001305, "grad_norm": 0.320946, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:29.165832+00:00", "epoch": 0, "step": 947, "train_loss": 4.038529396057129, "perplexity": 56.74283518005195, "lr": 0.001305, "grad_norm": 0.304338, "tokens_per_sec": 106319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:29.471860+00:00", "epoch": 0, "step": 948, "train_loss": 4.0811848640441895, "perplexity": 59.21559072413183, "lr": 0.001305, "grad_norm": 0.291295, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:29.777956+00:00", "epoch": 0, "step": 949, "train_loss": 4.088426113128662, "perplexity": 59.64594182553092, "lr": 0.001305, "grad_norm": 0.271931, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:30.084393+00:00", "epoch": 0, "step": 950, "train_loss": 4.189910411834717, "perplexity": 66.01687636464375, "lr": 0.001305, "grad_norm": 0.284754, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:30.392244+00:00", "epoch": 0, "step": 951, "train_loss": 4.02869176864624, "perplexity": 56.18735708246593, "lr": 0.001305, "grad_norm": 0.259961, "tokens_per_sec": 106390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:30.700024+00:00", "epoch": 0, "step": 952, "train_loss": 3.983640432357788, "perplexity": 53.7122144251833, "lr": 0.001305, "grad_norm": 0.258751, "tokens_per_sec": 106462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:31.007308+00:00", "epoch": 0, "step": 953, "train_loss": 4.085861682891846, "perplexity": 59.49317992614441, "lr": 0.001305, "grad_norm": 0.262114, "tokens_per_sec": 106638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:31.313332+00:00", "epoch": 0, "step": 954, "train_loss": 4.23184871673584, "perplexity": 68.84438837268756, "lr": 0.001305, "grad_norm": 0.270536, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:31.620153+00:00", "epoch": 0, "step": 955, "train_loss": 4.046256065368652, "perplexity": 57.182966488118915, "lr": 0.001305, "grad_norm": 0.289484, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:31.925856+00:00", "epoch": 0, "step": 956, "train_loss": 4.016034126281738, "perplexity": 55.48063973442009, "lr": 0.001305, "grad_norm": 0.282964, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:32.232828+00:00", "epoch": 0, "step": 957, "train_loss": 4.099643230438232, "perplexity": 60.31876385913456, "lr": 0.001305, "grad_norm": 0.334234, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:32.540648+00:00", "epoch": 0, "step": 958, "train_loss": 3.96360182762146, "perplexity": 52.64660887224775, "lr": 0.001305, "grad_norm": 0.342256, "tokens_per_sec": 106452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:32.848468+00:00", "epoch": 0, "step": 959, "train_loss": 4.023388385772705, "perplexity": 55.89016277936453, "lr": 0.001305, "grad_norm": 0.324034, "tokens_per_sec": 106452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:33.156948+00:00", "epoch": 0, "step": 960, "train_loss": 4.029219150543213, "perplexity": 56.21699709254535, "lr": 0.001305, "grad_norm": 0.37562, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:33.463257+00:00", "epoch": 0, "step": 961, "train_loss": 4.036494731903076, "perplexity": 56.62749994133902, "lr": 0.001305, "grad_norm": 0.418343, "tokens_per_sec": 106512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:33.769322+00:00", "epoch": 0, "step": 962, "train_loss": 4.060443878173828, "perplexity": 58.00005032241154, "lr": 0.001305, "grad_norm": 0.418824, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:34.076251+00:00", "epoch": 0, "step": 963, "train_loss": 4.045047283172607, "perplexity": 57.113886496054675, "lr": 0.001305, "grad_norm": 0.383354, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:34.384392+00:00", "epoch": 0, "step": 964, "train_loss": 4.0119309425354, "perplexity": 55.253458876364384, "lr": 0.001305, "grad_norm": 0.316024, "tokens_per_sec": 106342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:34.692350+00:00", "epoch": 0, "step": 965, "train_loss": 4.066823482513428, "perplexity": 58.37125049128316, "lr": 0.001305, "grad_norm": 0.29973, "tokens_per_sec": 106406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:34.999016+00:00", "epoch": 0, "step": 966, "train_loss": 4.119204998016357, "perplexity": 61.510321993149695, "lr": 0.001305, "grad_norm": 0.297306, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:35.306683+00:00", "epoch": 0, "step": 967, "train_loss": 4.078423976898193, "perplexity": 59.052328638661, "lr": 0.001305, "grad_norm": 0.340338, "tokens_per_sec": 106505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:35.613326+00:00", "epoch": 0, "step": 968, "train_loss": 4.142634391784668, "perplexity": 62.96848680399619, "lr": 0.001305, "grad_norm": 0.345859, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:35.921599+00:00", "epoch": 0, "step": 969, "train_loss": 4.040563583374023, "perplexity": 56.8583782139446, "lr": 0.001305, "grad_norm": 0.307125, "tokens_per_sec": 106295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:36.229101+00:00", "epoch": 0, "step": 970, "train_loss": 4.217467308044434, "perplexity": 67.86139444831839, "lr": 0.001305, "grad_norm": 0.324748, "tokens_per_sec": 106620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:36.536549+00:00", "epoch": 0, "step": 971, "train_loss": 4.1250128746032715, "perplexity": 61.86860577899475, "lr": 0.001305, "grad_norm": 0.307396, "tokens_per_sec": 106523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:36.843123+00:00", "epoch": 0, "step": 972, "train_loss": 4.101600646972656, "perplexity": 60.436948435335076, "lr": 0.001305, "grad_norm": 0.33505, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:37.149392+00:00", "epoch": 0, "step": 973, "train_loss": 4.122751712799072, "perplexity": 61.72886889411942, "lr": 0.001305, "grad_norm": 0.347991, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:37.456434+00:00", "epoch": 0, "step": 974, "train_loss": 4.136460304260254, "perplexity": 62.58091154814022, "lr": 0.001305, "grad_norm": 0.35427, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:37.763659+00:00", "epoch": 0, "step": 975, "train_loss": 4.111347198486328, "perplexity": 61.02888022755616, "lr": 0.001305, "grad_norm": 0.364281, "tokens_per_sec": 106657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:38.070589+00:00", "epoch": 0, "step": 976, "train_loss": 4.060197830200195, "perplexity": 57.98578128306521, "lr": 0.001305, "grad_norm": 0.304701, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:38.377277+00:00", "epoch": 0, "step": 977, "train_loss": 4.191137790679932, "perplexity": 66.09795382832442, "lr": 0.001305, "grad_norm": 0.316182, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:38.683110+00:00", "epoch": 0, "step": 978, "train_loss": 4.081794738769531, "perplexity": 59.25171583103577, "lr": 0.001305, "grad_norm": 0.276534, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:38.989369+00:00", "epoch": 0, "step": 979, "train_loss": 4.186179161071777, "perplexity": 65.77100982444999, "lr": 0.001305, "grad_norm": 0.286025, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:39.296301+00:00", "epoch": 0, "step": 980, "train_loss": 4.056589126586914, "perplexity": 57.77690489779176, "lr": 0.001305, "grad_norm": 0.258015, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:39.604276+00:00", "epoch": 0, "step": 981, "train_loss": 3.993875503540039, "perplexity": 54.26478574277941, "lr": 0.001305, "grad_norm": 0.271237, "tokens_per_sec": 106343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:39.911926+00:00", "epoch": 0, "step": 982, "train_loss": 4.040572166442871, "perplexity": 56.858866235413736, "lr": 0.001305, "grad_norm": 0.273844, "tokens_per_sec": 106510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:40.219077+00:00", "epoch": 0, "step": 983, "train_loss": 4.018616676330566, "perplexity": 55.62410643850869, "lr": 0.001305, "grad_norm": 0.260116, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:40.525968+00:00", "epoch": 0, "step": 984, "train_loss": 4.067449569702148, "perplexity": 58.407807426125494, "lr": 0.001305, "grad_norm": 0.240109, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:40.833521+00:00", "epoch": 0, "step": 985, "train_loss": 4.013698577880859, "perplexity": 55.351213214791926, "lr": 0.001305, "grad_norm": 0.24884, "tokens_per_sec": 106544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:41.140515+00:00", "epoch": 0, "step": 986, "train_loss": 4.059089660644531, "perplexity": 57.92155879686023, "lr": 0.001305, "grad_norm": 0.259124, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:41.448596+00:00", "epoch": 0, "step": 987, "train_loss": 4.050209045410156, "perplexity": 57.40945697461755, "lr": 0.001305, "grad_norm": 0.273215, "tokens_per_sec": 106363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:41.756701+00:00", "epoch": 0, "step": 988, "train_loss": 4.086430072784424, "perplexity": 59.52700486025645, "lr": 0.001305, "grad_norm": 0.281142, "tokens_per_sec": 106358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:42.062741+00:00", "epoch": 0, "step": 989, "train_loss": 4.131570339202881, "perplexity": 62.27564006914183, "lr": 0.001305, "grad_norm": 0.299275, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:42.370104+00:00", "epoch": 0, "step": 990, "train_loss": 3.925708532333374, "perplexity": 50.688980119811795, "lr": 0.001305, "grad_norm": 0.323507, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:42.677307+00:00", "epoch": 0, "step": 991, "train_loss": 4.166865348815918, "perplexity": 64.51290935511469, "lr": 0.001305, "grad_norm": 0.334402, "tokens_per_sec": 106590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:42.984471+00:00", "epoch": 0, "step": 992, "train_loss": 4.064894676208496, "perplexity": 58.258772164632674, "lr": 0.001305, "grad_norm": 0.299228, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:43.292177+00:00", "epoch": 0, "step": 993, "train_loss": 3.997981548309326, "perplexity": 54.48805745055968, "lr": 0.001305, "grad_norm": 0.288528, "tokens_per_sec": 106491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:43.599091+00:00", "epoch": 0, "step": 994, "train_loss": 4.147075176239014, "perplexity": 63.24873808852352, "lr": 0.001305, "grad_norm": 0.305169, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:43.905965+00:00", "epoch": 0, "step": 995, "train_loss": 4.10845947265625, "perplexity": 60.85289976769322, "lr": 0.001305, "grad_norm": 0.337436, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:44.213361+00:00", "epoch": 0, "step": 996, "train_loss": 4.103476524353027, "perplexity": 60.55042714263683, "lr": 0.001305, "grad_norm": 0.303433, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:44.520991+00:00", "epoch": 0, "step": 997, "train_loss": 4.144894123077393, "perplexity": 63.110939555935126, "lr": 0.001305, "grad_norm": 0.302066, "tokens_per_sec": 106517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:44.828718+00:00", "epoch": 0, "step": 998, "train_loss": 4.100291728973389, "perplexity": 60.35789317542047, "lr": 0.001305, "grad_norm": 0.306294, "tokens_per_sec": 106485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:45.137718+00:00", "epoch": 0, "step": 999, "train_loss": 4.099578857421875, "perplexity": 60.31488108333633, "lr": 0.001305, "grad_norm": 0.301317, "tokens_per_sec": 106046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:45.446134+00:00", "epoch": 0, "step": 1000, "train_loss": 4.119718074798584, "perplexity": 61.54188960884449, "lr": 0.001305, "grad_norm": 0.34976, "tokens_per_sec": 106304} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:05:51.321503+00:00", "step": 1000, "epoch": 0, "val_loss": 4.085229611396789, "val_ppl": 59.45558786464963, "eval_train_loss": 4.119718074798584, "eval_train_ppl": 61.54188960884449} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:05:52.305763+00:00", "step": 1000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_4p0852_epoch_0000_step_0001000.pt", "val_loss": 4.085229611396789} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:05:54.041541+00:00", "step": 1000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0000_step_0001000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:55.000592+00:00", "epoch": 0, "step": 1001, "train_loss": 4.175808429718018, "perplexity": 65.09244105965612, "lr": 0.001305, "grad_norm": 0.344176, "tokens_per_sec": 3430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:55.305983+00:00", "epoch": 0, "step": 1002, "train_loss": 4.136459827423096, "perplexity": 62.580881707243314, "lr": 0.001305, "grad_norm": 0.374436, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:55.611175+00:00", "epoch": 0, "step": 1003, "train_loss": 4.074098110198975, "perplexity": 58.797427867569326, "lr": 0.001305, "grad_norm": 0.379657, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:55.917427+00:00", "epoch": 0, "step": 1004, "train_loss": 3.942171812057495, "perplexity": 51.5303941914434, "lr": 0.001305, "grad_norm": 0.336247, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:56.224460+00:00", "epoch": 0, "step": 1005, "train_loss": 4.140922546386719, "perplexity": 62.86078669891007, "lr": 0.001305, "grad_norm": 0.320371, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:56.530724+00:00", "epoch": 0, "step": 1006, "train_loss": 4.085350036621094, "perplexity": 59.462748248291135, "lr": 0.001305, "grad_norm": 0.271938, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:56.836749+00:00", "epoch": 0, "step": 1007, "train_loss": 4.028532028198242, "perplexity": 56.17838240570228, "lr": 0.001305, "grad_norm": 0.273777, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:57.142013+00:00", "epoch": 0, "step": 1008, "train_loss": 4.057186126708984, "perplexity": 57.811408015226476, "lr": 0.001305, "grad_norm": 0.257275, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:57.447907+00:00", "epoch": 0, "step": 1009, "train_loss": 4.019870281219482, "perplexity": 55.69388081589422, "lr": 0.001305, "grad_norm": 0.266593, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:57.754452+00:00", "epoch": 0, "step": 1010, "train_loss": 3.9841508865356445, "perplexity": 53.7396390483498, "lr": 0.001305, "grad_norm": 0.268567, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:58.061099+00:00", "epoch": 0, "step": 1011, "train_loss": 3.996062994003296, "perplexity": 54.38361937040995, "lr": 0.001305, "grad_norm": 0.269212, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:58.366948+00:00", "epoch": 0, "step": 1012, "train_loss": 4.081649303436279, "perplexity": 59.2430991645972, "lr": 0.001305, "grad_norm": 0.281955, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:58.673583+00:00", "epoch": 0, "step": 1013, "train_loss": 3.9913251399993896, "perplexity": 54.12656714044045, "lr": 0.001305, "grad_norm": 0.312406, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:58.979300+00:00", "epoch": 0, "step": 1014, "train_loss": 3.952852725982666, "perplexity": 52.083735732505765, "lr": 0.001305, "grad_norm": 0.334568, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:59.654777+00:00", "epoch": 0, "step": 1015, "train_loss": 4.055360794067383, "perplexity": 57.70597921570655, "lr": 0.001305, "grad_norm": 0.372643, "tokens_per_sec": 48511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:05:59.960754+00:00", "epoch": 0, "step": 1016, "train_loss": 4.066325664520264, "perplexity": 58.34219946416708, "lr": 0.001305, "grad_norm": 0.361004, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:00.267673+00:00", "epoch": 0, "step": 1017, "train_loss": 4.127869606018066, "perplexity": 62.04560046147816, "lr": 0.001305, "grad_norm": 0.327048, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:00.574477+00:00", "epoch": 0, "step": 1018, "train_loss": 4.1605119705200195, "perplexity": 64.10433372912931, "lr": 0.001305, "grad_norm": 0.333838, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:00.880916+00:00", "epoch": 0, "step": 1019, "train_loss": 4.041571617126465, "perplexity": 56.915722275812506, "lr": 0.001305, "grad_norm": 0.33874, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:01.187494+00:00", "epoch": 0, "step": 1020, "train_loss": 4.002892017364502, "perplexity": 54.7562773744065, "lr": 0.001305, "grad_norm": 0.314443, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:01.493140+00:00", "epoch": 0, "step": 1021, "train_loss": 4.109506130218506, "perplexity": 60.9166252589864, "lr": 0.001305, "grad_norm": 0.298307, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:01.800553+00:00", "epoch": 0, "step": 1022, "train_loss": 4.013674736022949, "perplexity": 55.34989355476294, "lr": 0.001305, "grad_norm": 0.291673, "tokens_per_sec": 106593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:02.107676+00:00", "epoch": 0, "step": 1023, "train_loss": 4.040554046630859, "perplexity": 56.85783597278046, "lr": 0.001305, "grad_norm": 0.309553, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:02.415140+00:00", "epoch": 0, "step": 1024, "train_loss": 4.148211479187012, "perplexity": 63.32064866443287, "lr": 0.001305, "grad_norm": 0.289539, "tokens_per_sec": 106575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:02.721440+00:00", "epoch": 0, "step": 1025, "train_loss": 3.995626211166382, "perplexity": 54.35987072573745, "lr": 0.001305, "grad_norm": 0.29875, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:03.027073+00:00", "epoch": 0, "step": 1026, "train_loss": 4.123196125030518, "perplexity": 61.75630805518704, "lr": 0.001305, "grad_norm": 0.319785, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:03.333226+00:00", "epoch": 0, "step": 1027, "train_loss": 4.041658878326416, "perplexity": 56.92068902673355, "lr": 0.001305, "grad_norm": 0.313797, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:03.639756+00:00", "epoch": 0, "step": 1028, "train_loss": 4.038524150848389, "perplexity": 56.742537552817474, "lr": 0.001305, "grad_norm": 0.301579, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:03.945614+00:00", "epoch": 0, "step": 1029, "train_loss": 4.098276615142822, "perplexity": 60.236387614982796, "lr": 0.001305, "grad_norm": 0.314863, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:04.252164+00:00", "epoch": 0, "step": 1030, "train_loss": 4.086469650268555, "perplexity": 59.52936083596815, "lr": 0.001305, "grad_norm": 0.320061, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:04.559591+00:00", "epoch": 0, "step": 1031, "train_loss": 4.043219566345215, "perplexity": 57.00959382239825, "lr": 0.001305, "grad_norm": 0.298318, "tokens_per_sec": 106526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:04.866957+00:00", "epoch": 0, "step": 1032, "train_loss": 4.078709125518799, "perplexity": 59.06916972970869, "lr": 0.001305, "grad_norm": 0.307241, "tokens_per_sec": 106609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:05.174245+00:00", "epoch": 0, "step": 1033, "train_loss": 4.0804033279418945, "perplexity": 59.169329681854215, "lr": 0.001305, "grad_norm": 0.312841, "tokens_per_sec": 106637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:05.481141+00:00", "epoch": 0, "step": 1034, "train_loss": 4.0096917152404785, "perplexity": 55.12987224402983, "lr": 0.001305, "grad_norm": 0.322231, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:05.788974+00:00", "epoch": 0, "step": 1035, "train_loss": 4.075043678283691, "perplexity": 58.853051132471975, "lr": 0.001305, "grad_norm": 0.36292, "tokens_per_sec": 106447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:06.095762+00:00", "epoch": 0, "step": 1036, "train_loss": 3.9888663291931152, "perplexity": 53.993643636091974, "lr": 0.001305, "grad_norm": 0.336624, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:06.402815+00:00", "epoch": 0, "step": 1037, "train_loss": 3.9273810386657715, "perplexity": 50.773828695155984, "lr": 0.001305, "grad_norm": 0.329007, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:06.708586+00:00", "epoch": 0, "step": 1038, "train_loss": 4.099475383758545, "perplexity": 60.30864040451584, "lr": 0.001305, "grad_norm": 0.276888, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:07.015786+00:00", "epoch": 0, "step": 1039, "train_loss": 3.8905913829803467, "perplexity": 48.93982014415777, "lr": 0.001305, "grad_norm": 0.277429, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:07.323724+00:00", "epoch": 0, "step": 1040, "train_loss": 4.090800762176514, "perplexity": 59.78774830817289, "lr": 0.001305, "grad_norm": 0.298488, "tokens_per_sec": 106484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:07.630499+00:00", "epoch": 0, "step": 1041, "train_loss": 4.033548355102539, "perplexity": 56.46089954353634, "lr": 0.001305, "grad_norm": 0.313304, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:07.936408+00:00", "epoch": 0, "step": 1042, "train_loss": 3.9263787269592285, "perplexity": 50.72296298817403, "lr": 0.001305, "grad_norm": 0.307243, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:08.244565+00:00", "epoch": 0, "step": 1043, "train_loss": 4.0732927322387695, "perplexity": 58.7500927789252, "lr": 0.001305, "grad_norm": 0.313359, "tokens_per_sec": 106335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:08.551735+00:00", "epoch": 0, "step": 1044, "train_loss": 4.093949317932129, "perplexity": 59.976290028525774, "lr": 0.001305, "grad_norm": 0.293009, "tokens_per_sec": 106679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:08.858610+00:00", "epoch": 0, "step": 1045, "train_loss": 3.9352307319641113, "perplexity": 51.17395606197558, "lr": 0.001305, "grad_norm": 0.257875, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:09.165590+00:00", "epoch": 0, "step": 1046, "train_loss": 3.9996559619903564, "perplexity": 54.579369425083364, "lr": 0.001305, "grad_norm": 0.256177, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:09.472753+00:00", "epoch": 0, "step": 1047, "train_loss": 4.047033309936523, "perplexity": 57.22742891504812, "lr": 0.001305, "grad_norm": 0.265738, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:09.779771+00:00", "epoch": 0, "step": 1048, "train_loss": 4.090292930603027, "perplexity": 59.75739390998759, "lr": 0.001305, "grad_norm": 0.263064, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:10.087244+00:00", "epoch": 0, "step": 1049, "train_loss": 3.996241331100464, "perplexity": 54.39331885208496, "lr": 0.001305, "grad_norm": 0.265118, "tokens_per_sec": 106573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:10.393041+00:00", "epoch": 0, "step": 1050, "train_loss": 4.015677452087402, "perplexity": 55.46085475054791, "lr": 0.001305, "grad_norm": 0.255435, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:10.699369+00:00", "epoch": 0, "step": 1051, "train_loss": 4.011677265167236, "perplexity": 55.23944410202536, "lr": 0.001305, "grad_norm": 0.250635, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:11.005280+00:00", "epoch": 0, "step": 1052, "train_loss": 4.1312360763549805, "perplexity": 62.25482711502995, "lr": 0.001305, "grad_norm": 0.264482, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:11.312364+00:00", "epoch": 0, "step": 1053, "train_loss": 4.179617404937744, "perplexity": 65.34084934481355, "lr": 0.001305, "grad_norm": 0.271191, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:11.619624+00:00", "epoch": 0, "step": 1054, "train_loss": 3.922513723373413, "perplexity": 50.527296922939136, "lr": 0.001305, "grad_norm": 0.270072, "tokens_per_sec": 106646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:11.926407+00:00", "epoch": 0, "step": 1055, "train_loss": 4.141220569610596, "perplexity": 62.87952346507457, "lr": 0.001305, "grad_norm": 0.277328, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:12.233314+00:00", "epoch": 0, "step": 1056, "train_loss": 4.082076072692871, "perplexity": 59.26838769378545, "lr": 0.001305, "grad_norm": 0.267922, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:12.540836+00:00", "epoch": 0, "step": 1057, "train_loss": 4.015110015869141, "perplexity": 55.429393179928205, "lr": 0.001305, "grad_norm": 0.265157, "tokens_per_sec": 106555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:12.847577+00:00", "epoch": 0, "step": 1058, "train_loss": 4.197315216064453, "perplexity": 66.50753277538054, "lr": 0.001305, "grad_norm": 0.275639, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:13.154746+00:00", "epoch": 0, "step": 1059, "train_loss": 4.040099143981934, "perplexity": 56.83197707467081, "lr": 0.001305, "grad_norm": 0.258414, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:13.461170+00:00", "epoch": 0, "step": 1060, "train_loss": 4.063260555267334, "perplexity": 58.16364802837149, "lr": 0.001305, "grad_norm": 0.298675, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:13.767976+00:00", "epoch": 0, "step": 1061, "train_loss": 3.9149951934814453, "perplexity": 50.14883046917339, "lr": 0.001305, "grad_norm": 0.318322, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:14.073485+00:00", "epoch": 0, "step": 1062, "train_loss": 4.09365177154541, "perplexity": 59.95844695484168, "lr": 0.001305, "grad_norm": 0.347827, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:14.380864+00:00", "epoch": 0, "step": 1063, "train_loss": 3.9932949542999268, "perplexity": 54.233291505513904, "lr": 0.001305, "grad_norm": 0.299093, "tokens_per_sec": 106605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:14.689081+00:00", "epoch": 0, "step": 1064, "train_loss": 3.9376380443573, "perplexity": 51.29729616012092, "lr": 0.001305, "grad_norm": 0.309054, "tokens_per_sec": 106315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:14.997449+00:00", "epoch": 0, "step": 1065, "train_loss": 4.013996124267578, "perplexity": 55.3676852187556, "lr": 0.001305, "grad_norm": 0.312817, "tokens_per_sec": 106263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:15.304342+00:00", "epoch": 0, "step": 1066, "train_loss": 4.0129075050354, "perplexity": 55.307443687779205, "lr": 0.001305, "grad_norm": 0.300681, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:15.610802+00:00", "epoch": 0, "step": 1067, "train_loss": 4.048673152923584, "perplexity": 57.32134989980927, "lr": 0.001305, "grad_norm": 0.298034, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:15.917479+00:00", "epoch": 0, "step": 1068, "train_loss": 3.9818856716156006, "perplexity": 53.61804498651484, "lr": 0.001305, "grad_norm": 0.334692, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:16.224196+00:00", "epoch": 0, "step": 1069, "train_loss": 4.105094909667969, "perplexity": 60.648500403494594, "lr": 0.001305, "grad_norm": 0.351041, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:16.532344+00:00", "epoch": 0, "step": 1070, "train_loss": 4.114541530609131, "perplexity": 61.22413843383541, "lr": 0.001305, "grad_norm": 0.35362, "tokens_per_sec": 106397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:16.840533+00:00", "epoch": 0, "step": 1071, "train_loss": 4.065068244934082, "perplexity": 58.26888494307713, "lr": 0.001305, "grad_norm": 0.303477, "tokens_per_sec": 106266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:17.147447+00:00", "epoch": 0, "step": 1072, "train_loss": 4.062060356140137, "perplexity": 58.09388194374488, "lr": 0.001305, "grad_norm": 0.358275, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:17.453959+00:00", "epoch": 0, "step": 1073, "train_loss": 4.050140380859375, "perplexity": 57.405515115378336, "lr": 0.001305, "grad_norm": 0.334166, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:17.761105+00:00", "epoch": 0, "step": 1074, "train_loss": 3.952265501022339, "perplexity": 52.053159841198465, "lr": 0.001305, "grad_norm": 0.32521, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:18.068795+00:00", "epoch": 0, "step": 1075, "train_loss": 4.038602352142334, "perplexity": 56.746975066183225, "lr": 0.001305, "grad_norm": 0.361096, "tokens_per_sec": 106499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:18.375719+00:00", "epoch": 0, "step": 1076, "train_loss": 3.946882963180542, "perplexity": 51.773734422133266, "lr": 0.001305, "grad_norm": 0.435308, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:18.682460+00:00", "epoch": 0, "step": 1077, "train_loss": 3.992971658706665, "perplexity": 54.215760955289596, "lr": 0.001305, "grad_norm": 0.426653, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:18.989314+00:00", "epoch": 0, "step": 1078, "train_loss": 3.8808445930480957, "perplexity": 48.46513110301964, "lr": 0.001305, "grad_norm": 0.350254, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:19.295687+00:00", "epoch": 0, "step": 1079, "train_loss": 3.944791078567505, "perplexity": 51.66554294523268, "lr": 0.001305, "grad_norm": 0.331953, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:19.603094+00:00", "epoch": 0, "step": 1080, "train_loss": 4.175220489501953, "perplexity": 65.05418184396322, "lr": 0.001305, "grad_norm": 0.347979, "tokens_per_sec": 106658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:19.910429+00:00", "epoch": 0, "step": 1081, "train_loss": 3.9233102798461914, "perplexity": 50.56756080245306, "lr": 0.001305, "grad_norm": 0.334302, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:20.216912+00:00", "epoch": 0, "step": 1082, "train_loss": 3.9936976432800293, "perplexity": 54.25513505214044, "lr": 0.001305, "grad_norm": 0.307717, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:20.523435+00:00", "epoch": 0, "step": 1083, "train_loss": 4.0445966720581055, "perplexity": 57.08815614163437, "lr": 0.001305, "grad_norm": 0.313016, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:20.829973+00:00", "epoch": 0, "step": 1084, "train_loss": 4.04019021987915, "perplexity": 56.837153333686125, "lr": 0.001305, "grad_norm": 0.311645, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:21.136866+00:00", "epoch": 0, "step": 1085, "train_loss": 4.042624473571777, "perplexity": 56.97567791765375, "lr": 0.001305, "grad_norm": 0.276221, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:21.442954+00:00", "epoch": 0, "step": 1086, "train_loss": 4.022892475128174, "perplexity": 55.862453124042055, "lr": 0.001305, "grad_norm": 0.270867, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:21.751274+00:00", "epoch": 0, "step": 1087, "train_loss": 4.102748394012451, "perplexity": 60.50635458672918, "lr": 0.001305, "grad_norm": 0.296449, "tokens_per_sec": 106279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:22.058317+00:00", "epoch": 0, "step": 1088, "train_loss": 4.052631855010986, "perplexity": 57.548717791254056, "lr": 0.001305, "grad_norm": 0.274466, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:22.365476+00:00", "epoch": 0, "step": 1089, "train_loss": 3.9754726886749268, "perplexity": 53.27529358283177, "lr": 0.001305, "grad_norm": 0.261026, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:22.671716+00:00", "epoch": 0, "step": 1090, "train_loss": 4.056843280792236, "perplexity": 57.791591007331235, "lr": 0.001305, "grad_norm": 0.260583, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:22.977925+00:00", "epoch": 0, "step": 1091, "train_loss": 3.953531503677368, "perplexity": 52.119101011801284, "lr": 0.001305, "grad_norm": 0.241263, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:23.284712+00:00", "epoch": 0, "step": 1092, "train_loss": 3.896726608276367, "perplexity": 49.24099992486872, "lr": 0.001305, "grad_norm": 0.2532, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:23.593137+00:00", "epoch": 0, "step": 1093, "train_loss": 4.076786041259766, "perplexity": 58.95568389559288, "lr": 0.001305, "grad_norm": 0.270804, "tokens_per_sec": 106244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:23.901382+00:00", "epoch": 0, "step": 1094, "train_loss": 3.9421770572662354, "perplexity": 51.53066447982626, "lr": 0.001305, "grad_norm": 0.280349, "tokens_per_sec": 106306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:24.208272+00:00", "epoch": 0, "step": 1095, "train_loss": 4.027419567108154, "perplexity": 56.115920890616145, "lr": 0.001305, "grad_norm": 0.280926, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:24.514805+00:00", "epoch": 0, "step": 1096, "train_loss": 4.068047046661377, "perplexity": 58.442715172550315, "lr": 0.001305, "grad_norm": 0.272082, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:24.821584+00:00", "epoch": 0, "step": 1097, "train_loss": 4.019339084625244, "perplexity": 55.66430427225999, "lr": 0.001305, "grad_norm": 0.247425, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:25.128651+00:00", "epoch": 0, "step": 1098, "train_loss": 4.042336940765381, "perplexity": 56.95929789609538, "lr": 0.001305, "grad_norm": 0.23416, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:25.435881+00:00", "epoch": 0, "step": 1099, "train_loss": 4.160254001617432, "perplexity": 64.08779893732883, "lr": 0.001305, "grad_norm": 0.260132, "tokens_per_sec": 106657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:25.744284+00:00", "epoch": 0, "step": 1100, "train_loss": 3.9835762977600098, "perplexity": 53.70876972437879, "lr": 0.001305, "grad_norm": 0.289332, "tokens_per_sec": 106308} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:06:31.649256+00:00", "step": 1100, "epoch": 0, "val_loss": 4.0298722863197325, "val_ppl": 56.25372641790752, "eval_train_loss": 3.9835762977600098, "eval_train_ppl": 53.70876972437879} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:06:32.545265+00:00", "step": 1100, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_4p0299_epoch_0000_step_0001100.pt", "val_loss": 4.0298722863197325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:33.508755+00:00", "epoch": 0, "step": 1101, "train_loss": 4.065237045288086, "perplexity": 58.27872158167362, "lr": 0.001305, "grad_norm": 0.275023, "tokens_per_sec": 4220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:33.814096+00:00", "epoch": 0, "step": 1102, "train_loss": 3.9240753650665283, "perplexity": 50.60626409962389, "lr": 0.001305, "grad_norm": 0.28874, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:34.119239+00:00", "epoch": 0, "step": 1103, "train_loss": 3.972588539123535, "perplexity": 53.12186103630809, "lr": 0.001305, "grad_norm": 0.291523, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:34.424245+00:00", "epoch": 0, "step": 1104, "train_loss": 4.029563903808594, "perplexity": 56.23638142707723, "lr": 0.001305, "grad_norm": 0.262309, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:34.729178+00:00", "epoch": 0, "step": 1105, "train_loss": 3.9741370677948, "perplexity": 53.204185485628045, "lr": 0.001305, "grad_norm": 0.297561, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:35.034383+00:00", "epoch": 0, "step": 1106, "train_loss": 4.025834083557129, "perplexity": 56.027020514758064, "lr": 0.001305, "grad_norm": 0.335932, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:35.340867+00:00", "epoch": 0, "step": 1107, "train_loss": 4.0179314613342285, "perplexity": 55.58600502194092, "lr": 0.001305, "grad_norm": 0.347863, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:35.646840+00:00", "epoch": 0, "step": 1108, "train_loss": 3.9555633068084717, "perplexity": 52.225104416985495, "lr": 0.001305, "grad_norm": 0.303484, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:35.953779+00:00", "epoch": 0, "step": 1109, "train_loss": 4.073945045471191, "perplexity": 58.788428744020194, "lr": 0.001305, "grad_norm": 0.310347, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:36.260045+00:00", "epoch": 0, "step": 1110, "train_loss": 3.9808261394500732, "perplexity": 53.56126502860661, "lr": 0.001305, "grad_norm": 0.32441, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:36.565502+00:00", "epoch": 0, "step": 1111, "train_loss": 4.019888877868652, "perplexity": 55.694916545087175, "lr": 0.001305, "grad_norm": 0.29369, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:36.871588+00:00", "epoch": 0, "step": 1112, "train_loss": 3.9627673625946045, "perplexity": 52.60269544301996, "lr": 0.001305, "grad_norm": 0.268839, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:37.179169+00:00", "epoch": 0, "step": 1113, "train_loss": 3.957451105117798, "perplexity": 52.32378799885053, "lr": 0.001305, "grad_norm": 0.283282, "tokens_per_sec": 106534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:37.486317+00:00", "epoch": 0, "step": 1114, "train_loss": 3.991145610809326, "perplexity": 54.116850713898, "lr": 0.001305, "grad_norm": 0.27654, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:37.798847+00:00", "epoch": 0, "step": 1115, "train_loss": 3.96671199798584, "perplexity": 52.81060368866062, "lr": 0.001305, "grad_norm": 0.300709, "tokens_per_sec": 104848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:38.105089+00:00", "epoch": 0, "step": 1116, "train_loss": 3.975013256072998, "perplexity": 53.25082279785114, "lr": 0.001305, "grad_norm": 0.271338, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:38.411629+00:00", "epoch": 0, "step": 1117, "train_loss": 4.0641608238220215, "perplexity": 58.216034509129805, "lr": 0.001305, "grad_norm": 0.26441, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:38.718981+00:00", "epoch": 0, "step": 1118, "train_loss": 3.956299066543579, "perplexity": 52.263543685276154, "lr": 0.001305, "grad_norm": 0.295284, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:39.027165+00:00", "epoch": 0, "step": 1119, "train_loss": 4.098645210266113, "perplexity": 60.258594546134184, "lr": 0.001305, "grad_norm": 0.325217, "tokens_per_sec": 106327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:39.335323+00:00", "epoch": 0, "step": 1120, "train_loss": 3.9792821407318115, "perplexity": 53.47863031441666, "lr": 0.001305, "grad_norm": 0.315927, "tokens_per_sec": 106392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:39.642044+00:00", "epoch": 0, "step": 1121, "train_loss": 4.007856369018555, "perplexity": 55.02878263688254, "lr": 0.001305, "grad_norm": 0.284538, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:39.948639+00:00", "epoch": 0, "step": 1122, "train_loss": 3.9406521320343018, "perplexity": 51.452143953524505, "lr": 0.001305, "grad_norm": 0.284085, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:40.255164+00:00", "epoch": 0, "step": 1123, "train_loss": 4.074721813201904, "perplexity": 58.83411143853238, "lr": 0.001305, "grad_norm": 0.338222, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:40.562237+00:00", "epoch": 0, "step": 1124, "train_loss": 3.9503707885742188, "perplexity": 51.954627446038266, "lr": 0.001305, "grad_norm": 0.3418, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:40.868987+00:00", "epoch": 0, "step": 1125, "train_loss": 4.009263515472412, "perplexity": 55.106270698970086, "lr": 0.001305, "grad_norm": 0.331123, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:41.175065+00:00", "epoch": 0, "step": 1126, "train_loss": 3.9580438137054443, "perplexity": 52.35480994991587, "lr": 0.001305, "grad_norm": 0.32416, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:41.481404+00:00", "epoch": 0, "step": 1127, "train_loss": 4.062933444976807, "perplexity": 58.14462521201592, "lr": 0.001305, "grad_norm": 0.325259, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:41.787453+00:00", "epoch": 0, "step": 1128, "train_loss": 4.118834495544434, "perplexity": 61.487536488106024, "lr": 0.001305, "grad_norm": 0.312133, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:42.094930+00:00", "epoch": 0, "step": 1129, "train_loss": 3.9882400035858154, "perplexity": 53.95983662265554, "lr": 0.001305, "grad_norm": 0.316421, "tokens_per_sec": 106571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:42.402588+00:00", "epoch": 0, "step": 1130, "train_loss": 4.033890247344971, "perplexity": 56.480206387332984, "lr": 0.001305, "grad_norm": 0.325733, "tokens_per_sec": 106564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:42.710252+00:00", "epoch": 0, "step": 1131, "train_loss": 3.8935468196868896, "perplexity": 49.08467263076564, "lr": 0.001305, "grad_norm": 0.341946, "tokens_per_sec": 106450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:43.016828+00:00", "epoch": 0, "step": 1132, "train_loss": 3.9750757217407227, "perplexity": 53.25414924994754, "lr": 0.001305, "grad_norm": 0.286998, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:43.322808+00:00", "epoch": 0, "step": 1133, "train_loss": 4.043927192687988, "perplexity": 57.049949589495775, "lr": 0.001305, "grad_norm": 0.281388, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:43.628801+00:00", "epoch": 0, "step": 1134, "train_loss": 3.8860692977905273, "perplexity": 48.719009746290304, "lr": 0.001305, "grad_norm": 0.269034, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:43.936562+00:00", "epoch": 0, "step": 1135, "train_loss": 3.999072790145874, "perplexity": 54.54754955267369, "lr": 0.001305, "grad_norm": 0.269742, "tokens_per_sec": 106473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:44.243091+00:00", "epoch": 0, "step": 1136, "train_loss": 3.9206602573394775, "perplexity": 50.433733029840106, "lr": 0.001305, "grad_norm": 0.271073, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:44.550579+00:00", "epoch": 0, "step": 1137, "train_loss": 3.9546356201171875, "perplexity": 52.17667834824963, "lr": 0.001305, "grad_norm": 0.316953, "tokens_per_sec": 106568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:44.857525+00:00", "epoch": 0, "step": 1138, "train_loss": 3.97786283493042, "perplexity": 53.402781623137365, "lr": 0.001305, "grad_norm": 0.3152, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:45.164150+00:00", "epoch": 0, "step": 1139, "train_loss": 3.965411901473999, "perplexity": 52.74198941926755, "lr": 0.001305, "grad_norm": 0.302934, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:45.471137+00:00", "epoch": 0, "step": 1140, "train_loss": 4.000321865081787, "perplexity": 54.615726099579376, "lr": 0.001305, "grad_norm": 0.309345, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:45.778467+00:00", "epoch": 0, "step": 1141, "train_loss": 3.877406120300293, "perplexity": 48.29877124632733, "lr": 0.001305, "grad_norm": 0.324341, "tokens_per_sec": 106563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:46.085649+00:00", "epoch": 0, "step": 1142, "train_loss": 3.9834272861480713, "perplexity": 53.70076709028444, "lr": 0.001305, "grad_norm": 0.294317, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:46.392921+00:00", "epoch": 0, "step": 1143, "train_loss": 3.9570212364196777, "perplexity": 52.301300473910565, "lr": 0.001305, "grad_norm": 0.304733, "tokens_per_sec": 106642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:46.699326+00:00", "epoch": 0, "step": 1144, "train_loss": 4.048433303833008, "perplexity": 57.307603074812846, "lr": 0.001305, "grad_norm": 0.301958, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:47.005962+00:00", "epoch": 0, "step": 1145, "train_loss": 3.9733192920684814, "perplexity": 53.16069417969027, "lr": 0.001305, "grad_norm": 0.324665, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:47.312296+00:00", "epoch": 0, "step": 1146, "train_loss": 3.9180920124053955, "perplexity": 50.30437303565784, "lr": 0.001305, "grad_norm": 0.280077, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:47.620808+00:00", "epoch": 0, "step": 1147, "train_loss": 3.957914352416992, "perplexity": 52.34803246748335, "lr": 0.001305, "grad_norm": 0.285996, "tokens_per_sec": 106213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:47.928616+00:00", "epoch": 0, "step": 1148, "train_loss": 4.000593662261963, "perplexity": 54.63057251744235, "lr": 0.001305, "grad_norm": 0.275417, "tokens_per_sec": 106456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:48.234836+00:00", "epoch": 0, "step": 1149, "train_loss": 4.01425838470459, "perplexity": 55.38220787635341, "lr": 0.001305, "grad_norm": 0.252783, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:48.542382+00:00", "epoch": 0, "step": 1150, "train_loss": 4.038933753967285, "perplexity": 56.76578423380471, "lr": 0.001305, "grad_norm": 0.251578, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:48.849461+00:00", "epoch": 0, "step": 1151, "train_loss": 3.974726915359497, "perplexity": 53.235577102092556, "lr": 0.001305, "grad_norm": 0.264006, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:49.156966+00:00", "epoch": 0, "step": 1152, "train_loss": 3.9231770038604736, "perplexity": 50.56082181002467, "lr": 0.001305, "grad_norm": 0.301932, "tokens_per_sec": 106560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:49.463984+00:00", "epoch": 0, "step": 1153, "train_loss": 3.887131929397583, "perplexity": 48.77080762206703, "lr": 0.001305, "grad_norm": 0.321907, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:49.771205+00:00", "epoch": 0, "step": 1154, "train_loss": 4.078949451446533, "perplexity": 59.08336728867691, "lr": 0.001305, "grad_norm": 0.346813, "tokens_per_sec": 106660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:50.077109+00:00", "epoch": 0, "step": 1155, "train_loss": 3.9534287452697754, "perplexity": 52.11374561113708, "lr": 0.001305, "grad_norm": 0.333901, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:50.383284+00:00", "epoch": 0, "step": 1156, "train_loss": 3.8315320014953613, "perplexity": 46.13316019433103, "lr": 0.001305, "grad_norm": 0.291047, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:50.690490+00:00", "epoch": 0, "step": 1157, "train_loss": 4.051914215087891, "perplexity": 57.50743334929778, "lr": 0.001305, "grad_norm": 0.301999, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:50.997207+00:00", "epoch": 0, "step": 1158, "train_loss": 4.074249267578125, "perplexity": 58.8063162044185, "lr": 0.001305, "grad_norm": 0.316312, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:51.304208+00:00", "epoch": 0, "step": 1159, "train_loss": 3.910815954208374, "perplexity": 49.93968384867666, "lr": 0.001305, "grad_norm": 0.297192, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:51.611680+00:00", "epoch": 0, "step": 1160, "train_loss": 3.971980094909668, "perplexity": 53.08954917830904, "lr": 0.001305, "grad_norm": 0.27865, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:51.918703+00:00", "epoch": 0, "step": 1161, "train_loss": 3.9942257404327393, "perplexity": 54.28379460132847, "lr": 0.001305, "grad_norm": 0.251298, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:52.224824+00:00", "epoch": 0, "step": 1162, "train_loss": 4.018632888793945, "perplexity": 55.62500824960758, "lr": 0.001305, "grad_norm": 0.273548, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:52.531812+00:00", "epoch": 0, "step": 1163, "train_loss": 3.9964544773101807, "perplexity": 54.40491381749976, "lr": 0.001305, "grad_norm": 0.316375, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:52.838486+00:00", "epoch": 0, "step": 1164, "train_loss": 3.94421124458313, "perplexity": 51.63559419110365, "lr": 0.001305, "grad_norm": 0.317718, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:53.146066+00:00", "epoch": 0, "step": 1165, "train_loss": 4.024001598358154, "perplexity": 55.92444584091176, "lr": 0.001305, "grad_norm": 0.296986, "tokens_per_sec": 106536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:53.453480+00:00", "epoch": 0, "step": 1166, "train_loss": 3.9986233711242676, "perplexity": 54.523040354185234, "lr": 0.001305, "grad_norm": 0.27672, "tokens_per_sec": 106592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:53.760188+00:00", "epoch": 0, "step": 1167, "train_loss": 4.014675140380859, "perplexity": 55.40529353605595, "lr": 0.001305, "grad_norm": 0.271594, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:54.065994+00:00", "epoch": 0, "step": 1168, "train_loss": 4.034335613250732, "perplexity": 56.50536634788673, "lr": 0.001305, "grad_norm": 0.275868, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:54.373164+00:00", "epoch": 0, "step": 1169, "train_loss": 3.88932466506958, "perplexity": 48.87786644464599, "lr": 0.001305, "grad_norm": 0.296926, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:54.681066+00:00", "epoch": 0, "step": 1170, "train_loss": 4.046751976013184, "perplexity": 57.21133116247692, "lr": 0.001305, "grad_norm": 0.311264, "tokens_per_sec": 106484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:54.989280+00:00", "epoch": 0, "step": 1171, "train_loss": 3.989255905151367, "perplexity": 54.01468235937532, "lr": 0.001305, "grad_norm": 0.354274, "tokens_per_sec": 106255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:55.296503+00:00", "epoch": 0, "step": 1172, "train_loss": 4.021414279937744, "perplexity": 55.77993851588191, "lr": 0.001305, "grad_norm": 0.339619, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:55.603648+00:00", "epoch": 0, "step": 1173, "train_loss": 4.008979797363281, "perplexity": 55.09063826975262, "lr": 0.001305, "grad_norm": 0.263562, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:55.909973+00:00", "epoch": 0, "step": 1174, "train_loss": 4.0093674659729, "perplexity": 55.1119993211307, "lr": 0.001305, "grad_norm": 0.242684, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:56.216608+00:00", "epoch": 0, "step": 1175, "train_loss": 3.994229793548584, "perplexity": 54.28401462028236, "lr": 0.001305, "grad_norm": 0.267153, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:56.524409+00:00", "epoch": 0, "step": 1176, "train_loss": 3.907608985900879, "perplexity": 49.779785396962716, "lr": 0.001305, "grad_norm": 0.280323, "tokens_per_sec": 106459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:56.832824+00:00", "epoch": 0, "step": 1177, "train_loss": 3.9769110679626465, "perplexity": 53.35197879966089, "lr": 0.001305, "grad_norm": 0.296596, "tokens_per_sec": 106247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:57.139299+00:00", "epoch": 0, "step": 1178, "train_loss": 4.002623558044434, "perplexity": 54.741579514390125, "lr": 0.001305, "grad_norm": 0.308101, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:57.446282+00:00", "epoch": 0, "step": 1179, "train_loss": 3.9723052978515625, "perplexity": 53.10681686348451, "lr": 0.001305, "grad_norm": 0.267509, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:57.753089+00:00", "epoch": 0, "step": 1180, "train_loss": 3.9363627433776855, "perplexity": 51.231918365120194, "lr": 0.001305, "grad_norm": 0.280544, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:58.060973+00:00", "epoch": 0, "step": 1181, "train_loss": 3.9820985794067383, "perplexity": 53.62946190136997, "lr": 0.001305, "grad_norm": 0.274586, "tokens_per_sec": 106372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:58.368602+00:00", "epoch": 0, "step": 1182, "train_loss": 3.971637010574341, "perplexity": 53.07133810976119, "lr": 0.001305, "grad_norm": 0.304022, "tokens_per_sec": 106517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:58.677122+00:00", "epoch": 0, "step": 1183, "train_loss": 4.032461643218994, "perplexity": 56.39957613951764, "lr": 0.001305, "grad_norm": 0.291643, "tokens_per_sec": 106211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:58.983862+00:00", "epoch": 0, "step": 1184, "train_loss": 3.927835702896118, "perplexity": 50.79691898766869, "lr": 0.001305, "grad_norm": 0.297748, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:59.290845+00:00", "epoch": 0, "step": 1185, "train_loss": 3.976775646209717, "perplexity": 53.34475427035983, "lr": 0.001305, "grad_norm": 0.281607, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:59.597405+00:00", "epoch": 0, "step": 1186, "train_loss": 3.9740099906921387, "perplexity": 53.19742488145518, "lr": 0.001305, "grad_norm": 0.262197, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:06:59.904523+00:00", "epoch": 0, "step": 1187, "train_loss": 3.998835802078247, "perplexity": 54.53462396597683, "lr": 0.001305, "grad_norm": 0.286632, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:00.212337+00:00", "epoch": 0, "step": 1188, "train_loss": 4.024345874786377, "perplexity": 55.94370262401425, "lr": 0.001305, "grad_norm": 0.300402, "tokens_per_sec": 106454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:00.520705+00:00", "epoch": 0, "step": 1189, "train_loss": 3.91267466545105, "perplexity": 50.03259361997132, "lr": 0.001305, "grad_norm": 0.289477, "tokens_per_sec": 106262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:00.828289+00:00", "epoch": 0, "step": 1190, "train_loss": 3.9555208683013916, "perplexity": 52.22288810855068, "lr": 0.001305, "grad_norm": 0.264439, "tokens_per_sec": 106596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:01.134971+00:00", "epoch": 0, "step": 1191, "train_loss": 3.91196608543396, "perplexity": 49.997154081286304, "lr": 0.001305, "grad_norm": 0.240989, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:01.442240+00:00", "epoch": 0, "step": 1192, "train_loss": 3.968780279159546, "perplexity": 52.91994390021463, "lr": 0.001305, "grad_norm": 0.256181, "tokens_per_sec": 106643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:01.749417+00:00", "epoch": 0, "step": 1193, "train_loss": 4.004902362823486, "perplexity": 54.86646713060414, "lr": 0.001305, "grad_norm": 0.246857, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:02.057380+00:00", "epoch": 0, "step": 1194, "train_loss": 3.93361496925354, "perplexity": 51.091337855700885, "lr": 0.001305, "grad_norm": 0.266385, "tokens_per_sec": 106402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:02.365022+00:00", "epoch": 0, "step": 1195, "train_loss": 3.94195294380188, "perplexity": 51.519117058103824, "lr": 0.001305, "grad_norm": 0.276527, "tokens_per_sec": 106514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:02.684814+00:00", "epoch": 0, "step": 1196, "train_loss": 3.8910462856292725, "perplexity": 48.96208806246281, "lr": 0.001305, "grad_norm": 0.272858, "tokens_per_sec": 102469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:02.991228+00:00", "epoch": 0, "step": 1197, "train_loss": 4.020414352416992, "perplexity": 55.72419049688823, "lr": 0.001305, "grad_norm": 0.276292, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:03.297679+00:00", "epoch": 0, "step": 1198, "train_loss": 4.040043354034424, "perplexity": 56.82880651009657, "lr": 0.001305, "grad_norm": 0.255511, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:03.605513+00:00", "epoch": 0, "step": 1199, "train_loss": 4.0620269775390625, "perplexity": 58.09194288359637, "lr": 0.001305, "grad_norm": 0.240406, "tokens_per_sec": 106446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:03.914424+00:00", "epoch": 0, "step": 1200, "train_loss": 4.015449047088623, "perplexity": 55.448188660640504, "lr": 0.001305, "grad_norm": 0.249083, "tokens_per_sec": 106134} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:07:09.785819+00:00", "step": 1200, "epoch": 0, "val_loss": 3.984813761711121, "val_ppl": 53.77527353031592, "eval_train_loss": 4.015449047088623, "eval_train_ppl": 55.448188660640504} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:07:10.684693+00:00", "step": 1200, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p9848_epoch_0000_step_0001200.pt", "val_loss": 3.984813761711121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:11.888560+00:00", "epoch": 0, "step": 1201, "train_loss": 4.042606353759766, "perplexity": 56.974645538433926, "lr": 0.001305, "grad_norm": 0.249248, "tokens_per_sec": 4109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:12.194204+00:00", "epoch": 0, "step": 1202, "train_loss": 4.000355243682861, "perplexity": 54.61754912653811, "lr": 0.001305, "grad_norm": 0.240098, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:12.500949+00:00", "epoch": 0, "step": 1203, "train_loss": 3.922884225845337, "perplexity": 50.546020879770666, "lr": 0.001305, "grad_norm": 0.27509, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:12.807629+00:00", "epoch": 0, "step": 1204, "train_loss": 3.892256498336792, "perplexity": 49.02137847339059, "lr": 0.001305, "grad_norm": 0.280056, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:13.113730+00:00", "epoch": 0, "step": 1205, "train_loss": 3.8931095600128174, "perplexity": 49.06321457452239, "lr": 0.001305, "grad_norm": 0.302137, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:13.419368+00:00", "epoch": 0, "step": 1206, "train_loss": 3.966620922088623, "perplexity": 52.80579413456767, "lr": 0.001305, "grad_norm": 0.333558, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:13.725502+00:00", "epoch": 0, "step": 1207, "train_loss": 3.9712255001068115, "perplexity": 53.04950319156006, "lr": 0.001305, "grad_norm": 0.301191, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:14.032376+00:00", "epoch": 0, "step": 1208, "train_loss": 3.9165267944335938, "perplexity": 50.22569731531028, "lr": 0.001305, "grad_norm": 0.321708, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:14.339459+00:00", "epoch": 0, "step": 1209, "train_loss": 3.9282867908477783, "perplexity": 50.819838034669836, "lr": 0.001305, "grad_norm": 0.353085, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:14.644845+00:00", "epoch": 0, "step": 1210, "train_loss": 3.970038652420044, "perplexity": 52.98657885959822, "lr": 0.001305, "grad_norm": 0.305178, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:14.950795+00:00", "epoch": 0, "step": 1211, "train_loss": 3.9898152351379395, "perplexity": 54.04490284175723, "lr": 0.001305, "grad_norm": 0.35505, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:15.256641+00:00", "epoch": 0, "step": 1212, "train_loss": 3.8350584506988525, "perplexity": 46.29613363023183, "lr": 0.001305, "grad_norm": 0.465911, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:15.563898+00:00", "epoch": 0, "step": 1213, "train_loss": 3.993906259536743, "perplexity": 54.266454736016506, "lr": 0.001305, "grad_norm": 0.4419, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:15.870644+00:00", "epoch": 0, "step": 1214, "train_loss": 4.002096176147461, "perplexity": 54.71271740768779, "lr": 0.001305, "grad_norm": 0.402718, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:16.176709+00:00", "epoch": 0, "step": 1215, "train_loss": 4.117249965667725, "perplexity": 61.39018479819468, "lr": 0.001305, "grad_norm": 0.382001, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:16.483435+00:00", "epoch": 0, "step": 1216, "train_loss": 3.9734249114990234, "perplexity": 53.166309278463295, "lr": 0.001305, "grad_norm": 0.295867, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:16.790354+00:00", "epoch": 0, "step": 1217, "train_loss": 4.00321102142334, "perplexity": 54.773747635529745, "lr": 0.001305, "grad_norm": 0.301111, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:17.097779+00:00", "epoch": 0, "step": 1218, "train_loss": 3.9852166175842285, "perplexity": 53.79694157934294, "lr": 0.001305, "grad_norm": 0.258927, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:17.404198+00:00", "epoch": 0, "step": 1219, "train_loss": 3.9142253398895264, "perplexity": 50.11023806906115, "lr": 0.001305, "grad_norm": 0.243185, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:17.710773+00:00", "epoch": 0, "step": 1220, "train_loss": 3.9788405895233154, "perplexity": 53.455021973102006, "lr": 0.001305, "grad_norm": 0.257487, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:18.017006+00:00", "epoch": 0, "step": 1221, "train_loss": 3.9305758476257324, "perplexity": 50.936300773435136, "lr": 0.001305, "grad_norm": 0.272127, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:18.324174+00:00", "epoch": 0, "step": 1222, "train_loss": 3.966813564300537, "perplexity": 52.81596773945304, "lr": 0.001305, "grad_norm": 0.28559, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:18.631777+00:00", "epoch": 0, "step": 1223, "train_loss": 3.97586989402771, "perplexity": 53.29645901784774, "lr": 0.001305, "grad_norm": 0.261063, "tokens_per_sec": 106527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:18.937444+00:00", "epoch": 0, "step": 1224, "train_loss": 3.8268918991088867, "perplexity": 45.919593476520134, "lr": 0.001305, "grad_norm": 0.24607, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:19.243836+00:00", "epoch": 0, "step": 1225, "train_loss": 3.9189891815185547, "perplexity": 50.349524816765374, "lr": 0.001305, "grad_norm": 0.272503, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:19.550431+00:00", "epoch": 0, "step": 1226, "train_loss": 4.029341697692871, "perplexity": 56.22388674744654, "lr": 0.001305, "grad_norm": 0.276245, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:19.857262+00:00", "epoch": 0, "step": 1227, "train_loss": 3.9333834648132324, "perplexity": 51.07951135312279, "lr": 0.001305, "grad_norm": 0.248681, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:20.165121+00:00", "epoch": 0, "step": 1228, "train_loss": 3.978090763092041, "perplexity": 53.414955008254125, "lr": 0.001305, "grad_norm": 0.267195, "tokens_per_sec": 106439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:20.471910+00:00", "epoch": 0, "step": 1229, "train_loss": 3.9508719444274902, "perplexity": 51.980671337166534, "lr": 0.001305, "grad_norm": 0.277391, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:20.778097+00:00", "epoch": 0, "step": 1230, "train_loss": 3.9391963481903076, "perplexity": 51.37729524860473, "lr": 0.001305, "grad_norm": 0.276029, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:21.084372+00:00", "epoch": 0, "step": 1231, "train_loss": 3.975619077682495, "perplexity": 53.28309307105302, "lr": 0.001305, "grad_norm": 0.26601, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:21.390514+00:00", "epoch": 0, "step": 1232, "train_loss": 3.9831125736236572, "perplexity": 53.68386944540019, "lr": 0.001305, "grad_norm": 0.281982, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:21.697368+00:00", "epoch": 0, "step": 1233, "train_loss": 3.9811155796051025, "perplexity": 53.57677005324018, "lr": 0.001305, "grad_norm": 0.314614, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:22.003740+00:00", "epoch": 0, "step": 1234, "train_loss": 3.865290403366089, "perplexity": 47.717127635277976, "lr": 0.001305, "grad_norm": 0.271441, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:22.309578+00:00", "epoch": 0, "step": 1235, "train_loss": 3.9591586589813232, "perplexity": 52.41321000992488, "lr": 0.001305, "grad_norm": 0.258188, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:22.615576+00:00", "epoch": 0, "step": 1236, "train_loss": 3.9891796112060547, "perplexity": 54.01056152335274, "lr": 0.001305, "grad_norm": 0.294193, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:22.921662+00:00", "epoch": 0, "step": 1237, "train_loss": 3.9514811038970947, "perplexity": 52.01234550167714, "lr": 0.001305, "grad_norm": 0.288299, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:23.227979+00:00", "epoch": 0, "step": 1238, "train_loss": 3.961827516555786, "perplexity": 52.55328023305876, "lr": 0.001305, "grad_norm": 0.291722, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:23.533818+00:00", "epoch": 0, "step": 1239, "train_loss": 3.95064640045166, "perplexity": 51.968948731918026, "lr": 0.001305, "grad_norm": 0.301623, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:23.840883+00:00", "epoch": 0, "step": 1240, "train_loss": 3.8777377605438232, "perplexity": 48.31479171895556, "lr": 0.001305, "grad_norm": 0.286006, "tokens_per_sec": 106774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:24.147361+00:00", "epoch": 0, "step": 1241, "train_loss": 3.8916730880737305, "perplexity": 48.992787239100515, "lr": 0.001305, "grad_norm": 0.288812, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:24.453370+00:00", "epoch": 0, "step": 1242, "train_loss": 4.078278541564941, "perplexity": 59.043740968056646, "lr": 0.001305, "grad_norm": 0.288972, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:24.759718+00:00", "epoch": 0, "step": 1243, "train_loss": 3.997345447540283, "perplexity": 54.45340857656773, "lr": 0.001305, "grad_norm": 0.31082, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:25.066160+00:00", "epoch": 0, "step": 1244, "train_loss": 3.9953675270080566, "perplexity": 54.34581050698843, "lr": 0.001305, "grad_norm": 0.330697, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:25.373120+00:00", "epoch": 0, "step": 1245, "train_loss": 4.002609729766846, "perplexity": 54.740822537866855, "lr": 0.001305, "grad_norm": 0.312979, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:25.679448+00:00", "epoch": 0, "step": 1246, "train_loss": 3.9236223697662354, "perplexity": 50.58334489136012, "lr": 0.001305, "grad_norm": 0.292745, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:25.985971+00:00", "epoch": 0, "step": 1247, "train_loss": 3.9238667488098145, "perplexity": 50.595707911375676, "lr": 0.001305, "grad_norm": 0.30414, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:26.291959+00:00", "epoch": 0, "step": 1248, "train_loss": 3.9507033824920654, "perplexity": 51.97191011302645, "lr": 0.001305, "grad_norm": 0.293643, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:26.599460+00:00", "epoch": 0, "step": 1249, "train_loss": 3.891249656677246, "perplexity": 48.972046546222366, "lr": 0.001305, "grad_norm": 0.321556, "tokens_per_sec": 106563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:26.907661+00:00", "epoch": 0, "step": 1250, "train_loss": 3.911470890045166, "perplexity": 49.972401850233574, "lr": 0.001305, "grad_norm": 0.297075, "tokens_per_sec": 106378} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:07:28.134806+00:00", "step": 1250, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0000_step_0001250.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:29.126434+00:00", "epoch": 0, "step": 1251, "train_loss": 3.891645669937134, "perplexity": 48.991443966582835, "lr": 0.001305, "grad_norm": 0.322808, "tokens_per_sec": 14767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:29.431957+00:00", "epoch": 0, "step": 1252, "train_loss": 3.9804677963256836, "perplexity": 53.5420751560348, "lr": 0.001305, "grad_norm": 0.340148, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:29.738541+00:00", "epoch": 0, "step": 1253, "train_loss": 3.889078140258789, "perplexity": 48.86581832301032, "lr": 0.001305, "grad_norm": 0.322295, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:30.044642+00:00", "epoch": 0, "step": 1254, "train_loss": 3.800429344177246, "perplexity": 44.72038080719694, "lr": 0.001305, "grad_norm": 0.302156, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:30.350387+00:00", "epoch": 0, "step": 1255, "train_loss": 3.962371826171875, "perplexity": 52.581893275317206, "lr": 0.001305, "grad_norm": 0.263783, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:30.656176+00:00", "epoch": 0, "step": 1256, "train_loss": 3.881155490875244, "perplexity": 48.48020114947299, "lr": 0.001305, "grad_norm": 0.251525, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:30.962858+00:00", "epoch": 0, "step": 1257, "train_loss": 3.9356038570404053, "perplexity": 51.193053910957076, "lr": 0.001305, "grad_norm": 0.262654, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:31.269527+00:00", "epoch": 0, "step": 1258, "train_loss": 3.908952474594116, "perplexity": 49.84670892122819, "lr": 0.001305, "grad_norm": 0.274198, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:31.576250+00:00", "epoch": 0, "step": 1259, "train_loss": 3.9014363288879395, "perplexity": 49.47345825435339, "lr": 0.001305, "grad_norm": 0.268255, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:31.882797+00:00", "epoch": 0, "step": 1260, "train_loss": 3.7619211673736572, "perplexity": 43.03101640311504, "lr": 0.001305, "grad_norm": 0.258674, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:32.189355+00:00", "epoch": 0, "step": 1261, "train_loss": 4.03720235824585, "perplexity": 56.667585233054425, "lr": 0.001305, "grad_norm": 0.279533, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:32.495508+00:00", "epoch": 0, "step": 1262, "train_loss": 4.162903308868408, "perplexity": 64.25781231715136, "lr": 0.001305, "grad_norm": 0.281769, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:32.803837+00:00", "epoch": 0, "step": 1263, "train_loss": 3.8753156661987305, "perplexity": 48.19791034115163, "lr": 0.001305, "grad_norm": 0.301018, "tokens_per_sec": 106276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:33.111311+00:00", "epoch": 0, "step": 1264, "train_loss": 3.943568706512451, "perplexity": 51.60242701276224, "lr": 0.001305, "grad_norm": 0.307486, "tokens_per_sec": 106572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:33.833545+00:00", "epoch": 0, "step": 1265, "train_loss": 3.8443665504455566, "perplexity": 46.72907446231921, "lr": 0.001305, "grad_norm": 0.286419, "tokens_per_sec": 45370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:34.140612+00:00", "epoch": 0, "step": 1266, "train_loss": 3.8905091285705566, "perplexity": 48.93579479369026, "lr": 0.001305, "grad_norm": 0.268265, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:34.446749+00:00", "epoch": 0, "step": 1267, "train_loss": 3.9593939781188965, "perplexity": 52.42554529260899, "lr": 0.001305, "grad_norm": 0.247893, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:34.752958+00:00", "epoch": 0, "step": 1268, "train_loss": 3.970998764038086, "perplexity": 53.03747631927296, "lr": 0.001305, "grad_norm": 0.252058, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:35.060126+00:00", "epoch": 0, "step": 1269, "train_loss": 3.9126293659210205, "perplexity": 50.03032721832804, "lr": 0.001305, "grad_norm": 0.25531, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:35.366252+00:00", "epoch": 0, "step": 1270, "train_loss": 3.9176886081695557, "perplexity": 50.28408413108415, "lr": 0.001305, "grad_norm": 0.303093, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:35.674278+00:00", "epoch": 0, "step": 1271, "train_loss": 4.012296676635742, "perplexity": 55.273670646278745, "lr": 0.001305, "grad_norm": 0.257397, "tokens_per_sec": 106301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:35.980724+00:00", "epoch": 0, "step": 1272, "train_loss": 3.992933750152588, "perplexity": 54.21370575313871, "lr": 0.001305, "grad_norm": 0.25025, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:36.288021+00:00", "epoch": 0, "step": 1273, "train_loss": 3.9297447204589844, "perplexity": 50.8939838179103, "lr": 0.001305, "grad_norm": 0.293197, "tokens_per_sec": 106633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:36.594112+00:00", "epoch": 0, "step": 1274, "train_loss": 4.020823001861572, "perplexity": 55.746966809831314, "lr": 0.001305, "grad_norm": 0.305324, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:36.900107+00:00", "epoch": 0, "step": 1275, "train_loss": 3.895653486251831, "perplexity": 49.1881866659431, "lr": 0.001305, "grad_norm": 0.278993, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:37.206806+00:00", "epoch": 0, "step": 1276, "train_loss": 3.8859915733337402, "perplexity": 48.715223234876746, "lr": 0.001305, "grad_norm": 0.293493, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:37.514335+00:00", "epoch": 0, "step": 1277, "train_loss": 4.030503749847412, "perplexity": 56.289259812277045, "lr": 0.001305, "grad_norm": 0.286657, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:37.821171+00:00", "epoch": 0, "step": 1278, "train_loss": 4.005977153778076, "perplexity": 54.92546881474485, "lr": 0.001305, "grad_norm": 0.275562, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:38.128272+00:00", "epoch": 0, "step": 1279, "train_loss": 3.9648003578186035, "perplexity": 52.70974525062854, "lr": 0.001305, "grad_norm": 0.273557, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:38.434103+00:00", "epoch": 0, "step": 1280, "train_loss": 3.873019218444824, "perplexity": 48.087353350968385, "lr": 0.001305, "grad_norm": 0.2635, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:38.740006+00:00", "epoch": 0, "step": 1281, "train_loss": 3.879960536956787, "perplexity": 48.42230414215906, "lr": 0.001305, "grad_norm": 0.266092, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:39.047606+00:00", "epoch": 0, "step": 1282, "train_loss": 3.8726327419281006, "perplexity": 48.06877230894632, "lr": 0.001305, "grad_norm": 0.276468, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:39.354564+00:00", "epoch": 0, "step": 1283, "train_loss": 4.08735466003418, "perplexity": 59.582068221484626, "lr": 0.001305, "grad_norm": 0.280384, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:39.661394+00:00", "epoch": 0, "step": 1284, "train_loss": 3.7936296463012695, "perplexity": 44.41732723284407, "lr": 0.001305, "grad_norm": 0.256668, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:39.968154+00:00", "epoch": 0, "step": 1285, "train_loss": 3.972723960876465, "perplexity": 53.129055378972765, "lr": 0.001305, "grad_norm": 0.262839, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:40.274404+00:00", "epoch": 0, "step": 1286, "train_loss": 3.9563393592834473, "perplexity": 52.26564956907208, "lr": 0.001305, "grad_norm": 0.272393, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:40.581819+00:00", "epoch": 0, "step": 1287, "train_loss": 3.910008668899536, "perplexity": 49.89938454428443, "lr": 0.001305, "grad_norm": 0.295289, "tokens_per_sec": 106593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:40.891595+00:00", "epoch": 0, "step": 1288, "train_loss": 3.8689587116241455, "perplexity": 47.89249021398406, "lr": 0.001305, "grad_norm": 0.256163, "tokens_per_sec": 105779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:41.200649+00:00", "epoch": 0, "step": 1289, "train_loss": 3.9702858924865723, "perplexity": 52.99968088448658, "lr": 0.001305, "grad_norm": 0.255843, "tokens_per_sec": 106028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:41.508121+00:00", "epoch": 0, "step": 1290, "train_loss": 3.972867965698242, "perplexity": 53.13670677002919, "lr": 0.001305, "grad_norm": 0.289377, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:41.814495+00:00", "epoch": 0, "step": 1291, "train_loss": 3.938405990600586, "perplexity": 51.336704855941036, "lr": 0.001305, "grad_norm": 0.298353, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:42.121173+00:00", "epoch": 0, "step": 1292, "train_loss": 3.9641072750091553, "perplexity": 52.6732256893069, "lr": 0.001305, "grad_norm": 0.303956, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:42.431700+00:00", "epoch": 0, "step": 1293, "train_loss": 3.9900155067443848, "perplexity": 54.05572758517772, "lr": 0.001305, "grad_norm": 0.275313, "tokens_per_sec": 105524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:42.739002+00:00", "epoch": 0, "step": 1294, "train_loss": 3.833019495010376, "perplexity": 46.201834034230465, "lr": 0.001305, "grad_norm": 0.265656, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:43.046299+00:00", "epoch": 0, "step": 1295, "train_loss": 3.8913357257843018, "perplexity": 48.976261707934285, "lr": 0.001305, "grad_norm": 0.290556, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:43.352525+00:00", "epoch": 0, "step": 1296, "train_loss": 3.996345043182373, "perplexity": 54.398960388967765, "lr": 0.001305, "grad_norm": 0.304468, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:43.658577+00:00", "epoch": 0, "step": 1297, "train_loss": 3.9623560905456543, "perplexity": 52.581065872808516, "lr": 0.001305, "grad_norm": 0.328029, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:43.964867+00:00", "epoch": 0, "step": 1298, "train_loss": 3.894336462020874, "perplexity": 49.12344727325947, "lr": 0.001305, "grad_norm": 0.31162, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:44.271284+00:00", "epoch": 0, "step": 1299, "train_loss": 3.8234188556671143, "perplexity": 45.7603893549898, "lr": 0.001305, "grad_norm": 0.299086, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:44.578499+00:00", "epoch": 0, "step": 1300, "train_loss": 3.8599767684936523, "perplexity": 47.464248690050404, "lr": 0.001305, "grad_norm": 0.293634, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:07:50.508274+00:00", "step": 1300, "epoch": 0, "val_loss": 3.956408154964447, "val_ppl": 52.269245343712555, "eval_train_loss": 3.8599767684936523, "eval_train_ppl": 47.464248690050404} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:07:51.409101+00:00", "step": 1300, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p9564_epoch_0000_step_0001300.pt", "val_loss": 3.956408154964447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:52.397475+00:00", "epoch": 0, "step": 1301, "train_loss": 3.742302417755127, "perplexity": 42.195028994101555, "lr": 0.001305, "grad_norm": 0.306396, "tokens_per_sec": 4191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:52.701866+00:00", "epoch": 0, "step": 1302, "train_loss": 3.949939727783203, "perplexity": 51.93223666947409, "lr": 0.001305, "grad_norm": 0.311219, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:53.007717+00:00", "epoch": 0, "step": 1303, "train_loss": 3.9489405155181885, "perplexity": 51.88037125823714, "lr": 0.001305, "grad_norm": 0.306029, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:53.313661+00:00", "epoch": 0, "step": 1304, "train_loss": 3.986130475997925, "perplexity": 53.84612683779445, "lr": 0.001305, "grad_norm": 0.300838, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:53.619797+00:00", "epoch": 0, "step": 1305, "train_loss": 4.065579891204834, "perplexity": 58.298705628929945, "lr": 0.001305, "grad_norm": 0.280484, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:53.926368+00:00", "epoch": 0, "step": 1306, "train_loss": 3.9453346729278564, "perplexity": 51.69363567783594, "lr": 0.001305, "grad_norm": 0.294142, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:54.232154+00:00", "epoch": 0, "step": 1307, "train_loss": 3.9531445503234863, "perplexity": 52.098937252332234, "lr": 0.001305, "grad_norm": 0.270911, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:54.537384+00:00", "epoch": 0, "step": 1308, "train_loss": 3.8853769302368164, "perplexity": 48.685289959283296, "lr": 0.001305, "grad_norm": 0.304701, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:54.843203+00:00", "epoch": 0, "step": 1309, "train_loss": 3.9277522563934326, "perplexity": 50.792680339284615, "lr": 0.001305, "grad_norm": 0.280098, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:55.150482+00:00", "epoch": 0, "step": 1310, "train_loss": 3.8822615146636963, "perplexity": 48.53385106878765, "lr": 0.001305, "grad_norm": 0.292837, "tokens_per_sec": 106697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:55.456614+00:00", "epoch": 0, "step": 1311, "train_loss": 3.940244436264038, "perplexity": 51.4311714075633, "lr": 0.001305, "grad_norm": 0.264741, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:55.762325+00:00", "epoch": 0, "step": 1312, "train_loss": 3.843362331390381, "perplexity": 46.6821717895339, "lr": 0.001305, "grad_norm": 0.27411, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:56.068896+00:00", "epoch": 0, "step": 1313, "train_loss": 3.925168037414551, "perplexity": 50.6615903862894, "lr": 0.001305, "grad_norm": 0.329857, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:56.374922+00:00", "epoch": 0, "step": 1314, "train_loss": 3.942599058151245, "perplexity": 51.55241505490007, "lr": 0.001305, "grad_norm": 0.317781, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:56.730964+00:00", "epoch": 0, "step": 1315, "train_loss": 3.8791630268096924, "perplexity": 48.383702258003105, "lr": 0.001305, "grad_norm": 0.350555, "tokens_per_sec": 92034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:57.037191+00:00", "epoch": 0, "step": 1316, "train_loss": 3.9956274032592773, "perplexity": 54.35993552779177, "lr": 0.001305, "grad_norm": 0.380628, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:57.343954+00:00", "epoch": 0, "step": 1317, "train_loss": 3.9120829105377197, "perplexity": 50.00299534519603, "lr": 0.001305, "grad_norm": 0.356397, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:57.650385+00:00", "epoch": 0, "step": 1318, "train_loss": 3.960561513900757, "perplexity": 52.486789738205424, "lr": 0.001305, "grad_norm": 0.353811, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:57.956372+00:00", "epoch": 0, "step": 1319, "train_loss": 3.8446924686431885, "perplexity": 46.744306800158185, "lr": 0.001305, "grad_norm": 0.312317, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:58.262459+00:00", "epoch": 0, "step": 1320, "train_loss": 3.9340879917144775, "perplexity": 51.1155109228171, "lr": 0.001305, "grad_norm": 0.285279, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:58.568223+00:00", "epoch": 0, "step": 1321, "train_loss": 3.8817832469940186, "perplexity": 48.51064444686734, "lr": 0.001305, "grad_norm": 0.262211, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:58.874494+00:00", "epoch": 0, "step": 1322, "train_loss": 3.9716830253601074, "perplexity": 53.0737802322011, "lr": 0.001305, "grad_norm": 0.265571, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:59.181332+00:00", "epoch": 0, "step": 1323, "train_loss": 3.96697998046875, "perplexity": 52.82475790181698, "lr": 0.001305, "grad_norm": 0.257037, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:59.488178+00:00", "epoch": 0, "step": 1324, "train_loss": 3.895038366317749, "perplexity": 49.15793933562518, "lr": 0.001305, "grad_norm": 0.244889, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:07:59.794674+00:00", "epoch": 0, "step": 1325, "train_loss": 3.9246535301208496, "perplexity": 50.63553133288653, "lr": 0.001305, "grad_norm": 0.229091, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:00.100223+00:00", "epoch": 0, "step": 1326, "train_loss": 3.970543146133423, "perplexity": 53.013316999570996, "lr": 0.001305, "grad_norm": 0.23844, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:00.407069+00:00", "epoch": 0, "step": 1327, "train_loss": 3.9498374462127686, "perplexity": 51.926925230387155, "lr": 0.001305, "grad_norm": 0.238178, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:00.714887+00:00", "epoch": 0, "step": 1328, "train_loss": 3.975191593170166, "perplexity": 53.26032024185884, "lr": 0.001305, "grad_norm": 0.241773, "tokens_per_sec": 106453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:01.021272+00:00", "epoch": 0, "step": 1329, "train_loss": 3.825463056564331, "perplexity": 45.8540284599295, "lr": 0.001305, "grad_norm": 0.246974, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:01.327334+00:00", "epoch": 0, "step": 1330, "train_loss": 3.8283400535583496, "perplexity": 45.98614031354986, "lr": 0.001305, "grad_norm": 0.241995, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:01.633359+00:00", "epoch": 0, "step": 1331, "train_loss": 3.9562227725982666, "perplexity": 52.25955644543545, "lr": 0.001305, "grad_norm": 0.23914, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:01.939224+00:00", "epoch": 0, "step": 1332, "train_loss": 3.9695117473602295, "perplexity": 52.95866731710707, "lr": 0.001305, "grad_norm": 0.256783, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:02.245672+00:00", "epoch": 0, "step": 1333, "train_loss": 3.9048023223876953, "perplexity": 49.6402661729396, "lr": 0.001305, "grad_norm": 0.272858, "tokens_per_sec": 106928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:02.552423+00:00", "epoch": 0, "step": 1334, "train_loss": 4.026869297027588, "perplexity": 56.08505047261926, "lr": 0.001305, "grad_norm": 0.286377, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:02.859479+00:00", "epoch": 0, "step": 1335, "train_loss": 3.765479326248169, "perplexity": 43.184400316348764, "lr": 0.001305, "grad_norm": 0.301345, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:03.166156+00:00", "epoch": 0, "step": 1336, "train_loss": 3.9465062618255615, "perplexity": 51.75423485921052, "lr": 0.001305, "grad_norm": 0.28684, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:03.472490+00:00", "epoch": 0, "step": 1337, "train_loss": 4.004875183105469, "perplexity": 54.86497589576467, "lr": 0.001305, "grad_norm": 0.297222, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:03.778227+00:00", "epoch": 0, "step": 1338, "train_loss": 3.9784188270568848, "perplexity": 53.43248140490848, "lr": 0.001305, "grad_norm": 0.290252, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:04.083908+00:00", "epoch": 0, "step": 1339, "train_loss": 3.9597275257110596, "perplexity": 52.443034623609144, "lr": 0.001305, "grad_norm": 0.297337, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:04.391132+00:00", "epoch": 0, "step": 1340, "train_loss": 3.9673728942871094, "perplexity": 52.84551755725876, "lr": 0.001305, "grad_norm": 0.299019, "tokens_per_sec": 106714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:04.697873+00:00", "epoch": 0, "step": 1341, "train_loss": 3.8968915939331055, "perplexity": 49.24912465379321, "lr": 0.001305, "grad_norm": 0.265864, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:05.004003+00:00", "epoch": 0, "step": 1342, "train_loss": 3.8943653106689453, "perplexity": 49.124864438743465, "lr": 0.001305, "grad_norm": 0.248233, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:05.310143+00:00", "epoch": 0, "step": 1343, "train_loss": 3.885502338409424, "perplexity": 48.69139587538784, "lr": 0.001305, "grad_norm": 0.254225, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:05.616362+00:00", "epoch": 0, "step": 1344, "train_loss": 3.942514181137085, "perplexity": 51.548039625526776, "lr": 0.001305, "grad_norm": 0.261569, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:05.924340+00:00", "epoch": 0, "step": 1345, "train_loss": 3.9248149394989014, "perplexity": 50.64370504214521, "lr": 0.001305, "grad_norm": 0.297406, "tokens_per_sec": 106398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:06.231974+00:00", "epoch": 0, "step": 1346, "train_loss": 3.9600412845611572, "perplexity": 52.45949167148548, "lr": 0.001305, "grad_norm": 0.287978, "tokens_per_sec": 106516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:06.538671+00:00", "epoch": 0, "step": 1347, "train_loss": 3.8157002925872803, "perplexity": 45.40854451827572, "lr": 0.001305, "grad_norm": 0.247351, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:06.844847+00:00", "epoch": 0, "step": 1348, "train_loss": 3.892072916030884, "perplexity": 49.0123798417117, "lr": 0.001305, "grad_norm": 0.231476, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:07.151881+00:00", "epoch": 0, "step": 1349, "train_loss": 3.9396543502807617, "perplexity": 51.4008315466559, "lr": 0.001305, "grad_norm": 0.24074, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:07.458808+00:00", "epoch": 0, "step": 1350, "train_loss": 3.8414125442504883, "perplexity": 46.59124016877646, "lr": 0.001305, "grad_norm": 0.248818, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:07.766433+00:00", "epoch": 0, "step": 1351, "train_loss": 3.8157286643981934, "perplexity": 45.40983285919083, "lr": 0.001305, "grad_norm": 0.283978, "tokens_per_sec": 106455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:08.074227+00:00", "epoch": 0, "step": 1352, "train_loss": 3.956246852874756, "perplexity": 52.26081488515559, "lr": 0.001305, "grad_norm": 0.279191, "tokens_per_sec": 106461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:08.381990+00:00", "epoch": 0, "step": 1353, "train_loss": 3.9375481605529785, "perplexity": 51.2926855712024, "lr": 0.001305, "grad_norm": 0.286903, "tokens_per_sec": 106472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:08.689334+00:00", "epoch": 0, "step": 1354, "train_loss": 3.9181714057922363, "perplexity": 50.308367028752265, "lr": 0.001305, "grad_norm": 0.244634, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:08.996366+00:00", "epoch": 0, "step": 1355, "train_loss": 3.865727663040161, "perplexity": 47.7379969732829, "lr": 0.001305, "grad_norm": 0.240988, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:09.303295+00:00", "epoch": 0, "step": 1356, "train_loss": 3.9643394947052, "perplexity": 52.68545887010397, "lr": 0.001305, "grad_norm": 0.25835, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:09.610687+00:00", "epoch": 0, "step": 1357, "train_loss": 3.938591480255127, "perplexity": 51.34622816680055, "lr": 0.001305, "grad_norm": 0.273511, "tokens_per_sec": 106600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:09.918028+00:00", "epoch": 0, "step": 1358, "train_loss": 3.928586483001709, "perplexity": 50.83507062382257, "lr": 0.001305, "grad_norm": 0.275381, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:10.225703+00:00", "epoch": 0, "step": 1359, "train_loss": 4.016517639160156, "perplexity": 55.50747182454237, "lr": 0.001305, "grad_norm": 0.276419, "tokens_per_sec": 106502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:10.532276+00:00", "epoch": 0, "step": 1360, "train_loss": 3.904574394226074, "perplexity": 49.628953047667316, "lr": 0.001305, "grad_norm": 0.293711, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:10.838943+00:00", "epoch": 0, "step": 1361, "train_loss": 3.884296178817749, "perplexity": 48.63270168561648, "lr": 0.001305, "grad_norm": 0.2788, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:11.146148+00:00", "epoch": 0, "step": 1362, "train_loss": 3.8235793113708496, "perplexity": 45.767732459572706, "lr": 0.001305, "grad_norm": 0.257028, "tokens_per_sec": 106664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:11.453906+00:00", "epoch": 0, "step": 1363, "train_loss": 3.8554434776306152, "perplexity": 47.249566421382625, "lr": 0.001305, "grad_norm": 0.262712, "tokens_per_sec": 106473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:11.761844+00:00", "epoch": 0, "step": 1364, "train_loss": 3.7509658336639404, "perplexity": 42.56217013144182, "lr": 0.001305, "grad_norm": 0.25181, "tokens_per_sec": 106412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:12.068393+00:00", "epoch": 0, "step": 1365, "train_loss": 3.9013192653656006, "perplexity": 49.4676670560435, "lr": 0.001305, "grad_norm": 0.247236, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:12.374928+00:00", "epoch": 0, "step": 1366, "train_loss": 3.9473962783813477, "perplexity": 51.80031748917712, "lr": 0.001305, "grad_norm": 0.244049, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:12.681708+00:00", "epoch": 0, "step": 1367, "train_loss": 3.8508355617523193, "perplexity": 47.03234524682922, "lr": 0.001305, "grad_norm": 0.262304, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:12.990437+00:00", "epoch": 0, "step": 1368, "train_loss": 3.8856496810913086, "perplexity": 48.69857072480863, "lr": 0.001305, "grad_norm": 0.262613, "tokens_per_sec": 106140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:13.301931+00:00", "epoch": 0, "step": 1369, "train_loss": 3.9083480834960938, "perplexity": 49.81659111647384, "lr": 0.001305, "grad_norm": 0.284178, "tokens_per_sec": 105197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:13.610681+00:00", "epoch": 0, "step": 1370, "train_loss": 3.9838151931762695, "perplexity": 53.72160203600808, "lr": 0.001305, "grad_norm": 0.259918, "tokens_per_sec": 106192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:13.917994+00:00", "epoch": 0, "step": 1371, "train_loss": 3.8870623111724854, "perplexity": 48.76741240318974, "lr": 0.001305, "grad_norm": 0.247189, "tokens_per_sec": 106565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:14.224202+00:00", "epoch": 0, "step": 1372, "train_loss": 3.9110872745513916, "perplexity": 49.95323533914306, "lr": 0.001305, "grad_norm": 0.22319, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:14.531374+00:00", "epoch": 0, "step": 1373, "train_loss": 3.86746883392334, "perplexity": 47.82118938872208, "lr": 0.001305, "grad_norm": 0.263282, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:14.839201+00:00", "epoch": 0, "step": 1374, "train_loss": 3.8689823150634766, "perplexity": 47.89362065481234, "lr": 0.001305, "grad_norm": 0.283542, "tokens_per_sec": 106450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:15.147331+00:00", "epoch": 0, "step": 1375, "train_loss": 3.924271583557129, "perplexity": 50.61619495865768, "lr": 0.001305, "grad_norm": 0.301579, "tokens_per_sec": 106345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:15.454044+00:00", "epoch": 0, "step": 1376, "train_loss": 3.9225573539733887, "perplexity": 50.52950150731235, "lr": 0.001305, "grad_norm": 0.320238, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:15.760128+00:00", "epoch": 0, "step": 1377, "train_loss": 3.9356958866119385, "perplexity": 51.19776540256893, "lr": 0.001305, "grad_norm": 0.313985, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:16.065695+00:00", "epoch": 0, "step": 1378, "train_loss": 3.8931124210357666, "perplexity": 49.06335494570605, "lr": 0.001305, "grad_norm": 0.28347, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:16.371725+00:00", "epoch": 0, "step": 1379, "train_loss": 3.9271090030670166, "perplexity": 50.76001828481261, "lr": 0.001305, "grad_norm": 0.283151, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:16.679225+00:00", "epoch": 0, "step": 1380, "train_loss": 3.7984180450439453, "perplexity": 44.630525137644966, "lr": 0.001305, "grad_norm": 0.294786, "tokens_per_sec": 106620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:16.986348+00:00", "epoch": 0, "step": 1381, "train_loss": 3.8316729068756104, "perplexity": 46.13966106280322, "lr": 0.001305, "grad_norm": 0.286231, "tokens_per_sec": 106637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:17.293359+00:00", "epoch": 0, "step": 1382, "train_loss": 3.941598653793335, "perplexity": 51.500867582675376, "lr": 0.001305, "grad_norm": 0.278415, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:17.599316+00:00", "epoch": 0, "step": 1383, "train_loss": 4.005208969116211, "perplexity": 54.88329211387702, "lr": 0.001305, "grad_norm": 0.266454, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:17.906200+00:00", "epoch": 0, "step": 1384, "train_loss": 3.917719841003418, "perplexity": 50.2856546700558, "lr": 0.001305, "grad_norm": 0.238832, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:18.213289+00:00", "epoch": 0, "step": 1385, "train_loss": 3.8162732124328613, "perplexity": 45.43456742839821, "lr": 0.001305, "grad_norm": 0.263927, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:18.523271+00:00", "epoch": 0, "step": 1386, "train_loss": 3.924065351486206, "perplexity": 50.6057573522707, "lr": 0.001305, "grad_norm": 0.277831, "tokens_per_sec": 105711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:18.830573+00:00", "epoch": 0, "step": 1387, "train_loss": 3.9412567615509033, "perplexity": 51.48326284520343, "lr": 0.001305, "grad_norm": 0.323255, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:19.137601+00:00", "epoch": 0, "step": 1388, "train_loss": 3.8922133445739746, "perplexity": 49.01926306209528, "lr": 0.001305, "grad_norm": 0.32316, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:19.445198+00:00", "epoch": 0, "step": 1389, "train_loss": 3.8814265727996826, "perplexity": 48.493345037152444, "lr": 0.001305, "grad_norm": 0.350065, "tokens_per_sec": 106530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:19.752104+00:00", "epoch": 0, "step": 1390, "train_loss": 4.0469465255737305, "perplexity": 57.222462684594134, "lr": 0.001305, "grad_norm": 0.354007, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:20.074340+00:00", "epoch": 0, "step": 1391, "train_loss": 3.937535285949707, "perplexity": 51.29202520247594, "lr": 0.001305, "grad_norm": 0.34084, "tokens_per_sec": 101638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:20.381556+00:00", "epoch": 0, "step": 1392, "train_loss": 3.93231201171875, "perplexity": 51.02481136208918, "lr": 0.001305, "grad_norm": 0.338725, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:20.689118+00:00", "epoch": 0, "step": 1393, "train_loss": 3.8769266605377197, "perplexity": 48.275619479546805, "lr": 0.001305, "grad_norm": 0.321786, "tokens_per_sec": 106543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:20.995874+00:00", "epoch": 0, "step": 1394, "train_loss": 3.9629123210906982, "perplexity": 52.61032120333788, "lr": 0.001305, "grad_norm": 0.320447, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:21.302962+00:00", "epoch": 0, "step": 1395, "train_loss": 3.7948758602142334, "perplexity": 44.472715229495975, "lr": 0.001305, "grad_norm": 0.273663, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:21.610402+00:00", "epoch": 0, "step": 1396, "train_loss": 3.8826003074645996, "perplexity": 48.550296773815795, "lr": 0.001305, "grad_norm": 0.258769, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:21.918063+00:00", "epoch": 0, "step": 1397, "train_loss": 3.9441823959350586, "perplexity": 51.63410459550539, "lr": 0.001305, "grad_norm": 0.269552, "tokens_per_sec": 106506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:22.224335+00:00", "epoch": 0, "step": 1398, "train_loss": 3.860729217529297, "perplexity": 47.499976558219096, "lr": 0.001305, "grad_norm": 0.266914, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:22.545811+00:00", "epoch": 0, "step": 1399, "train_loss": 3.849350929260254, "perplexity": 46.96257130604346, "lr": 0.001305, "grad_norm": 0.303377, "tokens_per_sec": 101930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:22.852806+00:00", "epoch": 0, "step": 1400, "train_loss": 3.777787685394287, "perplexity": 43.71921401266293, "lr": 0.001305, "grad_norm": 0.313731, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:08:28.745440+00:00", "step": 1400, "epoch": 0, "val_loss": 3.930932033061981, "val_ppl": 50.95444677342562, "eval_train_loss": 3.777787685394287, "eval_train_ppl": 43.71921401266293} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:08:29.786065+00:00", "step": 1400, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p9309_epoch_0000_step_0001400.pt", "val_loss": 3.930932033061981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:30.900196+00:00", "epoch": 0, "step": 1401, "train_loss": 3.9809813499450684, "perplexity": 53.56957894425106, "lr": 0.001305, "grad_norm": 0.330306, "tokens_per_sec": 4072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:31.205991+00:00", "epoch": 0, "step": 1402, "train_loss": 3.9455254077911377, "perplexity": 51.70349639673101, "lr": 0.001305, "grad_norm": 0.310314, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:31.511778+00:00", "epoch": 0, "step": 1403, "train_loss": 3.840859889984131, "perplexity": 46.565498434917004, "lr": 0.001305, "grad_norm": 0.258367, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:31.816958+00:00", "epoch": 0, "step": 1404, "train_loss": 4.046140670776367, "perplexity": 57.176368263722395, "lr": 0.001305, "grad_norm": 0.275634, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:32.123296+00:00", "epoch": 0, "step": 1405, "train_loss": 3.9166202545166016, "perplexity": 50.230391632512706, "lr": 0.001305, "grad_norm": 0.294237, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:32.428863+00:00", "epoch": 0, "step": 1406, "train_loss": 3.946789026260376, "perplexity": 51.768871185398474, "lr": 0.001305, "grad_norm": 0.280628, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:32.735957+00:00", "epoch": 0, "step": 1407, "train_loss": 3.80704665184021, "perplexity": 45.01729061413147, "lr": 0.001305, "grad_norm": 0.270404, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:33.041647+00:00", "epoch": 0, "step": 1408, "train_loss": 3.9612247943878174, "perplexity": 52.521614749765796, "lr": 0.001305, "grad_norm": 0.27044, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:33.346893+00:00", "epoch": 0, "step": 1409, "train_loss": 3.9161365032196045, "perplexity": 50.206098491804724, "lr": 0.001305, "grad_norm": 0.282357, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:33.652164+00:00", "epoch": 0, "step": 1410, "train_loss": 3.7953786849975586, "perplexity": 44.49508283591728, "lr": 0.001305, "grad_norm": 0.295559, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:33.959504+00:00", "epoch": 0, "step": 1411, "train_loss": 3.830894708633423, "perplexity": 46.10376922696212, "lr": 0.001305, "grad_norm": 0.264077, "tokens_per_sec": 106527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:34.264915+00:00", "epoch": 0, "step": 1412, "train_loss": 3.816598415374756, "perplexity": 45.4493452861608, "lr": 0.001305, "grad_norm": 0.267544, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:34.571564+00:00", "epoch": 0, "step": 1413, "train_loss": 3.898730516433716, "perplexity": 49.339773299626614, "lr": 0.001305, "grad_norm": 0.27771, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:34.877600+00:00", "epoch": 0, "step": 1414, "train_loss": 3.8201684951782227, "perplexity": 45.61189305745981, "lr": 0.001305, "grad_norm": 0.264383, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:35.247479+00:00", "epoch": 0, "step": 1415, "train_loss": 3.982886552810669, "perplexity": 53.67173714471166, "lr": 0.001305, "grad_norm": 0.269863, "tokens_per_sec": 88589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:35.553189+00:00", "epoch": 0, "step": 1416, "train_loss": 3.9502906799316406, "perplexity": 51.950465598060134, "lr": 0.001305, "grad_norm": 0.28074, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:35.859094+00:00", "epoch": 0, "step": 1417, "train_loss": 3.9159793853759766, "perplexity": 50.19821083753456, "lr": 0.001305, "grad_norm": 0.252394, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:36.165418+00:00", "epoch": 0, "step": 1418, "train_loss": 3.928513526916504, "perplexity": 50.831362031362566, "lr": 0.001305, "grad_norm": 0.24701, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:36.471529+00:00", "epoch": 0, "step": 1419, "train_loss": 3.8111345767974854, "perplexity": 45.20169457792498, "lr": 0.001305, "grad_norm": 0.256972, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:36.778216+00:00", "epoch": 0, "step": 1420, "train_loss": 3.9411470890045166, "perplexity": 51.477616854281656, "lr": 0.001305, "grad_norm": 0.245359, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:37.085496+00:00", "epoch": 0, "step": 1421, "train_loss": 3.81789231300354, "perplexity": 45.50819014766004, "lr": 0.001305, "grad_norm": 0.271932, "tokens_per_sec": 106540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:37.392300+00:00", "epoch": 0, "step": 1422, "train_loss": 3.94575834274292, "perplexity": 51.715541350962106, "lr": 0.001305, "grad_norm": 0.285126, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:37.697948+00:00", "epoch": 0, "step": 1423, "train_loss": 3.8565633296966553, "perplexity": 47.30250858412151, "lr": 0.001305, "grad_norm": 0.286634, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:38.003021+00:00", "epoch": 0, "step": 1424, "train_loss": 3.8766684532165527, "perplexity": 48.26315597031722, "lr": 0.001305, "grad_norm": 0.286951, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:38.309630+00:00", "epoch": 0, "step": 1425, "train_loss": 3.860743522644043, "perplexity": 47.500656055694336, "lr": 0.001305, "grad_norm": 0.290375, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:38.616464+00:00", "epoch": 0, "step": 1426, "train_loss": 3.8619258403778076, "perplexity": 47.55685013679772, "lr": 0.001305, "grad_norm": 0.305749, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:38.923501+00:00", "epoch": 0, "step": 1427, "train_loss": 3.9285011291503906, "perplexity": 50.830731839931374, "lr": 0.001305, "grad_norm": 0.319256, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:39.229501+00:00", "epoch": 0, "step": 1428, "train_loss": 3.9085545539855957, "perplexity": 49.82687783434227, "lr": 0.001305, "grad_norm": 0.303925, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:39.535275+00:00", "epoch": 0, "step": 1429, "train_loss": 3.8086137771606445, "perplexity": 45.08789365756938, "lr": 0.001305, "grad_norm": 0.283588, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:39.842367+00:00", "epoch": 0, "step": 1430, "train_loss": 3.8849661350250244, "perplexity": 48.66529438260809, "lr": 0.001305, "grad_norm": 0.250074, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:40.149320+00:00", "epoch": 0, "step": 1431, "train_loss": 3.8605923652648926, "perplexity": 47.49347652365032, "lr": 0.001305, "grad_norm": 0.282096, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:40.456474+00:00", "epoch": 0, "step": 1432, "train_loss": 3.819920539855957, "perplexity": 45.60058474785332, "lr": 0.001305, "grad_norm": 0.283354, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:40.762703+00:00", "epoch": 0, "step": 1433, "train_loss": 3.9199068546295166, "perplexity": 50.39575042859596, "lr": 0.001305, "grad_norm": 0.305439, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:41.068599+00:00", "epoch": 0, "step": 1434, "train_loss": 3.824361801147461, "perplexity": 45.80355925754454, "lr": 0.001305, "grad_norm": 0.306827, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:41.375855+00:00", "epoch": 0, "step": 1435, "train_loss": 3.9376840591430664, "perplexity": 51.29965664852241, "lr": 0.001305, "grad_norm": 0.306263, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:41.682928+00:00", "epoch": 0, "step": 1436, "train_loss": 3.901390314102173, "perplexity": 49.47118179614645, "lr": 0.001305, "grad_norm": 0.258194, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:41.990027+00:00", "epoch": 0, "step": 1437, "train_loss": 3.8990044593811035, "perplexity": 49.35329143406201, "lr": 0.001305, "grad_norm": 0.270335, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:42.297787+00:00", "epoch": 0, "step": 1438, "train_loss": 3.8330905437469482, "perplexity": 46.205116732780326, "lr": 0.001305, "grad_norm": 0.301733, "tokens_per_sec": 106468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:42.604126+00:00", "epoch": 0, "step": 1439, "train_loss": 3.862614870071411, "perplexity": 47.58962951036022, "lr": 0.001305, "grad_norm": 0.276047, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:42.910318+00:00", "epoch": 0, "step": 1440, "train_loss": 3.913723945617676, "perplexity": 50.085119380441775, "lr": 0.001305, "grad_norm": 0.287603, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:43.216360+00:00", "epoch": 0, "step": 1441, "train_loss": 3.9226531982421875, "perplexity": 50.53434470252963, "lr": 0.001305, "grad_norm": 0.278511, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:43.523290+00:00", "epoch": 0, "step": 1442, "train_loss": 3.8332178592681885, "perplexity": 46.21099973579198, "lr": 0.001305, "grad_norm": 0.283873, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:43.830903+00:00", "epoch": 0, "step": 1443, "train_loss": 3.861902952194214, "perplexity": 47.555761659337335, "lr": 0.001305, "grad_norm": 0.263131, "tokens_per_sec": 106522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:44.137095+00:00", "epoch": 0, "step": 1444, "train_loss": 3.755072593688965, "perplexity": 42.73732215781754, "lr": 0.001305, "grad_norm": 0.258344, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:44.442538+00:00", "epoch": 0, "step": 1445, "train_loss": 3.822735071182251, "perplexity": 45.72910980618139, "lr": 0.001305, "grad_norm": 0.287869, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:44.747864+00:00", "epoch": 0, "step": 1446, "train_loss": 3.912078380584717, "perplexity": 50.00276883449015, "lr": 0.001305, "grad_norm": 0.268028, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:45.054454+00:00", "epoch": 0, "step": 1447, "train_loss": 3.900160551071167, "perplexity": 49.410381358401146, "lr": 0.001305, "grad_norm": 0.282994, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:45.360752+00:00", "epoch": 0, "step": 1448, "train_loss": 3.8706324100494385, "perplexity": 47.9727149166903, "lr": 0.001305, "grad_norm": 0.307039, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:45.669029+00:00", "epoch": 0, "step": 1449, "train_loss": 3.873840570449829, "perplexity": 48.126866219824265, "lr": 0.001305, "grad_norm": 0.302661, "tokens_per_sec": 106294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:45.975423+00:00", "epoch": 0, "step": 1450, "train_loss": 3.855370044708252, "perplexity": 47.24609687503093, "lr": 0.001305, "grad_norm": 0.313907, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:46.282271+00:00", "epoch": 0, "step": 1451, "train_loss": 3.8552358150482178, "perplexity": 47.239755473121086, "lr": 0.001305, "grad_norm": 0.30955, "tokens_per_sec": 106733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:46.588556+00:00", "epoch": 0, "step": 1452, "train_loss": 3.842806100845337, "perplexity": 46.656212959891846, "lr": 0.001305, "grad_norm": 0.277357, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:46.895661+00:00", "epoch": 0, "step": 1453, "train_loss": 3.8495280742645264, "perplexity": 46.970891227832645, "lr": 0.001305, "grad_norm": 0.256697, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:47.202248+00:00", "epoch": 0, "step": 1454, "train_loss": 3.829272747039795, "perplexity": 46.02905129514279, "lr": 0.001305, "grad_norm": 0.244665, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:47.509024+00:00", "epoch": 0, "step": 1455, "train_loss": 3.815847873687744, "perplexity": 45.41524645577368, "lr": 0.001305, "grad_norm": 0.277779, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:47.815381+00:00", "epoch": 0, "step": 1456, "train_loss": 3.92773699760437, "perplexity": 50.79190531040242, "lr": 0.001305, "grad_norm": 0.241643, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:48.121624+00:00", "epoch": 0, "step": 1457, "train_loss": 3.909775733947754, "perplexity": 49.887762587184206, "lr": 0.001305, "grad_norm": 0.253556, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:48.428108+00:00", "epoch": 0, "step": 1458, "train_loss": 3.9528725147247314, "perplexity": 52.084766414315894, "lr": 0.001305, "grad_norm": 0.249831, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:48.736323+00:00", "epoch": 0, "step": 1459, "train_loss": 3.8094706535339355, "perplexity": 45.12654496569744, "lr": 0.001305, "grad_norm": 0.234701, "tokens_per_sec": 106316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:49.044593+00:00", "epoch": 0, "step": 1460, "train_loss": 3.974689245223999, "perplexity": 53.23357174846102, "lr": 0.001305, "grad_norm": 0.254377, "tokens_per_sec": 106367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:49.351036+00:00", "epoch": 0, "step": 1461, "train_loss": 3.7351632118225098, "perplexity": 41.89486274203525, "lr": 0.001305, "grad_norm": 0.270935, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:49.658025+00:00", "epoch": 0, "step": 1462, "train_loss": 3.8905467987060547, "perplexity": 48.937638246432186, "lr": 0.001305, "grad_norm": 0.271312, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:49.964052+00:00", "epoch": 0, "step": 1463, "train_loss": 3.843728542327881, "perplexity": 46.699270442095184, "lr": 0.001305, "grad_norm": 0.271269, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:50.270074+00:00", "epoch": 0, "step": 1464, "train_loss": 3.8654818534851074, "perplexity": 47.726263959590256, "lr": 0.001305, "grad_norm": 0.276706, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:50.577931+00:00", "epoch": 0, "step": 1465, "train_loss": 3.880723476409912, "perplexity": 48.45926152473028, "lr": 0.001305, "grad_norm": 0.284109, "tokens_per_sec": 106439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:50.887644+00:00", "epoch": 0, "step": 1466, "train_loss": 3.83473539352417, "perplexity": 46.28117974771209, "lr": 0.001305, "grad_norm": 0.318872, "tokens_per_sec": 105802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:51.194797+00:00", "epoch": 0, "step": 1467, "train_loss": 3.9028263092041016, "perplexity": 49.542273202131376, "lr": 0.001305, "grad_norm": 0.332733, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:51.501225+00:00", "epoch": 0, "step": 1468, "train_loss": 3.95103120803833, "perplexity": 51.98895062585478, "lr": 0.001305, "grad_norm": 0.357178, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:51.807396+00:00", "epoch": 0, "step": 1469, "train_loss": 3.767331838607788, "perplexity": 43.2644740976125, "lr": 0.001305, "grad_norm": 0.30147, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:52.114467+00:00", "epoch": 0, "step": 1470, "train_loss": 3.912966728210449, "perplexity": 50.047208411438156, "lr": 0.001305, "grad_norm": 0.274928, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:52.421351+00:00", "epoch": 0, "step": 1471, "train_loss": 3.928906202316284, "perplexity": 50.85132617622737, "lr": 0.001305, "grad_norm": 0.273367, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:52.728159+00:00", "epoch": 0, "step": 1472, "train_loss": 3.9309616088867188, "perplexity": 50.9559538154989, "lr": 0.001305, "grad_norm": 0.298446, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:53.034915+00:00", "epoch": 0, "step": 1473, "train_loss": 3.977001190185547, "perplexity": 53.35678721525575, "lr": 0.001305, "grad_norm": 0.270171, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:53.341349+00:00", "epoch": 0, "step": 1474, "train_loss": 3.893927574157715, "perplexity": 49.10336539777026, "lr": 0.001305, "grad_norm": 0.246988, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:53.647486+00:00", "epoch": 0, "step": 1475, "train_loss": 3.9215755462646484, "perplexity": 50.479915599110306, "lr": 0.001305, "grad_norm": 0.276419, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:53.955381+00:00", "epoch": 0, "step": 1476, "train_loss": 3.8472650051116943, "perplexity": 46.86471304241777, "lr": 0.001305, "grad_norm": 0.284348, "tokens_per_sec": 106430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:54.262939+00:00", "epoch": 0, "step": 1477, "train_loss": 3.8773086071014404, "perplexity": 48.29406170826724, "lr": 0.001305, "grad_norm": 0.283887, "tokens_per_sec": 106539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:54.570414+00:00", "epoch": 0, "step": 1478, "train_loss": 3.915910005569458, "perplexity": 50.19472821619227, "lr": 0.001305, "grad_norm": 0.274652, "tokens_per_sec": 106577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:54.876876+00:00", "epoch": 0, "step": 1479, "train_loss": 3.7651290893554688, "perplexity": 43.16927819448591, "lr": 0.001305, "grad_norm": 0.279884, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:55.183842+00:00", "epoch": 0, "step": 1480, "train_loss": 3.8467085361480713, "perplexity": 46.838641538784636, "lr": 0.001305, "grad_norm": 0.273413, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:55.490034+00:00", "epoch": 0, "step": 1481, "train_loss": 3.801633358001709, "perplexity": 44.77425719137124, "lr": 0.001305, "grad_norm": 0.267294, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:55.797151+00:00", "epoch": 0, "step": 1482, "train_loss": 3.982469081878662, "perplexity": 53.64933543094095, "lr": 0.001305, "grad_norm": 0.236277, "tokens_per_sec": 106696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:56.104542+00:00", "epoch": 0, "step": 1483, "train_loss": 3.818399667739868, "perplexity": 45.5312848015688, "lr": 0.001305, "grad_norm": 0.254222, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:56.412365+00:00", "epoch": 0, "step": 1484, "train_loss": 3.8494746685028076, "perplexity": 46.96838277859143, "lr": 0.001305, "grad_norm": 0.249427, "tokens_per_sec": 106450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:56.719321+00:00", "epoch": 0, "step": 1485, "train_loss": 3.769493341445923, "perplexity": 43.358091521879906, "lr": 0.001305, "grad_norm": 0.262747, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:57.025436+00:00", "epoch": 0, "step": 1486, "train_loss": 3.8666083812713623, "perplexity": 47.78005921731203, "lr": 0.001305, "grad_norm": 0.290479, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:57.332215+00:00", "epoch": 0, "step": 1487, "train_loss": 3.9927732944488525, "perplexity": 54.20500755268659, "lr": 0.001305, "grad_norm": 0.2684, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:57.640003+00:00", "epoch": 0, "step": 1488, "train_loss": 3.8282699584960938, "perplexity": 45.982917025151295, "lr": 0.001305, "grad_norm": 0.270188, "tokens_per_sec": 106463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:57.947089+00:00", "epoch": 0, "step": 1489, "train_loss": 3.949442148208618, "perplexity": 51.90640267701135, "lr": 0.001305, "grad_norm": 0.28317, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:58.253660+00:00", "epoch": 0, "step": 1490, "train_loss": 3.830040693283081, "perplexity": 46.064412708261266, "lr": 0.001305, "grad_norm": 0.294399, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:58.559726+00:00", "epoch": 0, "step": 1491, "train_loss": 3.83681583404541, "perplexity": 46.3775652167829, "lr": 0.001305, "grad_norm": 0.270122, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:58.866069+00:00", "epoch": 0, "step": 1492, "train_loss": 3.884396553039551, "perplexity": 48.63758340019735, "lr": 0.001305, "grad_norm": 0.282698, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:59.173230+00:00", "epoch": 0, "step": 1493, "train_loss": 3.7694575786590576, "perplexity": 43.35654094342059, "lr": 0.001305, "grad_norm": 0.287688, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:59.480112+00:00", "epoch": 0, "step": 1494, "train_loss": 3.7856028079986572, "perplexity": 44.06222361558243, "lr": 0.001305, "grad_norm": 0.281698, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:08:59.787247+00:00", "epoch": 0, "step": 1495, "train_loss": 3.8200888633728027, "perplexity": 45.608261044680816, "lr": 0.001305, "grad_norm": 0.244941, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:00.093941+00:00", "epoch": 0, "step": 1496, "train_loss": 3.8466134071350098, "perplexity": 46.834186036969015, "lr": 0.001305, "grad_norm": 0.243694, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:00.400484+00:00", "epoch": 0, "step": 1497, "train_loss": 3.915541172027588, "perplexity": 50.17621813058114, "lr": 0.001305, "grad_norm": 0.260159, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:00.707206+00:00", "epoch": 0, "step": 1498, "train_loss": 3.890018939971924, "perplexity": 48.911812903322186, "lr": 0.001305, "grad_norm": 0.264205, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:01.014151+00:00", "epoch": 0, "step": 1499, "train_loss": 3.7952942848205566, "perplexity": 44.49132760152367, "lr": 0.001305, "grad_norm": 0.266172, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:01.322157+00:00", "epoch": 0, "step": 1500, "train_loss": 3.8253726959228516, "perplexity": 45.84988524769792, "lr": 0.001305, "grad_norm": 0.252718, "tokens_per_sec": 106447} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:09:07.190782+00:00", "step": 1500, "epoch": 0, "val_loss": 3.897708606719971, "val_ppl": 49.28937825999364, "eval_train_loss": 3.8253726959228516, "eval_train_ppl": 45.84988524769792} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:09:08.089490+00:00", "step": 1500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p8977_epoch_0000_step_0001500.pt", "val_loss": 3.897708606719971} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:09:09.874476+00:00", "step": 1500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0000_step_0001500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:10.947187+00:00", "epoch": 0, "step": 1501, "train_loss": 3.938241720199585, "perplexity": 51.328272447464805, "lr": 0.001305, "grad_norm": 0.278138, "tokens_per_sec": 3404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:11.251190+00:00", "epoch": 0, "step": 1502, "train_loss": 4.012598037719727, "perplexity": 55.29033048976867, "lr": 0.001305, "grad_norm": 0.244916, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:11.556636+00:00", "epoch": 0, "step": 1503, "train_loss": 3.8838672637939453, "perplexity": 48.61184686200866, "lr": 0.001305, "grad_norm": 0.242382, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:11.863346+00:00", "epoch": 0, "step": 1504, "train_loss": 3.9130125045776367, "perplexity": 50.04949944326426, "lr": 0.001305, "grad_norm": 0.245778, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:12.168342+00:00", "epoch": 0, "step": 1505, "train_loss": 3.9715614318847656, "perplexity": 53.067327199144444, "lr": 0.001305, "grad_norm": 0.291025, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:12.472248+00:00", "epoch": 0, "step": 1506, "train_loss": 3.7544913291931152, "perplexity": 42.712487688196475, "lr": 0.001305, "grad_norm": 0.308697, "tokens_per_sec": 107824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:12.778354+00:00", "epoch": 0, "step": 1507, "train_loss": 3.8159661293029785, "perplexity": 45.420617381249116, "lr": 0.001305, "grad_norm": 0.309747, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:13.084659+00:00", "epoch": 0, "step": 1508, "train_loss": 3.8182640075683594, "perplexity": 45.52510843861631, "lr": 0.001305, "grad_norm": 0.312678, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:13.391566+00:00", "epoch": 0, "step": 1509, "train_loss": 3.897353172302246, "perplexity": 49.27186223161586, "lr": 0.001305, "grad_norm": 0.299578, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:13.697932+00:00", "epoch": 0, "step": 1510, "train_loss": 3.9278178215026855, "perplexity": 50.79601067609607, "lr": 0.001305, "grad_norm": 0.279719, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:14.003084+00:00", "epoch": 0, "step": 1511, "train_loss": 3.989109754562378, "perplexity": 54.00678865858299, "lr": 0.001305, "grad_norm": 0.299412, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:14.308092+00:00", "epoch": 0, "step": 1512, "train_loss": 3.9922332763671875, "perplexity": 54.175743770688214, "lr": 0.001305, "grad_norm": 0.295565, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:14.615049+00:00", "epoch": 0, "step": 1513, "train_loss": 3.8283188343048096, "perplexity": 45.98516453233193, "lr": 0.001305, "grad_norm": 0.271128, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:14.922174+00:00", "epoch": 0, "step": 1514, "train_loss": 3.8696975708007812, "perplexity": 47.92788909565449, "lr": 0.001305, "grad_norm": 0.278745, "tokens_per_sec": 106689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:15.597221+00:00", "epoch": 0, "step": 1515, "train_loss": 3.917184591293335, "perplexity": 50.25874648991255, "lr": 0.001305, "grad_norm": 0.294405, "tokens_per_sec": 48542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:15.904340+00:00", "epoch": 0, "step": 1516, "train_loss": 3.9963340759277344, "perplexity": 54.39836378498866, "lr": 0.001305, "grad_norm": 0.31589, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:16.209846+00:00", "epoch": 0, "step": 1517, "train_loss": 3.809305191040039, "perplexity": 45.119078832726004, "lr": 0.001305, "grad_norm": 0.310101, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:16.515643+00:00", "epoch": 0, "step": 1518, "train_loss": 3.8939805030822754, "perplexity": 49.105964454875114, "lr": 0.001305, "grad_norm": 0.28379, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:16.820789+00:00", "epoch": 0, "step": 1519, "train_loss": 3.886533737182617, "perplexity": 48.741642028815036, "lr": 0.001305, "grad_norm": 0.247125, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:17.127640+00:00", "epoch": 0, "step": 1520, "train_loss": 3.77876877784729, "perplexity": 43.76212765126327, "lr": 0.001305, "grad_norm": 0.278675, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:17.434926+00:00", "epoch": 0, "step": 1521, "train_loss": 3.8917768001556396, "perplexity": 48.99786864656073, "lr": 0.001305, "grad_norm": 0.296246, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:17.741698+00:00", "epoch": 0, "step": 1522, "train_loss": 3.9185445308685303, "perplexity": 50.32714184449755, "lr": 0.001305, "grad_norm": 0.263718, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:18.048100+00:00", "epoch": 0, "step": 1523, "train_loss": 3.8298747539520264, "perplexity": 46.056769444607596, "lr": 0.001305, "grad_norm": 0.257562, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:18.353318+00:00", "epoch": 0, "step": 1524, "train_loss": 3.8186540603637695, "perplexity": 45.54286909799638, "lr": 0.001305, "grad_norm": 0.279754, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:18.659809+00:00", "epoch": 0, "step": 1525, "train_loss": 3.8400919437408447, "perplexity": 46.52975236261488, "lr": 0.001305, "grad_norm": 0.288091, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:18.965784+00:00", "epoch": 0, "step": 1526, "train_loss": 3.8074233531951904, "perplexity": 45.034251882969265, "lr": 0.001305, "grad_norm": 0.266701, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:19.271996+00:00", "epoch": 0, "step": 1527, "train_loss": 3.8223512172698975, "perplexity": 45.711559876993825, "lr": 0.001305, "grad_norm": 0.249641, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:19.578264+00:00", "epoch": 0, "step": 1528, "train_loss": 3.9222500324249268, "perplexity": 50.51397508858984, "lr": 0.001305, "grad_norm": 0.28196, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:19.883994+00:00", "epoch": 0, "step": 1529, "train_loss": 3.8547332286834717, "perplexity": 47.21601938136402, "lr": 0.001305, "grad_norm": 0.267951, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:20.189257+00:00", "epoch": 0, "step": 1530, "train_loss": 3.873392105102539, "perplexity": 48.10528782699339, "lr": 0.001305, "grad_norm": 0.291191, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:20.495015+00:00", "epoch": 0, "step": 1531, "train_loss": 3.7857539653778076, "perplexity": 44.068884449228094, "lr": 0.001305, "grad_norm": 0.296736, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:20.803243+00:00", "epoch": 0, "step": 1532, "train_loss": 3.8192498683929443, "perplexity": 45.570011990250244, "lr": 0.001305, "grad_norm": 0.283393, "tokens_per_sec": 106312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:21.110184+00:00", "epoch": 0, "step": 1533, "train_loss": 3.851529598236084, "perplexity": 47.064998740394735, "lr": 0.001305, "grad_norm": 0.26786, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:21.416161+00:00", "epoch": 0, "step": 1534, "train_loss": 3.7972569465637207, "perplexity": 44.57873477540989, "lr": 0.001305, "grad_norm": 0.238324, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:21.722006+00:00", "epoch": 0, "step": 1535, "train_loss": 3.940467596054077, "perplexity": 51.442650057715014, "lr": 0.001305, "grad_norm": 0.248032, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:22.028513+00:00", "epoch": 0, "step": 1536, "train_loss": 3.8495233058929443, "perplexity": 46.97066725370373, "lr": 0.001305, "grad_norm": 0.252629, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:22.334906+00:00", "epoch": 0, "step": 1537, "train_loss": 3.879605531692505, "perplexity": 48.405117020219805, "lr": 0.001305, "grad_norm": 0.271189, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:22.641661+00:00", "epoch": 0, "step": 1538, "train_loss": 3.8194081783294678, "perplexity": 45.57722674702457, "lr": 0.001305, "grad_norm": 0.268147, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:22.948872+00:00", "epoch": 0, "step": 1539, "train_loss": 3.858485698699951, "perplexity": 47.39352891970718, "lr": 0.001305, "grad_norm": 0.272391, "tokens_per_sec": 106665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:23.254792+00:00", "epoch": 0, "step": 1540, "train_loss": 3.846342086791992, "perplexity": 46.82148069323557, "lr": 0.001305, "grad_norm": 0.284687, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:23.560688+00:00", "epoch": 0, "step": 1541, "train_loss": 3.785419225692749, "perplexity": 44.054135313424936, "lr": 0.001305, "grad_norm": 0.28917, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:23.866893+00:00", "epoch": 0, "step": 1542, "train_loss": 3.8960299491882324, "perplexity": 49.20670768115077, "lr": 0.001305, "grad_norm": 0.270508, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:24.174062+00:00", "epoch": 0, "step": 1543, "train_loss": 3.9193031787872314, "perplexity": 50.365336912384514, "lr": 0.001305, "grad_norm": 0.28692, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:24.480669+00:00", "epoch": 0, "step": 1544, "train_loss": 3.928361654281616, "perplexity": 50.8236427246665, "lr": 0.001305, "grad_norm": 0.331193, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:24.786905+00:00", "epoch": 0, "step": 1545, "train_loss": 3.8645195960998535, "perplexity": 47.68036109835012, "lr": 0.001305, "grad_norm": 0.304584, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:25.092585+00:00", "epoch": 0, "step": 1546, "train_loss": 3.829580068588257, "perplexity": 46.04319918832592, "lr": 0.001305, "grad_norm": 0.26529, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:25.399482+00:00", "epoch": 0, "step": 1547, "train_loss": 3.87807035446167, "perplexity": 48.33086359737967, "lr": 0.001305, "grad_norm": 0.255412, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:25.706529+00:00", "epoch": 0, "step": 1548, "train_loss": 3.8939437866210938, "perplexity": 49.10416149073685, "lr": 0.001305, "grad_norm": 0.288059, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:26.013369+00:00", "epoch": 0, "step": 1549, "train_loss": 3.85375714302063, "perplexity": 47.16995498684424, "lr": 0.001305, "grad_norm": 0.264607, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:26.319234+00:00", "epoch": 0, "step": 1550, "train_loss": 3.8633639812469482, "perplexity": 47.62529278987897, "lr": 0.001305, "grad_norm": 0.24005, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:26.625908+00:00", "epoch": 0, "step": 1551, "train_loss": 3.8320395946502686, "perplexity": 46.15658301478867, "lr": 0.001305, "grad_norm": 0.244892, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:26.932275+00:00", "epoch": 0, "step": 1552, "train_loss": 3.7451705932617188, "perplexity": 42.316225466054014, "lr": 0.001305, "grad_norm": 0.241888, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:27.239063+00:00", "epoch": 0, "step": 1553, "train_loss": 3.887176990509033, "perplexity": 48.77300533838021, "lr": 0.001305, "grad_norm": 0.266038, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:27.546343+00:00", "epoch": 0, "step": 1554, "train_loss": 3.787431478500366, "perplexity": 44.142872621902946, "lr": 0.001305, "grad_norm": 0.253568, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:27.852248+00:00", "epoch": 0, "step": 1555, "train_loss": 3.7547950744628906, "perplexity": 42.72546337484406, "lr": 0.001305, "grad_norm": 0.231632, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:28.157780+00:00", "epoch": 0, "step": 1556, "train_loss": 3.915958881378174, "perplexity": 50.19718158408178, "lr": 0.001305, "grad_norm": 0.282786, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:28.463917+00:00", "epoch": 0, "step": 1557, "train_loss": 3.772371292114258, "perplexity": 43.483053701596376, "lr": 0.001305, "grad_norm": 0.287436, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:28.770394+00:00", "epoch": 0, "step": 1558, "train_loss": 3.9184300899505615, "perplexity": 50.32138268973389, "lr": 0.001305, "grad_norm": 0.285992, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:29.077441+00:00", "epoch": 0, "step": 1559, "train_loss": 3.6802680492401123, "perplexity": 39.65702268280875, "lr": 0.001305, "grad_norm": 0.290514, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:29.384223+00:00", "epoch": 0, "step": 1560, "train_loss": 3.8796682357788086, "perplexity": 48.40815231401665, "lr": 0.001305, "grad_norm": 0.2845, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:29.710069+00:00", "epoch": 0, "step": 1561, "train_loss": 3.7764134407043457, "perplexity": 43.65917437898048, "lr": 0.001305, "grad_norm": 0.287548, "tokens_per_sec": 100513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:30.015389+00:00", "epoch": 0, "step": 1562, "train_loss": 3.8005988597869873, "perplexity": 44.72796225238583, "lr": 0.001305, "grad_norm": 0.273457, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:30.322554+00:00", "epoch": 0, "step": 1563, "train_loss": 4.002350330352783, "perplexity": 54.72662464211776, "lr": 0.001305, "grad_norm": 0.305132, "tokens_per_sec": 106679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:30.629921+00:00", "epoch": 0, "step": 1564, "train_loss": 3.9326117038726807, "perplexity": 51.04010538934571, "lr": 0.001305, "grad_norm": 0.283388, "tokens_per_sec": 106609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:30.938238+00:00", "epoch": 0, "step": 1565, "train_loss": 3.8382325172424316, "perplexity": 46.4433140958139, "lr": 0.001305, "grad_norm": 0.31288, "tokens_per_sec": 106280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:31.244751+00:00", "epoch": 0, "step": 1566, "train_loss": 3.9362199306488037, "perplexity": 51.22460231747742, "lr": 0.001305, "grad_norm": 0.338739, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:31.550867+00:00", "epoch": 0, "step": 1567, "train_loss": 3.9252078533172607, "perplexity": 50.663607563400944, "lr": 0.001305, "grad_norm": 0.394715, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:31.857634+00:00", "epoch": 0, "step": 1568, "train_loss": 3.8432445526123047, "perplexity": 46.6766739441537, "lr": 0.001305, "grad_norm": 0.419989, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:32.165199+00:00", "epoch": 0, "step": 1569, "train_loss": 3.783287525177002, "perplexity": 43.960325113606125, "lr": 0.001305, "grad_norm": 0.360535, "tokens_per_sec": 106541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:32.472962+00:00", "epoch": 0, "step": 1570, "train_loss": 3.896313190460205, "perplexity": 49.2206470256296, "lr": 0.001305, "grad_norm": 0.367711, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:32.780460+00:00", "epoch": 0, "step": 1571, "train_loss": 3.892850637435913, "perplexity": 49.05051264505267, "lr": 0.001305, "grad_norm": 0.351871, "tokens_per_sec": 106508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:33.086240+00:00", "epoch": 0, "step": 1572, "train_loss": 3.858304738998413, "perplexity": 47.384953376796275, "lr": 0.001305, "grad_norm": 0.296174, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:33.393029+00:00", "epoch": 0, "step": 1573, "train_loss": 3.8341424465179443, "perplexity": 46.253745595039234, "lr": 0.001305, "grad_norm": 0.261067, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:33.700601+00:00", "epoch": 0, "step": 1574, "train_loss": 3.889526605606079, "perplexity": 48.88773786390505, "lr": 0.001305, "grad_norm": 0.273458, "tokens_per_sec": 106542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:34.007973+00:00", "epoch": 0, "step": 1575, "train_loss": 3.9506239891052246, "perplexity": 51.96778405085519, "lr": 0.001305, "grad_norm": 0.264535, "tokens_per_sec": 106603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:34.314897+00:00", "epoch": 0, "step": 1576, "train_loss": 3.926068067550659, "perplexity": 50.70720786985582, "lr": 0.001305, "grad_norm": 0.251479, "tokens_per_sec": 106763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:34.621050+00:00", "epoch": 0, "step": 1577, "train_loss": 3.946178436279297, "perplexity": 51.73727127959559, "lr": 0.001305, "grad_norm": 0.24723, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:34.927754+00:00", "epoch": 0, "step": 1578, "train_loss": 3.796250343322754, "perplexity": 44.53388425363412, "lr": 0.001305, "grad_norm": 0.272203, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:35.234853+00:00", "epoch": 0, "step": 1579, "train_loss": 3.8090038299560547, "perplexity": 45.1054837468384, "lr": 0.001305, "grad_norm": 0.245986, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:35.542064+00:00", "epoch": 0, "step": 1580, "train_loss": 3.8303964138031006, "perplexity": 46.08080167987818, "lr": 0.001305, "grad_norm": 0.244059, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:35.849314+00:00", "epoch": 0, "step": 1581, "train_loss": 3.865084409713745, "perplexity": 47.70729922220714, "lr": 0.001305, "grad_norm": 0.266776, "tokens_per_sec": 106566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:36.157218+00:00", "epoch": 0, "step": 1582, "train_loss": 3.8469388484954834, "perplexity": 46.849430298611644, "lr": 0.001305, "grad_norm": 0.263283, "tokens_per_sec": 106427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:36.464156+00:00", "epoch": 0, "step": 1583, "train_loss": 3.8738274574279785, "perplexity": 48.12623513531365, "lr": 0.001305, "grad_norm": 0.238285, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:36.771463+00:00", "epoch": 0, "step": 1584, "train_loss": 3.811659812927246, "perplexity": 45.22544237709884, "lr": 0.001305, "grad_norm": 0.233579, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:37.080274+00:00", "epoch": 0, "step": 1585, "train_loss": 3.820681095123291, "perplexity": 45.63527970482052, "lr": 0.001305, "grad_norm": 0.24576, "tokens_per_sec": 106110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:37.387671+00:00", "epoch": 0, "step": 1586, "train_loss": 3.8524234294891357, "perplexity": 47.10708571371773, "lr": 0.001305, "grad_norm": 0.241155, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:37.694423+00:00", "epoch": 0, "step": 1587, "train_loss": 3.846748113632202, "perplexity": 46.84049533106083, "lr": 0.001305, "grad_norm": 0.24583, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:38.000863+00:00", "epoch": 0, "step": 1588, "train_loss": 3.861158609390259, "perplexity": 47.52037704113629, "lr": 0.001305, "grad_norm": 0.260247, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:38.307981+00:00", "epoch": 0, "step": 1589, "train_loss": 3.791830539703369, "perplexity": 44.33748756792764, "lr": 0.001305, "grad_norm": 0.262885, "tokens_per_sec": 106696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:38.615278+00:00", "epoch": 0, "step": 1590, "train_loss": 3.8103723526000977, "perplexity": 45.16725387998889, "lr": 0.001305, "grad_norm": 0.296509, "tokens_per_sec": 106696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:38.922263+00:00", "epoch": 0, "step": 1591, "train_loss": 3.8094639778137207, "perplexity": 45.12624371451452, "lr": 0.001305, "grad_norm": 0.301676, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:39.229262+00:00", "epoch": 0, "step": 1592, "train_loss": 3.8086462020874023, "perplexity": 45.089355652921306, "lr": 0.001305, "grad_norm": 0.298688, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:39.535984+00:00", "epoch": 0, "step": 1593, "train_loss": 3.941235303878784, "perplexity": 51.48215814608185, "lr": 0.001305, "grad_norm": 0.32014, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:39.844225+00:00", "epoch": 0, "step": 1594, "train_loss": 3.9460151195526123, "perplexity": 51.72882240774244, "lr": 0.001305, "grad_norm": 0.282658, "tokens_per_sec": 106307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:40.150203+00:00", "epoch": 0, "step": 1595, "train_loss": 3.7922775745391846, "perplexity": 44.35731240026822, "lr": 0.001305, "grad_norm": 0.263742, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:40.457847+00:00", "epoch": 0, "step": 1596, "train_loss": 3.782021999359131, "perplexity": 43.90472737482001, "lr": 0.001305, "grad_norm": 0.245241, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:40.765139+00:00", "epoch": 0, "step": 1597, "train_loss": 3.7862281799316406, "perplexity": 44.08978751148155, "lr": 0.001305, "grad_norm": 0.237849, "tokens_per_sec": 106636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:41.072395+00:00", "epoch": 0, "step": 1598, "train_loss": 3.850966215133667, "perplexity": 47.03849058321418, "lr": 0.001305, "grad_norm": 0.245427, "tokens_per_sec": 106647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:41.378802+00:00", "epoch": 0, "step": 1599, "train_loss": 3.8003318309783936, "perplexity": 44.71602019242273, "lr": 0.001305, "grad_norm": 0.22252, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:41.686079+00:00", "epoch": 0, "step": 1600, "train_loss": 3.9312894344329834, "perplexity": 50.97266121730084, "lr": 0.001305, "grad_norm": 0.2674, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:09:47.556664+00:00", "step": 1600, "epoch": 0, "val_loss": 3.875067102909088, "val_ppl": 48.18593159880282, "eval_train_loss": 3.9312894344329834, "eval_train_ppl": 50.97266121730084} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:09:48.508730+00:00", "step": 1600, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p8751_epoch_0000_step_0001600.pt", "val_loss": 3.875067102909088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:49.593605+00:00", "epoch": 0, "step": 1601, "train_loss": 3.7849087715148926, "perplexity": 44.031653434475025, "lr": 0.001305, "grad_norm": 0.261422, "tokens_per_sec": 4144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:49.899452+00:00", "epoch": 0, "step": 1602, "train_loss": 3.8081188201904297, "perplexity": 45.06558261229043, "lr": 0.001305, "grad_norm": 0.254003, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:50.206210+00:00", "epoch": 0, "step": 1603, "train_loss": 3.7648983001708984, "perplexity": 43.159316341561095, "lr": 0.001305, "grad_norm": 0.255001, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:50.513447+00:00", "epoch": 0, "step": 1604, "train_loss": 3.8864667415618896, "perplexity": 48.73837666163591, "lr": 0.001305, "grad_norm": 0.231147, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:50.821076+00:00", "epoch": 0, "step": 1605, "train_loss": 3.835444211959839, "perplexity": 46.313996330267145, "lr": 0.001305, "grad_norm": 0.236096, "tokens_per_sec": 106519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:51.128627+00:00", "epoch": 0, "step": 1606, "train_loss": 3.8623952865600586, "perplexity": 47.57918075963704, "lr": 0.001305, "grad_norm": 0.251905, "tokens_per_sec": 106545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:51.434414+00:00", "epoch": 0, "step": 1607, "train_loss": 3.813816785812378, "perplexity": 45.323097712117146, "lr": 0.001305, "grad_norm": 0.2638, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:51.740696+00:00", "epoch": 0, "step": 1608, "train_loss": 3.875675916671753, "perplexity": 48.21527678909845, "lr": 0.001305, "grad_norm": 0.266093, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:52.047066+00:00", "epoch": 0, "step": 1609, "train_loss": 3.7539453506469727, "perplexity": 42.68917395124558, "lr": 0.001305, "grad_norm": 0.243149, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:52.353548+00:00", "epoch": 0, "step": 1610, "train_loss": 3.759290933609009, "perplexity": 42.91798348750469, "lr": 0.001305, "grad_norm": 0.269346, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:52.659965+00:00", "epoch": 0, "step": 1611, "train_loss": 3.8525197505950928, "perplexity": 47.11162333884312, "lr": 0.001305, "grad_norm": 0.290237, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:52.965449+00:00", "epoch": 0, "step": 1612, "train_loss": 3.7683115005493164, "perplexity": 43.306879424357035, "lr": 0.001305, "grad_norm": 0.324395, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:53.271981+00:00", "epoch": 0, "step": 1613, "train_loss": 3.698925495147705, "perplexity": 40.40386687634303, "lr": 0.001305, "grad_norm": 0.332893, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:53.577218+00:00", "epoch": 0, "step": 1614, "train_loss": 3.852726697921753, "perplexity": 47.12137397224667, "lr": 0.001305, "grad_norm": 0.278511, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:53.883004+00:00", "epoch": 0, "step": 1615, "train_loss": 3.9192328453063965, "perplexity": 50.3617946674967, "lr": 0.001305, "grad_norm": 0.278168, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:54.189435+00:00", "epoch": 0, "step": 1616, "train_loss": 3.9035913944244385, "perplexity": 49.58019176675725, "lr": 0.001305, "grad_norm": 0.282733, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:54.495693+00:00", "epoch": 0, "step": 1617, "train_loss": 3.9929158687591553, "perplexity": 54.21273634520391, "lr": 0.001305, "grad_norm": 0.277212, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:54.801532+00:00", "epoch": 0, "step": 1618, "train_loss": 3.8949708938598633, "perplexity": 49.15462264052765, "lr": 0.001305, "grad_norm": 0.253587, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:55.107984+00:00", "epoch": 0, "step": 1619, "train_loss": 3.867678642272949, "perplexity": 47.831223726151165, "lr": 0.001305, "grad_norm": 0.265699, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:55.413915+00:00", "epoch": 0, "step": 1620, "train_loss": 3.7896244525909424, "perplexity": 44.23978302000452, "lr": 0.001305, "grad_norm": 0.271212, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:55.719366+00:00", "epoch": 0, "step": 1621, "train_loss": 3.7894725799560547, "perplexity": 44.23306471776646, "lr": 0.001305, "grad_norm": 0.294042, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:56.025666+00:00", "epoch": 0, "step": 1622, "train_loss": 3.8945295810699463, "perplexity": 49.132934862771684, "lr": 0.001305, "grad_norm": 0.280502, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:56.331572+00:00", "epoch": 0, "step": 1623, "train_loss": 3.873663902282715, "perplexity": 48.11836448559527, "lr": 0.001305, "grad_norm": 0.288876, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:56.636827+00:00", "epoch": 0, "step": 1624, "train_loss": 3.8530585765838623, "perplexity": 47.1370151461397, "lr": 0.001305, "grad_norm": 0.261209, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:56.942516+00:00", "epoch": 0, "step": 1625, "train_loss": 3.7989377975463867, "perplexity": 44.65372799411955, "lr": 0.001305, "grad_norm": 0.24998, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:57.248522+00:00", "epoch": 0, "step": 1626, "train_loss": 3.784761667251587, "perplexity": 44.02517666692616, "lr": 0.001305, "grad_norm": 0.247976, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:57.555617+00:00", "epoch": 0, "step": 1627, "train_loss": 3.8670167922973633, "perplexity": 47.79957710570847, "lr": 0.001305, "grad_norm": 0.28583, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:57.861593+00:00", "epoch": 0, "step": 1628, "train_loss": 3.9952995777130127, "perplexity": 54.34211787293325, "lr": 0.001305, "grad_norm": 0.310347, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:58.167223+00:00", "epoch": 0, "step": 1629, "train_loss": 3.837122678756714, "perplexity": 46.391798110925286, "lr": 0.001305, "grad_norm": 0.336096, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:58.472931+00:00", "epoch": 0, "step": 1630, "train_loss": 3.853071451187134, "perplexity": 47.13762202041573, "lr": 0.001305, "grad_norm": 0.304381, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:58.778921+00:00", "epoch": 0, "step": 1631, "train_loss": 3.857499122619629, "perplexity": 47.34679465495746, "lr": 0.001305, "grad_norm": 0.261964, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:59.085803+00:00", "epoch": 0, "step": 1632, "train_loss": 3.836777448654175, "perplexity": 46.37578502996431, "lr": 0.001305, "grad_norm": 0.260407, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:59.392530+00:00", "epoch": 0, "step": 1633, "train_loss": 3.8714804649353027, "perplexity": 48.01341566775965, "lr": 0.001305, "grad_norm": 0.266978, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:09:59.700173+00:00", "epoch": 0, "step": 1634, "train_loss": 3.762929916381836, "perplexity": 43.07444579923001, "lr": 0.001305, "grad_norm": 0.275817, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:00.006756+00:00", "epoch": 0, "step": 1635, "train_loss": 3.779975175857544, "perplexity": 43.81495405341251, "lr": 0.001305, "grad_norm": 0.255629, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:00.311872+00:00", "epoch": 0, "step": 1636, "train_loss": 3.8913776874542236, "perplexity": 48.97831687678094, "lr": 0.001305, "grad_norm": 0.266641, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:00.617820+00:00", "epoch": 0, "step": 1637, "train_loss": 3.789836883544922, "perplexity": 44.24918191758835, "lr": 0.001305, "grad_norm": 0.257506, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:00.924229+00:00", "epoch": 0, "step": 1638, "train_loss": 3.7077853679656982, "perplexity": 40.763430490103794, "lr": 0.001305, "grad_norm": 0.249426, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:01.230861+00:00", "epoch": 0, "step": 1639, "train_loss": 3.848512887954712, "perplexity": 46.92323121808629, "lr": 0.001305, "grad_norm": 0.262968, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:01.537355+00:00", "epoch": 0, "step": 1640, "train_loss": 3.809899091720581, "perplexity": 45.14588304308172, "lr": 0.001305, "grad_norm": 0.286715, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:01.843018+00:00", "epoch": 0, "step": 1641, "train_loss": 3.853943347930908, "perplexity": 47.178739081875825, "lr": 0.001305, "grad_norm": 0.256522, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:02.147889+00:00", "epoch": 0, "step": 1642, "train_loss": 3.8108468055725098, "perplexity": 45.188688702353446, "lr": 0.001305, "grad_norm": 0.241194, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:02.454877+00:00", "epoch": 0, "step": 1643, "train_loss": 3.7995080947875977, "perplexity": 44.67920115495065, "lr": 0.001305, "grad_norm": 0.27504, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:02.760558+00:00", "epoch": 0, "step": 1644, "train_loss": 3.7963812351226807, "perplexity": 44.539713755410475, "lr": 0.001305, "grad_norm": 0.288875, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:03.067647+00:00", "epoch": 0, "step": 1645, "train_loss": 3.780669689178467, "perplexity": 43.84539469214919, "lr": 0.001305, "grad_norm": 0.296652, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:03.373797+00:00", "epoch": 0, "step": 1646, "train_loss": 3.7721188068389893, "perplexity": 43.47207625669293, "lr": 0.001305, "grad_norm": 0.293406, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:03.679824+00:00", "epoch": 0, "step": 1647, "train_loss": 3.9375240802764893, "perplexity": 51.29145044402313, "lr": 0.001305, "grad_norm": 0.269543, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:03.984669+00:00", "epoch": 0, "step": 1648, "train_loss": 3.8862502574920654, "perplexity": 48.727826721487766, "lr": 0.001305, "grad_norm": 0.260292, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:04.291224+00:00", "epoch": 0, "step": 1649, "train_loss": 3.8437633514404297, "perplexity": 46.700896030548435, "lr": 0.001305, "grad_norm": 0.232571, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:04.597736+00:00", "epoch": 0, "step": 1650, "train_loss": 3.744981050491333, "perplexity": 42.30820549153502, "lr": 0.001305, "grad_norm": 0.235836, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:04.904509+00:00", "epoch": 0, "step": 1651, "train_loss": 3.7839462757110596, "perplexity": 43.98929354168822, "lr": 0.001305, "grad_norm": 0.236829, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:05.210135+00:00", "epoch": 0, "step": 1652, "train_loss": 3.917736291885376, "perplexity": 50.286481920229434, "lr": 0.001305, "grad_norm": 0.237126, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:05.517042+00:00", "epoch": 0, "step": 1653, "train_loss": 3.779695987701416, "perplexity": 43.80272314462139, "lr": 0.001305, "grad_norm": 0.252054, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:05.823120+00:00", "epoch": 0, "step": 1654, "train_loss": 3.8357815742492676, "perplexity": 46.32962356197284, "lr": 0.001305, "grad_norm": 0.271815, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:06.130132+00:00", "epoch": 0, "step": 1655, "train_loss": 3.847412109375, "perplexity": 46.8716075485981, "lr": 0.001305, "grad_norm": 0.279591, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:06.436801+00:00", "epoch": 0, "step": 1656, "train_loss": 3.8021345138549805, "perplexity": 44.79670169606627, "lr": 0.001305, "grad_norm": 0.259484, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:06.743111+00:00", "epoch": 0, "step": 1657, "train_loss": 3.8773033618927, "perplexity": 48.293808396497006, "lr": 0.001305, "grad_norm": 0.274428, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:07.049413+00:00", "epoch": 0, "step": 1658, "train_loss": 3.855360746383667, "perplexity": 47.24565756752923, "lr": 0.001305, "grad_norm": 0.284457, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:07.356416+00:00", "epoch": 0, "step": 1659, "train_loss": 3.806023597717285, "perplexity": 44.97125903977662, "lr": 0.001305, "grad_norm": 0.298766, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:07.663710+00:00", "epoch": 0, "step": 1660, "train_loss": 3.8513667583465576, "perplexity": 47.0573353051727, "lr": 0.001305, "grad_norm": 0.295871, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:07.971203+00:00", "epoch": 0, "step": 1661, "train_loss": 3.756047248840332, "perplexity": 42.778996614827335, "lr": 0.001305, "grad_norm": 0.288215, "tokens_per_sec": 106506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:08.277909+00:00", "epoch": 0, "step": 1662, "train_loss": 3.9398434162139893, "perplexity": 51.41055061158403, "lr": 0.001305, "grad_norm": 0.309541, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:08.583937+00:00", "epoch": 0, "step": 1663, "train_loss": 3.9130685329437256, "perplexity": 50.05230371350024, "lr": 0.001305, "grad_norm": 0.309232, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:08.889930+00:00", "epoch": 0, "step": 1664, "train_loss": 3.8734254837036133, "perplexity": 48.10689354100342, "lr": 0.001305, "grad_norm": 0.295858, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:09.196867+00:00", "epoch": 0, "step": 1665, "train_loss": 3.774061679840088, "perplexity": 43.556619081345445, "lr": 0.001305, "grad_norm": 0.282098, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:09.503609+00:00", "epoch": 0, "step": 1666, "train_loss": 3.7622249126434326, "perplexity": 43.044088856048475, "lr": 0.001305, "grad_norm": 0.281804, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:09.810812+00:00", "epoch": 0, "step": 1667, "train_loss": 3.7158048152923584, "perplexity": 41.091644964120086, "lr": 0.001305, "grad_norm": 0.26203, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:10.116738+00:00", "epoch": 0, "step": 1668, "train_loss": 3.786376476287842, "perplexity": 44.09632635114625, "lr": 0.001305, "grad_norm": 0.28509, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:10.422718+00:00", "epoch": 0, "step": 1669, "train_loss": 3.7669224739074707, "perplexity": 43.246766773763085, "lr": 0.001305, "grad_norm": 0.289939, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:10.728282+00:00", "epoch": 0, "step": 1670, "train_loss": 3.937223196029663, "perplexity": 51.27601997609643, "lr": 0.001305, "grad_norm": 0.256276, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:11.035209+00:00", "epoch": 0, "step": 1671, "train_loss": 3.9297149181365967, "perplexity": 50.89246708159821, "lr": 0.001305, "grad_norm": 0.296758, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:11.341818+00:00", "epoch": 0, "step": 1672, "train_loss": 3.927865505218506, "perplexity": 50.798432876383245, "lr": 0.001305, "grad_norm": 0.289065, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:11.650055+00:00", "epoch": 0, "step": 1673, "train_loss": 3.8579912185668945, "perplexity": 47.370099554376175, "lr": 0.001305, "grad_norm": 0.287504, "tokens_per_sec": 106309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:11.956210+00:00", "epoch": 0, "step": 1674, "train_loss": 3.830554962158203, "perplexity": 46.088108294396875, "lr": 0.001305, "grad_norm": 0.267988, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:12.263068+00:00", "epoch": 0, "step": 1675, "train_loss": 3.7342536449432373, "perplexity": 41.85677388727976, "lr": 0.001305, "grad_norm": 0.259843, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:12.568713+00:00", "epoch": 0, "step": 1676, "train_loss": 3.7741384506225586, "perplexity": 43.55996308543336, "lr": 0.001305, "grad_norm": 0.2387, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:12.875597+00:00", "epoch": 0, "step": 1677, "train_loss": 3.8667638301849365, "perplexity": 47.78748715292517, "lr": 0.001305, "grad_norm": 0.231652, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:13.182789+00:00", "epoch": 0, "step": 1678, "train_loss": 3.920365810394287, "perplexity": 50.41888515727756, "lr": 0.001305, "grad_norm": 0.228125, "tokens_per_sec": 106670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:13.490162+00:00", "epoch": 0, "step": 1679, "train_loss": 3.7661471366882324, "perplexity": 43.21324894136225, "lr": 0.001305, "grad_norm": 0.239053, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:13.795680+00:00", "epoch": 0, "step": 1680, "train_loss": 3.860140323638916, "perplexity": 47.47201234701545, "lr": 0.001305, "grad_norm": 0.250881, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:14.101620+00:00", "epoch": 0, "step": 1681, "train_loss": 3.739506483078003, "perplexity": 42.077219220296335, "lr": 0.001305, "grad_norm": 0.264576, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:14.407904+00:00", "epoch": 0, "step": 1682, "train_loss": 3.672706365585327, "perplexity": 39.35827974586766, "lr": 0.001305, "grad_norm": 0.261599, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:14.715593+00:00", "epoch": 0, "step": 1683, "train_loss": 3.7775073051452637, "perplexity": 43.70695772684148, "lr": 0.001305, "grad_norm": 0.266644, "tokens_per_sec": 106498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:15.021228+00:00", "epoch": 0, "step": 1684, "train_loss": 3.796267509460449, "perplexity": 44.5346487349849, "lr": 0.001305, "grad_norm": 0.241142, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:15.328179+00:00", "epoch": 0, "step": 1685, "train_loss": 3.7634403705596924, "perplexity": 43.09643894281554, "lr": 0.001305, "grad_norm": 0.231913, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:15.634125+00:00", "epoch": 0, "step": 1686, "train_loss": 3.781550168991089, "perplexity": 43.88401667749616, "lr": 0.001305, "grad_norm": 0.230047, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:15.940439+00:00", "epoch": 0, "step": 1687, "train_loss": 3.8873493671417236, "perplexity": 48.78141338946158, "lr": 0.001305, "grad_norm": 0.243786, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:16.246525+00:00", "epoch": 0, "step": 1688, "train_loss": 3.8136932849884033, "perplexity": 45.31750061783461, "lr": 0.001305, "grad_norm": 0.296555, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:16.554776+00:00", "epoch": 0, "step": 1689, "train_loss": 3.829066276550293, "perplexity": 46.019548635433736, "lr": 0.001305, "grad_norm": 0.35436, "tokens_per_sec": 106304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:16.862147+00:00", "epoch": 0, "step": 1690, "train_loss": 3.8106892108917236, "perplexity": 45.1815677665078, "lr": 0.001305, "grad_norm": 0.350105, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:17.169495+00:00", "epoch": 0, "step": 1691, "train_loss": 3.813032388687134, "perplexity": 45.287560344091375, "lr": 0.001305, "grad_norm": 0.378198, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:17.475849+00:00", "epoch": 0, "step": 1692, "train_loss": 3.7120423316955566, "perplexity": 40.9373288120177, "lr": 0.001305, "grad_norm": 0.389816, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:17.782344+00:00", "epoch": 0, "step": 1693, "train_loss": 3.746424436569214, "perplexity": 42.36931665921621, "lr": 0.001305, "grad_norm": 0.352309, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:18.088666+00:00", "epoch": 0, "step": 1694, "train_loss": 3.7728874683380127, "perplexity": 43.50550441381685, "lr": 0.001305, "grad_norm": 0.329132, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:18.395262+00:00", "epoch": 0, "step": 1695, "train_loss": 3.78609299659729, "perplexity": 44.083827709836946, "lr": 0.001305, "grad_norm": 0.344179, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:18.702024+00:00", "epoch": 0, "step": 1696, "train_loss": 3.868150234222412, "perplexity": 47.85378586583695, "lr": 0.001305, "grad_norm": 0.320756, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:19.008691+00:00", "epoch": 0, "step": 1697, "train_loss": 3.803957223892212, "perplexity": 44.8784275525257, "lr": 0.001305, "grad_norm": 0.3256, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:19.314700+00:00", "epoch": 0, "step": 1698, "train_loss": 3.8199689388275146, "perplexity": 45.602791822667186, "lr": 0.001305, "grad_norm": 0.354075, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:19.621353+00:00", "epoch": 0, "step": 1699, "train_loss": 3.765629529953003, "perplexity": 43.19088726043664, "lr": 0.001305, "grad_norm": 0.276888, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:19.928700+00:00", "epoch": 0, "step": 1700, "train_loss": 3.849963665008545, "perplexity": 46.9913557700488, "lr": 0.001305, "grad_norm": 0.2807, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:10:25.769820+00:00", "step": 1700, "epoch": 0, "val_loss": 3.8609135031700133, "val_ppl": 47.50873092846073, "eval_train_loss": 3.849963665008545, "eval_train_ppl": 46.9913557700488} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:10:26.735598+00:00", "step": 1700, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p8609_epoch_0000_step_0001700.pt", "val_loss": 3.8609135031700133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:27.808347+00:00", "epoch": 0, "step": 1701, "train_loss": 3.859851360321045, "perplexity": 47.4582966585825, "lr": 0.001305, "grad_norm": 0.282198, "tokens_per_sec": 4158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:28.114370+00:00", "epoch": 0, "step": 1702, "train_loss": 3.876859664916992, "perplexity": 48.272385332791806, "lr": 0.001305, "grad_norm": 0.254153, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:28.419595+00:00", "epoch": 0, "step": 1703, "train_loss": 3.7976832389831543, "perplexity": 44.597742403228565, "lr": 0.001305, "grad_norm": 0.285418, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:28.725020+00:00", "epoch": 0, "step": 1704, "train_loss": 3.854552745819092, "perplexity": 47.207498467904024, "lr": 0.001305, "grad_norm": 0.275904, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:29.029685+00:00", "epoch": 0, "step": 1705, "train_loss": 3.8616716861724854, "perplexity": 47.54476489916561, "lr": 0.001305, "grad_norm": 0.315225, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:29.334569+00:00", "epoch": 0, "step": 1706, "train_loss": 3.8533005714416504, "perplexity": 47.14842344173531, "lr": 0.001305, "grad_norm": 0.293941, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:29.640854+00:00", "epoch": 0, "step": 1707, "train_loss": 3.8215880393981934, "perplexity": 45.67668713475984, "lr": 0.001305, "grad_norm": 0.294745, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:29.946560+00:00", "epoch": 0, "step": 1708, "train_loss": 3.838740587234497, "perplexity": 46.46691654537852, "lr": 0.001305, "grad_norm": 0.279407, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:30.251775+00:00", "epoch": 0, "step": 1709, "train_loss": 3.902284622192383, "perplexity": 49.515444063360796, "lr": 0.001305, "grad_norm": 0.260568, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:30.556508+00:00", "epoch": 0, "step": 1710, "train_loss": 3.8379123210906982, "perplexity": 46.42844550592634, "lr": 0.001305, "grad_norm": 0.263994, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:30.862901+00:00", "epoch": 0, "step": 1711, "train_loss": 3.8688313961029053, "perplexity": 47.886393144762984, "lr": 0.001305, "grad_norm": 0.248291, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:31.168468+00:00", "epoch": 0, "step": 1712, "train_loss": 3.8174350261688232, "perplexity": 45.48738460884578, "lr": 0.001305, "grad_norm": 0.239987, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:31.474551+00:00", "epoch": 0, "step": 1713, "train_loss": 3.7929635047912598, "perplexity": 44.387748860195266, "lr": 0.001305, "grad_norm": 0.259495, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:31.780406+00:00", "epoch": 0, "step": 1714, "train_loss": 3.8714001178741455, "perplexity": 48.00955808588939, "lr": 0.001305, "grad_norm": 0.254822, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:32.121445+00:00", "epoch": 0, "step": 1715, "train_loss": 3.8697781562805176, "perplexity": 47.93175154321652, "lr": 0.001305, "grad_norm": 0.249018, "tokens_per_sec": 96083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:32.426428+00:00", "epoch": 0, "step": 1716, "train_loss": 3.840284585952759, "perplexity": 46.5387168204686, "lr": 0.001305, "grad_norm": 0.233033, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:32.732604+00:00", "epoch": 0, "step": 1717, "train_loss": 3.774111747741699, "perplexity": 43.558799924458796, "lr": 0.001305, "grad_norm": 0.239839, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:33.039187+00:00", "epoch": 0, "step": 1718, "train_loss": 3.8530619144439697, "perplexity": 47.137172483164726, "lr": 0.001305, "grad_norm": 0.2534, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:33.345846+00:00", "epoch": 0, "step": 1719, "train_loss": 3.7885520458221436, "perplexity": 44.19236540727252, "lr": 0.001305, "grad_norm": 0.245348, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:33.651121+00:00", "epoch": 0, "step": 1720, "train_loss": 3.7780728340148926, "perplexity": 43.73168226380042, "lr": 0.001305, "grad_norm": 0.221975, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:33.956911+00:00", "epoch": 0, "step": 1721, "train_loss": 3.8273868560791016, "perplexity": 45.942327325055054, "lr": 0.001305, "grad_norm": 0.242619, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:34.262701+00:00", "epoch": 0, "step": 1722, "train_loss": 3.755051612854004, "perplexity": 42.736425502520994, "lr": 0.001305, "grad_norm": 0.244381, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:34.570627+00:00", "epoch": 0, "step": 1723, "train_loss": 3.7614030838012695, "perplexity": 43.008728514406464, "lr": 0.001305, "grad_norm": 0.246368, "tokens_per_sec": 106414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:34.877311+00:00", "epoch": 0, "step": 1724, "train_loss": 3.7798385620117188, "perplexity": 43.808968732882754, "lr": 0.001305, "grad_norm": 0.281605, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:35.182713+00:00", "epoch": 0, "step": 1725, "train_loss": 3.864368438720703, "perplexity": 47.67315440461574, "lr": 0.001305, "grad_norm": 0.273975, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:35.488449+00:00", "epoch": 0, "step": 1726, "train_loss": 3.793869733810425, "perplexity": 44.42799255855725, "lr": 0.001305, "grad_norm": 0.226373, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:35.794640+00:00", "epoch": 0, "step": 1727, "train_loss": 3.789435625076294, "perplexity": 44.231430120381724, "lr": 0.001305, "grad_norm": 0.232585, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:36.100876+00:00", "epoch": 0, "step": 1728, "train_loss": 3.7268829345703125, "perplexity": 41.54939392574127, "lr": 0.001305, "grad_norm": 0.254238, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:36.407569+00:00", "epoch": 0, "step": 1729, "train_loss": 3.893671751022339, "perplexity": 49.09080522753626, "lr": 0.001305, "grad_norm": 0.238736, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:36.713093+00:00", "epoch": 0, "step": 1730, "train_loss": 3.7861669063568115, "perplexity": 44.08708605535205, "lr": 0.001305, "grad_norm": 0.227381, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:37.018729+00:00", "epoch": 0, "step": 1731, "train_loss": 3.8666281700134277, "perplexity": 47.78100473393501, "lr": 0.001305, "grad_norm": 0.267239, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:37.324613+00:00", "epoch": 0, "step": 1732, "train_loss": 3.891294479370117, "perplexity": 48.97424165441894, "lr": 0.001305, "grad_norm": 0.29912, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:37.631682+00:00", "epoch": 0, "step": 1733, "train_loss": 3.877856731414795, "perplexity": 48.32054011374617, "lr": 0.001305, "grad_norm": 0.313229, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:37.938745+00:00", "epoch": 0, "step": 1734, "train_loss": 3.8406457901000977, "perplexity": 46.55552983427834, "lr": 0.001305, "grad_norm": 0.270921, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:38.245271+00:00", "epoch": 0, "step": 1735, "train_loss": 3.8183481693267822, "perplexity": 45.52894007303117, "lr": 0.001305, "grad_norm": 0.261089, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:38.551092+00:00", "epoch": 0, "step": 1736, "train_loss": 3.756075143814087, "perplexity": 42.78018995045912, "lr": 0.001305, "grad_norm": 0.275303, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:38.857223+00:00", "epoch": 0, "step": 1737, "train_loss": 3.851613759994507, "perplexity": 47.06895998013901, "lr": 0.001305, "grad_norm": 0.252472, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:39.164386+00:00", "epoch": 0, "step": 1738, "train_loss": 3.777238607406616, "perplexity": 43.69521534378365, "lr": 0.001305, "grad_norm": 0.224161, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:39.473103+00:00", "epoch": 0, "step": 1739, "train_loss": 3.8781113624572754, "perplexity": 48.33284558986017, "lr": 0.001305, "grad_norm": 0.234784, "tokens_per_sec": 106143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:39.778499+00:00", "epoch": 0, "step": 1740, "train_loss": 3.740718126296997, "perplexity": 42.12823269646399, "lr": 0.001305, "grad_norm": 0.254007, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:40.084185+00:00", "epoch": 0, "step": 1741, "train_loss": 3.7436282634735107, "perplexity": 42.25101019564803, "lr": 0.001305, "grad_norm": 0.256049, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:40.390825+00:00", "epoch": 0, "step": 1742, "train_loss": 3.7759339809417725, "perplexity": 43.63824657901851, "lr": 0.001305, "grad_norm": 0.241002, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:40.697463+00:00", "epoch": 0, "step": 1743, "train_loss": 3.7896177768707275, "perplexity": 44.239487688576496, "lr": 0.001305, "grad_norm": 0.250458, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:41.003644+00:00", "epoch": 0, "step": 1744, "train_loss": 3.9118216037750244, "perplexity": 49.98993093134149, "lr": 0.001305, "grad_norm": 0.279423, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:41.309887+00:00", "epoch": 0, "step": 1745, "train_loss": 3.8031551837921143, "perplexity": 44.84244768457642, "lr": 0.001305, "grad_norm": 0.249844, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:41.615764+00:00", "epoch": 0, "step": 1746, "train_loss": 3.7866063117980957, "perplexity": 44.106462417582875, "lr": 0.001305, "grad_norm": 0.234417, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:41.922824+00:00", "epoch": 0, "step": 1747, "train_loss": 3.7898197174072266, "perplexity": 44.24842233655821, "lr": 0.001305, "grad_norm": 0.232468, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:42.228649+00:00", "epoch": 0, "step": 1748, "train_loss": 3.7737507820129395, "perplexity": 43.543079527933045, "lr": 0.001305, "grad_norm": 0.232825, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:42.536066+00:00", "epoch": 0, "step": 1749, "train_loss": 3.824381113052368, "perplexity": 45.8044438200666, "lr": 0.001305, "grad_norm": 0.241923, "tokens_per_sec": 106591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:42.842992+00:00", "epoch": 0, "step": 1750, "train_loss": 3.7585830688476562, "perplexity": 42.88761410934, "lr": 0.001305, "grad_norm": 0.244008, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:10:43.974107+00:00", "step": 1750, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0000_step_0001750.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:45.167446+00:00", "epoch": 0, "step": 1751, "train_loss": 3.8452751636505127, "perplexity": 46.77155241152143, "lr": 0.001305, "grad_norm": 0.26287, "tokens_per_sec": 14096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:45.472819+00:00", "epoch": 0, "step": 1752, "train_loss": 3.826101779937744, "perplexity": 45.88332585514152, "lr": 0.001305, "grad_norm": 0.263423, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:45.779367+00:00", "epoch": 0, "step": 1753, "train_loss": 3.8416426181793213, "perplexity": 46.601960831676514, "lr": 0.001305, "grad_norm": 0.273222, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:46.085934+00:00", "epoch": 0, "step": 1754, "train_loss": 3.7769954204559326, "perplexity": 43.684590529565185, "lr": 0.001305, "grad_norm": 0.288623, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:46.392423+00:00", "epoch": 0, "step": 1755, "train_loss": 3.8203790187835693, "perplexity": 45.62149644846774, "lr": 0.001305, "grad_norm": 0.303008, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:46.697444+00:00", "epoch": 0, "step": 1756, "train_loss": 3.902008533477783, "perplexity": 49.50177529503967, "lr": 0.001305, "grad_norm": 0.301853, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:47.003153+00:00", "epoch": 0, "step": 1757, "train_loss": 3.7493231296539307, "perplexity": 42.49231047896736, "lr": 0.001305, "grad_norm": 0.284268, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:47.308767+00:00", "epoch": 0, "step": 1758, "train_loss": 3.7748916149139404, "perplexity": 43.59278325210055, "lr": 0.001305, "grad_norm": 0.267805, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:47.615156+00:00", "epoch": 0, "step": 1759, "train_loss": 3.816617488861084, "perplexity": 45.45021217189398, "lr": 0.001305, "grad_norm": 0.264159, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:47.922108+00:00", "epoch": 0, "step": 1760, "train_loss": 3.9668290615081787, "perplexity": 52.81678624581416, "lr": 0.001305, "grad_norm": 0.254821, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:48.228618+00:00", "epoch": 0, "step": 1761, "train_loss": 3.8173160552978516, "perplexity": 45.48197325698364, "lr": 0.001305, "grad_norm": 0.266121, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:48.534799+00:00", "epoch": 0, "step": 1762, "train_loss": 3.783352851867676, "perplexity": 43.96319698997081, "lr": 0.001305, "grad_norm": 0.284974, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:48.841376+00:00", "epoch": 0, "step": 1763, "train_loss": 3.9114112854003906, "perplexity": 49.96942335173978, "lr": 0.001305, "grad_norm": 0.275876, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:49.565339+00:00", "epoch": 0, "step": 1764, "train_loss": 3.878417730331421, "perplexity": 48.347655489537836, "lr": 0.001305, "grad_norm": 0.255016, "tokens_per_sec": 45262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:49.871244+00:00", "epoch": 0, "step": 1765, "train_loss": 3.806636333465576, "perplexity": 44.99882298167983, "lr": 0.001305, "grad_norm": 0.268223, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:50.177000+00:00", "epoch": 0, "step": 1766, "train_loss": 3.764983654022217, "perplexity": 43.16300031264944, "lr": 0.001305, "grad_norm": 0.304129, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:50.483551+00:00", "epoch": 0, "step": 1767, "train_loss": 3.911195755004883, "perplexity": 49.95865458270169, "lr": 0.001305, "grad_norm": 0.321515, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:50.790525+00:00", "epoch": 0, "step": 1768, "train_loss": 3.848334312438965, "perplexity": 46.91485262599897, "lr": 0.001305, "grad_norm": 0.290896, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:51.096991+00:00", "epoch": 0, "step": 1769, "train_loss": 3.8541061878204346, "perplexity": 47.18642228808516, "lr": 0.001305, "grad_norm": 0.338761, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:51.402919+00:00", "epoch": 0, "step": 1770, "train_loss": 3.8276522159576416, "perplexity": 45.95452019313132, "lr": 0.001305, "grad_norm": 0.318126, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:51.708834+00:00", "epoch": 0, "step": 1771, "train_loss": 3.801088333129883, "perplexity": 44.74986075652021, "lr": 0.001305, "grad_norm": 0.277331, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:52.014002+00:00", "epoch": 0, "step": 1772, "train_loss": 3.8065216541290283, "perplexity": 44.993662842401186, "lr": 0.001305, "grad_norm": 0.305123, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:52.320064+00:00", "epoch": 0, "step": 1773, "train_loss": 3.7875564098358154, "perplexity": 44.148387794431926, "lr": 0.001305, "grad_norm": 0.307402, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:52.626506+00:00", "epoch": 0, "step": 1774, "train_loss": 3.8078036308288574, "perplexity": 45.05138065834795, "lr": 0.001305, "grad_norm": 0.35393, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:52.932969+00:00", "epoch": 0, "step": 1775, "train_loss": 3.9418551921844482, "perplexity": 51.51408122721706, "lr": 0.001305, "grad_norm": 0.310709, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:53.238732+00:00", "epoch": 0, "step": 1776, "train_loss": 3.7719995975494385, "perplexity": 43.46689429024248, "lr": 0.001305, "grad_norm": 0.292975, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:53.544417+00:00", "epoch": 0, "step": 1777, "train_loss": 3.9057910442352295, "perplexity": 49.6893708600604, "lr": 0.001305, "grad_norm": 0.303862, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:53.851226+00:00", "epoch": 0, "step": 1778, "train_loss": 3.9101037979125977, "perplexity": 49.90413164927863, "lr": 0.001305, "grad_norm": 0.275214, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:54.159053+00:00", "epoch": 0, "step": 1779, "train_loss": 3.957690954208374, "perplexity": 52.3363393169685, "lr": 0.001305, "grad_norm": 0.262324, "tokens_per_sec": 106451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:54.466339+00:00", "epoch": 0, "step": 1780, "train_loss": 3.7308897972106934, "perplexity": 41.71621062260195, "lr": 0.001305, "grad_norm": 0.271692, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:54.773183+00:00", "epoch": 0, "step": 1781, "train_loss": 3.8202929496765137, "perplexity": 45.61757001598057, "lr": 0.001305, "grad_norm": 0.291804, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:55.079549+00:00", "epoch": 0, "step": 1782, "train_loss": 3.726994514465332, "perplexity": 41.55403026140974, "lr": 0.001305, "grad_norm": 0.286963, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:55.410004+00:00", "epoch": 0, "step": 1783, "train_loss": 3.891400098800659, "perplexity": 48.97941455910852, "lr": 0.001305, "grad_norm": 0.30258, "tokens_per_sec": 99161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:55.716792+00:00", "epoch": 0, "step": 1784, "train_loss": 3.869588851928711, "perplexity": 47.92267871284991, "lr": 0.001305, "grad_norm": 0.248071, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:56.024237+00:00", "epoch": 0, "step": 1785, "train_loss": 3.8030261993408203, "perplexity": 44.836664079072754, "lr": 0.001305, "grad_norm": 0.239035, "tokens_per_sec": 106586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:56.331231+00:00", "epoch": 0, "step": 1786, "train_loss": 3.7852210998535156, "perplexity": 44.04540791548433, "lr": 0.001305, "grad_norm": 0.269382, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:56.637647+00:00", "epoch": 0, "step": 1787, "train_loss": 3.8331894874572754, "perplexity": 46.20968866464419, "lr": 0.001305, "grad_norm": 0.263057, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:56.943955+00:00", "epoch": 0, "step": 1788, "train_loss": 3.824859142303467, "perplexity": 45.826344918308635, "lr": 0.001305, "grad_norm": 0.253244, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:57.249748+00:00", "epoch": 0, "step": 1789, "train_loss": 3.7850563526153564, "perplexity": 44.03815215387641, "lr": 0.001305, "grad_norm": 0.260315, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:57.556709+00:00", "epoch": 0, "step": 1790, "train_loss": 3.807689666748047, "perplexity": 45.04624671171029, "lr": 0.001305, "grad_norm": 0.239162, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:57.863802+00:00", "epoch": 0, "step": 1791, "train_loss": 3.7097835540771484, "perplexity": 40.84496484404217, "lr": 0.001305, "grad_norm": 0.29537, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:58.169007+00:00", "epoch": 0, "step": 1792, "train_loss": 3.9516310691833496, "perplexity": 52.020146132856404, "lr": 0.001305, "grad_norm": 0.32742, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:58.475453+00:00", "epoch": 0, "step": 1793, "train_loss": 3.834636926651001, "perplexity": 46.2766228090132, "lr": 0.001305, "grad_norm": 0.311901, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:58.780861+00:00", "epoch": 0, "step": 1794, "train_loss": 3.867267370223999, "perplexity": 47.811556125409965, "lr": 0.001305, "grad_norm": 0.275547, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:59.086292+00:00", "epoch": 0, "step": 1795, "train_loss": 3.775338888168335, "perplexity": 43.61228549922548, "lr": 0.001305, "grad_norm": 0.282453, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:59.392746+00:00", "epoch": 0, "step": 1796, "train_loss": 3.6850030422210693, "perplexity": 39.845243667716815, "lr": 0.001305, "grad_norm": 0.297102, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:10:59.699770+00:00", "epoch": 0, "step": 1797, "train_loss": 3.7878470420837402, "perplexity": 44.161220604342745, "lr": 0.001305, "grad_norm": 0.268196, "tokens_per_sec": 106727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:00.007165+00:00", "epoch": 0, "step": 1798, "train_loss": 3.7908196449279785, "perplexity": 44.29268968018155, "lr": 0.001305, "grad_norm": 0.26232, "tokens_per_sec": 106600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:00.314153+00:00", "epoch": 0, "step": 1799, "train_loss": 3.822765350341797, "perplexity": 45.73049446615617, "lr": 0.001305, "grad_norm": 0.257273, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:00.621668+00:00", "epoch": 0, "step": 1800, "train_loss": 3.755681037902832, "perplexity": 42.763333346576836, "lr": 0.001305, "grad_norm": 0.280209, "tokens_per_sec": 106614} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:11:06.527270+00:00", "step": 1800, "epoch": 0, "val_loss": 3.8397106170654296, "val_ppl": 46.51201270935501, "eval_train_loss": 3.755681037902832, "eval_train_ppl": 42.763333346576836} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:11:07.505960+00:00", "step": 1800, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p8397_epoch_0000_step_0001800.pt", "val_loss": 3.8397106170654296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:08.527329+00:00", "epoch": 0, "step": 1801, "train_loss": 3.6708128452301025, "perplexity": 39.28382455547761, "lr": 0.001305, "grad_norm": 0.24929, "tokens_per_sec": 4145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:08.831521+00:00", "epoch": 0, "step": 1802, "train_loss": 3.7782325744628906, "perplexity": 43.73866854029756, "lr": 0.001305, "grad_norm": 0.221744, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:09.136535+00:00", "epoch": 0, "step": 1803, "train_loss": 3.933375358581543, "perplexity": 51.079097292447415, "lr": 0.001305, "grad_norm": 0.257941, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:09.442929+00:00", "epoch": 0, "step": 1804, "train_loss": 3.7672388553619385, "perplexity": 43.260451413404944, "lr": 0.001305, "grad_norm": 0.275843, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:09.748353+00:00", "epoch": 0, "step": 1805, "train_loss": 3.827801465988159, "perplexity": 45.96137941852964, "lr": 0.001305, "grad_norm": 0.231451, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:10.053641+00:00", "epoch": 0, "step": 1806, "train_loss": 3.7536544799804688, "perplexity": 42.67675872846548, "lr": 0.001305, "grad_norm": 0.244555, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:10.359730+00:00", "epoch": 0, "step": 1807, "train_loss": 3.786804676055908, "perplexity": 44.115212431081346, "lr": 0.001305, "grad_norm": 0.257164, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:10.665369+00:00", "epoch": 0, "step": 1808, "train_loss": 3.768270254135132, "perplexity": 43.30509320770903, "lr": 0.001305, "grad_norm": 0.252703, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:10.971775+00:00", "epoch": 0, "step": 1809, "train_loss": 3.81628680229187, "perplexity": 45.43518488195923, "lr": 0.001305, "grad_norm": 0.253805, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:11.277327+00:00", "epoch": 0, "step": 1810, "train_loss": 3.876113176345825, "perplexity": 48.23636399526902, "lr": 0.001305, "grad_norm": 0.26257, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:11.582563+00:00", "epoch": 0, "step": 1811, "train_loss": 3.777804374694824, "perplexity": 43.719943661853485, "lr": 0.001305, "grad_norm": 0.244661, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:11.888829+00:00", "epoch": 0, "step": 1812, "train_loss": 3.7626824378967285, "perplexity": 43.06378711958812, "lr": 0.001305, "grad_norm": 0.267548, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:12.195380+00:00", "epoch": 0, "step": 1813, "train_loss": 3.6370198726654053, "perplexity": 37.97848718347246, "lr": 0.001305, "grad_norm": 0.283704, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:12.502175+00:00", "epoch": 0, "step": 1814, "train_loss": 3.7708871364593506, "perplexity": 43.41856594832173, "lr": 0.001305, "grad_norm": 0.289086, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:12.869445+00:00", "epoch": 0, "step": 1815, "train_loss": 3.8446738719940186, "perplexity": 46.7434375207668, "lr": 0.001305, "grad_norm": 0.294127, "tokens_per_sec": 89220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:13.174797+00:00", "epoch": 0, "step": 1816, "train_loss": 3.700044870376587, "perplexity": 40.449119286563864, "lr": 0.001305, "grad_norm": 0.258346, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:13.480731+00:00", "epoch": 0, "step": 1817, "train_loss": 3.751025438308716, "perplexity": 42.56470711008048, "lr": 0.001305, "grad_norm": 0.248295, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:13.787251+00:00", "epoch": 0, "step": 1818, "train_loss": 3.6707088947296143, "perplexity": 39.27974119449136, "lr": 0.001305, "grad_norm": 0.284843, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:14.093142+00:00", "epoch": 0, "step": 1819, "train_loss": 3.781254529953003, "perplexity": 43.87104476661439, "lr": 0.001305, "grad_norm": 0.283256, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:14.398974+00:00", "epoch": 0, "step": 1820, "train_loss": 3.7973673343658447, "perplexity": 44.58365599557956, "lr": 0.001305, "grad_norm": 0.272607, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:14.705059+00:00", "epoch": 0, "step": 1821, "train_loss": 3.9641337394714355, "perplexity": 52.674619676346815, "lr": 0.001305, "grad_norm": 0.240803, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:15.011108+00:00", "epoch": 0, "step": 1822, "train_loss": 3.8306076526641846, "perplexity": 46.09053676412073, "lr": 0.001305, "grad_norm": 0.249246, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:15.318186+00:00", "epoch": 0, "step": 1823, "train_loss": 3.8235247135162354, "perplexity": 45.765233707783715, "lr": 0.001305, "grad_norm": 0.250235, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:15.623962+00:00", "epoch": 0, "step": 1824, "train_loss": 3.8122053146362305, "perplexity": 45.25011966334409, "lr": 0.001305, "grad_norm": 0.247704, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:15.930399+00:00", "epoch": 0, "step": 1825, "train_loss": 3.7935633659362793, "perplexity": 44.41438333374562, "lr": 0.001305, "grad_norm": 0.254641, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:16.236035+00:00", "epoch": 0, "step": 1826, "train_loss": 3.7592766284942627, "perplexity": 42.917369545217504, "lr": 0.001305, "grad_norm": 0.247949, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:16.542800+00:00", "epoch": 0, "step": 1827, "train_loss": 3.841779947280884, "perplexity": 46.60836107654846, "lr": 0.001305, "grad_norm": 0.2522, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:16.849716+00:00", "epoch": 0, "step": 1828, "train_loss": 3.748512029647827, "perplexity": 42.457858939389254, "lr": 0.001305, "grad_norm": 0.272572, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:17.157078+00:00", "epoch": 0, "step": 1829, "train_loss": 3.902437686920166, "perplexity": 49.523023711401, "lr": 0.001305, "grad_norm": 0.265959, "tokens_per_sec": 106611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:17.463484+00:00", "epoch": 0, "step": 1830, "train_loss": 3.8157825469970703, "perplexity": 45.412279724921, "lr": 0.001305, "grad_norm": 0.250006, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:17.770994+00:00", "epoch": 0, "step": 1831, "train_loss": 3.760984182357788, "perplexity": 42.99071586897447, "lr": 0.001305, "grad_norm": 0.252333, "tokens_per_sec": 106503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:18.077102+00:00", "epoch": 0, "step": 1832, "train_loss": 3.8704662322998047, "perplexity": 47.96474358122926, "lr": 0.001305, "grad_norm": 0.250424, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:18.383947+00:00", "epoch": 0, "step": 1833, "train_loss": 3.8523592948913574, "perplexity": 47.104064616602436, "lr": 0.001305, "grad_norm": 0.242249, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:18.690404+00:00", "epoch": 0, "step": 1834, "train_loss": 3.749450206756592, "perplexity": 42.49771062177839, "lr": 0.001305, "grad_norm": 0.29341, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:18.997503+00:00", "epoch": 0, "step": 1835, "train_loss": 3.708146572113037, "perplexity": 40.77815706974729, "lr": 0.001305, "grad_norm": 0.284244, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:19.303688+00:00", "epoch": 0, "step": 1836, "train_loss": 3.676281452178955, "perplexity": 39.49924082803478, "lr": 0.001305, "grad_norm": 0.256568, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:19.610965+00:00", "epoch": 0, "step": 1837, "train_loss": 3.731905698776245, "perplexity": 41.75861172030642, "lr": 0.001305, "grad_norm": 0.259493, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:19.917209+00:00", "epoch": 0, "step": 1838, "train_loss": 3.7446188926696777, "perplexity": 42.292886018197045, "lr": 0.001305, "grad_norm": 0.249935, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:20.224318+00:00", "epoch": 0, "step": 1839, "train_loss": 3.8201963901519775, "perplexity": 45.61316541776571, "lr": 0.001305, "grad_norm": 0.293767, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:20.531859+00:00", "epoch": 0, "step": 1840, "train_loss": 3.750722885131836, "perplexity": 42.55183097067802, "lr": 0.001305, "grad_norm": 0.304084, "tokens_per_sec": 106608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:20.839726+00:00", "epoch": 0, "step": 1841, "train_loss": 3.8120005130767822, "perplexity": 45.24085331718528, "lr": 0.001305, "grad_norm": 0.304431, "tokens_per_sec": 106377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:21.146391+00:00", "epoch": 0, "step": 1842, "train_loss": 3.806159257888794, "perplexity": 44.97736026232801, "lr": 0.001305, "grad_norm": 0.26187, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:21.452729+00:00", "epoch": 0, "step": 1843, "train_loss": 3.686312675476074, "perplexity": 39.89746050887039, "lr": 0.001305, "grad_norm": 0.252664, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:21.758963+00:00", "epoch": 0, "step": 1844, "train_loss": 3.6691203117370605, "perplexity": 39.21739160254066, "lr": 0.001305, "grad_norm": 0.266435, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:22.064521+00:00", "epoch": 0, "step": 1845, "train_loss": 3.751953125, "perplexity": 42.60421214369869, "lr": 0.001305, "grad_norm": 0.269898, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:22.371302+00:00", "epoch": 0, "step": 1846, "train_loss": 3.8294360637664795, "perplexity": 46.036569223017686, "lr": 0.001305, "grad_norm": 0.252289, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:22.677724+00:00", "epoch": 0, "step": 1847, "train_loss": 3.8636341094970703, "perplexity": 47.63815946462978, "lr": 0.001305, "grad_norm": 0.219905, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:22.983204+00:00", "epoch": 0, "step": 1848, "train_loss": 3.7671003341674805, "perplexity": 43.254459339026596, "lr": 0.001305, "grad_norm": 0.247949, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:23.289135+00:00", "epoch": 0, "step": 1849, "train_loss": 3.8098156452178955, "perplexity": 45.142115934209336, "lr": 0.001305, "grad_norm": 0.254376, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:23.595802+00:00", "epoch": 0, "step": 1850, "train_loss": 3.737637758255005, "perplexity": 41.9986619001248, "lr": 0.001305, "grad_norm": 0.249165, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:23.902465+00:00", "epoch": 0, "step": 1851, "train_loss": 3.8195836544036865, "perplexity": 45.58522516159259, "lr": 0.001305, "grad_norm": 0.226283, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:24.209278+00:00", "epoch": 0, "step": 1852, "train_loss": 3.817837715148926, "perplexity": 45.50570556593763, "lr": 0.001305, "grad_norm": 0.235502, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:24.515995+00:00", "epoch": 0, "step": 1853, "train_loss": 3.7788631916046143, "perplexity": 43.76625959321634, "lr": 0.001305, "grad_norm": 0.250013, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:24.821751+00:00", "epoch": 0, "step": 1854, "train_loss": 3.7264225482940674, "perplexity": 41.53026955762766, "lr": 0.001305, "grad_norm": 0.266421, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:25.128272+00:00", "epoch": 0, "step": 1855, "train_loss": 3.710932731628418, "perplexity": 40.89192994117244, "lr": 0.001305, "grad_norm": 0.318203, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:25.434936+00:00", "epoch": 0, "step": 1856, "train_loss": 3.7301650047302246, "perplexity": 41.6859859814485, "lr": 0.001305, "grad_norm": 0.388492, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:25.742459+00:00", "epoch": 0, "step": 1857, "train_loss": 3.91791033744812, "perplexity": 50.29523482095334, "lr": 0.001305, "grad_norm": 0.413715, "tokens_per_sec": 106549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:26.049417+00:00", "epoch": 0, "step": 1858, "train_loss": 3.8514392375946045, "perplexity": 47.060746109055465, "lr": 0.001305, "grad_norm": 0.382095, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:26.356199+00:00", "epoch": 0, "step": 1859, "train_loss": 3.8255233764648438, "perplexity": 45.85679445378571, "lr": 0.001305, "grad_norm": 0.398813, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:26.662547+00:00", "epoch": 0, "step": 1860, "train_loss": 3.7277233600616455, "perplexity": 41.58432777313764, "lr": 0.001305, "grad_norm": 0.39649, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:26.968191+00:00", "epoch": 0, "step": 1861, "train_loss": 3.7786970138549805, "perplexity": 43.75898721895739, "lr": 0.001305, "grad_norm": 0.32285, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:27.274900+00:00", "epoch": 0, "step": 1862, "train_loss": 3.7078757286071777, "perplexity": 40.76711406625443, "lr": 0.001305, "grad_norm": 0.336292, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:27.582728+00:00", "epoch": 0, "step": 1863, "train_loss": 3.8299570083618164, "perplexity": 46.06055797280454, "lr": 0.001305, "grad_norm": 0.331992, "tokens_per_sec": 106451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:27.889089+00:00", "epoch": 0, "step": 1864, "train_loss": 3.778010368347168, "perplexity": 43.72895062038494, "lr": 0.001305, "grad_norm": 0.325156, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:28.195544+00:00", "epoch": 0, "step": 1865, "train_loss": 3.793804168701172, "perplexity": 44.42507972786234, "lr": 0.001305, "grad_norm": 0.28789, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:28.502720+00:00", "epoch": 0, "step": 1866, "train_loss": 3.807375907897949, "perplexity": 45.03211527018916, "lr": 0.001305, "grad_norm": 0.291136, "tokens_per_sec": 106679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:28.809775+00:00", "epoch": 0, "step": 1867, "train_loss": 3.9368269443511963, "perplexity": 51.25570579214635, "lr": 0.001305, "grad_norm": 0.256664, "tokens_per_sec": 106714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:29.117353+00:00", "epoch": 0, "step": 1868, "train_loss": 3.7444465160369873, "perplexity": 42.285596341221506, "lr": 0.001305, "grad_norm": 0.266278, "tokens_per_sec": 106536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:29.424372+00:00", "epoch": 0, "step": 1869, "train_loss": 3.7651658058166504, "perplexity": 43.170863246711555, "lr": 0.001305, "grad_norm": 0.273221, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:29.730297+00:00", "epoch": 0, "step": 1870, "train_loss": 3.9266605377197266, "perplexity": 50.73725927927816, "lr": 0.001305, "grad_norm": 0.260003, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:30.036318+00:00", "epoch": 0, "step": 1871, "train_loss": 3.802525758743286, "perplexity": 44.81423160563796, "lr": 0.001305, "grad_norm": 0.241282, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:30.343324+00:00", "epoch": 0, "step": 1872, "train_loss": 3.735825538635254, "perplexity": 41.92262002412692, "lr": 0.001305, "grad_norm": 0.243245, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:30.650442+00:00", "epoch": 0, "step": 1873, "train_loss": 3.727058172225952, "perplexity": 41.55667558211763, "lr": 0.001305, "grad_norm": 0.233467, "tokens_per_sec": 106696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:30.957652+00:00", "epoch": 0, "step": 1874, "train_loss": 3.8302199840545654, "perplexity": 46.07267237277239, "lr": 0.001305, "grad_norm": 0.236547, "tokens_per_sec": 106664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:31.265126+00:00", "epoch": 0, "step": 1875, "train_loss": 3.6525838375091553, "perplexity": 38.57420687760545, "lr": 0.001305, "grad_norm": 0.244628, "tokens_per_sec": 106572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:31.571036+00:00", "epoch": 0, "step": 1876, "train_loss": 3.7482378482818604, "perplexity": 42.446219381377105, "lr": 0.001305, "grad_norm": 0.26184, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:31.876981+00:00", "epoch": 0, "step": 1877, "train_loss": 3.667057991027832, "perplexity": 39.13659610542666, "lr": 0.001305, "grad_norm": 0.269198, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:32.182756+00:00", "epoch": 0, "step": 1878, "train_loss": 3.739840030670166, "perplexity": 42.091256316341834, "lr": 0.001305, "grad_norm": 0.245275, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:32.489803+00:00", "epoch": 0, "step": 1879, "train_loss": 3.8202359676361084, "perplexity": 45.61497070782037, "lr": 0.001305, "grad_norm": 0.229129, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:32.796949+00:00", "epoch": 0, "step": 1880, "train_loss": 3.8509721755981445, "perplexity": 47.03877095530195, "lr": 0.001305, "grad_norm": 0.238998, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:33.103488+00:00", "epoch": 0, "step": 1881, "train_loss": 3.708238363265991, "perplexity": 40.781900315595855, "lr": 0.001305, "grad_norm": 0.245476, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:33.410242+00:00", "epoch": 0, "step": 1882, "train_loss": 3.7655951976776123, "perplexity": 43.189404444455214, "lr": 0.001305, "grad_norm": 0.242098, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:33.717859+00:00", "epoch": 0, "step": 1883, "train_loss": 3.798445463180542, "perplexity": 44.63174884025532, "lr": 0.001305, "grad_norm": 0.251301, "tokens_per_sec": 106522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:34.025139+00:00", "epoch": 0, "step": 1884, "train_loss": 3.7146613597869873, "perplexity": 41.044685349692536, "lr": 0.001305, "grad_norm": 0.237343, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:34.333220+00:00", "epoch": 0, "step": 1885, "train_loss": 3.745542526245117, "perplexity": 42.33196719329008, "lr": 0.001305, "grad_norm": 0.260667, "tokens_per_sec": 106362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:34.639928+00:00", "epoch": 0, "step": 1886, "train_loss": 3.7713003158569336, "perplexity": 43.43650931190298, "lr": 0.001305, "grad_norm": 0.249928, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:34.946545+00:00", "epoch": 0, "step": 1887, "train_loss": 3.837488889694214, "perplexity": 46.40879040599605, "lr": 0.001305, "grad_norm": 0.281529, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:35.253460+00:00", "epoch": 0, "step": 1888, "train_loss": 3.7515294551849365, "perplexity": 42.58616584812414, "lr": 0.001305, "grad_norm": 0.235885, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:35.561847+00:00", "epoch": 0, "step": 1889, "train_loss": 3.8078880310058594, "perplexity": 45.05518316331349, "lr": 0.001305, "grad_norm": 0.225167, "tokens_per_sec": 106256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:35.868977+00:00", "epoch": 0, "step": 1890, "train_loss": 3.8086118698120117, "perplexity": 45.08780765931907, "lr": 0.001305, "grad_norm": 0.242944, "tokens_per_sec": 106747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:36.177777+00:00", "epoch": 0, "step": 1891, "train_loss": 3.742417573928833, "perplexity": 42.199888291973636, "lr": 0.001305, "grad_norm": 0.260822, "tokens_per_sec": 106059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:36.483986+00:00", "epoch": 0, "step": 1892, "train_loss": 3.7409305572509766, "perplexity": 42.13718298775097, "lr": 0.001305, "grad_norm": 0.259829, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:36.791814+00:00", "epoch": 0, "step": 1893, "train_loss": 3.899059295654297, "perplexity": 49.35599785883853, "lr": 0.001305, "grad_norm": 0.258276, "tokens_per_sec": 106449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:37.098543+00:00", "epoch": 0, "step": 1894, "train_loss": 3.755934476852417, "perplexity": 42.77417261434924, "lr": 0.001305, "grad_norm": 0.283959, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:37.405887+00:00", "epoch": 0, "step": 1895, "train_loss": 3.766242742538452, "perplexity": 43.21738057826923, "lr": 0.001305, "grad_norm": 0.261784, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:37.712623+00:00", "epoch": 0, "step": 1896, "train_loss": 3.753488779067993, "perplexity": 42.66968773645376, "lr": 0.001305, "grad_norm": 0.264501, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:38.019916+00:00", "epoch": 0, "step": 1897, "train_loss": 3.8402180671691895, "perplexity": 46.5356212245956, "lr": 0.001305, "grad_norm": 0.290857, "tokens_per_sec": 106635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:38.327457+00:00", "epoch": 0, "step": 1898, "train_loss": 3.619370460510254, "perplexity": 37.31406974591809, "lr": 0.001305, "grad_norm": 0.303806, "tokens_per_sec": 106548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:38.634960+00:00", "epoch": 0, "step": 1899, "train_loss": 3.7245142459869385, "perplexity": 41.45109281901925, "lr": 0.001305, "grad_norm": 0.275455, "tokens_per_sec": 106561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:38.942981+00:00", "epoch": 0, "step": 1900, "train_loss": 3.797734260559082, "perplexity": 44.600017908378234, "lr": 0.001305, "grad_norm": 0.28654, "tokens_per_sec": 106438} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:11:44.805597+00:00", "step": 1900, "epoch": 0, "val_loss": 3.820321524143219, "val_ppl": 45.61887353233972, "eval_train_loss": 3.797734260559082, "eval_train_ppl": 44.600017908378234} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:11:45.790707+00:00", "step": 1900, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p8203_epoch_0000_step_0001900.pt", "val_loss": 3.820321524143219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:46.808936+00:00", "epoch": 0, "step": 1901, "train_loss": 3.855802536010742, "perplexity": 47.2665348203042, "lr": 0.001305, "grad_norm": 0.254215, "tokens_per_sec": 4166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:47.113110+00:00", "epoch": 0, "step": 1902, "train_loss": 3.786438226699829, "perplexity": 44.09904940153944, "lr": 0.001305, "grad_norm": 0.264689, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:47.418953+00:00", "epoch": 0, "step": 1903, "train_loss": 3.6312105655670166, "perplexity": 37.75849809904949, "lr": 0.001305, "grad_norm": 0.298897, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:47.725466+00:00", "epoch": 0, "step": 1904, "train_loss": 3.7942469120025635, "perplexity": 44.44475298910659, "lr": 0.001305, "grad_norm": 0.285624, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:48.032295+00:00", "epoch": 0, "step": 1905, "train_loss": 3.8795385360717773, "perplexity": 48.4018741979873, "lr": 0.001305, "grad_norm": 0.275051, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:48.338192+00:00", "epoch": 0, "step": 1906, "train_loss": 3.859269857406616, "perplexity": 47.43070754311466, "lr": 0.001305, "grad_norm": 0.237854, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:48.643419+00:00", "epoch": 0, "step": 1907, "train_loss": 3.7701199054718018, "perplexity": 43.3852666548496, "lr": 0.001305, "grad_norm": 0.250301, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:48.949659+00:00", "epoch": 0, "step": 1908, "train_loss": 3.769172430038452, "perplexity": 43.34417964806327, "lr": 0.001305, "grad_norm": 0.254852, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:49.257280+00:00", "epoch": 0, "step": 1909, "train_loss": 3.764594316482544, "perplexity": 43.146198607282365, "lr": 0.001305, "grad_norm": 0.253255, "tokens_per_sec": 106521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:49.564467+00:00", "epoch": 0, "step": 1910, "train_loss": 3.76428484916687, "perplexity": 43.13284833486087, "lr": 0.001305, "grad_norm": 0.259456, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:49.871203+00:00", "epoch": 0, "step": 1911, "train_loss": 3.7927820682525635, "perplexity": 44.379696031242354, "lr": 0.001305, "grad_norm": 0.259308, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:50.176727+00:00", "epoch": 0, "step": 1912, "train_loss": 3.6573867797851562, "perplexity": 38.7599221995561, "lr": 0.001305, "grad_norm": 0.252524, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:50.482464+00:00", "epoch": 0, "step": 1913, "train_loss": 3.759253740310669, "perplexity": 42.91638725582544, "lr": 0.001305, "grad_norm": 0.256041, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:50.789982+00:00", "epoch": 0, "step": 1914, "train_loss": 3.8327860832214355, "perplexity": 46.19105123996081, "lr": 0.001305, "grad_norm": 0.243188, "tokens_per_sec": 106558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:51.133899+00:00", "epoch": 0, "step": 1915, "train_loss": 3.8229589462280273, "perplexity": 45.739348558790404, "lr": 0.001305, "grad_norm": 0.272755, "tokens_per_sec": 95278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:51.441528+00:00", "epoch": 0, "step": 1916, "train_loss": 3.7920234203338623, "perplexity": 44.346040235279645, "lr": 0.001305, "grad_norm": 0.306446, "tokens_per_sec": 106520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:51.749049+00:00", "epoch": 0, "step": 1917, "train_loss": 3.817286968231201, "perplexity": 45.480650339036124, "lr": 0.001305, "grad_norm": 0.298512, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:52.055773+00:00", "epoch": 0, "step": 1918, "train_loss": 3.773174285888672, "perplexity": 43.5179843456792, "lr": 0.001305, "grad_norm": 0.29152, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:52.362076+00:00", "epoch": 0, "step": 1919, "train_loss": 3.741852283477783, "perplexity": 42.176039839373146, "lr": 0.001305, "grad_norm": 0.289601, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:52.668592+00:00", "epoch": 0, "step": 1920, "train_loss": 3.875833511352539, "perplexity": 48.222875859023816, "lr": 0.001305, "grad_norm": 0.29015, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:52.974769+00:00", "epoch": 0, "step": 1921, "train_loss": 3.703463315963745, "perplexity": 40.587629008985445, "lr": 0.001305, "grad_norm": 0.227282, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:53.282021+00:00", "epoch": 0, "step": 1922, "train_loss": 3.7943766117095947, "perplexity": 44.45051783438964, "lr": 0.001305, "grad_norm": 0.255288, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:53.589647+00:00", "epoch": 0, "step": 1923, "train_loss": 3.802773952484131, "perplexity": 44.825355597818756, "lr": 0.001305, "grad_norm": 0.245595, "tokens_per_sec": 106519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:53.896586+00:00", "epoch": 0, "step": 1924, "train_loss": 3.663728713989258, "perplexity": 39.00651619083271, "lr": 0.001305, "grad_norm": 0.26599, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:54.202863+00:00", "epoch": 0, "step": 1925, "train_loss": 3.855375289916992, "perplexity": 47.24634469132113, "lr": 0.001305, "grad_norm": 0.275243, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:54.508860+00:00", "epoch": 0, "step": 1926, "train_loss": 3.870464324951172, "perplexity": 47.96465209582841, "lr": 0.001305, "grad_norm": 0.239822, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:54.816635+00:00", "epoch": 0, "step": 1927, "train_loss": 3.8122048377990723, "perplexity": 45.250098086410766, "lr": 0.001305, "grad_norm": 0.219613, "tokens_per_sec": 106473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:55.124026+00:00", "epoch": 0, "step": 1928, "train_loss": 3.7277469635009766, "perplexity": 41.58530931787922, "lr": 0.001305, "grad_norm": 0.244943, "tokens_per_sec": 106594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:55.431022+00:00", "epoch": 0, "step": 1929, "train_loss": 3.8477871417999268, "perplexity": 46.88918921787923, "lr": 0.001305, "grad_norm": 0.234395, "tokens_per_sec": 106737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:55.737934+00:00", "epoch": 0, "step": 1930, "train_loss": 3.8156111240386963, "perplexity": 45.4044956847847, "lr": 0.001305, "grad_norm": 0.217755, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:56.044846+00:00", "epoch": 0, "step": 1931, "train_loss": 3.727630138397217, "perplexity": 41.58045139357256, "lr": 0.001305, "grad_norm": 0.252579, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:56.351906+00:00", "epoch": 0, "step": 1932, "train_loss": 3.763427972793579, "perplexity": 43.09590464655726, "lr": 0.001305, "grad_norm": 0.27998, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:56.659048+00:00", "epoch": 0, "step": 1933, "train_loss": 3.8160221576690674, "perplexity": 45.42316229552077, "lr": 0.001305, "grad_norm": 0.292931, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:56.966697+00:00", "epoch": 0, "step": 1934, "train_loss": 3.8156280517578125, "perplexity": 45.40526428583958, "lr": 0.001305, "grad_norm": 0.252414, "tokens_per_sec": 106511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:57.274476+00:00", "epoch": 0, "step": 1935, "train_loss": 3.80500864982605, "perplexity": 44.925638710292795, "lr": 0.001305, "grad_norm": 0.253737, "tokens_per_sec": 106467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:57.581608+00:00", "epoch": 0, "step": 1936, "train_loss": 3.7178916931152344, "perplexity": 41.177487747247184, "lr": 0.001305, "grad_norm": 0.254344, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:57.888750+00:00", "epoch": 0, "step": 1937, "train_loss": 3.7611660957336426, "perplexity": 42.998537166606354, "lr": 0.001305, "grad_norm": 0.24863, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:58.208160+00:00", "epoch": 0, "step": 1938, "train_loss": 3.69592022895813, "perplexity": 40.2826247749366, "lr": 0.001305, "grad_norm": 0.259541, "tokens_per_sec": 102588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:11:58.516089+00:00", "epoch": 0, "step": 1939, "train_loss": 3.71496319770813, "perplexity": 41.0570760620918, "lr": 0.001305, "grad_norm": 0.284347, "tokens_per_sec": 106414} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:12:04.429219+00:00", "step": 1939, "epoch": 1, "val_loss": 3.817453074455261, "val_ppl": 45.488205585601094, "eval_train_loss": 3.71496319770813, "eval_train_ppl": 41.0570760620918} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:12:05.402146+00:00", "step": 1939, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p8175_epoch_0001_step_0001939.pt", "val_loss": 3.817453074455261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:06.445183+00:00", "epoch": 1, "step": 1940, "train_loss": 3.7520132064819336, "perplexity": 42.606771944798446, "lr": 0.001305, "grad_norm": 0.268619, "tokens_per_sec": 4133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:06.750883+00:00", "epoch": 1, "step": 1941, "train_loss": 3.6951756477355957, "perplexity": 40.25264225253158, "lr": 0.001305, "grad_norm": 0.241488, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:07.056147+00:00", "epoch": 1, "step": 1942, "train_loss": 3.7105867862701416, "perplexity": 40.877786014472285, "lr": 0.001305, "grad_norm": 0.261169, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:07.363112+00:00", "epoch": 1, "step": 1943, "train_loss": 3.7559475898742676, "perplexity": 42.77473351668693, "lr": 0.001305, "grad_norm": 0.278188, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:07.668584+00:00", "epoch": 1, "step": 1944, "train_loss": 3.6834840774536133, "perplexity": 39.78476608971805, "lr": 0.001305, "grad_norm": 0.234455, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:07.974839+00:00", "epoch": 1, "step": 1945, "train_loss": 3.8026373386383057, "perplexity": 44.8192322518765, "lr": 0.001305, "grad_norm": 0.274229, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:08.280983+00:00", "epoch": 1, "step": 1946, "train_loss": 3.71553635597229, "perplexity": 41.080615009665806, "lr": 0.001305, "grad_norm": 0.305757, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:08.587472+00:00", "epoch": 1, "step": 1947, "train_loss": 3.787060260772705, "perplexity": 44.12648904616083, "lr": 0.001305, "grad_norm": 0.303448, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:08.894290+00:00", "epoch": 1, "step": 1948, "train_loss": 3.7988216876983643, "perplexity": 44.648543557536414, "lr": 0.001305, "grad_norm": 0.311829, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:09.201015+00:00", "epoch": 1, "step": 1949, "train_loss": 3.7113728523254395, "perplexity": 40.90993128697256, "lr": 0.001305, "grad_norm": 0.315251, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:09.506814+00:00", "epoch": 1, "step": 1950, "train_loss": 3.645080327987671, "perplexity": 38.28584815317835, "lr": 0.001305, "grad_norm": 0.325672, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:09.813492+00:00", "epoch": 1, "step": 1951, "train_loss": 3.9379055500030518, "perplexity": 51.31102031201778, "lr": 0.001305, "grad_norm": 0.292195, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:10.120081+00:00", "epoch": 1, "step": 1952, "train_loss": 3.738614797592163, "perplexity": 42.039716297523945, "lr": 0.001305, "grad_norm": 0.244028, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:10.427129+00:00", "epoch": 1, "step": 1953, "train_loss": 3.7939789295196533, "perplexity": 44.43284416959702, "lr": 0.001305, "grad_norm": 0.270255, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:10.735532+00:00", "epoch": 1, "step": 1954, "train_loss": 3.7170801162719727, "perplexity": 41.1440826089797, "lr": 0.001305, "grad_norm": 0.265442, "tokens_per_sec": 106252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:11.042067+00:00", "epoch": 1, "step": 1955, "train_loss": 3.762704849243164, "perplexity": 43.064752247854955, "lr": 0.001305, "grad_norm": 0.269316, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:11.348420+00:00", "epoch": 1, "step": 1956, "train_loss": 3.815295696258545, "perplexity": 45.39017610401747, "lr": 0.001305, "grad_norm": 0.271945, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:11.655404+00:00", "epoch": 1, "step": 1957, "train_loss": 3.784497022628784, "perplexity": 44.01352718220805, "lr": 0.001305, "grad_norm": 0.275441, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:11.961883+00:00", "epoch": 1, "step": 1958, "train_loss": 3.8073227405548096, "perplexity": 45.029721095910794, "lr": 0.001305, "grad_norm": 0.2673, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:12.269485+00:00", "epoch": 1, "step": 1959, "train_loss": 3.781811475753784, "perplexity": 43.89548536618593, "lr": 0.001305, "grad_norm": 0.227097, "tokens_per_sec": 106527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:12.577239+00:00", "epoch": 1, "step": 1960, "train_loss": 3.835116147994995, "perplexity": 46.298804869025034, "lr": 0.001305, "grad_norm": 0.238182, "tokens_per_sec": 106537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:12.884339+00:00", "epoch": 1, "step": 1961, "train_loss": 3.7122323513031006, "perplexity": 40.94510844629057, "lr": 0.001305, "grad_norm": 0.232198, "tokens_per_sec": 106641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:13.191167+00:00", "epoch": 1, "step": 1962, "train_loss": 3.8573389053344727, "perplexity": 47.339209487710676, "lr": 0.001305, "grad_norm": 0.262281, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:13.510094+00:00", "epoch": 1, "step": 1963, "train_loss": 3.801100254058838, "perplexity": 44.75039421961072, "lr": 0.001305, "grad_norm": 0.259299, "tokens_per_sec": 102743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:13.816977+00:00", "epoch": 1, "step": 1964, "train_loss": 3.7416255474090576, "perplexity": 42.1664780939428, "lr": 0.001305, "grad_norm": 0.267916, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:14.125436+00:00", "epoch": 1, "step": 1965, "train_loss": 3.8055107593536377, "perplexity": 44.948201965660914, "lr": 0.001305, "grad_norm": 0.266552, "tokens_per_sec": 106232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:14.445890+00:00", "epoch": 1, "step": 1966, "train_loss": 3.7770497798919678, "perplexity": 43.68696526381382, "lr": 0.001305, "grad_norm": 0.261657, "tokens_per_sec": 102255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:14.753255+00:00", "epoch": 1, "step": 1967, "train_loss": 3.71480393409729, "perplexity": 41.05053768458425, "lr": 0.001305, "grad_norm": 0.266652, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:15.059555+00:00", "epoch": 1, "step": 1968, "train_loss": 3.772963762283325, "perplexity": 43.50882374701229, "lr": 0.001305, "grad_norm": 0.250537, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:15.367104+00:00", "epoch": 1, "step": 1969, "train_loss": 3.615159273147583, "perplexity": 37.15726360870054, "lr": 0.001305, "grad_norm": 0.270743, "tokens_per_sec": 106546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:15.685601+00:00", "epoch": 1, "step": 1970, "train_loss": 3.650726318359375, "perplexity": 38.5026210562429, "lr": 0.001305, "grad_norm": 0.281323, "tokens_per_sec": 102938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:16.004388+00:00", "epoch": 1, "step": 1971, "train_loss": 3.707576036453247, "perplexity": 40.75489831260457, "lr": 0.001305, "grad_norm": 0.273024, "tokens_per_sec": 102736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:16.312269+00:00", "epoch": 1, "step": 1972, "train_loss": 3.7577221393585205, "perplexity": 42.85070678721522, "lr": 0.001305, "grad_norm": 0.264266, "tokens_per_sec": 106431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:16.632833+00:00", "epoch": 1, "step": 1973, "train_loss": 3.7850823402404785, "perplexity": 44.03929661573652, "lr": 0.001305, "grad_norm": 0.266809, "tokens_per_sec": 102220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:16.939106+00:00", "epoch": 1, "step": 1974, "train_loss": 3.726041078567505, "perplexity": 41.51443003839606, "lr": 0.001305, "grad_norm": 0.261803, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:17.257054+00:00", "epoch": 1, "step": 1975, "train_loss": 3.667558193206787, "perplexity": 39.15617721292353, "lr": 0.001305, "grad_norm": 0.256756, "tokens_per_sec": 103062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:17.563727+00:00", "epoch": 1, "step": 1976, "train_loss": 3.8259353637695312, "perplexity": 45.87569076318684, "lr": 0.001305, "grad_norm": 0.268211, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:17.881622+00:00", "epoch": 1, "step": 1977, "train_loss": 3.802678108215332, "perplexity": 44.82105955026689, "lr": 0.001305, "grad_norm": 0.218577, "tokens_per_sec": 103079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:18.199127+00:00", "epoch": 1, "step": 1978, "train_loss": 3.781323194503784, "perplexity": 43.874057255620016, "lr": 0.001305, "grad_norm": 0.255433, "tokens_per_sec": 103204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:18.519304+00:00", "epoch": 1, "step": 1979, "train_loss": 3.790426254272461, "perplexity": 44.2752687767886, "lr": 0.001305, "grad_norm": 0.241247, "tokens_per_sec": 102344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:18.842503+00:00", "epoch": 1, "step": 1980, "train_loss": 3.7333500385284424, "perplexity": 41.81896892087, "lr": 0.001305, "grad_norm": 0.241, "tokens_per_sec": 101440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:19.165822+00:00", "epoch": 1, "step": 1981, "train_loss": 3.6283812522888184, "perplexity": 37.65181846522958, "lr": 0.001305, "grad_norm": 0.26246, "tokens_per_sec": 101296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:19.472598+00:00", "epoch": 1, "step": 1982, "train_loss": 3.648836851119995, "perplexity": 38.42994030070318, "lr": 0.001305, "grad_norm": 0.2782, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:19.790513+00:00", "epoch": 1, "step": 1983, "train_loss": 3.740732192993164, "perplexity": 42.1288253056814, "lr": 0.001305, "grad_norm": 0.277005, "tokens_per_sec": 103071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:20.108700+00:00", "epoch": 1, "step": 1984, "train_loss": 3.6964519023895264, "perplexity": 40.30404767076397, "lr": 0.001305, "grad_norm": 0.272722, "tokens_per_sec": 102985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:20.428127+00:00", "epoch": 1, "step": 1985, "train_loss": 3.748215675354004, "perplexity": 42.445278234851, "lr": 0.001305, "grad_norm": 0.248439, "tokens_per_sec": 102584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:20.745794+00:00", "epoch": 1, "step": 1986, "train_loss": 3.7553508281707764, "perplexity": 42.749214808898216, "lr": 0.001305, "grad_norm": 0.25947, "tokens_per_sec": 103152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:21.064930+00:00", "epoch": 1, "step": 1987, "train_loss": 3.7457430362701416, "perplexity": 42.34045602811114, "lr": 0.001305, "grad_norm": 0.268722, "tokens_per_sec": 102677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:21.388141+00:00", "epoch": 1, "step": 1988, "train_loss": 3.7628612518310547, "perplexity": 43.071488213300874, "lr": 0.001305, "grad_norm": 0.287766, "tokens_per_sec": 101382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:21.707033+00:00", "epoch": 1, "step": 1989, "train_loss": 3.666372537612915, "perplexity": 39.1097789839717, "lr": 0.001305, "grad_norm": 0.274378, "tokens_per_sec": 102757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:22.029858+00:00", "epoch": 1, "step": 1990, "train_loss": 3.800607919692993, "perplexity": 44.728367485355356, "lr": 0.001305, "grad_norm": 0.270231, "tokens_per_sec": 101559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:22.337358+00:00", "epoch": 1, "step": 1991, "train_loss": 3.8160321712493896, "perplexity": 45.42361714628224, "lr": 0.001305, "grad_norm": 0.279214, "tokens_per_sec": 106501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:22.655830+00:00", "epoch": 1, "step": 1992, "train_loss": 3.80523681640625, "perplexity": 44.935890409143695, "lr": 0.001305, "grad_norm": 0.266628, "tokens_per_sec": 102892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:22.974266+00:00", "epoch": 1, "step": 1993, "train_loss": 3.8401801586151123, "perplexity": 46.533857159918675, "lr": 0.001305, "grad_norm": 0.307627, "tokens_per_sec": 102903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:23.292365+00:00", "epoch": 1, "step": 1994, "train_loss": 3.814326047897339, "perplexity": 45.34618492557717, "lr": 0.001305, "grad_norm": 0.247218, "tokens_per_sec": 103012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:23.610100+00:00", "epoch": 1, "step": 1995, "train_loss": 3.7296371459960938, "perplexity": 41.663987476219695, "lr": 0.001305, "grad_norm": 0.258638, "tokens_per_sec": 103130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:23.928592+00:00", "epoch": 1, "step": 1996, "train_loss": 3.7769978046417236, "perplexity": 43.68469468186937, "lr": 0.001305, "grad_norm": 0.257546, "tokens_per_sec": 102886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:24.247968+00:00", "epoch": 1, "step": 1997, "train_loss": 3.721832752227783, "perplexity": 41.34009086435167, "lr": 0.001305, "grad_norm": 0.243904, "tokens_per_sec": 102600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:24.568861+00:00", "epoch": 1, "step": 1998, "train_loss": 3.705291271209717, "perplexity": 40.66188922985624, "lr": 0.001305, "grad_norm": 0.239493, "tokens_per_sec": 102116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:24.887925+00:00", "epoch": 1, "step": 1999, "train_loss": 3.737880229949951, "perplexity": 42.008846621564636, "lr": 0.001305, "grad_norm": 0.250604, "tokens_per_sec": 102700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:25.206722+00:00", "epoch": 1, "step": 2000, "train_loss": 3.689115285873413, "perplexity": 40.009434382798894, "lr": 0.001305, "grad_norm": 0.268571, "tokens_per_sec": 102849} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:12:31.096430+00:00", "step": 2000, "epoch": 1, "val_loss": 3.8102170705795286, "val_ppl": 45.16024076206251, "eval_train_loss": 3.689115285873413, "eval_train_ppl": 40.009434382798894} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:12:32.141437+00:00", "step": 2000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p8102_epoch_0001_step_0002000.pt", "val_loss": 3.8102170705795286} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:12:33.867419+00:00", "step": 2000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0001_step_0002000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:34.889918+00:00", "epoch": 1, "step": 2001, "train_loss": 3.766174554824829, "perplexity": 43.21443378436753, "lr": 0.001305, "grad_norm": 0.275111, "tokens_per_sec": 3384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:35.195151+00:00", "epoch": 1, "step": 2002, "train_loss": 3.7868897914886475, "perplexity": 44.11896747628168, "lr": 0.001305, "grad_norm": 0.273971, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:35.501126+00:00", "epoch": 1, "step": 2003, "train_loss": 3.785768747329712, "perplexity": 44.069535878173184, "lr": 0.001305, "grad_norm": 0.311002, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:35.806808+00:00", "epoch": 1, "step": 2004, "train_loss": 3.7737362384796143, "perplexity": 43.542446262309824, "lr": 0.001305, "grad_norm": 0.309472, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:36.113684+00:00", "epoch": 1, "step": 2005, "train_loss": 3.719578266143799, "perplexity": 41.246995185674876, "lr": 0.001305, "grad_norm": 0.299765, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:36.419717+00:00", "epoch": 1, "step": 2006, "train_loss": 3.690187692642212, "perplexity": 40.052363785823744, "lr": 0.001305, "grad_norm": 0.290133, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:36.728148+00:00", "epoch": 1, "step": 2007, "train_loss": 3.7305986881256104, "perplexity": 41.70406842213266, "lr": 0.001305, "grad_norm": 0.294181, "tokens_per_sec": 106241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:37.034329+00:00", "epoch": 1, "step": 2008, "train_loss": 3.7320609092712402, "perplexity": 41.76509359811656, "lr": 0.001305, "grad_norm": 0.308982, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:37.340657+00:00", "epoch": 1, "step": 2009, "train_loss": 3.861403226852417, "perplexity": 47.53200277703993, "lr": 0.001305, "grad_norm": 0.322246, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:37.646912+00:00", "epoch": 1, "step": 2010, "train_loss": 3.8101890087127686, "perplexity": 45.158973499184356, "lr": 0.001305, "grad_norm": 0.27318, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:37.954208+00:00", "epoch": 1, "step": 2011, "train_loss": 3.767575979232788, "perplexity": 43.275038002845726, "lr": 0.001305, "grad_norm": 0.263583, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:38.260990+00:00", "epoch": 1, "step": 2012, "train_loss": 3.694000720977783, "perplexity": 40.20537611863654, "lr": 0.001305, "grad_norm": 0.232289, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:38.567724+00:00", "epoch": 1, "step": 2013, "train_loss": 3.6431567668914795, "perplexity": 38.21227377023253, "lr": 0.001305, "grad_norm": 0.252285, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:38.874063+00:00", "epoch": 1, "step": 2014, "train_loss": 3.7657597064971924, "perplexity": 43.19651006685149, "lr": 0.001305, "grad_norm": 0.244766, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:39.294868+00:00", "epoch": 1, "step": 2015, "train_loss": 3.730278253555298, "perplexity": 41.690707137710575, "lr": 0.001305, "grad_norm": 0.237015, "tokens_per_sec": 77870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:39.600345+00:00", "epoch": 1, "step": 2016, "train_loss": 3.7782249450683594, "perplexity": 43.73833484201196, "lr": 0.001305, "grad_norm": 0.225241, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:39.905849+00:00", "epoch": 1, "step": 2017, "train_loss": 3.773439884185791, "perplexity": 43.52954418328411, "lr": 0.001305, "grad_norm": 0.224422, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:40.211729+00:00", "epoch": 1, "step": 2018, "train_loss": 3.66789174079895, "perplexity": 39.169239839933994, "lr": 0.001305, "grad_norm": 0.245273, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:40.517692+00:00", "epoch": 1, "step": 2019, "train_loss": 3.8372268676757812, "perplexity": 46.39663187403196, "lr": 0.001305, "grad_norm": 0.25416, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:40.824533+00:00", "epoch": 1, "step": 2020, "train_loss": 3.755014181137085, "perplexity": 42.7348258346788, "lr": 0.001305, "grad_norm": 0.238404, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:41.131139+00:00", "epoch": 1, "step": 2021, "train_loss": 3.777853012084961, "perplexity": 43.72207013752281, "lr": 0.001305, "grad_norm": 0.229637, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:41.437644+00:00", "epoch": 1, "step": 2022, "train_loss": 3.8097500801086426, "perplexity": 45.139156283472175, "lr": 0.001305, "grad_norm": 0.211766, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:41.744113+00:00", "epoch": 1, "step": 2023, "train_loss": 3.6541759967803955, "perplexity": 38.63567207692547, "lr": 0.001305, "grad_norm": 0.240592, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:42.049874+00:00", "epoch": 1, "step": 2024, "train_loss": 3.783524751663208, "perplexity": 43.970754904127666, "lr": 0.001305, "grad_norm": 0.278083, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:42.356633+00:00", "epoch": 1, "step": 2025, "train_loss": 3.7974047660827637, "perplexity": 44.585324869604214, "lr": 0.001305, "grad_norm": 0.318213, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:42.663787+00:00", "epoch": 1, "step": 2026, "train_loss": 3.739889144897461, "perplexity": 42.093323646638936, "lr": 0.001305, "grad_norm": 0.361183, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:42.970563+00:00", "epoch": 1, "step": 2027, "train_loss": 3.836426019668579, "perplexity": 46.35949009829839, "lr": 0.001305, "grad_norm": 0.347297, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:43.276963+00:00", "epoch": 1, "step": 2028, "train_loss": 3.7384660243988037, "perplexity": 42.03346237990162, "lr": 0.001305, "grad_norm": 0.301747, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:43.583294+00:00", "epoch": 1, "step": 2029, "train_loss": 3.821293830871582, "perplexity": 45.66325064059993, "lr": 0.001305, "grad_norm": 0.301536, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:43.889582+00:00", "epoch": 1, "step": 2030, "train_loss": 3.8270065784454346, "perplexity": 45.92485980699844, "lr": 0.001305, "grad_norm": 0.298658, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:44.196689+00:00", "epoch": 1, "step": 2031, "train_loss": 3.750810384750366, "perplexity": 42.55555440255278, "lr": 0.001305, "grad_norm": 0.311848, "tokens_per_sec": 106643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:44.503781+00:00", "epoch": 1, "step": 2032, "train_loss": 3.682055711746216, "perplexity": 39.72797945985084, "lr": 0.001305, "grad_norm": 0.324192, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:44.810275+00:00", "epoch": 1, "step": 2033, "train_loss": 3.745119094848633, "perplexity": 42.31404630370691, "lr": 0.001305, "grad_norm": 0.31064, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:45.116222+00:00", "epoch": 1, "step": 2034, "train_loss": 3.7572734355926514, "perplexity": 42.83148382673954, "lr": 0.001305, "grad_norm": 0.29636, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:45.423040+00:00", "epoch": 1, "step": 2035, "train_loss": 3.7455666065216064, "perplexity": 42.332986571037836, "lr": 0.001305, "grad_norm": 0.268258, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:45.729943+00:00", "epoch": 1, "step": 2036, "train_loss": 3.6395227909088135, "perplexity": 38.07366329122255, "lr": 0.001305, "grad_norm": 0.260889, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:46.037760+00:00", "epoch": 1, "step": 2037, "train_loss": 3.8184406757354736, "perplexity": 45.53315198658035, "lr": 0.001305, "grad_norm": 0.254056, "tokens_per_sec": 106453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:46.344603+00:00", "epoch": 1, "step": 2038, "train_loss": 3.7920138835906982, "perplexity": 44.345617320500196, "lr": 0.001305, "grad_norm": 0.266563, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:46.651119+00:00", "epoch": 1, "step": 2039, "train_loss": 3.7450740337371826, "perplexity": 42.31213962870928, "lr": 0.001305, "grad_norm": 0.266221, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:46.957401+00:00", "epoch": 1, "step": 2040, "train_loss": 3.7666420936584473, "perplexity": 43.234642934247596, "lr": 0.001305, "grad_norm": 0.233308, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:47.264086+00:00", "epoch": 1, "step": 2041, "train_loss": 3.7654879093170166, "perplexity": 43.184770972620505, "lr": 0.001305, "grad_norm": 0.219305, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:47.570671+00:00", "epoch": 1, "step": 2042, "train_loss": 3.6340103149414062, "perplexity": 37.86436055550222, "lr": 0.001305, "grad_norm": 0.21878, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:47.877331+00:00", "epoch": 1, "step": 2043, "train_loss": 3.8257312774658203, "perplexity": 45.866329118353036, "lr": 0.001305, "grad_norm": 0.223142, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:48.183531+00:00", "epoch": 1, "step": 2044, "train_loss": 3.789006233215332, "perplexity": 44.21244158134314, "lr": 0.001305, "grad_norm": 0.242448, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:48.490159+00:00", "epoch": 1, "step": 2045, "train_loss": 3.7562215328216553, "perplexity": 42.78645295841613, "lr": 0.001305, "grad_norm": 0.24567, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:48.796658+00:00", "epoch": 1, "step": 2046, "train_loss": 3.7402565479278564, "perplexity": 42.10879170263746, "lr": 0.001305, "grad_norm": 0.224569, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:49.103462+00:00", "epoch": 1, "step": 2047, "train_loss": 3.686020851135254, "perplexity": 39.88581915745445, "lr": 0.001305, "grad_norm": 0.236252, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:49.410319+00:00", "epoch": 1, "step": 2048, "train_loss": 3.754876136779785, "perplexity": 42.72892694027604, "lr": 0.001305, "grad_norm": 0.226293, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:49.717308+00:00", "epoch": 1, "step": 2049, "train_loss": 3.7984912395477295, "perplexity": 44.63379196634154, "lr": 0.001305, "grad_norm": 0.229281, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:50.023803+00:00", "epoch": 1, "step": 2050, "train_loss": 3.6323397159576416, "perplexity": 37.80115720166617, "lr": 0.001305, "grad_norm": 0.259885, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:50.329567+00:00", "epoch": 1, "step": 2051, "train_loss": 3.7950212955474854, "perplexity": 44.4791836040097, "lr": 0.001305, "grad_norm": 0.243959, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:50.636414+00:00", "epoch": 1, "step": 2052, "train_loss": 3.7659060955047607, "perplexity": 43.202834023958204, "lr": 0.001305, "grad_norm": 0.28964, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:50.943821+00:00", "epoch": 1, "step": 2053, "train_loss": 3.7367799282073975, "perplexity": 41.96264963439775, "lr": 0.001305, "grad_norm": 0.317159, "tokens_per_sec": 106596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:51.251257+00:00", "epoch": 1, "step": 2054, "train_loss": 3.7929773330688477, "perplexity": 44.38836267055197, "lr": 0.001305, "grad_norm": 0.34155, "tokens_per_sec": 106585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:51.558342+00:00", "epoch": 1, "step": 2055, "train_loss": 3.757290840148926, "perplexity": 42.83222929619739, "lr": 0.001305, "grad_norm": 0.284347, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:51.864079+00:00", "epoch": 1, "step": 2056, "train_loss": 3.7304911613464355, "perplexity": 41.69958435905949, "lr": 0.001305, "grad_norm": 0.286517, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:52.171056+00:00", "epoch": 1, "step": 2057, "train_loss": 3.7718653678894043, "perplexity": 43.461060135366225, "lr": 0.001305, "grad_norm": 0.278181, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:52.477091+00:00", "epoch": 1, "step": 2058, "train_loss": 3.7817904949188232, "perplexity": 43.89456441191316, "lr": 0.001305, "grad_norm": 0.265458, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:52.784085+00:00", "epoch": 1, "step": 2059, "train_loss": 3.787038564682007, "perplexity": 44.12553168423783, "lr": 0.001305, "grad_norm": 0.238793, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:53.090818+00:00", "epoch": 1, "step": 2060, "train_loss": 3.6997110843658447, "perplexity": 40.43562018942939, "lr": 0.001305, "grad_norm": 0.238397, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:53.396969+00:00", "epoch": 1, "step": 2061, "train_loss": 3.7038676738739014, "perplexity": 40.60404425642363, "lr": 0.001305, "grad_norm": 0.261521, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:53.703490+00:00", "epoch": 1, "step": 2062, "train_loss": 3.793795347213745, "perplexity": 44.42468783430863, "lr": 0.001305, "grad_norm": 0.286272, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:54.010094+00:00", "epoch": 1, "step": 2063, "train_loss": 3.6183600425720215, "perplexity": 37.276385981880026, "lr": 0.001305, "grad_norm": 0.264929, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:54.316959+00:00", "epoch": 1, "step": 2064, "train_loss": 3.790018320083618, "perplexity": 44.257211064363894, "lr": 0.001305, "grad_norm": 0.263341, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:54.623560+00:00", "epoch": 1, "step": 2065, "train_loss": 3.673731803894043, "perplexity": 39.39865993384167, "lr": 0.001305, "grad_norm": 0.223601, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:54.931161+00:00", "epoch": 1, "step": 2066, "train_loss": 3.6963891983032227, "perplexity": 40.30152052151256, "lr": 0.001305, "grad_norm": 0.248332, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:55.239617+00:00", "epoch": 1, "step": 2067, "train_loss": 3.8687801361083984, "perplexity": 47.88393855142519, "lr": 0.001305, "grad_norm": 0.271417, "tokens_per_sec": 106232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:55.546321+00:00", "epoch": 1, "step": 2068, "train_loss": 3.7020509243011475, "perplexity": 40.53034384423603, "lr": 0.001305, "grad_norm": 0.252355, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:55.853125+00:00", "epoch": 1, "step": 2069, "train_loss": 3.7051520347595215, "perplexity": 40.65622800687508, "lr": 0.001305, "grad_norm": 0.272484, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:56.159810+00:00", "epoch": 1, "step": 2070, "train_loss": 3.6984779834747314, "perplexity": 40.38578971945365, "lr": 0.001305, "grad_norm": 0.301142, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:56.467064+00:00", "epoch": 1, "step": 2071, "train_loss": 3.7766919136047363, "perplexity": 43.67133396887801, "lr": 0.001305, "grad_norm": 0.24422, "tokens_per_sec": 106591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:56.774195+00:00", "epoch": 1, "step": 2072, "train_loss": 3.8232948780059814, "perplexity": 45.754716440609634, "lr": 0.001305, "grad_norm": 0.250792, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:57.081549+00:00", "epoch": 1, "step": 2073, "train_loss": 3.7451977729797363, "perplexity": 42.317375624760174, "lr": 0.001305, "grad_norm": 0.272007, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:57.388649+00:00", "epoch": 1, "step": 2074, "train_loss": 3.6813488006591797, "perplexity": 39.699905234864694, "lr": 0.001305, "grad_norm": 0.269325, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:57.695021+00:00", "epoch": 1, "step": 2075, "train_loss": 3.764998435974121, "perplexity": 43.16363835075982, "lr": 0.001305, "grad_norm": 0.253211, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:58.001738+00:00", "epoch": 1, "step": 2076, "train_loss": 3.880506753921509, "perplexity": 48.44876045093698, "lr": 0.001305, "grad_norm": 0.271378, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:58.308889+00:00", "epoch": 1, "step": 2077, "train_loss": 3.6834816932678223, "perplexity": 39.78467123555711, "lr": 0.001305, "grad_norm": 0.278416, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:58.616878+00:00", "epoch": 1, "step": 2078, "train_loss": 3.7239327430725098, "perplexity": 41.426995894633635, "lr": 0.001305, "grad_norm": 0.244, "tokens_per_sec": 106393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:58.925312+00:00", "epoch": 1, "step": 2079, "train_loss": 3.7469398975372314, "perplexity": 42.39116201792679, "lr": 0.001305, "grad_norm": 0.251964, "tokens_per_sec": 106240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:59.232050+00:00", "epoch": 1, "step": 2080, "train_loss": 3.703200101852417, "perplexity": 40.5769471781509, "lr": 0.001305, "grad_norm": 0.243604, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:59.538363+00:00", "epoch": 1, "step": 2081, "train_loss": 3.785958766937256, "perplexity": 44.077910749755105, "lr": 0.001305, "grad_norm": 0.230803, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:12:59.846182+00:00", "epoch": 1, "step": 2082, "train_loss": 3.655290126800537, "perplexity": 38.678741226896086, "lr": 0.001305, "grad_norm": 0.245753, "tokens_per_sec": 106452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:00.155468+00:00", "epoch": 1, "step": 2083, "train_loss": 3.5881094932556152, "perplexity": 36.165639871355424, "lr": 0.001305, "grad_norm": 0.243889, "tokens_per_sec": 105947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:00.463668+00:00", "epoch": 1, "step": 2084, "train_loss": 3.6067850589752197, "perplexity": 36.847399968254116, "lr": 0.001305, "grad_norm": 0.259482, "tokens_per_sec": 106321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:00.770827+00:00", "epoch": 1, "step": 2085, "train_loss": 3.7855653762817383, "perplexity": 44.06057432176937, "lr": 0.001305, "grad_norm": 0.273977, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:01.078159+00:00", "epoch": 1, "step": 2086, "train_loss": 3.822711706161499, "perplexity": 45.72804135706402, "lr": 0.001305, "grad_norm": 0.291454, "tokens_per_sec": 106621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:01.385645+00:00", "epoch": 1, "step": 2087, "train_loss": 3.70865797996521, "perplexity": 40.799016672897686, "lr": 0.001305, "grad_norm": 0.299295, "tokens_per_sec": 106567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:01.692434+00:00", "epoch": 1, "step": 2088, "train_loss": 3.6916770935058594, "perplexity": 40.11206225747926, "lr": 0.001305, "grad_norm": 0.295836, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:02.001536+00:00", "epoch": 1, "step": 2089, "train_loss": 3.642658233642578, "perplexity": 38.193228429004314, "lr": 0.001305, "grad_norm": 0.281807, "tokens_per_sec": 106011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:02.308923+00:00", "epoch": 1, "step": 2090, "train_loss": 3.8128015995025635, "perplexity": 45.277109670965935, "lr": 0.001305, "grad_norm": 0.280576, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:02.615784+00:00", "epoch": 1, "step": 2091, "train_loss": 3.7090344429016113, "perplexity": 40.814378881986386, "lr": 0.001305, "grad_norm": 0.28829, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:02.922309+00:00", "epoch": 1, "step": 2092, "train_loss": 3.704423666000366, "perplexity": 40.62662606240427, "lr": 0.001305, "grad_norm": 0.278302, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:03.229549+00:00", "epoch": 1, "step": 2093, "train_loss": 3.7250308990478516, "perplexity": 41.472514186233354, "lr": 0.001305, "grad_norm": 0.260688, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:03.537100+00:00", "epoch": 1, "step": 2094, "train_loss": 3.655836582183838, "perplexity": 38.699883209307764, "lr": 0.001305, "grad_norm": 0.264868, "tokens_per_sec": 106545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:03.845336+00:00", "epoch": 1, "step": 2095, "train_loss": 3.8164756298065186, "perplexity": 45.44376510506369, "lr": 0.001305, "grad_norm": 0.278581, "tokens_per_sec": 106308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:04.153825+00:00", "epoch": 1, "step": 2096, "train_loss": 3.679856538772583, "perplexity": 39.64070676018045, "lr": 0.001305, "grad_norm": 0.280671, "tokens_per_sec": 106221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:04.461105+00:00", "epoch": 1, "step": 2097, "train_loss": 3.7029786109924316, "perplexity": 40.56796075046737, "lr": 0.001305, "grad_norm": 0.266569, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:04.768400+00:00", "epoch": 1, "step": 2098, "train_loss": 3.754000425338745, "perplexity": 42.69152510908702, "lr": 0.001305, "grad_norm": 0.274308, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:05.075463+00:00", "epoch": 1, "step": 2099, "train_loss": 3.83347225189209, "perplexity": 46.222756968681125, "lr": 0.001305, "grad_norm": 0.301664, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:05.383818+00:00", "epoch": 1, "step": 2100, "train_loss": 3.743954658508301, "perplexity": 42.26480296641425, "lr": 0.001305, "grad_norm": 0.272152, "tokens_per_sec": 106324} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:13:11.243436+00:00", "step": 2100, "epoch": 1, "val_loss": 3.7971917510032656, "val_ppl": 44.575828534549856, "eval_train_loss": 3.743954658508301, "eval_train_ppl": 42.26480296641425} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:13:12.122231+00:00", "step": 2100, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p7972_epoch_0001_step_0002100.pt", "val_loss": 3.7971917510032656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:13.103397+00:00", "epoch": 1, "step": 2101, "train_loss": 3.6706182956695557, "perplexity": 39.27618264806271, "lr": 0.001305, "grad_norm": 0.260214, "tokens_per_sec": 4245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:13.409430+00:00", "epoch": 1, "step": 2102, "train_loss": 3.6323609352111816, "perplexity": 37.80195932251512, "lr": 0.001305, "grad_norm": 0.275585, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:13.715890+00:00", "epoch": 1, "step": 2103, "train_loss": 3.6966371536254883, "perplexity": 40.31151473702952, "lr": 0.001305, "grad_norm": 0.282465, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:14.021792+00:00", "epoch": 1, "step": 2104, "train_loss": 3.6842434406280518, "perplexity": 39.81498864948741, "lr": 0.001305, "grad_norm": 0.26479, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:14.327549+00:00", "epoch": 1, "step": 2105, "train_loss": 3.7296664714813232, "perplexity": 41.665209310784384, "lr": 0.001305, "grad_norm": 0.24125, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:14.632671+00:00", "epoch": 1, "step": 2106, "train_loss": 3.6835501194000244, "perplexity": 39.78739363987143, "lr": 0.001305, "grad_norm": 0.251639, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:14.939427+00:00", "epoch": 1, "step": 2107, "train_loss": 3.741177797317505, "perplexity": 42.14760227565582, "lr": 0.001305, "grad_norm": 0.234094, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:15.246107+00:00", "epoch": 1, "step": 2108, "train_loss": 3.859626293182373, "perplexity": 47.44761655746229, "lr": 0.001305, "grad_norm": 0.263794, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:15.552177+00:00", "epoch": 1, "step": 2109, "train_loss": 3.806885242462158, "perplexity": 45.01002498763777, "lr": 0.001305, "grad_norm": 0.258101, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:15.858066+00:00", "epoch": 1, "step": 2110, "train_loss": 3.6330039501190186, "perplexity": 37.82627436253375, "lr": 0.001305, "grad_norm": 0.238796, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:16.163045+00:00", "epoch": 1, "step": 2111, "train_loss": 3.7442290782928467, "perplexity": 42.276402856084985, "lr": 0.001305, "grad_norm": 0.228282, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:16.468209+00:00", "epoch": 1, "step": 2112, "train_loss": 3.843224287033081, "perplexity": 46.67572802390483, "lr": 0.001305, "grad_norm": 0.246864, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:16.774698+00:00", "epoch": 1, "step": 2113, "train_loss": 3.731964588165283, "perplexity": 41.761070931847726, "lr": 0.001305, "grad_norm": 0.260619, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:17.080967+00:00", "epoch": 1, "step": 2114, "train_loss": 3.564229726791382, "perplexity": 35.31224285124055, "lr": 0.001305, "grad_norm": 0.254698, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:17.438168+00:00", "epoch": 1, "step": 2115, "train_loss": 3.6794955730438232, "perplexity": 39.626400405783365, "lr": 0.001305, "grad_norm": 0.213891, "tokens_per_sec": 91735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:17.744848+00:00", "epoch": 1, "step": 2116, "train_loss": 3.750812530517578, "perplexity": 42.55564571696407, "lr": 0.001305, "grad_norm": 0.21057, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:18.050722+00:00", "epoch": 1, "step": 2117, "train_loss": 3.7384610176086426, "perplexity": 42.03325192770259, "lr": 0.001305, "grad_norm": 0.218256, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:18.356285+00:00", "epoch": 1, "step": 2118, "train_loss": 3.6397852897644043, "perplexity": 38.083658896124675, "lr": 0.001305, "grad_norm": 0.237681, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:18.662462+00:00", "epoch": 1, "step": 2119, "train_loss": 3.6617939472198486, "perplexity": 38.93112063943912, "lr": 0.001305, "grad_norm": 0.240008, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:18.968898+00:00", "epoch": 1, "step": 2120, "train_loss": 3.7768406867980957, "perplexity": 43.67783157601372, "lr": 0.001305, "grad_norm": 0.224625, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:19.276008+00:00", "epoch": 1, "step": 2121, "train_loss": 3.723149299621582, "perplexity": 41.39455289629508, "lr": 0.001305, "grad_norm": 0.238671, "tokens_per_sec": 106642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:19.583055+00:00", "epoch": 1, "step": 2122, "train_loss": 3.79728102684021, "perplexity": 44.57980825659364, "lr": 0.001305, "grad_norm": 0.252537, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:19.889866+00:00", "epoch": 1, "step": 2123, "train_loss": 3.721794366836548, "perplexity": 41.33850403924579, "lr": 0.001305, "grad_norm": 0.276453, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:20.196610+00:00", "epoch": 1, "step": 2124, "train_loss": 3.706970453262329, "perplexity": 40.73022530277292, "lr": 0.001305, "grad_norm": 0.280457, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:20.502973+00:00", "epoch": 1, "step": 2125, "train_loss": 3.7729263305664062, "perplexity": 43.507195167518766, "lr": 0.001305, "grad_norm": 0.281062, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:20.809378+00:00", "epoch": 1, "step": 2126, "train_loss": 3.6740500926971436, "perplexity": 39.411202082062566, "lr": 0.001305, "grad_norm": 0.289899, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:21.115587+00:00", "epoch": 1, "step": 2127, "train_loss": 3.6870803833007812, "perplexity": 39.92810186178725, "lr": 0.001305, "grad_norm": 0.26241, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:21.421785+00:00", "epoch": 1, "step": 2128, "train_loss": 3.872706651687622, "perplexity": 48.07232519164291, "lr": 0.001305, "grad_norm": 0.247963, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:21.728248+00:00", "epoch": 1, "step": 2129, "train_loss": 3.741081476211548, "perplexity": 42.14354276750249, "lr": 0.001305, "grad_norm": 0.28645, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:22.034031+00:00", "epoch": 1, "step": 2130, "train_loss": 3.828184127807617, "perplexity": 45.97897044909596, "lr": 0.001305, "grad_norm": 0.279156, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:22.340201+00:00", "epoch": 1, "step": 2131, "train_loss": 3.776559591293335, "perplexity": 43.665555659333336, "lr": 0.001305, "grad_norm": 0.283632, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:22.647773+00:00", "epoch": 1, "step": 2132, "train_loss": 3.7520782947540283, "perplexity": 42.609545236217265, "lr": 0.001305, "grad_norm": 0.305399, "tokens_per_sec": 106537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:22.954837+00:00", "epoch": 1, "step": 2133, "train_loss": 3.6439731121063232, "perplexity": 38.243480913241, "lr": 0.001305, "grad_norm": 0.26199, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:23.261176+00:00", "epoch": 1, "step": 2134, "train_loss": 3.695331573486328, "perplexity": 40.25891916534722, "lr": 0.001305, "grad_norm": 0.280614, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:23.567173+00:00", "epoch": 1, "step": 2135, "train_loss": 3.845527410507202, "perplexity": 46.78335187672616, "lr": 0.001305, "grad_norm": 0.267673, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:23.872454+00:00", "epoch": 1, "step": 2136, "train_loss": 3.659863233566284, "perplexity": 38.8560283074938, "lr": 0.001305, "grad_norm": 0.259284, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:24.178933+00:00", "epoch": 1, "step": 2137, "train_loss": 3.6971936225891113, "perplexity": 40.333953086400804, "lr": 0.001305, "grad_norm": 0.263071, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:24.484484+00:00", "epoch": 1, "step": 2138, "train_loss": 3.705650806427002, "perplexity": 40.67651123944191, "lr": 0.001305, "grad_norm": 0.28013, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:24.791402+00:00", "epoch": 1, "step": 2139, "train_loss": 3.6918835639953613, "perplexity": 40.12034506965721, "lr": 0.001305, "grad_norm": 0.279982, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:25.097048+00:00", "epoch": 1, "step": 2140, "train_loss": 3.8309571743011475, "perplexity": 46.1066492196409, "lr": 0.001305, "grad_norm": 0.275331, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:25.402667+00:00", "epoch": 1, "step": 2141, "train_loss": 3.7402656078338623, "perplexity": 42.1091732060605, "lr": 0.001305, "grad_norm": 0.263548, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:25.708621+00:00", "epoch": 1, "step": 2142, "train_loss": 3.783473253250122, "perplexity": 43.96849053833404, "lr": 0.001305, "grad_norm": 0.219278, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:26.015644+00:00", "epoch": 1, "step": 2143, "train_loss": 3.7253572940826416, "perplexity": 41.48605281829495, "lr": 0.001305, "grad_norm": 0.241778, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:26.323985+00:00", "epoch": 1, "step": 2144, "train_loss": 3.826399326324463, "perplexity": 45.89698030427562, "lr": 0.001305, "grad_norm": 0.236638, "tokens_per_sec": 106274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:26.630611+00:00", "epoch": 1, "step": 2145, "train_loss": 3.7811636924743652, "perplexity": 43.86705981251692, "lr": 0.001305, "grad_norm": 0.235657, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:26.936519+00:00", "epoch": 1, "step": 2146, "train_loss": 3.690757989883423, "perplexity": 40.07521205292736, "lr": 0.001305, "grad_norm": 0.237009, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:27.242717+00:00", "epoch": 1, "step": 2147, "train_loss": 3.7965781688690186, "perplexity": 44.548485991847485, "lr": 0.001305, "grad_norm": 0.252427, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:27.549685+00:00", "epoch": 1, "step": 2148, "train_loss": 3.6841611862182617, "perplexity": 39.811713825781446, "lr": 0.001305, "grad_norm": 0.259998, "tokens_per_sec": 106747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:27.856549+00:00", "epoch": 1, "step": 2149, "train_loss": 3.7310752868652344, "perplexity": 41.72394926579604, "lr": 0.001305, "grad_norm": 0.263151, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:28.163072+00:00", "epoch": 1, "step": 2150, "train_loss": 3.6568446159362793, "perplexity": 38.73891366650239, "lr": 0.001305, "grad_norm": 0.25493, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:28.469350+00:00", "epoch": 1, "step": 2151, "train_loss": 3.715189218521118, "perplexity": 41.06635686459003, "lr": 0.001305, "grad_norm": 0.236536, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:28.775101+00:00", "epoch": 1, "step": 2152, "train_loss": 3.7011075019836426, "perplexity": 40.49212464457493, "lr": 0.001305, "grad_norm": 0.243776, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:29.081018+00:00", "epoch": 1, "step": 2153, "train_loss": 3.7052738666534424, "perplexity": 40.6611815338755, "lr": 0.001305, "grad_norm": 0.265636, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:29.389499+00:00", "epoch": 1, "step": 2154, "train_loss": 3.757744789123535, "perplexity": 42.851677356646206, "lr": 0.001305, "grad_norm": 0.262087, "tokens_per_sec": 106225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:29.697087+00:00", "epoch": 1, "step": 2155, "train_loss": 3.8617970943450928, "perplexity": 47.55072777513754, "lr": 0.001305, "grad_norm": 0.270962, "tokens_per_sec": 106540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:30.003606+00:00", "epoch": 1, "step": 2156, "train_loss": 3.6715996265411377, "perplexity": 39.314744496484685, "lr": 0.001305, "grad_norm": 0.288604, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:30.310223+00:00", "epoch": 1, "step": 2157, "train_loss": 3.681001901626587, "perplexity": 39.68613576459074, "lr": 0.001305, "grad_norm": 0.292038, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:30.615896+00:00", "epoch": 1, "step": 2158, "train_loss": 3.7161812782287598, "perplexity": 41.10711735765346, "lr": 0.001305, "grad_norm": 0.305187, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:30.921443+00:00", "epoch": 1, "step": 2159, "train_loss": 3.63411021232605, "perplexity": 37.86814329503267, "lr": 0.001305, "grad_norm": 0.310024, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:31.227782+00:00", "epoch": 1, "step": 2160, "train_loss": 3.728485107421875, "perplexity": 41.61601659295086, "lr": 0.001305, "grad_norm": 0.322456, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:31.535233+00:00", "epoch": 1, "step": 2161, "train_loss": 3.751225471496582, "perplexity": 42.57322231576777, "lr": 0.001305, "grad_norm": 0.277837, "tokens_per_sec": 106517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:31.841586+00:00", "epoch": 1, "step": 2162, "train_loss": 3.747769832611084, "perplexity": 42.426358533497954, "lr": 0.001305, "grad_norm": 0.262679, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:32.147408+00:00", "epoch": 1, "step": 2163, "train_loss": 3.702106237411499, "perplexity": 40.53258576562092, "lr": 0.001305, "grad_norm": 0.263325, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:32.453066+00:00", "epoch": 1, "step": 2164, "train_loss": 3.7062265872955322, "perplexity": 40.69993874031831, "lr": 0.001305, "grad_norm": 0.247061, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:32.759993+00:00", "epoch": 1, "step": 2165, "train_loss": 3.7333896160125732, "perplexity": 41.820624043201406, "lr": 0.001305, "grad_norm": 0.266265, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:33.067569+00:00", "epoch": 1, "step": 2166, "train_loss": 3.6997087001800537, "perplexity": 40.43552378351321, "lr": 0.001305, "grad_norm": 0.226605, "tokens_per_sec": 106536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:33.375888+00:00", "epoch": 1, "step": 2167, "train_loss": 3.7020726203918457, "perplexity": 40.531223203791406, "lr": 0.001305, "grad_norm": 0.240583, "tokens_per_sec": 106281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:33.682216+00:00", "epoch": 1, "step": 2168, "train_loss": 3.637441396713257, "perplexity": 37.994499403652156, "lr": 0.001305, "grad_norm": 0.248057, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:33.988230+00:00", "epoch": 1, "step": 2169, "train_loss": 3.724196434020996, "perplexity": 41.43792125887055, "lr": 0.001305, "grad_norm": 0.270133, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:34.295755+00:00", "epoch": 1, "step": 2170, "train_loss": 3.7505433559417725, "perplexity": 42.544192360625395, "lr": 0.001305, "grad_norm": 0.269046, "tokens_per_sec": 106642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:34.603595+00:00", "epoch": 1, "step": 2171, "train_loss": 3.785081148147583, "perplexity": 44.039244116835185, "lr": 0.001305, "grad_norm": 0.252631, "tokens_per_sec": 106357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:34.910898+00:00", "epoch": 1, "step": 2172, "train_loss": 3.633159875869751, "perplexity": 37.83217291261712, "lr": 0.001305, "grad_norm": 0.251741, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:35.217529+00:00", "epoch": 1, "step": 2173, "train_loss": 3.78244948387146, "perplexity": 43.92349997800346, "lr": 0.001305, "grad_norm": 0.241741, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:35.523486+00:00", "epoch": 1, "step": 2174, "train_loss": 3.686476945877075, "perplexity": 39.904015019048366, "lr": 0.001305, "grad_norm": 0.26044, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:35.831776+00:00", "epoch": 1, "step": 2175, "train_loss": 3.763187885284424, "perplexity": 43.08555910012381, "lr": 0.001305, "grad_norm": 0.27366, "tokens_per_sec": 106290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:36.139776+00:00", "epoch": 1, "step": 2176, "train_loss": 3.714521646499634, "perplexity": 41.03895126234736, "lr": 0.001305, "grad_norm": 0.283987, "tokens_per_sec": 106390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:36.448676+00:00", "epoch": 1, "step": 2177, "train_loss": 3.721832036972046, "perplexity": 41.34006129562508, "lr": 0.001305, "grad_norm": 0.297457, "tokens_per_sec": 106085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:36.755973+00:00", "epoch": 1, "step": 2178, "train_loss": 3.7148473262786865, "perplexity": 41.05231899560909, "lr": 0.001305, "grad_norm": 0.278678, "tokens_per_sec": 106628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:37.063266+00:00", "epoch": 1, "step": 2179, "train_loss": 3.69397234916687, "perplexity": 40.204235435489316, "lr": 0.001305, "grad_norm": 0.259023, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:37.371075+00:00", "epoch": 1, "step": 2180, "train_loss": 3.7960004806518555, "perplexity": 44.5227582884086, "lr": 0.001305, "grad_norm": 0.227851, "tokens_per_sec": 106525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:37.677705+00:00", "epoch": 1, "step": 2181, "train_loss": 3.7280044555664062, "perplexity": 41.59601858378335, "lr": 0.001305, "grad_norm": 0.272874, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:37.985155+00:00", "epoch": 1, "step": 2182, "train_loss": 3.733832359313965, "perplexity": 41.839143943835, "lr": 0.001305, "grad_norm": 0.268554, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:38.292059+00:00", "epoch": 1, "step": 2183, "train_loss": 3.7815206050872803, "perplexity": 43.88271931382604, "lr": 0.001305, "grad_norm": 0.250338, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:38.599686+00:00", "epoch": 1, "step": 2184, "train_loss": 3.6190526485443115, "perplexity": 37.30221277229868, "lr": 0.001305, "grad_norm": 0.237447, "tokens_per_sec": 106518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:38.907028+00:00", "epoch": 1, "step": 2185, "train_loss": 3.715334892272949, "perplexity": 41.07233959062102, "lr": 0.001305, "grad_norm": 0.280939, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:39.213598+00:00", "epoch": 1, "step": 2186, "train_loss": 3.7282769680023193, "perplexity": 41.60735556079535, "lr": 0.001305, "grad_norm": 0.251112, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:39.520889+00:00", "epoch": 1, "step": 2187, "train_loss": 3.6673684120178223, "perplexity": 39.14874681215419, "lr": 0.001305, "grad_norm": 0.234243, "tokens_per_sec": 106635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:39.827849+00:00", "epoch": 1, "step": 2188, "train_loss": 3.668525218963623, "perplexity": 39.19406055896116, "lr": 0.001305, "grad_norm": 0.236066, "tokens_per_sec": 106751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:40.134762+00:00", "epoch": 1, "step": 2189, "train_loss": 3.7734367847442627, "perplexity": 43.52940926621624, "lr": 0.001305, "grad_norm": 0.246368, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:40.443028+00:00", "epoch": 1, "step": 2190, "train_loss": 3.721897840499878, "perplexity": 41.34278170700448, "lr": 0.001305, "grad_norm": 0.270065, "tokens_per_sec": 106355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:40.750068+00:00", "epoch": 1, "step": 2191, "train_loss": 3.7256269454956055, "perplexity": 41.49724109945581, "lr": 0.001305, "grad_norm": 0.280938, "tokens_per_sec": 106665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:41.056352+00:00", "epoch": 1, "step": 2192, "train_loss": 3.7837698459625244, "perplexity": 43.98153320628756, "lr": 0.001305, "grad_norm": 0.286905, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:41.362734+00:00", "epoch": 1, "step": 2193, "train_loss": 3.789473056793213, "perplexity": 44.233085809740366, "lr": 0.001305, "grad_norm": 0.321733, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:41.669852+00:00", "epoch": 1, "step": 2194, "train_loss": 3.670583486557007, "perplexity": 39.274815502795114, "lr": 0.001305, "grad_norm": 0.329303, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:41.977032+00:00", "epoch": 1, "step": 2195, "train_loss": 3.6701667308807373, "perplexity": 39.25845091075506, "lr": 0.001305, "grad_norm": 0.289977, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:42.284228+00:00", "epoch": 1, "step": 2196, "train_loss": 3.6767218112945557, "perplexity": 39.51663850912529, "lr": 0.001305, "grad_norm": 0.253589, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:42.591743+00:00", "epoch": 1, "step": 2197, "train_loss": 3.8452634811401367, "perplexity": 46.77100600556678, "lr": 0.001305, "grad_norm": 0.260344, "tokens_per_sec": 106558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:42.897638+00:00", "epoch": 1, "step": 2198, "train_loss": 3.9155216217041016, "perplexity": 50.175237178874355, "lr": 0.001305, "grad_norm": 0.285602, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:43.205192+00:00", "epoch": 1, "step": 2199, "train_loss": 3.785266876220703, "perplexity": 44.04742420039874, "lr": 0.001305, "grad_norm": 0.264803, "tokens_per_sec": 106544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:43.513505+00:00", "epoch": 1, "step": 2200, "train_loss": 3.7313430309295654, "perplexity": 41.73512210121584, "lr": 0.001305, "grad_norm": 0.275474, "tokens_per_sec": 106343} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:13:49.337555+00:00", "step": 2200, "epoch": 1, "val_loss": 3.7864866971969606, "val_ppl": 44.10118695619071, "eval_train_loss": 3.7313430309295654, "eval_train_ppl": 41.73512210121584} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:13:50.199077+00:00", "step": 2200, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p7865_epoch_0001_step_0002200.pt", "val_loss": 3.7864866971969606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:51.186983+00:00", "epoch": 1, "step": 2201, "train_loss": 3.878709316253662, "perplexity": 48.361755040767385, "lr": 0.001305, "grad_norm": 0.285289, "tokens_per_sec": 4270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:51.493555+00:00", "epoch": 1, "step": 2202, "train_loss": 3.7174277305603027, "perplexity": 41.15838736609975, "lr": 0.001305, "grad_norm": 0.287044, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:51.799343+00:00", "epoch": 1, "step": 2203, "train_loss": 3.7906851768493652, "perplexity": 44.286734127728465, "lr": 0.001305, "grad_norm": 0.265916, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:52.105444+00:00", "epoch": 1, "step": 2204, "train_loss": 3.707655668258667, "perplexity": 40.758143827958314, "lr": 0.001305, "grad_norm": 0.271874, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:52.411165+00:00", "epoch": 1, "step": 2205, "train_loss": 3.6483027935028076, "perplexity": 38.40942197782852, "lr": 0.001305, "grad_norm": 0.29416, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:52.717323+00:00", "epoch": 1, "step": 2206, "train_loss": 3.8671481609344482, "perplexity": 47.805856883479976, "lr": 0.001305, "grad_norm": 0.296338, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:53.023130+00:00", "epoch": 1, "step": 2207, "train_loss": 3.7601542472839355, "perplexity": 42.95505116776513, "lr": 0.001305, "grad_norm": 0.294838, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:53.329289+00:00", "epoch": 1, "step": 2208, "train_loss": 3.681333065032959, "perplexity": 39.69928053690994, "lr": 0.001305, "grad_norm": 0.293248, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:53.635686+00:00", "epoch": 1, "step": 2209, "train_loss": 3.737847328186035, "perplexity": 42.0074644791483, "lr": 0.001305, "grad_norm": 0.254303, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:53.942418+00:00", "epoch": 1, "step": 2210, "train_loss": 3.7195370197296143, "perplexity": 41.245293930113164, "lr": 0.001305, "grad_norm": 0.252785, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:54.249218+00:00", "epoch": 1, "step": 2211, "train_loss": 3.8074753284454346, "perplexity": 45.03659261030986, "lr": 0.001305, "grad_norm": 0.258394, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:54.554730+00:00", "epoch": 1, "step": 2212, "train_loss": 3.8000874519348145, "perplexity": 44.705093869312016, "lr": 0.001305, "grad_norm": 0.267078, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:54.860584+00:00", "epoch": 1, "step": 2213, "train_loss": 3.754058837890625, "perplexity": 42.694018902846004, "lr": 0.001305, "grad_norm": 0.291303, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:55.166762+00:00", "epoch": 1, "step": 2214, "train_loss": 3.7345316410064697, "perplexity": 41.86841152317356, "lr": 0.001305, "grad_norm": 0.300581, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:55.517316+00:00", "epoch": 1, "step": 2215, "train_loss": 3.747114658355713, "perplexity": 42.39857097947666, "lr": 0.001305, "grad_norm": 0.314015, "tokens_per_sec": 93473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:55.823097+00:00", "epoch": 1, "step": 2216, "train_loss": 3.7059402465820312, "perplexity": 40.68828635917514, "lr": 0.001305, "grad_norm": 0.288309, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:56.130496+00:00", "epoch": 1, "step": 2217, "train_loss": 3.6494760513305664, "perplexity": 38.45451257910176, "lr": 0.001305, "grad_norm": 0.283962, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:56.436603+00:00", "epoch": 1, "step": 2218, "train_loss": 3.6985132694244385, "perplexity": 40.387214795541, "lr": 0.001305, "grad_norm": 0.263639, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:56.743013+00:00", "epoch": 1, "step": 2219, "train_loss": 3.723743438720703, "perplexity": 41.41915432627343, "lr": 0.001305, "grad_norm": 0.280225, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:57.049248+00:00", "epoch": 1, "step": 2220, "train_loss": 3.7648839950561523, "perplexity": 43.158698947004424, "lr": 0.001305, "grad_norm": 0.310738, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:57.356460+00:00", "epoch": 1, "step": 2221, "train_loss": 3.7962639331817627, "perplexity": 44.53448946695461, "lr": 0.001305, "grad_norm": 0.302038, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:57.662951+00:00", "epoch": 1, "step": 2222, "train_loss": 3.6120193004608154, "perplexity": 37.04077379897484, "lr": 0.001305, "grad_norm": 0.267093, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:57.970757+00:00", "epoch": 1, "step": 2223, "train_loss": 3.6825761795043945, "perplexity": 39.7486619740813, "lr": 0.001305, "grad_norm": 0.261506, "tokens_per_sec": 106461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:58.277187+00:00", "epoch": 1, "step": 2224, "train_loss": 3.731684923171997, "perplexity": 41.749393455192774, "lr": 0.001305, "grad_norm": 0.285036, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:58.583163+00:00", "epoch": 1, "step": 2225, "train_loss": 3.585620880126953, "perplexity": 36.07574948275809, "lr": 0.001305, "grad_norm": 0.249182, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:58.889790+00:00", "epoch": 1, "step": 2226, "train_loss": 3.87566876411438, "perplexity": 48.21493192779828, "lr": 0.001305, "grad_norm": 0.281343, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:59.197464+00:00", "epoch": 1, "step": 2227, "train_loss": 3.6785271167755127, "perplexity": 39.588042546872266, "lr": 0.001305, "grad_norm": 0.277602, "tokens_per_sec": 106503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:59.503891+00:00", "epoch": 1, "step": 2228, "train_loss": 3.7243118286132812, "perplexity": 41.442703246801834, "lr": 0.001305, "grad_norm": 0.249322, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:13:59.810746+00:00", "epoch": 1, "step": 2229, "train_loss": 3.764667272567749, "perplexity": 43.149346499851795, "lr": 0.001305, "grad_norm": 0.239199, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:00.117861+00:00", "epoch": 1, "step": 2230, "train_loss": 3.7343695163726807, "perplexity": 41.86162417250128, "lr": 0.001305, "grad_norm": 0.226108, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:00.424829+00:00", "epoch": 1, "step": 2231, "train_loss": 3.616334915161133, "perplexity": 37.20097293713457, "lr": 0.001305, "grad_norm": 0.250116, "tokens_per_sec": 106689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:00.731520+00:00", "epoch": 1, "step": 2232, "train_loss": 3.650254011154175, "perplexity": 38.484440284691615, "lr": 0.001305, "grad_norm": 0.249134, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:01.038025+00:00", "epoch": 1, "step": 2233, "train_loss": 3.6386666297912598, "perplexity": 38.04108005135163, "lr": 0.001305, "grad_norm": 0.252939, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:01.344477+00:00", "epoch": 1, "step": 2234, "train_loss": 3.8558247089385986, "perplexity": 47.26758286938991, "lr": 0.001305, "grad_norm": 0.262987, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:01.650894+00:00", "epoch": 1, "step": 2235, "train_loss": 3.6352293491363525, "perplexity": 37.910546651284655, "lr": 0.001305, "grad_norm": 0.246661, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:01.957312+00:00", "epoch": 1, "step": 2236, "train_loss": 3.6467537879943848, "perplexity": 38.34997162795643, "lr": 0.001305, "grad_norm": 0.215955, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:02.264643+00:00", "epoch": 1, "step": 2237, "train_loss": 3.6939198970794678, "perplexity": 40.20212669472272, "lr": 0.001305, "grad_norm": 0.227217, "tokens_per_sec": 106622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:02.570716+00:00", "epoch": 1, "step": 2238, "train_loss": 3.704012393951416, "perplexity": 40.60992090207987, "lr": 0.001305, "grad_norm": 0.24161, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:02.878001+00:00", "epoch": 1, "step": 2239, "train_loss": 3.7854056358337402, "perplexity": 44.053536628005304, "lr": 0.001305, "grad_norm": 0.247876, "tokens_per_sec": 106636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:03.184680+00:00", "epoch": 1, "step": 2240, "train_loss": 3.7535831928253174, "perplexity": 42.67371653218069, "lr": 0.001305, "grad_norm": 0.243129, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:03.491945+00:00", "epoch": 1, "step": 2241, "train_loss": 3.6921589374542236, "perplexity": 40.131394669163186, "lr": 0.001305, "grad_norm": 0.269527, "tokens_per_sec": 106591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:03.799026+00:00", "epoch": 1, "step": 2242, "train_loss": 3.70651912689209, "perplexity": 40.71184682568564, "lr": 0.001305, "grad_norm": 0.256407, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:04.106044+00:00", "epoch": 1, "step": 2243, "train_loss": 3.8060317039489746, "perplexity": 44.971623588699316, "lr": 0.001305, "grad_norm": 0.214274, "tokens_per_sec": 106730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:04.412339+00:00", "epoch": 1, "step": 2244, "train_loss": 3.6880428791046143, "perplexity": 39.96655099287963, "lr": 0.001305, "grad_norm": 0.253817, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:04.719141+00:00", "epoch": 1, "step": 2245, "train_loss": 3.7756245136260986, "perplexity": 43.624744057391304, "lr": 0.001305, "grad_norm": 0.256774, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:05.025985+00:00", "epoch": 1, "step": 2246, "train_loss": 3.8264293670654297, "perplexity": 45.898359104282086, "lr": 0.001305, "grad_norm": 0.251264, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:05.333886+00:00", "epoch": 1, "step": 2247, "train_loss": 3.7238574028015137, "perplexity": 41.42387489110648, "lr": 0.001305, "grad_norm": 0.274255, "tokens_per_sec": 106424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:05.641687+00:00", "epoch": 1, "step": 2248, "train_loss": 3.763885974884033, "perplexity": 43.115647181691465, "lr": 0.001305, "grad_norm": 0.265696, "tokens_per_sec": 106458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:05.948747+00:00", "epoch": 1, "step": 2249, "train_loss": 3.732543468475342, "perplexity": 41.78525259200532, "lr": 0.001305, "grad_norm": 0.254573, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:06.255267+00:00", "epoch": 1, "step": 2250, "train_loss": 3.8897712230682373, "perplexity": 48.89969812105621, "lr": 0.001305, "grad_norm": 0.268565, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:14:07.191949+00:00", "step": 2250, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0001_step_0002250.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:08.411345+00:00", "epoch": 1, "step": 2251, "train_loss": 3.6810786724090576, "perplexity": 39.68918261723976, "lr": 0.001305, "grad_norm": 0.287066, "tokens_per_sec": 15197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:08.718383+00:00", "epoch": 1, "step": 2252, "train_loss": 3.7069082260131836, "perplexity": 40.72769085175203, "lr": 0.001305, "grad_norm": 0.303155, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:09.026441+00:00", "epoch": 1, "step": 2253, "train_loss": 3.6617331504821777, "perplexity": 38.928753826258365, "lr": 0.001305, "grad_norm": 0.254027, "tokens_per_sec": 106371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:09.405016+00:00", "epoch": 1, "step": 2254, "train_loss": 3.7099053859710693, "perplexity": 40.84994136660968, "lr": 0.001305, "grad_norm": 0.277773, "tokens_per_sec": 86556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:09.712389+00:00", "epoch": 1, "step": 2255, "train_loss": 3.736588478088379, "perplexity": 41.95461664911344, "lr": 0.001305, "grad_norm": 0.285627, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:10.019692+00:00", "epoch": 1, "step": 2256, "train_loss": 3.665903329849243, "perplexity": 39.09143267648847, "lr": 0.001305, "grad_norm": 0.265532, "tokens_per_sec": 106632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:10.326352+00:00", "epoch": 1, "step": 2257, "train_loss": 3.7456135749816895, "perplexity": 42.33497493292257, "lr": 0.001305, "grad_norm": 0.261914, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:10.632505+00:00", "epoch": 1, "step": 2258, "train_loss": 3.654238224029541, "perplexity": 38.63807634332237, "lr": 0.001305, "grad_norm": 0.268128, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:10.939056+00:00", "epoch": 1, "step": 2259, "train_loss": 3.8182904720306396, "perplexity": 45.526313252073685, "lr": 0.001305, "grad_norm": 0.26295, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:11.246087+00:00", "epoch": 1, "step": 2260, "train_loss": 3.7875359058380127, "perplexity": 44.14748258526583, "lr": 0.001305, "grad_norm": 0.241973, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:11.553444+00:00", "epoch": 1, "step": 2261, "train_loss": 3.6752281188964844, "perplexity": 39.45765686776779, "lr": 0.001305, "grad_norm": 0.237343, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:11.860924+00:00", "epoch": 1, "step": 2262, "train_loss": 3.7779948711395264, "perplexity": 43.72827294900825, "lr": 0.001305, "grad_norm": 0.248212, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:12.169688+00:00", "epoch": 1, "step": 2263, "train_loss": 3.6282711029052734, "perplexity": 37.64767136904053, "lr": 0.001305, "grad_norm": 0.243979, "tokens_per_sec": 106127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:12.476948+00:00", "epoch": 1, "step": 2264, "train_loss": 3.7130160331726074, "perplexity": 40.97720896207814, "lr": 0.001305, "grad_norm": 0.253953, "tokens_per_sec": 106646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:12.783474+00:00", "epoch": 1, "step": 2265, "train_loss": 3.6138906478881836, "perplexity": 37.110154853512874, "lr": 0.001305, "grad_norm": 0.26528, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:13.090448+00:00", "epoch": 1, "step": 2266, "train_loss": 3.757814407348633, "perplexity": 42.854660718213196, "lr": 0.001305, "grad_norm": 0.277693, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:13.397885+00:00", "epoch": 1, "step": 2267, "train_loss": 3.642491579055786, "perplexity": 38.18686388265738, "lr": 0.001305, "grad_norm": 0.281919, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:13.704408+00:00", "epoch": 1, "step": 2268, "train_loss": 3.7188198566436768, "perplexity": 41.215724932002146, "lr": 0.001305, "grad_norm": 0.299081, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:14.012002+00:00", "epoch": 1, "step": 2269, "train_loss": 3.624539375305176, "perplexity": 37.507442325742204, "lr": 0.001305, "grad_norm": 0.315544, "tokens_per_sec": 106532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:14.319564+00:00", "epoch": 1, "step": 2270, "train_loss": 3.692352771759033, "perplexity": 40.13917426410175, "lr": 0.001305, "grad_norm": 0.321224, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:14.626808+00:00", "epoch": 1, "step": 2271, "train_loss": 3.7068235874176025, "perplexity": 40.724243863073276, "lr": 0.001305, "grad_norm": 0.243893, "tokens_per_sec": 106581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:14.934247+00:00", "epoch": 1, "step": 2272, "train_loss": 3.737292766571045, "perplexity": 41.98417521006929, "lr": 0.001305, "grad_norm": 0.247054, "tokens_per_sec": 106582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:15.253516+00:00", "epoch": 1, "step": 2273, "train_loss": 3.684293270111084, "perplexity": 39.81697265921943, "lr": 0.001305, "grad_norm": 0.282969, "tokens_per_sec": 102633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:15.559877+00:00", "epoch": 1, "step": 2274, "train_loss": 3.665609121322632, "perplexity": 39.07993333536265, "lr": 0.001305, "grad_norm": 0.255599, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:15.867543+00:00", "epoch": 1, "step": 2275, "train_loss": 3.6499807834625244, "perplexity": 38.47392670627396, "lr": 0.001305, "grad_norm": 0.251248, "tokens_per_sec": 106505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:16.174125+00:00", "epoch": 1, "step": 2276, "train_loss": 3.712824821472168, "perplexity": 40.969374389328095, "lr": 0.001305, "grad_norm": 0.242427, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:16.481066+00:00", "epoch": 1, "step": 2277, "train_loss": 3.7751810550689697, "perplexity": 43.605402580225245, "lr": 0.001305, "grad_norm": 0.247951, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:16.787634+00:00", "epoch": 1, "step": 2278, "train_loss": 3.6319403648376465, "perplexity": 37.786064281088535, "lr": 0.001305, "grad_norm": 0.236404, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:17.095866+00:00", "epoch": 1, "step": 2279, "train_loss": 3.624769449234009, "perplexity": 37.51607280314442, "lr": 0.001305, "grad_norm": 0.219961, "tokens_per_sec": 106310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:17.404226+00:00", "epoch": 1, "step": 2280, "train_loss": 3.6053452491760254, "perplexity": 36.79438489567408, "lr": 0.001305, "grad_norm": 0.235942, "tokens_per_sec": 106323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:17.711093+00:00", "epoch": 1, "step": 2281, "train_loss": 3.7409555912017822, "perplexity": 42.13823786112074, "lr": 0.001305, "grad_norm": 0.22912, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:18.016862+00:00", "epoch": 1, "step": 2282, "train_loss": 3.698967933654785, "perplexity": 40.40558159251827, "lr": 0.001305, "grad_norm": 0.243013, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:18.322857+00:00", "epoch": 1, "step": 2283, "train_loss": 3.7152152061462402, "perplexity": 41.067424095544695, "lr": 0.001305, "grad_norm": 0.243414, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:18.629909+00:00", "epoch": 1, "step": 2284, "train_loss": 3.7492308616638184, "perplexity": 42.48838997975531, "lr": 0.001305, "grad_norm": 0.251004, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:18.937504+00:00", "epoch": 1, "step": 2285, "train_loss": 3.630401611328125, "perplexity": 37.72796555333928, "lr": 0.001305, "grad_norm": 0.278775, "tokens_per_sec": 106530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:19.257008+00:00", "epoch": 1, "step": 2286, "train_loss": 3.683668613433838, "perplexity": 39.79210848797391, "lr": 0.001305, "grad_norm": 0.264528, "tokens_per_sec": 102559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:19.564561+00:00", "epoch": 1, "step": 2287, "train_loss": 3.7262179851531982, "perplexity": 41.52177486412602, "lr": 0.001305, "grad_norm": 0.257725, "tokens_per_sec": 106544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:19.871356+00:00", "epoch": 1, "step": 2288, "train_loss": 3.6322669982910156, "perplexity": 37.79840848965988, "lr": 0.001305, "grad_norm": 0.26196, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:20.177981+00:00", "epoch": 1, "step": 2289, "train_loss": 3.730442523956299, "perplexity": 41.69755624942786, "lr": 0.001305, "grad_norm": 0.234936, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:20.484879+00:00", "epoch": 1, "step": 2290, "train_loss": 3.7717535495758057, "perplexity": 43.45620066460866, "lr": 0.001305, "grad_norm": 0.240657, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:20.804572+00:00", "epoch": 1, "step": 2291, "train_loss": 3.7984728813171387, "perplexity": 44.63297257641777, "lr": 0.001305, "grad_norm": 0.2363, "tokens_per_sec": 102446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:21.111865+00:00", "epoch": 1, "step": 2292, "train_loss": 3.700138807296753, "perplexity": 40.45291913072311, "lr": 0.001305, "grad_norm": 0.26229, "tokens_per_sec": 106636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:21.418684+00:00", "epoch": 1, "step": 2293, "train_loss": 3.7327308654785156, "perplexity": 41.79308375686342, "lr": 0.001305, "grad_norm": 0.295646, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:21.725428+00:00", "epoch": 1, "step": 2294, "train_loss": 3.7805092334747314, "perplexity": 43.83836001288061, "lr": 0.001305, "grad_norm": 0.287907, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:22.032457+00:00", "epoch": 1, "step": 2295, "train_loss": 3.714132070541382, "perplexity": 41.02296658740837, "lr": 0.001305, "grad_norm": 0.299373, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:22.339509+00:00", "epoch": 1, "step": 2296, "train_loss": 3.787869930267334, "perplexity": 44.162231386035096, "lr": 0.001305, "grad_norm": 0.28952, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:22.655688+00:00", "epoch": 1, "step": 2297, "train_loss": 3.7269678115844727, "perplexity": 41.55292066390523, "lr": 0.001305, "grad_norm": 0.263551, "tokens_per_sec": 103639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:22.962952+00:00", "epoch": 1, "step": 2298, "train_loss": 3.7835376262664795, "perplexity": 43.97132101379682, "lr": 0.001305, "grad_norm": 0.282569, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:23.285074+00:00", "epoch": 1, "step": 2299, "train_loss": 3.726925849914551, "perplexity": 41.551177070546345, "lr": 0.001305, "grad_norm": 0.295824, "tokens_per_sec": 101726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:23.592578+00:00", "epoch": 1, "step": 2300, "train_loss": 3.674064874649048, "perplexity": 39.41178466086205, "lr": 0.001305, "grad_norm": 0.258996, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:14:29.443729+00:00", "step": 2300, "epoch": 1, "val_loss": 3.7748889088630677, "val_ppl": 43.592665287970995, "eval_train_loss": 3.674064874649048, "eval_train_ppl": 39.41178466086205} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:14:30.304225+00:00", "step": 2300, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p7749_epoch_0001_step_0002300.pt", "val_loss": 3.7748889088630677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:31.507996+00:00", "epoch": 1, "step": 2301, "train_loss": 3.737673282623291, "perplexity": 42.00015390255873, "lr": 0.001305, "grad_norm": 0.243078, "tokens_per_sec": 4140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:31.814473+00:00", "epoch": 1, "step": 2302, "train_loss": 3.718515157699585, "perplexity": 41.203168457204846, "lr": 0.001305, "grad_norm": 0.247438, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:32.121006+00:00", "epoch": 1, "step": 2303, "train_loss": 3.6727218627929688, "perplexity": 39.35888969402754, "lr": 0.001305, "grad_norm": 0.250296, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:32.426053+00:00", "epoch": 1, "step": 2304, "train_loss": 3.719740390777588, "perplexity": 41.25368288176975, "lr": 0.001305, "grad_norm": 0.253186, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:32.731790+00:00", "epoch": 1, "step": 2305, "train_loss": 3.6516480445861816, "perplexity": 38.538126292411214, "lr": 0.001305, "grad_norm": 0.264125, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:33.039458+00:00", "epoch": 1, "step": 2306, "train_loss": 3.781432867050171, "perplexity": 43.87886929906926, "lr": 0.001305, "grad_norm": 0.264458, "tokens_per_sec": 106507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:33.346396+00:00", "epoch": 1, "step": 2307, "train_loss": 3.6979174613952637, "perplexity": 40.363158935738646, "lr": 0.001305, "grad_norm": 0.241793, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:33.651516+00:00", "epoch": 1, "step": 2308, "train_loss": 3.783233880996704, "perplexity": 43.95796696125092, "lr": 0.001305, "grad_norm": 0.222478, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:33.957818+00:00", "epoch": 1, "step": 2309, "train_loss": 3.6649656295776367, "perplexity": 39.0547938102723, "lr": 0.001305, "grad_norm": 0.213588, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:34.263620+00:00", "epoch": 1, "step": 2310, "train_loss": 3.612797498703003, "perplexity": 37.069610082751886, "lr": 0.001305, "grad_norm": 0.23284, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:34.570142+00:00", "epoch": 1, "step": 2311, "train_loss": 3.662574291229248, "perplexity": 38.96151216258862, "lr": 0.001305, "grad_norm": 0.252064, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:34.875733+00:00", "epoch": 1, "step": 2312, "train_loss": 3.609138011932373, "perplexity": 36.93420224776538, "lr": 0.001305, "grad_norm": 0.255003, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:35.182778+00:00", "epoch": 1, "step": 2313, "train_loss": 3.7053182125091553, "perplexity": 40.662984728746736, "lr": 0.001305, "grad_norm": 0.231232, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:35.490175+00:00", "epoch": 1, "step": 2314, "train_loss": 3.696913719177246, "perplexity": 40.32266505517083, "lr": 0.001305, "grad_norm": 0.257483, "tokens_per_sec": 106600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:35.856912+00:00", "epoch": 1, "step": 2315, "train_loss": 3.7617523670196533, "perplexity": 43.02375336533196, "lr": 0.001305, "grad_norm": 0.278397, "tokens_per_sec": 89349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:36.163059+00:00", "epoch": 1, "step": 2316, "train_loss": 3.646315097808838, "perplexity": 38.33315156145604, "lr": 0.001305, "grad_norm": 0.258854, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:36.468801+00:00", "epoch": 1, "step": 2317, "train_loss": 3.6559221744537354, "perplexity": 38.70319576191884, "lr": 0.001305, "grad_norm": 0.260925, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:36.774735+00:00", "epoch": 1, "step": 2318, "train_loss": 3.644556999206543, "perplexity": 38.26581730874575, "lr": 0.001305, "grad_norm": 0.230499, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:37.081437+00:00", "epoch": 1, "step": 2319, "train_loss": 3.7296266555786133, "perplexity": 41.663550405889694, "lr": 0.001305, "grad_norm": 0.219594, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:37.387849+00:00", "epoch": 1, "step": 2320, "train_loss": 3.7301764488220215, "perplexity": 41.68646304242847, "lr": 0.001305, "grad_norm": 0.251992, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:37.694352+00:00", "epoch": 1, "step": 2321, "train_loss": 3.789654016494751, "perplexity": 44.24109094002775, "lr": 0.001305, "grad_norm": 0.244234, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:37.999942+00:00", "epoch": 1, "step": 2322, "train_loss": 3.7200214862823486, "perplexity": 41.265280736558694, "lr": 0.001305, "grad_norm": 0.236961, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:38.306035+00:00", "epoch": 1, "step": 2323, "train_loss": 3.649022102355957, "perplexity": 38.43706015410229, "lr": 0.001305, "grad_norm": 0.247743, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:38.612405+00:00", "epoch": 1, "step": 2324, "train_loss": 3.659681558609009, "perplexity": 38.848969781409274, "lr": 0.001305, "grad_norm": 0.24318, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:38.919288+00:00", "epoch": 1, "step": 2325, "train_loss": 3.862100601196289, "perplexity": 47.565161937119385, "lr": 0.001305, "grad_norm": 0.269603, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:39.227601+00:00", "epoch": 1, "step": 2326, "train_loss": 3.7475082874298096, "perplexity": 42.415263574844225, "lr": 0.001305, "grad_norm": 0.286773, "tokens_per_sec": 106282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:39.534147+00:00", "epoch": 1, "step": 2327, "train_loss": 3.6325693130493164, "perplexity": 37.80983723383856, "lr": 0.001305, "grad_norm": 0.274265, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:39.840197+00:00", "epoch": 1, "step": 2328, "train_loss": 3.6604950428009033, "perplexity": 38.880585661963856, "lr": 0.001305, "grad_norm": 0.259394, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:40.146469+00:00", "epoch": 1, "step": 2329, "train_loss": 3.737396478652954, "perplexity": 41.98852970209038, "lr": 0.001305, "grad_norm": 0.28249, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:40.454018+00:00", "epoch": 1, "step": 2330, "train_loss": 3.775913953781128, "perplexity": 43.637372637595334, "lr": 0.001305, "grad_norm": 0.341123, "tokens_per_sec": 106603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:40.760941+00:00", "epoch": 1, "step": 2331, "train_loss": 3.740987539291382, "perplexity": 42.139584118824565, "lr": 0.001305, "grad_norm": 0.326922, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:41.068249+00:00", "epoch": 1, "step": 2332, "train_loss": 3.722437858581543, "perplexity": 41.3651135859391, "lr": 0.001305, "grad_norm": 0.299287, "tokens_per_sec": 106630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:41.375712+00:00", "epoch": 1, "step": 2333, "train_loss": 3.689211130142212, "perplexity": 40.01326924155405, "lr": 0.001305, "grad_norm": 0.322551, "tokens_per_sec": 106576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:41.682216+00:00", "epoch": 1, "step": 2334, "train_loss": 3.7690515518188477, "perplexity": 43.33894059744713, "lr": 0.001305, "grad_norm": 0.290069, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:41.989347+00:00", "epoch": 1, "step": 2335, "train_loss": 3.808022975921631, "perplexity": 45.06126354145936, "lr": 0.001305, "grad_norm": 0.268237, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:42.296382+00:00", "epoch": 1, "step": 2336, "train_loss": 3.770416021347046, "perplexity": 43.39811562335569, "lr": 0.001305, "grad_norm": 0.274373, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:42.603450+00:00", "epoch": 1, "step": 2337, "train_loss": 3.747757911682129, "perplexity": 42.425852774906616, "lr": 0.001305, "grad_norm": 0.269177, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:42.909677+00:00", "epoch": 1, "step": 2338, "train_loss": 3.7566871643066406, "perplexity": 42.806380317087175, "lr": 0.001305, "grad_norm": 0.265994, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:43.215374+00:00", "epoch": 1, "step": 2339, "train_loss": 3.6587882041931152, "perplexity": 38.81427938042074, "lr": 0.001305, "grad_norm": 0.270967, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:43.521753+00:00", "epoch": 1, "step": 2340, "train_loss": 3.705498695373535, "perplexity": 40.670324363024, "lr": 0.001305, "grad_norm": 0.288925, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:43.828881+00:00", "epoch": 1, "step": 2341, "train_loss": 3.6850311756134033, "perplexity": 39.846364665358216, "lr": 0.001305, "grad_norm": 0.318606, "tokens_per_sec": 106633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:44.136051+00:00", "epoch": 1, "step": 2342, "train_loss": 3.622641086578369, "perplexity": 37.436309907098575, "lr": 0.001305, "grad_norm": 0.325764, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:44.443190+00:00", "epoch": 1, "step": 2343, "train_loss": 3.810938596725464, "perplexity": 45.19283681456701, "lr": 0.001305, "grad_norm": 0.281637, "tokens_per_sec": 106688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:44.749910+00:00", "epoch": 1, "step": 2344, "train_loss": 3.7482705116271973, "perplexity": 42.447605839542064, "lr": 0.001305, "grad_norm": 0.264541, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:45.056013+00:00", "epoch": 1, "step": 2345, "train_loss": 3.5914511680603027, "perplexity": 36.2866958310837, "lr": 0.001305, "grad_norm": 0.288895, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:45.362422+00:00", "epoch": 1, "step": 2346, "train_loss": 3.8334946632385254, "perplexity": 46.22379289450896, "lr": 0.001305, "grad_norm": 0.311348, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:45.670283+00:00", "epoch": 1, "step": 2347, "train_loss": 3.796215057373047, "perplexity": 44.53231286095829, "lr": 0.001305, "grad_norm": 0.284897, "tokens_per_sec": 106438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:45.977755+00:00", "epoch": 1, "step": 2348, "train_loss": 3.5793821811676025, "perplexity": 35.85138434300016, "lr": 0.001305, "grad_norm": 0.24413, "tokens_per_sec": 106572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:46.284814+00:00", "epoch": 1, "step": 2349, "train_loss": 3.713606119155884, "perplexity": 41.00139617428606, "lr": 0.001305, "grad_norm": 0.260366, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:46.591233+00:00", "epoch": 1, "step": 2350, "train_loss": 3.73543643951416, "perplexity": 41.90631114261308, "lr": 0.001305, "grad_norm": 0.248948, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:46.897256+00:00", "epoch": 1, "step": 2351, "train_loss": 3.579854726791382, "perplexity": 35.86832976120416, "lr": 0.001305, "grad_norm": 0.2298, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:47.203180+00:00", "epoch": 1, "step": 2352, "train_loss": 3.7528419494628906, "perplexity": 42.6420966435152, "lr": 0.001305, "grad_norm": 0.231916, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:47.509811+00:00", "epoch": 1, "step": 2353, "train_loss": 3.6808884143829346, "perplexity": 39.681632149987834, "lr": 0.001305, "grad_norm": 0.245425, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:47.817570+00:00", "epoch": 1, "step": 2354, "train_loss": 3.666802167892456, "perplexity": 39.12658533925042, "lr": 0.001305, "grad_norm": 0.236002, "tokens_per_sec": 106473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:48.124455+00:00", "epoch": 1, "step": 2355, "train_loss": 3.7583510875701904, "perplexity": 42.87766613974741, "lr": 0.001305, "grad_norm": 0.231371, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:48.431115+00:00", "epoch": 1, "step": 2356, "train_loss": 3.709071159362793, "perplexity": 40.8158774690555, "lr": 0.001305, "grad_norm": 0.254128, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:48.737561+00:00", "epoch": 1, "step": 2357, "train_loss": 3.7047576904296875, "perplexity": 40.640198614645826, "lr": 0.001305, "grad_norm": 0.250251, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:49.044014+00:00", "epoch": 1, "step": 2358, "train_loss": 3.705486536026001, "perplexity": 40.66982984142228, "lr": 0.001305, "grad_norm": 0.224563, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:49.351004+00:00", "epoch": 1, "step": 2359, "train_loss": 3.5690701007843018, "perplexity": 35.48358165057576, "lr": 0.001305, "grad_norm": 0.252373, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:49.657383+00:00", "epoch": 1, "step": 2360, "train_loss": 3.6529762744903564, "perplexity": 38.5893477936381, "lr": 0.001305, "grad_norm": 0.256735, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:49.963748+00:00", "epoch": 1, "step": 2361, "train_loss": 3.6301794052124023, "perplexity": 37.71958310001024, "lr": 0.001305, "grad_norm": 0.245106, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:50.270050+00:00", "epoch": 1, "step": 2362, "train_loss": 3.613981008529663, "perplexity": 37.11350830241846, "lr": 0.001305, "grad_norm": 0.265659, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:50.575955+00:00", "epoch": 1, "step": 2363, "train_loss": 3.6718475818634033, "perplexity": 39.32449400529745, "lr": 0.001305, "grad_norm": 0.275623, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:50.881775+00:00", "epoch": 1, "step": 2364, "train_loss": 3.7038354873657227, "perplexity": 40.60273737505316, "lr": 0.001305, "grad_norm": 0.265562, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:51.189344+00:00", "epoch": 1, "step": 2365, "train_loss": 3.735459327697754, "perplexity": 41.907270312933036, "lr": 0.001305, "grad_norm": 0.251335, "tokens_per_sec": 106540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:51.495919+00:00", "epoch": 1, "step": 2366, "train_loss": 3.8206920623779297, "perplexity": 45.63578020129808, "lr": 0.001305, "grad_norm": 0.30089, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:51.802966+00:00", "epoch": 1, "step": 2367, "train_loss": 3.6999192237854004, "perplexity": 40.4440373118821, "lr": 0.001305, "grad_norm": 0.313439, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:52.109218+00:00", "epoch": 1, "step": 2368, "train_loss": 3.6464319229125977, "perplexity": 38.337630097462274, "lr": 0.001305, "grad_norm": 0.342634, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:52.416064+00:00", "epoch": 1, "step": 2369, "train_loss": 3.800525188446045, "perplexity": 44.724667204805854, "lr": 0.001305, "grad_norm": 0.334667, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:52.722936+00:00", "epoch": 1, "step": 2370, "train_loss": 3.692105531692505, "perplexity": 40.12925147869191, "lr": 0.001305, "grad_norm": 0.349099, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:53.030627+00:00", "epoch": 1, "step": 2371, "train_loss": 3.7017972469329834, "perplexity": 40.520063517277116, "lr": 0.001305, "grad_norm": 0.310863, "tokens_per_sec": 106442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:53.338789+00:00", "epoch": 1, "step": 2372, "train_loss": 3.6986069679260254, "perplexity": 40.39099919434409, "lr": 0.001305, "grad_norm": 0.287146, "tokens_per_sec": 106331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:53.645934+00:00", "epoch": 1, "step": 2373, "train_loss": 3.7299869060516357, "perplexity": 41.67856242351208, "lr": 0.001305, "grad_norm": 0.275984, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:53.952343+00:00", "epoch": 1, "step": 2374, "train_loss": 3.770934581756592, "perplexity": 43.42062600395852, "lr": 0.001305, "grad_norm": 0.258733, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:54.258649+00:00", "epoch": 1, "step": 2375, "train_loss": 3.645561456680298, "perplexity": 38.30427300525347, "lr": 0.001305, "grad_norm": 0.273046, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:54.565472+00:00", "epoch": 1, "step": 2376, "train_loss": 3.720527410507202, "perplexity": 41.28616312373605, "lr": 0.001305, "grad_norm": 0.251173, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:54.872247+00:00", "epoch": 1, "step": 2377, "train_loss": 3.6728270053863525, "perplexity": 39.363028207325875, "lr": 0.001305, "grad_norm": 0.236604, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:55.178806+00:00", "epoch": 1, "step": 2378, "train_loss": 3.563473701477051, "perplexity": 35.28555599098129, "lr": 0.001305, "grad_norm": 0.26137, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:55.484521+00:00", "epoch": 1, "step": 2379, "train_loss": 3.717219829559326, "perplexity": 41.14983138559703, "lr": 0.001305, "grad_norm": 0.223681, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:55.790809+00:00", "epoch": 1, "step": 2380, "train_loss": 3.6107616424560547, "perplexity": 36.99421845479106, "lr": 0.001305, "grad_norm": 0.261303, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:56.097435+00:00", "epoch": 1, "step": 2381, "train_loss": 3.723788261413574, "perplexity": 41.42101088591447, "lr": 0.001305, "grad_norm": 0.262436, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:56.404152+00:00", "epoch": 1, "step": 2382, "train_loss": 3.6274733543395996, "perplexity": 37.61764996956124, "lr": 0.001305, "grad_norm": 0.260489, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:56.711097+00:00", "epoch": 1, "step": 2383, "train_loss": 3.7419791221618652, "perplexity": 42.181389732045666, "lr": 0.001305, "grad_norm": 0.302396, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:57.017595+00:00", "epoch": 1, "step": 2384, "train_loss": 3.795236110687256, "perplexity": 44.48873943238403, "lr": 0.001305, "grad_norm": 0.287859, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:57.323678+00:00", "epoch": 1, "step": 2385, "train_loss": 3.6552047729492188, "perplexity": 38.67543998825695, "lr": 0.001305, "grad_norm": 0.305086, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:57.630125+00:00", "epoch": 1, "step": 2386, "train_loss": 3.7488794326782227, "perplexity": 42.47346095136543, "lr": 0.001305, "grad_norm": 0.302458, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:57.937145+00:00", "epoch": 1, "step": 2387, "train_loss": 3.7079660892486572, "perplexity": 40.770797975270405, "lr": 0.001305, "grad_norm": 0.269568, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:58.244241+00:00", "epoch": 1, "step": 2388, "train_loss": 3.656968355178833, "perplexity": 38.74370748692258, "lr": 0.001305, "grad_norm": 0.24853, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:58.551205+00:00", "epoch": 1, "step": 2389, "train_loss": 3.6514647006988525, "perplexity": 38.53106121021599, "lr": 0.001305, "grad_norm": 0.244085, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:58.857370+00:00", "epoch": 1, "step": 2390, "train_loss": 3.8231756687164307, "perplexity": 45.749262378463044, "lr": 0.001305, "grad_norm": 0.255211, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:59.163610+00:00", "epoch": 1, "step": 2391, "train_loss": 3.7100038528442383, "perplexity": 40.85396393064658, "lr": 0.001305, "grad_norm": 0.224822, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:59.470413+00:00", "epoch": 1, "step": 2392, "train_loss": 3.697042226791382, "perplexity": 40.327847157615345, "lr": 0.001305, "grad_norm": 0.227951, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:14:59.777419+00:00", "epoch": 1, "step": 2393, "train_loss": 3.812974452972412, "perplexity": 45.28493665291829, "lr": 0.001305, "grad_norm": 0.214323, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:00.085076+00:00", "epoch": 1, "step": 2394, "train_loss": 3.6346275806427, "perplexity": 37.88774014154051, "lr": 0.001305, "grad_norm": 0.229968, "tokens_per_sec": 106509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:00.391629+00:00", "epoch": 1, "step": 2395, "train_loss": 3.7258665561676025, "perplexity": 41.50718547262307, "lr": 0.001305, "grad_norm": 0.224708, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:00.698498+00:00", "epoch": 1, "step": 2396, "train_loss": 3.7599403858184814, "perplexity": 42.94586571981542, "lr": 0.001305, "grad_norm": 0.215718, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:01.005165+00:00", "epoch": 1, "step": 2397, "train_loss": 3.862147331237793, "perplexity": 47.567384711045605, "lr": 0.001305, "grad_norm": 0.218234, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:01.312517+00:00", "epoch": 1, "step": 2398, "train_loss": 3.6705033779144287, "perplexity": 39.27166937665531, "lr": 0.001305, "grad_norm": 0.274997, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:01.619845+00:00", "epoch": 1, "step": 2399, "train_loss": 3.6843512058258057, "perplexity": 39.819279550813555, "lr": 0.001305, "grad_norm": 0.264631, "tokens_per_sec": 106623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:01.927086+00:00", "epoch": 1, "step": 2400, "train_loss": 3.7259321212768555, "perplexity": 41.509906984990515, "lr": 0.001305, "grad_norm": 0.210836, "tokens_per_sec": 106711} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:15:07.707659+00:00", "step": 2400, "epoch": 1, "val_loss": 3.761365032196045, "val_ppl": 43.00709199438413, "eval_train_loss": 3.7259321212768555, "eval_train_ppl": 41.509906984990515} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:15:08.631145+00:00", "step": 2400, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p7614_epoch_0001_step_0002400.pt", "val_loss": 3.761365032196045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:09.643297+00:00", "epoch": 1, "step": 2401, "train_loss": 3.8021581172943115, "perplexity": 44.79775906477571, "lr": 0.001305, "grad_norm": 0.249642, "tokens_per_sec": 4247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:09.948505+00:00", "epoch": 1, "step": 2402, "train_loss": 3.7617578506469727, "perplexity": 43.023989292208164, "lr": 0.001305, "grad_norm": 0.257496, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:10.254271+00:00", "epoch": 1, "step": 2403, "train_loss": 3.7179627418518066, "perplexity": 41.1804134596597, "lr": 0.001305, "grad_norm": 0.224369, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:10.559718+00:00", "epoch": 1, "step": 2404, "train_loss": 3.6132898330688477, "perplexity": 37.08786521917085, "lr": 0.001305, "grad_norm": 0.257144, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:10.864981+00:00", "epoch": 1, "step": 2405, "train_loss": 3.6496009826660156, "perplexity": 38.45931705282072, "lr": 0.001305, "grad_norm": 0.255313, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:11.170970+00:00", "epoch": 1, "step": 2406, "train_loss": 3.829680919647217, "perplexity": 46.04784292788106, "lr": 0.001305, "grad_norm": 0.24407, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:11.476703+00:00", "epoch": 1, "step": 2407, "train_loss": 3.6733055114746094, "perplexity": 39.3818681631326, "lr": 0.001305, "grad_norm": 0.286386, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:11.782173+00:00", "epoch": 1, "step": 2408, "train_loss": 3.7826874256134033, "perplexity": 43.93395245559152, "lr": 0.001305, "grad_norm": 0.286511, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:12.087377+00:00", "epoch": 1, "step": 2409, "train_loss": 3.6278085708618164, "perplexity": 37.63026214114406, "lr": 0.001305, "grad_norm": 0.247186, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:12.392568+00:00", "epoch": 1, "step": 2410, "train_loss": 3.7517824172973633, "perplexity": 42.596939897252895, "lr": 0.001305, "grad_norm": 0.221409, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:12.698859+00:00", "epoch": 1, "step": 2411, "train_loss": 3.770664930343628, "perplexity": 43.40891914926067, "lr": 0.001305, "grad_norm": 0.259965, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:13.007911+00:00", "epoch": 1, "step": 2412, "train_loss": 3.707023859024048, "perplexity": 40.732400589566105, "lr": 0.001305, "grad_norm": 0.28518, "tokens_per_sec": 106028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:13.315972+00:00", "epoch": 1, "step": 2413, "train_loss": 3.671720266342163, "perplexity": 39.31948770554225, "lr": 0.001305, "grad_norm": 0.251309, "tokens_per_sec": 106371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:13.622314+00:00", "epoch": 1, "step": 2414, "train_loss": 3.5736358165740967, "perplexity": 35.64596000422794, "lr": 0.001305, "grad_norm": 0.242606, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:13.928807+00:00", "epoch": 1, "step": 2415, "train_loss": 3.5733642578125, "perplexity": 35.636281345695004, "lr": 0.001305, "grad_norm": 0.251203, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:14.234113+00:00", "epoch": 1, "step": 2416, "train_loss": 3.7211806774139404, "perplexity": 41.31314281932264, "lr": 0.001305, "grad_norm": 0.258221, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:14.540985+00:00", "epoch": 1, "step": 2417, "train_loss": 3.7485885620117188, "perplexity": 42.461108464044976, "lr": 0.001305, "grad_norm": 0.265414, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:14.847738+00:00", "epoch": 1, "step": 2418, "train_loss": 3.6532676219940186, "perplexity": 38.600592341741695, "lr": 0.001305, "grad_norm": 0.239673, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:15.155354+00:00", "epoch": 1, "step": 2419, "train_loss": 3.6802210807800293, "perplexity": 39.65516009726359, "lr": 0.001305, "grad_norm": 0.291772, "tokens_per_sec": 106521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:15.460981+00:00", "epoch": 1, "step": 2420, "train_loss": 3.687527656555176, "perplexity": 39.945964628319835, "lr": 0.001305, "grad_norm": 0.274884, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:15.766034+00:00", "epoch": 1, "step": 2421, "train_loss": 3.784491539001465, "perplexity": 44.01328582908972, "lr": 0.001305, "grad_norm": 0.242079, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:16.072296+00:00", "epoch": 1, "step": 2422, "train_loss": 3.5699830055236816, "perplexity": 35.515989570856036, "lr": 0.001305, "grad_norm": 0.264823, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:16.378879+00:00", "epoch": 1, "step": 2423, "train_loss": 3.729463815689087, "perplexity": 41.65676647030698, "lr": 0.001305, "grad_norm": 0.273676, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:16.684772+00:00", "epoch": 1, "step": 2424, "train_loss": 3.757124185562134, "perplexity": 42.82509170349545, "lr": 0.001305, "grad_norm": 0.29458, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:16.990732+00:00", "epoch": 1, "step": 2425, "train_loss": 3.6881070137023926, "perplexity": 39.96911431375005, "lr": 0.001305, "grad_norm": 0.297192, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:17.295972+00:00", "epoch": 1, "step": 2426, "train_loss": 3.6538772583007812, "perplexity": 38.62413183883275, "lr": 0.001305, "grad_norm": 0.301733, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:17.602129+00:00", "epoch": 1, "step": 2427, "train_loss": 3.664146900177002, "perplexity": 39.02283158833491, "lr": 0.001305, "grad_norm": 0.271446, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:17.908776+00:00", "epoch": 1, "step": 2428, "train_loss": 3.7384536266326904, "perplexity": 42.03294126209646, "lr": 0.001305, "grad_norm": 0.290244, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:18.215604+00:00", "epoch": 1, "step": 2429, "train_loss": 3.702552556991577, "perplexity": 40.5506802899492, "lr": 0.001305, "grad_norm": 0.302519, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:18.522209+00:00", "epoch": 1, "step": 2430, "train_loss": 3.7915334701538086, "perplexity": 44.324318206672146, "lr": 0.001305, "grad_norm": 0.239742, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:18.827876+00:00", "epoch": 1, "step": 2431, "train_loss": 3.758819580078125, "perplexity": 42.89775871133373, "lr": 0.001305, "grad_norm": 0.27927, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:19.133797+00:00", "epoch": 1, "step": 2432, "train_loss": 3.7182300090789795, "perplexity": 41.19142110550486, "lr": 0.001305, "grad_norm": 0.259583, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:19.439615+00:00", "epoch": 1, "step": 2433, "train_loss": 3.6235601902008057, "perplexity": 37.4707335721792, "lr": 0.001305, "grad_norm": 0.257029, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:19.746158+00:00", "epoch": 1, "step": 2434, "train_loss": 3.7068963050842285, "perplexity": 40.727205342736745, "lr": 0.001305, "grad_norm": 0.239143, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:20.052517+00:00", "epoch": 1, "step": 2435, "train_loss": 3.542062520980835, "perplexity": 34.53808129627408, "lr": 0.001305, "grad_norm": 0.263408, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:20.358342+00:00", "epoch": 1, "step": 2436, "train_loss": 3.6513731479644775, "perplexity": 38.52753374768071, "lr": 0.001305, "grad_norm": 0.254493, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:20.664317+00:00", "epoch": 1, "step": 2437, "train_loss": 3.4886860847473145, "perplexity": 32.74289803746844, "lr": 0.001305, "grad_norm": 0.242275, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:20.971705+00:00", "epoch": 1, "step": 2438, "train_loss": 3.7083382606506348, "perplexity": 40.78597452427619, "lr": 0.001305, "grad_norm": 0.256818, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:21.279179+00:00", "epoch": 1, "step": 2439, "train_loss": 3.65002703666687, "perplexity": 38.47570628982329, "lr": 0.001305, "grad_norm": 0.26019, "tokens_per_sec": 106572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:21.585219+00:00", "epoch": 1, "step": 2440, "train_loss": 3.7125868797302246, "perplexity": 40.95962722469422, "lr": 0.001305, "grad_norm": 0.226118, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:21.891528+00:00", "epoch": 1, "step": 2441, "train_loss": 3.7589216232299805, "perplexity": 42.902136357190756, "lr": 0.001305, "grad_norm": 0.228943, "tokens_per_sec": 106921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:22.196658+00:00", "epoch": 1, "step": 2442, "train_loss": 3.6608097553253174, "perplexity": 38.892823794874076, "lr": 0.001305, "grad_norm": 0.249456, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:22.502041+00:00", "epoch": 1, "step": 2443, "train_loss": 3.67744779586792, "perplexity": 39.54533739528796, "lr": 0.001305, "grad_norm": 0.227209, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:22.808292+00:00", "epoch": 1, "step": 2444, "train_loss": 3.737612247467041, "perplexity": 41.99759049483255, "lr": 0.001305, "grad_norm": 0.243183, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:23.114137+00:00", "epoch": 1, "step": 2445, "train_loss": 3.6968586444854736, "perplexity": 40.32044435797414, "lr": 0.001305, "grad_norm": 0.258061, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:23.421620+00:00", "epoch": 1, "step": 2446, "train_loss": 3.714876651763916, "perplexity": 41.05352289243577, "lr": 0.001305, "grad_norm": 0.264529, "tokens_per_sec": 106568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:23.728026+00:00", "epoch": 1, "step": 2447, "train_loss": 3.6925501823425293, "perplexity": 40.14709894409639, "lr": 0.001305, "grad_norm": 0.253908, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:24.033559+00:00", "epoch": 1, "step": 2448, "train_loss": 3.6790337562561035, "perplexity": 39.60810449384388, "lr": 0.001305, "grad_norm": 0.253458, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:24.339505+00:00", "epoch": 1, "step": 2449, "train_loss": 3.7148802280426025, "perplexity": 41.05366971153723, "lr": 0.001305, "grad_norm": 0.233437, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:24.646398+00:00", "epoch": 1, "step": 2450, "train_loss": 3.683384418487549, "perplexity": 39.78080137862721, "lr": 0.001305, "grad_norm": 0.247867, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:24.953633+00:00", "epoch": 1, "step": 2451, "train_loss": 3.6375715732574463, "perplexity": 37.999445718222816, "lr": 0.001305, "grad_norm": 0.218005, "tokens_per_sec": 106592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:25.259817+00:00", "epoch": 1, "step": 2452, "train_loss": 3.7482964992523193, "perplexity": 42.44870896634371, "lr": 0.001305, "grad_norm": 0.218409, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:25.566183+00:00", "epoch": 1, "step": 2453, "train_loss": 3.6760454177856445, "perplexity": 39.48991874889863, "lr": 0.001305, "grad_norm": 0.233085, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:25.872286+00:00", "epoch": 1, "step": 2454, "train_loss": 3.7797372341156006, "perplexity": 43.8045298871432, "lr": 0.001305, "grad_norm": 0.260064, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:26.179269+00:00", "epoch": 1, "step": 2455, "train_loss": 3.704648733139038, "perplexity": 40.635770809938506, "lr": 0.001305, "grad_norm": 0.255578, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:26.485728+00:00", "epoch": 1, "step": 2456, "train_loss": 3.8186450004577637, "perplexity": 45.54245648575223, "lr": 0.001305, "grad_norm": 0.258351, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:26.793090+00:00", "epoch": 1, "step": 2457, "train_loss": 3.782233953475952, "perplexity": 43.91403414880471, "lr": 0.001305, "grad_norm": 0.249403, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:27.099723+00:00", "epoch": 1, "step": 2458, "train_loss": 3.62286114692688, "perplexity": 37.44454906102598, "lr": 0.001305, "grad_norm": 0.24408, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:27.406388+00:00", "epoch": 1, "step": 2459, "train_loss": 3.786388874053955, "perplexity": 44.09687305047573, "lr": 0.001305, "grad_norm": 0.242209, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:27.712161+00:00", "epoch": 1, "step": 2460, "train_loss": 3.7637152671813965, "perplexity": 43.10828763679676, "lr": 0.001305, "grad_norm": 0.250324, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:28.018658+00:00", "epoch": 1, "step": 2461, "train_loss": 3.7221271991729736, "perplexity": 41.35226512006882, "lr": 0.001305, "grad_norm": 0.232055, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:28.325128+00:00", "epoch": 1, "step": 2462, "train_loss": 3.6827757358551025, "perplexity": 39.756594863513286, "lr": 0.001305, "grad_norm": 0.216293, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:28.632114+00:00", "epoch": 1, "step": 2463, "train_loss": 3.658989667892456, "perplexity": 38.82209983647452, "lr": 0.001305, "grad_norm": 0.208046, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:28.937569+00:00", "epoch": 1, "step": 2464, "train_loss": 3.6619932651519775, "perplexity": 38.938881083272555, "lr": 0.001305, "grad_norm": 0.227276, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:29.243900+00:00", "epoch": 1, "step": 2465, "train_loss": 3.680718421936035, "perplexity": 39.67488714555785, "lr": 0.001305, "grad_norm": 0.213791, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:29.550607+00:00", "epoch": 1, "step": 2466, "train_loss": 3.6837196350097656, "perplexity": 39.79413879585277, "lr": 0.001305, "grad_norm": 0.217562, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:29.857109+00:00", "epoch": 1, "step": 2467, "train_loss": 3.6635613441467285, "perplexity": 38.99998822266798, "lr": 0.001305, "grad_norm": 0.208736, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:30.163297+00:00", "epoch": 1, "step": 2468, "train_loss": 3.672758102416992, "perplexity": 39.36031607123756, "lr": 0.001305, "grad_norm": 0.244169, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:30.469966+00:00", "epoch": 1, "step": 2469, "train_loss": 3.702924966812134, "perplexity": 40.56578457383669, "lr": 0.001305, "grad_norm": 0.273342, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:30.776449+00:00", "epoch": 1, "step": 2470, "train_loss": 3.749392032623291, "perplexity": 42.49523842620531, "lr": 0.001305, "grad_norm": 0.272764, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:31.083141+00:00", "epoch": 1, "step": 2471, "train_loss": 3.6455066204071045, "perplexity": 38.302172599264225, "lr": 0.001305, "grad_norm": 0.245277, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:31.389525+00:00", "epoch": 1, "step": 2472, "train_loss": 3.760859489440918, "perplexity": 42.98535556541732, "lr": 0.001305, "grad_norm": 0.240525, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:31.696394+00:00", "epoch": 1, "step": 2473, "train_loss": 3.7258496284484863, "perplexity": 41.506482856592946, "lr": 0.001305, "grad_norm": 0.259164, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:32.003061+00:00", "epoch": 1, "step": 2474, "train_loss": 3.619457483291626, "perplexity": 37.31731706134485, "lr": 0.001305, "grad_norm": 0.24736, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:32.310545+00:00", "epoch": 1, "step": 2475, "train_loss": 3.7354366779327393, "perplexity": 41.906321133857425, "lr": 0.001305, "grad_norm": 0.270283, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:32.617144+00:00", "epoch": 1, "step": 2476, "train_loss": 3.6551756858825684, "perplexity": 38.674315049516935, "lr": 0.001305, "grad_norm": 0.290833, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:32.922506+00:00", "epoch": 1, "step": 2477, "train_loss": 3.72367525100708, "perplexity": 41.41633014512804, "lr": 0.001305, "grad_norm": 0.297517, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:33.230261+00:00", "epoch": 1, "step": 2478, "train_loss": 3.774596929550171, "perplexity": 43.57993898951123, "lr": 0.001305, "grad_norm": 0.296939, "tokens_per_sec": 106475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:33.539366+00:00", "epoch": 1, "step": 2479, "train_loss": 3.7387197017669678, "perplexity": 42.044126670600384, "lr": 0.001305, "grad_norm": 0.300551, "tokens_per_sec": 106010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:33.854320+00:00", "epoch": 1, "step": 2480, "train_loss": 3.6248669624328613, "perplexity": 37.519731293784496, "lr": 0.001305, "grad_norm": 0.309093, "tokens_per_sec": 104094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:34.165953+00:00", "epoch": 1, "step": 2481, "train_loss": 3.6720564365386963, "perplexity": 39.33270796745633, "lr": 0.001305, "grad_norm": 0.297731, "tokens_per_sec": 105094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:34.472179+00:00", "epoch": 1, "step": 2482, "train_loss": 3.6931800842285156, "perplexity": 40.17239564382116, "lr": 0.001305, "grad_norm": 0.269976, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:34.778648+00:00", "epoch": 1, "step": 2483, "train_loss": 3.6750752925872803, "perplexity": 39.45162716045954, "lr": 0.001305, "grad_norm": 0.286969, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:35.084919+00:00", "epoch": 1, "step": 2484, "train_loss": 3.7301626205444336, "perplexity": 41.685886594431516, "lr": 0.001305, "grad_norm": 0.297782, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:35.392970+00:00", "epoch": 1, "step": 2485, "train_loss": 3.7218964099884033, "perplexity": 41.34272256572315, "lr": 0.001305, "grad_norm": 0.280468, "tokens_per_sec": 106373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:35.700857+00:00", "epoch": 1, "step": 2486, "train_loss": 3.604565382003784, "perplexity": 36.76570134890315, "lr": 0.001305, "grad_norm": 0.275191, "tokens_per_sec": 106429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:36.008242+00:00", "epoch": 1, "step": 2487, "train_loss": 3.709521532058716, "perplexity": 40.83426396590499, "lr": 0.001305, "grad_norm": 0.238489, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:36.314789+00:00", "epoch": 1, "step": 2488, "train_loss": 3.7779619693756104, "perplexity": 43.72683423536347, "lr": 0.001305, "grad_norm": 0.247178, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:36.620981+00:00", "epoch": 1, "step": 2489, "train_loss": 3.7365591526031494, "perplexity": 41.953386327662564, "lr": 0.001305, "grad_norm": 0.230367, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:36.928023+00:00", "epoch": 1, "step": 2490, "train_loss": 3.741591453552246, "perplexity": 42.16504050058311, "lr": 0.001305, "grad_norm": 0.231104, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:37.234973+00:00", "epoch": 1, "step": 2491, "train_loss": 3.668365716934204, "perplexity": 39.187809525300366, "lr": 0.001305, "grad_norm": 0.222374, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:37.541620+00:00", "epoch": 1, "step": 2492, "train_loss": 3.7562716007232666, "perplexity": 42.78859523996247, "lr": 0.001305, "grad_norm": 0.235503, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:37.848861+00:00", "epoch": 1, "step": 2493, "train_loss": 3.6403589248657227, "perplexity": 38.10551128670404, "lr": 0.001305, "grad_norm": 0.227767, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:38.155040+00:00", "epoch": 1, "step": 2494, "train_loss": 3.6978447437286377, "perplexity": 40.360223927717946, "lr": 0.001305, "grad_norm": 0.240091, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:38.461646+00:00", "epoch": 1, "step": 2495, "train_loss": 3.699289560317993, "perplexity": 40.41857919497656, "lr": 0.001305, "grad_norm": 0.251285, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:38.768482+00:00", "epoch": 1, "step": 2496, "train_loss": 3.6984705924987793, "perplexity": 40.385491230156084, "lr": 0.001305, "grad_norm": 0.207062, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:39.075797+00:00", "epoch": 1, "step": 2497, "train_loss": 3.706688642501831, "perplexity": 40.718748704195534, "lr": 0.001305, "grad_norm": 0.222084, "tokens_per_sec": 106627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:39.382850+00:00", "epoch": 1, "step": 2498, "train_loss": 3.7025556564331055, "perplexity": 40.55080597460646, "lr": 0.001305, "grad_norm": 0.235356, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:39.689679+00:00", "epoch": 1, "step": 2499, "train_loss": 3.681668758392334, "perplexity": 39.712609558865054, "lr": 0.001305, "grad_norm": 0.243792, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:39.996944+00:00", "epoch": 1, "step": 2500, "train_loss": 3.725203275680542, "perplexity": 41.47966369476438, "lr": 0.001305, "grad_norm": 0.241157, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:15:45.809829+00:00", "step": 2500, "epoch": 1, "val_loss": 3.7530035614967345, "val_ppl": 42.64898867638385, "eval_train_loss": 3.725203275680542, "eval_train_ppl": 41.47966369476438} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:15:46.744099+00:00", "step": 2500, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p7530_epoch_0001_step_0002500.pt", "val_loss": 3.7530035614967345} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:15:48.460583+00:00", "step": 2500, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0001_step_0002500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:49.521247+00:00", "epoch": 1, "step": 2501, "train_loss": 3.69669508934021, "perplexity": 40.313850281102376, "lr": 0.001305, "grad_norm": 0.231842, "tokens_per_sec": 3440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:49.827154+00:00", "epoch": 1, "step": 2502, "train_loss": 3.655797004699707, "perplexity": 38.69835159560309, "lr": 0.001305, "grad_norm": 0.239018, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:50.133719+00:00", "epoch": 1, "step": 2503, "train_loss": 3.723731756210327, "perplexity": 41.418670449399706, "lr": 0.001305, "grad_norm": 0.24721, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:50.440921+00:00", "epoch": 1, "step": 2504, "train_loss": 3.663855791091919, "perplexity": 39.011473340858615, "lr": 0.001305, "grad_norm": 0.234692, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:50.746766+00:00", "epoch": 1, "step": 2505, "train_loss": 3.7599680423736572, "perplexity": 42.947053470944745, "lr": 0.001305, "grad_norm": 0.237294, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:51.053411+00:00", "epoch": 1, "step": 2506, "train_loss": 3.6954994201660156, "perplexity": 40.265677058386025, "lr": 0.001305, "grad_norm": 0.247203, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:51.360304+00:00", "epoch": 1, "step": 2507, "train_loss": 3.654498815536499, "perplexity": 38.64814640989256, "lr": 0.0013028489010989011, "grad_norm": 0.25562, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:51.666554+00:00", "epoch": 1, "step": 2508, "train_loss": 3.669217586517334, "perplexity": 39.221206651242696, "lr": 0.001300697802197802, "grad_norm": 0.281673, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:51.974040+00:00", "epoch": 1, "step": 2509, "train_loss": 3.7564034461975098, "perplexity": 42.794237094512404, "lr": 0.0012985467032967032, "grad_norm": 0.30475, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:52.280114+00:00", "epoch": 1, "step": 2510, "train_loss": 3.8002336025238037, "perplexity": 44.71162802258538, "lr": 0.0012963956043956044, "grad_norm": 0.259172, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:52.586329+00:00", "epoch": 1, "step": 2511, "train_loss": 3.684605598449707, "perplexity": 39.82941057039404, "lr": 0.0012942445054945054, "grad_norm": 0.260789, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:52.892326+00:00", "epoch": 1, "step": 2512, "train_loss": 3.606159210205078, "perplexity": 36.824346283113805, "lr": 0.0012920934065934065, "grad_norm": 0.255686, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:53.198651+00:00", "epoch": 1, "step": 2513, "train_loss": 3.569984197616577, "perplexity": 35.51603190924011, "lr": 0.0012899423076923077, "grad_norm": 0.265404, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:53.506085+00:00", "epoch": 1, "step": 2514, "train_loss": 3.7325618267059326, "perplexity": 41.78601970234907, "lr": 0.0012877912087912087, "grad_norm": 0.277349, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:53.812922+00:00", "epoch": 1, "step": 2515, "train_loss": 3.6439049243927, "perplexity": 38.24087326662228, "lr": 0.0012856401098901098, "grad_norm": 0.239151, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:54.119561+00:00", "epoch": 1, "step": 2516, "train_loss": 3.6459174156188965, "perplexity": 38.31791018060945, "lr": 0.001283489010989011, "grad_norm": 0.258862, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:54.425726+00:00", "epoch": 1, "step": 2517, "train_loss": 3.632798910140991, "perplexity": 37.81851925914989, "lr": 0.001281337912087912, "grad_norm": 0.246234, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:54.732020+00:00", "epoch": 1, "step": 2518, "train_loss": 3.594961166381836, "perplexity": 36.414285861948535, "lr": 0.001279186813186813, "grad_norm": 0.287782, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:55.039251+00:00", "epoch": 1, "step": 2519, "train_loss": 3.7401812076568604, "perplexity": 42.105619334364306, "lr": 0.0012770357142857143, "grad_norm": 0.303766, "tokens_per_sec": 106657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:55.346613+00:00", "epoch": 1, "step": 2520, "train_loss": 3.7339324951171875, "perplexity": 41.84333374989126, "lr": 0.0012748846153846154, "grad_norm": 0.296822, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:55.652505+00:00", "epoch": 1, "step": 2521, "train_loss": 3.7693443298339844, "perplexity": 43.35163114411926, "lr": 0.0012727335164835164, "grad_norm": 0.270076, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:55.958781+00:00", "epoch": 1, "step": 2522, "train_loss": 3.641554355621338, "perplexity": 38.1510910251372, "lr": 0.0012705824175824176, "grad_norm": 0.247902, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:56.264751+00:00", "epoch": 1, "step": 2523, "train_loss": 3.701375961303711, "perplexity": 40.50299659209766, "lr": 0.0012684313186813187, "grad_norm": 0.265703, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:56.570862+00:00", "epoch": 1, "step": 2524, "train_loss": 3.6858832836151123, "perplexity": 39.88033254162286, "lr": 0.0012662802197802197, "grad_norm": 0.26384, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:56.877849+00:00", "epoch": 1, "step": 2525, "train_loss": 3.658352851867676, "perplexity": 38.79738517136565, "lr": 0.0012641291208791209, "grad_norm": 0.253489, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:57.184958+00:00", "epoch": 1, "step": 2526, "train_loss": 3.892390012741089, "perplexity": 49.02792397048463, "lr": 0.001261978021978022, "grad_norm": 0.247033, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:57.491102+00:00", "epoch": 1, "step": 2527, "train_loss": 3.629167079925537, "perplexity": 37.68141793326908, "lr": 0.001259826923076923, "grad_norm": 0.275565, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:57.797271+00:00", "epoch": 1, "step": 2528, "train_loss": 3.722541570663452, "perplexity": 41.36940387046083, "lr": 0.0012576758241758241, "grad_norm": 0.238341, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:58.103529+00:00", "epoch": 1, "step": 2529, "train_loss": 3.673025131225586, "perplexity": 39.37082781295041, "lr": 0.0012555247252747253, "grad_norm": 0.268714, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:58.410578+00:00", "epoch": 1, "step": 2530, "train_loss": 3.764031410217285, "perplexity": 43.121918176208744, "lr": 0.0012533736263736263, "grad_norm": 0.238133, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:58.738285+00:00", "epoch": 1, "step": 2531, "train_loss": 3.606448173522949, "perplexity": 36.834988705955055, "lr": 0.0012512225274725274, "grad_norm": 0.227004, "tokens_per_sec": 99941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:59.045056+00:00", "epoch": 1, "step": 2532, "train_loss": 3.6187310218811035, "perplexity": 37.29021731520727, "lr": 0.0012490714285714286, "grad_norm": 0.228312, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:59.352270+00:00", "epoch": 1, "step": 2533, "train_loss": 3.706789970397949, "perplexity": 40.72287485837802, "lr": 0.0012469203296703295, "grad_norm": 0.206488, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:59.658670+00:00", "epoch": 1, "step": 2534, "train_loss": 3.7389419078826904, "perplexity": 42.0534701707298, "lr": 0.0012447692307692307, "grad_norm": 0.214603, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:15:59.964878+00:00", "epoch": 1, "step": 2535, "train_loss": 3.7446887493133545, "perplexity": 42.295840560461684, "lr": 0.0012426181318681319, "grad_norm": 0.228149, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:00.272300+00:00", "epoch": 1, "step": 2536, "train_loss": 3.6380507946014404, "perplexity": 38.01766022771321, "lr": 0.0012404670329670328, "grad_norm": 0.253613, "tokens_per_sec": 106590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:00.578609+00:00", "epoch": 1, "step": 2537, "train_loss": 3.7442421913146973, "perplexity": 42.27695723111416, "lr": 0.001238315934065934, "grad_norm": 0.235211, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:00.885519+00:00", "epoch": 1, "step": 2538, "train_loss": 3.6488733291625977, "perplexity": 38.43134217527135, "lr": 0.0012361648351648352, "grad_norm": 0.240692, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:01.192529+00:00", "epoch": 1, "step": 2539, "train_loss": 3.591029644012451, "perplexity": 36.27140333947674, "lr": 0.0012340137362637361, "grad_norm": 0.264048, "tokens_per_sec": 106733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:01.499011+00:00", "epoch": 1, "step": 2540, "train_loss": 3.5721981525421143, "perplexity": 35.594749909920125, "lr": 0.0012318626373626373, "grad_norm": 0.241418, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:01.805447+00:00", "epoch": 1, "step": 2541, "train_loss": 3.6417274475097656, "perplexity": 38.157695241079935, "lr": 0.0012297115384615385, "grad_norm": 0.217742, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:02.112316+00:00", "epoch": 1, "step": 2542, "train_loss": 3.6436381340026855, "perplexity": 38.23067232994542, "lr": 0.0012275604395604394, "grad_norm": 0.238208, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:02.419616+00:00", "epoch": 1, "step": 2543, "train_loss": 3.672893524169922, "perplexity": 39.36564667516751, "lr": 0.0012254093406593406, "grad_norm": 0.23669, "tokens_per_sec": 106632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:02.726570+00:00", "epoch": 1, "step": 2544, "train_loss": 3.6599578857421875, "perplexity": 38.85970628918181, "lr": 0.0012232582417582418, "grad_norm": 0.240352, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:03.033158+00:00", "epoch": 1, "step": 2545, "train_loss": 3.648470878601074, "perplexity": 38.41587857190945, "lr": 0.0012211071428571427, "grad_norm": 0.242032, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:03.338143+00:00", "epoch": 1, "step": 2546, "train_loss": 3.6743247509002686, "perplexity": 39.42202817867949, "lr": 0.0012189560439560439, "grad_norm": 0.252705, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:03.645270+00:00", "epoch": 1, "step": 2547, "train_loss": 3.6921286582946777, "perplexity": 40.13017954265779, "lr": 0.001216804945054945, "grad_norm": 0.287576, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:03.952616+00:00", "epoch": 1, "step": 2548, "train_loss": 3.789670944213867, "perplexity": 44.24183984712721, "lr": 0.0012146538461538462, "grad_norm": 0.246361, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:04.259403+00:00", "epoch": 1, "step": 2549, "train_loss": 3.7916972637176514, "perplexity": 44.33157883932233, "lr": 0.0012125027472527472, "grad_norm": 0.269041, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:04.566145+00:00", "epoch": 1, "step": 2550, "train_loss": 3.680793285369873, "perplexity": 39.677857455029084, "lr": 0.0012103516483516483, "grad_norm": 0.287781, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:04.872571+00:00", "epoch": 1, "step": 2551, "train_loss": 3.7400968074798584, "perplexity": 42.10206576260286, "lr": 0.0012082005494505495, "grad_norm": 0.270112, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:05.179179+00:00", "epoch": 1, "step": 2552, "train_loss": 3.6513149738311768, "perplexity": 38.52529250698847, "lr": 0.0012060494505494504, "grad_norm": 0.269685, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:05.486328+00:00", "epoch": 1, "step": 2553, "train_loss": 3.6014485359191895, "perplexity": 36.65128671563289, "lr": 0.0012038983516483516, "grad_norm": 0.242354, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:05.792873+00:00", "epoch": 1, "step": 2554, "train_loss": 3.6433753967285156, "perplexity": 38.220629026740596, "lr": 0.0012017472527472528, "grad_norm": 0.247779, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:06.099640+00:00", "epoch": 1, "step": 2555, "train_loss": 3.6049511432647705, "perplexity": 36.77988686815253, "lr": 0.0011995961538461537, "grad_norm": 0.253848, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:06.405820+00:00", "epoch": 1, "step": 2556, "train_loss": 3.6320340633392334, "perplexity": 37.78960494456738, "lr": 0.001197445054945055, "grad_norm": 0.215016, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:06.711241+00:00", "epoch": 1, "step": 2557, "train_loss": 3.6889708042144775, "perplexity": 40.00365417092355, "lr": 0.001195293956043956, "grad_norm": 0.24086, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:07.017319+00:00", "epoch": 1, "step": 2558, "train_loss": 3.599529981613159, "perplexity": 36.58103664252153, "lr": 0.001193142857142857, "grad_norm": 0.223755, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:07.324378+00:00", "epoch": 1, "step": 2559, "train_loss": 3.6403136253356934, "perplexity": 38.1037851640478, "lr": 0.0011909917582417582, "grad_norm": 0.233716, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:07.631829+00:00", "epoch": 1, "step": 2560, "train_loss": 3.681597948074341, "perplexity": 39.709797595913024, "lr": 0.0011888406593406594, "grad_norm": 0.233338, "tokens_per_sec": 106635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:07.938690+00:00", "epoch": 1, "step": 2561, "train_loss": 3.7965099811553955, "perplexity": 44.54544843600551, "lr": 0.0011866895604395603, "grad_norm": 0.224101, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:08.245343+00:00", "epoch": 1, "step": 2562, "train_loss": 3.671841621398926, "perplexity": 39.32425961374638, "lr": 0.0011845384615384615, "grad_norm": 0.206526, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:08.553088+00:00", "epoch": 1, "step": 2563, "train_loss": 3.8071305751800537, "perplexity": 45.021068774046206, "lr": 0.0011823873626373627, "grad_norm": 0.20634, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:08.860577+00:00", "epoch": 1, "step": 2564, "train_loss": 3.6768696308135986, "perplexity": 39.522480271376516, "lr": 0.0011802362637362636, "grad_norm": 0.222884, "tokens_per_sec": 106567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:09.169276+00:00", "epoch": 1, "step": 2565, "train_loss": 3.736539602279663, "perplexity": 41.952566133406066, "lr": 0.0011780851648351648, "grad_norm": 0.224824, "tokens_per_sec": 106149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:09.476327+00:00", "epoch": 1, "step": 2566, "train_loss": 3.6299939155578613, "perplexity": 37.71258715642923, "lr": 0.001175934065934066, "grad_norm": 0.228839, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:09.782925+00:00", "epoch": 1, "step": 2567, "train_loss": 3.7052483558654785, "perplexity": 40.66014424832607, "lr": 0.0011737829670329669, "grad_norm": 0.238242, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:10.090089+00:00", "epoch": 1, "step": 2568, "train_loss": 3.6791059970855713, "perplexity": 39.6109659195208, "lr": 0.001171631868131868, "grad_norm": 0.238122, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:10.397614+00:00", "epoch": 1, "step": 2569, "train_loss": 3.6572940349578857, "perplexity": 38.756327583960264, "lr": 0.0011694807692307692, "grad_norm": 0.226158, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:10.705962+00:00", "epoch": 1, "step": 2570, "train_loss": 3.7075822353363037, "perplexity": 40.75515094823622, "lr": 0.0011673296703296702, "grad_norm": 0.226036, "tokens_per_sec": 106323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:11.012957+00:00", "epoch": 1, "step": 2571, "train_loss": 3.6881113052368164, "perplexity": 39.969285842948075, "lr": 0.0011651785714285713, "grad_norm": 0.215269, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:11.319192+00:00", "epoch": 1, "step": 2572, "train_loss": 3.6677932739257812, "perplexity": 39.16538315724341, "lr": 0.0011630274725274725, "grad_norm": 0.229068, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:11.625920+00:00", "epoch": 1, "step": 2573, "train_loss": 3.5874459743499756, "perplexity": 36.14165124489991, "lr": 0.0011608763736263737, "grad_norm": 0.230569, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:11.933412+00:00", "epoch": 1, "step": 2574, "train_loss": 3.6925160884857178, "perplexity": 40.145730197986644, "lr": 0.0011587252747252746, "grad_norm": 0.217931, "tokens_per_sec": 106566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:12.241239+00:00", "epoch": 1, "step": 2575, "train_loss": 3.6529414653778076, "perplexity": 38.58800455606615, "lr": 0.0011565741758241758, "grad_norm": 0.240513, "tokens_per_sec": 106449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:12.549139+00:00", "epoch": 1, "step": 2576, "train_loss": 3.6080620288848877, "perplexity": 36.894483044701005, "lr": 0.001154423076923077, "grad_norm": 0.229459, "tokens_per_sec": 106425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:12.855650+00:00", "epoch": 1, "step": 2577, "train_loss": 3.558762788772583, "perplexity": 35.119719743717255, "lr": 0.001152271978021978, "grad_norm": 0.257241, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:13.162315+00:00", "epoch": 1, "step": 2578, "train_loss": 3.6275737285614014, "perplexity": 37.62142600140791, "lr": 0.001150120879120879, "grad_norm": 0.226176, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:13.469720+00:00", "epoch": 1, "step": 2579, "train_loss": 3.707277297973633, "perplexity": 40.74272507464366, "lr": 0.0011479697802197803, "grad_norm": 0.209184, "tokens_per_sec": 106597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:13.777045+00:00", "epoch": 1, "step": 2580, "train_loss": 3.6862685680389404, "perplexity": 39.8957007729482, "lr": 0.0011458186813186812, "grad_norm": 0.217431, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:14.084406+00:00", "epoch": 1, "step": 2581, "train_loss": 3.634220838546753, "perplexity": 37.8723327363372, "lr": 0.0011436675824175824, "grad_norm": 0.229305, "tokens_per_sec": 106556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:14.391185+00:00", "epoch": 1, "step": 2582, "train_loss": 3.6992759704589844, "perplexity": 40.418029915916286, "lr": 0.0011415164835164836, "grad_norm": 0.231528, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:14.706492+00:00", "epoch": 1, "step": 2583, "train_loss": 3.7095835208892822, "perplexity": 40.83679531263208, "lr": 0.0011393653846153845, "grad_norm": 0.229217, "tokens_per_sec": 103924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:15.012775+00:00", "epoch": 1, "step": 2584, "train_loss": 3.6054017543792725, "perplexity": 36.796464028611325, "lr": 0.0011372142857142857, "grad_norm": 0.261509, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:15.321719+00:00", "epoch": 1, "step": 2585, "train_loss": 3.6514923572540283, "perplexity": 38.53212686137238, "lr": 0.0011350631868131868, "grad_norm": 0.250596, "tokens_per_sec": 106066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:15.630077+00:00", "epoch": 1, "step": 2586, "train_loss": 3.737379550933838, "perplexity": 41.9878189380693, "lr": 0.0011329120879120878, "grad_norm": 0.239138, "tokens_per_sec": 106265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:15.951763+00:00", "epoch": 1, "step": 2587, "train_loss": 3.6045432090759277, "perplexity": 36.76488615469721, "lr": 0.001130760989010989, "grad_norm": 0.229305, "tokens_per_sec": 101863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:16.258563+00:00", "epoch": 1, "step": 2588, "train_loss": 3.7314884662628174, "perplexity": 41.74119230400723, "lr": 0.0011286098901098901, "grad_norm": 0.239258, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:16.565781+00:00", "epoch": 1, "step": 2589, "train_loss": 3.6147639751434326, "perplexity": 37.14257831928036, "lr": 0.001126458791208791, "grad_norm": 0.24543, "tokens_per_sec": 106663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:16.873754+00:00", "epoch": 1, "step": 2590, "train_loss": 3.5777323246002197, "perplexity": 35.79228346848126, "lr": 0.0011243076923076922, "grad_norm": 0.221686, "tokens_per_sec": 106457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:17.192536+00:00", "epoch": 1, "step": 2591, "train_loss": 3.6841282844543457, "perplexity": 39.810403971720426, "lr": 0.0011221565934065934, "grad_norm": 0.21257, "tokens_per_sec": 102736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:17.499719+00:00", "epoch": 1, "step": 2592, "train_loss": 3.5990710258483887, "perplexity": 36.56425141700581, "lr": 0.0011200054945054946, "grad_norm": 0.221015, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:17.806538+00:00", "epoch": 1, "step": 2593, "train_loss": 3.623603105545044, "perplexity": 37.47234167611523, "lr": 0.0011178543956043955, "grad_norm": 0.226608, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:18.127848+00:00", "epoch": 1, "step": 2594, "train_loss": 3.667009115219116, "perplexity": 39.13468331938648, "lr": 0.0011157032967032967, "grad_norm": 0.21363, "tokens_per_sec": 101982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:18.433997+00:00", "epoch": 1, "step": 2595, "train_loss": 3.673654079437256, "perplexity": 39.39559781340254, "lr": 0.0011135521978021977, "grad_norm": 0.226497, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:18.740891+00:00", "epoch": 1, "step": 2596, "train_loss": 3.61793851852417, "perplexity": 37.26067639998755, "lr": 0.0011114010989010988, "grad_norm": 0.199016, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:19.056165+00:00", "epoch": 1, "step": 2597, "train_loss": 3.742126941680908, "perplexity": 42.18762542565568, "lr": 0.00110925, "grad_norm": 0.210061, "tokens_per_sec": 103936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:19.363234+00:00", "epoch": 1, "step": 2598, "train_loss": 3.6595263481140137, "perplexity": 38.84294048149687, "lr": 0.0011070989010989012, "grad_norm": 0.222862, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:19.683111+00:00", "epoch": 1, "step": 2599, "train_loss": 3.689077377319336, "perplexity": 40.00791771173958, "lr": 0.0011049478021978021, "grad_norm": 0.206594, "tokens_per_sec": 102440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:19.989968+00:00", "epoch": 1, "step": 2600, "train_loss": 3.7034120559692383, "perplexity": 40.58554854066825, "lr": 0.0011027967032967033, "grad_norm": 0.194246, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:16:25.825422+00:00", "step": 2600, "epoch": 1, "val_loss": 3.730707883834839, "val_ppl": 41.70862257610424, "eval_train_loss": 3.7034120559692383, "eval_train_ppl": 40.58554854066825} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:16:26.748248+00:00", "step": 2600, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p7307_epoch_0001_step_0002600.pt", "val_loss": 3.730707883834839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:27.801860+00:00", "epoch": 1, "step": 2601, "train_loss": 3.6062538623809814, "perplexity": 36.82783195257621, "lr": 0.0011006456043956044, "grad_norm": 0.199702, "tokens_per_sec": 4195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:28.107568+00:00", "epoch": 1, "step": 2602, "train_loss": 3.6984355449676514, "perplexity": 40.38407584319813, "lr": 0.0010984945054945054, "grad_norm": 0.219399, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:28.414177+00:00", "epoch": 1, "step": 2603, "train_loss": 3.6117753982543945, "perplexity": 37.03174057417424, "lr": 0.0010963434065934066, "grad_norm": 0.219343, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:28.720092+00:00", "epoch": 1, "step": 2604, "train_loss": 3.639601469039917, "perplexity": 38.07665897374037, "lr": 0.0010941923076923077, "grad_norm": 0.201917, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:29.025349+00:00", "epoch": 1, "step": 2605, "train_loss": 3.5779688358306885, "perplexity": 35.80074974663114, "lr": 0.0010920412087912087, "grad_norm": 0.219767, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:29.331017+00:00", "epoch": 1, "step": 2606, "train_loss": 3.693589925765991, "perplexity": 40.18886333455724, "lr": 0.0010898901098901099, "grad_norm": 0.216522, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:29.637840+00:00", "epoch": 1, "step": 2607, "train_loss": 3.744493007659912, "perplexity": 42.287562312922006, "lr": 0.001087739010989011, "grad_norm": 0.218368, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:29.944768+00:00", "epoch": 1, "step": 2608, "train_loss": 3.6466026306152344, "perplexity": 38.34417518485327, "lr": 0.001085587912087912, "grad_norm": 0.214213, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:30.250635+00:00", "epoch": 1, "step": 2609, "train_loss": 3.5650651454925537, "perplexity": 35.341755685369094, "lr": 0.0010834368131868131, "grad_norm": 0.199718, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:30.555685+00:00", "epoch": 1, "step": 2610, "train_loss": 3.7031142711639404, "perplexity": 40.573464580297355, "lr": 0.0010812857142857143, "grad_norm": 0.217693, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:30.861577+00:00", "epoch": 1, "step": 2611, "train_loss": 3.67598032951355, "perplexity": 39.48734850196948, "lr": 0.0010791346153846153, "grad_norm": 0.212039, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:31.167534+00:00", "epoch": 1, "step": 2612, "train_loss": 3.6680266857147217, "perplexity": 39.174525886359554, "lr": 0.0010769835164835164, "grad_norm": 0.243067, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:31.473589+00:00", "epoch": 1, "step": 2613, "train_loss": 3.625840425491333, "perplexity": 37.55627314935034, "lr": 0.0010748324175824176, "grad_norm": 0.232167, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:31.779811+00:00", "epoch": 1, "step": 2614, "train_loss": 3.658851385116577, "perplexity": 38.81673177990712, "lr": 0.0010726813186813186, "grad_norm": 0.227961, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:32.085956+00:00", "epoch": 1, "step": 2615, "train_loss": 3.8425724506378174, "perplexity": 46.64531299949028, "lr": 0.0010705302197802197, "grad_norm": 0.239854, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:32.391503+00:00", "epoch": 1, "step": 2616, "train_loss": 3.660212278366089, "perplexity": 38.86959316935012, "lr": 0.001068379120879121, "grad_norm": 0.248261, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:32.697835+00:00", "epoch": 1, "step": 2617, "train_loss": 3.6628623008728027, "perplexity": 38.97273506989411, "lr": 0.0010662280219780218, "grad_norm": 0.230839, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:33.003573+00:00", "epoch": 1, "step": 2618, "train_loss": 3.6361398696899414, "perplexity": 37.945080702804425, "lr": 0.001064076923076923, "grad_norm": 0.207576, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:33.311428+00:00", "epoch": 1, "step": 2619, "train_loss": 3.56196928024292, "perplexity": 35.23251156187764, "lr": 0.0010619258241758242, "grad_norm": 0.221848, "tokens_per_sec": 106440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:33.618014+00:00", "epoch": 1, "step": 2620, "train_loss": 3.5963664054870605, "perplexity": 36.46549261086711, "lr": 0.0010597747252747251, "grad_norm": 0.224663, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:33.924021+00:00", "epoch": 1, "step": 2621, "train_loss": 3.5881428718566895, "perplexity": 36.16684704996814, "lr": 0.0010576236263736263, "grad_norm": 0.225896, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:34.230210+00:00", "epoch": 1, "step": 2622, "train_loss": 3.7071969509124756, "perplexity": 40.73945164792725, "lr": 0.0010554725274725275, "grad_norm": 0.218264, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:34.537030+00:00", "epoch": 1, "step": 2623, "train_loss": 3.6959228515625, "perplexity": 40.28273042046291, "lr": 0.0010533214285714286, "grad_norm": 0.228504, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:34.843044+00:00", "epoch": 1, "step": 2624, "train_loss": 3.6343681812286377, "perplexity": 37.877913358533625, "lr": 0.0010511703296703296, "grad_norm": 0.232152, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:35.149187+00:00", "epoch": 1, "step": 2625, "train_loss": 3.5731475353240967, "perplexity": 35.62855899895765, "lr": 0.0010490192307692308, "grad_norm": 0.230916, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:35.455823+00:00", "epoch": 1, "step": 2626, "train_loss": 3.6731910705566406, "perplexity": 39.377361523865574, "lr": 0.0010468681318681317, "grad_norm": 0.24024, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:35.762110+00:00", "epoch": 1, "step": 2627, "train_loss": 3.67606258392334, "perplexity": 39.49059664409986, "lr": 0.0010447170329670329, "grad_norm": 0.198827, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:36.068278+00:00", "epoch": 1, "step": 2628, "train_loss": 3.687473773956299, "perplexity": 39.94381229391822, "lr": 0.001042565934065934, "grad_norm": 0.203247, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:36.374368+00:00", "epoch": 1, "step": 2629, "train_loss": 3.697274684906006, "perplexity": 40.33722278261028, "lr": 0.0010404148351648352, "grad_norm": 0.209892, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:36.681464+00:00", "epoch": 1, "step": 2630, "train_loss": 3.6329915523529053, "perplexity": 37.82580540413829, "lr": 0.0010382637362637362, "grad_norm": 0.219591, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:36.988050+00:00", "epoch": 1, "step": 2631, "train_loss": 3.605389356613159, "perplexity": 36.79600783748438, "lr": 0.0010361126373626373, "grad_norm": 0.202618, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:37.293948+00:00", "epoch": 1, "step": 2632, "train_loss": 3.543519973754883, "perplexity": 34.58845561885075, "lr": 0.0010339615384615383, "grad_norm": 0.240502, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:37.600392+00:00", "epoch": 1, "step": 2633, "train_loss": 3.7029366493225098, "perplexity": 40.566258486804124, "lr": 0.0010318104395604395, "grad_norm": 0.222492, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:37.906313+00:00", "epoch": 1, "step": 2634, "train_loss": 3.656864881515503, "perplexity": 38.73969874098131, "lr": 0.0010296593406593406, "grad_norm": 0.235556, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:38.212973+00:00", "epoch": 1, "step": 2635, "train_loss": 3.6116058826446533, "perplexity": 37.02546364812454, "lr": 0.0010275082417582418, "grad_norm": 0.254132, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:38.519185+00:00", "epoch": 1, "step": 2636, "train_loss": 3.6943206787109375, "perplexity": 40.21824219783108, "lr": 0.001025357142857143, "grad_norm": 0.274693, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:38.825116+00:00", "epoch": 1, "step": 2637, "train_loss": 3.595749616622925, "perplexity": 36.443008035930035, "lr": 0.001023206043956044, "grad_norm": 0.248788, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:39.130631+00:00", "epoch": 1, "step": 2638, "train_loss": 3.779536724090576, "perplexity": 43.79574752026507, "lr": 0.001021054945054945, "grad_norm": 0.239246, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:39.436402+00:00", "epoch": 1, "step": 2639, "train_loss": 3.596766471862793, "perplexity": 36.48008414693209, "lr": 0.001018903846153846, "grad_norm": 0.222109, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:39.743346+00:00", "epoch": 1, "step": 2640, "train_loss": 3.6383860111236572, "perplexity": 38.03040651182072, "lr": 0.0010167527472527472, "grad_norm": 0.211496, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:40.050125+00:00", "epoch": 1, "step": 2641, "train_loss": 3.633364677429199, "perplexity": 37.83992179409136, "lr": 0.0010146016483516484, "grad_norm": 0.212814, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:40.357103+00:00", "epoch": 1, "step": 2642, "train_loss": 3.640183687210083, "perplexity": 38.09883435128155, "lr": 0.0010124505494505495, "grad_norm": 0.209117, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:40.663863+00:00", "epoch": 1, "step": 2643, "train_loss": 3.647969961166382, "perplexity": 38.396640207382134, "lr": 0.0010102994505494505, "grad_norm": 0.206355, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:40.969769+00:00", "epoch": 1, "step": 2644, "train_loss": 3.619518756866455, "perplexity": 37.31960369681868, "lr": 0.0010081483516483517, "grad_norm": 0.207571, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:41.276274+00:00", "epoch": 1, "step": 2645, "train_loss": 3.6613619327545166, "perplexity": 38.9143054646322, "lr": 0.0010059972527472526, "grad_norm": 0.211638, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:41.582590+00:00", "epoch": 1, "step": 2646, "train_loss": 3.744215965270996, "perplexity": 42.275848488325295, "lr": 0.0010038461538461538, "grad_norm": 0.209229, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:41.889677+00:00", "epoch": 1, "step": 2647, "train_loss": 3.691575288772583, "perplexity": 40.10797886753771, "lr": 0.001001695054945055, "grad_norm": 0.216038, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:42.195930+00:00", "epoch": 1, "step": 2648, "train_loss": 3.5779905319213867, "perplexity": 35.80152649137084, "lr": 0.0009995439560439561, "grad_norm": 0.215913, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:42.502855+00:00", "epoch": 1, "step": 2649, "train_loss": 3.5937042236328125, "perplexity": 36.36854394288257, "lr": 0.000997392857142857, "grad_norm": 0.215138, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:42.809904+00:00", "epoch": 1, "step": 2650, "train_loss": 3.63083553314209, "perplexity": 37.74434009296743, "lr": 0.0009952417582417582, "grad_norm": 0.212023, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:43.116751+00:00", "epoch": 1, "step": 2651, "train_loss": 3.5694267749786377, "perplexity": 35.49623998578969, "lr": 0.0009930906593406592, "grad_norm": 0.218474, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:43.423079+00:00", "epoch": 1, "step": 2652, "train_loss": 3.663449287414551, "perplexity": 38.99561825627943, "lr": 0.0009909395604395604, "grad_norm": 0.20613, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:43.731342+00:00", "epoch": 1, "step": 2653, "train_loss": 3.632631778717041, "perplexity": 37.8121991243357, "lr": 0.0009887884615384615, "grad_norm": 0.199187, "tokens_per_sec": 106299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:44.037244+00:00", "epoch": 1, "step": 2654, "train_loss": 3.663247585296631, "perplexity": 38.98775355067809, "lr": 0.0009866373626373627, "grad_norm": 0.219702, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:44.343676+00:00", "epoch": 1, "step": 2655, "train_loss": 3.6851935386657715, "perplexity": 39.852834767989656, "lr": 0.0009844862637362639, "grad_norm": 0.225547, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:44.649995+00:00", "epoch": 1, "step": 2656, "train_loss": 3.658677577972412, "perplexity": 38.8099857408825, "lr": 0.0009823351648351648, "grad_norm": 0.205874, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:44.956332+00:00", "epoch": 1, "step": 2657, "train_loss": 3.672971248626709, "perplexity": 39.368706467580225, "lr": 0.0009801840659340658, "grad_norm": 0.21482, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:45.263068+00:00", "epoch": 1, "step": 2658, "train_loss": 3.576421022415161, "perplexity": 35.74537972817035, "lr": 0.000978032967032967, "grad_norm": 0.204674, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:45.570452+00:00", "epoch": 1, "step": 2659, "train_loss": 3.5080108642578125, "perplexity": 33.38180076685691, "lr": 0.0009758818681318681, "grad_norm": 0.241766, "tokens_per_sec": 106605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:45.877018+00:00", "epoch": 1, "step": 2660, "train_loss": 3.714323043823242, "perplexity": 41.03080162608693, "lr": 0.0009737307692307693, "grad_norm": 0.209365, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:46.182839+00:00", "epoch": 1, "step": 2661, "train_loss": 3.6312363147735596, "perplexity": 37.75947036293325, "lr": 0.0009715796703296702, "grad_norm": 0.224898, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:46.488826+00:00", "epoch": 1, "step": 2662, "train_loss": 3.569650173187256, "perplexity": 35.504170668032316, "lr": 0.0009694285714285714, "grad_norm": 0.221629, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:46.796034+00:00", "epoch": 1, "step": 2663, "train_loss": 3.7274374961853027, "perplexity": 41.5724420149407, "lr": 0.0009672774725274726, "grad_norm": 0.214798, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:47.102127+00:00", "epoch": 1, "step": 2664, "train_loss": 3.7415034770965576, "perplexity": 42.16133113293681, "lr": 0.0009651263736263735, "grad_norm": 0.210215, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:47.409000+00:00", "epoch": 1, "step": 2665, "train_loss": 3.6658926010131836, "perplexity": 39.091013273165814, "lr": 0.0009629752747252747, "grad_norm": 0.206594, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:47.714063+00:00", "epoch": 1, "step": 2666, "train_loss": 3.723627805709839, "perplexity": 41.41436518164817, "lr": 0.0009608241758241758, "grad_norm": 0.19627, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:48.020878+00:00", "epoch": 1, "step": 2667, "train_loss": 3.751626968383789, "perplexity": 42.59031876386161, "lr": 0.0009586730769230768, "grad_norm": 0.211464, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:48.327026+00:00", "epoch": 1, "step": 2668, "train_loss": 3.747586727142334, "perplexity": 42.41859074641517, "lr": 0.000956521978021978, "grad_norm": 0.190144, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:48.634208+00:00", "epoch": 1, "step": 2669, "train_loss": 3.645112991333008, "perplexity": 38.28709871748179, "lr": 0.0009543708791208791, "grad_norm": 0.208437, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:48.941179+00:00", "epoch": 1, "step": 2670, "train_loss": 3.632519006729126, "perplexity": 37.80793520790269, "lr": 0.0009522197802197802, "grad_norm": 0.191019, "tokens_per_sec": 106807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:49.246934+00:00", "epoch": 1, "step": 2671, "train_loss": 3.6874897480010986, "perplexity": 39.94445036326154, "lr": 0.0009500686813186813, "grad_norm": 0.211306, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:49.553314+00:00", "epoch": 1, "step": 2672, "train_loss": 3.7446370124816895, "perplexity": 42.29365236428413, "lr": 0.0009479175824175824, "grad_norm": 0.205987, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:49.860201+00:00", "epoch": 1, "step": 2673, "train_loss": 3.5735230445861816, "perplexity": 35.64194036511273, "lr": 0.0009457664835164834, "grad_norm": 0.207333, "tokens_per_sec": 106774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:50.168434+00:00", "epoch": 1, "step": 2674, "train_loss": 3.538294792175293, "perplexity": 34.408196012400566, "lr": 0.0009436153846153845, "grad_norm": 0.194986, "tokens_per_sec": 106310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:50.475104+00:00", "epoch": 1, "step": 2675, "train_loss": 3.667367458343506, "perplexity": 39.14870947701764, "lr": 0.0009414642857142857, "grad_norm": 0.214021, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:50.781632+00:00", "epoch": 1, "step": 2676, "train_loss": 3.6127865314483643, "perplexity": 37.069203533128125, "lr": 0.0009393131868131868, "grad_norm": 0.208416, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:51.087835+00:00", "epoch": 1, "step": 2677, "train_loss": 3.6620798110961914, "perplexity": 38.942251231336755, "lr": 0.0009371620879120878, "grad_norm": 0.208594, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:51.393692+00:00", "epoch": 1, "step": 2678, "train_loss": 3.639875650405884, "perplexity": 38.0871003154545, "lr": 0.000935010989010989, "grad_norm": 0.21482, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:51.699868+00:00", "epoch": 1, "step": 2679, "train_loss": 3.6930620670318604, "perplexity": 40.16765489005509, "lr": 0.0009328598901098902, "grad_norm": 0.18963, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:52.007023+00:00", "epoch": 1, "step": 2680, "train_loss": 3.6956846714019775, "perplexity": 40.273137015789764, "lr": 0.0009307087912087911, "grad_norm": 0.237134, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:52.313946+00:00", "epoch": 1, "step": 2681, "train_loss": 3.476525068283081, "perplexity": 32.347122509307276, "lr": 0.0009285576923076923, "grad_norm": 0.204245, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:52.619951+00:00", "epoch": 1, "step": 2682, "train_loss": 3.6784844398498535, "perplexity": 39.586353086974235, "lr": 0.0009264065934065933, "grad_norm": 0.211203, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:52.927216+00:00", "epoch": 1, "step": 2683, "train_loss": 3.727956533432007, "perplexity": 41.5940252615528, "lr": 0.0009242554945054944, "grad_norm": 0.213131, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:53.233321+00:00", "epoch": 1, "step": 2684, "train_loss": 3.7526779174804688, "perplexity": 42.63510254951137, "lr": 0.0009221043956043956, "grad_norm": 0.210995, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:53.539925+00:00", "epoch": 1, "step": 2685, "train_loss": 3.609036684036255, "perplexity": 36.930459972358456, "lr": 0.0009199532967032966, "grad_norm": 0.218011, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:53.846799+00:00", "epoch": 1, "step": 2686, "train_loss": 3.618981122970581, "perplexity": 37.29954480554379, "lr": 0.0009178021978021977, "grad_norm": 0.198346, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:54.155124+00:00", "epoch": 1, "step": 2687, "train_loss": 3.68715238571167, "perplexity": 39.93097688488656, "lr": 0.0009156510989010989, "grad_norm": 0.228515, "tokens_per_sec": 106278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:54.461703+00:00", "epoch": 1, "step": 2688, "train_loss": 3.642254590988159, "perplexity": 38.17781512384338, "lr": 0.0009135, "grad_norm": 0.207987, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:54.768179+00:00", "epoch": 1, "step": 2689, "train_loss": 3.6021132469177246, "perplexity": 36.6756573278329, "lr": 0.0009113489010989011, "grad_norm": 0.217298, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:55.074087+00:00", "epoch": 1, "step": 2690, "train_loss": 3.5170047283172607, "perplexity": 33.68338632183053, "lr": 0.0009091978021978022, "grad_norm": 0.232397, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:55.380480+00:00", "epoch": 1, "step": 2691, "train_loss": 3.6628992557525635, "perplexity": 38.9741753292447, "lr": 0.0009070467032967032, "grad_norm": 0.207218, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:55.688383+00:00", "epoch": 1, "step": 2692, "train_loss": 3.6072301864624023, "perplexity": 36.863805409796804, "lr": 0.0009048956043956043, "grad_norm": 0.214036, "tokens_per_sec": 106424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:55.995304+00:00", "epoch": 1, "step": 2693, "train_loss": 3.685948610305786, "perplexity": 39.882937876868816, "lr": 0.0009027445054945054, "grad_norm": 0.225551, "tokens_per_sec": 106763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:56.301165+00:00", "epoch": 1, "step": 2694, "train_loss": 3.6683778762817383, "perplexity": 39.188286026392454, "lr": 0.0009005934065934066, "grad_norm": 0.213589, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:56.607178+00:00", "epoch": 1, "step": 2695, "train_loss": 3.6648383140563965, "perplexity": 39.049821845352284, "lr": 0.0008984423076923077, "grad_norm": 0.206315, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:56.913641+00:00", "epoch": 1, "step": 2696, "train_loss": 3.639728307723999, "perplexity": 38.08148887336142, "lr": 0.0008962912087912087, "grad_norm": 0.22493, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:57.221096+00:00", "epoch": 1, "step": 2697, "train_loss": 3.6113381385803223, "perplexity": 37.01555162700527, "lr": 0.0008941401098901099, "grad_norm": 0.224151, "tokens_per_sec": 106576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:57.528279+00:00", "epoch": 1, "step": 2698, "train_loss": 3.579314708709717, "perplexity": 35.84896544358539, "lr": 0.0008919890109890108, "grad_norm": 0.198181, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:57.835236+00:00", "epoch": 1, "step": 2699, "train_loss": 3.614760637283325, "perplexity": 37.14245434275681, "lr": 0.000889837912087912, "grad_norm": 0.207333, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:16:58.141956+00:00", "epoch": 1, "step": 2700, "train_loss": 3.637483835220337, "perplexity": 37.99611186769914, "lr": 0.0008876868131868132, "grad_norm": 0.231909, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:17:03.917876+00:00", "step": 2700, "epoch": 1, "val_loss": 3.702568781375885, "val_ppl": 40.55133820510729, "eval_train_loss": 3.637483835220337, "eval_train_ppl": 37.99611186769914} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:17:04.869498+00:00", "step": 2700, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p7026_epoch_0001_step_0002700.pt", "val_loss": 3.702568781375885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:05.921672+00:00", "epoch": 1, "step": 2701, "train_loss": 3.6260945796966553, "perplexity": 37.56581944717196, "lr": 0.0008855357142857142, "grad_norm": 0.212775, "tokens_per_sec": 4212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:06.227047+00:00", "epoch": 1, "step": 2702, "train_loss": 3.6345677375793457, "perplexity": 37.885472890947156, "lr": 0.0008833846153846153, "grad_norm": 0.246979, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:06.533066+00:00", "epoch": 1, "step": 2703, "train_loss": 3.684537172317505, "perplexity": 39.82668529112204, "lr": 0.0008812335164835165, "grad_norm": 0.249408, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:06.838454+00:00", "epoch": 1, "step": 2704, "train_loss": 3.591592311859131, "perplexity": 36.291817834641236, "lr": 0.0008790824175824175, "grad_norm": 0.229575, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:07.143831+00:00", "epoch": 1, "step": 2705, "train_loss": 3.5989322662353516, "perplexity": 36.55917812762017, "lr": 0.0008769313186813186, "grad_norm": 0.220936, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:07.449424+00:00", "epoch": 1, "step": 2706, "train_loss": 3.6452736854553223, "perplexity": 38.29325172356887, "lr": 0.0008747802197802198, "grad_norm": 0.243526, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:07.755816+00:00", "epoch": 1, "step": 2707, "train_loss": 3.708550214767456, "perplexity": 40.794620195695636, "lr": 0.0008726291208791208, "grad_norm": 0.212197, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:08.060911+00:00", "epoch": 1, "step": 2708, "train_loss": 3.603153944015503, "perplexity": 36.713845445658144, "lr": 0.0008704780219780219, "grad_norm": 0.220151, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:08.366480+00:00", "epoch": 1, "step": 2709, "train_loss": 3.7734005451202393, "perplexity": 43.527831805373935, "lr": 0.000868326923076923, "grad_norm": 0.226354, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:08.671597+00:00", "epoch": 1, "step": 2710, "train_loss": 3.6603522300720215, "perplexity": 38.87503341590009, "lr": 0.0008661758241758241, "grad_norm": 0.232267, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:08.977670+00:00", "epoch": 1, "step": 2711, "train_loss": 3.587308645248413, "perplexity": 36.13668828519266, "lr": 0.0008640247252747252, "grad_norm": 0.211223, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:09.285056+00:00", "epoch": 1, "step": 2712, "train_loss": 3.7491955757141113, "perplexity": 42.48689076301415, "lr": 0.0008618736263736263, "grad_norm": 0.199967, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:09.590569+00:00", "epoch": 1, "step": 2713, "train_loss": 3.668529748916626, "perplexity": 39.19423810661563, "lr": 0.0008597225274725275, "grad_norm": 0.206122, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:09.896519+00:00", "epoch": 1, "step": 2714, "train_loss": 3.6964058876037598, "perplexity": 40.30219313131332, "lr": 0.0008575714285714286, "grad_norm": 0.204529, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:10.241521+00:00", "epoch": 1, "step": 2715, "train_loss": 3.691148519515991, "perplexity": 40.090865667166646, "lr": 0.0008554203296703296, "grad_norm": 0.210639, "tokens_per_sec": 94979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:10.546785+00:00", "epoch": 1, "step": 2716, "train_loss": 3.5899600982666016, "perplexity": 36.23263015288354, "lr": 0.0008532692307692307, "grad_norm": 0.216287, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:10.852123+00:00", "epoch": 1, "step": 2717, "train_loss": 3.5834507942199707, "perplexity": 35.99754689108255, "lr": 0.0008511181318681317, "grad_norm": 0.216536, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:11.157710+00:00", "epoch": 1, "step": 2718, "train_loss": 3.7003250122070312, "perplexity": 40.46045236424111, "lr": 0.0008489670329670329, "grad_norm": 0.208669, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:11.463671+00:00", "epoch": 1, "step": 2719, "train_loss": 3.7317044734954834, "perplexity": 41.75020967731886, "lr": 0.0008468159340659341, "grad_norm": 0.219089, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:11.769925+00:00", "epoch": 1, "step": 2720, "train_loss": 3.5664713382720947, "perplexity": 35.39148796542176, "lr": 0.0008446648351648351, "grad_norm": 0.241762, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:12.075360+00:00", "epoch": 1, "step": 2721, "train_loss": 3.5804905891418457, "perplexity": 35.89114433436869, "lr": 0.0008425137362637362, "grad_norm": 0.20227, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:12.380188+00:00", "epoch": 1, "step": 2722, "train_loss": 3.671872138977051, "perplexity": 39.32545971322332, "lr": 0.0008403626373626374, "grad_norm": 0.210313, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:12.685198+00:00", "epoch": 1, "step": 2723, "train_loss": 3.653010845184326, "perplexity": 38.59068187723113, "lr": 0.0008382115384615384, "grad_norm": 0.230023, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:12.990817+00:00", "epoch": 1, "step": 2724, "train_loss": 3.551957845687866, "perplexity": 34.881543357194346, "lr": 0.0008360604395604395, "grad_norm": 0.201674, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:13.296762+00:00", "epoch": 1, "step": 2725, "train_loss": 3.586164951324463, "perplexity": 36.095382599402576, "lr": 0.0008339093406593407, "grad_norm": 0.205598, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:13.602516+00:00", "epoch": 1, "step": 2726, "train_loss": 3.7480416297912598, "perplexity": 42.43789146535067, "lr": 0.0008317582417582417, "grad_norm": 0.20097, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:13.908816+00:00", "epoch": 1, "step": 2727, "train_loss": 3.689098596572876, "perplexity": 40.0087666588961, "lr": 0.0008296071428571428, "grad_norm": 0.204524, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:14.215769+00:00", "epoch": 1, "step": 2728, "train_loss": 3.6245625019073486, "perplexity": 37.5083097554697, "lr": 0.000827456043956044, "grad_norm": 0.209796, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:14.522989+00:00", "epoch": 1, "step": 2729, "train_loss": 3.60695481300354, "perplexity": 36.853655493767164, "lr": 0.000825304945054945, "grad_norm": 0.206831, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:14.829613+00:00", "epoch": 1, "step": 2730, "train_loss": 3.6696901321411133, "perplexity": 39.23974484053023, "lr": 0.0008231538461538461, "grad_norm": 0.208543, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:15.136038+00:00", "epoch": 1, "step": 2731, "train_loss": 3.680500030517578, "perplexity": 39.66622343675127, "lr": 0.0008210027472527472, "grad_norm": 0.191688, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:15.442165+00:00", "epoch": 1, "step": 2732, "train_loss": 3.724804401397705, "perplexity": 41.46312182293877, "lr": 0.0008188516483516484, "grad_norm": 0.209865, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:15.749213+00:00", "epoch": 1, "step": 2733, "train_loss": 3.673086643218994, "perplexity": 39.37324966553704, "lr": 0.0008167005494505494, "grad_norm": 0.212138, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:16.055730+00:00", "epoch": 1, "step": 2734, "train_loss": 3.627095937728882, "perplexity": 37.6034551224606, "lr": 0.0008145494505494505, "grad_norm": 0.217884, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:16.363064+00:00", "epoch": 1, "step": 2735, "train_loss": 3.7244269847869873, "perplexity": 41.44747590473101, "lr": 0.0008123983516483516, "grad_norm": 0.191725, "tokens_per_sec": 106621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:16.668906+00:00", "epoch": 1, "step": 2736, "train_loss": 3.679537057876587, "perplexity": 39.628044334476044, "lr": 0.0008102472527472528, "grad_norm": 0.213371, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:16.975115+00:00", "epoch": 1, "step": 2737, "train_loss": 3.628201961517334, "perplexity": 37.64506844677525, "lr": 0.0008080961538461538, "grad_norm": 0.216987, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:17.280775+00:00", "epoch": 1, "step": 2738, "train_loss": 3.581511974334717, "perplexity": 35.92782174544073, "lr": 0.0008059450549450549, "grad_norm": 0.208176, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:17.587139+00:00", "epoch": 1, "step": 2739, "train_loss": 3.6148316860198975, "perplexity": 37.1450933609594, "lr": 0.000803793956043956, "grad_norm": 0.202467, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:17.893420+00:00", "epoch": 1, "step": 2740, "train_loss": 3.6119368076324463, "perplexity": 37.03771832680818, "lr": 0.0008016428571428571, "grad_norm": 0.218686, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:18.200136+00:00", "epoch": 1, "step": 2741, "train_loss": 3.565984010696411, "perplexity": 35.37424491923894, "lr": 0.0007994917582417582, "grad_norm": 0.196407, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:18.506162+00:00", "epoch": 1, "step": 2742, "train_loss": 3.5302045345306396, "perplexity": 34.130947858280095, "lr": 0.0007973406593406593, "grad_norm": 0.226778, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:18.812256+00:00", "epoch": 1, "step": 2743, "train_loss": 3.680772542953491, "perplexity": 39.67703444892421, "lr": 0.0007951895604395604, "grad_norm": 0.221358, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:19.118481+00:00", "epoch": 1, "step": 2744, "train_loss": 3.5497026443481445, "perplexity": 34.80296708983815, "lr": 0.0007930384615384615, "grad_norm": 0.189642, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:19.425954+00:00", "epoch": 1, "step": 2745, "train_loss": 3.4826691150665283, "perplexity": 32.546476536426205, "lr": 0.0007908873626373626, "grad_norm": 0.198504, "tokens_per_sec": 106573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:19.733939+00:00", "epoch": 1, "step": 2746, "train_loss": 3.6676666736602783, "perplexity": 39.16042512318802, "lr": 0.0007887362637362637, "grad_norm": 0.203354, "tokens_per_sec": 106394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:20.041207+00:00", "epoch": 1, "step": 2747, "train_loss": 3.6397249698638916, "perplexity": 38.08136176289102, "lr": 0.0007865851648351649, "grad_norm": 0.19591, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:20.347258+00:00", "epoch": 1, "step": 2748, "train_loss": 3.6725690364837646, "perplexity": 39.35287507978858, "lr": 0.0007844340659340659, "grad_norm": 0.204053, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:20.654476+00:00", "epoch": 1, "step": 2749, "train_loss": 3.6652488708496094, "perplexity": 39.06585730649316, "lr": 0.0007822829670329671, "grad_norm": 0.202067, "tokens_per_sec": 106660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:20.961734+00:00", "epoch": 1, "step": 2750, "train_loss": 3.6202008724212646, "perplexity": 37.345068663037026, "lr": 0.000780131868131868, "grad_norm": 0.214659, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:17:21.937332+00:00", "step": 2750, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0001_step_0002750.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:22.986785+00:00", "epoch": 1, "step": 2751, "train_loss": 3.6164207458496094, "perplexity": 37.20416605928574, "lr": 0.0007779807692307692, "grad_norm": 0.178984, "tokens_per_sec": 16180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:23.292056+00:00", "epoch": 1, "step": 2752, "train_loss": 3.5341923236846924, "perplexity": 34.2673266261227, "lr": 0.0007758296703296703, "grad_norm": 0.195903, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:23.597497+00:00", "epoch": 1, "step": 2753, "train_loss": 3.7438249588012695, "perplexity": 42.25932158932596, "lr": 0.0007736785714285714, "grad_norm": 0.19738, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:23.903704+00:00", "epoch": 1, "step": 2754, "train_loss": 3.5901618003845215, "perplexity": 36.239939088212274, "lr": 0.0007715274725274725, "grad_norm": 0.199715, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:24.210589+00:00", "epoch": 1, "step": 2755, "train_loss": 3.714491844177246, "perplexity": 41.03772822451616, "lr": 0.0007693763736263735, "grad_norm": 0.212868, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:24.516606+00:00", "epoch": 1, "step": 2756, "train_loss": 3.73483943939209, "perplexity": 41.88130053615581, "lr": 0.0007672252747252747, "grad_norm": 0.217825, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:24.822448+00:00", "epoch": 1, "step": 2757, "train_loss": 3.6334686279296875, "perplexity": 37.84385547735092, "lr": 0.0007650741758241758, "grad_norm": 0.217047, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:25.128355+00:00", "epoch": 1, "step": 2758, "train_loss": 3.6172356605529785, "perplexity": 37.234496637972796, "lr": 0.0007629230769230768, "grad_norm": 0.206542, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:25.434715+00:00", "epoch": 1, "step": 2759, "train_loss": 3.62436842918396, "perplexity": 37.501031121960715, "lr": 0.000760771978021978, "grad_norm": 0.215124, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:25.740661+00:00", "epoch": 1, "step": 2760, "train_loss": 3.635465145111084, "perplexity": 37.91948685957604, "lr": 0.0007586208791208791, "grad_norm": 0.20944, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:26.047603+00:00", "epoch": 1, "step": 2761, "train_loss": 3.646144151687622, "perplexity": 38.32659921794745, "lr": 0.0007564697802197801, "grad_norm": 0.214548, "tokens_per_sec": 106696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:26.353275+00:00", "epoch": 1, "step": 2762, "train_loss": 3.5894033908843994, "perplexity": 36.21246479382376, "lr": 0.0007543186813186813, "grad_norm": 0.189355, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:26.659185+00:00", "epoch": 1, "step": 2763, "train_loss": 3.563824415206909, "perplexity": 35.29793329025054, "lr": 0.0007521675824175824, "grad_norm": 0.207457, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:26.965792+00:00", "epoch": 1, "step": 2764, "train_loss": 3.5735654830932617, "perplexity": 35.64345298794776, "lr": 0.0007500164835164835, "grad_norm": 0.190201, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:27.319040+00:00", "epoch": 1, "step": 2765, "train_loss": 3.6233553886413574, "perplexity": 37.46306029328651, "lr": 0.0007478653846153846, "grad_norm": 0.192723, "tokens_per_sec": 92761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:27.625054+00:00", "epoch": 1, "step": 2766, "train_loss": 3.553934335708618, "perplexity": 34.950554557056094, "lr": 0.0007457142857142857, "grad_norm": 0.195848, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:27.932224+00:00", "epoch": 1, "step": 2767, "train_loss": 3.7121479511260986, "perplexity": 40.941652817720225, "lr": 0.0007435631868131867, "grad_norm": 0.197405, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:28.238213+00:00", "epoch": 1, "step": 2768, "train_loss": 3.633047580718994, "perplexity": 37.82792478158316, "lr": 0.0007414120879120879, "grad_norm": 0.197696, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:28.544708+00:00", "epoch": 1, "step": 2769, "train_loss": 3.6511178016662598, "perplexity": 38.51769714048284, "lr": 0.0007392609890109889, "grad_norm": 0.209822, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:28.851018+00:00", "epoch": 1, "step": 2770, "train_loss": 3.544374704360962, "perplexity": 34.61803206863299, "lr": 0.0007371098901098901, "grad_norm": 0.188104, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:29.157159+00:00", "epoch": 1, "step": 2771, "train_loss": 3.7305867671966553, "perplexity": 41.703571273859104, "lr": 0.0007349587912087912, "grad_norm": 0.199277, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:29.463919+00:00", "epoch": 1, "step": 2772, "train_loss": 3.6491353511810303, "perplexity": 38.44141335249666, "lr": 0.0007328076923076923, "grad_norm": 0.168464, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:29.771046+00:00", "epoch": 1, "step": 2773, "train_loss": 3.5864310264587402, "perplexity": 36.10498796099177, "lr": 0.0007306565934065934, "grad_norm": 0.196356, "tokens_per_sec": 106693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:30.077433+00:00", "epoch": 1, "step": 2774, "train_loss": 3.6510753631591797, "perplexity": 38.51606254160525, "lr": 0.0007285054945054944, "grad_norm": 0.180984, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:30.384235+00:00", "epoch": 1, "step": 2775, "train_loss": 3.688718795776367, "perplexity": 39.99357418269162, "lr": 0.0007263543956043955, "grad_norm": 0.191278, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:30.689851+00:00", "epoch": 1, "step": 2776, "train_loss": 3.678562879562378, "perplexity": 39.58945835091668, "lr": 0.0007242032967032967, "grad_norm": 0.186832, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:30.996577+00:00", "epoch": 1, "step": 2777, "train_loss": 3.5701491832733154, "perplexity": 35.521892028493916, "lr": 0.0007220521978021977, "grad_norm": 0.197173, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:31.303564+00:00", "epoch": 1, "step": 2778, "train_loss": 3.6008262634277344, "perplexity": 36.628486722773225, "lr": 0.0007199010989010989, "grad_norm": 0.188513, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:31.610911+00:00", "epoch": 1, "step": 2779, "train_loss": 3.6499481201171875, "perplexity": 38.47267003964306, "lr": 0.00071775, "grad_norm": 0.199204, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:31.917418+00:00", "epoch": 1, "step": 2780, "train_loss": 3.631385564804077, "perplexity": 37.76510638561506, "lr": 0.000715598901098901, "grad_norm": 0.196994, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:32.223799+00:00", "epoch": 1, "step": 2781, "train_loss": 3.4741344451904297, "perplexity": 32.269885090818654, "lr": 0.0007134478021978022, "grad_norm": 0.189926, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:32.530018+00:00", "epoch": 1, "step": 2782, "train_loss": 3.6497626304626465, "perplexity": 38.465534419180486, "lr": 0.0007112967032967033, "grad_norm": 0.18411, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:32.836092+00:00", "epoch": 1, "step": 2783, "train_loss": 3.6669650077819824, "perplexity": 39.132957226869266, "lr": 0.0007091456043956044, "grad_norm": 0.195439, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:33.143448+00:00", "epoch": 1, "step": 2784, "train_loss": 3.555386781692505, "perplexity": 35.00135523334548, "lr": 0.0007069945054945055, "grad_norm": 0.179701, "tokens_per_sec": 106612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:33.451573+00:00", "epoch": 1, "step": 2785, "train_loss": 3.6135199069976807, "perplexity": 37.09639915171391, "lr": 0.0007048434065934066, "grad_norm": 0.184762, "tokens_per_sec": 106347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:33.758464+00:00", "epoch": 1, "step": 2786, "train_loss": 3.5876071453094482, "perplexity": 36.14747669894238, "lr": 0.0007026923076923077, "grad_norm": 0.187515, "tokens_per_sec": 106774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:34.066534+00:00", "epoch": 1, "step": 2787, "train_loss": 3.562370538711548, "perplexity": 35.24665174225671, "lr": 0.0007005412087912088, "grad_norm": 0.189807, "tokens_per_sec": 106366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:34.372345+00:00", "epoch": 1, "step": 2788, "train_loss": 3.732758045196533, "perplexity": 41.7942196965322, "lr": 0.0006983901098901098, "grad_norm": 0.192161, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:34.679391+00:00", "epoch": 1, "step": 2789, "train_loss": 3.5400586128234863, "perplexity": 34.46893945351411, "lr": 0.0006962390109890109, "grad_norm": 0.194361, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:34.986518+00:00", "epoch": 1, "step": 2790, "train_loss": 3.7579379081726074, "perplexity": 42.859953630955864, "lr": 0.0006940879120879121, "grad_norm": 0.188825, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:35.294321+00:00", "epoch": 1, "step": 2791, "train_loss": 3.6847362518310547, "perplexity": 39.83461475752706, "lr": 0.0006919368131868131, "grad_norm": 0.183463, "tokens_per_sec": 106398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:35.600484+00:00", "epoch": 1, "step": 2792, "train_loss": 3.7180445194244385, "perplexity": 41.18378123161464, "lr": 0.0006897857142857142, "grad_norm": 0.192075, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:35.906637+00:00", "epoch": 1, "step": 2793, "train_loss": 3.6467862129211426, "perplexity": 38.35121514313796, "lr": 0.0006876346153846153, "grad_norm": 0.179516, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:36.213129+00:00", "epoch": 1, "step": 2794, "train_loss": 3.654216766357422, "perplexity": 38.63724726904392, "lr": 0.0006854835164835164, "grad_norm": 0.195988, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:36.520532+00:00", "epoch": 1, "step": 2795, "train_loss": 3.6638448238372803, "perplexity": 39.011045494442804, "lr": 0.0006833324175824176, "grad_norm": 0.188414, "tokens_per_sec": 106596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:36.828165+00:00", "epoch": 1, "step": 2796, "train_loss": 3.585829257965088, "perplexity": 36.0832676527272, "lr": 0.0006811813186813186, "grad_norm": 0.186867, "tokens_per_sec": 106517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:37.135087+00:00", "epoch": 1, "step": 2797, "train_loss": 3.6472671031951904, "perplexity": 38.369662304673, "lr": 0.0006790302197802198, "grad_norm": 0.189275, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:37.441231+00:00", "epoch": 1, "step": 2798, "train_loss": 3.655207395553589, "perplexity": 38.67554141876788, "lr": 0.0006768791208791209, "grad_norm": 0.175939, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:37.748772+00:00", "epoch": 1, "step": 2799, "train_loss": 3.6079390048980713, "perplexity": 36.88994441749105, "lr": 0.0006747280219780219, "grad_norm": 0.180114, "tokens_per_sec": 106548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:38.055871+00:00", "epoch": 1, "step": 2800, "train_loss": 3.5891737937927246, "perplexity": 36.204151471618545, "lr": 0.0006725769230769231, "grad_norm": 0.197617, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:17:43.884076+00:00", "step": 2800, "epoch": 1, "val_loss": 3.6796430230140684, "val_ppl": 39.63224374813387, "eval_train_loss": 3.5891737937927246, "eval_train_ppl": 36.204151471618545} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:17:44.797832+00:00", "step": 2800, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p6796_epoch_0001_step_0002800.pt", "val_loss": 3.6796430230140684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:45.861811+00:00", "epoch": 1, "step": 2801, "train_loss": 3.664118766784668, "perplexity": 39.021733759146755, "lr": 0.0006704258241758242, "grad_norm": 0.184553, "tokens_per_sec": 4198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:46.167772+00:00", "epoch": 1, "step": 2802, "train_loss": 3.510648012161255, "perplexity": 33.46994969255333, "lr": 0.0006682747252747253, "grad_norm": 0.186099, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:46.472880+00:00", "epoch": 1, "step": 2803, "train_loss": 3.644373893737793, "perplexity": 38.25881126977193, "lr": 0.0006661236263736264, "grad_norm": 0.180173, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:46.778047+00:00", "epoch": 1, "step": 2804, "train_loss": 3.6435675621032715, "perplexity": 38.227974413982864, "lr": 0.0006639725274725275, "grad_norm": 0.177947, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:47.084017+00:00", "epoch": 1, "step": 2805, "train_loss": 3.5972273349761963, "perplexity": 36.49690034677473, "lr": 0.0006618214285714285, "grad_norm": 0.193647, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:47.390641+00:00", "epoch": 1, "step": 2806, "train_loss": 3.5577199459075928, "perplexity": 35.083114484646096, "lr": 0.0006596703296703296, "grad_norm": 0.175961, "tokens_per_sec": 106870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:47.696714+00:00", "epoch": 1, "step": 2807, "train_loss": 3.5371670722961426, "perplexity": 34.36941507687762, "lr": 0.0006575192307692307, "grad_norm": 0.190677, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:48.002574+00:00", "epoch": 1, "step": 2808, "train_loss": 3.6223092079162598, "perplexity": 37.42388765610634, "lr": 0.0006553681318681318, "grad_norm": 0.183691, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:48.308683+00:00", "epoch": 1, "step": 2809, "train_loss": 3.6554107666015625, "perplexity": 38.68340770401739, "lr": 0.0006532170329670328, "grad_norm": 0.198232, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:48.614976+00:00", "epoch": 1, "step": 2810, "train_loss": 3.574366569519043, "perplexity": 35.67201791426414, "lr": 0.000651065934065934, "grad_norm": 0.182967, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:48.922176+00:00", "epoch": 1, "step": 2811, "train_loss": 3.699491500854492, "perplexity": 40.42674216873363, "lr": 0.0006489148351648351, "grad_norm": 0.203922, "tokens_per_sec": 106611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:49.227514+00:00", "epoch": 1, "step": 2812, "train_loss": 3.6752240657806396, "perplexity": 39.45749694163764, "lr": 0.0006467637362637362, "grad_norm": 0.187377, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:49.533549+00:00", "epoch": 1, "step": 2813, "train_loss": 3.615407705307007, "perplexity": 37.16649581467808, "lr": 0.0006446126373626373, "grad_norm": 0.202635, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:49.839793+00:00", "epoch": 1, "step": 2814, "train_loss": 3.551945924758911, "perplexity": 34.88112753927261, "lr": 0.0006424615384615385, "grad_norm": 0.178445, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:50.361668+00:00", "epoch": 1, "step": 2815, "train_loss": 3.771622657775879, "perplexity": 43.45051297652967, "lr": 0.0006403104395604395, "grad_norm": 0.19486, "tokens_per_sec": 62789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:50.667620+00:00", "epoch": 1, "step": 2816, "train_loss": 3.6082584857940674, "perplexity": 36.90173193282951, "lr": 0.0006381593406593406, "grad_norm": 0.190918, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:50.974678+00:00", "epoch": 1, "step": 2817, "train_loss": 3.678372859954834, "perplexity": 39.58193629226991, "lr": 0.0006360082417582418, "grad_norm": 0.191972, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:51.281945+00:00", "epoch": 1, "step": 2818, "train_loss": 3.6146433353424072, "perplexity": 37.13809771629731, "lr": 0.0006338571428571428, "grad_norm": 0.192758, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:51.587960+00:00", "epoch": 1, "step": 2819, "train_loss": 3.6121959686279297, "perplexity": 37.04731830267612, "lr": 0.000631706043956044, "grad_norm": 0.19902, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:51.919473+00:00", "epoch": 1, "step": 2820, "train_loss": 3.578234910964966, "perplexity": 35.81027670331414, "lr": 0.000629554945054945, "grad_norm": 0.184159, "tokens_per_sec": 98897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:52.226342+00:00", "epoch": 1, "step": 2821, "train_loss": 3.6970322132110596, "perplexity": 40.327443333500476, "lr": 0.0006274038461538462, "grad_norm": 0.18089, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:52.532538+00:00", "epoch": 1, "step": 2822, "train_loss": 3.588480234146118, "perplexity": 36.17905043865616, "lr": 0.0006252527472527472, "grad_norm": 0.200272, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:52.838342+00:00", "epoch": 1, "step": 2823, "train_loss": 3.669865846633911, "perplexity": 39.24664043820288, "lr": 0.0006231016483516482, "grad_norm": 0.186431, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:53.144906+00:00", "epoch": 1, "step": 2824, "train_loss": 3.6035571098327637, "perplexity": 36.7286501973464, "lr": 0.0006209505494505494, "grad_norm": 0.20591, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:53.451027+00:00", "epoch": 1, "step": 2825, "train_loss": 3.695614814758301, "perplexity": 40.27032376787068, "lr": 0.0006187994505494505, "grad_norm": 0.185201, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:53.756758+00:00", "epoch": 1, "step": 2826, "train_loss": 3.5007834434509277, "perplexity": 33.14140620816973, "lr": 0.0006166483516483516, "grad_norm": 0.21434, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:54.063493+00:00", "epoch": 1, "step": 2827, "train_loss": 3.655040740966797, "perplexity": 38.66909649944638, "lr": 0.0006144972527472527, "grad_norm": 0.194858, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:54.368782+00:00", "epoch": 1, "step": 2828, "train_loss": 3.5493907928466797, "perplexity": 34.79211542443777, "lr": 0.0006123461538461537, "grad_norm": 0.200497, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:54.675451+00:00", "epoch": 1, "step": 2829, "train_loss": 3.5372512340545654, "perplexity": 34.372307789012616, "lr": 0.0006101950549450549, "grad_norm": 0.190956, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:54.981209+00:00", "epoch": 1, "step": 2830, "train_loss": 3.571054697036743, "perplexity": 35.55407215820504, "lr": 0.000608043956043956, "grad_norm": 0.193612, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:55.287950+00:00", "epoch": 1, "step": 2831, "train_loss": 3.574458599090576, "perplexity": 35.67530094585419, "lr": 0.0006058928571428571, "grad_norm": 0.19487, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:55.594758+00:00", "epoch": 1, "step": 2832, "train_loss": 3.574367046356201, "perplexity": 35.67203492401184, "lr": 0.0006037417582417582, "grad_norm": 0.207766, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:55.900509+00:00", "epoch": 1, "step": 2833, "train_loss": 3.564803123474121, "perplexity": 35.33249658030714, "lr": 0.0006015906593406594, "grad_norm": 0.202732, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:56.206113+00:00", "epoch": 1, "step": 2834, "train_loss": 3.6331090927124023, "perplexity": 37.830251724209674, "lr": 0.0005994395604395604, "grad_norm": 0.201855, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:56.512592+00:00", "epoch": 1, "step": 2835, "train_loss": 3.5433502197265625, "perplexity": 34.582584587505785, "lr": 0.0005972884615384615, "grad_norm": 0.200309, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:56.819159+00:00", "epoch": 1, "step": 2836, "train_loss": 3.5311076641082764, "perplexity": 34.161786450329885, "lr": 0.0005951373626373627, "grad_norm": 0.21291, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:57.125723+00:00", "epoch": 1, "step": 2837, "train_loss": 3.658719301223755, "perplexity": 38.81160505345344, "lr": 0.0005929862637362637, "grad_norm": 0.194252, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:57.431819+00:00", "epoch": 1, "step": 2838, "train_loss": 3.659074544906616, "perplexity": 38.82539508023532, "lr": 0.0005908351648351649, "grad_norm": 0.219107, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:57.737604+00:00", "epoch": 1, "step": 2839, "train_loss": 3.4393417835235596, "perplexity": 31.16643715296912, "lr": 0.0005886840659340658, "grad_norm": 0.195906, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:58.043947+00:00", "epoch": 1, "step": 2840, "train_loss": 3.67287540435791, "perplexity": 39.364933383512415, "lr": 0.0005865329670329669, "grad_norm": 0.21736, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:58.350803+00:00", "epoch": 1, "step": 2841, "train_loss": 3.582611560821533, "perplexity": 35.967349220688895, "lr": 0.0005843818681318681, "grad_norm": 0.193783, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:58.657596+00:00", "epoch": 1, "step": 2842, "train_loss": 3.6227848529815674, "perplexity": 37.44169237762288, "lr": 0.0005822307692307691, "grad_norm": 0.211389, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:58.964443+00:00", "epoch": 1, "step": 2843, "train_loss": 3.6173312664031982, "perplexity": 37.23805664385736, "lr": 0.0005800796703296703, "grad_norm": 0.192223, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:59.270578+00:00", "epoch": 1, "step": 2844, "train_loss": 3.5760374069213867, "perplexity": 35.73166987649966, "lr": 0.0005779285714285714, "grad_norm": 0.201987, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:59.576514+00:00", "epoch": 1, "step": 2845, "train_loss": 3.565422296524048, "perplexity": 35.35438028417698, "lr": 0.0005757774725274725, "grad_norm": 0.19158, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:17:59.884184+00:00", "epoch": 1, "step": 2846, "train_loss": 3.691697835922241, "perplexity": 40.11289428720566, "lr": 0.0005736263736263736, "grad_norm": 0.194905, "tokens_per_sec": 106503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:00.190864+00:00", "epoch": 1, "step": 2847, "train_loss": 3.4898009300231934, "perplexity": 32.77942165797411, "lr": 0.0005714752747252746, "grad_norm": 0.191422, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:00.496735+00:00", "epoch": 1, "step": 2848, "train_loss": 3.5590262413024902, "perplexity": 35.12897334162174, "lr": 0.0005693241758241758, "grad_norm": 0.194706, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:00.802805+00:00", "epoch": 1, "step": 2849, "train_loss": 3.614757537841797, "perplexity": 37.14233922206976, "lr": 0.0005671730769230769, "grad_norm": 0.19714, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:01.109245+00:00", "epoch": 1, "step": 2850, "train_loss": 3.526810646057129, "perplexity": 34.01530757386804, "lr": 0.000565021978021978, "grad_norm": 0.186962, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:01.414962+00:00", "epoch": 1, "step": 2851, "train_loss": 3.5214390754699707, "perplexity": 33.833081805637704, "lr": 0.0005628708791208791, "grad_norm": 0.194781, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:01.721943+00:00", "epoch": 1, "step": 2852, "train_loss": 3.6574292182922363, "perplexity": 38.761567147693114, "lr": 0.0005607197802197803, "grad_norm": 0.18726, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:02.028870+00:00", "epoch": 1, "step": 2853, "train_loss": 3.611708879470825, "perplexity": 37.02927734976405, "lr": 0.0005585686813186813, "grad_norm": 0.189251, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:02.335131+00:00", "epoch": 1, "step": 2854, "train_loss": 3.5683391094207764, "perplexity": 35.457652936829724, "lr": 0.0005564175824175824, "grad_norm": 0.170997, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:02.641596+00:00", "epoch": 1, "step": 2855, "train_loss": 3.6567344665527344, "perplexity": 38.73464683404163, "lr": 0.0005542664835164836, "grad_norm": 0.184532, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:02.947487+00:00", "epoch": 1, "step": 2856, "train_loss": 3.6714541912078857, "perplexity": 39.30902715927862, "lr": 0.0005521153846153845, "grad_norm": 0.181566, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:03.253933+00:00", "epoch": 1, "step": 2857, "train_loss": 3.606855630874634, "perplexity": 36.85000045101777, "lr": 0.0005499642857142856, "grad_norm": 0.178479, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:03.560447+00:00", "epoch": 1, "step": 2858, "train_loss": 3.6489899158477783, "perplexity": 38.43582301926091, "lr": 0.0005478131868131867, "grad_norm": 0.192421, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:03.868117+00:00", "epoch": 1, "step": 2859, "train_loss": 3.6102283000946045, "perplexity": 36.97449313160318, "lr": 0.0005456620879120878, "grad_norm": 0.1832, "tokens_per_sec": 106505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:04.174996+00:00", "epoch": 1, "step": 2860, "train_loss": 3.586477756500244, "perplexity": 36.10667518799948, "lr": 0.000543510989010989, "grad_norm": 0.18113, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:04.480960+00:00", "epoch": 1, "step": 2861, "train_loss": 3.6267354488372803, "perplexity": 37.58990193763618, "lr": 0.00054135989010989, "grad_norm": 0.203515, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:04.786988+00:00", "epoch": 1, "step": 2862, "train_loss": 3.635840654373169, "perplexity": 37.93372865190043, "lr": 0.0005392087912087912, "grad_norm": 0.177778, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:05.093982+00:00", "epoch": 1, "step": 2863, "train_loss": 3.5271289348602295, "perplexity": 34.02613598859299, "lr": 0.0005370576923076923, "grad_norm": 0.199317, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:05.401059+00:00", "epoch": 1, "step": 2864, "train_loss": 3.6734113693237305, "perplexity": 39.38603726365284, "lr": 0.0005349065934065933, "grad_norm": 0.183251, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:05.707702+00:00", "epoch": 1, "step": 2865, "train_loss": 3.5799009799957275, "perplexity": 35.86998882475835, "lr": 0.0005327554945054945, "grad_norm": 0.203078, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:06.013470+00:00", "epoch": 1, "step": 2866, "train_loss": 3.6509451866149902, "perplexity": 38.51104898001894, "lr": 0.0005306043956043955, "grad_norm": 0.190974, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:06.319672+00:00", "epoch": 1, "step": 2867, "train_loss": 3.573315382003784, "perplexity": 35.63453963618869, "lr": 0.0005284532967032967, "grad_norm": 0.187129, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:06.626707+00:00", "epoch": 1, "step": 2868, "train_loss": 3.5602810382843018, "perplexity": 35.173080738471846, "lr": 0.0005263021978021978, "grad_norm": 0.197005, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:06.933996+00:00", "epoch": 1, "step": 2869, "train_loss": 3.578748941421509, "perplexity": 35.828689008034154, "lr": 0.0005241510989010989, "grad_norm": 0.190302, "tokens_per_sec": 106636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:07.240944+00:00", "epoch": 1, "step": 2870, "train_loss": 3.8056113719940186, "perplexity": 44.95272455045183, "lr": 0.000522, "grad_norm": 0.199232, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:07.547277+00:00", "epoch": 1, "step": 2871, "train_loss": 3.622377872467041, "perplexity": 37.42645743876621, "lr": 0.0005198489010989011, "grad_norm": 0.198998, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:07.853465+00:00", "epoch": 1, "step": 2872, "train_loss": 3.6123709678649902, "perplexity": 37.05380212242943, "lr": 0.0005176978021978022, "grad_norm": 0.208254, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:08.160150+00:00", "epoch": 1, "step": 2873, "train_loss": 3.521169662475586, "perplexity": 33.82396796150833, "lr": 0.0005155467032967032, "grad_norm": 0.194497, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:08.466213+00:00", "epoch": 1, "step": 2874, "train_loss": 3.622417688369751, "perplexity": 37.42794763662096, "lr": 0.0005133956043956044, "grad_norm": 0.194412, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:08.773209+00:00", "epoch": 1, "step": 2875, "train_loss": 3.588793992996216, "perplexity": 36.19040371692205, "lr": 0.0005112445054945054, "grad_norm": 0.182279, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:09.080084+00:00", "epoch": 1, "step": 2876, "train_loss": 3.543261766433716, "perplexity": 34.57952577930669, "lr": 0.0005090934065934065, "grad_norm": 0.178226, "tokens_per_sec": 106779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:09.386810+00:00", "epoch": 1, "step": 2877, "train_loss": 3.6646242141723633, "perplexity": 39.04146217795752, "lr": 0.0005069423076923076, "grad_norm": 0.178187, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:09.693174+00:00", "epoch": 1, "step": 2878, "train_loss": 3.592444658279419, "perplexity": 36.32276422230618, "lr": 0.0005047912087912087, "grad_norm": 0.183391, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:09.999909+00:00", "epoch": 1, "step": 2879, "train_loss": 3.5995254516601562, "perplexity": 36.58087093252008, "lr": 0.0005026401098901099, "grad_norm": 0.174846, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:10.307894+00:00", "epoch": 1, "step": 2880, "train_loss": 3.5510594844818115, "perplexity": 34.850221203245475, "lr": 0.0005004890109890109, "grad_norm": 0.188482, "tokens_per_sec": 106453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:10.616276+00:00", "epoch": 1, "step": 2881, "train_loss": 3.6709959506988525, "perplexity": 39.29101829717376, "lr": 0.0004983379120879121, "grad_norm": 0.175809, "tokens_per_sec": 106199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:10.922850+00:00", "epoch": 1, "step": 2882, "train_loss": 3.590348482131958, "perplexity": 36.246705054889844, "lr": 0.0004961868131868132, "grad_norm": 0.186361, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:11.229651+00:00", "epoch": 1, "step": 2883, "train_loss": 3.5279695987701416, "perplexity": 34.0547525598864, "lr": 0.0004940357142857142, "grad_norm": 0.183137, "tokens_per_sec": 106807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:11.535539+00:00", "epoch": 1, "step": 2884, "train_loss": 3.6233623027801514, "perplexity": 37.46331931898049, "lr": 0.0004918846153846154, "grad_norm": 0.169462, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:11.842518+00:00", "epoch": 1, "step": 2885, "train_loss": 3.656856060028076, "perplexity": 38.73935700072328, "lr": 0.0004897335164835164, "grad_norm": 0.184125, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:12.148737+00:00", "epoch": 1, "step": 2886, "train_loss": 3.6539247035980225, "perplexity": 38.62596441572176, "lr": 0.0004875824175824176, "grad_norm": 0.17681, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:12.456414+00:00", "epoch": 1, "step": 2887, "train_loss": 3.503610610961914, "perplexity": 33.23523508755268, "lr": 0.0004854313186813187, "grad_norm": 0.179889, "tokens_per_sec": 106502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:12.761966+00:00", "epoch": 1, "step": 2888, "train_loss": 3.6044206619262695, "perplexity": 36.76038099874409, "lr": 0.00048328021978021984, "grad_norm": 0.17928, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:13.067433+00:00", "epoch": 1, "step": 2889, "train_loss": 3.641812562942505, "perplexity": 38.160943188045955, "lr": 0.0004811291208791209, "grad_norm": 0.189684, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:13.373634+00:00", "epoch": 1, "step": 2890, "train_loss": 3.5405209064483643, "perplexity": 34.484877908318936, "lr": 0.00047897802197802185, "grad_norm": 0.178816, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:13.680703+00:00", "epoch": 1, "step": 2891, "train_loss": 3.627394676208496, "perplexity": 37.61469039959338, "lr": 0.000476826923076923, "grad_norm": 0.178045, "tokens_per_sec": 106656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:13.987232+00:00", "epoch": 1, "step": 2892, "train_loss": 3.612009286880493, "perplexity": 37.04040289006827, "lr": 0.0004746758241758241, "grad_norm": 0.177353, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:14.293088+00:00", "epoch": 1, "step": 2893, "train_loss": 3.549910306930542, "perplexity": 34.81019511432826, "lr": 0.00047252472527472525, "grad_norm": 0.181112, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:14.599496+00:00", "epoch": 1, "step": 2894, "train_loss": 3.59232234954834, "perplexity": 36.31832190277764, "lr": 0.0004703736263736263, "grad_norm": 0.175614, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:14.907672+00:00", "epoch": 1, "step": 2895, "train_loss": 3.4914374351501465, "perplexity": 32.83310926761821, "lr": 0.00046822252747252737, "grad_norm": 0.174826, "tokens_per_sec": 106331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:15.214582+00:00", "epoch": 1, "step": 2896, "train_loss": 3.5601255893707275, "perplexity": 35.167613546229546, "lr": 0.00046607142857142854, "grad_norm": 0.178789, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:15.523377+00:00", "epoch": 1, "step": 2897, "train_loss": 3.613848924636841, "perplexity": 37.10860652949533, "lr": 0.0004639203296703296, "grad_norm": 0.176624, "tokens_per_sec": 106116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:15.830744+00:00", "epoch": 1, "step": 2898, "train_loss": 3.670793056488037, "perplexity": 39.28304718569775, "lr": 0.00046176923076923077, "grad_norm": 0.172902, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:16.137689+00:00", "epoch": 1, "step": 2899, "train_loss": 3.5601885318756104, "perplexity": 35.16982715358116, "lr": 0.00045961813186813183, "grad_norm": 0.184538, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:16.446501+00:00", "epoch": 1, "step": 2900, "train_loss": 3.5467278957366943, "perplexity": 34.699590847212875, "lr": 0.000457467032967033, "grad_norm": 0.16841, "tokens_per_sec": 106176} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:18:22.241104+00:00", "step": 2900, "epoch": 1, "val_loss": 3.6558138847351076, "val_ppl": 38.69900483066127, "eval_train_loss": 3.5467278957366943, "eval_train_ppl": 34.699590847212875} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:18:23.158233+00:00", "step": 2900, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p6558_epoch_0001_step_0002900.pt", "val_loss": 3.6558138847351076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:24.247447+00:00", "epoch": 1, "step": 2901, "train_loss": 3.589948892593384, "perplexity": 36.23222414414503, "lr": 0.00045531593406593406, "grad_norm": 0.174431, "tokens_per_sec": 4200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:24.553326+00:00", "epoch": 1, "step": 2902, "train_loss": 3.6411080360412598, "perplexity": 38.134067245517016, "lr": 0.0004531648351648351, "grad_norm": 0.172992, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:24.859916+00:00", "epoch": 1, "step": 2903, "train_loss": 3.6235532760620117, "perplexity": 37.47047449522222, "lr": 0.0004510137362637363, "grad_norm": 0.173259, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:25.165570+00:00", "epoch": 1, "step": 2904, "train_loss": 3.6151411533355713, "perplexity": 37.156590332168925, "lr": 0.00044886263736263734, "grad_norm": 0.172137, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:25.472231+00:00", "epoch": 1, "step": 2905, "train_loss": 3.5897223949432373, "perplexity": 36.224018559824664, "lr": 0.0004467115384615385, "grad_norm": 0.178137, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:25.778061+00:00", "epoch": 1, "step": 2906, "train_loss": 3.610830307006836, "perplexity": 36.99675873339531, "lr": 0.0004445604395604396, "grad_norm": 0.16642, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:26.084124+00:00", "epoch": 1, "step": 2907, "train_loss": 3.5463216304779053, "perplexity": 34.68549647217878, "lr": 0.0004424093406593405, "grad_norm": 0.189406, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:26.388905+00:00", "epoch": 1, "step": 2908, "train_loss": 3.5539016723632812, "perplexity": 34.94941297366695, "lr": 0.0004402582417582417, "grad_norm": 0.168664, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:26.694959+00:00", "epoch": 1, "step": 2909, "train_loss": 3.5764126777648926, "perplexity": 35.74508144672233, "lr": 0.00043810714285714275, "grad_norm": 0.188627, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:27.000585+00:00", "epoch": 1, "step": 2910, "train_loss": 3.4976062774658203, "perplexity": 33.036277553693104, "lr": 0.0004359560439560439, "grad_norm": 0.170204, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:27.307403+00:00", "epoch": 1, "step": 2911, "train_loss": 3.6339075565338135, "perplexity": 37.86046987401056, "lr": 0.000433804945054945, "grad_norm": 0.17386, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:27.613923+00:00", "epoch": 1, "step": 2912, "train_loss": 3.5663182735443115, "perplexity": 35.386071191519584, "lr": 0.00043165384615384604, "grad_norm": 0.170916, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:27.919913+00:00", "epoch": 1, "step": 2913, "train_loss": 3.5859084129333496, "perplexity": 36.086123935676035, "lr": 0.0004295027472527472, "grad_norm": 0.178067, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:28.225524+00:00", "epoch": 1, "step": 2914, "train_loss": 3.5246341228485107, "perplexity": 33.94135297862683, "lr": 0.00042735164835164827, "grad_norm": 0.177756, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:28.531570+00:00", "epoch": 1, "step": 2915, "train_loss": 3.5474326610565186, "perplexity": 34.724054535020876, "lr": 0.00042520054945054944, "grad_norm": 0.174475, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:28.839036+00:00", "epoch": 1, "step": 2916, "train_loss": 3.5411217212677, "perplexity": 34.50560315940479, "lr": 0.0004230494505494505, "grad_norm": 0.175877, "tokens_per_sec": 106575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:29.145057+00:00", "epoch": 1, "step": 2917, "train_loss": 3.587932586669922, "perplexity": 36.15924249737184, "lr": 0.00042089835164835167, "grad_norm": 0.175075, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:29.451187+00:00", "epoch": 1, "step": 2918, "train_loss": 3.5770139694213867, "perplexity": 35.76658112909899, "lr": 0.0004187472527472527, "grad_norm": 0.171529, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:29.758052+00:00", "epoch": 1, "step": 2919, "train_loss": 3.5977208614349365, "perplexity": 36.51491697823434, "lr": 0.0004165961538461539, "grad_norm": 0.178441, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:30.065726+00:00", "epoch": 1, "step": 2920, "train_loss": 3.587397336959839, "perplexity": 36.1398934520557, "lr": 0.00041444505494505496, "grad_norm": 0.174313, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:30.379622+00:00", "epoch": 1, "step": 2921, "train_loss": 3.6220474243164062, "perplexity": 37.41409197830505, "lr": 0.000412293956043956, "grad_norm": 0.186039, "tokens_per_sec": 104337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:30.686121+00:00", "epoch": 1, "step": 2922, "train_loss": 3.4792466163635254, "perplexity": 32.43527666195099, "lr": 0.0004101428571428572, "grad_norm": 0.181198, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:30.992597+00:00", "epoch": 1, "step": 2923, "train_loss": 3.542869806289673, "perplexity": 34.56597463932818, "lr": 0.00040799175824175824, "grad_norm": 0.17265, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:31.299603+00:00", "epoch": 1, "step": 2924, "train_loss": 3.6028401851654053, "perplexity": 36.702327958679675, "lr": 0.0004058406593406594, "grad_norm": 0.175553, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:31.606616+00:00", "epoch": 1, "step": 2925, "train_loss": 3.6682651042938232, "perplexity": 39.18386693465384, "lr": 0.00040368956043956036, "grad_norm": 0.177543, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:31.913404+00:00", "epoch": 1, "step": 2926, "train_loss": 3.5489072799682617, "perplexity": 34.77529705483945, "lr": 0.0004015384615384614, "grad_norm": 0.176536, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:32.219888+00:00", "epoch": 1, "step": 2927, "train_loss": 3.525230646133423, "perplexity": 33.96160582604698, "lr": 0.0003993873626373626, "grad_norm": 0.171405, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:32.526279+00:00", "epoch": 1, "step": 2928, "train_loss": 3.5058207511901855, "perplexity": 33.30877084985552, "lr": 0.00039723626373626365, "grad_norm": 0.166226, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:32.833693+00:00", "epoch": 1, "step": 2929, "train_loss": 3.641616106033325, "perplexity": 38.15344694346448, "lr": 0.0003950851648351648, "grad_norm": 0.17716, "tokens_per_sec": 106593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:33.141271+00:00", "epoch": 1, "step": 2930, "train_loss": 3.612907648086548, "perplexity": 37.0736935023397, "lr": 0.0003929340659340659, "grad_norm": 0.173267, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:33.448828+00:00", "epoch": 1, "step": 2931, "train_loss": 3.5019071102142334, "perplexity": 33.17866703527101, "lr": 0.00039078296703296694, "grad_norm": 0.177332, "tokens_per_sec": 106478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:33.755234+00:00", "epoch": 1, "step": 2932, "train_loss": 3.517826557159424, "perplexity": 33.71107967825238, "lr": 0.0003886318681318681, "grad_norm": 0.167764, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:34.062201+00:00", "epoch": 1, "step": 2933, "train_loss": 3.630939245223999, "perplexity": 37.748254840058536, "lr": 0.00038648076923076917, "grad_norm": 0.176319, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:34.368786+00:00", "epoch": 1, "step": 2934, "train_loss": 3.541971206665039, "perplexity": 34.53492761900174, "lr": 0.00038432967032967034, "grad_norm": 0.165179, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:34.675648+00:00", "epoch": 1, "step": 2935, "train_loss": 3.6678247451782227, "perplexity": 39.1666157602994, "lr": 0.0003821785714285714, "grad_norm": 0.187417, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:34.982281+00:00", "epoch": 1, "step": 2936, "train_loss": 3.5428929328918457, "perplexity": 34.56677404211608, "lr": 0.00038002747252747257, "grad_norm": 0.173418, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:35.289311+00:00", "epoch": 1, "step": 2937, "train_loss": 3.5314605236053467, "perplexity": 34.17384288810496, "lr": 0.0003778763736263736, "grad_norm": 0.188329, "tokens_per_sec": 106727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:35.595405+00:00", "epoch": 1, "step": 2938, "train_loss": 3.456564426422119, "perplexity": 31.707854506501928, "lr": 0.0003757252747252747, "grad_norm": 0.1976, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:35.901567+00:00", "epoch": 1, "step": 2939, "train_loss": 3.5979855060577393, "perplexity": 36.52458173347106, "lr": 0.00037357417582417585, "grad_norm": 0.173748, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:36.207550+00:00", "epoch": 1, "step": 2940, "train_loss": 3.6250951290130615, "perplexity": 37.528293019276504, "lr": 0.0003714230769230769, "grad_norm": 0.196165, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:36.515299+00:00", "epoch": 1, "step": 2941, "train_loss": 3.5138256549835205, "perplexity": 33.576474397051456, "lr": 0.0003692719780219781, "grad_norm": 0.184906, "tokens_per_sec": 106421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:36.822678+00:00", "epoch": 1, "step": 2942, "train_loss": 3.6017706394195557, "perplexity": 36.66309412487876, "lr": 0.00036712087912087903, "grad_norm": 0.179101, "tokens_per_sec": 106605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:37.130240+00:00", "epoch": 1, "step": 2943, "train_loss": 3.5385661125183105, "perplexity": 34.417532922534825, "lr": 0.0003649697802197801, "grad_norm": 0.1738, "tokens_per_sec": 106546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:37.436630+00:00", "epoch": 1, "step": 2944, "train_loss": 3.603727340698242, "perplexity": 36.734903079459016, "lr": 0.00036281868131868126, "grad_norm": 0.163184, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:37.743246+00:00", "epoch": 1, "step": 2945, "train_loss": 3.534782886505127, "perplexity": 34.28756961197285, "lr": 0.0003606675824175823, "grad_norm": 0.176649, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:38.049686+00:00", "epoch": 1, "step": 2946, "train_loss": 3.6051406860351562, "perplexity": 36.78685889053134, "lr": 0.0003585164835164835, "grad_norm": 0.171567, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:38.358125+00:00", "epoch": 1, "step": 2947, "train_loss": 3.646855592727661, "perplexity": 38.35387603532937, "lr": 0.00035636538461538455, "grad_norm": 0.161974, "tokens_per_sec": 106239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:38.665419+00:00", "epoch": 1, "step": 2948, "train_loss": 3.673877000808716, "perplexity": 39.40438091303052, "lr": 0.0003542142857142857, "grad_norm": 0.179205, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:38.982847+00:00", "epoch": 1, "step": 2949, "train_loss": 3.68540358543396, "perplexity": 39.861206606343835, "lr": 0.0003520631868131868, "grad_norm": 0.176285, "tokens_per_sec": 103230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:39.289590+00:00", "epoch": 1, "step": 2950, "train_loss": 3.5845024585723877, "perplexity": 36.03542414150837, "lr": 0.00034991208791208784, "grad_norm": 0.175003, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:39.596412+00:00", "epoch": 1, "step": 2951, "train_loss": 3.6823294162750244, "perplexity": 39.738854675979475, "lr": 0.000347760989010989, "grad_norm": 0.177612, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:39.903318+00:00", "epoch": 1, "step": 2952, "train_loss": 3.574490785598755, "perplexity": 35.67644922769935, "lr": 0.00034560989010989007, "grad_norm": 0.171782, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:40.221958+00:00", "epoch": 1, "step": 2953, "train_loss": 3.6154162883758545, "perplexity": 37.1668148186395, "lr": 0.00034345879120879124, "grad_norm": 0.168514, "tokens_per_sec": 102836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:40.529059+00:00", "epoch": 1, "step": 2954, "train_loss": 3.6617109775543213, "perplexity": 38.927890671377604, "lr": 0.0003413076923076923, "grad_norm": 0.167404, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:40.836668+00:00", "epoch": 1, "step": 2955, "train_loss": 3.5559959411621094, "perplexity": 35.02268313572219, "lr": 0.00033915659340659347, "grad_norm": 0.16619, "tokens_per_sec": 106525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:41.155900+00:00", "epoch": 1, "step": 2956, "train_loss": 3.629368543624878, "perplexity": 37.68901013587326, "lr": 0.0003370054945054945, "grad_norm": 0.17664, "tokens_per_sec": 102647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:41.462551+00:00", "epoch": 1, "step": 2957, "train_loss": 3.656621217727661, "perplexity": 38.73026042918033, "lr": 0.0003348543956043956, "grad_norm": 0.16142, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:41.769235+00:00", "epoch": 1, "step": 2958, "train_loss": 3.652588129043579, "perplexity": 38.574372420497355, "lr": 0.00033270329670329675, "grad_norm": 0.16476, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:42.092158+00:00", "epoch": 1, "step": 2959, "train_loss": 3.5485799312591553, "perplexity": 34.7639152692478, "lr": 0.0003305521978021977, "grad_norm": 0.171097, "tokens_per_sec": 101473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:42.398867+00:00", "epoch": 1, "step": 2960, "train_loss": 3.584066390991211, "perplexity": 36.01971368692522, "lr": 0.00032840109890109877, "grad_norm": 0.175599, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:42.707538+00:00", "epoch": 1, "step": 2961, "train_loss": 3.6306140422821045, "perplexity": 37.73598099238702, "lr": 0.00032624999999999993, "grad_norm": 0.168234, "tokens_per_sec": 106102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:43.015183+00:00", "epoch": 1, "step": 2962, "train_loss": 3.6549315452575684, "perplexity": 38.66487423055992, "lr": 0.0003240989010989011, "grad_norm": 0.173935, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:43.336726+00:00", "epoch": 1, "step": 2963, "train_loss": 3.54327654838562, "perplexity": 34.58003693597158, "lr": 0.00032194780219780205, "grad_norm": 0.16372, "tokens_per_sec": 101909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:43.643178+00:00", "epoch": 1, "step": 2964, "train_loss": 3.5269010066986084, "perplexity": 34.01838135775299, "lr": 0.0003197967032967032, "grad_norm": 0.177176, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:43.950654+00:00", "epoch": 1, "step": 2965, "train_loss": 3.550889253616333, "perplexity": 34.844289124853674, "lr": 0.0003176456043956044, "grad_norm": 0.170803, "tokens_per_sec": 106574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:44.273951+00:00", "epoch": 1, "step": 2966, "train_loss": 3.6083803176879883, "perplexity": 36.906228014597374, "lr": 0.00031549450549450556, "grad_norm": 0.169573, "tokens_per_sec": 101355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:44.581810+00:00", "epoch": 1, "step": 2967, "train_loss": 3.579723596572876, "perplexity": 35.863626647652026, "lr": 0.0003133434065934065, "grad_norm": 0.190642, "tokens_per_sec": 106439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:44.889486+00:00", "epoch": 1, "step": 2968, "train_loss": 3.5408849716186523, "perplexity": 34.497434936916676, "lr": 0.0003111923076923077, "grad_norm": 0.16849, "tokens_per_sec": 106501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:45.206285+00:00", "epoch": 1, "step": 2969, "train_loss": 3.491743803024292, "perplexity": 32.84316981854223, "lr": 0.00030904120879120885, "grad_norm": 0.188386, "tokens_per_sec": 103434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:45.513233+00:00", "epoch": 1, "step": 2970, "train_loss": 3.511592149734497, "perplexity": 33.501564851813995, "lr": 0.0003068901098901098, "grad_norm": 0.180897, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:45.819714+00:00", "epoch": 1, "step": 2971, "train_loss": 3.655012607574463, "perplexity": 38.66800862188627, "lr": 0.00030473901098901097, "grad_norm": 0.172891, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:46.136762+00:00", "epoch": 1, "step": 2972, "train_loss": 3.579275369644165, "perplexity": 35.847555206522735, "lr": 0.00030258791208791214, "grad_norm": 0.175098, "tokens_per_sec": 103353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:46.443520+00:00", "epoch": 1, "step": 2973, "train_loss": 3.603158473968506, "perplexity": 36.714011758029265, "lr": 0.0003004368131868133, "grad_norm": 0.183812, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:46.761625+00:00", "epoch": 1, "step": 2974, "train_loss": 3.5457568168640137, "perplexity": 34.665911163111964, "lr": 0.00029828571428571426, "grad_norm": 0.171757, "tokens_per_sec": 103010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:47.069484+00:00", "epoch": 1, "step": 2975, "train_loss": 3.639867067337036, "perplexity": 38.0867734126532, "lr": 0.0002961346153846154, "grad_norm": 0.171907, "tokens_per_sec": 106439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:47.386329+00:00", "epoch": 1, "step": 2976, "train_loss": 3.5311920642852783, "perplexity": 34.16466983283028, "lr": 0.0002939835164835164, "grad_norm": 0.173369, "tokens_per_sec": 103419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:47.692616+00:00", "epoch": 1, "step": 2977, "train_loss": 3.646315813064575, "perplexity": 38.33317897947243, "lr": 0.00029183241758241754, "grad_norm": 0.175597, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:48.012454+00:00", "epoch": 1, "step": 2978, "train_loss": 3.6994285583496094, "perplexity": 40.4241976883961, "lr": 0.0002896813186813187, "grad_norm": 0.173489, "tokens_per_sec": 102452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:48.319807+00:00", "epoch": 1, "step": 2979, "train_loss": 3.6279830932617188, "perplexity": 37.636830037907764, "lr": 0.00028753021978021966, "grad_norm": 0.159379, "tokens_per_sec": 106614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:48.640726+00:00", "epoch": 1, "step": 2980, "train_loss": 3.5940608978271484, "perplexity": 36.38151797760671, "lr": 0.00028537912087912083, "grad_norm": 0.177113, "tokens_per_sec": 102158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:48.947905+00:00", "epoch": 1, "step": 2981, "train_loss": 3.521538734436035, "perplexity": 33.83645374360837, "lr": 0.000283228021978022, "grad_norm": 0.170506, "tokens_per_sec": 106617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:49.255518+00:00", "epoch": 1, "step": 2982, "train_loss": 3.566404104232788, "perplexity": 35.389108532719106, "lr": 0.00028107692307692295, "grad_norm": 0.167685, "tokens_per_sec": 106524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:49.574503+00:00", "epoch": 1, "step": 2983, "train_loss": 3.6494336128234863, "perplexity": 38.45288066162573, "lr": 0.0002789258241758241, "grad_norm": 0.168192, "tokens_per_sec": 102726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:49.881735+00:00", "epoch": 1, "step": 2984, "train_loss": 3.6024703979492188, "perplexity": 36.68875841607259, "lr": 0.0002767747252747253, "grad_norm": 0.162059, "tokens_per_sec": 106656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:50.189177+00:00", "epoch": 1, "step": 2985, "train_loss": 3.516339063644409, "perplexity": 33.660971942551974, "lr": 0.00027462362637362646, "grad_norm": 0.169099, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:50.496129+00:00", "epoch": 1, "step": 2986, "train_loss": 3.6008853912353516, "perplexity": 36.63065254891912, "lr": 0.0002724725274725274, "grad_norm": 0.172651, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:50.815427+00:00", "epoch": 1, "step": 2987, "train_loss": 3.6782805919647217, "perplexity": 39.57828431504636, "lr": 0.0002703214285714286, "grad_norm": 0.165045, "tokens_per_sec": 102626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:51.122883+00:00", "epoch": 1, "step": 2988, "train_loss": 3.6696083545684814, "perplexity": 39.2365360406522, "lr": 0.00026817032967032975, "grad_norm": 0.174498, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:51.438159+00:00", "epoch": 1, "step": 2989, "train_loss": 3.601321220397949, "perplexity": 36.64662073499275, "lr": 0.0002660192307692307, "grad_norm": 0.164734, "tokens_per_sec": 103934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:51.746647+00:00", "epoch": 1, "step": 2990, "train_loss": 3.576899290084839, "perplexity": 35.76247967648495, "lr": 0.00026386813186813187, "grad_norm": 0.160336, "tokens_per_sec": 106280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:52.067172+00:00", "epoch": 1, "step": 2991, "train_loss": 3.5779664516448975, "perplexity": 35.80066439109404, "lr": 0.00026171703296703304, "grad_norm": 0.170428, "tokens_per_sec": 102181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:52.374281+00:00", "epoch": 1, "step": 2992, "train_loss": 3.5178797245025635, "perplexity": 33.71287205444076, "lr": 0.0002595659340659342, "grad_norm": 0.171044, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:52.681300+00:00", "epoch": 1, "step": 2993, "train_loss": 3.6463584899902344, "perplexity": 38.334814956611005, "lr": 0.00025741483516483516, "grad_norm": 0.163164, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:52.999632+00:00", "epoch": 1, "step": 2994, "train_loss": 3.633035182952881, "perplexity": 37.82745580272631, "lr": 0.0002552637362637361, "grad_norm": 0.161426, "tokens_per_sec": 102936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:53.306668+00:00", "epoch": 1, "step": 2995, "train_loss": 3.6035542488098145, "perplexity": 36.72854511598561, "lr": 0.0002531126373626373, "grad_norm": 0.170824, "tokens_per_sec": 106727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:53.623619+00:00", "epoch": 1, "step": 2996, "train_loss": 3.6104841232299805, "perplexity": 36.98395327237489, "lr": 0.00025096153846153844, "grad_norm": 0.159901, "tokens_per_sec": 103382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:53.939318+00:00", "epoch": 1, "step": 2997, "train_loss": 3.5694308280944824, "perplexity": 35.49638385645397, "lr": 0.0002488104395604394, "grad_norm": 0.164418, "tokens_per_sec": 103795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:54.246066+00:00", "epoch": 1, "step": 2998, "train_loss": 3.557258367538452, "perplexity": 35.06692461461225, "lr": 0.00024665934065934056, "grad_norm": 0.161146, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:54.565315+00:00", "epoch": 1, "step": 2999, "train_loss": 3.457716703414917, "perplexity": 31.74441179575753, "lr": 0.00024450824175824173, "grad_norm": 0.163934, "tokens_per_sec": 102642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:18:54.872556+00:00", "epoch": 1, "step": 3000, "train_loss": 3.535979747772217, "perplexity": 34.32863164386012, "lr": 0.0002423571428571429, "grad_norm": 0.162677, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T14:19:00.653462+00:00", "step": 3000, "epoch": 1, "val_loss": 3.633824050426483, "val_ppl": 37.85730842555165, "eval_train_loss": 3.535979747772217, "eval_train_ppl": 34.32863164386012} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T14:19:01.644652+00:00", "step": 3000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p6338_epoch_0001_step_0003000.pt", "val_loss": 3.633824050426483} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:19:03.613510+00:00", "step": 3000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0001_step_0003000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:04.634562+00:00", "epoch": 1, "step": 3001, "train_loss": 3.5624687671661377, "perplexity": 35.25011413643685, "lr": 0.00024020604395604385, "grad_norm": 0.163466, "tokens_per_sec": 3357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:04.938797+00:00", "epoch": 1, "step": 3002, "train_loss": 3.480405807495117, "perplexity": 32.4728971474622, "lr": 0.00023805494505494502, "grad_norm": 0.157509, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:05.244160+00:00", "epoch": 1, "step": 3003, "train_loss": 3.5661213397979736, "perplexity": 35.37910316609385, "lr": 0.0002359038461538462, "grad_norm": 0.157316, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:05.549435+00:00", "epoch": 1, "step": 3004, "train_loss": 3.5062954425811768, "perplexity": 33.3245859899809, "lr": 0.00023375274725274714, "grad_norm": 0.159985, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:05.855328+00:00", "epoch": 1, "step": 3005, "train_loss": 3.488708972930908, "perplexity": 32.74364747150667, "lr": 0.0002316016483516483, "grad_norm": 0.166071, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:06.162592+00:00", "epoch": 1, "step": 3006, "train_loss": 3.5021474361419678, "perplexity": 33.18664168742671, "lr": 0.00022945054945054948, "grad_norm": 0.165074, "tokens_per_sec": 106647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:06.470252+00:00", "epoch": 1, "step": 3007, "train_loss": 3.485398769378662, "perplexity": 32.635438528953806, "lr": 0.00022729945054945065, "grad_norm": 0.156051, "tokens_per_sec": 106507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:06.776052+00:00", "epoch": 1, "step": 3008, "train_loss": 3.530278444290161, "perplexity": 34.13347056165359, "lr": 0.0002251483516483516, "grad_norm": 0.160016, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:07.081632+00:00", "epoch": 1, "step": 3009, "train_loss": 3.561765193939209, "perplexity": 35.22532182251114, "lr": 0.00022299725274725277, "grad_norm": 0.163802, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:07.388288+00:00", "epoch": 1, "step": 3010, "train_loss": 3.5169293880462646, "perplexity": 33.68084870197067, "lr": 0.00022084615384615372, "grad_norm": 0.160877, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:07.694205+00:00", "epoch": 1, "step": 3011, "train_loss": 3.539679765701294, "perplexity": 34.455883468264496, "lr": 0.0002186950549450549, "grad_norm": 0.156744, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:08.000654+00:00", "epoch": 1, "step": 3012, "train_loss": 3.6680545806884766, "perplexity": 39.175618673972586, "lr": 0.00021654395604395606, "grad_norm": 0.168828, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:08.306490+00:00", "epoch": 1, "step": 3013, "train_loss": 3.4933876991271973, "perplexity": 32.89720497934042, "lr": 0.000214392857142857, "grad_norm": 0.155975, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:08.612093+00:00", "epoch": 1, "step": 3014, "train_loss": 3.638273000717163, "perplexity": 38.026108922962386, "lr": 0.00021224175824175818, "grad_norm": 0.156602, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:09.067677+00:00", "epoch": 1, "step": 3015, "train_loss": 3.5272669792175293, "perplexity": 34.030833428887036, "lr": 0.00021009065934065934, "grad_norm": 0.15788, "tokens_per_sec": 71925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:09.374443+00:00", "epoch": 1, "step": 3016, "train_loss": 3.6390395164489746, "perplexity": 38.0552677075773, "lr": 0.0002079395604395603, "grad_norm": 0.164769, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:09.680438+00:00", "epoch": 1, "step": 3017, "train_loss": 3.573543071746826, "perplexity": 35.64265417912591, "lr": 0.00020578846153846146, "grad_norm": 0.15397, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:09.986975+00:00", "epoch": 1, "step": 3018, "train_loss": 3.470782518386841, "perplexity": 32.16189987840281, "lr": 0.00020363736263736263, "grad_norm": 0.159379, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:10.293215+00:00", "epoch": 1, "step": 3019, "train_loss": 3.6369032859802246, "perplexity": 37.97405965564398, "lr": 0.0002014862637362638, "grad_norm": 0.164239, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:10.599195+00:00", "epoch": 1, "step": 3020, "train_loss": 3.5999326705932617, "perplexity": 36.59577038921782, "lr": 0.00019933516483516475, "grad_norm": 0.165392, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:10.904928+00:00", "epoch": 1, "step": 3021, "train_loss": 3.631678581237793, "perplexity": 37.776173803795594, "lr": 0.00019718406593406592, "grad_norm": 0.1561, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:11.212314+00:00", "epoch": 1, "step": 3022, "train_loss": 3.643906831741333, "perplexity": 38.240946205369184, "lr": 0.0001950329670329671, "grad_norm": 0.159662, "tokens_per_sec": 106603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:11.519003+00:00", "epoch": 1, "step": 3023, "train_loss": 3.5705509185791016, "perplexity": 35.53616529349536, "lr": 0.00019288186813186804, "grad_norm": 0.163727, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:11.825648+00:00", "epoch": 1, "step": 3024, "train_loss": 3.475938320159912, "perplexity": 32.32814846292243, "lr": 0.0001907307692307692, "grad_norm": 0.156654, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:12.131992+00:00", "epoch": 1, "step": 3025, "train_loss": 3.525851011276245, "perplexity": 33.982680958958746, "lr": 0.00018857967032967038, "grad_norm": 0.153332, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:12.439249+00:00", "epoch": 1, "step": 3026, "train_loss": 3.5766825675964355, "perplexity": 35.75472998269478, "lr": 0.00018642857142857155, "grad_norm": 0.159519, "tokens_per_sec": 106647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:12.747020+00:00", "epoch": 1, "step": 3027, "train_loss": 3.5064921379089355, "perplexity": 33.33114142503522, "lr": 0.0001842774725274725, "grad_norm": 0.161141, "tokens_per_sec": 106468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:13.054553+00:00", "epoch": 1, "step": 3028, "train_loss": 3.575387716293335, "perplexity": 35.70846288495502, "lr": 0.00018212637362637345, "grad_norm": 0.156557, "tokens_per_sec": 106551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:13.360735+00:00", "epoch": 1, "step": 3029, "train_loss": 3.5614283084869385, "perplexity": 35.21345692270587, "lr": 0.00017997527472527462, "grad_norm": 0.162011, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:13.666522+00:00", "epoch": 1, "step": 3030, "train_loss": 3.5633862018585205, "perplexity": 35.28246865336485, "lr": 0.00017782417582417579, "grad_norm": 0.159055, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:13.996391+00:00", "epoch": 1, "step": 3031, "train_loss": 3.588548183441162, "perplexity": 36.18150886315199, "lr": 0.00017567307692307695, "grad_norm": 0.156924, "tokens_per_sec": 99287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:14.301524+00:00", "epoch": 1, "step": 3032, "train_loss": 3.551520586013794, "perplexity": 34.86629439903504, "lr": 0.0001735219780219779, "grad_norm": 0.154916, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:14.608162+00:00", "epoch": 1, "step": 3033, "train_loss": 3.5414834022521973, "perplexity": 34.51808543709132, "lr": 0.00017137087912087907, "grad_norm": 0.14949, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:14.914971+00:00", "epoch": 1, "step": 3034, "train_loss": 3.6216626167297363, "perplexity": 37.39969752158913, "lr": 0.00016921978021978024, "grad_norm": 0.154996, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:15.221601+00:00", "epoch": 1, "step": 3035, "train_loss": 3.560178518295288, "perplexity": 35.16947497945531, "lr": 0.0001670686813186812, "grad_norm": 0.157181, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:15.527712+00:00", "epoch": 1, "step": 3036, "train_loss": 3.6389758586883545, "perplexity": 38.052845271559484, "lr": 0.00016491758241758236, "grad_norm": 0.153065, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:15.833504+00:00", "epoch": 1, "step": 3037, "train_loss": 3.624215602874756, "perplexity": 37.495300415695496, "lr": 0.00016276648351648353, "grad_norm": 0.152887, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:16.139744+00:00", "epoch": 1, "step": 3038, "train_loss": 3.4928057193756104, "perplexity": 32.87806504222679, "lr": 0.0001606153846153847, "grad_norm": 0.150841, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:16.446716+00:00", "epoch": 1, "step": 3039, "train_loss": 3.4465179443359375, "perplexity": 31.390896934416553, "lr": 0.00015846428571428565, "grad_norm": 0.154224, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:16.755258+00:00", "epoch": 1, "step": 3040, "train_loss": 3.504969358444214, "perplexity": 33.280424072809204, "lr": 0.00015631318681318682, "grad_norm": 0.157221, "tokens_per_sec": 106261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:17.062416+00:00", "epoch": 1, "step": 3041, "train_loss": 3.5845377445220947, "perplexity": 36.036695708106386, "lr": 0.000154162087912088, "grad_norm": 0.153853, "tokens_per_sec": 106624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:17.368950+00:00", "epoch": 1, "step": 3042, "train_loss": 3.704293966293335, "perplexity": 40.62135714260237, "lr": 0.00015201098901098894, "grad_norm": 0.154107, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:17.675112+00:00", "epoch": 1, "step": 3043, "train_loss": 3.538130760192871, "perplexity": 34.4025524306737, "lr": 0.0001498598901098901, "grad_norm": 0.160195, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:17.981415+00:00", "epoch": 1, "step": 3044, "train_loss": 3.5125603675842285, "perplexity": 33.53401737291735, "lr": 0.00014770879120879106, "grad_norm": 0.153931, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:18.288409+00:00", "epoch": 1, "step": 3045, "train_loss": 3.6013641357421875, "perplexity": 36.64819347108377, "lr": 0.00014555769230769223, "grad_norm": 0.151812, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:18.594752+00:00", "epoch": 1, "step": 3046, "train_loss": 3.531008720397949, "perplexity": 34.158406523641084, "lr": 0.0001434065934065934, "grad_norm": 0.158523, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:18.900271+00:00", "epoch": 1, "step": 3047, "train_loss": 3.4971764087677, "perplexity": 33.02207934397196, "lr": 0.00014125549450549435, "grad_norm": 0.150451, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:19.206112+00:00", "epoch": 1, "step": 3048, "train_loss": 3.5345230102539062, "perplexity": 34.278660244635255, "lr": 0.00013910439560439552, "grad_norm": 0.148345, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:19.513085+00:00", "epoch": 1, "step": 3049, "train_loss": 3.5830979347229004, "perplexity": 35.984847055551356, "lr": 0.00013695329670329669, "grad_norm": 0.155888, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:19.818969+00:00", "epoch": 1, "step": 3050, "train_loss": 3.520291805267334, "perplexity": 33.79428837654644, "lr": 0.00013480219780219785, "grad_norm": 0.151681, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:20.125927+00:00", "epoch": 1, "step": 3051, "train_loss": 3.6405179500579834, "perplexity": 38.11157150481351, "lr": 0.0001326510989010988, "grad_norm": 0.159146, "tokens_per_sec": 106696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T14:19:20.431537+00:00", "epoch": 1, "step": 3052, "train_loss": 3.5320591926574707, "perplexity": 34.19430783548832, "lr": 0.00013049999999999997, "grad_norm": 0.149736, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T14:19:21.425647+00:00", "step": 3052, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/last_epoch_0001_step_0003052.pt", "category": "last"} +{"stage": "pretraining", "type": "training_complete", "timestamp": "2026-04-08T14:19:21.428173+00:00", "epochs": 2, "total_steps": 3052, "total_time_hours": 0.35, "best_val_loss": 3.633824050426483, "best_checkpoint_path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints/best_loss_3p6338_epoch_0001_step_0003000.pt", "checkpoint_dir": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_135803/checkpoints", "stopped_early": false, "stop_reason": null, "run_dir": "runs/big_run/exp_c/run_20260408_135803"}